text
stringlengths 14
1.76M
|
|---|
# Where have all the solar-like stars gone? Rotation period detectability at
various inclinations and metallicities
Timo Reinhold Max-Planck-Institut für Sonnensystemforschung
Justus-von-Liebig-Weg 3, 37077 Göttingen, Germany Alexander I. Shapiro Max-
Planck-Institut für Sonnensystemforschung
Justus-von-Liebig-Weg 3, 37077 Göttingen, Germany Veronika Witzke Max-Planck-
Institut für Sonnensystemforschung
Justus-von-Liebig-Weg 3, 37077 Göttingen, Germany Nina-E. Nèmec Max-Planck-
Institut für Sonnensystemforschung
Justus-von-Liebig-Weg 3, 37077 Göttingen, Germany Institut für Astrophysik,
Georg-August-Universität Göttingen, Friedrich-Hund-Platz 1, D-37077 Göttingen,
Germany Emre Işık Dept. of Computer Science, Turkish-German University
Şahinkaya Cd. 108, Beykoz, 34820 Istanbul, Turkey Sami K. Solanki Max-Planck-
Institut für Sonnensystemforschung
Justus-von-Liebig-Weg 3, 37077 Göttingen, Germany School of Space Research,
Kyung Hee University,
Yongin, Gyeonggi, 446-701, Korea
###### Abstract
The plethora of photometric data collected by the Kepler space telescope has
promoted the detection of tens of thousands of stellar rotation periods.
However, these periods are not found to an equal extent among different
spectral types. Interestingly, early G-type stars with near-solar rotation
periods are strongly underrepresented among those stars with known rotation
periods. In this study we investigate whether the small number of such stars
can be explained by difficulties in the period determination from photometric
time series. For that purpose, we generate model light curves of early G-type
stars with solar rotation periods for different inclination angles,
metallicities and (magnitude-dependent) noise levels. We find that the
detectability is determined by the predominant type of activity (i.e. spot or
faculae domination) on the surface, which defines the degree of irregularity
of the light curve, and further depends on the level of photometric noise.
These two effects significantly complicate the period detection and explain
the lack of solar-like stars with known near-solar rotation periods. We
conclude that the rotation periods of the majority of solar-like stars with
near-solar rotation periods remain undetected to date. Finally, we promote the
use of new techniques to recover more periods of near-solar rotators.
††journal: ApJL
## 1 Introduction
Stellar brightness variations at the timescale of stellar rotation are caused
by transits of magnetic features (such as dark spots or bright faculae)
rotating across the visible disk. These variations have routinely been
observed by transit photometry missions. In particular, the Kepler telescope
obtained light curves of roughly 150,000 main-sequence stars. Some of these
light curves exhibit clear signatures of stellar rotation, which can be
extracted by standard frequency analysis tools such as Lomb-Scargle
periodograms, auto-correlation functions, or wavelet transforms. The biggest
survey of rotation periods based on the Kepler data has been published by
McQuillan et al. 2014 (hereafter MMA14), who detected rotation periods for
34,030 presumably main-sequence stars.
However, for the majority of the main-sequence stars the light curves were
either too noisy or too irregular for the rotation period to be determined.
MMA14 found that the fraction of stars with detectable periods strongly
depends on the effective temperature. Interestingly, this fraction appeared to
be lowest for stars with near-solar effective temperature (between 5500–6000
K, hereafter referred to as early G-type stars), reaching only 16% (see Table
3 in MMA14). On the contrary, van Saders et al. (2019) used Galactic evolution
models to predict that $\sim$59% of the early G-type stars should have
detectable rotation periods.
The lack of stars with known rotation period becomes even more severe for
early G-type stars of near-solar age. Recently, Reinhold et al. (2020) showed
that only a dozen Kepler stars with near-solar fundamental parameters and
rotation periods between 20–30 days (i.e. encompassing the solar rotation
period of $\sim$25 days) exhibit rotational variability levels similar to that
of the Sun. In contrast, the majority of these stars are substantially more
variable than the Sun and also show more regular light curves patterns. It has
been proposed that such a conspicuous difference between the Sun and other
solar-like stars can be explained by a detection bias towards more active
stars in bulk rotation period measurements (see the discussion in Amazo-Gómez
et al. 2020), thus missing the majority of early G-type stars with near-solar
rotation periods and small variabilities.
In this Letter we address the question whether the ”missing” solar-like stars
(i.e. stars with solar fundamental parameters and rotation periods) do not
exist or simply go undetected. Our approach is based on the solar paradigm,
i.e. we build on the comprehensive understanding of solar brightness
variability (see, e.g., reviews by Ermolli et al. 2013; Solanki et al. 2013),
and extend solar models to solar-like stars. Namely, we combine two recently
developed physics-based models by Witzke et al. (2020) and Nèmec et al.
(2020). This allows calculating light curves of stars with a solar
distribution of active regions and solar effective temperature, but various
metallicities and observed at arbitrary inclination angles. These light curves
are used to identify obstacles in the period determination of solar-like
stars, and to discuss possible limitations of period measurements in real data
sets. We further compare the number of actual period measurements in Kepler
data to predictions from Galactic evolution models using the detection rate
obtained from the model light curves.
## 2 Methods
### 2.1 The curious case of the Sun
The morphology of the solar light curve (as it would be observed in the Total
Solar Irradiance or in the broad-band spectral passband like those of CoRoT,
Kepler, or TESS) changes significantly depending on the phase of the solar
activity cycle. While it appears to be quite regular at 11-year cycle minima
when activity is low, the regularity disappears at periods of intermediate and
high solar activity (Lanza & Shkolnik, 2014; Aigrain et al., 2015; He et al.,
2015). In particular, Amazo-Gómez et al. (2020) showed that if the Sun were
observed by Kepler, the standard frequency analysis tools would most probably
fail to detect the correct rotation period (unless observations are done
during epochs of low solar activity). The causes for this inability are
manifold: solar rotational variability is mainly brought about by spots (see,
e.g., Shapiro et al. 2016). The relatively short lifetimes of sunspots from
days to weeks (see, e.g., Solanki 2003) implies that most of the spots transit
the visible solar disk only once, which leads to irregularities in the solar
light curve, and hampers the detection of the solar rotation period.
Furthermore, the brightness changes of dark spots and bright faculae partly
compensate each other, which decreases the amplitude of the rotational signal,
further hindering the period determination (Shapiro et al., 2017; Witzke et
al., 2020; Nèmec et al., 2020). The exception from this general tendency are
epochs of low solar activity with a small number of active regions. At these
times the rotational variability is attributed to long-lived facular features
and the light curve pattern becomes more periodic.
### 2.2 The model
While the irregularity of the solar light curve is quite well understood, the
situation gets more complicated for other early G-type stars. Their light
curves look different, partly because the stars are observed at various
inclinations. For example, faculae appear brighter at the limb and therefore
contribute more strongly to the variability when the star is observed out of
the ecliptic plane (Nèmec et al., 2020). Additionally, stellar metallicity
$\rm[Fe/H]$ affects facular (and to a smaller degree spot) contrasts (Witzke
et al., 2018), which eventually has an impact on the period detectability
(Witzke et al., 2020).
To synthesize the light curves of solar-like stars, we built on recent
calculations by Nèmec et al. (2020) and by Witzke et al. (2018). Nèmec et al.
(2020) utilized a semi-empirical sunspot-group record by Jiang et al. (2011)
and the Surface Flux Transport Model by Cameron et al. (2010) to reconstruct
the distribution of active regions on the solar surface from the year 2010
back to 1700 with a daily cadence. By applying an appropriate geometrical
transformation, Nèmec et al. (2020) calculated the distribution of active
regions on the solar disk as it would be observed at arbitrary inclinations.
Witzke et al. (2018) calculated the brightness contrasts of faculae and spots
relative to the quiet Sun (i.e. free from any apparent manifestations of
magnetic activity) as a function of wavelength and position on the visible
disk for stars with different metallicities and solar effective temperature.
All in all, by combining the reconstructed disk distribution of active regions
with their brightness contrasts, we generated light curves with a time span of
310 years as they would be seen in the passband of the Kepler telescope. The
light curves were calculated for ten inclination angles $0^{\circ}\leq i\leq
90^{\circ}$ (with a step of $10^{\circ}$), and nine different metallicities
$\rm-0.4\leq[Fe/H]\leq 0.4$ dex (with a step of 0.1 dex). The solar record
from 1700–2010 covers epochs of both low solar activity (like the Dalton
minimum around 1790–1830), and very high solar activity (like the modern grand
maximum around 1950–2000, see Solanki et al. 2004; Usoskin et al. 2007), which
allows studying rotation period detectability during activity cycles of very
different strengths. We note that, by assuming a solar disk distribution of
active regions, we only account for the metallicity effect on the contrasts of
magnetic features. A change in metallicity also affects the depth of the
convective zone, which in turn could influence the stellar dynamo, in
particular the length of the stellar activity cycle or the emergence latitudes
of magnetic bipoles (Schuessler & Solanki, 1992). However, this effects is
rather weak, e.g. doubling the metallicity of a star with solar temperature
will deepen the convective zone by only ca. 8% (van Saders & Pinsonneault,
2012; Karoff et al., 2018). Therefore, we expect these effects to be
relatively small. Studying them is beyond the scope of the present paper, but
would be an interesting future exercise.
### 2.3 Monte Carlo approach
We take a Monte Carlo approach to analyze light curves with different
realizations of inclination angles and metallicities. The distribution of
inclination angles is uniform in $\cos{i}$, where $i=0^{\circ}$ denotes a
pole-on view and $i=90^{\circ}$ an equator-on view. The input distribution of
metallicities was adapted for solar-like stars in the Kepler field (see Fig.
A.1 and Reinhold et al. 2020). For each (random) parameter combination
$(i,[{\rm Fe/H}])$, we chose the model light curve from the grid with the
closest parameters in metallicity and inclination angle.
Following the observing strategy of the Kepler telescope, we pick a random
4-year segment of the full time series (see top row of Fig. 1 for example).
This light curve is then cut into 90-day segments (i.e. similar to the Kepler
observing quarters), where each 90-day chunk is normalized by its median, and
appended to the previous one, to form a 4-year time series. These Keplerized
light curves (Fig. 1, middle row) will be analyzed for rotation in the next
step. The detrending is necessary because it filters out brightness variations
on the activity-cycle timescale, and renders the light curves comparable to
detrended Kepler data.
In addition to the various inclination and metallicity combinations, we study
the impact of noise on the period detectability. The model light curves are by
definition noise-free. In real observations, the visual stellar magnitude
defines the noise level. We use the distribution of Kepler magnitudes $Kp$ of
solar-like stars to compute different noise realizations $\sigma$ (see
Reinhold et al. 2020 for details). A noise time series with zero mean and
standard deviation $\sigma$ is then added to the time series in the Monte-
Carlo simulation. In total, we conducted 50,000 Monte Carlo runs, both for the
noise-free and the noisy cases to study them separately.
## 3 Results
### 3.1 Period detection
From among the various period detection methods, we chose the auto-correlation
function (ACF) to search for periodicity in the time series (i.e. the same
method as employed by MMA14). The ACF returns peaks of different power as a
measure of the periodicity in the light curve. To quantify the strength of the
periodicity, we adapt the measure of MMA14, where the local peak height (LPH)
is computed as the difference between the highest peak and the mean of the two
troughs on either side (see Fig. 1, bottom row). We only search for peaks at
periods less than 70 days, consistent with MMA14. If the highest ACF peak lies
between 24–30 days and $LPH>0.1$, we count it as a detection. If the peak lies
outside this period range or is smaller ($LPH<0.1$), it is counted as a false
or non-detection.
Fig. 1 illustrates the difficulty of detecting the correct rotation period of
the Sun from the photometric time series obtained in the Kepler passband. The
top row of Fig. 1 shows the modeled light curve computed for a star with solar
metallicity, $\rm[Fe/H]=0$, as it would be observed outside of its equatorial
plane at $i=40^{\circ}$. The bottom row gives the ACF and the computed LPH for
three different 4-year segments of the same light curve. Depending on the
selected segment, the ACF shows the highest peak at different periods. The
first panel shows a peak with a moderate LPH but outside the range of 24–30
days (red dashed lines), i.e. a false detection. The second panel shows a peak
close to the model rotation period of 27 days, although with a rather low LPH.
The last panel shows a clear peak within the range 24–30 days, although this
period is found to be the first harmonic of the highest peak at twice the
correct rotation period (so that this panel corresponds to a false detection
again). The light curve segment shown in this panel corresponds to an epoch of
relatively low magnetic activity when the rotational variability of the Sun
becomes faculae-dominated. Since faculae have significantly longer lifetimes
than spots, this segment shows a more stable periodicity but even in such
cases the correct rotation period is not necessarily associated with the
highest ACF peak.
We now consider how the apparent magnitude of a star affects the period
detection. For that purpose, Fig. 2 shows the same light curve as Fig. 1, but
with different noise levels to simulate the star as observed at different
magnitudes. To demonstrate the effect of noise on the ACF, we chose a segment
during solar minimum111The chosen segment slightly differs from the one chosen
in the third panel Fig. 1 to demonstrate the effect of noise on the LPH. where
the correct period was detected, and added Poisson noise to the light curve,
representative of a solar-like star at 11th, 13th, and 15th Kepler magnitude
(see Reinhold et al. 2020 for details). In all cases, the correct period was
detected. While from 11th to 13th magnitude the LPH only slightly decreases,
it decreases by more than half at 15th magnitude.
Figure 1: Top row: Model light curve (black) with an inclination of
$i=40^{\circ}$ and solar metallicity $\rm[Fe/H]=0$, with three randomly chosen
4-year segments (green). Middle row: Keplerized light curves of the chosen
segment from the top row (see Sect. 2.3 for details). Bottom row: auto-
correlation function (ACF) of the selected 4-year segment. The measured period
is indicated by the red asterisk, and the local peak height (LPH) is shown as
the vertical gray line between the peak and the two troughs on either side.
The vertical dashed red lines indicate the period detection window from 24 to
30 days. Figure 2: Same as Fig. 1 but choosing the same light curve segment
for three different noise realizations corresponding to 11th (left), 13th
(middle), and 15th (right) Kepler magnitude. The LPH decreases towards fainter
stars.
### 3.2 LPH dependence on spot area
The two examples in Sect. 3.1 illustrated how the period detection is affected
by the activity level (Fig. 1) and the amount of observational noise (Fig. 2).
Fig. 3 combines both of these effects by showing the LPH when the highest peak
was found between 24–30 days, as a function of the sunspot coverage on the
visible solar disc, averaged over the 4-year segment (in ppm), for different
magnitudes. The red diamonds show the median LPH values for the selected spot
area bins to better illustrate the LPH dependence.
The upper left panel shows only stars brighter than 12th Kepler magnitude
(i.e. with small noise levels). The LPH increases with decreasing spot area.
As mentioned above, small spot coverages are typically found during activity
minima when variability becomes faculae-dominated. Consequently, the light
curves become more regular. A similar trend is found for stars between 12th
and 13th magnitude but with larger scatter (upper right panel). Between 13th
to 14th magnitude (lower left panel), the noise level becomes comparable to
the variability amplitude during epochs of small spot coverages. As a result,
the LPH drops for small coverages and the increase of the LPH with decreasing
spot area can only be identified down to spot areas of 100 ppm. For the
faintest stars down to 15th magnitude (lower right panel), the larger noise
further decreases the LPH for small spot areas, and for spot areas above 100
ppm no trend can be identified any longer.
Figure 3: Local peak height (LPH) vs. spot area fraction for different Kepler
magnitudes. The red diamonds show the median LPH values for the selected spot
area bins. The very few peaks with $\rm LPH<0.01$ were excluded from the
analysis.
### 3.3 Period distribution
As shown in the previous section, the position of the highest peak and the
associated LPH determine the period detection. The percentages of correct and
false detections are given in Table 1. The period distribution is shown in
Fig. 4 for different LPH constraints for the noise-free (black) and the noisy
(red) case. From the upper left to the lower right panel, the LPH threshold
increases from 0.1 to 0.4. Consequently, the detection fraction decreases, but
also the number of false detections drops. The decrease of detections is even
stronger for the noisy case. We note that also the false detections decrease
more strongly for the noisy case (see Table 1). This is caused by the fact
that a peak is more easily found in the noise-free case, but the associated
period lies outside the range of 24–30 days, and is therefore counted as false
detection.
When measuring rotation periods in real data, the period is a priori unknown,
and one has to assign a certain LPH threshold, for which periods are
considered as significant. The upper left panel in Fig. 4 shows that even for
small values ($0.1<LPH<0.2$), most detections are found at the correct (model)
rotation period of 27 days. However, the number of false detections is also
quite high (see Table 1). Further increasing the LPH threshold significantly
decreases the number of false detections but also lowers the number of correct
detections. Finding an optimal LPH threshold that compromises between
discarding correct detections and not having too many false periods is non-
trivial. We stress that MMA14 required $LPH>0.3$ to count the period as a real
detection. As seen in the lower left panel, this threshold eliminates almost
all false detections but strongly decreases the number of real detections (see
discussion below).
Figure 4: Rotation period distribution for different local peak heights of the noise-free (black) and noisy (red) cases. The blue dashed-dotted line indicates the model rotation period of 27 days. LPH | Detection | False detection
---|---|---
| Noise-free | Noisy | Noise-free | Noisy
$>$0.1 | 23.8% | 17.3% | 35.2% | 27.4%
$>$0.2 | 15.5% | 7.4% | 9.0% | 4.5%
$>$0.3 | 9.5% | 2.9% | 1.9% | 0.5%
$>$0.4 | 6.1% | 1.0% | 0.4% | 0.0%
Table 1: Detections and false detections for different LPH for both the noise-
free and noisy cases.
### 3.4 Detection rate
We now turn to the question how the period detectability is affected by
stellar inclination and metallicity. For that purpose, we define the detection
rate as the number of detections divided by the number of different Monte
Carlo runs at a given parameter. In Fig. 5 we show the detection rate as a
function of the inclination of the rotation axis of the model star (integrated
over all metallicities) for different LPH values. The error bars indicate the
square root of the number of detections divided by the number of models. As
before, we consider the noise-free (left panel) and the noisy (right panel)
cases separately.
The noise-free case qualitatively displays the same behavior for all LPH
thresholds. As expected, the detection rate is zero for the pole-on view.
However, when increasing the inclination angle the detection rate steeply
increases to the maximum at an inclination near $20^{\circ}$, and gradually
decreases towards the equator-on view at $90^{\circ}$. This result might be
surprising at first glance but can be explained by the dominant contribution
of faculae to brightness variability for stars with near-equatorial activity
belts (similar to those on the Sun) observed close to the pole-on view
(Shapiro et al., 2016). In such stars each facular feature spends roughly half
a rotation period on the far-side of a star and the remaining half of the time
near the limb on the visible disk. Faculae appear especially bright near the
limb, and usually last for several stellar rotations. Consequently, the light
curves of such stars appear more regular, leading to higher LPH values. We
note that the calculations are performed assuming a solar latitudinal
distribution of active regions. A change of the distribution would affect the
visibility of the active regions, and consequently, the inclination angle
corresponding to the maximum of the detection rate. However, we expect that a
solar distribution is typical for stars with solar rotation period and
temperature (see Sect. 2.2).
In the noisy case (right panel) the detection rates are generally smaller (cf.
Table 1 and Fig. 4). While the curves have a similar shape to the noise-free
cases, their maxima are shifted to higher inclinations. Such a shift is caused
by the decrease of the amplitude of brightness variability with decreasing
inclination (Nèmec et al., 2020), and consequently a decrease of the signal-
to-noise ratio. Consequently, the detection peak near $20^{\circ}$ is
suppressed, leaving a residual peak near $40^{\circ}$. As already shown in
Figs. 3 and 4, the noise decreases the LPH such that only a few cases with
$LPH>0.3$ remain.
Figure 5: The detection rate as a function of inclination angle, integrated
over all metallicities, for different LPH thresholds. The error bars show the
square root of the number of detections divided by the number of models at a
given inclination.
In Fig. 6 we show the detection rate as a function of metallicity (integrated
over all inclinations) for different LPH values. We note that the qualitative
shape of the $LPH>0.1$ curve is consistent with the one found in Witzke et al.
(2020), who considered the noise-free case (see Sect. C in the appendix for
the difference between the calculations in this study and those employed in
Witzke et al. 2020). Fig. 6 shows that for both the noise-free (left panel)
and the noisy (right panel) case, the detection rate increases with
metallicity. This is caused by the stronger contribution of the faculae to the
overall variability. Only for the cases $LPH>0.1$ and $LPH>0.2$, does the
detection rate show a minimum at $\rm[Fe/H]=-0.3$ dex or $-0.2$ dex (noisy),
increasing again towards smaller metallicity values. We expect that this trend
continues towards even smaller metallicities.
Figure 6: The detection rate as a function of metallicity, integrated over all
inclinations, for different LPH thresholds. The error bars show the square
root of the number of detections divided by the number of models at a given
metallicity.
### 3.5 Comparison with observations
We now compare our detection rates (see Table 1) to period detections of
solar-like stars in the MMA14 sample. As mentioned above, MMA14 used a
relatively conservative detection threshold of $LPH>0.3$. The bottom left
panel of Fig. 4 and Table 1 indicate that this threshold represents only the
tip of the iceberg: for $LPH>0.3$ (noisy case), the rotation periods can be
correctly detected for only 2.9% of our modeled light curves.
Galactic evolution models (van Saders et al., 2019) predict that 16% of the
(dwarf) stars in the Kepler field with effective temperatures $5500<T_{\rm
eff}<6000$ K should have rotation periods between 24–30 days (van Saders,
private communications). Using the latest Kepler parameter catalog (Mathur et
al., 2017), we select stars in this temperature range, with surface gravities
$\log g>4.2$ to exclude more evolved stars, and brighter than 15th Kepler
magnitude (following the selection criteria used in Reinhold et al. 2020). We
further restrict the catalog metallicities to $\rm-0.45<[Fe/H]<0.45$ dex,
which corresponds to the range of simulated metallicities (see Fig. A.1).
Selecting such stars from tables 1 and 2 in MMA14 yields $N=16890$ stars.
Among those, only 16% will have periods between 24–30 days, and according to
our analysis, only 2.9% of these stars will have detectable periods. Thus, we
estimate that $N_{det}=N*0.16*0.029=78$ stars should have detectable periods.
MMA14 found 455 stars in this parameter range with periods $24\leq P_{\rm
rot}\leq 30$ days. However, the vast majority of these stars exhibits
variability levels much higher than the Sun, and represent a regime of
variability very different from that of the Sun (Reinhold et al., 2020; Zhang
et al., 2020; Isık et al., 2020). Consequently, the light curves of these
stars cannot be accounted for by our model. To correct for such stars, we
followed the approach of Witzke et al. (2020) and selected the stars with
variability (regressed to solar values of effective temperature, metallicity,
and rotation period, see Fig. S8 and its detailed discussion in Reinhold et
al. 2020) below 0.18%, which corresponds to the maximum variability of the Sun
over the last 140 years. All in all, only 73 out of the 455 stars satisfied
such a criterion. This number is gratifyingly close to our estimate of 78
stars.
## 4 Conclusions
In this study we identified biases in the period determination of stars with
solar-like variability. The detection rates among these stars are lower than
for stars of other spectral types. In particular, only 2.9% of them would be
detectable using the thresholds set in MMA14. This is mainly caused by the
small variability amplitudes of the rotational tracers and their relatively
short lifetimes compared to the rotation period.
The very low detection rate explains the large discrepancy between the number
of measured rotation periods (MMA14), and those predicted by Galactic
evolution models (van Saders et al., 2019). The predicted number of stars with
detectable periods (78), and that for which rotation periods have actually
been measured (73), is remarkably similar. Fig. 4 shows that many more
rotation periods of solar-like stars may be measured when lowering the
thresholds in the automated period surveys. However, this will also add a
number of false periods, depending on how the thresholds are set.
Our study revealed that the rotation periods of most solar-like stars will go
undetected using standard frequency analysis tools. Thus, we emphasize the
importance of alternative methods for period detection such as the GPS method
(Shapiro et al., 2020; Amazo-Gómez et al., 2020) or new approaches based on
Gaussian process regression (Foreman-Mackey et al., 2017; Angus et al., 2018;
Kosiarek & Crossfield, 2020).
We would like to thank Jennifer van Saders for providing model numbers and for
helpful discussion. This work has received funding from the European Research
Council (ERC) under the European Union’s Horizon 2020 research and innovation
programme (grant agreement No. 715947). This work has been partially supported
by the BK21 plus programme through the National Research Foundation (NRF)
funded by the Ministry of Education of Korea. We would like to thank the
International Space Science Institute, Bern, for their support of science team
446 and the resulting helpful discussions.
## Appendix A Monte Carlo input distribution
Fig. A.1 shows the distributions of input parameters used in the Monte Carlo
simulation. The first panel shows the distribution of inclination angles. It
can be shown that isotropic inclination angles $i$ exhibit a uniform
distribution in $\cos i$ (see e.g.
http://keatonb.github.io/archivers/uniforminclination for a detailed
derivation). The last bin of the distribution ($85-90^{\circ}$) only contains
half the number of realizations because no inclination angles greater than
$90^{\circ}$ exist. The same argument applies to the first bin from
$0-5^{\circ}$.
The middle panel shows the distribution of metallicities of the solar-like
stars in the Kepler field. The catalog values were adapted from Mathur et al.
(2017) and the selection of solar-like stars can be found in Reinhold et al.
(2020). We note that the Sun ($\rm[Fe/H]=0$) is slightly more metal-rich than
the peak of the distribution.
The last panel shows the apparent magnitudes of the stars in the Kepler field.
It is obvious that the majority of stars is very faint. Since the stellar
magnitudes define the noise in the light curves, it is crucial to adapt this
distribution for the noise model (see Reinhold et al. 2020) to make realistic
predictions about stars in the Kepler field.
Figure A.1: Input distributions of inclination angles (left) and metallicities
(middle), and Kepler magnitudes (right) of the Monte Carlo simulation.
## Appendix B Generating light curves
The total spectral flux at a certain time is composed of fluxes emerging from
surface areas with different levels of magnetic activity. Following the
detailed description in Shapiro et al. (2014), we decompose the spectral flux
from a star, $F$, into contributions from the quiet stellar region ($F_{\rm
Q}$), faculae ($F_{\rm fac}$), and spots ($F_{\rm spot}$):
$F(\lambda)=F_{\rm Q}(\lambda)+F_{\rm fac}(\lambda)+F_{\rm spot}(\lambda),$
(B1)
where $\lambda$ is the wavelength. For the quiet stellar region, the disc-
integrated flux $\rm F_{Q}(\lambda)$ is obtained by integrating
$F_{Q}(\lambda)=\int_{0}^{1}I_{Q}(\lambda,\mu)\omega(\mu)d\mu,$ (B2)
where $\omega(\mu)=2\pi\mu(r_{star}/d_{star})^{2}$ is a weighting function
with the stellar radius, $r_{star}$, and the distance between the star and the
observer, $d_{star}$. The emergent intensity, $I_{Q}(\lambda,\mu)$, also
depends on $\mu$, which is the cosine of the angle between the observer’s
direction and the local stellar radius. In this formulation the stellar disc
center is associated with $\mu=1$ and the limb with $\mu=0$.
Both faculae and spots are taken into account through their contrast with
respect to the quiet regions. Therefore, the contribution of faculae is
defined as
$F_{Fac}(\lambda)=\int_{0}^{1}\alpha_{Fac}(\mu)\left[I_{Fac}(\lambda,\mu)-I_{Q}(\lambda,\mu)\right]\omega(\mu)d\mu,$
(B3)
where the fractional coverage of the ring corresponding to a given $\mu$ by
faculae is given by the function $\alpha_{Fac}(\mu)$.
The contribution from spots consists of those from spot umbrae and spot
penumbrae:
$\displaystyle F_{\rm spot}(\lambda)$ $\displaystyle=$
$\displaystyle\int_{0}^{1}\alpha_{\rm pen}(\mu)\left(I_{\rm
pen}(\lambda,\mu)-I_{Q}(\lambda,\mu)\right)\omega(\mu)d\mu$ (B4)
$\displaystyle+$ $\displaystyle\int_{0}^{1}\alpha_{\rm umb}(\mu)\left(I_{\rm
umb}(\lambda,\mu)-I_{Q}(\lambda,\mu)\right)\omega(\mu)d\mu,$
where $I_{\rm umb}$ and $I_{\rm pen}$ are the emergent intensities from the
spot umbrae and spot penumbrae, respectively, and the $\alpha_{\rm umb}$ and
$\alpha_{\rm pen}$ denote the corresponding surface coverages.
The surface coverages for the magnetic features (i.e. $\alpha_{\rm Fac}$,
$\alpha_{\rm umb}$ and $\alpha_{\rm pen}$) used in this work are taking from
Nèmec et al. (2020). Furthermore, calculations of the emergent intensities for
all stellar regions follow the approach used in Witzke et al. (2020), but with
a small modification which is explained in Appendix C.
## Appendix C Calculating emergent intensities
The model atmospheres and corresponding emergent spectra computed by Unruh et
al. (1999) for the solar faculae, spots, and quiet regions (hereafter,
original models) proved to be very successful in reproducing the solar
brightness variations with high accuracy (Krivova et al., 2003; Solanki et
al., 2013; Ermolli et al., 2013).
Here, we extend the intensities for different surface components to different
metallicities following the approach outlined in Witzke et al. (2018, 2020),
but with slight modifications to cover a broader metallicity range. In our
modeling approach, we aim to match the intensity contrasts for the solar
metallicity as closely as possible to the original models. Thus, in the first
step we searched for the model parameters (input parameters for calculating
stellar atmospheres with ATLAS9), such as convection settings, surface gravity
and continuum opacity sources for the quiet Sun, spot-umbra and spot-penumbra,
to match the original models and spectra by Unruh et al. (1999). The closest
match is achieved by the parameters listed in Table 2. Then for the generation
of the faculae model we assumed that the temperature difference,
$\Delta\text{T}$, and pressure difference, $\Delta P$, as a function of column
mass between the original facular and quiet Sun models are the same as between
our new facular and quiet Sun models.
Finally, to calculate atmospheric models for different metallicity values, we
first generated atmosphere models for the quiet regions and the spots assuming
radiative equilibrium. Then we followed up on the Witzke et al. (2018)
approach and assumed that a change of the metallicity value has the same
effect on the temperature and pressure structures of the quiet Sun and
faculae. Using the quiet stellar atmosphere models for different
metallicities, we applied the solar $\Delta\text{T}$ and $\Delta P$ with
column mass to calculate the facular models. Using these atmospheric models
for the quiet regions and magnetic features, we generated the emergent
intensities $I_{\lambda,\mu}$ for each metallicity value using the MPS-ATLAS
code (Witzke et al., in prep.).
model | effective | surface | mixing | over
---|---|---|---|---
| temperature [K] | gravity | -length | -shoot
Quiet region | 5777 | 4.43777 | 1.25 | on
Spot umbra | 4500 | 4.0 | 1.25 | on
Spot penumbra | 5450 | 4.0 | 1.25 | on
Table 2: Input parameters for model atmospheres in radiative equilibrium.
## References
* Aigrain et al. (2015) Aigrain, S., Llama, J., Ceillier, T., et al. 2015, MNRAS, 450, 3211, doi: 10.1093/mnras/stv853
* Amazo-Gómez et al. (2020) Amazo-Gómez, E. M., Shapiro, A. I., Solanki, S. K., et al. 2020, A&A, 636, A69, doi: 10.1051/0004-6361/201936925
* Angus et al. (2018) Angus, R., Morton, T., Aigrain, S., Foreman-Mackey, D., & Rajpaul, V. 2018, MNRAS, 474, 2094, doi: 10.1093/mnras/stx2109
* Cameron et al. (2010) Cameron, R. H., Jiang, J., Schmitt, D., & Schüssler, M. 2010, ApJ, 719, 264, doi: 10.1088/0004-637X/719/1/264
* Ermolli et al. (2013) Ermolli, I., Matthes, K., Dudok de Wit, T., et al. 2013, Atmosph. Chem. Phys., 13, 3945, doi: 10.5194/acp-13-3945-2013
* Foreman-Mackey et al. (2017) Foreman-Mackey, D., Agol, E., Ambikasaran, S., & Angus, R. 2017, AJ, 154, 220, doi: 10.3847/1538-3881/aa9332
* He et al. (2015) He, H., Wang, H., & Yun, D. 2015, ApJS, 221, 18, doi: 10.1088/0067-0049/221/1/18
* Isık et al. (2020) Isık, E., Shapiro, A. I., Solanki, S. K., & Krivova, N. A. 2020, ApJ, 901, L12, doi: 10.3847/2041-8213/abb409
* Jiang et al. (2011) Jiang, J., Cameron, R. H., Schmitt, D., & Schüssler, M. 2011, A&A, 528, A82, doi: 10.1051/0004-6361/201016167
* Karoff et al. (2018) Karoff, C., Metcalfe, T. S., Santos, Â. R. G., et al. 2018, ApJ, 852, 46, doi: 10.3847/1538-4357/aaa026
* Kosiarek & Crossfield (2020) Kosiarek, M. R., & Crossfield, I. J. M. 2020, AJ, 159, 271, doi: 10.3847/1538-3881/ab8d3a
* Krivova et al. (2003) Krivova, N. A., Solanki, S. K., Fligge, M., & Unruh, Y. C. 2003, A&A, 399, L1, doi: 10.1051/0004-6361:20030029
* Lanza & Shkolnik (2014) Lanza, A. F., & Shkolnik, E. L. 2014, MNRAS, 443, 1451, doi: 10.1093/mnras/stu1206
* Mathur et al. (2017) Mathur, S., Huber, D., Batalha, N. M., et al. 2017, Astrophys. J., Suppl. Ser., 229, 30, doi: 10.3847/1538-4365/229/2/30
* McQuillan et al. (2014) McQuillan, A., Mazeh, T., & Aigrain, S. 2014, Astrophys. J., Suppl. Ser., 211, 24, doi: 10.1088/0067-0049/211/2/24
* Nèmec et al. (2020) Nèmec, N. E., Shapiro, A. I., Krivova, N. A., et al. 2020, A&A, 636, A43, doi: 10.1051/0004-6361/202037588
* Reinhold et al. (2020) Reinhold, T., Shapiro, A. I., Solanki, S. K., et al. 2020, Science, 368, 518. https://arxiv.org/abs/2005.01401
* Schuessler & Solanki (1992) Schuessler, M., & Solanki, S. K. 1992, A&A, 264, L13
* Shapiro et al. (2020) Shapiro, A. I., Amazo-Gómez, E. M., Krivova, N. A., & Solanki, S. K. 2020, A&A, 633, A32, doi: 10.1051/0004-6361/201936018
* Shapiro et al. (2017) Shapiro, A. I., Solanki, S. K., Krivova, N. A., et al. 2017, Nature Astronomy, 1, 612, doi: 10.1038/s41550-017-0217-y
* Shapiro et al. (2014) —. 2014, A&A, 569, A38, doi: 10.1051/0004-6361/201323086
* Shapiro et al. (2016) Shapiro, A. I., Solanki, S. K., Krivova, N. A., Yeo, K. L., & Schmutz, W. K. 2016, A&A, 589, A46, doi: 10.1051/0004-6361/201527527
* Solanki (2003) Solanki, S. K. 2003, A&A Rev., 11, 153, doi: 10.1007/s00159-003-0018-4
* Solanki et al. (2013) Solanki, S. K., Krivova, N. A., & Haigh, J. D. 2013, Ann. Rev. Astron. Astrophys., 51, 311, doi: 10.1146/annurev-astro-082812-141007
* Solanki et al. (2004) Solanki, S. K., Usoskin, I. G., Kromer, B., Schüssler, M., & Beer, J. 2004, Nature, 431, 1084, doi: 10.1038/nature02995
* Unruh et al. (1999) Unruh, Y. C., Solanki, S. K., & Fligge, M. 1999, A&A, 345, 635
* Usoskin et al. (2007) Usoskin, I. G., Solanki, S. K., & Kovaltsov, G. A. 2007, A&A, 471, 301, doi: 10.1051/0004-6361:20077704
* van Saders & Pinsonneault (2012) van Saders, J. L., & Pinsonneault, M. H. 2012, ApJ, 746, 16, doi: 10.1088/0004-637X/746/1/16
* van Saders et al. (2019) van Saders, J. L., Pinsonneault, M. H., & Barbieri, M. 2019, ApJ, 872, 128, doi: 10.3847/1538-4357/aafafe
* Witzke et al. (2020) Witzke, V., Reinhold, T., Shapiro, A. I., Krivova, N. A., & Solanki, S. K. 2020, A&A, 634, L9, doi: 10.1051/0004-6361/201936608
* Witzke et al. (in prep.) Witzke, V., Shapiro, A. I., Cernetic, M., et al. in prep., A&A
* Witzke et al. (2018) Witzke, V., Shapiro, A. I., Solanki, S. K., Krivova, N. A., & Schmutz, W. 2018, A&A, 619, A146, doi: 10.1051/0004-6361/201833936
* Zhang et al. (2020) Zhang, J., Shapiro, A. I., Bi, S., et al. 2020, ApJ, 894, L11, doi: 10.3847/2041-8213/ab8795
|
# One Model to Serve All: Star Topology Adaptive Recommender for Multi-Domain
CTR Prediction
Xiang-Rong Sheng, Liqin Zhao and Guorui Zhou, Xinyao Ding, Binding Dai,
Qiang Luo
Siran Yang, Jingshan Lv, Chi Zhang, Hongbo Deng, Xiaoqiang Zhu Alibaba
GroupBeijing, China
(2021)
###### Abstract.
Traditional industry recommendation systems usually use data in a single
domain to train models and then serve the domain. However, a large-scale
commercial platform often contains multiple domains, and its recommendation
system often needs to make click-through rate (CTR) predictions for multiple
domains. Generally, different domains may share some common user groups and
items, and each domain may have its own unique user groups and items.
Moreover, even the same user may have different behaviors in different
domains. In order to leverage all the data from different domains, a single
model can be trained to serve all domains. However, it is difficult for a
single model to capture the characteristics of various domains and serve all
domains well. On the other hand, training an individual model for each domain
separately does not fully use the data from all domains. In this paper, we
propose the Star Topology Adaptive Recommender (STAR) model to train a single
model to serve all domains by leveraging data from all domains simultaneously,
capturing the characteristics of each domain, and modeling the commonalities
between different domains. Essentially, the network of each domain consists of
two factorized networks: one centered network shared by all domains and the
domain-specific network tailored for each domain. For each domain, we combine
these two factorized networks and generate a unified network by element-wise
multiplying the weights of the shared network and those of the domain-specific
network, although these two factorized networks can be combined using other
functions, which is open for further research. Most importantly, STAR can
learn the shared network from all the data and adapt domain-specific
parameters according to the characteristics of each domain. The experimental
results from production data validate the superiority of the proposed STAR
model. Since late 2020, STAR has been deployed in the display advertising
system of Alibaba, obtaining 8.0% improvement on CTR and 6.0% increase on RPM
(Revenue Per Mille).
Multi-Domain Learning, Recommender System, Display Advertising
††copyright: acmcopyright††journalyear: 2021††copyright:
acmcopyright††conference: Proceedings of the 30th ACM International Conference
on Information and Knowledge Management; November 1–5, 2021; Virtual Event,
QLD, Australia††booktitle: Proceedings of the 30th ACM International
Conference on Information and Knowledge Management (CIKM ’21), November 1–5,
2021, Virtual Event, QLD, Australia††price: 15.00††doi:
10.1145/3459637.3481941††isbn: 978-1-4503-8446-9/21/11††ccs: Information
systems Information retrieval Figure 1. Two representative business domains,
Banner and Guess What You Like on Taobao mobile app home. A business domain is
referred to as a specific spot that items are presented to users in the mobile
app and PC websites.
## 1\. Introduction
Traditional CTR prediction models (Zhou et al., 2018, 2019; Rendle, 2010; Guo
et al., 2017b; Cheng et al., 2016) focus on single-domain prediction, where
the CTR model serves for a single business domain after trained with examples
collected from this domain. Each business domain is a specific spot that items
are presented to users on the mobile app or PC websites. At large commercial
companies like Alibaba and Amazon, there are often many business domains that
need CTR prediction to enhance user satisfaction and improve business revenue.
For example, in Alibaba, the business domains range from Guess What You Like
in Taobao App homepage, Banner of Taobao App homepage to other domains (Zhu et
al., 2017). Figure 1 shows two representative business domains in Alibaba.
* •
Banner: In banner, the items to be recommended appears in the top banner of
the Taobao home page. The item can be a single commodity, a store, or a brand.
* •
Guess What You Like: In Guess What You Like, items are all single commodities
and displayed to users in the left or right column.
Since different business domains have overlapping user groups and items, there
exist commonalities among these domains. Enabling information sharing is
beneficial for learning the CTR model of each domain. However, the specific
user group may be different and the users’ behaviors also change in various
domains. These distinctions result in domain-specific data distributions.
Simply mixing all the data and training a single shared CTR model can not work
well on all domains.
Besides mixing data and training a shared model, another simple solution is to
build a separate model per business domain. This strategy also has some
downsides: (1) some business domains have much less data than other domains.
Splitting the data neglects the domain commonalities and causes much less
training data, making the models hard to learn. (2) Maintaining multiple
models cause a tremendous amount of resource consumption and require much more
human cost. It is unduly burdensome when the number of business domains is up
to hundreds.
Figure 2. (a): Single shared model for all domains, square nodes indicate the
shared model. (b): One model per domain where each model is learned
separately. The circle node indicates the domain-specific model. (c): The
proposed Star Topology Adaptive Recommender (STAR), where each domain has
specific parameters and also shares a common centered model. The edges mean
the combination of the center shared parameters with the domain-specific
parameters.
This paper aims to learn an effective and efficient CTR model to handle
multiple domains simultaneously. We formulate multi-domain CTR prediction as
the problem that the recommender needs to make CTR prediction for $M$ business
domains $D_{1},D_{2},\dots,D_{M}$ simultaneously. The model takes input as
$({\bf x},y,p)$, where ${\bf x}$ is the common feature used by multiple
business domains like historical user behavior, user profile feature, item
feature, and context feature. $y\in\\{0,1\\}$ is the clicked label, and $p$ is
the domain indicator that indicates which domain this sample is collected.
Note that $({\bf x},y)$ is drawn from the domain-specific distribution
$D_{p}$, and the distribution varies with different domains. Multi-domain CTR
prediction aims to construct an effective and efficient model that gives
accurate CTR prediction for each domain and at a trivial cost on resource
consumption. To achieve this goal, the model should make full use of the
domain commonalities and capture the domain distinction.
One possible strategy to improve learning with multiple domains is multi-task
learning (Ruder, 2017; Caruana, 1998; Ma et al., 2018b). As shown in Figure 3,
the difference between multi-domain CTR prediction and multi-task learning is
that multi-domain CTR prediction solves the same task, i.e., CTR prediction,
across different domains, in which the label spaces of different domains are
the same and the data distribution is different. By contrast, most multi-task
learning approaches (Ma et al., 2018b; Misra et al., 2016; Ma et al., 2018a,
2019; Tang et al., 2020) address various tasks in the same domain, where the
label space might be different, e.g., jointly estimate CTR and conversion rate
(CVR) (Ma et al., 2018a; Wen et al., 2020). Due to the heterogeneity of tasks,
existing multi-task learning approaches focus on sharing information in the
bottom layers but keeping separate task-specific output layers (Ruder, 2017).
Directly adapting multi-task approaches to multi-domain CTR prediction can not
sufficiently exploit the domain relationship in the label space and neglect
the distinct data distribution of different domains.
Figure 3. Comparison of multi-task learning with multi-domain learning. Most
multi-task learning approaches focus on tackling different tasks within a
single domain. In contrast, multi-domain learning makes predictions for
multiple domains addressing the same task, e.g., CTR prediction, where the
label spaces are of the same. Directly adapting multi-task approaches to
multi-domain CTR prediction can not sufficiently exploit the domain
relationship in the label space and neglects the distinct data distribution of
different domains.
To fully exploit the domain relationship, we propose Star Topology Adaptive
Recommender (STAR) for multi-domain CTR prediction. The proposed STAR model
has the star topology, as illustrated in Figure 4. STAR consists of shared
centered parameters and multiple sets of domain-specific parameters. The final
model of each domain is obtained by combining the shared centered parameters
and the domain-specific parameters. The centered parameters are used to learn
general behaviors among all domains, in which the common knowledge can be
learned and transferred among all domains. The domain-specific parameters
capture specific behaviors in different domains to facilitate more refined CTR
prediction. The star topology facilitates effective information transformation
across multiple domains to learn domain commonalities and distinctions. This
paper implements the STAR model with the element-wise product of weights in
each layer as the combination strategy. Since embedding layers contribute most
parameters in industrial recommender, the added domain-specific parameters are
negligible to the total amount of parameters. Thus, using the STAR model to
serve multiple domains only adds little computational and memory costs while
yielding much better performance.
The main contributions of this work can be summarized as follows:
* •
We propose Star Topology Adaptive Recommender (STAR) to tackle multi-domain
CTR prediction. The star topology facilitates effective information
transformation across multiple domains to learn domain commonalities while
capturing domain distinction.
* •
Different domains have different data distributions, this leads to inaccurate
statistics when using batch normalization. We propose Partitioned
Normalization (PN) that privatizes normalization for examples from different
domains to address this issue. PN can lead to more accurate moments within the
domain, which improves model performance.
* •
In multi-domain CTR prediction, features that depict the domain information
are of importance. We propose an auxiliary network that treats the domain
indicator directly as the input and learns its embeddings to depict the
domain. The embeddings are then fed to the auxiliary network, which is much
simpler than the original network. This makes the domain indicator influence
the final prediction in a direct manner.
* •
We evaluate STAR on the industrial production dataset and deploy it in the
display advertising system of Alibaba in 2020. The consistent superiority
validates the efficacy of STAR. Up to now, the deployment of STAR brings 6%
CTR and 8% RPM lift. We believe the lessons learned in our deployment
generalize to other setups and are thus of interest to researchers and
industrial practitioners.
## 2\. Related Work
Our work is closely related to traditional single-domain CTR prediction, where
the recommender is trained on a single business domain and then serve for this
business domain. Besides, our work is also related to multi-task learning and
multi-domain learning. In this section, we give a brief introduction.
### 2.1. Single-Domain CTR Prediction
Inspired by the success within deep learning, recent CTR prediction model has
made the transition from traditional shallow approaches (Friedman, 2001;
Rendle, 2010; Koren, 2008; Koren et al., 2009; Zhou et al., 2008) to modern
deep approaches (Guo et al., 2017b; Cheng et al., 2016; Qu et al., 2016; Zhou
et al., 2018, 2019; Pi et al., 2019). Most deep CTR models follow the
embedding and MLP paradigm. Wide & Deep (Cheng et al., 2016) and deepFM (Guo
et al., 2017b) combine low-order and high-order features to improve the
expression power of the model. PNN (Qu et al., 2016) introduces a product
layer to capture interactive patterns between inter-field categories. In these
models, the user’s history behaviors are transformed into low-dimensional
vectors after the embedding and pooling. DIN (Zhou et al., 2018) employs the
mechanism of attention to activate historical behaviors locally w.r.t. the
given the target item, and successfully captures the diversity characteristic
of user interest. DIEN (Zhou et al., 2019) further proposes an auxiliary loss
to capture latent interest from historical behaviors. Additionally, DIEN
integrates the attention mechanism with GRU to model the dynamic evolution of
user interest. MIND (Li et al., 2019) and DMIN (Xiao et al., 2020) argue that
a single vector might be insufficient to capture complicated pattern lying in
the user and items. Capsule network and the dynamic routing mechanism are
introduced in MIND to learn multiple representations to aggregate raw
features. Moreover, inspired by the success of the self-attention architecture
in the tasks of sequence to sequence learning (Vaswani et al., 2017),
Transformer is introduced in (Feng et al., 2019) for feature aggregation. MIMN
(Pi et al., 2019) proposes a memory-based architecture to aggregate features
and tackle the challenge of long-term user interest modeling. SIM (Pi et al.,
2020) extracts user interests with two cascaded search units, which achieves
better ability to model lifelong sequential behavior data in both scalability
and accuracy.
### 2.2. Multi-Task Learning
Multi-task learning (Caruana, 1998; Ruder, 2017) aims to improve
generalization by sharing knowledge across multiple related tasks. The shared
knowledge and task-specific knowledge are explored to facilitate the learning
of each task. Multi-task learning has been used successfully on multiple
application domains, ranging from natural language processing (Collobert and
Weston, 2008), speech recognition (Deng et al., 2013), recommender system
(Yuan et al., 2020) to computer vision (Kendall et al., 2018). In early
literature on MTL for linear models, Argyriou et al. (2008) propose a method
to learn sparse representations shared across multiple tasks. In the context
of deep learning, multi-task learning is typically done with parameter sharing
of hidden layers (Caruana, 1998; Ma et al., 2018a). Misra et al. (2016)
propose cross-stitch units to learn unique combinations of task-specific
hidden-layers for each task. Ma et al. (2018b) proposes Multi-gate Mixture-of-
Experts (MMoE) to model task relationships by sharing the expert sub-models
across all tasks, while also having a gating network trained to optimize each
task. Kendall et al. (2018) propose a principled approach to multi-task deep
learning which weighs multiple loss functions by considering the homoscedastic
uncertainty of each task. In multi-task learning, different tasks may
conflict, necessitating a trade-off, optimize a proxy objective that minimizes
a weighted linear combination of per-task losses may not be optimal. To
address this issue, Sener and Koltun (2018) explicitly cast multi-task
learning as multi-objective optimization, with the overall objective of
finding a Pareto optimal solution. Note that (Kendall et al., 2018; Sener and
Koltun, 2018) are complementary to this work and could be potentially combined
to achieve better performance.
### 2.3. Multi-Domain Learning
In real-world applications, it is oftentimes that the data are collected from
multiple domains (Dredze et al., 2010; Joshi et al., 2012; Li et al., 2020).
Multi-domain learning enables knowledge transfer between domains to improve
learning. As such, it contrasts with the domain adaptation (DA) problem
(Bickel et al., 2007; Ben-David et al., 2010), where knowledge transfer is
only one way, i.e., from the source domain to the target domain. Wang et al.
(2019) propose Transferable Normalization in place of existing normalization
techniques for domain adaptation and reveals that BN (Ioffe and Szegedy, 2015)
is the constraint of transferability.
Multi-domain CTR prediction can be seen as a special kind of multi-domain
learning problem, in which each domain corresponds to a business domain and
the task is the CTR prediction. Compared with traditional multi-domain
learning, our work focuses on CTR prediction. The proposed model makes full
use of the domain indicator that is directly fed as the ID feature and
learning its semantic embeddings to facilitates the model learning, which is
neglected by previous literature. The difference between multi-domain learning
and multi-task learning is that multi-domain learning makes prediction for
multiple domains addressing the same problem, e.g., CTR prediction, where the
label spaces are of the same. In contrast, multi-task learning focuses on
tackling different problems (Yang and Hospedales, 2015). For example, in the
field of video recommendation, a multi-task learning problem can be as
simultaneously predicting CTR and expected watch time of videos for a single
business domain and multi-domain CTR prediction makes CTR predictions for
multiple business domains, e.g., multiple video platforms.
## 3\. The Proposed Approach
Figure 4. Comparison of model for single-domain CTR prediction and the Star
Topology Adaptive Recommender (STAR) for multi-domain CTR prediction. In STAR,
the partitioned normalization (PN) privatizes normalization for examples from
different domains. The normalized features are then fed as input to the
following star topology fully-connected neural network (star topology FCN).
The star topology FCN consists of shared centered FCN and multiple domain-
specific FCNs. The final combined model of each domain is obtained by the
element-wise product of weights in each layer.
In this section, we first give a brief introduction about the background of
multi-domain CTR prediction. Next is the architecture overview of the proposed
method, star topology adaptive recommender (STAR) for multi-domain CTR
prediction. Then we introduce STAR in detail, including the proposed star
topology network, partitioned normalization, and auxiliary network.
### 3.1. Multi-Domain CTR Prediction
In sequential recommender systems, the model takes input as the user
historical behavior, user profile feature, target item feature, and other
features like context feature. The predicted CTR $\hat{y}$ of a user $u$
clicking on an item $m$ is calculated via:
$\hat{y}=\mathrm{f}(E(u_{1}),\dots,E(u_{i});E(m_{1}),\dots,E(m_{j});E(c_{1}),\dots,E(c_{k})),$
where $\\{u_{1},\dots,u_{i}\\}$ is the set of user features including user
historical behavior and user profile feature. $\\{m_{1},\dots,m_{j}\\}$ is the
set of target item feature and $\\{c_{1},\dots,c_{k}\\}$ is the set of other
features. The $E(\cdot)\in\mathbb{R}^{d}$ means the embedding layer which maps
the sparse IDs into learnable dense vectors.
After mapping the raw features to low-dimensional embeddings, the common
practice is to aggregate these embeddings to obtain fixed-length vectors.
Different kinds of aggregation methods like (Zhou et al., 2018, 2019) can be
employed to aggregate these embeddings to extract user interest and get the
fixed-length representation. The obtained representation is then fed into the
following deep neural network, e.g., a multi-layer fully-connected network, to
get the final CTR prediction.
Traditional CTR models (Guo et al., 2017b; Lian et al., 2018; Cheng et al.,
2016; Zhou et al., 2018, 2019) are usually trained on data from a single
business domain. However, real-world recommender often has to deal with
multiple business domains. Concretely, the recommender needs to make CTR
prediction for $M$ domains $D_{1},D_{2},\dots,D_{M}$ simultaneously. The model
takes input as $({\bf x},y,p)$, where ${\bf x}$ is the common feature used by
multiple domains like user historical behavior and user profile feature,
target item feature as mentioned above. $y\in\\{0,1\\}$ is the clicked label
and $p\in\\{1,2,\dots,M\\}$ is the domain indicator that indicates which
domain this sample is collected. Note that $({\bf x},y)$ is drawn from the
domain-specific distribution $D_{p}$ and the distribution varies for different
domains. The goal of multi-domain CTR prediction is to construct a single CTR
model that can give accurate CTR prediction to serve all domains at low
resource consumption and human cost.
### 3.2. Architecture Overview
As mentioned above, ignoring domain indicator $p$ and learning a single shared
CTR model neglect the domain differences. This leads to inferior model
performance. On the other hand, training separate models for each domain
performs much worse since splitting the domains provides much less data for
each model. Besides, it is infeasible to maintain each domain a separate model
in production due to the resource consumption and human cost.
To this end, we propose Star Topology Adaptive Recommender (STAR) for multi-
domain CTR prediction to better utilize the similarity among different domains
while capturing the domain distinction. As shown in Figure 4, STAR consists of
three main components: (1) the partitioned normalization (PN) which privatizes
normalization for examples from different domains, (2) the star topology
fully-connected neural network (star topology FCN), (3) the auxiliary network
that treats the domain indicator directly as the input feature and learns its
semantic embeddings to capture the domain distinction.
During training, a domain indicator $p$ is first sampled and then a mini-batch
of $B$ instances
$({\bf x}_{1},p),({\bf x}_{2},p),\dots,({\bf x}_{B},p)$
is sampled from this domain. STAR first embeds these input features as low-
dimensional vectors by an embedding layer. In industrial recommender, the
model is often trained with billions of features (Jiang et al., 2019) and the
parameters of embedding are usually much more than other parts of the model.
This makes it difficult for different domains to learn domain-specific
embeddings with limited data. For example, for models used in our daily tasks,
the embeddings parameters are 10,000 times more than the parameters of fully-
connected layers (Jiang et al., 2019). Thus, in the proposed STAR model, we
let all business domains share the same embedding layer, i.e., the same ID
features in different domains share the same embedding. Sharing embedding
layer across multiple domains can significantly reduce the computational and
memory cost.
The embeddings are then pooled and concatenated to obtain $B$ fixed-length
representations. After that, the $B$ extracted representations are processed
by the proposed partitioned normalization (PN) layer that privatizes
normalization statistics for different domains. The normalized vectors are
then fed as input to the proposed star topology FCN to get the output. The
star topology FCN consists of shared centered FCN and multiple domain-specific
FCNs. The final model of each domain is obtained by combining the shared
centered FCN and domain-specific FCN.
In multi-domain CTR prediction, features that depict the domain information is
of importance. In the STAR model, the auxiliary network treats the domain
indicator as input and fed with other features depicting the domain to the
auxiliary network. The output of the auxiliary network is added with the
output of the star topology FCN to get the final prediction. We make the
auxiliary network much simpler than the star topology FCN to let the model
capture the domain distinction in a direct and easy manner. In what follows we
will describe these components in detail.
### 3.3. Partitioned Normalization
As mentioned above, the raw features are first transformed into low-
dimensional embeddings and then pooled and aggregated to get the intermediate
representation. Denote the intermediate representation of an instance as ${\bf
z}$, to train deep networks fast and stably, a standard practice is applying
normalization layer to the intermediate representation ${\bf z}$. Among all
normalization methods, batch normalization (BN) (Ioffe and Szegedy, 2015) is a
representative method that is proved to be crucial to the successful training
of very deep neural networks (Ioffe and Szegedy, 2015; Radford et al., 2016).
BN uses a global normalization for all examples, which accumulates
normalization moments and learns shared parameters across all samples.
Concretely, the normalization of BN in training is given as
(1) ${\bf z}^{\prime}=\gamma\frac{{\bf
z}-\mu}{\sqrt{\sigma^{2}+\epsilon}}+\beta,$
where ${\bf z}^{\prime}$ is the output, $\gamma,\beta$ are the learnable scale
and bias parameters, $\mu,\sigma^{2}$ are mean and variances of current mini-
batch. During testing, moving averaged statistics of mean $E$ and variance
$Var$ across all samples are used instead
(2) ${\bf z}^{\prime}=\gamma\frac{{\bf z}-E}{\sqrt{Var+\epsilon}}+\beta.$
In other words, BN assumes all samples are i.i.d. and use the shared
statistics across all training samples.
However, in multi-domain CTR prediction, samples are only assumed to be
locally i.i.d. within a specific domain. Thus, data from different domains
have different normalization moments. Sharing global moments and parameters of
BN layers during testing will obscure domain differences and lead to degraded
model performance. To capture the unique data characteristic of each domain,
we propose partitioned normalization (PN) which privatizes normalization
statistics and parameters for different domains. Concretely, during training,
suppose the current mini-batch is sampled from the $p$-th domain, we compute
the mean and variances of the current mini-batch and normalize the feature as:
(3)
$z^{\prime}=(\gamma*\gamma_{p})\frac{z-\mu}{\sqrt{\sigma^{2}+\epsilon}}+(\beta+\beta_{p}),$
where $\gamma,\beta$ are the global scale and bias, and $\gamma_{p},\beta_{p}$
are the domain-specific scale and bias parameters. For each mini-batch, it
receives the final scale by element-wise multiplying the shared $\gamma$ with
the domain-specific $\gamma_{p}$, i.e., PN adaptively scales the
representation according to the domain indicator. Similarly, the bias of PN is
also adaptive conditioned on the domain, which is implemented by the addition
of global bias $\beta$ and domain-specific bias $\beta_{p}$. Note that
compared with BN, PN also uses the moments of the current mini-batch during
training, but PN introduces domain-specific scale and bias
$\gamma_{p},\beta_{p}$ to capture the domain distinction.
Besides the modification of the scale and bias, PN also let different domains
to accumulate the domain-specific moving average of mean $E_{p}$ and variance
$Var_{p}$. During testing, PN transforms instance ${\bf z}$ from the $p$-th
domain as:
(4) $\displaystyle{\bf z}^{\prime}=(\gamma*\gamma_{p})\frac{{\bf
z}-E_{p}}{\sqrt{Var_{p}+\epsilon}}+(\beta+\beta_{p}).$
From Equation 4, we can see that PN uses the domain-specific mean $E_{p}$ and
variance $Var_{p}$ to normalize the intermediate representation ${\bf z}$.
Thus PN adaptively alters the intermediate representation conditioned on the
domain indicator to capture the distinctive characteristics.
### 3.4. Star Topology FCN
Figure 5. An illustration on how STAR generates the parameters of fully-
connected network (FCN) for different domains. STAR consists of a shared
centered FCN and independent FCNs per domain. For each domain, the final
weights of a neural network layer are obtained by element-wise multiplying the
weights of the shared FCN and the domain-specific FCN. The shared parameters
are updated through the gradient of all examples, while the domain-specific
parameters are only updated through examples within this domain.
After the PN layer, the representation ${\bf z}^{\prime}$ is fed as input to
the following star topology multi-layer fully-connected neural network (star
topology FCN). As depicted in Figure 5, the proposed star topology FCN
consists of a shared centered FCN and independent FCNs per domain, thus the
total number of FCN is $M+1$. The final model of $p$-th domain is obtained by
combining the shared centered FCN and domain-specific FCN, in which the
centered parameters learn general behaviors among all domains, and the domain-
specific parameters capture specific behaviors in different domains to
facilitate more refined CTR prediction.
Specifically, for the shared FCN, let $W$ be the weights and $b$ be the bias
in a neural network layer respectively. For the specific FCN of the $p$-th
domain, let $W_{p}$ be the weights and $b_{p}$ be the bias in the
corresponding layer. Denote the input dimension as $c$ and the output
dimension as $d$, i.e, $W,W_{p}\in\mathbb{R}^{c\times
d},b,b_{p}\in\mathbb{R}^{d}$. The final weights $W^{\star}_{i}$ and bias
$b^{\star}_{i}$ for the $p$-th domain is obtained by:
(5) $W^{\star}_{p}=W_{p}\otimes W,b^{\star}_{p}=b_{p}+b,$
where $\otimes$ denotes the element-wise multiplication. Let
$in_{p}\in\mathbb{R}^{c\times 1}$ denote the input of this neural network
layer from the $p$-th domain, the final output $out_{p}\in\mathbb{R}^{d}\times
1$ is given by:
(6) $out_{p}=\phi((W^{\star}_{p})^{\top}in_{p}+b^{\star}_{p}),$
where $\phi$ denotes the activation function of this layer. The combination of
shared parameters of domain-specific parameters is employed in all layers. By
this means, STAR can modulate its parameters conditioned on the domain.
Note that we implement the combination strategy of the shared centered FCN and
domain-specific FCN by element-wise product of between weights and addition of
bias in each layer, other strategies can also be investigated for better
performance. The shared parameters are updated through the gradient of all
examples while the domain-specific parameters are only updated through
examples within this domain. This helps captures the domain differences for
more refined CTR prediction while learning the domain commonality through the
shared centered parameters. As mentioned above, most of the parameters in
industrial recommenders are contributed by the embedding layer, the increased
$M$ FCNs is negligible to the total amount of parameters. Thus STAR uses one
model to effectively serve all business domains in a parameter efficient and
memory friendly manner.
### 3.5. Auxiliary Network
In the traditional way of CTR modeling, all features are treated equally and
fed to the complicated model. In multi-domain CTR prediction, however, it may
be hard for the model to automatically learn the domain distinction. We argue
that a good multi-domain CTR model should have the following characteristic:
(1) have informative features regarding the domain characteristic (2) make
these features easily and directly influence the final CTR prediction. The
intuition behind is that features that depict the information of domains are
of importance since it can reduce the difficulty for the model to capture the
distinction among domains.
To this end, we propose an auxiliary network to learn the domain distinction.
To augment informative features regarding the domain characteristic, we treat
the domain indicator directly as the ID feature input. The domain indicator is
first mapped into embedding vector and concatenated with other features. The
auxiliary network then computes forward pass with respect to the concatenated
features to gets the one-dimensional output. Denote the one-dimensional output
of star topology FCN as $s_{m}$ and the output of the auxiliary network as
$s_{a}$. $s_{m}$ and $s_{a}$ are added to get the final logit. Sigmoid is then
applied to get the CTR prediction:
(7) $\textrm{Sigmoid}(s_{m}+s_{a}).$
In our implementation, the auxiliary network is much simpler than the main
network, which is a two-layer fully connected neural network. The simple
architecture makes the domain features directly influence the final
prediction.
Denote $\hat{y}^{p}_{i}$ the predicted probability for the $i$-th instance in
the $p$-th domain and $y^{p}_{i}\in\\{0,1\\}$ the ground truth. We minimize
the cross entropy loss function between the $\hat{y}^{p}_{i}$ and label
$y^{p}_{i}$ in all domains as:
(8)
$\displaystyle\min\sum_{p=1}^{M}\sum_{i=1}^{N_{p}}-y^{p}_{i}\mathrm{log}(\hat{y}^{p}_{i})-(1-y^{p}_{i})\mathrm{log}(1-\hat{y}^{p}_{i}).$
## 4\. Experiments
Table 1. The example percentage and average click-through rate (CTR) of each
domain.
| 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 | 10 | 11 | 12 | 13 | 14 | 15 | 16 | 17 | 18 | 19
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---
Percentage | 0.99% | 1.61% | 3.40% | 3.85% | 2.79% | 0.56% | 4.27% | 16.76% | 10.00% | 12.16% | 0.76% | 1.31% | 3.34% | 28.76% | 1.17% | 0.46% | 1.05% | 0.91% | 5.85%
CTR | 2.14% | 2.69% | 2.97% | 3.63% | 2.77% | 3.45% | 3.59% | 3.24% | 3.23% | 2.08% | 12.05% | 3.52% | 1.27% | 3.75% | 12.03% | 4.02% | 1.63% | 4.64% | 1.42%
We evaluate the efficacy of STAR in this section. We begin by introducing the
setup including the used production dataset, compared methods and
implementation details in Sec. 4.1. The results and discussion are elaborated
in Sec. 4.2. We also perform in-depth ablation studies in Sec. 4.3.
Experimental results on production environment are shown in Sec. 4.4.
### 4.1. Experimental Settings
Dataset. Due to the lack of public dataset on multi-domain CTR prediction, we
use Alibaba production data regarding user click behavior on 19 business
domains to perform the offline evaluation. The training data is collected from
traffic logs of the online display advertising system of Alibaba. Data of one
day from 19 business domains are used for training and the data of the
following day is used for testing. The training dataset consists of billions
of examples. Table 1 shows the example percentage and average CTR (# Click/#
Impression, i.e., ratio of positive examples) of each domain in the training
set. As shown in Table1, different domains have different domain-specific data
distribution, which can be reflected from the different CTR. It can be seen
that domain with the highest CTR (domain #15) is 12.03% while domain with the
lowest CTR (domain #13) is only 1.27%. In this dataset, the majority of items
are available in most of the business domains while only some of users are
overlapping, e.g., domain #1 and domain #2 have the same set of items but only
have 8.52% overlapping users.
Compared models. To verify the effectiveness of the proposed approach, we
compare STAR with the following models:
* •
Base. We refer to Base as the model composed of embedding layer, pooling &
concatenation layer, batch normalization, and a 7-layer fully-connected
network. Specifically, the pooling & concatenation layer is based on DIEN
(Zhou et al., 2019), which extracts user interest after the embedding layer.
We mix all samples from different domains and train the base model.
* •
Shared Bottom. The Shared Bottom model is a multi-task model that shares the
parameters of the bottom layers. In our implementation, we let the Shared
Bottom share the embedding layer. Each domain will also have a specific
7-layer fully-connected network that is not shared.
* •
MulANN. MulANN (Sebag et al., 2019) adds domain discriminator module to the
Base model. The domain discriminator classifies which domain the examples are
from. MulANN adopts a adversarial loss to let the domain discriminator
indiscriminates with respect to the shift between the domains.
* •
MMoE. MMoE (Ma et al., 2018b) implicitly models task relationships for multi-
task learning, where different tasks may have different label spaces. Here we
adapt MMoE for multi-domain CTR prediction, where each expert is a 7-layer
fully-connected network. The number of experts is equal to the number of
domains. Besides, MMoE also learns gating networks per domain that takes the
input features and outputs softmax gates assembling the experts with different
weights.
* •
Cross-Stitch. Cross-Stitch (Misra et al., 2016) uses linear cross-stitch units
to learn an optimal combination of task-specific representations. In the
cross-stitch method, each domain have a 7-layer fully-connected network and
the cross-stitch units are added in each hidden layer to learn task-specific
representations.
To give a fair comparison, all compared methods and the STAR model are trained
with the proposed auxiliary network in Sec. 4.2. The ablation study about the
auxiliary network is performed in Sec. 4.3.
Implementation details. All models are trained with Adam (Kingma and Ba,
2015), the learning rate is set to 0.001 and the batch size is 2000. We
minimize the cross-entropy loss for samples from all domains to train the
model.
Metrics. Area under the ROC curve (AUC) is the common metric used to evaluate
the performance of CTR prediction. An variation of user weighted AUC (Zhou et
al., 2018) measures the goodness of intra-user order by averaging AUC over
users and is shown to be more relevant to online performance in recommender
system. It is calculated as follows:
(9)
$\textrm{AUC}=\frac{\sum_{i}^{n}\\#\textrm{impression}_{i}\times\textrm{AUC}_{i}}{\sum_{i}^{n}\\#\textrm{impression}_{i}},$
where $n$ is the number of users, $\textrm{impression}_{i}$ and
$\textrm{AUC}_{i}$ are the number of impressions and AUC of the $i$-th user,
respectively. We use this weighted AUC as the evaluation metric and still
refer it to as AUC for simplicity. Concretely, we use the AUC of each domain
and overall AUC (mixing samples from all domains to calculate the overall AUC)
as the metrics.
### 4.2. Results
We evaluate all approaches on the Alibaba production dataset. To give a fair
comparison, all compared methods and STAR model are trained with the proposed
auxiliary network. As illustrated in Table 2, the consistent improvement
validates the efficacy of STAR. Note that the performance of MulANN is worse
than the Base model, which proves obscuring domain difference hurts the
modeling of multi-domain CTR prediction. Besides, the shared Bottom model,
MMoE, Cross-Stitch and STAR all achieve better overall performance than the
Base model. This demonstrates the importance of exploiting domain relationship
and capturing domain distinction to enhance the prediction performance.
Although the Shared Bottom, MMoE, and Cross-Stitch achieve better overall
performance than the Base model, it is notable that in some domains, the AUCs
of Shared Bottom, MMoE, and Cross-Stitch are worse than the Base model, e.g.,
domain # 5, #6, and #16. We hypothesize this is because the learning of these
models conflicts in different domains. In contrast, STAR avoids this issue by
its star topology, where the the domain-specific parameters are only updated
through examples within this domain. The proposed STAR model exhibits superior
performance across all domains compared with the Base model. STAR also
achieves consistent improvement over the Shared Bottom, which demonstrates the
importance of information sharing on top specific layers for multi-domain
learning, where all domains share the same label space. STAR also outperforms
MMoE and Cross-Stitch, which shows the superiority of explicitly modeling
domain relationships compared with implicitly modeling domain relationships by
the gate networks or cross-stitch units.
Table 2. Results of different approaches on offline Alibaba production
dataset.
| Base | Shared Bottom | MulANN | MMoE | Cross-Stitch | STAR
---|---|---|---|---|---|---
#1 | 0.6134 | 0.6186 | 0.6143 | 0.6143 | 0.6183 | 0.6306
#2 | 0.6321 | 0.6320 | 0.6321 | 0.6355 | 0.6337 | 0.6417
#3 | 0.6281 | 0.6293 | 0.6282 | 0.6311 | 0.6307 | 0.6372
#4 | 0.6326 | 0.6361 | 0.6333 | 0.6373 | 0.6372 | 0.6451
#5 | 0.6308 | 0.6292 | 0.6302 | 0.6336 | 0.6322 | 0.6388
#6 | 0.6378 | 0.6383 | 0.6336 | 0.6412 | 0.6368 | 0.6494
#7 | 0.6305 | 0.6329 | 0.6310 | 0.6340 | 0.6352 | 0.6410
#8 | 0.6297 | 0.6278 | 0.6297 | 0.6330 | 0.6328 | 0.6411
#9 | 0.6264 | 0.6283 | 0.6258 | 0.6292 | 0.6278 | 0.6368
#10 | 0.6392 | 0.6434 | 0.6375 | 0.6431 | 0.6278 | 0.6577
#11 | 0.6469 | 0.6529 | 0.6445 | 0.6508 | 0.6548 | 0.6719
#12 | 0.6506 | 0.6575 | 0.6498 | 0.6518 | 0.6570 | 0.6676
#13 | 0.6558 | 0.6612 | 0.6538 | 0.6603 | 0.6637 | 0.6739
#14 | 0.6362 | 0.6405 | 0.6371 | 0.6412 | 0.6411 | 0.6486
#15 | 0.6745 | 0.6888 | 0.6710 | 0.6787 | 0.6819 | 0.7021
#16 | 0.6638 | 0.6627 | 0.6517 | 0.6634 | 0.6727 | 0.6901
#17 | 0.6524 | 0.6658 | 0.6499 | 0.6519 | 0.6575 | 0.6715
#18 | 0.6493 | 0.6480 | 0.6375 | 0.6500 | 0.6610 | 0.6754
#19 | 0.6330 | 0.6375 | 0.6306 | 0.6374 | 0.6381 | 0.6476
Overall AUC | 0.6364 | 0.6398 | 0.6353 | 0.6403 | 0.6415 | 0.6506
### 4.3. Ablation Study
To investigate the effect of each component, we conduct several ablation
studies.
#### 4.3.1. STAR Topology FCN and PN
Table 3. Ablation study of partitioned normalization (PN) and star topology
fully-connected neural networks (STAR FCN). All models are trained with the
proposed auxiliary network.
| Base (BN) | Base (PN) | STAR FCN (BN) | STAR FCN (PN)
---|---|---|---|---
Overall AUC | 0.6364 | 0.6485 | 0.6455 | 0.6506
We analyze the influence of different components of STAR. Concretely, the
separate effects of star topology FCN and PN are investigated. We compare (a)
the Base model trained with BN, (b) Base model trained with PN, (c) STAR FCN
with BN and (d) STAR model (STAR FCN + PN). The result is reported in Table 3.
We observe that using star topology FCN and PN separately can outperform the
Base model. Bring them together can further boost performance. The result
validates the effect of both star topology FCN and PN.
#### 4.3.2. Normalization
Normalization methods are very effective components in deep learning, which
have been shown by many practices to ease optimization and enable very deep
networks to converge. We analyze the effect of different normalization methods
including Batch Normalization (BN) (Ioffe and Szegedy, 2015), Layer
Normalization (LN) (Ba et al., 2016) and the proposed Partitioned
Normalization (PN) on multi-domain CTR prediction. BN accumulates global
statistics and learns global parameters for samples from all domains. LN is a
representative instance-based normalization method, which operates along the
channel dimension and avoids mixing statistics for samples from different
domains.
The result is shown in Table 4. Our first observation is that both LN and PN
outperforms BN. This observation validates that data from different domains
have distinct distribution and need specific normalization. Using global
normalization obscures domain differences, which will hurt performance for
multi-domain CTR prediction. We also observe that PN outperforms LN, which
validates that domain-specific normalization is better than the instance-
specific normalization, since PN leads to more accurate moments within the
domain.
Table 4. Ablation study of normalization methods for multi-domain CTR
prediction. STAR FCN is trained BN, LN, and PN respectively.
| STAR FCN (BN) | STAR FCN (LN) | STAR FCN (PN)
---|---|---|---
Overall AUC | 0.6455 | 0.6463 | 0.6506
#### 4.3.3. Auxiliary network
We conduct experiment to assess the effect of the auxiliary network for
different models. All methods are trained with and without the proposed
auxiliary network. The result is illustrated in Figure 6. We observe that the
auxiliary network improves all methods consistently. The result validates the
importance of making full utilization of domain features and using it to
capture the domain distinction. We also observe the improvement of the
auxiliary network for MulANN is slightly weaker than the other methods. The
reason may due to the fact that the adversarial loss for obscuring domain
differences contradicts with the domain feature to capture the domain
differences.
Figure 6. The performance of different methods trained with (w/) and without
(w/o) the auxiliary network.
#### 4.3.4. Ability to Capture Domain Distinction
Figure 7. Predicted CTR over CTR (PCOC) of the Base model and STAR in all
domains. Each circle means PCOC of a specific domain.
Cost-per-click (CPC) is a widely used performance-dependent payment model in
display advertising, where advertisers bid for clicks. In CPC, the display
systems compute the effective cost per mille (eCPM) as the product of bid
times its CTR. The systems allocate impressions according to the descending
order of the eCPM. In CPC, the CTR model needs to be well-calibrated (Guo et
al., 2017a) in order to achieve a competitive advertising system, i.e., the
predicted CTR should be as close as to the actual CTR.
We show that STAR is more well-calibrated and is capable of capturing domain
distinctions. We compute the predicted CTR over CTR (PCOC) in each domain.
Note that the closer PCOC is to 1.0, the more accurate the CTR prediction is.
For the simplicity of illustration, we show the PCOCs of the Base model and
STAR in Figure 7. We can see that the PCOCs of STAR in different domains are
more compact and concentrated around 1.0 compared with the Base model. The
result validates the ability of STAR to capture the domain distinction.
### 4.4. Production
Online serving and challenges. One of the challenges in industrial recommender
is that the distribution of features and CTR exhibits large shifts over time.
To capture the dynamic change of data in real-time, it is important to use
real-time examples to update the CTR models continuously to prevent them from
becoming stale. However, for multi-domain CTR prediction, the percentage of
examples of each domain changes over time. For example, some business domains
have traffic spike in the morning while some business domains have traffic
spike in the evening. If we train the model directly in the chronological
order, the changes in data percentage over time will cause the instability of
model learning. To address this issue, we redesign the data pipeline and
maintain a buffer that stores a sliding window of history samples to avoid the
sudden change of example percentage. Specifically, samples in the buffer are
shuffled firstly and then sampled to construct a mini-batch. After fed to the
model, this mini-batch of samples are removed from the buffer and new arriving
data is added to this buffer. We empirically found this training manner is
more stable than the traditional way of online updates.
Note that during serving, the weights of FCN for each domain are pre-computed
to achieve faster inferences. By this means, the computational time of STAR
equals the Shared Bottom model. The systematical optimization makes STAR
capable of serving main traffic of multiple business domains stably. Since
2020, STAR is deployed and serves more than 60 business domains on the display
advertising system of Alibaba. We compute the overall improvements of all
domains. Table 5 shows the improvement of STAR over the previous production
model, the Base model. The introduction of STAR brings +8.0% overall CTR lift
and +6.0% overall RPM lift in our online A / B test.
Table 5. CTR and RPM gains in online display advertising system of Alibaba. | CTR | RPM
---|---|---
Overall | +8.0% | +6.0%
## 5\. Conclusion
In this paper, we propose the star topology adaptive recommender to address
the problem of multi-domain CTR prediction. Instead of keeping unique models
for different domains or simply mixing all samples and maintaining a shared
model, STAR has the star topology, which consists of shared centered
parameters and domain-specific parameters. The shared parameters learn
commonalities, which is updated through all examples. The domain-specific
parameters capture domain distinction for more refined prediction, which is
learned using examples within the specific domain. By this means, STAR can
adaptively modulate its parameters conditioned on the domain for more refined
prediction. The experiments demonstrate that the superiority of STAR on multi-
domain CTR prediction. Since 2020, STAR is deployed in the advertising system
of Alibaba, obtaining 8.0% improvement on CTR and 6.0% on RPM.
## References
* (1)
* Argyriou et al. (2008) Andreas Argyriou, Theodoros Evgeniou, and Massimiliano Pontil. 2008. Convex multi-task feature learning. _Machine Learning_ 73, 3 (2008), 243–272.
* Ba et al. (2016) Lei Jimmy Ba, Jamie Ryan Kiros, and Geoffrey E. Hinton. 2016\. Layer Normalization. _CoRR_ abs/1607.06450 (2016).
* Ben-David et al. (2010) Shai Ben-David, John Blitzer, Koby Crammer, Alex Kulesza, Fernando Pereira, and Jennifer Wortman Vaughan. 2010. A theory of learning from different domains. _Machine Learning_ 79, 1-2 (2010), 151–175.
* Bickel et al. (2007) Steffen Bickel, Michael Brückner, and Tobias Scheffer. 2007\. Discriminative learning for differing training and test distributions. In _Proceedings of the 24th International Conference on Machine Learning_ , Vol. 227. 81–88.
* Caruana (1998) Rich Caruana. 1998\. Multitask Learning. In _Learning to Learn_. 95–133.
* Cheng et al. (2016) Heng-Tze Cheng, Levent Koc, Jeremiah Harmsen, Tal Shaked, Tushar Chandra, Hrishi Aradhye, Glen Anderson, Greg Corrado, Wei Chai, Mustafa Ispir, et al. 2016\. Wide & deep learning for recommender systems. In _Proceedings of the 1st Workshop on Deep Learning for Recommender Systems_. ACM, 7–10.
* Collobert and Weston (2008) Ronan Collobert and Jason Weston. 2008. A Unified Architecture for Natural Language Processing: Deep Neural Networks with Multitask Learning. In _Proceedings of the 25-th International Conference on Machine Learning_ , William W. Cohen, Andrew McCallum, and Sam T. Roweis (Eds.), Vol. 307. 160–167.
* Deng et al. (2013) Li Deng, Geoffrey E. Hinton, and Brian Kingsbury. 2013\. New Types of Deep Neural Network Learning for Speech Recognition and Related Applications: An Overview. In _Proceedings of the 2013 IEEE International Conference on Acoustics, Speech and Signal Processing_. 8599–8603.
* Dredze et al. (2010) Mark Dredze, Alex Kulesza, and Koby Crammer. 2010\. Multi-Domain Learning by Confidence-Weighted Parameter Combination. _Maching Learning_ 79, 1-2 (2010), 123–149.
* Feng et al. (2019) Yufei Feng, Fuyu Lv, Weichen Shen, Menghan Wang, Fei Sun, Yu Zhu, and Keping Yang. 2019. Deep Session Interest Network for Click-Through Rate Prediction. In _Proceedings of the 28th International Joint Conference on Artificial Intelligence_. 2301–2307.
* Friedman (2001) Jerome H Friedman. 2001\. Greedy function approximation: a gradient boosting machine. _Annals of statistics_ (2001), 1189–1232.
* Guo et al. (2017a) Chuan Guo, Geoff Pleiss, Yu Sun, and Kilian Q. Weinberger. 2017a. On Calibration of Modern Neural Networks. In _Proceedings of the 34th International Conference on Machine Learning_ _(Proceedings of Machine Learning Research, Vol. 70)_. 1321–1330.
* Guo et al. (2017b) Huifeng Guo, Ruiming Tang, Yunming Ye, Zhenguo Li, and Xiuqiang He. 2017b. Deepfm: a factorization-machine based neural network for ctr prediction. In _Proceedings of the 26th International Joint Conference on Artificial Intelligence_. Melbourne, Australia., 2782–2788.
* Ioffe and Szegedy (2015) Sergey Ioffe and Christian Szegedy. 2015. Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift. In _Proceedings of the 32nd International Conference on Machine Learning_ , Vol. 37. 448–456.
* Jiang et al. (2019) Biye Jiang, Chao Deng, Huimin Yi, Zelin Hu, Guorui Zhou, Yang Zheng, Sui Huang, Xinyang Guo, Dongyue Wang, Yue Song, et al. 2019\. XDL: An Industrial Deep Learning Framework for High-Dimensional Sparse Data. In _Proceedings of the 1st International Workshop on Deep Learning Practice for High-Dimensional Sparse Data_. 1–9.
* Joshi et al. (2012) Mahesh Joshi, Mark Dredze, William W. Cohen, and Carolyn Penstein Rosé. 2012\. Multi-Domain Learning: When Do Domains Matter?. In _Proceedings of the 2012 Joint Conference on Empirical Methods in Natural Language Processing and Computational Natural Language Learning_. 1302–1312.
* Kendall et al. (2018) Alex Kendall, Yarin Gal, and Roberto Cipolla. 2018. Multi-task learning using uncertainty to weigh losses for scene geometry and semantics. 7482–7491.
* Kingma and Ba (2015) Diederik P. Kingma and Jimmy Ba. 2015. Adam: A Method for Stochastic Optimization. In _Proceedings of the 3rd International Conference on Learning Representations_.
* Koren (2008) Yehuda Koren. 2008\. Factorization meets the neighborhood: a multifaceted collaborative filtering model. In _Proceedings of the 14th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining_. Las Vegas, Nevada, USA, 426–434.
* Koren et al. (2009) Yehuda Koren, Robert M. Bell, and Chris Volinsky. 2009\. Matrix Factorization Techniques for Recommender Systems. _IEEE Computer_ 42, 8 (2009), 30–37.
* Li et al. (2019) Chao Li, Zhiyuan Liu, Mengmeng Wu, Yuchi Xu, Huan Zhao, Pipei Huang, Guoliang Kang, Qiwei Chen, Wei Li, and Dik Lun Lee. 2019\. Multi-Interest Network with Dynamic Routing for Recommendation at Tmall. In _Proceedings of the 28th ACM International Conference on Information and Knowledge Management_. 2615–2623.
* Li et al. (2020) Pengcheng Li, Runze Li, Qing Da, Anxiang Zeng, and Lijun Zhang. 2020. Improving Multi-Scenario Learning to Rank in E-commerce by Exploiting Task Relationships in the Label Space. In _Proceedings of The 29th ACM International Conference on Information and Knowledge Management0_. 2605–2612.
* Lian et al. (2018) Jianxun Lian, Xiaohuan Zhou, Fuzheng Zhang, Zhongxia Chen, Xing Xie, and Guangzhong Sun. 2018\. xDeepFM: Combining Explicit and Implicit Feature Interactions for Recommender Systems. In _Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining_. London, United Kingdom.
* Ma et al. (2019) Jiaqi Ma, Zhe Zhao, Jilin Chen, Ang Li, Lichan Hong, and Ed H. Chi. 2019\. SNR: Sub-Network Routing for Flexible Parameter Sharing in Multi-Task Learning. In _Proceedings of The 33rd AAAI Conference on Artificial Intelligence_. 216–223.
* Ma et al. (2018b) Jiaqi Ma, Zhe Zhao, Xinyang Yi, Jilin Chen, Lichan Hong, and Ed H. Chi. 2018b. Modeling Task Relationships in Multi-task Learning with Multi-gate Mixture-of-Experts. In _Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining_. 1930–1939.
* Ma et al. (2018a) Xiao Ma, Liqin Zhao, Guan Huang, Zhi Wang, Zelin Hu, Xiaoqiang Zhu, and Kun Gai. 2018a. Entire Space Multi-Task Model: An Effective Approach for Estimating Post-Click Conversion Rate. In _The 41st International ACM SIGIR Conference on Research & Development in Information Retrieval_. 1137–1140.
* Misra et al. (2016) Ishan Misra, Abhinav Shrivastava, Abhinav Gupta, and Martial Hebert. 2016. Cross-Stitch Networks for Multi-task Learning. 3994–4003.
* Pi et al. (2019) Qi Pi, Weijie Bian, Guorui Zhou, Xiaoqiang Zhu, and Kun Gai. 2019. Practice on Long Sequential User Behavior Modeling for Click-through Rate Prediction. In _Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining_. 1059–1068.
* Pi et al. (2020) Qi Pi, Guorui Zhou, Yujing Zhang, Zhe Wang, Lejian Ren, Ying Fan, Xiaoqiang Zhu, and Kun Gai. 2020\. Search-based User Interest Modeling with Lifelong Sequential Behavior Data for Click-Through Rate Prediction. In _Proceeding of The 29th ACM International Conference on Information and Knowledge Management_. 2685–2692.
* Qu et al. (2016) Yanru Qu, Han Cai, Kan Ren, Weinan Zhang, Yong Yu, Ying Wen, and Jun Wang. 2016. Product-based neural networks for user response prediction. In _Proceedings of the16th International Conference on Data Mining_. IEEE, 1149–1154.
* Radford et al. (2016) Alec Radford, Luke Metz, and Soumith Chintala. 2016. Unsupervised Representation Learning with Deep Convolutional Generative Adversarial Networks. In _Proceedings of the 4th International Conference on Learning Representations_.
* Rendle (2010) Steffen Rendle. 2010\. Factorization machines. In _Proceedings of the 10th International Conference on Data Mining_. IEEE, 995–1000.
* Ruder (2017) Sebastian Ruder. 2017\. An Overview of Multi-Task Learning in Deep Neural Networks. _CoRR_ abs/1706.05098 (2017).
* Sebag et al. (2019) Alice Schoenauer Sebag, Louise Heinrich, Marc Schoenauer, Michèle Sebag, Lani F. Wu, and Steven J. Altschuler. 2019. Multi-Domain Adversarial Learning. In _Proceedings of the 7th International Conference on Learning Representations_.
* Sener and Koltun (2018) Ozan Sener and Vladlen Koltun. 2018. Multi-Task Learning as Multi-Objective Optimization. In _Advances in Neural Information Processing Systems 31_. 525–536.
* Tang et al. (2020) Hongyan Tang, Junning Liu, Ming Zhao, and Xudong Gong. 2020\. Progressive Layered Extraction (PLE): A Novel Multi-Task Learning (MTL) Model for Personalized Recommendations. In _Proceedings of the 14th ACM Conference on Recommender Systems_. 269–278.
* Vaswani et al. (2017) Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N. Gomez, Lukasz Kaiser, and Illia Polosukhin. 2017\. Attention is All you Need. In _Advances in Neural Information Processing Systems 30_. 5998–6008.
* Wang et al. (2019) Ximei Wang, Ying Jin, Mingsheng Long, Jianmin Wang, and Michael I. Jordan. 2019. Transferable Normalization: Towards Improving Transferability of Deep Neural Networks. In _Advances in Neural Information Processing Systems 32 Vancouver, BC, Canada_. 1951–1961.
* Wen et al. (2020) Hong Wen, Jing Zhang, Yuan Wang, Fuyu Lv, Wentian Bao, Quan Lin, and Keping Yang. 2020. Entire Space Multi-Task Modeling via Post-Click Behavior Decomposition for Conversion Rate Prediction. In _Proceedings of the 43rd International ACM SIGIR conference on research and development in Information Retrieval_. 2377–2386.
* Xiao et al. (2020) Zhibo Xiao, Luwei Yang, Wen Jiang, Yi Wei, Yi Hu, and Hao Wang. 2020. Deep Multi-Interest Network for Click-through Rate Prediction. In _Proceedings of the 29th ACM International Conference on Information and Knowledge Management_. 2265–2268.
* Yang and Hospedales (2015) Yongxin Yang and Timothy M. Hospedales. 2015. A Unified Perspective on Multi-Domain and Multi-Task Learning. In _Proceeding of the 3rd International Conference on Learning Representations_.
* Yuan et al. (2020) Fajie Yuan, Guoxiao Zhang, Alexandros Karatzoglou, Xiangnan He, Joemon Jose, Beibei Kong, and Yudong Li. 2020. One Person, One Model, One World: Learning Continual User Representation without Forgetting. _CoRR_ abs/2009.13724 (2020). arXiv:2009.13724 https://arxiv.org/abs/2009.13724
* Zhou et al. (2019) Guorui Zhou, Na Mou, Ying Fan, Qi Pi, Weijie Bian, Chang Zhou, Xiaoqiang Zhu, and Kun Gai. 2019. Deep Interest Evolution Network for Click-Through Rate Prediction. In _Proceedings of the 33rd AAAI Conference on Artificial Intelligence_. Honolulu, Hawaii, USA, 5941–5948.
* Zhou et al. (2018) Guorui Zhou, Xiaoqiang Zhu, Chenru Song, Ying Fan, Han Zhu, Xiao Ma, Yanghui Yan, Junqi Jin, Han Li, and Kun Gai. 2018\. Deep interest network for click-through rate prediction. In _Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining_. ACM, 1059–1068.
* Zhou et al. (2008) Yunhong Zhou, Dennis Wilkinson, Robert Schreiber, and Rong Pan. 2008. Large-scale parallel collaborative filtering for the netflix prize. In _Proceedings of the International Conference on Algorithmic Applications in Management_. Springer, 337–348.
* Zhu et al. (2017) Han Zhu, Junqi Jin, Chang Tan, Fei Pan, Yifan Zeng, Han Li, and Kun Gai. 2017. Optimized Cost per Click in Taobao Display Advertising. In _Proceedings of the 23rd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining_. 2191–2200.
|
# TSQA: Tabular Scenario Based Question Answering
Xiao Li, Yawei Sun, Gong Cheng
###### Abstract
Scenario-based question answering (SQA) has attracted an increasing research
interest. Compared with the well-studied machine reading comprehension (MRC),
SQA is a more challenging task: a scenario may contain not only a textual
passage to read but also structured data like tables, i.e., tabular scenario
based question answering (TSQA). AI applications of TSQA such as answering
multiple-choice questions in high-school exams require synthesizing data in
multiple cells and combining tables with texts and domain knowledge to infer
answers. To support the study of this task, we construct GeoTSQA. This dataset
contains 1k real questions contextualized by tabular scenarios in the
geography domain. To solve the task, we extend state-of-the-art MRC methods
with TTGen, a novel table-to-text generator. It generates sentences from
variously synthesized tabular data and feeds the downstream MRC method with
the most useful sentences. Its sentence ranking model fuses the information in
the scenario, question, and domain knowledge. Our approach outperforms a
variety of strong baseline methods on GeoTSQA.
## 1 Introduction
Scenario-based question answering (SQA) is to answer questions contextualized
by scenarios (Lally et al. 2017). Compared with the well-studied task of
machine reading comprehension (MRC) which requires reading a passage to
extract or infer an answer (Rajpurkar et al. 2016; Lai et al. 2017), a SQA
task requires reading a scenario which commonly contains both a textual
passage and a set of structured data. One such prominent AI application of SQA
is answering multiple-choice questions in high-school geography exams (Ding et
al. 2018; Huang et al. 2019). Those questions are contextualized by scenarios
containing tables and diagrams, where the rich information cannot be captured
by current MRC methods but have to be manually interpreted using natural
language. Thus, one natural research question arises: can we solve SQA in a
fully automated manner?
Figure 1: Left: an example question contextualized by a tabular scenario in
GeoTSQA. Right: an overview of our approach.
#### Task and Challenges.
Specifically, we focus on questions contextualized by a scenario consisting of
a textual passage and a set of tables. We refer to this branch of SQA as TSQA,
short for _Tabular Scenario based Question Answering_. To support the study of
this task, we construct a dataset named GeoTSQA. It contains 1k real questions
contextualized by tabular scenarios in the geography domain, collected from
China’s high-school exams. Compared with existing datasets for table-based
question answering like WikiTableQuestions (Pasupat and Liang 2015), GeoTSQA
requires fundamentally different reading and reasoning skills, and poses new
research challenges.
For instance, Figure 1 shows a question in GeoTSQA. To answer it, tabular data
needs to be synthesized via a complex operation: identifying a monotonic
increase in ELP over the interval 2000–2003. Focusing on this particular
interval rather than many other intervals is implicitly suggested in the
question: after year 2000. Moreover, the passage in the scenario helps to link
ELP with educational level, and the retrieved domain knowledge bridges the gap
between educational level and rural labor which is the correct answer. To
conclude, TSQA methods need to _properly manipulate tabular data_ , and
_comprehend fused textual information_.
#### Our Approach.
To meet the challenges, considering that text reading has been extensively
studied in MRC research, we propose to extend state-of-the-art MRC methods
with a novel table-to-text generator named TTGen to specifically handle
tabular data. The basic idea is straightforward: feeding a MRC model with
sentences generated from tables _using templates that encapsulate many and
various predefined operations for manipulating tabular data_. However, the
potentially large number (e.g., hundreds) of generated sentences may easily
exceed the capacity of typical MRC models, and produce much noise information
influencing the accuracy of reading comprehension. To address this problem,
TTGen incorporates a sentence ranking model that fuses the information in the
scenario, question, and domain knowledge to effectively _select sentences that
are most useful for answering the question_. It outperforms a variety of
strong baseline methods in extensive experiments on GeoTSQA.
We summarize our contributions in the paper as follows.
* •
We construct and publish GeoTSQA, the first dataset dedicated to TSQA. It
requires reading and reasoning with tables, texts, and domain knowledge at
high school level.
* •
We extend MRC methods with TTGen to solve TSQA. TTGen performs question and
knowledge aware ranking of sentences generated from synthesized tabular data.
#### Outline.
The remainder of the paper is organized as follows. We discuss and compare
with related work in Section 2. We formally define the TSQA task and describe
the construction of the GeoTSQA dataset in Section 3. We introduce our
approach in Section 4. We present experiment settings in Section 5 and report
experiment results in Section 6. Finally we conclude the paper in Section 7.
Our code and data are available on Github.111https://github.com/nju-
websoft/TSQA
## 2 Related Work
### 2.1 SQA
SQA is an emerging AI task and has found application in many domains. The
pioneering WatsonPaths system provides recommendations for diagnosis and
treatment based on a medical scenario about a patient (Lally et al. 2017). In
the legal domain, SQA supports judgment prediction based on the fact
description of a legal case (Ye et al. 2018; Zhong et al. 2018; Yang et al.
2019b).
We focus on TSQA where a scenario contains both textual and tabular data. Such
questions are common in, for example, China’s high-school geography and
history exams where a scenario describes a concrete fact or event to
contextualize a set of questions. Previous efforts in this domain either
ignore tables (Cheng et al. 2016) or manually transform tables into triple-
structured knowledge (Ding et al. 2018) or natural language descriptions for
machine reading (Huang et al. 2019). In contrast, we aim at _solving TSQA in a
fully automated manner by generating texts from tables_.
### 2.2 Table-to-Text Generation
Table-to-text generation has been studied for decades. Early methods rely on
handcrafted rules to generate texts for specific domains such as stock market
summaries (Kukich 1983) and weather forecasts (Goldberg, Driedger, and
Kittredge 1994). They typically implement a pipeline of modules including
content planning, sentence planning, and surface realization. Today, it is
feasible to train neural generation models in an end-to-end fashion, thanks to
the availability of effective pre-trained language models (Devlin et al. 2019;
Radford et al. 2019) and large datasets (Lebret, Grangier, and Auli 2016;
Wiseman, Shieber, and Rush 2017; Dusek, Novikova, and Rieser 2019). Current
models often adopt an encoder-decoder architecture with a copy mechanism
(Wiseman, Shieber, and Rush 2017; Puduppully, Dong, and Lapata 2019a).
Moreover, they can be enhanced with entity representations (Puduppully, Dong,
and Lapata 2019b) and external background knowledge (Chen et al. 2019).
The above methods are targeted on surface-level description of tabular data,
which is insufficient for our task where data in multiple cells needs to be
_synthesized using various operations_ (e.g., extremum, monotonicity, trend).
Generating such natural language statements that are logically entailed from
tabular data, rather than superficial restatements, has recently attracted
research attention (Chen et al. 2020a, d). However, they are primarily focused
on high-fidelity generation, i.e., the generated text should be faithful to
the tabular data. Fidelity is necessary but insufficient for our task where
the generated text also needs to be useful for answering the question. It is
thus essential to _select the proper operation and data from a potentially
very large space_. To this end, our proposed generator TTGen features a
sentence ranking model that fuses the information in the scenario, question,
and domain knowledge.
### 2.3 Table-Based Question Answering
Similar to TSQA, there has been a line of research of answering questions over
tabular data (Pasupat and Liang 2015; Jauhar, Turney, and Hovy 2016; Yin et
al. 2016; Yu et al. 2020). Like our constructed dataset GeoTSQA, these
datasets also require performing various operations over multiple cells.
Differently, their questions can be answered solely on the basis of tabular
data, whereas the questions in GeoTSQA are more naturally contextualized by a
scenario containing _both_ a set of tables and a textual passage which are
equally important and are _dependent on each other_.
From this angle, the most similar dataset to GeoTSQA is HybridQA (Chen et al.
2020c), where table cells are linked with Wikipedia pages. However, GeoTSQA
has its _unique challenges_ due to the source of questions—high-school
geography exams. For example, table cells mainly contain non-linkable numeric
values; more complex operations (e.g., monotonicity) are needed; it would be
helpful to incorporate domain knowledge into question answering.
## 3 Task and Dataset
We firstly define the task of TSQA, and then we construct the GeoTSQA dataset
to support the study of TSQA.
### 3.1 Task Definition
A TSQA task consists of a scenario $\langle P,T\rangle$, a question $Q$, and a
set of options $O$ as candidate answers of which only one is correct. The
scenario contains a passage $P$ and a set of tables $T$. Each table in $T$ has
a header row, a header column, and a set of content cells. The goal is to
select an option from $O$ as the answer to $Q$ contextualized by $\langle
P,T\rangle$.
### 3.2 Dataset Construction
We constructed GeoTSQA. To the best of our knowledge, it is the first dataset
dedicated to the TSQA task.
#### Collecting Questions.
We collected multiple-choice questions contextualized by tabular scenarios in
the geography domain from China’s high-school exams. A related dataset is
GeoSQA (Huang et al. 2019). We not only collected all the questions from
GeoSQA but also reused the code for constructing GeoSQA to crawl much more
questions from the Web to expand our dataset.
However, many collected scenarios are not tabular. Indeed, each scenario is
associated with a set of image files. Each image file depicts either a table
or another kind of diagram such as a map or a histogram. Therefore, we need to
identify images depicting tables or table-like diagrams.
#### Identifying Tables.
We looked for tables, or charts that can be straightforwardly converted to
tables (e.g., histograms, line charts). We manually identified 200 such image
files as positive examples and another 200 image files as negative examples.
We used them to train an image classifier (Szegedy et al. 2016) to classify
all the remaining image files. Finally, for all the image files that were
classified as positive, we manually checked them for classification errors.
#### Extracting Tables.
We recruited 15 undergraduate students from a university in China as
annotators. For image files depicting tables, we used Baidu’s OCR tool to
extract tabular data. OCR errors were manually corrected by annotators. For
image files depicting charts, annotators manually extracted tabular data,
assisted with a tool we developed. The annotator used that tool to easily
click key points in the image, e.g., the origin, coordinate axes, data points.
The tool then automatically converted data points to data tables.
Annotators manually checked each extracted table and filtered out irregular
tables (e.g., with multi-level headers).
#### Filtering Questions.
Last but not least, annotators filtered out questions that can be answered
without using any table. Therefore, every question in GeoTSQA is
contextualized by a tabular scenario, and it is essential to employ the
information in the given tables to answer the question.
### 3.3 Dataset Statistics
Scenarios | 556 |
---|---|---
Chinese characters per passage | $52.42$ | $\pm 32.99$
Tables per scenario | $1.58$ | $\pm 0.93$
Cells per table | $26.98$ | $\pm 17.51$
Questions | 1,012 |
Chinese characters per question | $44.02$ | $\pm 15.89$
Table 1: Statistics about GeoTSQA.
GeoTSQA contains 556 scenarios and 1,012 multiple-choice questions. Each
question has four options. More statistics about the dataset are shown in
Table 1.
Out of the 878 tables in GeoTSQA, 96% only contain numeric content cells. It
differs from HybridQA (Chen et al. 2020c) where content cells are often
entities linked with Wikipedia pages, thereby providing extra background
knowledge for answering questions. For GeoTSQA, to obtain information that is
not explicitly given in the scenario but critical for answering questions, it
is essential to entail from tabular data via operations over multiple cells.
## 4 Approach
We propose a two-step approach to solve TSQA. As illustrated in Figure 1, the
first step (Section 4.2) is a table-to-text generator named TTGen. From the
tables $T$ in a scenario $\langle P,T\rangle$, TTGen generates top-$k$
sentences $S$ that are most useful for answering the question $Q$. The second
step (Section 4.1) is a MRC method based on K-BERT (Liu et al. 2020), a state-
of-the-art knowledge-enabled language model. It fuses the information in the
passage $P$, generated sentences $S$, question $Q$, and domain knowledge $K$
to rank the options in $O$.
### 4.1 MRC with Domain Knowledge
Our MRC method is based on K-BERT (Liu et al. 2020). This state-of-the-art
language model extends BERT (Devlin et al. 2019) with the capability to
utilize external knowledge such as domain knowledge.
#### MRC with K-BERT.
For each option $o_{i}\in O$, we concatenate the passage $P$, top-$k$
sentences $S=\\{s_{1},\ldots,s_{k}\\}$ generated from the tables $T$, question
$Q$, and $o_{i}$ in a standard way, starting with a [CLS] token and separating
with [SEP]:
$I^{\text{MRC}}_{i}=\text{[CLS] $P$ $s_{1}\cdots s_{k}$ $Q$ [SEP] $o_{i}$
[SEP] \emph{NUMS}${}_{i}$ [SEP]}\,,$
(1)
where _NUMS_ i is a concatenation of all the numeric tokens in $P$, $S$, $Q$,
and $o_{i}$. Each numeric token in the original position is replaced by a
special token [NUM].
We use K-BERT to obtain a vector representation for each token in
$I^{\text{MRC}}_{i}$ to capture its semantic features:
$\langle\mathbf{h}^{\text{MRC}}_{i1},\mathbf{h}^{\text{MRC}}_{i2},\ldots\rangle=\text{K-BERT}(I^{\text{MRC}}_{i},~{}K)\,,$
(2)
where $K$ is an external knowledge base we will explain later.
The vector representation for the [CLS] token, i.e.,
$\mathbf{h}^{\text{MRC}}_{i1}$, is used as an aggregate representation for
$I^{\text{MRC}}_{i}$. It is fed into two dense layers followed by a softmax
layer to obtain a correctness score $\hat{\omega}_{i}$ for each option
$o_{i}\in O$:
$\begin{split}\omega_{i}&=\mathbf{w}_{2}^{\intercal}\tanh(\mathbf{W}_{1}\mathbf{h}^{\text{MRC}}_{i1}+\mathbf{b}_{1})+b_{2}\,,\\\
\mathbf{\Omega}&=[\hat{\omega}_{1};\hat{\omega}_{2};\ldots]=\mathtt{softmax}([\omega_{1};\omega_{2};\ldots])\,,\end{split}$
(3)
where $\mathbf{W}_{1}$ is a trainable matrix, $\mathbf{w}_{2}$ and
$\mathbf{b}_{1}$ are trainable vectors, and $b_{2}$ is a trainable parameter.
In the training phase, we minimize the negative log-likelihood loss which
measures the difference between $\mathbf{\Omega}$ and the binary correctness
label on each option (we will detail in Section 5.1). In the test phase, we
choose the option in $O$ with the highest correctness score $\hat{\omega}$ as
the answer.
K-BERT extends BERT with an external knowledge base $K$. It helps to fuse the
information in $P$, $S$, $Q$, $O$, and $K$. We refer the reader to Liu et al.
(2020) for a detailed description of K-BERT. Briefly, each entry in $K$ is a
pair $\langle\text{entity},~{}\text{fact sentence}\rangle$, or a triple
$\langle\text{entity},~{}\text{property},~{}\text{value}\rangle$ which can be
converted into a pair by concatenating the property and the value into a fact
sentence. K-BERT employs $K$ to expand the input sequence into a tree of
tokens: fact sentences about an entity are retrieved from $K$ and inserted as
branches after each mention of the entity in the input sequence. In our
implementation, for each entity, we retrieve top-$\epsilon$ fact sentences
that are most relevant to the input sequence. The relevance of a fact sentence
to the input sequence is measured by the cosine similarity between their
average pre-trained BERT embedding vectors.
#### Domain Knowledge.
For the external knowledge base $K$, for our experiments we use domain
knowledge since all the questions in GeoTSQA are in the geography domain. We
obtain domain knowledge from two sources.
First, we import all the triples in Clinga (Hu et al. 2016), a large Chinese
geographical knowledge base.
Second, we reuse the corpus in (Huang et al. 2019). The corpus contains a
geography textbook providing a set of entity descriptions. We pair each entity
with each sentence in its description as a fact sentence. The corpus also
contains a subset of Chinese Wikipedia. We treat the title of each page as an
entity and pair it with each sentence in the page as a fact sentence.
### 4.2 Table-to-Text Generation (TTGen)
Below we describe the generation of sentences from tables to be fed into our
MRC method. We rely on templates that encapsulate predefined operations for
manipulating tabular data. It enables us to perform complex operations that
are needed for answering hard questions such as those in GeoTSQA. We generate
sentences from tables using all the applicable templates. However, it is
infeasible for a MRC model like K-BERT to jointly encode a large number (e.g.,
hundreds) of sentences. Therefore, we rank the generated sentences and select
$k$ top-ranked sentences that are most useful for answering the question. By
filtering the generated sentences, we can also reduce noise information that
may influence the accuracy of reading comprehension.
#### Sentence Generation.
By significantly extending the operations considered in Chen et al. (2020a,
b), we define six table-to-text templates that encapsulate different powerful
operations for synthesizing numeric tabular data. As we will show in the
experiments, these templates have covered most needs about tables in GeoTSQA.
One can easily add new templates to accommodate other applications.
* •
Extremum. This template reports the maximum or minimum value of a row or
column. An example sentence generated from the table in Figure 1 is: _ELP
reaches a maximum of 2.504 at Year 2000._
* •
Special values. This template reports or compares with a special value (e.g.,
under a column header that is mentioned in the question), e.g., _ELP at Year
2000 is 2.504._
* •
Comparison with average. This template reports a maximal sequence of cells
where all the values are above or below the average of the entire row or
column, e.g., _ELP is relatively large between Year 2000 and 2002._
* •
Monotonicity. This template reports a monotonic increase or decrease over a
maximal sequence of cells, e.g., _ELP decreases between Year 2000 and 2003._
* •
Trend. This template reports the overall trend of a row or column, e.g., _ELP
generally increases and then decreases._
* •
Range comparison. This template reports a comparison between two maximal
corresponding sequences of cells from different rows or columns.
For non-numeric tabular data, we simply concatenate each row header, each
column header, and the corresponding content cell into a sentence.
#### Sentence Ranking.
Let $\hat{S}$ be the set of sentences generated from the tables $T$ using all
the applicable templates. We compute a usefulness score for each sentence
$s_{j}\in\hat{S}$, and choose $k$ top-ranked sentences $S\subseteq\hat{S}$. To
select sentences that are most useful for answering the question, our ranking
model employs K-BERT to fuse the information in the passage $P$, question $Q$,
and domain knowledge $K$ to perform question and knowledge aware ranking.
Figure 2 presents an overview of the model. It integrates two complementary
rankers: sentence-level ranking directly assesses the usefulness of each
individual sentence; template-level ranking infers useful templates purely
from the passage and question.
Figure 2: Sentence ranking model in TTGen.
For sentence-level ranking, we concatenate the passage $P$, question $Q$, and
sentence $s_{j}$ in a standard way:
$I^{\text{SR}}_{j}=\text{[CLS] $P$ $Q$ [SEP] $s_{j}$ [SEP] \emph{NUMS}${}_{j}$
[SEP]}\,,$ (4)
where _NUMS_ j is a concatenation of all the numeric tokens in $P$, $Q$, and
$s_{j}$. Each numeric token in the original position is replaced by a special
token [NUM]. We use K-BERT to obtain a vector representation for each token in
$I^{\text{SR}}_{j}$:
$\langle\mathbf{h}^{\text{SR}}_{j1},\mathbf{h}^{\text{SR}}_{j2},\ldots\rangle=\text{K-BERT}(I^{\text{SR}}_{j},~{}K)\,.$
(5)
The vector representation for the [CLS] token, i.e.,
$\mathbf{h}^{\text{SR}}_{j1}$, is fed into two dense layers followed by a
softmax layer to obtain a usefulness score $\hat{\phi}_{j}$ for each sentence
$s_{j}\in\hat{S}$:
$\begin{split}\phi_{j}&=\mathbf{w}_{4}^{\intercal}\tanh(\mathbf{W}_{3}\mathbf{h}^{\text{SR}}_{j1}+\mathbf{b}_{3})+b_{4}\,,\\\
\mathbf{\Phi}&=[\hat{\phi}_{1};\hat{\phi}_{2};\ldots]=\mathtt{softmax}([\phi_{1};\phi_{2};\ldots])\,,\end{split}$
(6)
where $\mathbf{W}_{3}$ is a trainable matrix, $\mathbf{w}_{4}$ and
$\mathbf{b}_{3}$ are trainable vectors, and $b_{4}$ is a trainable parameter.
In the training phase, we minimize the negative log-likelihood loss which
measures the difference between $\mathbf{\Phi}$ and the binary usefulness
label on each generated sentence (we will detail in Section 5.1).
For template-level ranking, we concatenate the passage $P$ and question $Q$ in
a standard way:
$I^{\text{TR}}=\text{[CLS] $P$ $Q$ [SEP]}\,.$ (7)
We use K-BERT to obtain a vector representation for each token in
$I^{\text{TR}}$:
$\langle\mathbf{h}^{\text{TR}}_{1},\mathbf{h}^{\text{TR}}_{2},\ldots\rangle=\text{K-BERT}(I^{\text{TR}},~{}K)\,.$
(8)
The vector representation for the [CLS] token, i.e.,
$\mathbf{h}^{\text{TR}}_{1}$, is fed into two dense layers followed by a
sigmoid layer to obtain a usefulness score $\hat{\psi}$ for each of the six
templates:
$\begin{split}[\psi_{1};\ldots;\psi_{6}]&=\mathbf{W}_{6}\tanh(\mathbf{W}_{5}\mathbf{h}^{\text{TR}}_{1}+\mathbf{b}_{5})+\mathbf{b}_{6}\,,\\\
\mathbf{\Psi}&=[\hat{\psi}_{1};\ldots;\hat{\psi}_{6}]=\mathtt{sigmoid}([\psi_{1};\ldots;\psi_{6}])\,,\end{split}$
(9)
where $\mathbf{W}_{5}$ and $\mathbf{W}_{6}$ are trainable matrices,
$\mathbf{b}_{5}$ and $\mathbf{b}_{6}$ are trainable vectors. Let sentence
$s_{j}$ be generated by the $\tau_{j}$-th template. We derive usefulness
labels on templates for training from usefulness labels on generated
sentences: a template is labeled useful if and only if at least one sentence
it generates is labeled useful. Multiple sentences and hence multiple
templates may be labeled useful for answering a question. Therefore, in the
training phase, we formulate a multi-label binary classification task, and we
minimize the binary cross-entropy loss which measures the difference between
$\mathbf{\Psi}$ and the binary usefulness label on each template.
Finally, in the test phase, we compute:
$\text{usefulness score of }s_{j}=\hat{\phi}_{j}\cdot\hat{\psi}_{\tau_{j}}\,.$
(10)
## 5 Experiment Setup
We compared our approach with a variety of strong baseline methods for TSQA.
We also evaluated our sentence ranking model, which is the core component of
our approach.
### 5.1 Labeled Data
#### Correctness Labels on Options.
For each question, from its known correct answer, we derived a label for each
of the four options indicating whether it is the correct answer. These binary
correctness labels were used to train and evaluate TSQA methods.
#### Usefulness Labels on Generated Sentences.
The number of all the sentences $\hat{S}$ generated by our templates for a
question is in the range of 2–176, with a mean of 41.58 and a median of 38.
For each question, we asked an annotator (recruited in Section 3.2) to read
$\hat{S}$ and assign a label to each sentence indicating whether it is useful
for answering the question. These binary usefulness labels were used to train
and evaluate sentence ranking models.
#### Gold-Standard Sentences.
Furthermore, the annotator manually summarized the tables in one sentence
describing necessary information for answering the question. This gold-
standard sentence was used for comparison.
We randomly sampled 100 questions from GeoTSQA. For 92 questions, $\hat{S}$
fully covers the information in the gold-standard sentence. For 6 questions,
$\hat{S}$ partially covers that information. Therefore, our six templates show
good coverage of the various operations required by GeoTSQA.
### 5.2 Baselines
Our approach extends MRC methods. It is not our focus to compare existing MRC
methods. Instead, table-to-text generation is our major technical
contribution. Therefore, in the experiments we consistently used the MRC
method based on K-BERT described in Section 4.1, but fed it with sentences
generated from tables by the following different methods.
#### Supervised Methods.
Output of linearization for the table in Figure 1:
---
… ELP at Year 1998 is 2.465. ELP at Year 1999 is 2.476. ELP at Year 2000 is
2.504. ELP at Year 2001 is 2.490. ELP at Year 2002 is 2.482. ELP at Year 2003
is 2.473.
Table 2: Example output of Linearization.
Firstly, we compared with three table-to-text generators that achieved state-
of-the-art results on the recent LogicNLG dataset (Chen et al. 2020a) which,
similar to our GeoTSQA, requires synthesizing data in multiple cells. These
generators are open source. Field-Infusing employs LSTM to encode each table
into a sequence of vectors and then applies Transformer to generate text. GPT-
Linearization linearizes each table as a paragraph by horizontally scanning
the table and concatenating each content cell with its row header and column
header into a sentence. Table 2 illustrates such a paragraph. The resulting
paragraph is then fed into GPT-2 to generate a new text. Coarse-to-Fine is an
enhanced version of GPT-Linearization. It adopts a two-step text generation
process: generating a template and then filling it.
Furthermore, we implemented an enhanced version of GPT-Linearization and
Coarse-to-Fine, referred to as GPT-Linearization+ and Coarse-to-Fine+,
respectively. At the beginning of the paragraph fed into GPT-2, we inserted
the scenario passage and question to enable GPT-2 to perform question-aware
text generation.
All the above supervised table-to-text generators were trained based on
sentences with positive usefulness labels.
#### Unsupervised Methods.
We also compared with two naive table-to-text generators.
Recall that GPT-Linearization generates a paragraph from tables and then feeds
it into GPT-2 to generate a new text. We implemented Linearization. It
directly outputs the generated paragraph without feeding it into GPT-2.
Output of templation for the table in Figure 1:
---
… ELP at Year 2000 is 2.504. … ELP decreases between Year 2000 and 2003. … ELP
generally increases and then decreases. … ELP reaches a maximum of 2.504 at
Year 2000. … ELP is relatively large between Year 2000 and 2002. …
Table 3: Example output of Templation.
Besides, we implemented Templation. It generates a paragraph consisting of all
the sentences $\hat{S}$ generated by our templates. Sentences are sorted in
ascending order of length so that if the paragraph has to be truncated by the
maximum sequence length of K-BERT, the largest number of sentences can be
retained. Table 3 illustrates such a paragraph.
#### Gold-Standard Sentence.
Last but not least, we used manually annotated gold-standard sentence as a
reference.
### 5.3 Implementation Details
We performed 5-fold cross-validation. For each fold, we split GeoTSQA into 80%
for training and 20% for test. For model selection, we relied on an inner
holdout 80%/20% training/development split. We ran all the experiments on
TITAN RTX GPUs.
For K-BERT, we used BERT-wwm-ext (Cui et al. 2019), a pre-trained Chinese
language model as the underlying language model. We set $\text{maximum
sequence length}=256$, $\text{self-attention layer}=12$, $\text{hidden
units}=768$, $\text{epochs}=15$ for MRC and template-level ranking,
$\text{epochs}=5$ for sentence-level ranking, $\text{batch size}=8$ for MRC,
$\text{batch size}=16$ for template-level ranking and sentence-level ranking,
$\text{learning rate}=1e\text{--}5$, and $\text{attention heads}=12$. For
knowledge base retrieval we set $\epsilon=2$. Inspired by Jin et al. (2020),
for the K-BERT model in our MRC method (but not the one in TTGen), we coarse-
tuned it on $\text{C}^{3}$ (Sun et al. 2020), a Chinese MRC dataset.
For GPT-2, we used $\text{CDialGPT2}_{\text{LCCC-base}}$ (Wang et al. 2020), a
pre-trained Chinese GPT-2 model. For $\text{CDialGPT2}_{\text{LCCC-base}}$,
and for LSTM and Transformer in Field-Infusing, we followed the recommended
hyperparameter settings in their original implementation.
For our TTGen, by default we set $k=2$ to only select the top-2 generated
sentences for MRC. We will report a comparison in different settings of $k$.
### 5.4 Evaluation Metrics
To evaluate TSQA, we measured accuracy, i.e., the proportion of correctly
answered questions.
To evaluate sentence ranking, we measured the quality of the whole ranked list
of all the sentences $\hat{S}$ generated by our templates. We used two
standard information retrieval evaluation metrics: Mean Average Precision
(MAP) and Mean Reciprocal Rank (MRR).
## 6 Experiment Results
We report average results on the test sets over all the folds.
### 6.1 Results on TSQA
#### Comparison with Baselines.
| Accuracy
---|---
Field-Infusing | 0.353 ∙
GPT-Linearization | 0.370
Coarse-to-Fine | 0.367
GPT-Linearization+ | 0.348 ∙
Coarse-to-Fine+ | 0.359 ∘
Linearization | 0.235 ∙
Templation | 0.243 ∙
TTGen | 0.397
Gold-Standard Sentence | 0.418
Table 4: Accuracy of TSQA. We mark the results of baselines that are
significantly lower than TTGen under $p<0.01$ $(^{\bullet})$ or $p<0.05$
$(^{\circ})$.
Table 4 shows the accuracy of TSQA achieved by each method. Our TTGen
outperforms all the baselines by 2.7–16.2 percent of accuracy.
TTGen exceeds three state-of-the-art table-to-text generators, i.e., Field-
Infusing, GPT-Linearization, and Coarse-to-Fine, by 2.7–4.4 percent of
accuracy.
The enhanced version of these generators that we implemented, i.e., GPT-
Linearization+ and Coarse-to-Fine+, exhibit surprisingly worse performance
than their original version. Their generation methods are significantly
inferior to our TTGen by 3.8–5.1 percent of accuracy.
The two naive generators, i.e., Linearization and Templation, produce much
noise information for MRC and achieve accuracy even lower than random guess
(i.e., 0.25). It demonstrates the necessity of ranking and selecting generated
sentences.
The accuracy of using gold-standard sentence is 0.418. On the one hand,
compared with the accuracy 0.397 of our TTGen, it suggests that there is still
room for improving our templates and/or our sentence ranking model. On the
other hand, the achieved accuracy is not satisfying. To improve the overall
performance of our approach, we need to combine our TTGen with novel MRC
methods that are more powerful than K-BERT to meet the unique challenges
raised by the GeoTSQA dataset. This will be our future work.
#### Varying $k$.
| $k=1$ | $k=2$ | $k=3$ | $k=4$ | $k=5$
---|---|---|---|---|---
Accuracy | 0.390 | 0.397 | 0.352 | 0.343 | 0.330
Table 5: Accuracy of TSQA by varying $k$ in TTGen.
Table 5 shows the accuracy of TSQA achieved by our approach under different
settings of $k$. Increasing $k$ from 1 to 2 (the default value), the accuracy
remains stable. Further increasing $k$ to 3 or larger, the accuracy drops
substantially, probably influenced by the extra noise information. It is thus
important to rank generated sentences and only select those useful for
answering the question.
#### Ablation Study.
| Accuracy
---|---
TTGen | 0.397
TTGen w/o tabular data | 0.372
TTGen w/o domain knowledge | 0.380
Table 6: Accuracy of TSQA (ablation study).
To analyze the usefulness of tabular data and domain knowledge in TSQA, we
implemented two variants of our approach. The first variant ignored tabular
data. The second variant ignored domain knowledge.
Table 6 shows the accuracy of TSQA achieved by each variant. Compared with the
full version of our approach, the accuracy of both variants decrease, by 2.5
percent of accuracy without tabular data and by 1.7 percent of accuracy
without domain knowledge. The results reveal the usefulness of tabular data
and of domain knowledge.
### 6.2 Results on Sentence Ranking
We compared our sentence ranking model with a strong baseline method: RE2
(Yang et al. 2019a). This state-of-the-art text matcher is open source. We
employed it to compute the semantic relevance of each generated sentence in
$\hat{S}$ to the question. Specifically, we used RE2 as a text pair classifier
to predict a ranking score for each generated sentence conditioned on (i.e.,
paired with) a concatenation of the scenario passage and question. We followed
the recommended hyperparameter setting in its original implementation.
Table 7 shows the quality of sentence ranking computed by each method. Our
TTGen exceeds RE2 by 5.2 percent of MAP and by 6.0 percent of MRR. Paired
t-tests show that all these differences are statistically significant under
$p<0.01$.
### 6.3 Error Analysis
We randomly sampled 100 questions to which our approach provided incorrect
answers. We analyzed the question answering process and identified the
following three main causes of errors. Multiple causes could apply to a
question.
#### Knowledge Base.
For 76% of the errors, there is a lack of necessary domain or commonsense
knowledge for answering the question, such as the location of a particular
lake. It suggests expanding our knowledge base. However, this is orthogonal to
our technical contribution.
#### Reasoning Capabilities.
For 62% of the errors, more advanced reasoning skills are needed. For example,
some questions require multi-hop math calculations over a group of related
domain concepts. K-BERT as a language model cannot calculate. It is also
impracticable to encapsulate such extremely complex operations with predefined
templates. Therefore, it suggests incorporating specific calculators and
powerful reasoners into MRC models.
#### Sentence Ranking.
For 54% of the errors, our sentence ranking model chooses a sentence that is
not useful for answering the question. Indeed, some templates and their
generated sentences are linguistically similar though logically different,
e.g., _is relatively large_ , _reaches maximum_ , and _increases_. This
sometimes challenges our sentence ranking model as well as our MRC method. We
will focus on this problem in the future work.
| MAP | MRR
---|---|---
RE2 | 0.434 ∙ | 0.461 ∙
TTGen | 0.486 | 0.521
Table 7: Quality of sentence ranking. We mark the results of baselines that
are significantly lower than TTGen under $p<0.01$ $(^{\bullet})$.
## 7 Conclusion
Our study aims at solving TSQA in a fully automated manner to avoid manually
interpreting tabular data using natural language descriptions as done in
previous research. To support this study, we constructed and published the
first dataset GeoTSQA that is dedicated to the TSQA task. With only six
templates encapsulating predefined operations for synthesizing tabular data in
various ways, we covered most needs about tables in GeoTSQA but then, the
problem turned into selecting, among a large number of sentences generated
from templates, the most useful ones for answering the question. Our proposed
model effectively integrates sentence-level and template-level ranking, and
exploits the scenario passage, question, and domain knowledge by fusing their
information with K-BERT. Our approach has the potential to be adapted to other
AI applications that require table comprehension and explanation.
Although our approach outperformed a variety of strong baselines in the
experiments, its accuracy is still not satisfying. Following the results of
our error analysis, for the future work, we plan to enhance our sentence
ranking model with more powerful semantic matching techniques. We will also
extend our MRC method to perform math calculation and logical reasoning over
an expanded knowledge base.
## 8 Acknowledgments
This work was supported by the National Key R&D Program of China
(2018YFB1005100). We thank the annotators for their efforts and thank the
reviewers for their comments.
## References
* Chen et al. (2019) Chen, S.; Wang, J.; Feng, X.; Jiang, F.; Qin, B.; and Lin, C.-Y. 2019. Enhancing Neural Data-To-Text Generation Models with External Background Knowledge. In _Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP)_ , 3022–3032. Hong Kong, China: Association for Computational Linguistics.
* Chen et al. (2020a) Chen, W.; Chen, J.; Su, Y.; Chen, Z.; and Wang, W. Y. 2020a. Logical Natural Language Generation from Open-Domain Tables. In _Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics_ , 7929–7942. Online: Association for Computational Linguistics.
* Chen et al. (2020b) Chen, W.; Wang, H.; Chen, J.; Zhang, Y.; Wang, H.; Li, S.; Zhou, X.; and Wang, W. Y. 2020b. TabFact: A Large-scale Dataset for Table-based Fact Verification. In _8th International Conference on Learning Representations, ICLR 2020, Addis Ababa, Ethiopia, April 26-30, 2020_. OpenReview.net.
* Chen et al. (2020c) Chen, W.; Zha, H.; Chen, Z.; Xiong, W.; Wang, H.; and Wang, W. Y. 2020c. HybridQA: A Dataset of Multi-Hop Question Answering over Tabular and Textual Data. In Cohn, T.; He, Y.; and Liu, Y., eds., _Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: Findings, EMNLP 2020, Online Event, 16-20 November 2020_ , 1026–1036. Association for Computational Linguistics.
* Chen et al. (2020d) Chen, Z.; Chen, W.; Zha, H.; Zhou, X.; Zhang, Y.; Sundaresan, S.; and Wang, W. Y. 2020d. Logic2Text: High-Fidelity Natural Language Generation from Logical Forms. In Cohn, T.; He, Y.; and Liu, Y., eds., _Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: Findings, EMNLP 2020, Online Event, 16-20 November 2020_ , 2096–2111. Association for Computational Linguistics.
* Cheng et al. (2016) Cheng, G.; Zhu, W.; Wang, Z.; Chen, J.; and Qu, Y. 2016. Taking Up the Gaokao Challenge: An Information Retrieval Approach. In Kambhampati, S., ed., _Proceedings of the Twenty-Fifth International Joint Conference on Artificial Intelligence, IJCAI 2016, New York, NY, USA, 9-15 July 2016_ , 2479–2485. IJCAI/AAAI Press.
* Cui et al. (2019) Cui, Y.; Che, W.; Liu, T.; Qin, B.; Yang, Z.; Wang, S.; and Hu, G. 2019. Pre-Training with Whole Word Masking for Chinese BERT. _arXiv preprint arXiv:1906.08101_ .
* Devlin et al. (2019) Devlin, J.; Chang, M.-W.; Lee, K.; and Toutanova, K. 2019. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. In _Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers)_ , 4171–4186. Minneapolis, Minnesota: Association for Computational Linguistics.
* Ding et al. (2018) Ding, J.; Wang, Y.; Hu, W.; Shi, L.; and Qu, Y. 2018. Answering multiple-choice questions in geographical gaokao with a concept graph. In _European Semantic Web Conference_ , 161–176. Springer.
* Dusek, Novikova, and Rieser (2019) Dusek, O.; Novikova, J.; and Rieser, V. 2019. Evaluating the State-of-the-Art of End-to-End Natural Language Generation: The E2E NLG Challenge. _Comput. Speech Lang._ 59: 123–156.
* Goldberg, Driedger, and Kittredge (1994) Goldberg, E.; Driedger, N.; and Kittredge, R. I. 1994. Using natural-language processing to produce weather forecasts. _IEEE Expert_ 9(2): 45–53.
* Hu et al. (2016) Hu, W.; Li, H.; Sun, Z.; Qian, X.; Xue, L.; Cao, E.; and Qu, Y. 2016. Clinga: Bringing Chinese Physical and Human Geography in Linked Open Data. In Groth, P.; Simperl, E.; Gray, A.; Sabou, M.; Krötzsch, M.; Lecue, F.; Flöck, F.; and Gil, Y., eds., _The Semantic Web – ISWC 2016_ , 104–112. Cham: Springer International Publishing. ISBN 978-3-319-46547-0.
* Huang et al. (2019) Huang, Z.; Shen, Y.; Li, X.; Wei, Y.; Cheng, G.; Zhou, L.; Dai, X.; and Qu, Y. 2019\. GeoSQA: A Benchmark for Scenario-based Question Answering in the Geography Domain at High School Level. In _Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP)_ , 5866–5871. Hong Kong, China: Association for Computational Linguistics.
* Jauhar, Turney, and Hovy (2016) Jauhar, S. K.; Turney, P.; and Hovy, E. 2016. Tables as semi-structured knowledge for question answering. In _Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)_ , 474–483.
* Jin et al. (2020) Jin, D.; Gao, S.; Kao, J.; Chung, T.; and Hakkani-Tür, D. 2020. MMM: Multi-Stage Multi-Task Learning for Multi-Choice Reading Comprehension. In _The Thirty-Fourth AAAI Conference on Artificial Intelligence, AAAI 2020, The Thirty-Second Innovative Applications of Artificial Intelligence Conference, IAAI 2020, The Tenth AAAI Symposium on Educational Advances in Artificial Intelligence, EAAI 2020, New York, NY, USA, February 7-12, 2020_ , 8010–8017. AAAI Press.
* Kukich (1983) Kukich, K. 1983. Design of a Knowledge-Based Report Generator. In _21st Annual Meeting of the Association for Computational Linguistics_ , 145–150. Cambridge, Massachusetts, USA: Association for Computational Linguistics.
* Lai et al. (2017) Lai, G.; Xie, Q.; Liu, H.; Yang, Y.; and Hovy, E. 2017. RACE: Large-scale ReAding Comprehension Dataset From Examinations. In _Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing_ , 785–794. Copenhagen, Denmark: Association for Computational Linguistics.
* Lally et al. (2017) Lally, A.; Bagchi, S.; Barborak, M. A.; Buchanan, D. W.; Chu-Carroll, J.; Ferrucci, D. A.; Glass, M. R.; Kalyanpur, A.; Mueller, E. T.; Murdock, J. W.; et al. 2017. WatsonPaths: scenario-based question answering and inference over unstructured information. _AI Magazine_ 38(2): 59–76.
* Lebret, Grangier, and Auli (2016) Lebret, R.; Grangier, D.; and Auli, M. 2016. Neural Text Generation from Structured Data with Application to the Biography Domain. In _Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing_ , 1203–1213. Austin, Texas: Association for Computational Linguistics.
* Liu et al. (2020) Liu, W.; Zhou, P.; Zhao, Z.; Wang, Z.; Ju, Q.; Deng, H.; and Wang, P. 2020. K-BERT: Enabling Language Representation with Knowledge Graph. In _AAAI_ , 2901–2908.
* Pasupat and Liang (2015) Pasupat, P.; and Liang, P. 2015. Compositional Semantic Parsing on Semi-Structured Tables. In _Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing (Volume 1: Long Papers)_ , 1470–1480. Beijing, China: Association for Computational Linguistics.
* Puduppully, Dong, and Lapata (2019a) Puduppully, R.; Dong, L.; and Lapata, M. 2019a. Data-to-Text Generation with Content Selection and Planning. In _The Thirty-Third AAAI Conference on Artificial Intelligence, AAAI 2019, The Thirty-First Innovative Applications of Artificial Intelligence Conference, IAAI 2019, The Ninth AAAI Symposium on Educational Advances in Artificial Intelligence, EAAI 2019, Honolulu, Hawaii, USA, January 27 - February 1, 2019_ , 6908–6915. AAAI Press.
* Puduppully, Dong, and Lapata (2019b) Puduppully, R.; Dong, L.; and Lapata, M. 2019b. Data-to-text Generation with Entity Modeling. In _Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics_ , 2023–2035. Florence, Italy: Association for Computational Linguistics.
* Radford et al. (2019) Radford, A.; Wu, J.; Child, R.; Luan, D.; Amodei, D.; and Sutskever, I. 2019. Language models are unsupervised multitask learners. _OpenAI Blog_ 1(8): 9.
* Rajpurkar et al. (2016) Rajpurkar, P.; Zhang, J.; Lopyrev, K.; and Liang, P. 2016. SQuAD: 100,000+ Questions for Machine Comprehension of Text. In _Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing_ , 2383–2392. Austin, Texas: Association for Computational Linguistics.
* Sun et al. (2020) Sun, K.; Yu, D.; Yu, D.; and Cardie, C. 2020. Investigating Prior Knowledge for Challenging Chinese Machine Reading Comprehension. _Transactions of the Association for Computational Linguistics_ .
* Szegedy et al. (2016) Szegedy, C.; Vanhoucke, V.; Ioffe, S.; Shlens, J.; and Wojna, Z. 2016. Rethinking the Inception Architecture for Computer Vision. In _2016 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2016, Las Vegas, NV, USA, June 27-30, 2016_ , 2818–2826. IEEE Computer Society.
* Wang et al. (2020) Wang, Y.; Ke, P.; Zheng, Y.; Huang, K.; Jiang, Y.; Zhu, X.; and Huang, M. 2020. A Large-Scale Chinese Short-Text Conversation Dataset. In _Nlpcc_.
* Wiseman, Shieber, and Rush (2017) Wiseman, S.; Shieber, S.; and Rush, A. 2017. Challenges in Data-to-Document Generation. In _Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing_ , 2253–2263. Copenhagen, Denmark: Association for Computational Linguistics.
* Yang et al. (2019a) Yang, R.; Zhang, J.; Gao, X.; Ji, F.; and Chen, H. 2019a. Simple and Effective Text Matching with Richer Alignment Features. In _Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics_ , 4699–4709. Florence, Italy: Association for Computational Linguistics.
* Yang et al. (2019b) Yang, W.; Jia, W.; Zhou, X.; and Luo, Y. 2019b. Legal Judgment Prediction via Multi-Perspective Bi-Feedback Network. In Kraus, S., ed., _Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence, IJCAI 2019, Macao, China, August 10-16, 2019_ , 4085–4091. ijcai.org.
* Ye et al. (2018) Ye, H.; Jiang, X.; Luo, Z.; and Chao, W. 2018. Interpretable Charge Predictions for Criminal Cases: Learning to Generate Court Views from Fact Descriptions. In _Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long Papers)_ , 1854–1864. New Orleans, Louisiana: Association for Computational Linguistics.
* Yin et al. (2016) Yin, P.; Lu, Z.; Li, H.; and Kao, B. 2016. Neural Enquirer: Learning to Query Tables in Natural Language. In Kambhampati, S., ed., _Proceedings of the Twenty-Fifth International Joint Conference on Artificial Intelligence, IJCAI 2016, New York, NY, USA, 9-15 July 2016_ , 2308–2314. IJCAI/AAAI Press.
* Yu et al. (2020) Yu, X.; Chen, T.; Yu, Z.; Li, H.; Yang, Y.; Jiang, X.; and Jiang, A. 2020. Dataset and Enhanced Model for Eligibility Criteria-to-SQL Semantic Parsing. In _Proceedings of The 12th Language Resources and Evaluation Conference_ , 5829–5837.
* Zhong et al. (2018) Zhong, H.; Guo, Z.; Tu, C.; Xiao, C.; Liu, Z.; and Sun, M. 2018. Legal judgment prediction via topological learning. In _Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing_ , 3540–3549.
|
# An Explainable CNN Approach for Medical Codes Prediction from Clinical Text
Hu S.Y.ShuYuan Hu Teng F.Fei Teng Leeds joint School, SWJTU, Cheng Du, China
The school of information science and technology, SWJTU, Cheng Du, China
###### Abstract
Background Clinical notes are unstructured text documents generated by
clinicians during patient encounters, generally are annotated with ICD codes,
which give formatted information about the diagnosis and treatment. ICD code
has shown its potentials in many fields, but manual coding is labor-intensive
and error-prone, lead to researches of automatic coding. Two specific
challenges of this task are 1) given an annotated clinical notes, the reasons
behind specific diagnoses and treatments are lost; 2) explainability is
important for practical automatic coding method, the method should not only
explain its prediction output but also have explainable internal mechanics.
This study aims to develop an explainable CNN approach to address these two
challenges.
Method We develop CNN-based methods for automatic ICD coding based on clinical
text from the intensive care unit (ICU) stays. We come up with the Shallow and
Wide Attention convolutional Mechanism (SWAM), which allows our model to learn
local and low-level features for each label. The key idea behind our model
design is to look for the presence of informative snippets in the clinical
text that correlated with each code, and we infer that there exists a
correspondence between “informative snippet” and convolution filter.
Results We evaluate our approach on MIMIC-III, an open-access dataset of ICU
medical records. Our approach substantially outperforms previous results on
top-50 medical code prediction on MIMIC-III dataset. We attribute this
improvement to SWAM, by which the wide architecture gives the model ability to
more extensively learn the unique features of different codes, and we prove it
by ablation experiment. Besides, we perform manual analysis of the performance
imbalance between different codes, and preliminary conclude the
characteristics that determine the difficulty of learning specific codes.
Conclusions We present an explainable CNN approach for multi-label document
classification, which employs a wide convolution layer to learn local and low-
level features for each label, yields strong improvements over previous
metrics on the ICD-9 code prediction task while providing satisfactory
explanations for its internal mechanics.
ICD coding,
Machine learning,
Attentional Convolution for NLP,
###### keywords:
Research
## Background
Clinical notes are written by clinicians during patient encounters, they are
usually unstructured text narratives and accompanied by a set of metadata
codes from the International Classification of Diseases (ICD), which present a
standardized way of indicating diagnoses and procedures that were performed
during the encounter. There is much research that demonstrates the practical
application with ICD codes [7, 26, 3]. For example in work by Choi et al. [5],
they proposed the Doctor AI system based on the presence of ICD codes to
predict future patient states from learning patient representation from a
large dataset of patient records.
But manual coding is time-consuming and error-prone, so much research on
automatic coding has been done in the past three decades, some recent works
are Zhang et al. [40]; Kavuluru et al. [12] and Avati et al. [3]. And
automatic coding is considered a multi-label classification task. Two domain-
specific challenges are facing this task. First, a reasonable guess is that
for a certain code prediction task, most of the text is not informative, only
a few snippets are related to the code. However given the annotated text, the
connections between code and its corresponding informative snippets are lost,
in other words, the model has to learn the reasons behind specific diagnoses
and treatments; second, interpretability is a crucial obstacle for practical
automatic coding in both perspective of inferring and internal mechanics, the
method is supposed to explain its prediction as well as have an explainable
internal mechanics.
To address these two specific challenges together, in this paper, we develop
CNN-based methods for automatic ICD code assignment based on text discharge
summaries from intensive care unit (ICU) stays, we come up with Shallow and
Wide Attention convolutional Mechanism (SWAM), which allows our model to learn
local and low-level features for each label. Our model design is motivated by
the way human clinicians manual label the clinical notes, which is to look for
informative snippets that are relevant to each code. We consider the
“informative snippets” as local and low-level features. SWAM address the two
mentioned challenges in automatic coding: first, by transferring the base
representation (i.e. clinical notes in the word-embedding form) to the
convolution representation which represents the presence of “informative
snippets”, the model could filter out the irrelevant information in the text,
and through the attention mechanism the model could learn the correlation
between “informative snippets” and labels; second, SWAM gives “informative
snippets” extracted from clinical notes as explanations of its prediction
result, and provides a new perspective for understanding the internal
mechanics of the machine learning method.
We evaluate our approach on the MIMIC-III dataset [10], an open dataset of ICU
medical records. With the Shallow and Wide Attention CNN mechanism, the model
can learn non-generic features associated with specific labels that are not
informative for other labels, which the narrow one are failed to learn. With
the performance improvement gained from these specific labels, our approach
outperforms previous results on medical code prediction on MIMIC-III dataset.
### Related work
#### Automatic ICD coding
ICD coding has been a long-established task in the medical informatics
community for decades, from the perspective of data, the current approaches of
this task can be divided into two factions: much recent research focuses on
unstructured text data [34, 12], while the other incorporates structured data
as well [30]. We develop our methods on unstructured text data from the
MIMIC3. Also, from the perspective of the code set, many approaches [36, 24]
evaluate on a subset of the full ICD label space, while there are also methods
[21] developed on the full code set. We develop our methods on the top-50 code
set because the advantage of SWAM is learning specific features associated
with specific labels that are not generic feature for other labels, so instead
of carrying out a surprisingly large network to learn all non-generic features
on the full code set, using ensemble method to cover the whole code set is
preferred, which is discussed in later part.
A tendency in recent years is developing Neural Network-based methods for this
task. Shi et al. [32] applied attentional LSTMs to form a soft matching
between sentence representations from discharge summaries and the top 50
codes. Prakash et al. [24] generated predictions of the top 50 codes by memory
networks built from discharge summaries and Wikipedia. Mullenbach et al. [21]
applied a per-label mechanism to extract the most important snippet for each
code from discharge summaries. SWAM is compared with the published result from
all these papers, and it achieves state-of-the-art results across many
indicators. We attribute these improvements to the ability to learn non-
generic features associated with specific labels that are not informative for
other labels, which bring significant performance improvements on these
specific labels.
#### Attentional convolution for NLP and explainable text classification
Combing convolution with attention has been proved is efficient in different
tasks among NLP [2, 39, 29, 38, 4]. Yang et al. [37] and Mullenbach et al.
[21] utilize attentional convolution to select the most relevant parts of the
clinical text of each code. We refer to the per-label attention mechanism from
those of Mullenbach et al. [21], in which per-label parameter vectors are used
to compute attention over specific locations in the text. Our work differs in
that SWAM establishes the correspondence between the “informative snippet” and
convolution filter, which makes the network a wider one comparing to its of
Mullenbach et al. [21] and is better tuned to our goal of learning low-level
feature, a.k.a. informative snippet with explainable internal mechanics.
Figure 1: Internal mechanism of SWAM.
Attentional Convolution has also been applied to make explainable text
classification. Some prior works like Rushet al. [28] and Rocktäschel et al.
[27] employ attention to highlight salient features of the text. The per-label
attention mechanism [21] we referred extract snippet from the text as
automatically generated explanation of the prediction in the same medical
codes prediction task, and the informativeness of explanations are rated by a
physician. Their work illustrates that the neural network work in an
explainable way for this task: the model will try to find parts of the text
that are most relevant to each code. Our work differs in that instead of
making the model explainable by explaining its prediction, we take a further
step forward to make the internal mechanics of the method explainable by
opening the black box of the neural network to establish the correspondence
between the “informative snippet” and convolution filter. We also bring out a
preliminary analysis of the imbalance performance between the labels, provide
a rational explanation of why the model performs terribly on certain codes.
#### Neural network architecture design for text classification work
Hoa et al. [17] compared the deep CNN and shallow CNN under text
classification task, a practical rule is summarized that deep models do not
seem to bring a significant advantage over shallow networks for text
classification, another observation they made is that a global max-pooling
[6], which retrieves the most influential feature could already be good enough
for the text classification task. The authors believe one possible reason may
be related to these facts that images are represented as real and dense
values, as opposed to the discrete, artificial, and sparse representation of
text. Their work indicates that local and low-level features extracted by
shallow CNN work well for text classification tasks and inspires us to explore
the underlying correspondence between local and low-level features and
snippets in the text.
Gong and Ji [8] find that in CNN for the text classification task, the
convolution filters have learned division of labor. More than half of the
kernels have a preference for one specific label. Their work inspires us to
associate the width of the network with the learning of features of specific
labels that are not generic for other labels.
Table 1: Table of Notations Notation | Description
---|---
$\mathcal{L}$ | The set of ICD-9 codes.
$y_{i,\ell}\in\ {0,\ 1}$ | The true value of the label task for instance i and $\ell\in\mathcal{L}$, 1 indicates the label is true for instance i.
$d_{e}$ | The size of the input embedding
$d_{c}$ | The size of the convolution output, a.k.a. the number of convolution filters
$\mathbf{X}=[\mathbf{x}_{1},\mathbf{x}_{2},\ldots,\mathbf{x}_{N}]$ | The matrix of a document instance, where $\mathbf{N}$ is the length of the document and $\mathbf{x}_{i}$ is the vector representation of the word.
$\mathbf{W}_{c}\in\mathbb{R}^{k\times d_{e}\times d_{c}}$ | Convolution filters, where k is the width of filter window.
$\mathbf{H}\in\mathbb{R}^{d_{c}\times N}$ | Convolutional representation of the document.
$\ast$ | Convolution operator.
$g$ | An element-wise nonlinear transformation.
$\mathbf{b}_{c}\in\mathbb{R}^{d_{c}}$ | The bias in convolutional operation.
$\mathbf{u}_{\ell}\in\mathbb{R}^{d_{c}}$ | Attention parameter vector for label. $\ell$
$\boldsymbol{\alpha}_{\ell}\in\mathbb{R}^{N}$ | Attention result vector for label. $\ell$
$b_{\ell}$ | Scalar offset in linear layer for label. $\ell$
$\boldsymbol{\beta}_{\ell}\in\mathbb{R}^{d_{c}}$ | Vector of prediction weights.
$\sigma$ | Sigmoid function.
$\operatorname{SoftMax}()$ | $\operatorname{SoftMax}(\mathbf{x})=\frac{\operatorname{\exp}(\mathbf{x})}{\sum_{i}{\operatorname{\exp}(x_{i})}}$ , where $\operatorname{\exp}(\mathbf{x})$ is the element-wise exponentiation of the vector $\mathbf{x}$.
## Methods
In this paper, we use notations shown in Table 1.
We present SWAM, a CNN-based method for automatic ICD coding from the clinical
text, which provides a good explanation of its internal mechanics.
SWAM is motivated by the way human clinicians manual label the clinical notes,
to help the reader understand the method, firstly here is a brief introduction
of the way human clinicians manual label the clinical notes. Normally, human
clinicians will look for informative snippets that are relevant to each code.
For example, as shown in Figure 1, given code 96.04 in the figure, a human
clinician will look for the presence of relevant snippets in the clinical
notes. In this case, the relevant snippets are “intubation” and “endotracheal
intubation”, if the human clinician finds the relevant snippets, he/she will
give a positive prediction of code 96.04.
SWAM refers to the same idea of manual coding. As shown in Figure 1, the first
step, through the convolutional layer the model will extract informative
snippets that could be relevant to any code. In the second step, the attention
layer will assign importance weight to snippets to select the relevant
snippets of each code, and in the final step the model summary the weighted
score of all relevant snippets of each code to give the predictions of the
presence of each code.
Figure 2: The architecture of the model with per-label mechanism.
### The correspondence between “informative snippet” and convolution filter
Our explanation of the internal mechanics of SWAM builds on the correspondence
between “informative snippet” and convolution filter. Firstly, we classify the
”informative snippet” into two categories: ”generic snippet” and ”non-generic
snippet”. ”generic snippet” refers to snippets that are informative for
multiple labels, for example, in our task, ”experience fever” is likely to be
a ”generic snippet” since it is the symptom correlated with multiple
diagnoses. ”non-generic snippet” refers to snippets that are only considered
as informative to a specific or a few labels, for example, in top-50 code
task, “endotracheal intubation” will be considered as a ”non-generic snippet”
since it brings little information gain to the other 49 labels than it brings
to the code 96.04 “Insertion of endotracheal tube”.
Then we infer that there exists a correspondence between “informative snippet”
and convolution filter, which means one convolution filter can only generate a
high activation value for a specific “informative snippet”. Given that in the
CNN context, the “informative snippet” can be considered as a set of word
embedding sequences that are close in the embedding space. For example, ”large
mucus plug” and ”big mucus plug” are the same ”informative snippet” since they
have similar meanings and therefore are close in the embedding space. It is
most likely that for different ”informative snippets”, they will have very
little chance to be close in the embedding space. For each filter, it will be
”highly activated”
ouput exceeds threshold when the snippet in its window is close to its
parameters in embedding space, and this snippet can be considered as the
”informative snippet” corresponds to this filter.
Based on the above inference, an obvious conclusion is that the choice of the
width of the convolution layers, a.k.a. the number of convolution filters
should depend on the total numbers of ”informative snippets” in the task, more
accurately, the numbers of ”non-generic snippet” since it will much larger
than the numbers of ”generic snippet” in the large-scale coding task. Besides,
empirical guidance in our architecture design is that there could be multiple
”non-generic snippets” for each code [8].
Therefore we develop Shallow and Wide Attention CNN for this task: the
presence of the informative snippet of each code could be considered as a
local, low-level feature learned by the shallow CNN, and we also need the
convolutional layer to be wide since the model needs to learn the ”non-generic
snippets” of all codes.
The mechanism behind Shallow and Wide Attention CNN is general for a set of
similar text classification tasks that informative snippets relevant to each
label scattered at random locations in the input document. So SWAM can be
regarded as a general architecture with the following three characteristics,
and implement details can be varied (e.g. the attention layer in the model can
be either per-label attention mechanism [21] or the full connected layer in
textCNN [14]).
1\. The convolutional layer should be sufficiently wide, a.k.a. enough
convolution filters to not only extract all generic snippets that are
informative for multiple labels, but also all non-generic features that are
correlated to specific label and not informative to other labels, the certain
number of filters depends on task context, a.k.a. the total number of generic
features and non-generic features in the task.
2\. The network architecture should be shallow, this model is designed to
extract snippets of text, which can be considered as local, low-level
features, so a deeper network is unnecessary since informative snippets
relevant to each label scattered at random locations in the input document, it
is not likely that we can earn any benefit from the global, high-level
features by combining the adjacent snippets.
3\. Attention mechanism should be introduced to learn the correlations between
important/informative snippets and each code.
#### Word embedding
The word embedding model used in this paper is the word2vec CBOW method by
Mikolov et al. [20], we pre-train word embedding of size $d_{e}=100$ on the
preprocessed text from all discharge summaries in MIMIC3, which is the same
dataset for training our model. Details about the dataset can be found in
Dataset. We treat ICD code prediction as a multilabel text classification
problem [19]. For clinical note instance i, we want to determine
$y_{i,\ell}\in\ {0,\ 1}$ for all $\ell\in\mathcal{L}$. We train a neural
network which passes text through a convolutional layer to compute a base
representation of the text of each document [14], and makes $\ |\mathcal{L}|$
binary classification decisions.
#### Convolutional Layer
The input of convolutional layer is the clinical notes in form of pre-trained
embeddings representing by the matrix
$\mathbf{X}=[\mathbf{x}_{1},\mathbf{x}_{2},\ldots,\mathbf{x}_{N}]$. The
convolution of adjacent embeddings are computed with a convolution filter
$W_{c}\in\mathbb{R}^{k\times d_{e}\times d_{c}}$. At step n, we compute
$\mathbf{h}_{n}=g(\mathbf{W}_{c}\ast\mathbf{x}_{n:n+k-1}+\mathbf{b}_{c}),$
The input is padded on both sides with zeros so the base representation
$\mathbf{H}$ keeps the same length as $\mathbf{X}$.
#### Attention layer
Nowadays attention mechanism has been generalized and has been employed in
many different forms [35]. The core idea of the attention mechanism can be
regarded as ”giving weight to different parts of the input, to select the part
in the input that is more important for the current task”. So the full
connected layer in textCNN [14] can also be regarded as a kind of ”attention”
since it weighs input separately for each label.
As we mention in The correspondence between “informative snippet” and
convolution filter, SWAM can be regarded as a general CNN architecture, and
implement details can be varied. We adopt two different implements of the
attention layer in our model for different consider considerations. The first
one is the per-label attention mechanism by Mullenbach et al. [21], we adapt
it because it can extract snippets from the clinical text as explanations of
the model prediction, which can be used to verify our conjecture about the
correspondence between “informative snippet” and convolution filter. The
second one is the common full connected layer in textCNN [14], we adapt it
since the textCNN is the basis of many works so it can prove the versatility
of SWAM.
For the per-label attention mechanism (the implement shown in Figure 2), the
idea is to calculate the per-label representation of the document and use an
attention vector $\mathbf{\alpha}_{\ell}$ to represents importance
distribution over locations in the document. To obtain the per-label
representation of the document, formally a vector parameter
$\mathbf{u}_{\ell}\in\mathbb{R}^{d_{c}}$ is used to compute the matrix-vector
product, $\mathbf{H}^{\top}\mathbf{u}_{\ell}$, which can be taken as that the
base representation $\mathbf{H}$ is weighted for label $\ell$. The resulting
vector is then normalized using a SoftMax operation, obtaining
$\mathbf{\alpha}_{\ell}$, the attention vector, the value of the element in
the attention vector is the weighted sum of convolutional features from all
kernels in the same place of the document.
$\mathbf{\alpha}_{\ell}=\operatorname{SoftMax}(\mathbf{H}^{\top}\mathbf{u}_{\ell})$
$\mathbf{\alpha}_{\ell}$ is also taken as the location indicator of the most
important snippet for label $\ell$, every element in $\mathbf{\alpha}_{\ell}$
is corresponding to a location in the document, the value of the element is
seen as the importance of the corresponding location for label $\ell$. The
highest element value in $\mathbf{\alpha}_{\ell}$ means the snippet in this
location is most important (a.k.a. most informative) for the prediction of
label$\ell$. Therefore we obtain an explanation of the prediction in the form
of extracted snippets from the document.
$\alpha_{\ell,n}\mathbf{h}_{n}$, the element-wise vector product is then
computed, applies the attention vectors on the base representation to get the
vector document representations $\mathbf{v}_{\ell}$ for label $\ell$,
$\mathbf{v}_{\ell}=\sum_{n=1}^{N}{\alpha_{\ell,n}\mathbf{h}_{n}}$
For full connected attention, we instead use max-pooling to filter the base
representation down to a vector $\mathbf{v}\in d_{c}$ where every element in
$\mathbf{v}$ corresponds to the highest action value of a convolution filter
in the text,
$v_{j}=\max_{n}h_{n,j}$
#### Classification
Given the vector representation $v_{\ell}$, the likelihood for label $\ell$ is
computed using a linear layer and a non-linear function sigmoid:
$\hat{y}_{\ell}=\sigma\left(\boldsymbol{\beta}_{\ell}^{\top}\boldsymbol{v}_{\ell}+b_{\ell}\right)$
#### Loss function
The training procedure use BCE (binary cross-entropy) as the loss function,
the optimization goal is to minimize the loss.
$L_{\mathrm{BCE}}(X,y)=-\sum_{\ell=1}^{\mathcal{L}}y_{\ell}\log\left(\hat{y}_{\ell}\right)+\left(1-y_{\ell}\right)\log\left(1-\hat{y}_{\ell}\right)$
### Why Shallow and Wide Attention CNN
SWAM can learn a large scale of local and low-level features, it is suitable
for the multi-label text classification task that informative snippets
relevant to each label are not shared. Also, SWAM addresses the challenges of
interpretability: it provides a satisfactory explanation of the internal
mechanics of the deep learning method by establishing the correspondence
between “informative snippet” and convolution filter.
## Results
### Dataset
MIMIC-III [10] is an open-access dataset comprising health data in the form of
text and structured records of ICU admissions. Since MIMIC was built, it has
become the basis of many works on multi-label classification [36, 25].
Following previous works, we train our model on discharge summaries in MIMIC,
which summary records about one stay into a single document. We focus on the
raw text of the data and ignore the attached features like admission time.
Every discharge summary is corresponding to an admission, and each admission
is annotated with a set of ICD-9 codes, describing both diagnoses and
treatments that occurred during the patient’s stay. We train and evaluate SWAM
on a label set consisting of the 50 most frequent labels. We filter the
dataset down to the instances that have at least one of the top 50 most
frequent codes, after the filter, there are 8,067 summaries for training,
1,574 for validation, and 1,730 for testing. Other detailed statistics for the
setting are summarized in Table 2.
Table 2: Descriptive statistics for MIMIC3 | MIMIC-III full | MIMIC-III 50
---|---|---
Training Documents | 47,724 | 8,067
Vocabulary Size | 51,917 | 51,917
Mean Tokens per Document | 1,485 | 1,530
Mean Labels per Document | 15.9 | 5.7
Total Labels | 8922 | 50
#### Preprocessing
We remove the tokens that contain no alphabetic characters (e.g., removing
‘100’ but keeping ‘100ml’ ). For those tokens that appear too few times to
make their semantics difficult to learn, a threshold that only remains tokens
that appear in no fewer than 3 training documents is setting, and all tokens
that failed the threshold are replaced with an ‘UNK’ token. The distribution
of discharge summaries conforms to the long-tailed distribution, 90% of
discharge summaries are short than 2500 tokens, so we truncated discharge
summaries to a maximum length of 2500 tokens.
Table 3: Results on MIMIC-III, 50 labels. Model | AUC | F1 | P@5
---|---|---|---
Macro | Micro | Macro | Micro |
C-MemNN [24] | 0.833 | - | - | - | 0.42
Shi et al.(2017) [32] | - | 0.900 | - | 0.532 | -
CAML [21] | 0.875 | 0.909 | 0.532 | 0.614 | 0.609
Logistic regression | 0.828 | 0.862 | 0.477 | 0.530 | 0.545
SWAM-CAML | 0.900* | 0.924* | 0.593 | 0.648 | 0.625*
SWAM-textCNN | 0.892 | 0.919 | 0.603* | 0.652* | 0.620
### Baselines
As mentioned in The correspondence between “informative snippet” and
convolution filter, SWAM can be regarded as a general CNN architecture and
implement details can be varied. In model part Attention layer two different
implements of attention layer are adapt for different considerations, we name
those two implements as ”SWAM-textCNN”[14] and ”SWAM-CAML” [21] separately to
indicate the attention approaches they refers.
The baseline we compare against is a bag-of-words logistic regression model,
we also compare SWAM-CAML with the origin implement of CAML [21] at the same
setting.
For SWAM-textCNN and SWAM-CAML we initialize the embedding weights using the
same pre-trained word2vec vectors. The logistic regression model consists of
$\ |\mathcal{L}|$ binary one-vs-rest classifiers acting on unigram bag-of-
words features.
#### Parameter tuning
We tune the hyper-parameters of our SWAM-models using grid search.We sample
parameter values for the learning rate $\eta$ ,as well as filter size $k$ ,
number of filters $d_{c}$, and dropout probability $q$ as shown in Table 4. We
also adapt hyper-parameter tuning in the previous works as empirical guidance
[14, 21, 1]. We use a fixed batch size of 16, and train the model with early
stopping, in the case that the f1-macro does not improve for 10 epochs the
training will terminate.
Table 4: Hyper-parameter tuning ranges and optimal values for SWAM model | Range | Optimal Value
---|---|---
$\eta$ | 0.0001,0.0003, | 0.001
(learning rate) | 0.001,0.003 |
$k$ | 1-10 | 4
(filter size) | |
$d_{c}$ | 50-500 | 500
(number of filters) | |
$q$ | 0.2-0.8 | 0.2
(dropout probability) | |
### Evaluation Metrics
We focus on two metrics: Macro-averaged F1 and precision at n (denoted as
’P@n’), which is the fraction of the $n$ highest-scored labels that are
present in the ground truth. The reason we focus on Macro-averaged F1 is that
it pays attention to per-label performance, which can reflect the average
performance of the model on different label tasks. As for P@n, we choose it
because it reflects the performance of the model as a practical decision
support system which presents a fixed number of predicted codes to help user
annotated the clinical text. To facilitate comparison with both future and
prior work, we also report a variety of metrics includes the area under the
ROC curve (AUC) and micro-averaged F1. For recall, Macro-averaged values are
calculated by averaging metrics computed per-label. Micro-averaged values are
calculated by treating each (document, code) pair as a separate prediction.
### Results on quantitative evaluation
Our main quantitative evaluation involves predicting the 50-code set of ICD-9
codes based on the text of the MIMIC-III discharge summaries. These results
are shown in Table 3. We adopt two different implements of attention layer,
named ”SWAM-textCNN” and ”SWAM-CAML” The SWAM models give the strongest
results on all metrics, especially on F1-Macro, which emphasis average
performance over different labels. We attribute this improvement to SWAM, by
which the wide architecture gives the model ability to more extensively learn
the unique features of different codes.
### Ablation experiment on width of network
Figure 3: Comparing per-label performances of SWAM-CAML with CAML [21],the
only difference between two models are the width of the network. SWAM has 500
filters, while CAML has 50 filters.
According to our inference about the correspondence between the informative
snippet and convolution filter, since each ”non-generic snippet” has to
correspond to a convolution filter, if the network is too narrow, the model
will fail to learn the ”non-generic snippets” of some labels. Therefore the
impact of the width of the network can be observed from the perspective of
per-label performance.
We carry out an ablation experiment on the width of the network, we consider
the original implement of CAML [21] as a narrow variant of SWAM, the only
difference between SWAM-CAML and CAML are the width of the network, the former
has 500 convolution filters and the latter has 50. The experiment results are
in line with our expectations. For the narrow model(CAML), the performance of
5 labels is 0, while on the opposite, the wide model(SWAM-CAML) make
significant performance improvement that 4 of 5 labels that with a 0 precision
in the narrow model now have an average precision of 0.53, at the same time
the overall performance of the model is improved. As for the only ICD-9 code
285.9 ”Anemia, unspecified” that has 0 precision in both models, we make a
manual analysis in Section Analysis of the reason behind bad performance code.
### Secondary evaluation
#### Comparing informative snippets extracted by narrow and wide models
To verify our inference about The correspondence between “informative snippet”
and convolution filter, we also compare the informative snippets extracted by
both the narrow (CAML) and the wide (SWAM-CAML) model. In order to make the
cases representative, from the five labels that has 0 precision in the narrow
model, we pick up a label (276.1: Hyposmolality and/or hyponatremia) that has
a improved precision in the wide model, and the only label(285.9 Anemia,
unspecified) that has 0 precision in both models for analyse.
Table 5 shows the informative snippets extracted by the SWAM-CAML and the CAML
model during the prediction of code 276.1 in two random selected documents.
Through a simple analysis, it can be found that the word ”hyponatremia”
extracted by the SWAM model that appears in both the document and the code
description plays an important role in the prediction. While on the opposite,
the snippet extracted by the CAML model is not informative since it has 0
precision on this code.
The word ”hyponatremia”, as a local and low-level feature can be learned by a
single convolution filter according to our inference The correspondence
between “informative snippet” and convolution filter. Since the only
difference between the SWAM-CAML model and the CAML model is the number of
filters in the convolutional layer. Table 5 proves that the performance
difference between the narrow and the wide model comes from the learning of
non-generic ”informative snippet”.
Table 5: informative snippets extracted by the wide (SWAM-CAML) and narrow
(CAML) models for prediction of ICD code 276.1
ICD code 276.1: ”Hyposmolality and/or hyponatremia” SWAM-CAML …dehydration
and increased abd… SWAM-CAML …hyponatremia and possible initiation of
chemotherapy… CAML …peritonitis renal failure and ileus on the floor the
patient was followed by… CAML …renal failure and small bowel obstruction of
note the provided information on…
Table 6: ICD Code with 0 precision in both the wide (SWAM-CAML) and narrow (CAML) models before and after data shuffle Before shuffle | ICD Code with 0 precision
---|---
SWAM-CAML | 285.9 ”Anemia, unspecified”
CAML | 285.9 ”Anemia, unspecified”
| V15.82 ”History of tobacco use”
| 276.1: ”Hyposmolality and/or hyponatremia”
| 305.1 ”Tobacco use disorder”
| 311 ”Depressive disorder, not elsewhere classified”
After shuffle | ICD Code with 0 precision
SWAM-CAML | 285.9 ”Anemia, unspecified”
CAML | 285.9 ”Anemia, unspecified”
| V15.82 ”History of tobacco use”
| 272.0 ”Pure hypercholesterolemia”
| V45.81 ”Postsurgical aortocoronary bypass status”
#### Factors determine which ”non-generic snippet” will fail to be learned by
narrow model
According to our inference, a narrow network will fail to learn the ”non-
generic snippet” of a part of labels, which naturally raises a question: what
factors determine which ”non-generic snippet” will not be learned? The essence
of failing to learn different ”non-generic snippet” is that the model
converges to different local optimal parameters. The convergence result of the
model is related to the distribution of data during training, in other words,
the order that ”non-generic snippets” appear during training: once a filter
learns a specific ”non-generic snippet” for some labels, the loss function
will encourage it to keep its parameters unchanged, and after all the filters
have learned corresponding ”informative snippet”, it’s difficult for the model
to leave the current local optimal solution and learn new ”informative
snippet”.Therefore we shuffle the training dataset with a different random
seed and re-train the models with the shuffled dataset. The results are shown
in Table 6.
The results in Table 6 are in line with our expectations, after shuffle and
re-training, the 0 precision labels in the narrow model (CAML) change. On the
opposite, the only 0 precision label 285.9 in the wide model (SWAM-CAML) still
can not be learned. The distribution of data during training is a factor that
determines which ”non-generic snippet” will fail to be learned by the narrow
model. And the reason behind the bad performance ICD code 285.9 ”Anemia,
unspecified” is something beyond the local optimum.
#### Analysis of the reason behind bad performance code
The ICD-9 code 285.9 ”Anemia, unspecified” is failed to be predicted by both
narrow (CAML) and wide (SWAM) models. Through manual analyzing, we found there
are more than 50 codes in the ICD-9 that are in form of “Anemia + specific
reason”, which means the snippets related to anemia cannot are necessary but
not sufficient for prediction of code 285.9. The prediction of 285.9 Anemia,
unspecified is not only based on the presence of snippet related to ‘Anemia’,
it is also based on the information that all possible reasons are absent. This
is a blind spot in all current machine learning models. It is difficult for
models to learn inferences based on missing information.
## Discussion
For future work, we are considering several different directions. From the
application perspective, since our approach does work well on 50 labels task,
the next step is to apply the approach to the full code set. A major challenge
is for full code set, we may need tens of thousands of convolution filters, as
the number of filters in the network increases, unnecessary overlap in the
features captured by the network’s filters will also increase [23]. We plan to
address this challenge by adapting the ensemble method, we plan to cluster the
ICD codes and train a classifier for each clustered subset. From the
linguistic perspective, we plan to explore reasons behind hard-to-learn codes
such as ICD-9 code 285.9 ”Anemia, unspecified”, and leverage the hierarchy of
ICD codes to improve performance on these codes.
## Conclusion
We present SWAM, an explainable CNN approach for multi-label document
classification, which employs a wide convolution layer to learn local and low-
level features for each label, yields strong improvements over previous
metrics on the ICD-9 code prediction task, while providing satisfactory
explanations for its internal mechanics.
## Abbreviations
ICD:International Classification of Diseases;CNN:Convolutional Neural
Network;ICU:Intensive care unit;SWAM:Shallow and Wide Attention convolutional
Mechanism;MIMIC-III:Medical Information Mart for Intensive Care
III;NLP:Natural Language;CBOW:Continuous Bag-of-Words;BCE:Binary cross-
entropy;CAML:Convolutional Attention for Multi-Label
classification;F1:F-measure
## Availability of data and materials
The datasets is available from https://mimic.physionet.org/.
## Competing interests
The authors declare that they have no competing interests.
## Author’s contributions
Hu S.Y. conceived of the presented idea, developed the theory and carry out
the experiments. Teng F. encouraged Hu S.Y. to investigate per-label
performance of the model and supervised the findings of this work. All authors
discussed the results. Hu S.Y. wrote the manuscript under supervising from
Teng F., all authors have read and approved the final manuscript.
## Funding
No funding to declare.
## Ethics approval and consent to participate
Not applicable.
## Consent for publication
Not applicable.
## Acknowledgements
None.
## References
* [1] Ahmad Aghaebrahimian and Mark Cieliebak. Hyperparameter tuning for deep learning in natural language processing. In 4th Swiss Text Analytics Conference (SwissText 2019), Winterthur, June 18-19 2019. Swisstext, 2019.
* [2] Miltiadis Allamanis, Hao Peng, and Charles Sutton. A convolutional attention network for extreme summarization of source code. In International conference on machine learning, pages 2091–2100, 2016.
* [3] Anand Avati, Kenneth Jung, Stephanie Harman, Lance Downing, Andrew Ng, and Nigam H Shah. Improving palliative care with deep learning. BMC medical informatics and decision making, 18(4):122, 2018.
* [4] Dzmitry Bahdanau, Kyunghyun Cho, and Yoshua Bengio. Neural machine translation by jointly learning to align and translate. arXiv preprint arXiv:1409.0473, 2014.
* [5] Edward Choi, Mohammad Taha Bahadori, Andy Schuetz, Walter F Stewart, and Jimeng Sun. Doctor ai: Predicting clinical events via recurrent neural networks. In Machine Learning for Healthcare Conference, pages 301–318, 2016\.
* [6] Ronan Collobert and Jason Weston. A unified architecture for natural language processing: Deep neural networks with multitask learning. In Proceedings of the 25th international conference on Machine learning, pages 160–167, 2008.
* [7] Joshua C Denny, Marylyn D Ritchie, Melissa A Basford, Jill M Pulley, Lisa Bastarache, Kristin Brown-Gentry, Deede Wang, Dan R Masys, Dan M Roden, and Dana C Crawford. Phewas: demonstrating the feasibility of a phenome-wide scan to discover gene–disease associations. Bioinformatics, 26(9):1205–1210, 2010.
* [8] Linyuan Gong and Ruyi Ji. What does a textcnn learn? arXiv preprint arXiv:1801.06287, 2018.
* [9] G W Hunninghake and J E Gadek. The alveloar macrophage. In T J R Harris, editor, Cultured Human Cells and Tissues, pages 54–56. Academic Press, New York, 1995. Stoner G (Series Editor): Methods and Perspectives in Cell Biology, vol 1.
* [10] Alistair EW Johnson, Tom J Pollard, Lu Shen, H Lehman Li-Wei, Mengling Feng, Mohammad Ghassemi, Benjamin Moody, Peter Szolovits, Leo Anthony Celi, and Roger G Mark. Mimic-iii, a freely accessible critical care database. Scientific data, 3(1):1–9, 2016.
* [11] X Jones. Zeolites and synthetic mechanisms. In Y Smith, editor, Proceedings of the First National Conference on Porous Sieves: 27-30 June 1996; Baltimore, pages 16–27. Stoneham: Butterworth-Heinemann, 1996.
* [12] Ramakanth Kavuluru, Anthony Rios, and Yuan Lu. An empirical evaluation of supervised learning approaches in assigning diagnosis codes to electronic medical records. Artificial intelligence in medicine, 65(2):155–166, 2015.
* [13] S A Kharitonov and P J Barnes. Clinical aspects of exhaled nitric oxide. in press.
* [14] Yoon Kim. Convolutional neural networks for sentence classification. arXiv preprint arXiv:1408.5882, 2014.
* [15] R Kohavi. Wrappers for performance enhancement and obvious decision graphs. PhD thesis, Stanford University, Computer Science Department, 1995.
* [16] E V Koonin, S F Altschul, and P Bork. Brca1 protein products: functional motifs. Nat Genet, 13:266–267, 1996.
* [17] Hoa T Le, Christophe Cerisara, and Alexandre Denis. Do convolutional networks need to be deep for text classification? arXiv preprint arXiv:1707.04108, 2017.
* [18] L Margulis. Origin of Eukaryotic Cells. Yale University Press, New Haven, 1970.
* [19] Andrew Kachites McCallum. Multi-label text classification with a mixture model trained by em. In AAAI 99 workshop on text learning. Citeseer, 1999.
* [20] Tomas Mikolov, Ilya Sutskever, Kai Chen, Greg S Corrado, and Jeff Dean. Distributed representations of words and phrases and their compositionality. In Advances in neural information processing systems, pages 3111–3119, 2013.
* [21] James Mullenbach, Sarah Wiegreffe, Jon Duke, Jimeng Sun, and Jacob Eisenstein. Explainable prediction of medical codes from clinical text. arXiv preprint arXiv:1802.05695, 2018.
* [22] C A Orengo, J E Bray, T Hubbard, L LoConte, and I Sillitoe. Analysis and assessment of ab initio three-dimensional prediction, secondary structure, and contacts prediction. Proteins, Suppl 3:149–170, 1999.
* [23] Aaditya Prakash, James Storer, Dinei Florencio, and Cha Zhang. Repr: Improved training of convolutional filters. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 10666–10675, 2019.
* [24] Aaditya Prakash, Siyuan Zhao, Sadid A Hasan, Vivek Datla, Kathy Lee, Ashequl Qadir, Joey Liu, and Oladimeji Farri. Condensed memory networks for clinical diagnostic inferencing. In Proceedings of the Thirty-First AAAI Conference on Artificial Intelligence, pages 3274–3280, 2017.
* [25] Alvin Rajkomar, Eyal Oren, Kai Chen, Andrew M Dai, Nissan Hajaj, Michaela Hardt, Peter J Liu, Xiaobing Liu, Jake Marcus, Mimi Sun, et al. Scalable and accurate deep learning with electronic health records. NPJ Digital Medicine, 1(1):18, 2018.
* [26] Rajesh Ranganath, Adler J Perotte, Noémie Elhadad, and David M Blei. The survival filter: Joint survival analysis with a latent time series. In UAI, pages 742–751, 2015.
* [27] Tim Rocktäschel, Edward Grefenstette, Karl Moritz Hermann, Tomáš Kočiskỳ, and Phil Blunsom. Reasoning about entailment with neural attention. arXiv preprint arXiv:1509.06664, 2015.
* [28] Alexander M Rush, Sumit Chopra, and Jason Weston. A neural attention model for abstractive sentence summarization. arXiv preprint arXiv:1509.00685, 2015.
* [29] Cicero dos Santos, Ming Tan, Bing Xiang, and Bowen Zhou. Attentive pooling networks. arXiv preprint arXiv:1602.03609, 2016.
* [30] Elyne Scheurwegs, Boris Cule, Kim Luyckx, Léon Luyten, and Walter Daelemans. Selecting relevant features from the electronic health record for clinical code prediction. Journal of biomedical informatics, 74:92–103, 2017.
* [31] E Schnepf. From prey via endosymbiont to plastids: comparative studies in dinoflagellates. In R A Lewin, editor, Origins of Plastids, volume 2, pages 53–76. Chapman and Hall, New York, 2nd edition, 1993.
* [32] Haoran Shi, Pengtao Xie, Zhiting Hu, Ming Zhang, and Eric P Xing. Towards automated icd coding using deep learning. arXiv preprint arXiv:1711.04075, 2017.
* [33] Y Smith, editor. Proceedings of the First National Conference on Porous Sieves: 27-30 June 1996; Baltimore, Stoneham, 1996. Butterworth-Heinemann.
* [34] Michael Subotin and Anthony Davis. A system for predicting icd-10-pcs codes from electronic health records. In Proceedings of BioNLP 2014, pages 59–67, 2014.
* [35] Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, Łukasz Kaiser, and Illia Polosukhin. Attention is all you need. In Advances in neural information processing systems, pages 5998–6008, 2017.
* [36] Sen Wang, Xiaojun Chang, Xue Li, Guodong Long, Lina Yao, and Quan Z Sheng. Diagnosis code assignment using sparsity-based disease correlation embedding. IEEE Transactions on Knowledge and Data Engineering, 28(12):3191–3202, 2016.
* [37] Zichao Yang, Diyi Yang, Chris Dyer, Xiaodong He, Alex Smola, and Eduard Hovy. Hierarchical attention networks for document classification. In Proceedings of the 2016 conference of the North American chapter of the association for computational linguistics: human language technologies, pages 1480–1489, 2016.
* [38] Wenpeng Yin and Hinrich Schütze. Attentive convolution: Equipping cnns with rnn-style attention mechanisms. arXiv preprint arXiv:1710.00519, 2017.
* [39] Wenpeng Yin, Hinrich Schütze, Bing Xiang, and Bowen Zhou. Abcnn: Attention-based convolutional neural network for modeling sentence pairs. Transactions of the Association for Computational Linguistics, 4:259–272, 2016.
* [40] Danchen Zhang, Daqing He, Sanqiang Zhao, and Lei Li. Enhancing automatic icd-9-cm code assignment for medical texts with pubmed. In BioNLP 2017, pages 263–271, 2017.
* [41] N J Zvaifler, J A Burger, L Marinova-Mutafchieva, P Taylor, and R N Maini. Mesenchymal cells, stromal derived factor-1 and rheumatoid arthritis [abstract]. Arthritis Rheum, 42:s250, 1999.
|
labelinglabel
# Analysis of basic emotions in texts
based on bert vector representation
A. Artemov, A. Veselovskiy, I. Khasenevich, I. Bolokhov
Kognitivnie Sistemi
<EMAIL_ADDRESS>
###### Abstract
In the following paper the authors present a GAN-type model and the most
important stages of its development for the task of emotion recognition in
text. In particular, we propose an approach for generating a synthetic dataset
of all possible emotions combinations based on manually labelled incomplete
data.
Keywords: 7-dimensional emotional model, Pytorch, GAN, deep learning, cosine
similarity, Ekman, BERT, NLP, BRAIN2NLP, emotions, collisions, Big Data
## Introduction
There are several approaches to classification of human emotions. Authors
consider the methodology according to P.Ekman to be the most reasonable [4].
In accordance with it, the following seven human emotions can be
distinguished: fear, sadness, anger, disgust, calm, happiness, surprise. These
emotions were classified according to the analysis of facial expressions of a
person during events of different semantic meaning. The unambiguous perception
of facial expressions with different people and nations was the proof of the
sufficiency and validity of the seven basic emotions.
At the same time, although there is a number of solutions for emotional
recognition, a standard for the number of basic emotions for text data has not
yet been created. For example, in Rodrigo Masaru Ohashi’s work [7] the model
of bidirectional LSTM with a convolutional neural network layer was
successfully trained to identify 4 emotions: joy, fear, sadness and anger.
Chew-Yean Yam’s article “Emotion Detection and Recognition from Text Using
Deep Learning [2]” describes the neural network with 3 hidden layers of 125,
25 and 5 neurons for determining 5 emotions: anger, sadness, fear, happiness,
excitement. Jangwon Park [8] created a Huggingface Transformers based model
that predicts 7 emotions, the F1-measure of the model is 59.81 and 61.48 for
the train and test dataset, respectively. This model is based on the
GoEmotions [3] dataset with 27 emotion classes (see Table 5).
Let us assume, as a hypothesis, that the 7 basic emotions by Ekman are
applicable to texts. That is, the labelling of texts for this set of emotion
classes by a person is complete. Based on the possibility of obtaining such a
dataset, the task is to train a model that can predict the vector of seven
emotions for an arbitrary text. The result of this work is described in the
article.
A positive solution to this problem opens up prospects for combining different
data channels (audio and video) to determine emotions when machines are
communicating with humans. Moreover, of particular interest is the description
of information spread through means of mass communication (social networks),
taking into account the phenomenon of emotions social exchange.
## 1 Dataset vectorization
Initially, it is necessary to form a dataset for training the model. The
dataset should contain data about seven basic emotions for each certain piece
of text - a paragraph or a sentence. The authors couldn’t find a publicly
available dataset with this or similar data. Therefore, it was decided to form
it automatically using data on the emotional reaction of people, expressed in
texts. Such data were emoticons in texts from Twitter, YouTube and other
social networks.
The original texts were divided into sentences, emoticons were extracted from
each sentence. In order to vectorize emotions, a special dictionary was
compiled containing emoticons divided into emotion types according to uniquely
interpreted classes:
The classification of the texts’ emotions was carried out using the created
dictionary. By counting the number of encounters of each emotion in the text,
vectors of 7 emotions were generated for each sentence.
Text vectorization was carried out using our own NLP framework -
BRAIN2NLP111Russian cognitive processor for automated labelling of natural
text. The main functionality of BRAIN2NLP includes language detection, stem
extraction, morphological and syntactic labelling of text, extraction of
nominal entities and definition of a semantic vector of words.. The text
vectorization was based on the BERT language model trained by Google. Thus, at
the output, pairs of the following form were obtained: 512-dimensional vector
BERT to seven-dimensional vector of emotion. In total, more than 100,000 such
pairs were constructed. In the process, it turned out that the resulting
dataset contains a lot of conflicting data (collisions). Collisions are
examples of data, where the same properties of an object lead to different
classes or to a significant number of classes. The automation of the collision
search process required the development of a small script. The algorithm of
its work is given below.
* 1.
Determine the allowed fraction of the k-number of classes to which objects in
the same cluster can belong.
* 2.
Classify dataset objects by the vector of their properties (features).
* 3.
Assign the same ID to dataset objects belonging to the same cluster.
* 4.
Add up the values by class for objects with the same ID.
* 5.
Estimate for each ID the mathematical expectation of the vector of values by
class.
* 6.
Calculate the number of Z values of the class vector that is greater than or
equal to the expected value.
* 7.
Evaluate each cluster, and where $Z>k$ mark it as a collision.
* 8.
Assign a collision mark to all objects in the cluster.
Algorithm 1 Finding collisions in a dataset
For the purpose of "human" marking, 10,000 texts were selected, among which
there were only 6913 examples without collisions with a length of no more than
two sentences. The experts were asked to assign no more than 2 emotions to
each of these texts. But the dataset formed in this way also contained minor
examples of collisions, which were also eliminated [1]. As a result, a golden
dataset with 2813 examples was obtained.
## 2 Search for Architecture
BERT-vectors are successfully used for classification tasks of texts [9] and
their fragments [5]. Considering this fact, we have chosen these vectors as a
basic description of the space of text properties, which can be used to
predict the emotion vector.
During the search for the optimal solution for the classification of emotions,
different variants of embedding length BERT (lengths 768 and 512) and
different sizes of datasets were tested. Also, we have tested a certain number
of negative samples, various dictionaries for the extraction of emoticons,
normalization of the dataset, different model architectures, loss functions
and activation functions. As a result of this work, the following hypothesis
was formulated and confirmed in practice: It is possible to improve the
quality of classification by abandoning the attempt to normalize the
unbalanced initial data - the number of examples for a set of classes,
replacing it with uniformly generated BERT vectors for each set of classes.
Among all tested neural network architectures, Generative adversarial network
(GAN) [6] showed the best results - an architecture consisting of a generative
network and a discriminative network configured to work against each other.
Discriminative models learn the boundary between the classes; generative model
generates new data instances. A collision free dataset (Dataset_1) was used
for training the generative network and for final testing of the
discriminative network. In the formulation of the study, the vector of
emotions has 7 dimensions, we denote it as V1. Table 1 shows a list of
combinations of vector V1 values found in Dataset_1.
Table 1: Combinations of emotion vectors Dataset_1 # | fear | sadness | anger | disgust | calm | happiness | surprise
---|---|---|---|---|---|---|---
0 | 0 | 0 | 0 | 0 | 0 | 0 | 1
1 | 0 | 0 | 0 | 0 | 0 | 1 | 0
2 | 0 | 0 | 0 | 0 | 0 | 1 | 1
3 | 0 | 0 | 0 | 0 | 1 | 0 | 0
4 | 0 | 0 | 0 | 0 | 1 | 0 | 1
5 | 0 | 0 | 0 | 0 | 1 | 1 | 0
6 | 0 | 0 | 0 | 0 | 1 | 1 | 1
7 | 0 | 0 | 0 | 1 | 0 | 0 | 0
8 | 0 | 0 | 0 | 1 | 0 | 0 | 1
9 | 0 | 0 | 0 | 1 | 1 | 0 | 0
10 | 0 | 0 | 1 | 0 | 0 | 0 | 0
11 | 0 | 0 | 1 | 0 | 0 | 0 | 1
12 | 0 | 0 | 1 | 0 | 0 | 1 | 0
13 | 0 | 0 | 1 | 0 | 0 | 1 | 1
14 | 0 | 0 | 1 | 0 | 1 | 0 | 0
15 | 0 | 0 | 1 | 0 | 1 | 0 | 1
16 | 0 | 0 | 1 | 0 | 1 | 1 | 0
17 | 0 | 0 | 1 | 1 | 0 | 0 | 0
18 | 0 | 0 | 1 | 1 | 0 | 0 | 1
19 | 0 | 1 | 0 | 0 | 0 | 0 | 0
20 | 0 | 1 | 0 | 0 | 0 | 0 | 1
21 | 0 | 1 | 0 | 0 | 0 | 1 | 0
22 | 0 | 1 | 0 | 0 | 1 | 0 | 0
23 | 0 | 1 | 0 | 0 | 1 | 0 | 1
24 | 0 | 1 | 0 | 0 | 1 | 1 | 0
25 | 0 | 1 | 0 | 1 | 0 | 0 | 0
26 | 0 | 1 | 1 | 0 | 0 | 0 | 0
27 | 0 | 1 | 1 | 0 | 0 | 0 | 1
28 | 0 | 1 | 1 | 1 | 0 | 0 | 0
29 | 1 | 0 | 0 | 0 | 0 | 0 | 0
30 | 1 | 0 | 0 | 0 | 0 | 0 | 1
31 | 1 | 0 | 0 | 0 | 0 | 1 | 1
32 | 1 | 0 | 0 | 0 | 1 | 0 | 0
33 | 1 | 0 | 0 | 0 | 1 | 0 | 1
34 | 1 | 0 | 0 | 1 | 0 | 0 | 0
35 | 1 | 0 | 1 | 0 | 0 | 0 | 1
36 | 1 | 1 | 0 | 0 | 0 | 0 | 0
37 | 1 | 1 | 1 | 0 | 0 | 0 | 0
Since 37 combinations222Not entirely confident in ”Les 36 situations
dramatiques” concept, we decided to generate all possible combinations of
emotions. do not make up a complete list of 128 combinations, the missing ones
had to be generated.
For this purpose, based on Dataset_1, we have trained a generator model for
known combinations of emotions. Then the combinations V1 were initiated, where
the initial - has the values [0, 0, 0, 0, 0, 0, 0], and the final one is [1,
1, 1, 1, 1, 1, 1]. The total number of vectors V1 was 128. Using the
generator, 128 BERT - V2 vectors were obtained corresponding to various V1
combinations.
Thus, a new dataset (Dataset_2) was formed, where each vector V1 corresponded
to the generated vector V2 (Figure 1). Dataset_2 was used for fine-tuning the
Generator model and Discriminator training, predicting the emotion vector from
the BERT vector of the text. Figure 1 shows the process of sequential training
of the Generator, and Figure 2 shows the Discriminator.
Figure 1: The process of generator training with the formation of Dataset_2
Figure 2: The process of joint training of the Generator and the Discriminator
Figure 3 shows graphs of the "descent rate" when training the Generator and
the Discriminator. The large value of the generator error is explained by the
fact that a simple product of vectors is used, and the error is considered by
MSE. The opposite situation is with the discriminator, where not only the
output value is normalized, but also the error is calculated by the cosine
measure.
a)
b)
Figure 3: Gradient descent of a) Generator model; b) Discriminator model
In order to move to a non-negative real number, the output vector of the
discriminator value was normalized as follows
$M_{ij}^{\ast}=\dfrac{M_{ij}-min|M_{ij}|}{\sum_{M_{ij}\in
V}(M_{ij}-min|M_{ij}|)}$ (1)
where V is a vector of forecast values.
## 3 Results
Model training was performed on 70% of the data with testing on the remaining
30%. The evaluation of the quality of the model is illustrated in Figure 4.
Comparative results of the model training process are presented in Table 2.
Table 2: Accuracy rating Emotion | Accuracy | Figure 4: The matrix of the predictive distribution and the real value of the Discriminator based on Dataset_1
---|---|---
fear | 0.62
sadness | 0.70
anger | 0.63
disgust | 0.76
calm | 0.64
happiness | 0.68
surprise | 0.54
MEAN | 0.65
Table 3: The result of the model training process
Network | Dataset | Architecture | Learn
---|---|---|---
| Collisions
---
Train | Test | | 1-st layer
---
| 2-nd layer
---
| Weight Initialization
---
| Activation Function
---
Error | | Gradient descent
---
Epochs | | Train accuracy
---
| Test accuracy
---
| Generator
---
Yes | 4855 | 2082 | 7x128 | 128x512 | Random | No | MSE | ADAM | 10 | 0.99 | 0.99
| Discriminator
---
Yes | 89 | 39 | 512x7 | No | | Metod
---
FM*
| Cosine
---
similarity
| Cosine
---
Loss
ADAM | 50 | 0.75 | 0.81
| Generator
---
No | 1969 | 844 | 7x128 | 128x512 | Random | No | MSE | ADAM | 10 | 0.96 | 0.9
| Discriminator
---
No | 89 | 39 | 512x7 | No | | Metod
---
FM*
| Cosinus
---
similarity
| Cosine
---
Loss
ADAM | 50 | 0.74 | 0.87
*FM Method (frequency-matrix) is a method for forming a frequency matrix based on pairs of property and class vectors. The original dataset of the emotion vector and BERT vector pairs of text was divided into two matrices of emotion vectors and BERT vectors. Then, by multiplying the obtained matrices, a frequency matrix is formed.
Based on the input data, the model makes a forecast in the form of a
7-dimensional vector of emotions. To be able to compare it with gold dataset,
it was converted into 2-dimensional vector by selecting emotion classes with
maximum values. If at least one of the two emotions coincided with the gold
labelling, then it was considered that the emotion was predicted correctly.
The accuracy of the forecast for such an assessment varies depending on the
class of emotion (see Table 3). The average accuracy of emotion predictions is
0.65.
The results of the model for different words with different emotional
sentiment are shown in Table 4.
Table 4: An example of model’s evaluation the emotional sentiment of a sentence - | Example of sentence | Expert’s Emotion | Predicted emotion by the Model
---|---|---|---
fear | sadness | anger | disgust | calm | happiness | surprise
1 | | There is danger all around,
---
it’s scary to go outside.
| fear,
---
sadness
0.49 | 0.12 | 0.17 | 0.13 | 0.00 | 0.00 | 0.09
2 | | It is very sad that nothing
---
can be done. The city
burned down and everyone
died
| sadness,
---
fear
0.18 | 0.27 | 0.21 | 0.27 | 0.00 | 0.00 | 0.07
3 | | I am super angry,
---
I’m in a fury!
You’re dead, bastards!
anger | 0.21 | 0.15 | 0.22 | 0.19 | 0.00 | 0.12 | 0.11
4 | | The soup was terrible,
---
I’ve never tasted such
disgusting meal. I’ve been
sick of it for an hour.
| disgust,
---
anger
0.17 | 0.26 | 0.15 | 0.30 | 0.00 | 0.07 | 0.06
5 | | Relax and listen to nature, feel this
---
light summer wind. The
silence calms better than a
thousand words.
| calm,
---
happiness
0.03 | 0.20 | 0.19 | 0.14 | 0.14 | 0.30 | 0.00
6 | | What could be better than
---
watching the sunset together
with close friends!
Just wonderful!
| happiness,
---
calm
0.01 | 0.05 | 0.10 | 0.03 | 0.00 | 0.63 | 0.18
7 | | John won more than
---
$1 million in the lottery,
that was a pleasant surprise
for him.
| surprise,
---
happiness
0.00 | 0.04 | 0.03 | 0.15 | 0.07 | 0.47 | 0.25
Based on the study of the data on model’s performance, the following
conclusions were made:
* -
in all cases, the model correctly identified at least one of the dominant
emotions, a little less often – two of the two.
* -
the model does not define the calm emotion well, in most cases its value is
the lowest.
* -
Intends to show more emotions than necessary (see examples in the first five
sentences of Table 4).
That is, the model "gives a chance" to emotions that were not defined in the
gold dataset. We consider this to be a special property of the model rather
than a bug, since the model was trained on a larger variety (of generated
data) than those observed by experts. This is a very important consequence
that needs to be taken into account when reconstructing complete data based on
expert assessments.
## 4 Conclusion
Among all the architectures that were tested, GAN performed better than the
others (see Table 5). When compared with the models of emotions of some other
authors, our model showed the best result. The F1-measure value is 0.025
higher than the "goemotions-pytorch" model, and the accuracy value is 0.0053
higher than the "Emotion Detection and Recognition from Text Using Deep
Learning" model, but 0.19 lower than the "Bidirectional LSTM with a
Convolutional Neural Network"model. In the latter case, it is worth saying
that in our model the number of defined emotions is seven against four in
"Bidirectional LSTM with a Convolutional Neural Network".
Table 5: Comparing the quality of different emotion models Model | Emotions | F1 macro | Accuracy
---|---|---|---
| «Bidirectional LSTM with a Convolutional Neural
---
Network» by Rodrigo Masaru Ohashi (2019)
4 | - | 0.84
| «Emotion Detection and Recognition from Text Using
---
Deep Learning» by Chew-Yean Yam (2015)
5 | - | 0.64
«GoEmotions-pytorch» by Jangwon Park (2020) | 7 | 0.61 | -
GAN BERT-EMO model (2020) | 7 | 0.64 | 0.65
The hypothesis about the possibility of improving the quality of
classification by refusing to normalize the class distribution of a real
sample and replacing it with a sample obtained by generating all possible
combinations of classes (BERT vectors) for a finite number of variables was
fully justified. The authors have not encountered such an approach in other
publications. At the same time, it opens up new perspectives for machine
learning with minimizing human involvement by automatically generating "big
data" for training neural networks on "small" incomplete data.
## References
* [1] A.Artemov, & A. Veselovskiy. CLASSIFICATION-OF-BASIC-EMOTIONS-USING-BERT. 2021. https://github.com/cogsyscompany/datasets
* [2] Chew-Yean. Emotion Detection and Recognition from Text Using Deep Learning. November 29, 2015. https://devblogs.microsoft.com/cse/2015/11/29/emotion-detection-and-recognition-from-text-using-deep-learning/
* [3] D. Demszky, D. Movshovitz-Attias, Jeongwoo Ko, A. Cowen, G. Nemade, & Sujith Ravi. GoEmotions: A Dataset of Fine-Grained Emotions. 2020. https://arxiv.org/abs/2005.00547
* [4] P. E.Ekman, , & R. J.Davidson, (1994). The Nature of Emotion Fundamental Questions. Oxford Oxford University Press.
* [5] Y. Feng, Y. Wang, & Hang Li. A Sequence-to-Sequence Approach to Dialogue State Tracking. 2020. https://arxiv.org/pdf/2011.09553.pdf
* [6] Ian J. Goodfellow, Jean Pouget-Abadie, Mehdi Mirza, Bing Xu, David Warde-Farley, Sherjil Ozair, Aaron Courville, Yoshua Bengio. Generative Adversarial Networks. 2014. https://arxiv.org/abs/1406.2661
* [7] R. M. Ohashi. From Sentiment Analysis to Emotion Recognition: A NLP story. 2019. https://medium.com/neuronio/from-sentiment-analysis-to-emotion-recognition-a-nlp-story-bcc9d6ff61ae
* [8] Jangwon Park. GoEmotions Pytorch. 2020. https://github.com/monologg/GoEmotions-pytorch
* [9] M. Tan, & J. Jiang. A BERT-based Dual Embedding Model for Chinese Idiom Prediction. 2020. https://arxiv.org/abs/2011.02378
|
11institutetext: Instituto de Astrofísica de Canarias, Vía Láctea S/N, La
Laguna 38205, Tenerife, Spain
11email<EMAIL_ADDRESS><EMAIL_ADDRESS>22institutetext: Departamento de
Astrofísica, Universidad de La Laguna, 38205, Tenerife, Spain
# Mapping the Sun’s upper photosphere with artificial neural networks
H. Socas-Navarro 1122 A. Asensio Ramos 1122
(Received ; accepted )
We have developed an inversion procedure designed for high-resolution solar
spectro-polarimeters, such as Hinode/SP or DKIST/ViSP. The procedure is based
on artificial neural networks trained with profiles generated from random
atmospheric stratifications for a high generalization capability. When applied
to Hinode data we find a hot fine-scale network structure whose morphology
changes with height. In the middle layers this network resembles what is
observed in G-band filtergrams but it is not identical. Surprisingly, the
temperature enhancements in the middle and upper photosphere have a reversed
pattern. Hot pixels in the middle photosphere, possibly associated to small-
scale magnetic elements, appear cool at the $\log\tau_{500}$=$-3$ and $-4$
level, and viceversa. Finally, we find hot arcs on the limb side of magnetic
pores, which we interpret as the first direct observational evidence of the
”hot wall” effect in temperature.
###### Key Words.:
Sun: photosphere – Sun: faculae, plages – Sun: magnetic fields – Methods:
numerical – Methods: data analysis
## 1 Introduction
Inversion techniques allow us to retrieve information encoded in spectral
lines about the atmospheres where they form. A wide variety of strategies have
been employed for decades in solar physics to interpret spectroscopic and
spectropolarimetric observations (see e.g. the reviews by del Toro Iniesta &
Ruiz Cobo 2016; Bellot Rubio 2006; Socas-Navarro 2001). Most applications are
based on the least-squares fitting of the observed spectral lines with
synthetic profiles, which are computed from model atmospheres whose parameters
are iteratively adjusted until a satisfactory fit is attained. However,
advances in instrumentation are driving an increasing interest in the
exploration of alternative methods. Two-dimensional spectropolarimetry is now
very common and fast growing data rates motivate the exploration of new
algorithms that have the potential of being faster and/or more robust for
systematic application.
Artificial neural networks (ANNs) offer a promising new approach for many
purposes where profile fitting is inadequate because one needs a faster or a
more robust performance. The first applications of ANNs in solar physics are
almost 20 years old, dating back to Carroll & Staude (2001) and Socas-Navarro
(2002). However, while those first efforts produced encouraging results, ANN
inversions were not immediately adopted by the community, for two reasons
mainly. First, disentangling the magnetic filling factor from the intrinsic
field strength has proven extremely challenging, as noted since those early
works (Socas-Navarro 2003). The magnetic field tends to exhibit small-scale
structures in the solar photosphere. In arc-second resolution observations, it
is common to find pixels where the magnetic field occupies less than 10% of
the resolution element. This area fraction is referred to as the filling
factor and it introduces an important complication for ANN inversions. Second,
a more practical issue is the complexity involved in the coding of algorithms
for the training of an ANN model.
Those early problems have been largely resolved in recent years, leading to a
renewed interest in ANNs (e.g., Liu et al. 2020; Guo et al. 2020; Díaz Baso &
Asensio Ramos 2018; Asensio Ramos & Díaz Baso 2019; Felipe & Asensio Ramos
2019; Milić & Gafeira 2020). Current (and upcoming) instrumentation are
delivering very high resolution observations, mitigating the filling factor
problem. Furthermore, there has been a tremendous development in the field of
deep learning and many sophisticated tools have been made publicly available
to simplify the problem of building and training ANNs (e.g., Abadi et al.
2015; Paszke et al. 2019).
In this paper we use a relatively simple ANN model and take a different
approach to previous work for the training strategy. Instead of using a
simulation snapshot as the starting point for a training set, as in Asensio
Ramos & Díaz Baso (2019) or Milić & Gafeira (2020), we create a database of
profiles from random stratifications of the relevant parameters. This provides
a wide coverage of the parameter space and guarantees that the ANN is not
specialized on any particular scenario. Unlike Asensio Ramos & Díaz Baso
(2019), whose ANN performs a full inversion of the entire 2D field at once,
this ANN works on each pixel independently. In that regard, it is more similar
to a traditional inversion technique.
We created two different ANNs, one to invert photospheric observations from
DKIST/ViSP (Daniel K Inouye Solar Telescope/Visible Spectro-Polarimeter, see
Rimmele et al. 2020) and the other one for the Hinode satellite’s SOT/SP
(Solar Optical Telescope/Spectro-Polarimeter). DKIST/ViSP data are not yet
available so we focus here on the analysis of the Hinode inversions. After
testing the procedure with synthetic data and previous inversions of real
observations, we applied it to Hinode observations of active regions. In this
manner we obtained datacubes with a fairly unique combination of high spatial
resolution, large field of view and depth-dependent temperatures. These maps
show a fine hot network in active regions, particularly around sunspots and
pores.
We find some surprising results in this application, such as an
anticorrelation between hot pixels in middle and upper layers. Also, the
inversions reveal a series of hot arcs running along the limb side of pores in
the observed regions. We interpret these arcs as the first direct observation
of the ”hot wall” effect, a prediction of fluxtube models since the work of
Spruit (1976) which had not been directly observed thus far.
## 2 The ANN model and training set
All the calculations presented in this paper were produced with relatively
standard computer hardware. We employed a Linux workstation powered by eight
3·GHz Intel Xeon cores. The system is equipped with a GTX 1080 GPU that
handles most of the ANN-related processing. Our ANN model and codes are
publicly available in a
repository111https://github.com/hsocasnavarro/Paper_SNAR21.
The ANN is created and trained using PyTorch (Paszke et al. 2019). It is a
simple multilayer perceptron with six hidden layers between the input and
output layers. Each hidden layer comprises 300 neurons. The input layer has a
number of neurons that matches the number of spectral pixels in a given
profile (175 for DKIST/ViSP and 112 for Hinode SOT/SP). The output layer has 9
neurons, which correspond to the output parameters that we wish to retrieve.
These parameters are: 5 temperatures at different heights, 3 components of the
magnetic field vector and a single-valued line-of-sight velocity. In all
cases, the activation function chosen is a leaky ReLU (Maas et al. 2013). The
entire training procedure takes a few hours on our hardware described above.
For the training and validation sets we compute one million synthetic profiles
from randomized model atmospheres. A thousand models and profiles are taken as
the validation set and the rest are used for training. These models are
obtained as random variations from four different reference atmospheres,
namely: HSRA (Gingerich et al. 1971), VAL-C (Vernazza et al. 1981), FAL-C
(Fontenla et al. 1993) and the sunspot model M of Maltby et al. (1986). We
provide here a description of the randomization procedure in some detail
because the construction of this database is critical for the ability of the
ANN to perform adequately when faced with real observations and to exhibit
good generalization properties.
For each relevant parameter, we take the stratification in the reference
atmosphere and add a depth-dependent perturbation to it. The perturbation is
constructed by assigning values to certain layers and then interpolating in
depth. In the case of temperature, the parameter to which spectral lines are
most sensitive, we start by creating a perturbation at four layers. These
layers are not necessarily the same heights that the ANN will retrieve. They
are equispaced in the logarithm of continuum optical depth at 500 nm
($\log\tau_{500}$) and their actual location is different for the four
reference atmospheres. The perturbations at these four points are drawn from a
Gaussian distribution with a 1,500 K standard deviation. From these four
values, the depth-dependent perturbation is interpolated to the entire grid
and added to the reference model. With the new thermal stratification, the
model is set in hydrostatic equilibrium and the equation of state is solved to
compute plasma densities, ionization fractions, relevant molecules and
electron densities.
For the magnetic field, we have that $B_{z}$ (the line-of-sight component) is
linear in $\log\tau_{500}$, whereas $B_{x}$ and $B_{y}$ (the transverse
components) are constant with height. $B_{z}$ is defined by $B_{z}(0)$, its
value at $\log\tau_{500}$ $=0$, and its gradient. We construct three possible
scenarios with weak, strong and extreme fields, having probabilities of 45%,
45% and 10% respectively. The field strength $B_{z}(0)$ takes values from a
uniform distribution with a width of 500, 2000 and 6000 $G$ for the weak,
strong and extreme fields, respectively. The sign for each field component is
randomly set to $\pm 1$ except for $B_{x}$, which is always taken as positive.
Since the Zeeman effect has a 180-degree ambiguity in the transverse component
of the field, we restrict our solutions to the subspace with positive $B_{x}$.
The $B_{z}$ gradient is set to either 0 or a random value, with a 50%
probability. The random value is taken from a uniform probability distribution
between -150 and 150 $G$ per unit in $\log\tau_{500}$.
The filling factor ($\alpha$) is set to 1 in 50% of the models. The rest have
a uniform distribution between 0.1 and 1. In addition to the filling factor,
we consider a fixed amount of stray light in the instrument by adding an
average quiet Sun profile to Stokes I. The amount of stray light is fixed to
10%, which is a typical value for spectrographs.
This training set was built with the purpose of covering a sufficiently wide
range of profiles for the ANN to work with all possible observations of the
solar photosphere in the 630 nm spectral window observed by Hinode. The
statistical distribution of our random atmospheres is not necessarily optimal.
We relied on past experience and numerical experimentation to determine a
suitable set. A systematic analysis is beyond the scope of this work.
We used NICOLE (Socas-Navarro et al. 2015) to compute synthetic Stokes
profiles for the entire set of one million random models in the database. The
synthesis parameters for one of the training sets were defined to mimic
Hinode/SP observations. DKIST/ViSP will also feature a preset mode to observe
the same 630 nm window so we produced another similar set of profiles
simulating those observations in anticipation of its science operations. Both
training sets are publicly available in the repository mentioned above.
Figure 1: Comparison of inversions of a Hinode map performed with NICOLE
(left) and our ANN (right).
The Stokes profiles are fed as inputs to the ANN. For the outputs, we extract
a set of 9 parameters from the random model atmospheres in the database. These
parameters are: five temperatures ($T_{0}$, …, $T_{4}$), extracted at optical
depths $\log\tau_{500}$= $0$,…,$-4$, a bulk Doppler velocity ($v_{z}$) and the
three components of the pixel-averaged magnetic field ($F_{x}$, $F_{y}$,
$F_{z}$, where $F_{i}=\alpha B_{i}$ for $i=x,y,z$). We do not aim here at
disentangling the filling factor $\alpha$ from the intrinsic magnetic field
strength ($B_{i}$) in the magnetic element. We seek to retrieve the magnetic
flux density ($F$) in the resolution element, which simplifies the problem.
Figure 2: Comparison of inversions of a Hinode map performed with NICOLE
(left) and our ANN (right). Positive (negative) velocities are directed
downards (upwards). Positive (negative) magnetic polarity represents fields
pointing up (down) from the solar surface.
## 3 Comparisons with other inversions
After successfully training the ANN and observing a good recovery of the
validation set (see Figure 3), we tested it with real observations. As noted
in previous work (Socas-Navarro 2005), a good performance with the validation
set composed of synthetic observations does not guarantee a good operation
with real data.
Figure 3: Tests of the ANN performance with the validation set. Abscissas are
the ”true” values and ordinates are the ANN outputs. Each plot shows the
average (median) error in the recovery of that parameter.
Ideally, one would like to have inversions of Hinode/SP data to compare with
our ANN. Unfortunately, there are very few inversions of Hinode/SP maps that
yield the height stratification. The standard pipeline includes an inversion
carried out by the instrument team with the MERLIN code (Lites et al. 2007),
which is based on the Milne-Eddington approximation and therefore does not
provide information on the height dependence of any physical quantities. One
of the few inversions with the height stratification existing in the
literature is that of Socas-Navarro (2011) using NICOLE (later refined in
Socas-Navarro 2015).
We took the same Hinode/SP observations used for the NICOLE inversions and
processed them with our ANN. The NICOLE inversions took about 5 hours on a
dedicated parallel run over the eight cores of our workstation. The ANN
inversion was completed in half a second.
Figure 4: Comparison of inversions of a Hinode map (map1) performed with
MERLIN (left column) and our ANN (center). Scatter plots are shown in the
right column. Positive (negative) magnetic polarity represents fields pointing
up (down) from the solar surface. Figure 5: Comparison of inversions of a
Hinode map (map2) performed with MERLIN (left column) and our ANN (center).
Scatter plots are shown in the right column. Positive (negative) magnetic
polarity represents fields pointing up (down) from the solar surface.
An additional postprocessing renormalization was applied on each ANN
temperature output, so that the average value would match that of the NICOLE
inversions. This is done to remove (at least to first order) some small
residuals that arise in the application to real data.
We do not find these residuals in the validation tests so they must be due to
systematic differences between our synthetic training set and the real
observations, such as observation artifacts, differences in the PSF or an
inaccurate estimate of the stray light used in the synthesis. A detailed
analysis of these residuals is beyond the scope of this paper but for our
purposes here, this simple renormalization (the same for all observations)
resolves the issue.
The normalization factors for $T_{0}$,…,$T_{4}$ are 0.90,1.26,1.29,1.44 and
1.60, respectively. The growing trend of these factors indicates that the ANN
produces models that are, on average, steeper than those obtained with NICOLE.
The synthesis tests presented below demonstrate that the model atmospheres
obtained in this manner produce spectral profiles very similar to the
observations. For the magnetic field (see below), this calibration yields a
factor of 0.7 in all three components. We incorporate this normalization
factor in all subsequent inversions.
A comparison of the maps produced by the ANN and those from NICOLE (the 2015
version) is presented in Figs 1 and 2. The similarity between the spatial
structures in the images obtained with both techniques is remarkable. The
NICOLE inversions are much more noisy, especially in the higher layers. ANNs
are known to have good noise filtering properties. In this case, most of the
noise in the NICOLE data is ”inversion noise” produced by the specific
$\chi^{2}$ fitting procedure that seeks the best fit to the entire line
profile. The upper layers are probed only by the core of the spectral lines.
Since the core occupies very few pixels in the spectral profile, there is very
little information about those upper layers. For very similar profiles, the
$\chi^{2}$ minimization might reach slightly different solution where the core
is fitted with more or less accuracy, perhaps compensating it with a better
fit to other spectral regions. The end result is a pixel-to-pixel variation
that becomes more important in those layers where the profile is less
sensitive. This problem could be mitigated by fine-tuning the weights, giving
more weight to the pixels that carry the relevant information. However,
different layers would require a different optimization.
The ANN, on the other hand, ”learns” what are the optimal spectral points that
it needs to focus on for each layer. There is a direct, deterministic mapping
between the observations and the inversion result. For that reason the ANN
maps (right column in the figures) look cleaner. We can even see some of the
residual defects in the data reduction that are still present in the
observations, as they propagate directly into the results.
The similarity between both sets of images confirms that NICOLE and the ANN
are giving consistent results. This test should not be viewed as NICOLE giving
the ”correct” answer and our ANN being an approximation. Both techniques are
approximations and the difference between them is the sum of their respective
errors.
The accuracy of the ANN in recovering the magnetic field is not relevant for
the purposes of this paper. Nevertheless, we show here similar comparisons for
the sake of completeness. We processed two active region maps observed with
Hinode/SP (more details in section 4.2 below). These maps are 384$\times$384
(map1) and 871$\times$512 (map2) spatial pixels. The ANN inversions took 4 and
11 seconds, respectively. Standard inversions with the MERLIN code are
available for these maps. Figures 4 and 5 show the comparison of the magnetic
flux inferred by MERLIN and our ANN. For consistency with the training set,
the 180-degree ambiguity is resolved by choosing the solution that has a
positive component along the $x$ axis.
### 3.1 Reconstruction fits
A common problem with ANN-based inversions is that they are not based on
fitting the observations, unlike $\chi^{2}$ fitting methods. The quality of a
fit is usually a good indicator to assess the validity of the results. Our
approach suffers from this limitation too but it does provide enough
information to reconstruct a model atmosphere and, from there, synthesize
spectral profiles that can then be compared to the observations. It does not
provide the same information as a fit because the reconstruction of the
atmosphere implies additional approximations. Nevertheless, it is still useful
information.
We take the parameters from the ANN inversion and compute model atmospheres by
interpolating them in optical depth. For the temperature stratification, we
perform a cubic interpolation of the five temperatures between
$\log\tau_{500}$ $=0$ and $-4$. Above $\log\tau_{500}$$=-4$ we impose that the
stratification becomes flat. In the deeper layers below $\log\tau_{500}$ $=0$,
the temperature gradient usually becomes steeper. After some experimentation,
we concluded that a gradient that is 30% larger than between $\log\tau_{500}$
$=0$ and $-1$ works best in reproducing the observations. Hydrostatic
equilibrium is imposed and the plasma equation of state is solved numerically
to determine gas and electron densities, ionization stages and relevant
molecules. The magnetic field and bulk Doppler velocities are taken as
constant with height from the ANN inversion.
We computed the synthetic profiles from the models reconstructed from the ANN
outputs, obtaining the results shown in Fig 6. These ”reconstruction fits”
support the notion that the temperature stratification retrieved by the ANN is
consistent with the observations. The first panel shows the average profiles
over the entire region. The other three are selected representative samples of
profiles having a value of $\chi^{2}$ equal to the median over the region, and
the median plus/minus a standard deviation of all $\chi^{2}$ values.
Figure 6: Upper left: Average observed and synthetic profiles in the region.
Upper right: Sample profiles where the $\chi^{2}$ is equal to the median in
the region. Lower left: Profiles representative of a good reconstructed fit,
where the $\chi^{2}$ is equal to the median minus one standard deviation.
Lower right: Profiles representative of a poor reconstructed fit, where the
$\chi^{2}$ is equal to the median plus one standard deviation. In all cases,
the profile in blue is the observation and orange is the synthesis from the
reconstructed model atmosphere (reconstructed fit).
## 4 Results
We employed the ANN-based inversions described in the previous sections to
explore the thermal stratification of the solar photosphere. The maps
discussed above are in agreement with previous works in showing a rich thermal
structure, rapidly changing with height. In this paper, we compare the spatial
distribution to what is observed in the Ca II or G-Band filtergrams.
### 4.1 Quiet Sun
We start by considering the quiet Sun map inverted in the tests of section 3.
The maps are 200$\times$200 pixels but the field of view is not exactly square
because the slit stepping, which establishes the sampling in the
$x$-direction, does not necessarily match the pixel size. In this case the
sampling reported in the file headers is 0.15$\times$0.16 arc-seconds per
pixel.
The first recognizable pattern that stands out is the similarity of the mid-
photosphere temperature map to the Doppler velocity distribution. This is the
well known reversed granulation effect, a natural consequence of convective
motions. The tightest correlation in our dataset, shown in Figure 7, is
between temperature at $\log\tau_{500}$ $=-2$ and the Doppler velocity
($v_{z}$), with a Pearson’s correlation coefficient of 0.45 (recall that, as
noted above, the velocities retrieved by our procedure are at the base of the
photosphere). Reversed granulation is characterized also by an anticorrelation
with the $\log\tau_{500}$ $=0$ map, which in our data is of $-0.37$.
Figure 7: Scatter plot of the mid-photospheric temperatures and the Doppler
velocities retrieved by the ANN.
The scatter plot exhibits some vertical features. These are the result of the
ANN assigning nearly the same value of $v_{z}$ to many different profiles
instead of smearing them over the uncertainty range of that parameter. In
least-squares inversions, the solutions for similar profiles tend to be spread
over the error bar for that parameter because each inversion has followed a
different path on the $\chi^{2}$ hypersurface. However, an ANN might end up
assigning a specific value for a parameter (or a narrow range of values) as a
”sticky solution” for a range of input profiles. This means that the resulting
maps will usually be less noisy but the noise level should not be considered
an indication of the uncertainties.
Figure 8: Comparison of temperatures at different heights, as retrieved by the
ANN inversion, to Ca II filtergram (bottom right panel) in a quiet Sun region.
The temperature maps retrieved in the ANN inversions show a different network
structure at each atmospheric height. Bright photospheric networks have been
observed in the wings of the Ca II lines and in the G-Band, which are also
accessible to Hinode’s narrow band instrument (Hinode/NB). It is then of
interest to investigate whether these structures are related to those, both in
the quiet Sun and active regions.
Figure 8 shows a comparison of the temperature maps to a simultaneous Ca II
image from Hinode/NB. Alignment of SP and NB observations is not
straightforward. We made use of the pointing information stored in the data
headers to bring both datasets to a common reference frame. It should be
noted, however, that the alignment is only accurate to a few arc-seconds
222https://hesperia.gsfc.nasa.gov/ssw/hinode/sot/doc/guide/SAGv3.3.pdf.
The figure shows no obvious similarities between the Ca II intensity and the
photospheric temperatures retrieved at any of the heights. The very weak
magnetic fields in this quiet region do not appear to be correlated to the Ca
II filtergram, either.
### 4.2 Active regions
In this section we analyze two large-field active-region maps for which there
exists simultaneous G-Band and Ca II imaging. The datasets were acquired on
January 11, 2010 around 18:30 UT (map1) and January 22, 2012 around 06:30 UT
(map2). The spatial sampling is coarser than in the quiet Sun observations
(0.30 and 0.32 arc-seconds in the $x$ and $y$ directions, respectively) to
encompass a larger field of view. Map1 consists of 384$\times$384 spatial
pixels, while map2 is 871$\times$512\. The full maps are shown in Figs 9 to
10, with the various panels displaying temperatures at various heights, along
with the narrow-band images (the magnetic field was already introduced in
Figs. 4 and 5).
Figure 9: ANN inversion of map1 (active region) at different heights (except
$\log\tau_{500}$= 0) alongside G-band and Ca II filtergrams. The arrow points
in the direction to the nearest solar limb. Figure 10: ANN inversion of map2
(active region) at different heights (except $\log\tau_{500}$= 0) alongside
G-band and Ca II filtergrams. The arrow points in the direction to the nearest
solar limb.
The data shows a fine network of hot pixels that roughly follows, in the mid-
photosphere, the magnetic field distribution ($\log\tau_{500}$ $=-1$ and
$-2$). Higher up the structure is more patchy and does not follow the magnetic
field maps. Each layer exhibits a different structure and, more importantly,
they also differ from both the G-Band and Ca II H images. We discuss these
differences below.
A reversal of hot and cool areas between the middle and the upper photosphere
is also apparent. For instance, the lower left corner of both maps (Figs. 9
and 10) is cool at $\log\tau_{500}$ $=-1$ but hot at $\log\tau_{500}$ $=-4$
(upper left and upper right panels in both figures). The same anticorrelation
is apparent in the upper left and lower right corners of map2 (Fig. 10). In
fact, most of the region left of $x=200$ arc-sec in Fig. 10 has a reversed
appearance. The hot network at $\log\tau_{500}$ $=-1$ (upper left panel) is
seen as a dark shadow at $\log\tau_{500}$ $=-4$ (upper right panel). The same
is true about the area left of $x=-240$ in map1 (Fig. 9). We quantified this
by selecting only the pixels that are hot in either layer ($T>$5400 at
$\log\tau_{500}$ $=-1$ or $T>$5000 at $\log\tau_{500}$ $=-4$) and computing
the Pearson’s correlation coefficient. The values obtained are $-0.77$ in map1
and $-0.78$ in map2.
The temperature maps do not match the narrow-band images. There is some
similarity between the temperature at $\log\tau_{500}$ $=-1$ and the G-Band
image in the overall distribution of the hot network. However, a closer look
shows important differences (see discussion of Fig. 11 below).
The comparison with the Ca II H images is even more puzzling. The Ca emission
follows the pattern of the hot pixels in the mid-photosphere at
$\log\tau_{500}$ $=-1$ instead of the upper photosphere, as one would have
expected (recall that, as discussed above, the distribution of hot pixels at
$\log\tau_{500}$ $=-4$ is anticorrelated with that at $\log\tau_{500}$ $=-1$).
However, the Ca images show small-scale filamentary structure in the network,
as opposed to the chains of dots that appear in the temperature maps. The
appearance of filaments would suggest that we are seeing higher layers but the
brightness distribution follows that of the mid-photosphere. We speculate that
the most plausible explanation is that the this spectral band has
contributions to its response function from both the low photosphere and the
(low) chromosphere. Detailed radiative transfer modeling would be necessary to
confirm this point but it would require some knowledge of the chromospheric
conditions, which is not available from these data.
Figure 11 shows a zoom on two regions containing several pores in both maps.
The magnification is different in each figure because the area with pores is
larger in map2. Even though the G-Band bright points extend over the same area
as the hot network at $\log\tau_{500}$ $=-1$, they do not exhibit the same
features when seen at high resolution.
Figure 11: Enlargement of of the areas with pores in the active regions map1
(upper panels) and map2 (lower panels). The left panels show ANN temperature
reconstructions at $\log\tau_{500}$=-1 and the right panels, the respective
G-band filtergrams. Notice the bright arcs around the limb-side (indicated by
the orange arrow) of the pores in the left panels.
A very remarkable feature in these images is the presence of a bright arc
around the edge of pores, tracing the limb side (the arrows indicate the
direction to the closest limb). These arcs are visible around virtually every
pore in both datasets. They have a width of one or two pixels, suggesting that
they are not fully resolved in the observations, and their temperature is
always between 5,600 and 5,700 K. By contrast, the pores have temperatures
mostly of 4,600 to 4,800 K but in some cases, particularly the larger ones in
map2, they may go down to 3,600 K, such as in the feature at coordinates
(165,390) of map2.
The bright arcs are probably the pores’ ”hot walls”. The idea of a hot wall
seen in perspective was introduced in early fluxtube models to explain the
center-to-limb variation of faculae and G-band points (Spruit 1976; Knoelker &
Schuessler 1988; Topka et al. 1997). Spruit’s original work considered
unresolved fluxtubes and small pores of up to 1,000 km. The pores in our
observations are significantly larger, starting from roughly 3,500 km, but
there is no reason why the same effect should not take place in these.
## 5 Conclusions
ANN-based inversions are enabling the analysis of large spectroscopic (and
spectropolarimetric) datasets. One such application is presented in this
paper. The training strategy appears to be sufficiently robust for application
to real observations in various situations.
It is puzzling to find such a clear anticorrelation in the location of hot
points in the middle and upper photosphere. This is counterintuitive and
warrants further work to confirm it, since it appears to challenge the
generally accepted idea that small magnetic elements act as channels to
propagate energy into the upper atmosphere (e.g., Jefferies et al. 2006;
Rajaguru et al. 2019). One possibility is that the energy dissipation and
associated heating might occur at higher layers than we observe here. That
would explain the presence of hot points at intermediate heights that do not
exhibit a temperature enhancement in the upper photosphere. However, this
would not explain the patches with hot points in the upper layers that appear
as quiet lower down.
In our ANN approach, each pixel is inverted independently of the rest.
Therefore, the spatial distributions obtained cannot be artifacts of the
procedure, they must be present in the data somehow. A possible mundane
explanation for the pixels that are hot in middle layers and quiet at the top
could be that the ANN is not properly trained for such situations and the
closest models in the training set that reproduce the lower and mi layers are
quiet in the upper layers. However, that would not explain the opposite
scenario in the anticorrelation, i.e. the patches with quiet lower and middle
photosphere having enhanced temperature in the upper layers.
Another important result presented in this paper is the first observation of
the ”hot wall” effect, which has been a model prediction since the 1970
(Spruit 1976) and explains the bright appearance of small magnetic elements.
The original theoretical models considered smaller pores of up to 1,000 km but
we have detected it here in structures of at least 3,500 km. Hot walls are
believed to be responsible for the brightness of faculae and small magnetic
flux elements (e.g., Topka et al. 1997). This view, which is now the community
consensus, is strongly reinforced by our data.
Our results open the possibility of a future application to chromospheric
lines. The chromosphere is much more complicated to simulate due to NLTE
radiative transfer, much faster and vigorous dynamics, faster wave phase
velocities, etc. As a result, MHD numerical models are not yet sufficiently
realistic that they can be used to match the observations and this limits the
ability to train an ANN with chromospheric simulated profiles. However, our
training strategy does not require a full numerical model. In principle one
could apply a similar training using semiempirical 1D models with random
perturbations to produce a large number of NLTE profiles. The computational
effort involved in the database generation would be significantly higher than
here but still feasible.
An area of improvement that we have found with this technique is the ”sticky
solution”, where the ANN returns basically the same values within a broader
uncertainty range, creating the vertical features seen in the scatter plot of
Fig 7.
Finally, we would like to mention a negative result. In spite of our best
efforts with this approach, we have not been able to retrieve gradients in the
magnetic field or the Doppler velocity. We have encountered the same inability
to retrieve gradients in previous works with other ANNs. It is not clear to us
whether this inability is due to a specific problem with our methodology or an
intrinsic limitation of the procedure. An interesting line for future work
would be to explore these and other limitations.
###### Acknowledgements.
We acknowledge financial support from the Spanish Ministerio de Ciencia,
Innovación y Universidades through project PGC2018-102108-B-I00 and FEDER
funds. This research has made use of NASA’s Astrophysics Data System
Bibliographic Services. The Python Matplotlib (Hunter 2007), Numpy (Van Der
Walt et al. 2011), PyTorch (Paszke et al. 2019) and IPython (Pérez & Granger
2007) modules have been employed to generate the figures and calculations in
this paper.
## References
* Abadi et al. (2015) Abadi, M., Agarwal, A., Barham, P., et al. 2015, TensorFlow: Large-Scale Machine Learning on Heterogeneous Systems, software available from tensorflow.org
* Asensio Ramos & Díaz Baso (2019) Asensio Ramos, A. & Díaz Baso, C. J. 2019, A&A, 626, A102
* Bellot Rubio (2006) Bellot Rubio, L. R. 2006, in Astronomical Society of the Pacific Conference Series, Vol. 358, Solar Polarization 4, ed. R. Casini & B. W. Lites, 107
* Carroll & Staude (2001) Carroll, T. A. & Staude, J. 2001, A&A, 378, 316
* del Toro Iniesta & Ruiz Cobo (2016) del Toro Iniesta, J. C. & Ruiz Cobo, B. 2016, Living Reviews in Solar Physics, 13, 4
* Díaz Baso & Asensio Ramos (2018) Díaz Baso, C. J. & Asensio Ramos, A. 2018, A&A, 614, A5
* Felipe & Asensio Ramos (2019) Felipe, T. & Asensio Ramos, A. 2019, A&A, 632, A82
* Fontenla et al. (1993) Fontenla, J. M., Avrett, E. H., & Loeser, R. 1993, ApJ, 406, 319
* Gingerich et al. (1971) Gingerich, O., Noyes, R. W., Kalkofen, W., & Cuny, Y. 1971, Sol. Phys., 18, 347
* Guo et al. (2020) Guo, J., Bai, X., Deng, Y., et al. 2020, Sol. Phys., 295, 5
* Hunter (2007) Hunter, J. D. 2007, Computing In Science & Engineering, 9, 90
* Jefferies et al. (2006) Jefferies, S. M., McIntosh, S. W., Armstrong, J. D., et al. 2006, ApJ, 648, L151
* Knoelker & Schuessler (1988) Knoelker, M. & Schuessler, M. 1988, A&A, 202, 275
* Lites et al. (2007) Lites, B., Casini, R., Garcia, J., & Socas-Navarro, H. 2007, Mem. Soc. Astron. Italiana, 78, 148
* Liu et al. (2020) Liu, H., Xu, Y., Wang, J., et al. 2020, ApJ, 894, 70
* Maas et al. (2013) Maas, A. L., Hannun, A. Y., & Ng, A. Y. 2013, in in ICML Workshop on Deep Learning for Audio, Speech and Language Processing
* Maltby et al. (1986) Maltby, P., Avrett, E. H., Carlsson, M., et al. 1986, ApJ, 306, 284
* Milić & Gafeira (2020) Milić, I. & Gafeira, R. 2020, A&A, 644, A129
* Paszke et al. (2019) Paszke, A., Gross, S., Massa, F., et al. 2019, in Advances in Neural Information Processing Systems 32, ed. H. Wallach, H. Larochelle, A. Beygelzimer, F. d'Alché-Buc, E. Fox, & R. Garnett (Curran Associates, Inc.), 8024–8035
* Pérez & Granger (2007) Pérez, F. & Granger, B. E. 2007, Computing in Science and Engineering, 9, 21
* Rajaguru et al. (2019) Rajaguru, S. P., Sangeetha, C. R., & Tripathi, D. 2019, ApJ, 871, 155
* Rimmele et al. (2020) Rimmele, T. R., Warner, M., Keil, S. L., et al. 2020, Sol. Phys., 295, 172
* Socas-Navarro (2001) Socas-Navarro, H. 2001, in Astronomical Society of the Pacific Conference Series, Vol. 236, Advanced Solar Polarimetry – Theory, Observation, and Instrumentation, ed. M. Sigwarth, 487
* Socas-Navarro (2002) Socas-Navarro, H. 2002, in SOLMAG 2002. Proceedings of the Magnetic Coupling of the Solar Atmosphere Euroconference and IAU Colloquium 188, 11 - 15 June 2002, Santorini, Greece. Ed. H. Sawaya-Lacoste. ESA SP-505. Noordwijk, Netherlands: ESA Publications Division, ISBN 92-9092-815-8, 2002, p. 45 - 51, 45
* Socas-Navarro (2003) Socas-Navarro, H. 2003, Neural Networks, 16, 355
* Socas-Navarro (2005) Socas-Navarro, H. 2005, ApJ, 621, 545
* Socas-Navarro (2011) Socas-Navarro, H. 2011, A&A, 529, A37
* Socas-Navarro (2015) Socas-Navarro, H. 2015, A&A, 577, A25
* Socas-Navarro et al. (2015) Socas-Navarro, H., de la Cruz Rodríguez, J., Asensio Ramos, A., Trujillo Bueno, J., & Ruiz Cobo, B. 2015, A&A, 577, A7
* Spruit (1976) Spruit, H. C. 1976, Sol. Phys., 50, 269
* Topka et al. (1997) Topka, K. P., Tarbell, T. D., & Title, A. M. 1997, ApJ, 484, 479
* Van Der Walt et al. (2011) Van Der Walt, S., Colbert, S. C., & Varoquaux, G. 2011, Computing in Science & Engineering, 13, 22
* Vernazza et al. (1981) Vernazza, J. E., Avrett, E. H., & Loeser, R. 1981, ApJS, 45, 635
|
# ’Controlling by Showing: i-Mimic’: A Video-based Method to Control Robotic
Arms
Debarati B. Chakraborty<EMAIL_ADDRESS>Mukesh Sharma
<EMAIL_ADDRESS>Bhaskar Vijay<EMAIL_ADDRESS>Dept. of Computer
Science and Engineering, Indian Institute of Technology, Jodhpur, India
Department of Mechanical Engineering, Indian Institute of Technology, Jodhpur,
India
###### Abstract
A novel concept of vision-based intelligent control of robotic arms is
developed here in this work. This work enables the controlling of robotic arms
motion only with visual inputs, that is, controlling by showing the videos of
correct movements. This work can broadly be sub-divided into two segments. The
first part of this work is to develop an unsupervised vision-based method to
control robotic arm in 2-D plane, and the second one is with deep CNN in the
same task in 3-D plane. The first method is unsupervised, where our aim is to
perform mimicking of human arm motion in real-time by a manipulator.
Mimicking, here involves a series of steps, namely, tracking the motion of the
arm in videos, estimating motion parameters, and replicating the motion
parameters in the robot. We developed a network, namely the vision-to-motion
optical network (DON), where the input should be a video stream containing
hand movements of human, the the output would be out the velocity and torque
information of the hand movements shown in the videos. The output information
of the DON is then fed to the robotic arm by enabling it to generate motion
according to the real hand videos. The method has been tested with both live-
stream video feed as well as on recorded video obtained from a monocular
camera. Besides, the DON also enables a method to behave intelligently by
predicting the trajectory of human hand even if the hand gets occluded for
some time with varying degree of occlusion. This is why the mimicry of the arm
incorporates some intelligence to it and becomes intelligent mimic (i- mimic).
Alongside the unsupervised method another method has also been developed
deploying the deep neural network technique with CNN (Convolutional Neural
Network) to perform the mimicking, where labelled datasets are used for
training. The same dataset, as used in the unsupervised DON-based method, is
used in the deep CNN method, after manual annotations. Both the proposed
methods are validated with off-line as well as with on-line video datasets in
real-time, enabling the robotic arm in ’i- mimic’ and thereby showing the
effectiveness of the proposed method. The entire methodology is validated with
real-time 1-link and simulated n-link manipulators alongwith suitable
comparisons.
###### keywords:
Visio-based robot control, video processing , deep network , convolutional
neural network , robotic arm control
## 1 Introduction
Robotic arms are used to perform mechanical tasks in industries over decades.
It was mainly used for performing repetitive tasks in the industries to cut
down the labor cost [1, 2]. Normally, robotic arms are quite complex with five
or more degree of freedom as it aims to perform human tasks. Recently, the
application of robotic arms in conducting the domestic work has drawn
attention. Controlling of these robotic arms to perform different tasks is
still a major issue to be addressed. Here in this work we have defined a new
concept of controlling robotic arms only with visual information, that is, the
motion of different parts of the robotic arms could be controlled by providing
according human hand movement videos as the input to those arms.
The entire method could be subdivided into two parts. In the first part of the
this work we aim to find out a simple solution for unsupervised controlling
the robotic arm only with visual information. Here we aim to deal with the
issues of i) unavailability of sufficient training datasets, ii) domain
adaptation and iii) economic cost. On the way to search for a solution of the
two initial issues we concluded that the teaching/ training part should be
removed. But how could it be controlled then? ’Mimic’ is the solution that
stroked in our mind. The controlling could only be achieved by showing the arm
the desired movement and making it enabled to follow it. Visual mimic-based
controlling of 1-link and n-linked robotic arms is the primary contribution of
this work. The mechanism of this set-up is quite simple and the 1-link
manipulator is developed only by the authors. Either the recorded or real-time
video could be shown to the arm to achieve the control. It should be noted
that the real-time testing of this ’mimic’ with robotic arm is in a very
primary level where the arm is a simple 1-link robotic manipulator with a
degree of freedom of 120 degrees. The rest of the tests are conducted in the
form of simulation, where the simulated arm is able to mimic the motion of a
single joint (solder or elbow) or real hand.
Another contribution in this part of the work is development of the vision-to-
motion optical network (DON) to process the optical flow information of the
input video and to convert it into the physical force, to be fed the robotic
arm. The proposed DON is different from the existing deep networks in the
following manners: i) it does not require any labeled data set or manual
intervention, ii) it does neither require background estimation or a large
number of input frames for training, iii) the functioning of the intermediate
layers are simple which enables computational gain iv) all the intermediate
layers are not active simultaneously; some layer gets activated depending on
the values of the outputs of its previous layer v) the single network can
perform both estimation and prediction and vi) it produces torque and angular
velocity as the output.
In the second part of the work we focus on implementation rather than
definition. We implemented deep neural network with CNN and Refiner Network
for the purpose of i-mimic. To deal with unavailability of large set of
labelled data, we used a CNN network whose output are further fed into the
refinement network that smoothens the final output and enables to interpolate
to a larger range. The deep CNN is proven to be less effective with 2-D n-link
manipulators, but, it performs better in 3-D plane with n-link manipulators.
The rest of the article is organised as follows. Sec. 2 presents the
background research, Sec. 3 describes the layer-wise formulation of vision-to-
motion optical network (DON). The architecture of CNN and Refinement Network,
customized loss function, dataset, training are described in Sec. 4. The
experimental set-ups with four different variations in experimental studies
are described in Sec. 5. The real-time experimental results tested under
different scenarios like without occlusion, with low occlusion, with high
occlusion, and with multiple joints are given in Sec. 6 along with parameter
section and comparative study. The overall conclusion of this work with its
future scope are discussed in Sec. 7.
## 2 Background Research
Vision-based robotics to serve domestic purpose has drawn the attention of
researchers in several areas. Most of the approaches developed so far for this
purpose implied training the system through labeled data, i.e., with
supervised or semi-supervised learning. Automated driving [3], grasping [4]
and block stacking [5] are among few the applications where this kind of
learning were used. But gathering adequate amount of labeled data for training
is a challenging issue for this type of approaches.
Semi-supervised learning or reinforcement learning has appeared to be the
substitute of supervised learning in recent literature’s where the training is
carried out with less amount of labeled data or weakly labeled data. Rusu _et
al._ [6] used learning with progressive network for Jaco robot gripping to
have a faster algorithm with less amount of training data. KUKA IIWA robot
grasping with deep network and domain adaptation was developed by Bousmalis
_et al._ [7]. A method of training with weakly labeled images with adaptation
from real world to simulation using a PR2 robot was proposed by Tzeng _et al._
[8]. Zuo _et al._ [9] came up with a solution of semi supervised method of 3D
pose estimation where the training was carried out in a virtual environment.
Its real world implementation was done after domain adaptation. Domain
adaptation is another challenge while semi-supervised/ reinforcement learning
is carried out. There are many rich works carried out so far to deal with this
problem. Domain adaptation with back-propagation by inducing an ’inverse-
gradient layer’ to the deep network was formulated by Ganin and Lempitsky
[10]. In another work Ganin _et al._ [11] came up with a solution of carrying
out the training and testing of the network with the features that are non-
discriminative and domain invariant for training and test data. Bousmalis _et
al._ [12] came up with another solution of identifying the unique feature of
each domain to extract out the common features in the domains. They have
recently developed another way of domain adaptation with simultaneous
simulation [7]. Sing _et al._ [13] demonstrated that passively collected data
can be paired with interaction data to learn visual representations for end-
to-end control policies that generalize substantially better to unseen
environments. However, less amount of labeled data or synthetically labeled
data are always required in all of the aforementioned approaches.
Economic cost of a robotic arm controller is another major issue to be dealt
with to make the robotic arms be implementable for domestic purpose. The
controlling of the robotic arms are normally carried out with multiple sensors
which make it more costly. The robotic arms like PR2 [8], Jaco [6] or KUKA
IIWA [7] costs around USD 20,000/- to USD 50,000/-. A pocket friendly robotic
arm is developed recently [9] but but it still have the issues of synthetic
labeled data and domain adaptation.
## 3 DON: Vision-to-Motion Optical Network
Here we developed a network based on the information of optical flow from
frame-to- frame of a the input hand movement video sequence. The major
challenges that are to be addressed in the task of ’i-mimic’ are:
unavailability of sufficient number of labelled data, computation time and the
lag between input-video to robotic-arm gestures. The proposed deep flow
network is able to minimize all these parameters simultaneously. First of all,
no labelled data is required here and the process is fully automatic. The
computation complexity of this method is quite low and there the lag is as
less as around ten frames here.
The vision-to-motion optical network is a layer-wise network with multiple
layers between the input and output layer. Different feature of the optical
flow information is process in different layers of this network. The layer-
wise architecture is shown in Fig. 1. It can be noticed from the diagram that
the videos are fed in the input layer of the network, whereas we get velocity
and torque values that generates the physical motion at the output layer. That
is why it is named as a ’vision-to-motion optical network network’. The output
of the previous layer is the input to the next layer. All the layers of the
network may not be active at a time. Rather, the activation of some layer of
the network is dependent on the outputs of its previous layer. The layer-wise
working principles of this network are described in details in the following
sections.
Figure 1: Architecture of Deep Flow Network
### 3.1 Layer 1: Input Layer
The video sequence is the input that is fed to the network. But it is not the
entire sequence that is given as the input at a time since a on-line
processing is going on here with the input video frames. As we have already
stated that the proposed method is unsupervised, therefore the output is to be
produced only by automated processing of input data. Here, the frame that gets
generated on the current instant, say, at the instant $t$ is fed to the
network along-with $N$-number of previous frames that gets generated earlier
to the current frame. Let the current frame be denoted as $f_{t}$here. The
frames generated in the earlier instances be denoted by
$f_{t-1},f_{t-2},...,f_{t-N}$. Therefore, the input layer contains the frames:
$f_{t},f_{t-1},...,f_{t-N}$.
### 3.2 Layer 2: Frame Difference Computation Layer
The network is supposed to deal with the optical flow information. In case of
the video sequences that we are dealing with is captured by static cameras.
Therefore the changed information from frame to frame reflects the optical
flow of the sequence. Here two types of differences are computed here in this
layer. That is why two different colored of nodes (DO1 and DO2) are shown
there in Fig. 1. There are total $N+N=2N$ number of nodes present there in
layer 2. The difference operation carried out in DO1 the difference between
consecutive frames ($\delta 1$) given in Eqn (1). The difference between the
current to all its previous frames ($\delta 2$) is carried out in Eqn (2).
$\delta 1_{p}=|f_{t-p}-f_{t-(p-1)}|:p=0,...,N-1$ (1) $\delta
2_{p}=|f_{t}-f_{t-p}|:p=1,...,N$ (2)
Therefore $N$ number of binarized $\delta 1$ and $\delta 2$ frames, i.e., $2N$
number of difference frames in total are the output of this layer. $\delta
1_{p}$ are the binarized outputs from the nodes of type DO1, whereas, $\delta
2_{p}$ are the output from DO2 type of nodes.
### 3.3 Layer 3: Object Identification Layer
The third layer of this network is developed to find out the locations and the
shape of the moving hand in all the $N$-number of previous frames. As it can
be observed from Fig. 1 that this layer contains $N$ number of nodes, labeled
as $ObjectFinderF1,...,ObjectFinderFN$. The input fed to a certain node
$ObjectFinderFp$ are: ${\delta 2_{p}}:p=1,...,N$ and $\delta 2_{p}$. That is,
all the DO2- difference frames and only $p^{th}$ DO1 difference frame are the
input to the said node of layer 3. Let the location of the moving object
segment in the $p^{th}$-frame be represented by $l_{p}$. The operation that is
carried out in each node of the third layer is given by the Eqn. (3).
$l_{p}=(\cup\ _{p=1}^{N}\delta 2_{p})\cap\delta 1_{p}$ (3)
Please note that the union of $\delta 2_{p}\forall p=1,...,N$ is taken here to
have the entire moving obeject region as a subset of that union and
intersection of it to that of $\delta 1_{p}$ is carried out to extract out the
obvious moving region in the $p^{th}$-frame. The pixels those belong to the
set $l_{p}$ are in the region that definitely belong to the moving object in
the $p^{th}$ frame. For the sake of simplicity, here in this work we consider
only the skeleton and the locations of the corner pixels of the $l_{p}$
(moving hand) to be the output from each node of this layer as we need to find
out the angular velocity and torque from the hand movement video.
### 3.4 Layer 4: Estimation Layer
This layer contains two nodes and the operations and functioning of these two
nodes are different from each other. The location of the object in $N$-number
of frames are the input to this layer. Two types of estimations are performed
simultaneously in this layer with the two nodes. The path estimation node
gives the probable trajectory of the moving object as the output whereas the
trust factor estimator node computes the reliability of the estimated path.
The output of trust factor estimator node determines the activation of the
next layer, i.e., prediction layer. The working principles of the two nodes in
the forth layer are described below.
#### 3.4.1 Path Estimator
As discussed before the prediction of probable trajectory of the object is
carried out here. This is done by computing the optical velocity and
acceleration of the moving object from frame-to-frame displacement. Let
$\varsigma_{p}$ be the location $l_{p}$ (see Eqn. (3)) in the $p^{th}$-frame.
Then the velocity ($v_{p}$) and acceleration ($a_{p}$) of that object are
computed according to Eqn. (4). The velocity and acceleration values for all
the $N$ frames are stored in the sets $V$ and $A$ respectively.
$\displaystyle{V}$
$\displaystyle=\\{v_{p}:v_{p}=\varsigma_{p}-\varsigma_{p-1}\forall
p=1,...,N\\}$ (4) $\displaystyle A$
$\displaystyle=\\{a_{p}:a_{p}=v_{p}-v_{p-1}\forall p=2,...,N\\}$ (5)
Please note that signed difference between the locations and velocity are
taken while computing $v_{p}$ and $a_{p}$ in Eqn. (4). It is known from Sec.
3.3 that the input $\varsigma_{p}$ could be a scalar or vector component based
on the type of object representation. But the two components $v_{p}$ and
$a_{p}$ should always be a vector since these components contain both
magnitude and signs. Consideration of those signs helps in the incorporation
of the information of the direction and the change in the direction of the
moving hand.
Here the robotic arm with revolute joint is supposed to mimic the movement of
the arm shown in real time or recorded video. Therefore, the movement of the
arm is always supposed to be circular in nature with respect to any joint
(e.g. elbow) with maximum 180 degree of freedom. This phenomenon is kept in
mind and the determination of the radius w.r.t the angular motion is computed
by measuring the length of the skeleton of the arm. Let the skeleton of the
moving part of the arm be of length $r$. The angular velocity ($\omega_{p}$)
and torque ($\theta_{p}$) are then computed as:
$\omega_{p}=\frac{v_{p}}{r}$
and
$\theta_{p}=I\alpha_{p}$
where $I$ is mass moment of inertia of the manipulator arm and $\alpha_{p}$ is
the angular acceleration computed as:
$\alpha_{p}=\frac{a_{p}}{r}$
For any given one-link manipulator the algorithm computes $\alpha_{p}$ and
having $I$ of the manipulator one can compute the torque required to be
applied to at the joint.
#### 3.4.2 Trust Factor Computation
The working principle of this particular node is different from any other
nodes present there in a network. It takes input from the previous layer but
does not transmit its output to the next layer. Instead the output from this
layer determines which layer should be the fifth layer of this network. That
is the which path should be followed by the output information from the path
estimator node is decided with the output of this node. Since, only motion of
the moving hand of a static person is considered here, it can be assumed that
the size of the moving object will remain almost the same throughout the
sequence. This assumption is applied during formulation of the trust factor.
Let there be $M_{p}$ be the region of $l_{p}$ (see Eqn. (3)) in the
$p^{th}$-frame. Let, the set $\\{S\\}$ the regions of the object in all the
$N$ frames and the set $\\{S_{d}\\}$ contains the values of change in regions.
Those are computed according to Eqn. (3.4.2).
$\displaystyle{S}$ $\displaystyle=\\{M_{p}:p=1,...,N\\}$ $\displaystyle S_{d}$
$\displaystyle=\\{c_{p}:c_{p}=|M_{1}-M_{p}|\forall p=2,...,N\\}$ (6)
The trust factor ($\eta$) is computed as:
$\eta=1-\frac{max(S_{d})}{max(S)}$ (7)
In Eqn. (7) max(.) represents the element with maximum magnitude present in a
set. Physically, the effectiveness of measuring the $\eta$ is in determining
the amount of occlusion took place over the moving object. If huge amount of
occlusion is present there for some frames, then the estimation with those
frame may lead to a wrong trajectory. Therefore, prediction should be carried
out from the previous set of information and ignoring the wrong (occluded)
visual information. That is why the activation of the prediction layer is
necessary in this scenario. The path leading to prediction layer gets only
activated if the value of $\eta$ is low.
### 3.5 Layer 5: Prediction Layer
There is only one node in this layer. But, the input fed to this layer is not
only from the previous layer, but the output of layer 4 of the previous
execution of the network is also an input here. Please note that this layer
can not be active in the first execution of the network, but from the second
execution onward it could get activated any time. Here the velocity and
acceleration values from the frames without occlusion, or with minimal
occlusion are considered. There inputs that are provided to this node are: i)
the velocity and acceleration information (sets $V$ and $A$) from the previous
execution, ii) the velocity and acceleration information ($V$ and $A$ from
Eqn. (4)) from the previous layer and iii) Object regions and change in the
regions ($S$ and $S_{d}$ from Eqn. (3.4.2)). Let the velocity and acceleration
from the previous be denoted here as $\tilde{V}$ and $\tilde{A}$. We only
consider the information of the frames for with $c_{p}<0.05Xmax(S)$ ($C_{p}$
is as defined in Eqn. (3.4.2)). That is, frames maximum with $5\%$ change in
object size will be taken into account. Let $k$ number of frames out of the
$N$ frames failed to satisfy the criterion. Then only $N-k$ elements from the
sets $V$ and $A$ will be taken by merging it with the sets $\tilde{V}$ and
$\tilde{A}$ respectively. Therefore, the new sets will be
$V^{k}=\\{\tilde{V}|V(1:N-k)\\}$ and $A^{k}=\\{\tilde{A}|A(1:N-k)\\}$ with
$N+N-k=2N-k$ number of elements in each set. Now we need to predict the
information from the $(N-k+1)^{th}$ frame onward. As it is known, the
consecutive difference between the elements of $V^{k}$ forms $A^{k}$, i.e.,
$A^{k}$ could be said the first order derivative of $V^{k}$. We can similarly
compute the second order derivative of $V^{k}$ or the first order derivative
of $A^{k}$ and represent it by $A^{k^{\prime}}$. Now, the $(2N-k+1)^{th}$
element of the sets $V^{k}$ and $A^{k}$ will be approximated as:
$\displaystyle v_{2N-k+1}=v_{2N-k}+a_{2N-k-1}$ $\displaystyle
a_{2N-k+1}=a_{2N-k}+a^{\prime}_{2N-k-1}$ (8)
In Eqn. 3.5 the symbols $V_{p}$, $a_{p}$ and $a^{\prime}_{p}$ represents the
$p^{th}$ element of the sets $V^{k}$, $A^{k}$ and $A^{k^{\prime}}$
respectively. The element will get inserted to the sets $V^{k}$ and $A^{k}$ as
the $(2N-k+1)^{th}$ elements of them. The process will be repeated and the
next element will be approximated. The process will continue until the set is
going to have $2N$ number of elements. Once it is done, the last $N$ elements
of $V^{k}$ and $A^{k}$ will be stored in the sets $V$ and $A$ respectively and
will be given as the output to the output layer.
The experimentation those are carried out with this proposed DON are described
in the following section. Please note that, one additional layer, namely
object regression layer to this network is introduced while working with
multiple joints. It is shown in Fig. 9.
## 4 DNN:Deep Neural Network
In our second approach, we designed Convolutional Neural Network(CNN) for
which the input corresponds to the image(frames of video stream) and labels
correspond to coordinates$(x,y)$ of joints(shoulder, elbow and wrist joints).
We have used two networks for our purpose to get the joints coordinates from
which further joint angles, joint velocities and and joint torque can be
obtained in terms of pixel coordinates which are mapped to real value using
the mapping function or mapping factor. In [14] after using DNN based
regression a DNN based refiner network was added, which takes cropped image
around prediction as input to improve the prediction but we used a different
approach by mapping a simple neural network to refine the predictions of the
DNN based regression network. Our approach reduces the training time of the
network with similar accuracy for our dataset.
### 4.1 Convolutional Neural Network(CNN)
The architecture of CNN is shown in Fig. 2. The input to this network is the
images(from a stream of video feed) and labels as the pixel coordinates$(x,y)$
of the shoulder joint, elbow joint and wrist joint. The network has six hidden
layers wherein there are 2 Convolutional layers followed by max pooling and
flatten. Activation functions for all the layers other than the last layer are
ReLu activation. Final layer has a Linear Activation function. This network
uses Mean Square Error(MSE) Loss function.
Figure 2: CNN Network
### 4.2 Refiner Network
The architecture of Refiner Network is shown in Fig. 3. This network is a
simple neural network with input corresponding to the output of CNN Network
and labels as the pixel coordinates of the shoulder joint, elbow joint and
wrist joint respectively. Activation function for the last layer is Linear and
for rest is ReLu Activation. This network uses customized loss function that
includes MSE and simple error in link length.
Figure 3: Refiner Network
### 4.3 Customized Loss Function
For the error in coordinates of joint Mean Square Error(MSE) has been used and
another error for link length has also been taken into consideration. Let $d$
denote error in link length and $p$ denote the MSE in position. For real joint
coordinates $(x_{s},y_{s})$, $(x_{e},y_{e})$, $(x_{w},y_{w})$, and predicted
coordinates $(x_{sp},y_{sp})$, $(x_{ep},y_{ep})$, $(x_{wp},y_{wp})$ where
subscripts $s$, $e$ and $p$ represents shoulder, elbow and wrist joints
respectively and subscript $s$, $e$, $w$ followed by $p$ represents
corresponding predicted coordinates respectively. The respective error are
computed following the Eqns. (9)-(11).
$d=(\sqrt{(x_{s}-x_{e})^{2}-(y_{s}-y_{e})^{2}}-\sqrt{(x_{sp}-x_{ep})^{2}-(y_{sp}-y_{ep})^{2}})^{2}+\\\
(\sqrt{(x_{e}-x_{w})^{2}-(y_{e}-y_{w})^{2}}-\sqrt{(x_{ep}-x_{wp})^{2}-(y_{ep}-y_{wp})^{2}})^{2}$
(9)
$p=(x_{s}-x_{sp})^{2}+(y_{s}-y_{sp})^{2}+(x_{e}-x_{ep})^{2}+(y_{e}-y_{ep})^{2}+(x_{w}-x_{wp})^{2}+(y_{w}-y_{wp})^{2}$
(10) $loss=d+p$ (11)
### 4.4 Dataset
For the purpose of training the deep CNN, we created our own dataset of 300
images and manually labelled the pixel coordinates of the joints for each of
the images. From the total dataset, 224 images were used for training and rest
were used for testing.
### 4.5 Training
Our second approach requires training for which training and validation
dataset has been used as specified under sub-heading 4.4. For the optimization
purpose Adam optimizer has been used, with batch size of 14 images. The
training has been done on the sample dataset for the case of one-link and two-
link cases( forearm and arm) both on CNN network and Refinement network and
the combined network. CNN network has been trained on 10 epoch and Combined
Network on 100 epoch.
## 5 Experiment
The proposed method is tested with both real-time and recorded video
sequences. But the processing of both type of the sequences are carried out in
real-time since the ’mimic’ of robotic arm is a real-time task. The
experiments adressing four different challenges viz., i)presence of different
level of background noise, ii) variation in distance between arm and camera,
iii) variation in speed of hand movement, and iv) different number of links
are performed to test the proposed methodology and setups are accordingly
made. The experiments has been performed at frame rate of 30 fps and video
resolution of 240x320 pixels.
### 5.1 Experiment-1
In the first experiment, the recorded video of hand motion is given as input
to the network and the motion of the arm is mimicked by one-link manipulator
in virtual environment of PyBullet. This experiment requires preparation of
virtual environment and no physical setup is required. For the recorded video
even noisy data was used to test the method. The recorded data used for
testing is of [15] (lossy compressed AVI format devel-1). Additional setups
were not made for getting the recorded video. Videos used are available here
as ( Exp-1 Video-1, Exp-1 Video-2)
### 5.2 Experiment-2
In the second experiment, the motion of actual one-link manipulator against
stationary background is mimicked in simulation environment. This required
preparation of physical setup of the manipulator and camera. The experimental
setup consists of a camera(Model: HP HD 4310 H2W19AA) is mounted(fixed) at a
height of 32 cm(can be varied) above the manipulator. The manipulator’s link
length is 10.5 cm which is made up of paper to reduce the weight of the link
mounted on the servo motor(specifications: Model: SM-S2309S, Size: $22.9\times
12.3\times 22.2$mm, Weight: 9.9g, Rotation angle $\equiv 120$ , Micro analog
servo, 4 plastic gears + 1 metal gear). An Arduino UNO board has been used as
a controller to provide signal to the servo motor for the motion (Fig.2(a)).
The videos of experiment is available here as ( Exp-2 Video-1, Exp-2 Video-2)
(a) (b) (c)
Figure 4: (a) Experimental setup for experiment-2, (b) Experimental setup for
experiment-3, (c) Manipulator for experiment-4
### 5.3 Experiment-3
In the third experiment, the proposed method is used to mimic the forearm
motion of a standing person by an actual one-link manipulator. The one-link
manipulator used in Experiment-2 has been used here except the positioning has
been changed as shown in Fig.2(b). Here, the distance between the forearm and
the laptop(HP laptop AU030WM Pavilion) camera is varied. whereas the camera
was mounted at fixed distance from the manipulator in earlier setup. Also, the
background here is not stationary as noise is introduced while moving forearm
other body parts too move slightly.
### 5.4 Experiment-4
In the fourth experiment, we extend our method to n-link planar
manipulator(four-link, one-link is fixed). Here in case of human arm, we
considered forearm, thumb and index finger having three joints overall. One
additional layer to DON is added for the sake of experimentation here. It is
shown in Fig. 9. Please note no physical setup is prepared for this experiment
and the testing is done in simulation. The manipulator used in Pybullet is
shown in Fig.2(c)
### 5.5 Parameter Tuning
The actuator of our experimental setup enabled us to test the method using
position control. However, velocity control and torque control techniques can
also be used with the values obtained from the algorithm with the actuators
that enables velocity control and torque control. For the uniformity, we use
position control in simulator as well as on actual manipulator.
The joint angle computed in optical flow and the actual joint angle remains
the same. Same is not the case with angular velocity, angular acceleration,
torque and the link length. The term aspect ratio(ratio of value in optical
flow and actual value) has been introduced for mapping optical flow value to
actual value. These values shall be experimentally determined and is dependent
on the experimental setup.
Since the algorithm is completely unsupervised, the requirement of labelled
data and domain adaptation is not required. The issue of cost is also dealt
since the algorithm can easily run on low computation powered devices such as
mobile handset, laptops, computers, etc. There is only lag time between the
input of the image frame from video and the output signal to the manipulator
which is the processing time.
(a) (b)
Figure 5: (a) Computation time vs No of frames considered, (b) Average area
captured vs intensity level
This processing time is a function of number of input frames (Fig.3(a)) in
Input Layer, filters used, filter sizes, etc. The lag time can be decreased by
using optimal number of input frames in Input Layer and appropriate filter
with optimal filter size. For the case of experimentation the number of input
frames is $10$ with Gaussian filter size $15$x$15$ which has been determined
as optimal values through an iterative process.
Further the performance of Object Identification Layer is also dependent on
intensity of light and background noise(movement of other objects). The
algorithm works well for the intensity of light above 50%(determined using
experimental setup in Fig.2(a)) shown in Fig.3(b) and small background noise.
The algorithm captures the major motion in the video. So, the performance of
the algorithm is not affected until the major motion is of the arm.
## 6 Results and Discussions
### 6.1 Results of DON
The Experiment-1 performed on hand motion data set could hardly track the arm
joint angle due to extremely random and fast hand movement and very large
noise due to movement of other body part. This experiment was performed on
both the RGB video and depth video. The results are available here RGB
videos(Video-1 at 10fps, Video-2 at 30fps) and Depth videos (Video-1 at 10fps,
Video-2 at 30fps)
The Experiment-2 performed for mimicking actual one-link manipulator motion in
simulation result is shown in Fig.4. In addition we used [16] to generate
depth images from RGB image and tested our algorithm. In both the cases our
result obtained is the same.
(a) (b)
Figure 6: Plot of (a)Angular velocity vs Time, (b) Error vs Time Figure 7:
Visual _i_ -mimic in Real-time: (a) Without occlusion ($0.9<\eta<1$), (b) Low
occlusion to background and human hand ($0.7<\eta<0.5$) and (c) High occlusion
($0.4<\eta<0.3$)
The Experiment-3 is performed with variation of trust factor ($\eta$ in Eqn.
(7)) that is with various degrees of occlusion. Example frames with arm-mimic
for three different type of occlusions are shown in Fig.9, where there is no
occlusion present there in Fig. 9(a), low amount of occlusion is present there
in Fig. 9(b) and the amount of occlusion is quite high in Fig. 9(c). It is
also observed that the proposed method works well for moving hand. The videos
corresponding to the experimental results are: no occlusion, minor occlusion
and major occlusion.
Figure 8: Adjusted portion of Architecture of Deep Flow Network
In Experiment-4, since there are three rotating link, three-joints angles are
to be estimated. To accommodate multiple links, in the architecture of DON, an
additional layer is added between Object Identification Layer and Estimation
Layer to fit straight lines(Object Regression Layer) on objects as shown in
Fig.6. Rest of the network remains the same. The joint angle between forearm
and fixed link, thumb and forearm, index-finger and forearm are estimated
accordingly. The results obtained after Object Regression Layer is shown in
Fig.7.
The video expressing the obtained result is here.
Figure 9: Snapshot image after Object Regression Layer
In addition to, the algorithm has been tested on videos of different
resolutions at 30 fps. The average algorithm run time per loop execution for
video input of different resolution before giving signal to manipulator for is
presented in Table 1. This experiment has been performed on DELL Laptop with 8
GB RAM and Intel core i5, the algorithm run time will vary depending upon the
computation power of hardware used for testing.
Table 1: Average loop run time for videos of different resolutions. Resolution | Average loop run time(in milliseconds)
---|---
240x320 | 20
480x640 | 45
720x960 | 68
960x1280 | 124
### 6.2 Results of DNN
The CNN network and Refiner network has been tested on single link and two
link cases. The loss occurred during the training are shown in Fig. 10 and
Fig. 11. Likewise, the scattered plots of actual coordinate and predicted
coordinates by combined Networks of one-link, shoulders joint and wrist joint
obtained are also shown in Fig. 10 and Fig. 11.
Figure 10: Result for single link arm Figure 11: Results for two link arm
### 6.3 Discussions
As we stated earlier, four different experiments under different
circumstances, joints are performed here. In case of extremely random and very
fast hand movement, the method is found to be ineffective as in Experiment-1
while in Experiment-2 where the background noise, arm speed are limited, the
method performance i.e. mimicking is near to perfect. There is negligible lag
time because the arm control command is given to simulator which runs nearly
at 240Hz in Pybullet. In Experiment-3, the mimicking is performed with small
lag time. This lag time(as seen in video) has been caused due to hardware
limitation of the manipulator and setup. This can be reduced with good enough
hardware, since the algorithm has been fine-tuned with optimal parameter
values. The algorithm extended to n-link planar manipulator in Experiment-4 is
able to estimate the three-joint angles between lines accurately as depicted
in Fig.6. The algorithm can be extended to n-link planar manipulator just by
introducing Object Regression Layer and the outcome would be as desired.
Further, experimented with videos of different resolution shows the algorithm
run time increases with increase in video resolution. In addition, our
algorithm tracks/detects the objects on the basis of motion and not
probabilistic color distribution or the object features. This enables our
algorithm to run on both RGB and depth images independently and give the same
result.
In case of our second method, for the one-link(forearm) case the Predicted
Joint positions improved after adding a second network. The second network has
acted as a smoothing/refining network which refines the outcome of the first
network. For the two link cases, the performance of the combined network is
better compared to just use of CNN Network.
### 6.4 Comparative Study
Please note that, first approach’s application that we proposed here is new to
literature. Therefore, no similar method is available to compare with.
Therefore, we focus on comparing the proposed tracking algorithm. We could not
conduct any direct comparative study for our tracking method too since no
other tracking algorithm, formulated so far, gives torque and velocity as the
output. For example, we carried out the same experimentation with two other
robust and popular unsupervised tracking algorithms, namely, MoG2 and CAMSHIFT
for the sake of comparison. In this study we verified that the proposed
algorithm tracks down the object within about 30 ms, whereas the time consumed
by CAMSHIF and MoG2 are about 149 ms and 100ms respectively on HP laptop
AU030WM Pavilion. Besides, these methods is not robust enough since they loose
the object trajectory even to stationary background, therefore, its
performance get reduced with reduction of trust factor. Above all, none of the
algorithm enables us to compute values of kinematic and dynamic parameters of
motion like our algorithm for mimicking and hence failed to mimic. This is the
main cause why we failed to carry out suitable comparative study for this
application.
For the second approach of deep learning, we compare our results to that of
[14] using Percentage correct parts(PCP) at link length threshold of 0.5(PCP
0.5), for upper arm and lower arm. PCP 0.5 was calculated on our model using
our dataset on 20 images and also on 40 images. The comparison results are
shown in Table 2 and 3 respectively. The data taken from [14] is generalised
on large dataset that has been used, for our case the method has not been
generalised and results shown are obtained on our dataset.
Table 2: Comparison PCP(0.5) on 20 images Model | Upper Arm | Lower Arm
---|---|---
Deep Pose 1st | 0.5 | 0.27
Deep Pose 2nd | 0.56 | 0.35
Deep Pose 3rd | 0.56 | 0.35
CNN Network | 0.40 | 0.1
CNN + Refinement Network | 0.7 | 0.35
Table 3: Comparison PCP(0.5) on 40 images Model | Upper Arm | Lower Arm
---|---|---
Deep Pose 1st | 0.5 | 0.27
Deep Pose 2nd | 0.56 | 0.35
Deep Pose 3rd | 0.56 | 0.35
CNN Network | 0.25 | 0.125
CNN + Refinement Network | 0.566 | 0.275
From the above table, it is clear that our combined network comprising of CNN
and Refiner Network performs better compared to just CNN network and Deep Pose
Network for Upper Arm. However for lower the results are as good as that of
Deep pose for the case of 20 images. On increasing the number of images to 40,
performance decreases but still is better than that of Deep Pose and CNN
Network for Upper Arm. However, for lower arm combined is poor.
## 7 Conclusions and Future Work
In the proposed work we aimed to develop a method with which robotic arms
could be controlled only by showing video sequences in real-time. It is tested
both with unsupervised DON network and deep CNN with manually labelled
training samples. This approach is proven to be successful with adequate
amount of demonstration shown here. The unsupervised DON based method is
proved to be effective in achieving control over n-link manipulator in 2-D
plane. Control over n-link manipulator plane can be achieved better with
hybridization of CNN and refinement network as shown here in the study. The
proposed technique performs well both with real-arm manipulator and synthetic
arms. The mimic-based control therefore could be implemented to control
robotic arm in different tasks. The approach is in its entry level as of now
and more complex scenarios could be addressed in future with mimicking the
motion of more than one joint on manipulator in spatial 3D environment.
## References
* [1] R. Basu, S. Padage, Development of 5 dof robot arm-gripper for sorting and investigating rtm concepts, Materials Today 4 (2) (2017) 1634–1643.
* [2] M. H. Ali, K. Aizat, K. Yerkhan, T. Zhandos, O. Anuar, Vision-based robot manipulator for industrial applications, Procedia Computer Science, Elsevier 133 (2) (2018) 205–212.
* [3] A. Geiger, P. Lenz, C. Stiller, R. Urtasun, Vision meets robotics: The kitti dataset, The International Journal of Robotics Research 32 (11) (2013) 1231–1237.
* [4] S. Levine, P. Pastor, A. Krizhevsky, J. Ibarz, D. Quillen, Learning hand-eye coordination for robotic grasping with deep learning and large-scale data collection, The International Journal of Robotics Research 37 (4-5) (2018) 421–436.
* [5] A. Hundt, V. Jain, C. Paxton, G. D. Hager, Training frankenstein’s creature to stack: Hypertree architecture search, arXiv preprint arXiv:1810.11714.
* [6] A. A. Rusu, M. Vecerik, T. Rothorl, N. Heess, R. Pascanu, R. Hadsell, Sim-to-real robot learning from pixels with progressive nets, in: International Conference on Robotics Learning (CoRL), 2017.
* [7] K. Bousmalis, A. Irpan, P. Wohlhart, Y. Bai, M. Kelcey, M. Kalakrishnan, L. Downs, J. Ibarz, P. Pastor, e. a. K. Konolige, Using simulation and domain adaptation to improve efficiency of deep robotic grasping, in: International Conference on Robotics and Automation, 2018.
* [8] E. Tzeng, C. Devin, J. Hoffman, C. Finn, P. Abbeel, S. Levine, K. Saenko, T. Darrell, Adapting deep visuomotor representations with weak pairwise constraints, in: IEEE Conference on Computer Vision and Pattern Recognition (CVPR), IEEE, 2017.
* [9] Y. Zuo, W. Qiu, L. Xie, F. Zhong, Y. Wang, A. L. Yuille, Craves: Controlling robotic arm with a vision-based economic system, in: IEEE Conference on Computer Vision and Pattern Recognition (CVPR), IEEE, CA, USA, 2019, pp. 4214–4223.
* [10] Y. Ganin, V. S. Lempitsky, Unsupervised domain adaptation by backpropagation, in: International Conference on Machine Learning (ICML), 2015.
* [11] Y. Ganin, E. Ustinova, Ajakan, P. Germain, H. Larochelle, F. Laviolette, M. Marchand, Domain-adversarial neural networks, Journal of Machine Learning Research 17 (1) (2016) 1–35.
* [12] K. Bousmalis, G. Trigeorgis, N. Silberman, D. Krishnan, D. Erhan, Domain separation networks, CoRR abs/1608.06019.
* [13] A. Singh, L. Yang, S. Levine, GPLAC: generalizing vision-based robotic skills using weakly labeled images, CoRR abs/1708.02313.
* [14] A. Toshev, C. Szegedy, Deeppose: Human pose estimation via deep neural networks, in: Proceedings of the IEEE conference on computer vision and pattern recognition, 2014, pp. 1653–1660.
* [15] C. G. Dataset, Chalearn gesture dataset (cgd 2011), chalearn, california, 2011, data retrieved from ChaLearn Gesture Dataset , http://gesture.chalearn.org/data/cgd2011.
* [16] C. Godard, O. Mac Aodha, M. Firman, G. J. Brostow, Digging into self-supervised monocular depth prediction, in: The International Conference on Computer Vision (ICCV), 2019.
|
2.5cm2.5cm2.5cm2.5cm
# Iwahori-Hecke model for mod $p$ representations of ${\rm GL}_{2}(F)$
U. K. Anandavardhanan and Arindam Jana Department of Mathematics, Indian
Institute of Technology Bombay, Mumbai - 400076, India<EMAIL_ADDRESS>Department of Mathematics, Indian Institute of Technology Bombay, Mumbai -
400076, India<EMAIL_ADDRESS>
###### Abstract.
For a $p$-adic field $F$, the space of pro-$p$-Iwahori invariants of a
universal supersingular mod $p$ representation $\tau$ of ${\rm GL}_{2}(F)$ is
determined in the works of Breuil, Schein, and Hendel. The representation
$\tau$ is introduced by Barthel and Livné and this is defined in terms of the
spherical Hecke operator. In [AB13, AB15], an Iwahori-Hecke approach was
introduced to study these universal supersingular representations in which
they can be characterized via the Iwahori-Hecke operators. In this paper, we
construct a certain quotient $\pi$ of $\tau$, making use of the Iwahori-Hecke
operators. When $F$ is not totally ramified over $\mathbb{Q}_{p}$, the
representation $\pi$ is a non-trivial quotient of $\tau$. We determine a basis
for the space of invariants of $\pi$ under the pro-p Iwahori subgroup. A
pleasant feature of this ”new” representation $\pi$ is that its space of
pro-$p$-Iwahori invariants admits a more uniform description vis-à-vis the
description of the space of pro-$p$-Iwahori invariants of $\tau$.
###### 1991 Mathematics Subject Classification:
Primary 20G05; Secondary 22E50, 11F70
## 1\. Introduction
For a $p$-adic field $F$, the study of irreducible smooth mod $p$
representations of $\rm GL_{2}(F)$ started with the famous work of Barthel and
Livné [BL94]. They showed that there exist irreducible smooth representations,
called supersingular representations, which cannot be obtained as a
subquotient of a parabolically induced representation.
It is shown in [BL94] that a supersingular representation can be realized as
the quotient of a universal module constructed as follows. Let $G=\rm
GL_{2}(F)$ and let $K$ be its standard maximal compact subgroup. Let $Z$
denote the center of $G$. For an irreducible representation $\sigma$ of $KZ$,
let ind${}_{KZ}^{G}\sigma$ be the representation of $G$ compactly induced from
$\sigma$. Its endomorphism algebra is a polynomial algebra in one variable:
${\rm End}_{G}\left({\rm
ind}_{KZ}^{G}\sigma\right)\simeq\overline{\mathbb{F}}_{p}[T],$
where $T$ is the standard spherical Hecke operator and
$\overline{\mathbb{F}}_{p}$ denotes an algebraic closure of the finite field
$\mathbb{F}_{p}$ of $p$ elements [BL94, Proposition 8]. The universal module
in consideration is
$\tau=\frac{{\rm ind}_{KZ}^{G}\sigma}{(T)}$
and a supersingular representation of $G$ is an irreducible quotient of the
universal module for some $\sigma$ of $KZ$ up to a twist by a character
[BL94].
Explicitly constructing a supersingular representation of $\rm GL_{2}(F)$ is a
challenging problem when $F\neq\mathbb{Q}_{p}$ [BP12]. When
$F=\mathbb{Q}_{p}$, Breuil proved that the universal representation $\tau$
itself is irreducible [Bre03, Theorem 1.1]. The key step in Breuil’s proof of
the irreducibility of $\tau$ is the explicit computation of its
$I(1)$-invariant space, which is of dimension $2$, where $I(1)$ is the
pro-$p$-Iwahori subgroup of $K$ [Bre03, Theorem 3.2.4]. The space of
$I(1)$-invariants of $\tau$ is infinite dimensional when
$F\neq\mathbb{Q}_{p}$. An explicit basis for this infinite dimensional space
is computed by Schein when $F$ is totally ramified over $\mathbb{Q}_{p}$
[Sch11, §2] and by Hendel more generally for any $p$-adic field $F$ [Hen19,
Theorem 1.2].
One can also construct a universal module from the perspective of the Iwahori-
Hecke operators instead of the spherical Hecke operator $T$ [AB13, AB15]. For
this, instead of doing compact induction from an irreducible representation of
$KZ$, we start with a regular character $\chi$ of $IZ$, where $I$ is the
Iwahori subgroup $K$, and consider the compactly induced representation
ind${}_{IZ}^{G}\chi$. Its endomorphism algebra is [BL94, Proposition 13]:
${\rm End}_{G}({\rm
ind}_{IZ}^{G}\chi)\simeq\frac{\overline{\mathbb{F}}_{p}[T_{-1,0,},T_{1,2}]}{(T_{-1,0}T_{1,2},T_{1,2}T_{-1,0})},$
where $T_{-1,0}$ and $T_{1,2}$ are the Iwahori-Hecke operators. When $F$ is a
totally ramified extension of $\mathbb{Q}_{p}$, it is proved in [AB15,
Proposition 3.1 & Remark 1] that the image of one of these operators is equal
to the kernel of the other; i.e.,
(1) ${\rm Im}~{}T_{-1,0}={\rm Ker}~{}T_{1,2}~{}\&~{}{\rm Im}~{}T_{1,2}={\rm
Ker}~{}T_{-1,0}.$
Let $\mathbb{F}_{q}$ be the residue field of $F$ where $q=p^{f}$. Assume
$0<r<q-1$ and write $r=r_{0}+r_{1}p+\dots+r_{f-1}p^{f-1}$ with $0\leq
r_{i}\leq p-1$ for $0\leq i\leq f-1$. Let
$\sigma_{r}={\rm Sym}^{r_{0}}\overline{\mathbb{F}}_{p}^{2}\otimes{\rm
Sym}^{r_{1}}\overline{\mathbb{F}}_{p}^{2}\circ{\rm
Frob}\otimes\dots\otimes{\rm
Sym}^{r_{f-1}}\overline{\mathbb{F}}_{p}^{2}\circ{\rm Frob}^{f-1}$
be an irreducible representation of $\rm GL_{2}(\mathbb{F}_{q})$, where Frob
is the Frobenius morphism. We continue to denote the corresponding irreducible
representation of $K$, obtained via inflation, by $\sigma_{r}$. Similarly, let
$\chi_{r}$ be the character of $I$, valued in
$\overline{\mathbb{F}}_{p}^{\times}$, obtained via the character of the Borel
subgroup of $\rm GL_{2}(\mathbb{F}_{q})$ defined by
$\left(\begin{array}[]{cc}a&b\\\ \ 0&d\end{array}\right)\mapsto d^{r}.$
We fix a uniformizing element $\varpi$ of the ring of integers $\mathcal{O}$
of $F$. The representation $\sigma_{r}$ is treated as a representation of $KZ$
by making diag$(\varpi,\varpi)$ acting trivially and similarly the character
$\chi_{r}$ is treated as a character of $IZ$.
For $g\in G$ and $v\in\sigma_{r}$, let $g\otimes v$ be the function in ${\rm
ind}_{KZ}^{G}\sigma_{r}$ supported on $KZg^{-1}$ that sends $g^{-1}$ to
$\sigma_{r}(k)v$. Similarly, for $g\in G$, by $[g,1]$ we define the function
in ${\rm ind}_{IZ}^{G}\chi_{r}$ which is supported on $IZg^{-1}$ and sending
$g^{-1}$ to $1$. It can be seen that every element of ${\rm
ind}_{IZ}^{G}\chi_{r}$ (resp. ${\rm ind}_{KZ}^{G}\sigma_{r}$) is a finite sum
of these type of functions $[g,1]$ (resp. $g\otimes v$).
Now [AB15, Theorem 1.1] takes the form:
###### Theorem 1.1.
Let $F$ be a finite extension of $\mathbb{Q}_{p}$ with residue field
$\mathbb{F}_{q}$ and residue degree $f$. Let $0<r<q-1$ and
$r=r_{0}+r_{1}p+\dots+r_{f-1}p^{f-1}$ with $0\leq r_{i}\leq p-1$. Then
$\tau_{r}=\frac{{\rm ind}_{KZ}^{G}\sigma_{r}}{(T)}\simeq\frac{{\rm
ind}_{IZ}^{G}\chi_{r}}{({\rm Im}~{}T_{1,2},{\rm Ker}~{}T_{1,2})}.$
Moreover, this isomorphism is determined by
${\rm Id}\otimes\bigotimes_{j=0}^{f-1}x_{j}^{r_{j}}\mod
T\mapsto[\beta,1]\mod({\rm Im}~{}T_{1,2},{\rm Ker}~{}T_{1,2}).$
###### Remark 1.
Theorem 1.1 is stated and proved in [AB15, Theorem 4.1] when $F$ is a totally
ramified extension of $\mathbb{Q}_{p}$ (see [AB15, Remark 3]) and exactly the
same proof goes through in the general case as well.
###### Remark 2.
As mentioned earlier, the space of $I(1)$-invariants of $\tau_{r}$ is computed
by Hendel [Hen19, Theorem 1.2]. Stating an explicit basis for this space
involves four cases; (i) $e=1,f=1$, (ii) $e>1,f=1$, (iii) $e=1,f>1$, and (iv)
$e>1,f>1$.
In this paper, we study a new universal representation given by
$\pi_{r}=\frac{{\rm ind}_{IZ}^{G}\chi_{r}}{({\rm Ker}~{}T_{-1,0},{\rm
Ker}~{}T_{1,2})}$
which is a further quotient of $\tau_{r}$. Note that this representation
equals $\tau_{r}$ when $F$ is totally ramified over $\mathbb{Q}_{p}$ by (1).
We show that when $F$ is not totally ramified over $\mathbb{Q}_{p}$, we have
strict containments
(2) ${\rm Im}~{}T_{-1,0}\subsetneq{\rm Ker}~{}T_{1,2}~{}\&~{}{\rm
Im}~{}T_{1,2}\subsetneq{\rm Ker}~{}T_{-1,0}.$
and thus we have a new representation to investigate for its properties (cf.
Remark 4). At this stage, we also note that the representation $\pi_{r}$ is
indeed non-trivial (cf. Lemma 3.4).
The main result of this paper gives an explicit basis for the space of
$I(1)$-invariants of $\pi_{r}$. This space turns out to be infinite
dimensional as well as in the case of [Hen19, Theorem 1.2]. However, in this
case the basis can be written in a uniform manner whenever
$F\neq\mathbb{Q}_{p}$. Thus, the statement involves only two cases; (i)
$F=\mathbb{Q}_{p}$ and (ii) $F\neq\mathbb{Q}_{p}$. It is interesting to
compare our result with that of Hendel in this aspect (cf. Remark 2).
In order to state the theorem, we introduce a few more notations. Set
$I_{0}=\\{0\\}$, and for $n\in\mathbb{N}$, let
$I_{n}=\left\\{[{\mu}_{0}]+[{\mu}_{1}]\varpi+\dots+[\mu_{n-1}]{\varpi}^{n-1}\mid{\mu}_{i}\in\mathbb{F}_{q}\right\\}\subset\mathcal{O},$
where, for $x\in\mathbb{F}_{q}$, we denote its multiplicative representative
in $\mathcal{O}$ by $[x]$. If $0\leq m\leq n,$ let
$[\cdot]_{m}:I_{n}\rightarrow I_{m}$ be the truncation map defined by
$\sum\limits_{\begin{subarray}{c}i=0\end{subarray}}^{n-1}[\lambda_{i}]\varpi^{i}\mapsto\sum\limits_{\begin{subarray}{c}i=0\end{subarray}}^{m-1}[\lambda_{i}]\varpi^{i}.$
Let us denote
$\alpha=\left(\begin{array}[]{cc}1&0\\\
0&\varpi\end{array}\right),~{}\beta=\left(\begin{array}[]{cc}0&1\\\
\varpi&0\end{array}\right),~{}w=\left(\begin{array}[]{cc}0&1\\\
1&0\end{array}\right),$
and observe that $\beta=\alpha w$ normalizes $I(1)$. For any $n\in\mathbb{N}$,
we denote
$\displaystyle s_{n}^{k}$
$\displaystyle=\sum\limits_{\begin{subarray}{c}\mu\in
I_{n}\end{subarray}}\mu_{n-1}^{k}\left[\left(\begin{array}[]{cc}\varpi^{n}&\mu\\\
0&1\end{array}\right),1\right],$ $\displaystyle t_{n}^{s}$
$\displaystyle=\sum\limits_{\begin{subarray}{c}\mu\in
I_{n}\end{subarray}}\mu_{n-1}^{s}\left[\left(\begin{array}[]{cc}\varpi^{n-1}&[\mu]_{n-1}\\\
0&1\end{array}\right)\left(\begin{array}[]{cc}1&[\mu_{n-1}]\\\
0&1\end{array}\right)w,1\right],$
where $0\leq k,s\leq q-1$. For $0\leq l\leq f-1$ and $m\geq 1$, we define the
following sets
$\displaystyle\mathcal{S}_{m}^{l}$
$\displaystyle=\\{s_{n}^{q-1-r+p^{l}}\\}_{n\geq m}\cup\\{\beta
s_{n}^{q-1-r+p^{l}}\\}_{n\geq m}$ $\displaystyle\mathcal{S}_{m}$
$\displaystyle=\bigcup_{l=0}^{f-1}\mathcal{S}_{m}^{l},$
$\displaystyle\mathcal{T}_{m}^{l}$ $\displaystyle=\\{t_{n}^{r+p^{l}}\\}_{n\geq
m}\cup\\{\beta t_{n}^{r+p^{l}}\\}_{n\geq m},$ $\displaystyle\mathcal{T}_{m}$
$\displaystyle=\bigcup_{l=0}^{f-1}T_{m}^{l}.$
Now we state the main theorem of this paper.
###### Theorem 1.2.
Let $F$ be a finite extension of $\mathbb{Q}_{p}$ with ramification index $e$.
Let $\mathbb{F}_{q}$ be the residue field of $F$ with $q=p^{f}.$ Let $0<r<q-1$
and $r=r_{0}+r_{1}p+\dots+r_{f-1}p^{f-1}$ with $0<r_{j}<p-1$ for all $0\leq
j\leq f-1$. When $f=1$, we assume $2<r<p-3$. Then a basis of the space of
$I(1)$-invariants of the representation
$\pi_{r}=\frac{{\rm ind}_{IZ}^{G}\chi_{r}}{({\rm Ker}~{}T_{-1,0},{\rm
Ker}~{}T_{1,2})}$
as an $\overline{\mathbb{F}}_{p}$-vector space is given by the images of the
following sets in $\pi_{r}$:
1. (1)
$\left\\{\left[{\rm Id},1\right],\left[\beta,1\right]\right\\}$ when
$F=\mathbb{Q}_{p}$
2. (2)
$\mathcal{S}_{2}\bigcup\left\\{\left[{\rm
Id},1\right],\left[\beta,1\right]\right\\}\bigcup\mathcal{T}_{2}~{}$ when
$F\neq\mathbb{Q}_{p}$.
###### Remark 3.
The representation $\pi_{r}$ that we construct and investigate in this paper
is a quotient of the representation $\tau_{r}$ considered in [BL94, Bre03,
Sch11, Hen19];
$0\rightarrow\frac{{\rm Ker}~{}T_{-1,0}}{{\rm
Im~{}}T_{1,2}}\rightarrow\tau_{r}\rightarrow\pi_{r}\rightarrow 0.$
When $F$ is totally ramified over $\mathbb{Q}_{p}$, the representations
$\tau_{r}$ and $\pi_{r}$ are isomorphic by Theorem 1.1 together with the
equality of spaces in (1). However, $\pi_{r}$ is a “new” representation when
$F$ is not totally ramified over $\mathbb{Q}_{p}$. That there is no
isomorphism between $\tau_{r}$ and $\pi_{r}$ can be checked, for instance,
from the characterization of the space of $I(1)$-invariants of $\pi_{r}$ in
Theorem 1.2 and that of $\tau_{r}$ in [Hen19, Theorem 1.2]. We give more
details in §4.5.
Following the argument in [Hen19, Conclusion 3.10] word to word, we get the
following corollary to Theorem 1.2.
###### Corollary 1.3.
The representation $\pi_{r}$ is indecomposable; i.e., ${\rm
End}_{G}(\pi_{r})\simeq\overline{\mathbb{F}}_{p}.$
The plan of the paper is as follows. We collect many results about the
Iwahori-Hecke operators in Section 3. Several of these results are contained
in some form in [AB13, AB15]. Theorem 1.2 and the key ideas in its proof are
inspired by the work of Hendel [Hen19], though the Iwahori-Hecke approach
which is employed in this paper as in [AB13, AB15] seems to be more amenable
to carrying out the necessary calculations. We take up the proof in Section 4.
## 2\. Two basic results
As in the work of Hendel [Hen19], we will need to frequently make use of the
following two results in our computations.
The first one is the classical result in modular combinatorics due to Lucas
which gives a condition for a binomial coefficient ${n\choose r}$ to be zero
modulo $p$.
###### Theorem 2.1 (Lucas).
Let $n,r\in\mathbb{N}$ be such that
$n=\sum\limits_{\begin{subarray}{c}i=0\end{subarray}}^{k}n_{i}p^{i}$ and
$r=\sum\limits_{\begin{subarray}{c}i=0\end{subarray}}^{k}r_{i}p^{i},$ where
$0\leq n_{i}\leq p-1$ and $0\leq r_{i}\leq p-1.$ Then
${n\choose r}\equiv\prod_{i=0}^{k}{n_{i}\choose r_{i}}\mod p.$
###### Corollary 2.2.
Let $n,r\in\mathbb{N}.$ Then $p$ divides ${n\choose r}$ if and only if
$n_{i}<r_{i}$ for some $0\leq i\leq k.$
The next result gives a formula for adding multiplicative representatives in
$\mathcal{O}$ [Hen19, Lemma 1.7]. As in [Hen19], this formula will play a
crucial role in the calculations to follow.
###### Lemma 2.3.
Let $x,y\in\mathbb{F}_{q}$ with $q=p^{f}$. Then
$[x]+[y]\equiv[x+y]+\varpi^{e}[P_{0}(x,y)]\mod\varpi^{e+1},$
where $P_{0}(x,y)=\frac{x^{q^{e}}+y^{q^{e}}-(x+y)^{q^{e}}}{\varpi^{e}}.$
## 3\. Preliminaries on the Iwahori-Hecke operators
For $n\in\mathbb{N}\cup\\{0\\}$ and $\lambda\in I_{n}$, define
$g_{n,\lambda}^{0}=\left(\begin{array}[]{cc}\varpi^{n}&\lambda\\\
0&1\end{array}\right)~{}~{}\&~{}~{}g_{n,\lambda}^{1}=\left(\begin{array}[]{cc}1&0\\\
\varpi\lambda&\varpi^{n+1}\end{array}\right).$
We have the relations
$g_{0,0}^{0}={\rm Id},~{}g_{0,0}^{1}=\alpha,\beta
g_{n,\lambda}^{0}=g_{n,\lambda}^{1}w.$
Now $G$ acts transitively on the Bruhat-Tits tree of ${\rm SL}_{2}(F)$, whose
vertices are in a $G$-equivariant bijection with the cosets $G/{KZ}$ and whose
oriented edges are in a $G$-equivariant bijection with the cosets $G/{IZ}$. We
have the explicit Cartan decomposition given by
$G=\underset{\begin{subarray}{c}i\in\\{0,1\\}\\\ n\geq 0,~{}\lambda\in
I_{n}\end{subarray}}{\coprod}g_{n,\lambda}^{i}KZ$
and an explicit set of coset representatives of $G/IZ$ is given by
(3)
$\left\\{g_{n,\lambda}^{0},g_{n,\lambda}^{0}\left(\begin{array}[]{cc}1&\mu\\\
0&1\end{array}\right)w,g_{n,\lambda}^{1}w,g_{n,\lambda}^{1}w\left(\begin{array}[]{cc}1&\mu\\\
0&1\end{array}\right)w\right\\}_{n\geq 0,\lambda\in I_{n}},$
where $\mu\in I_{1}.$
Now we recall a few details about the Iwahori-Hecke algebra [BL94, §3.2]. By
definition, this algebra, denoted by $\mathcal{H}(IZ,\chi_{r})$, is the
endomorphism algebra of the compactly induced representation ${\rm
ind}_{IZ}^{G}\chi_{r}$. For $n\in\mathbb{Z}$, let $\phi_{n,n+1}$ denote the
convolution map supported on $IZ{\alpha}^{-n}I$ such that
$\phi_{n,n+1}(\alpha^{-n})=1$ [BL94, Lemma 9]. We denote by $T_{n,n+1}$ the
corresponding element in $\mathcal{H}(IZ,\chi_{r})$. By [BL94, Proposition
13], for $0<r<q-1$, we have:
$\mathcal{H}(IZ,\chi_{r})\simeq\frac{\overline{\mathbb{F}}_{p}[T_{-1,0,},T_{1,2}]}{(T_{-1,0}T_{1,2},T_{1,2}T_{-1,0})}.$
Substituting $n=1$ in [BL94, (16), (17)], we have the following explicit
formulas for $T_{-1,0}$ and $T_{1,2}$:
(4) $T_{-1,0}(\left[g,1\right])=\sum\limits_{\begin{subarray}{c}\lambda\in
I_{1}\end{subarray}}\left[gg_{1,\lambda}^{0},1\right],$ (5)
$T_{1,2}(\left[g,1\right])=\sum\limits_{\begin{subarray}{c}\lambda\in
I_{1}\end{subarray}}\left[g\beta\left(\begin{array}[]{cc}1&\lambda\\\
0&1\end{array}\right)w,1\right].$
The following proposition characterizes the kernel of the Iwahori-Hecke
operators $T_{-1,0}$ and $T_{1,2}$ [AB13, AB15].
###### Proposition 3.1.
We have:
1. (1)
${\rm Ker~{}}T_{-1,0}$ is generated as a $G$-module by the vectors
1. (a)
$(-1)^{q-1-r}s_{0}^{0}+t_{1}^{r}$,
2. (b)
$t_{1}^{s}$ where $0\leq s\leq r-1$,
3. (c)
$t_{1}^{s}$ where $s>r$ and ${q-1-r\choose q-1-s}\equiv 0\mod p$.
2. (2)
${\rm Ker~{}}T_{1,2}$ is generated as a $G$-module by the vectors
1. (a)
$t_{0}^{0}+s_{1}^{q-1-r}$,
2. (b)
$s_{1}^{k}$ where $0\leq k\leq q-2-r$,
3. (c)
$s_{1}^{k}$ where $k>q-1-r$ and ${r\choose q-1-k}\equiv 0\mod p$.
###### Proof.
We indicate the proof for Ker $T_{1,2}$, with the other case being similar. An
arbitrary vector in ind${}_{IZ}^{G}\chi_{r}$ is an
$\overline{\mathbb{F}}_{p}$-linear combination of vectors $[g,1]$, where $g$
is in the set of coset representatives (3) of $G/IZ$. Arguing as in the proof
of [AB15, Proposition 3.1], we can restrict our attention to the vectors
$\left\\{[{\rm
Id},1],[\beta,1],[g_{1,\mu}^{0},1],\left[\left(\begin{array}[]{cc}1&\mu\\\
0&1\end{array}\right)w,1\right]\right\\}$
for $\mu\in I_{1}$. Now the proof boils down to elementary linear algebra as
in [AB15, Lemma 3.2], where one is led to analyse the indices $i$ for which
$\sum_{\mu\in\mathbb{F}_{q}}\mu^{i}(\mu-\lambda)^{r}=0,$
for $\lambda\in\mathbb{F}_{q}$. Alternatively, this last step can be deduced
directly from the explicit formulas for the Iwahori-Hecke operators in [AB13,
p. 63-64]. ∎
###### Remark 4.
We remarked in (2) in Section 1 that we have strict containments
(6) ${\rm Im}~{}T_{-1,0}\subsetneq{\rm Ker}~{}T_{1,2}~{}\&~{}{\rm
Im}~{}T_{1,2}\subsetneq{\rm Ker}~{}T_{-1,0}.$
when $F$ is not a totally ramified extension of $\mathbb{Q}_{p}$. The reason
for this is that the third type of vectors in both (1) and (2) in Proposition
3.1 do not belong to the images of the Iwahori-Hecke operators. Note that such
vectors do not exist when $f=1$; i.e., when $q=p$. By the argument in [AB15,
Lemma 3.2], it can be shown that the first two types of vectors are indeed in
the image of the relevant Iwahori-Hecke operator.
###### Corollary 3.2.
A basis of the space of $I(1)$-invariants of ${\rm Ker~{}}T_{-1,0}$ is given
by $\\{t_{n}^{0},\beta t_{n}^{0}\\}_{n\geq 1}$ and that of ${\rm
Ker~{}}T_{1,2}$ is given by $\\{s_{n}^{0},\beta s_{n}^{0}\\}_{n\geq 1}$.
Moreover, the action of $I$ is given by
$\left(\begin{array}[]{cc}a&b\\\ \varpi c&d\end{array}\right)\cdot
v=\begin{cases}a^{r}v&\text{$v=t_{n}^{0}$ or $\beta s_{n}^{0}$,}\\\
d^{r}v&\text{$v=s_{n}^{0}$ or $\beta t_{n}^{0}$.}\end{cases}$
###### Proof.
The first part of Proposition 3.1 together with the observation that the space
of $I(1)$-invariants of the full induced representation is given by
$\left({\rm ind}_{IZ}^{G}\chi_{r}\right)^{I(1)}=\langle
s_{n}^{0},t_{n}^{0},\beta s_{n}^{0},\beta t_{n}^{0}\rangle_{n\geq 0}.$
For the second part, observe that since
$I/{I(1)}=\left\\{\left(\begin{array}[]{cc}a&0\\\ 0&d\end{array}\right)\mid
a,d\in{\mathbb{F}_{q}}^{\times}\right\\},$
it follows that
$\left(\begin{array}[]{cc}a&b\\\ \varpi
c&d\end{array}\right)~{}\&~{}\left(\begin{array}[]{cc}a&0\\\
0&d\end{array}\right)$
have the same action on any $I(1)$-invariant vector. Now, for any $k\geq 0$,
we have
$\displaystyle\left(\begin{array}[]{cc}a&0\\\ 0&d\end{array}\right)s_{n}^{k}$
$\displaystyle=\left(\begin{array}[]{cc}a&0\\\
0&d\end{array}\right)\sum\limits_{\begin{subarray}{c}\mu\in
I_{n}\end{subarray}}\mu_{n-1}^{k}\left[\left(\begin{array}[]{cc}\varpi^{n}&\mu\\\
0&1\end{array}\right),1\right]$
$\displaystyle=\sum\limits_{\begin{subarray}{c}\mu\in
I_{n}\end{subarray}}\mu_{n-1}^{k}\left[\left(\begin{array}[]{cc}\varpi^{n}&ad^{-1}\mu\\\
0&1\end{array}\right)\left(\begin{array}[]{cc}a&0\\\
0&d\end{array}\right),1\right]$ $\displaystyle=d^{r}(da^{-1})^{k}s_{n}^{k}.$
A similar computation gives
$\displaystyle\left(\begin{array}[]{cc}a&0\\\ 0&d\end{array}\right)t_{n}^{s}$
$\displaystyle=a^{r}(da^{-1})^{s}t_{n}^{s}.$
Similarly, we can check the action on $\beta s_{n}^{k}$ and $\beta t_{n}^{k}$.
∎
Next, we recall [AB15, Proposition 3.3], whose proof in [loc. cit.] is valid
for any $q$.
###### Proposition 3.3.
We have
${\rm Ker}~{}T_{-1,0}\cap{\rm Ker}~{}T_{1,2}=\\{0\\}.$
As a corollary to Proposition 3.3, we have the following lemma.
###### Lemma 3.4.
For the iwahori-Hecke operators $T_{-1,0}$ and $T_{1,2}$, we have
${\rm ind}_{IZ}^{G}\chi_{r}\neq{\rm Ker}~{}T_{-1,0}\oplus{\rm Ker}~{}T_{1,2}.$
###### Proof.
If possible, let
${\rm ind}_{IZ}^{G}\chi_{r}={\rm Ker}~{}T_{-1,0}\oplus{\rm Ker}~{}T_{1,2}.$
Then we get $[{\rm Id},1]=v_{1}+v_{2}$ for some $v_{1}\in{\rm Ker}~{}T_{-1,0}$
and $v_{2}\in{\rm Ker}~{}T_{1,2}$. Then, for an element $g\in I$,
$g(v_{1}+v_{2})=\left(\begin{array}[]{cc}a&b\\\ \varpi
c&d\end{array}\right)(v_{1}+v_{2})=d^{r}[{\rm Id},1]=d^{r}(v_{1}+v_{2})$
and this implies
$gv_{1}-d^{r}v_{1}=-gv_{2}+d^{r}v_{2}=0,$
by Proposition 3.3. In particular, both $v_{1}$ and $v_{2}$ are
$I(1)$-invariant. By Corollary 3.2, $v_{1}$ is a linear combination of vectors
of the form $\\{\beta t_{n}^{0}\\}_{n\geq 1}$ and $v_{2}$ is a linear
combination of vectors of the form $\\{s_{n}^{0}\\}_{n\geq 1}$. But $[{\rm
Id},1]$ cannot be written as a linear combination of these types of vectors. ∎
We end this section with two more results which immediately follow from
considerations similar to Proposition 3.1. We state these in a ready to use
format here (see also [AB13, p. 63-64]).
###### Lemma 3.5.
Let $0\leq i_{j}\leq q-1$ for $0\leq j\leq n-1$ and
$\mu=[\mu_{0}]+[\mu_{1}]\varpi+\dots+[\mu_{n-1}]\varpi^{n-1}\in I_{n}.$ Write
$i_{n-1}=i_{n-1,0}+i_{n-1,1}p+\dots+i_{n-1,f-1}p^{f-1}.$ Then
* (1)
$\sum\limits_{\begin{subarray}{c}\mu_{0}\end{subarray}}\dots\sum\limits_{\begin{subarray}{c}\mu_{n-1}\end{subarray}}\mu_{0}^{i_{0}}\dots\mu_{n-1}^{i_{n-1}}\left[g^{0}_{n,\mu},1\right]\in{\rm
Ker}~{}T_{1,2}$ if and only if $0\leq i_{n-1}\leq q-2-r$ or $i_{n-1}>q-1-r$
such that $i_{n-1,j}<p-1-r_{j}$ for some $0\leq j\leq f-2$.
* (2)
$\sum\limits_{\begin{subarray}{c}\mu_{0}\end{subarray}}\dots\sum\limits_{\begin{subarray}{c}\mu_{n-1}\end{subarray}}\mu_{0}^{i_{0}}\dots\mu_{n-1}^{i_{n-1}}\left[g^{0}_{n-1,[\mu]_{n-1}}\left(\begin{array}[]{cc}1&[\mu_{n-1}]\\\
0&1\end{array}\right)w,1\right]\in{\rm Ker}~{}T_{-1,0}$ if and only if $0\leq
i_{n-1}\leq r-1$ or $i_{n-1}>r$ such that $i_{n-1,j}<r_{j}$ for some $0\leq
j\leq f-2$.
###### Remark 5.
Note that in Lemma 3.5, the range for $j$ is $0\leq j\leq f-2$ because
$i_{n-1}>q-1-r\implies i_{n-1,f-1}\geq p-1-r_{f-1}.$
###### Remark 6.
Note that the condition
$i_{n-1}>q-1-r~{}\&~{}i_{n-1,j}<p-1-r_{j}\mbox{~{}for~{}some~{}}0\leq j\leq
f-2$
in Lemma 3.5 (1) is precisely what gives, by Theorem 2.1,
${r\choose q-1-i_{n-1}}\equiv 0\mod p$
which is related to the condition in (c) of Proposition 3.1 (2). Similarly,
the condition
$i_{n-1}>r~{}\&~{}i_{n-1,j}<r_{j}\mbox{~{}for~{}some~{}}0\leq j\leq f-2$
in Lemma 3.5 (2) is related to (c) of Proposition 3.1 (1).
The following lemma is [AB13, Lemma 3.1]. We note that its proof in [loc.
cit.] is valid for any $q$.
###### Lemma 3.6.
Let $\mu=[\mu_{0}]+\dots+[\mu_{n-1}]\varpi^{n-1}\in I_{n}.$ Then modulo $({\rm
Ker}~{}T_{-1,0},{\rm Ker}~{}T_{1,2})$, we have the identities
1. (1)
$\sum\limits_{\begin{subarray}{c}\mu_{n-1}\in
I_{1}\end{subarray}}\mu_{n-1}^{q-1-r}\left[g^{0}_{n,\mu},1\right]=-\left[g^{0}_{n-2,[\mu]_{n-2}}\left(\begin{array}[]{cc}1&[\mu_{n-2}]\\\
0&1\end{array}\right)w,1\right],$
2. (2)
$\sum\limits_{\begin{subarray}{c}\mu_{n-1}\in
I_{1}\end{subarray}}\mu_{n-1}^{r}\left[g^{0}_{n-1,[\mu]_{n-1}}\left(\begin{array}[]{cc}1&[\mu_{n-1}]\\\
0&1\end{array}\right)w,1\right]=(-1)^{r-1}\left[g^{0}_{n-1,[\mu]_{n-1}},1\right].$
###### Remark 7.
In fact, (1) is true modulo ${\rm Ker}~{}T_{1,2}$ and (2) is true modulo ${\rm
Ker}~{}T_{-1,0}$ (cf. [AB13, (4) & (5) on p. 62]).
## 4\. Proof of Theorem 1.2
In this section we take up the proof of Theorem 1.2. As mentioned in Section
1, several of the ideas of the proof here are already there in [Hen19].
### 4.1. A set of $I(1)$-invariants
First we make the following observation [Hen19, §2.1]. For
$a,b,c\in\mathcal{O}$, any matrix in $I(1)$ can be written as
$\left(\begin{array}[]{cc}1+\varpi a&b\\\ \varpi c&1+\varpi
d\end{array}\right)=\left(\begin{array}[]{cc}1&(1+\varpi d)^{-1}b\\\
0&1\end{array}\right)\left(\begin{array}[]{cc}1&0\\\ \varpi
ct^{-1}&1\end{array}\right)\left(\begin{array}[]{cc}t&0\\\ 0&1+\varpi
d\end{array}\right),$
where $t=1+\varpi(a-bc(1+\varpi d)^{-1}).$ Hence to prove that a certain
vector is $I(1)$-invariant modulo $({\rm Ker}~{}T_{-1,0},{\rm
Ker}~{}T_{1,2})$, it is enough to check for invariance under
$\left(\begin{array}[]{cc}1&b\\\
0&1\end{array}\right),\left(\begin{array}[]{cc}1&0\\\ \varpi
c&1\end{array}\right),\left(\begin{array}[]{cc}1+\varpi a&0\\\
0&1\end{array}\right),$
where $a,b,c\in\mathcal{O}.$
We first prove that the set of vectors $\mathcal{S}_{2}$ and $\mathcal{T}_{2}$
are $I(1)$-invariants when considered as vectors in $\pi_{r}$; i.e., when we
consider the images of these vectors modulo ${\rm Ker}~{}T_{-1,0}\oplus{\rm
Ker}~{}T_{1,2}$. The first step in achieving this is an inductive argument
which reduces the general case to the case $n=2$.
###### Lemma 4.1.
If $s_{n-1}^{k}$ (resp. $t_{n-1}^{s}$) is $I(1)$-invariant modulo $({\rm
Ker}~{}T_{-1,0},{\rm Ker}~{}T_{1,2})$, then, for all $n\geq 2$, the vector
$s_{n}^{k}$ (resp. $t_{n}^{s}$) is also $I(1)$-invariant modulo $({\rm
Ker}~{}T_{-1,0},{\rm Ker}~{}T_{1,2})$.
###### Proof.
We prove the case of $s_{n}^{k}$ and the case of $t_{n}^{s}$ is similar.
Assume that $s_{n-1}^{k}$ is $I(1)$-invariant modulo $({\rm
Ker}~{}T_{-1,0},{\rm Ker}~{}T_{1,2}).$
Now,
$\displaystyle\left(\begin{array}[]{cc}1&b\\\ 0&1\end{array}\right)s_{n}^{k}$
$\displaystyle=\sum\limits_{\begin{subarray}{c}\mu\in
I_{n}\end{subarray}}\mu_{n-1}^{k}\left[\left(\begin{array}[]{cc}1&b\\\
0&1\end{array}\right)\left(\begin{array}[]{cc}\varpi^{n}&\mu\\\
0&1\end{array}\right),1\right]$
$\displaystyle=\sum\limits_{\begin{subarray}{c}\mu\in
I_{n}\end{subarray}}\mu_{n-1}^{k}\left[\left(\begin{array}[]{cc}1&b\\\
0&1\end{array}\right)\left(\begin{array}[]{cc}\varpi&[\mu_{0}]\\\
0&1\end{array}\right)\left(\begin{array}[]{cc}\varpi^{n-1}&\displaystyle{\sum_{i=1}^{n-1}}[\mu_{i}]\varpi^{i-1}\\\
0&1\end{array}\right),1\right]$
$\displaystyle=\sum\limits_{\begin{subarray}{c}\mu\in
I_{n}\end{subarray}}\mu_{n-1}^{k}\left[\left(\begin{array}[]{cc}\varpi&[\mu_{0}+b_{0}]\\\
0&1\end{array}\right)\left(\begin{array}[]{cc}1&B(\mu_{0},b)\\\
0&1\end{array}\right)\left(\begin{array}[]{cc}\varpi^{n-1}&\displaystyle{\sum_{i=1}^{n-1}}[\mu_{i}]\varpi^{i-1},\\\
0&1\end{array}\right),1\right]$
where
(7)
$B(\mu_{0},b)=\varpi^{e-1}[P_{0}(\mu_{0},b_{0})]+[b_{1}]+[b_{2}]\varpi+\dots$
and
(8)
$P_{0}(\mu_{0},b_{0})=\frac{\mu_{0}^{q^{e}}+b_{0}^{q^{e}}-(\mu_{0}+b_{0})^{q^{e}}}{\varpi^{e}}$
is obtained from the formula in Lemma 2.3. Let
$\mu^{\prime}=[\mu_{1}]+[\mu_{2}]\varpi+\dots+[\mu_{n-1}]\varpi^{n-2}.$
We continue by making the substitution $\mu_{0}\rightarrow\mu_{0}-b_{0}$.
Thus,
$\displaystyle\left(\begin{array}[]{cc}1&b\\\ 0&1\end{array}\right)s_{n}^{k}$
$\displaystyle=\sum\limits_{\begin{subarray}{c}\mu_{0}\in
I_{1}\end{subarray}}\left(\begin{array}[]{cc}\varpi&[\mu_{0}]\\\
0&1\end{array}\right)\left\\{\left(\begin{array}[]{cc}1&B(\mu_{0}-b_{0},b)\\\
0&1\end{array}\right)\sum\limits_{\begin{subarray}{c}\mu^{\prime}\in
I_{n-1}\end{subarray}}{\mu^{\prime}}_{n-2}^{k}\left[\left(\begin{array}[]{cc}\varpi^{n-1}&\mu^{\prime}\\\
0&1\end{array}\right),1\right]\right\\}$
$\displaystyle=\sum\limits_{\begin{subarray}{c}\mu_{0}\in
I_{1}\end{subarray}}\left(\begin{array}[]{cc}\varpi&[\mu_{0}]\\\
0&1\end{array}\right)\left\\{\sum\limits_{\begin{subarray}{c}\mu^{\prime}\in
I_{n-1}\end{subarray}}{\mu^{\prime}}_{n-2}^{k}\left[\left(\begin{array}[]{cc}\varpi^{n-1}&\mu^{\prime}\\\
0&1\end{array}\right),1\right]+x_{\mu_{0}}\right\\},$
by our assumption, where $x_{\mu_{0}}\in({\rm Ker}~{}T_{-1,0},{\rm
Ker}~{}T_{1,2})$. Thus, we get
$\displaystyle\left(\begin{array}[]{cc}1&b\\\ 0&1\end{array}\right)s_{n}^{k}$
$\displaystyle=s_{n}^{k}+\sum\limits_{\begin{subarray}{c}\mu_{0}\in
I_{1}\end{subarray}}\left(\begin{array}[]{cc}\varpi&[\mu_{0}]\\\
0&1\end{array}\right)x_{\mu_{0}},$
and hence
$\left(\begin{array}[]{cc}1&b\\\
0&1\end{array}\right)s_{n}^{k}-s_{n}^{k}\in({\rm Ker}~{}T_{-1,0},{\rm
Ker}~{}T_{1,2}).$
Checking for invariance under
$\left(\begin{array}[]{cc}1&0\\\ \varpi
c&1\end{array}\right)~{}\&~{}\left(\begin{array}[]{cc}1+\varpi a&0\\\
0&1\end{array}\right)$
is even easier which we skip. ∎
Now we take the case $n=2$. Recall that $e$ (resp. $f$) is the ramification
index (resp. residue degree) of $F$ over $\mathbb{Q}_{p}$. We write
$r=r_{0}+r_{1}p+\dots+r_{f-1}p^{f-1}$
where $0\leq r_{j}\leq p-1$ for $0\leq j\leq f-1$.
We first observe that for $a,b,c\in\mathcal{O}$, we have
(9) $\left(\begin{array}[]{cc}1+\varpi a&b\\\
c\varpi&1+d\varpi\end{array}\right)\left(\begin{array}[]{cc}\varpi&[\mu]\\\
0&1\end{array}\right)=\left(\begin{array}[]{cc}\varpi&[\mu+b_{0}]\\\
0&1\end{array}\right)k,$
for $k\in I(1)$. Indeed,
$\displaystyle{\rm LHS}$
$\displaystyle=\left(\begin{array}[]{cc}\varpi(1+\varpi
a)&[\mu+b_{0}]+\varpi(*)\\\ c\varpi^{2}&1+\varpi(\Delta)\end{array}\right)$
$\displaystyle=\left(\begin{array}[]{cc}\varpi&[\mu+b_{0}]\\\
0&1\end{array}\right)\left(\begin{array}[]{cc}1+a\varpi-[\mu+b_{0}]c\varpi&(*)-(\mu+b_{0})\Delta\\\
c\varpi^{2}&1+\varpi\Delta\end{array}\right),$
where $*,\Delta\in\mathcal{O}$. Similarly, one can show that
(10) $\left(\begin{array}[]{cc}1+\varpi a&b\\\
c\varpi&1+d\varpi\end{array}\right)\left(\begin{array}[]{cc}1&[\mu]\\\
0&1\end{array}\right)w=\left(\begin{array}[]{cc}1&[\mu+b_{0}]\\\
0&1\end{array}\right)wk^{\prime}$
for some $k^{\prime}\in I(1)$.
###### Lemma 4.2.
Assume $0<r_{j}<p-1$, and if $f=1$, assume further that $2<r<p-3$. Then when
$(e,f)\neq(1,1)$, we have
$gs_{2}^{q-1-r+p^{l}}-s_{2}^{q-1-r+p^{l}}\in({\rm Ker}~{}T_{-1,0},{\rm
Ker}~{}T_{1,2})$
and
$gt_{2}^{r+p^{l}}-t_{2}^{r+p^{l}}\in({\rm Ker}~{}T_{-1,0},{\rm
Ker}~{}T_{1,2})$
for all $g\in I(1)$ and $0\leq l\leq f-1.$
###### Proof.
We have
$\displaystyle\left(\begin{array}[]{cc}1&b\\\
0&1\end{array}\right)s_{2}^{q-1-r+p^{l}}$
$\displaystyle=\sum\limits_{\begin{subarray}{c}\mu\in
I_{2}\end{subarray}}\mu_{1}^{q-1-r+p^{l}}\left[\left(\begin{array}[]{cc}1&b\\\
0&1\end{array}\right)\left(\begin{array}[]{cc}\varpi^{2}&[\mu_{0}]+[\mu_{1}]\varpi\\\
0&1\end{array}\right),1\right]$
$\displaystyle=\sum\limits_{\begin{subarray}{c}\mu\in
I_{2}\end{subarray}}\mu_{1}^{q-1-r+p^{l}}\left[\left(\begin{array}[]{cc}1&b\\\
0&1\end{array}\right)\left(\begin{array}[]{cc}\varpi&[\mu_{0}]\\\
0&1\end{array}\right)\left(\begin{array}[]{cc}\varpi&[\mu_{1}]\\\
0&1\end{array}\right),1\right]$
$\displaystyle=\sum\limits_{\begin{subarray}{c}\mu\in
I_{2}\end{subarray}}\mu_{1}^{q-1-r+p^{l}}\left[\left(\begin{array}[]{cc}\varpi&[\mu_{0}+b_{0}]\\\
0&1\end{array}\right)\left(\begin{array}[]{cc}1&B(\mu_{0},b)\\\
0&1\end{array}\right)\left(\begin{array}[]{cc}\varpi&[\mu_{1}]\\\
0&1\end{array}\right),1\right],$
where $B(\mu_{0},b)$ is given by (7) in the proof of Lemma 4.1. Now write
$B(\mu_{0},b)=[b_{1}+Z]+(*)\varpi$
where $Z=0$ for $e>1$ and $Z=P_{0}(\mu_{0},b_{0})$ for $e=1$.
To continue, the above expression equals
$\sum\limits_{\begin{subarray}{c}\mu\in
I_{2}\end{subarray}}\mu_{1}^{q-1-r+p^{l}}\left[\left(\begin{array}[]{cc}\varpi&[\mu_{0}+b_{0}]\\\
0&1\end{array}\right)\left(\begin{array}[]{cc}1&[b_{1}+Z]+(*)\varpi\\\
0&1\end{array}\right)\left(\begin{array}[]{cc}\varpi&[\mu_{1}]\\\
0&1\end{array}\right),1\right]$
which equals
$\sum\limits_{\begin{subarray}{c}\mu\in
I_{2}\end{subarray}}\mu_{1}^{q-1-r+p^{l}}\left[\left(\begin{array}[]{cc}\varpi&[\mu_{0}+b_{0}]\\\
0&1\end{array}\right)\left(\begin{array}[]{cc}\varpi&[\mu_{1}+b_{1}+Z]\\\
0&1\end{array}\right)k,1\right]$
for $k\in I(1)$, by (9). We continue by making the change of variables
$\mu_{1}\rightarrow\mu_{1}-b_{1}-Z~{}\&~{}\mu_{0}\rightarrow\mu_{0}-b_{0},$
and we get
$\displaystyle\sum\limits_{\begin{subarray}{c}\mu\in
I_{2}\end{subarray}}(\mu_{1}-b_{1}-Z)^{q-1-r+p^{l}}\left[\left(\begin{array}[]{cc}\varpi^{2}&\mu\\\
0&1\end{array}\right),1\right]$
$\displaystyle=s_{2}^{q-1-r+p^{l}}+\sum\limits_{\begin{subarray}{c}\mu\in
I_{2}\end{subarray}}\sum\limits_{\begin{subarray}{c}i=0\end{subarray}}^{q-1-r+p^{l}-1}{q-1-r+p^{l}\choose
i}(-b_{1}-Z)^{q-1-r+p^{l}-i}\mu_{1}^{i}\left[\left(\begin{array}[]{cc}\varpi^{2}&\mu\\\
0&1\end{array}\right),1\right].$
Now we read the above expression modulo ${\rm Ker}~{}T_{1,2}$. We claim that
only the term corresponding to $i=q-1-r$ remains amongst the $q-1-r+p^{l}$
terms in the inner summation in the above expression. By Lemma 3.6 (1), we
know that
$\sum_{\mu\in
I_{2}}\mu_{1}^{i}\left[\left(\begin{array}[]{cc}\varpi^{2}&\mu\\\
0&1\end{array}\right),1\right]\in{\rm Ker}~{}T_{1,2}$
precisely when $0\leq i\leq q-2-r$ or $i>q-1-r$ such that $i_{j}<p-1-r_{j}$
for some $0\leq j\leq f-2$. Note that if $i>q-1-r$ and $i_{j}\geq p-1-r_{j}$
for all $0\leq j\leq f-1$ (cf. Remark 5) then $i_{j}>p-1-r_{j}$ for some
$0\leq j\leq l-1$ (since $i\leq q-1-r+p^{l}-1$). If this is the case then
observe that
${q-1-r+p^{l}\choose i}\equiv 0\mod p$
by Corollary 2.2. Thus, modulo ${\rm Ker}~{}T_{1,2}$, we get
$\displaystyle\left(\begin{array}[]{cc}1&b\\\
0&1\end{array}\right)s_{2}^{q-1-r+p^{l}}$
$\displaystyle=s_{2}^{q-1-r+p^{l}}+\sum\limits_{\begin{subarray}{c}\mu\in
I_{2}\end{subarray}}{q-1-r+p^{l}\choose
q-1-r}(-b_{1}-Z)^{p^{l}}\mu_{1}^{q-1-r}\left[\left(\begin{array}[]{cc}\varpi^{2}&\mu\\\
0&1\end{array}\right),1\right]$
$\displaystyle=s_{2}^{q-1-r+p^{l}}+\sum\limits_{\begin{subarray}{c}\mu_{0}\in
I_{1}\end{subarray}}(p-r_{l})(-b_{1}-Z)^{p^{l}}\left[\left(\begin{array}[]{cc}1&[\mu_{0}]\\\
0&1\end{array}\right)w,1\right]$
by Lemma 3.6 (1) and the binomial coefficient here is computed via Theorem
2.1.
Now if $e>1$ then we have $Z=0$. Therefore, it follows, by Lemma 3.5 (2), that
$\left(\begin{array}[]{cc}1&b\\\
0&1\end{array}\right)s_{2}^{q-1-r+p^{l}}-s_{2}^{q-1-r+p^{l}}\in{\rm
Ker}~{}T_{-1,0},$
and thus we have proved
$\left(\begin{array}[]{cc}1&b\\\
0&1\end{array}\right)s_{2}^{q-1-r+p^{l}}\equiv s_{2}^{q-1-r+p^{l}}\mod({\rm
Ker}~{}T_{-1,0},{\rm Ker}~{}T_{1,2}).$
If $e=1$ then $Z=P_{0}(\mu_{0},b_{0})$. As $F$ is unramified over
$\mathbb{Q}_{p}$, we have $\varpi=p$. Now by Corollary 2.2, it follows that
$\displaystyle Z$
$\displaystyle=\frac{\mu_{0}^{q^{e}}+b_{0}^{q^{e}}-(\mu_{0}+b_{0})^{q^{e}}}{\varpi^{e}}$
$\displaystyle\equiv-\sum\limits_{\begin{subarray}{c}i=1\end{subarray}}^{p-1}\frac{1}{p}{p^{f}\choose
ip^{f-1}}b_{0}^{p^{f}-ip^{f-1}}\mu_{0}^{ip^{f-1}}\mod p.$
In this case, if further $f\neq 1$ we have, modulo ${\rm Ker}~{}T_{1,2}$,
$\displaystyle\left(\begin{array}[]{cc}1&b\\\
0&1\end{array}\right)s_{2}^{q-1-r+p^{l}}-s_{2}^{q-1-r+p^{l}}$
$\displaystyle=\sum\limits_{\begin{subarray}{c}\mu_{0}\in
I_{1}\end{subarray}}(p-r_{l})(-b_{1}-Z)^{p^{l}}\left[\left(\begin{array}[]{cc}1&[\mu_{0}]\\\
0&1\end{array}\right)w,1\right]$
$\displaystyle=\sum\limits_{\begin{subarray}{c}\mu_{0}\in
I_{1}\end{subarray}}r_{l}(b_{1}^{p^{l}}+Z^{p^{l}})\left[\left(\begin{array}[]{cc}1&[\mu_{0}]\\\
0&1\end{array}\right)w,1\right].$
Note that both
$\sum_{\mu_{0}\in I_{1}}\left[\left(\begin{array}[]{cc}1&[\mu_{0}]\\\
0&1\end{array}\right)w,1\right]~{}\&~{}\sum_{\mu_{0}\in
I_{1}}\mu_{0}^{ip^{l-1}}\left[\left(\begin{array}[]{cc}1&[\mu_{0}]\\\
0&1\end{array}\right)w,1\right]$
are in ${\rm Ker}~{}T_{-1,0}$, by Lemma 3.5 (2). Thus, once again we have
proved
$\left(\begin{array}[]{cc}1&b\\\
0&1\end{array}\right)s_{2}^{q-1-r+p^{l}}\equiv s_{2}^{q-1-r+p^{l}}\mod({\rm
Ker}~{}T_{-1,0},{\rm Ker}~{}T_{1,2}).$
Now we analyze invariance for the lower unipotent representative of $I(1)$. We
have
$\displaystyle\left(\begin{array}[]{cc}1&0\\\ \varpi
c&1\end{array}\right)s_{2}^{q-1-r+p^{l}}$
$\displaystyle=\sum\limits_{\begin{subarray}{c}\mu\in
I_{2}\end{subarray}}\mu_{1}^{q-1-r+p^{l}}\left[\left(\begin{array}[]{cc}1&0\\\
\varpi
c&1\end{array}\right)\left(\begin{array}[]{cc}\varpi^{2}&[\mu_{0}]+[\mu_{1}]\varpi\\\
0&1\end{array}\right),1\right]$
$\displaystyle=\sum\limits_{\begin{subarray}{c}\mu\in
I_{2}\end{subarray}}\mu_{1}^{q-1-r+p^{l}}\left[\left(\begin{array}[]{cc}1&0\\\
\varpi c&1\end{array}\right)\left(\begin{array}[]{cc}\varpi&[\mu_{0}]\\\
0&1\end{array}\right)\left(\begin{array}[]{cc}\varpi&[\mu_{1}]\\\
0&1\end{array}\right),1\right]$
which we express as
$\sum\limits_{\begin{subarray}{c}\mu\in
I_{2}\end{subarray}}\mu_{1}^{q-1-r+p^{l}}\left[\left(\begin{array}[]{cc}\varpi&[\mu_{0}]\\\
0&1\end{array}\right)\left(\begin{array}[]{cc}1-\varpi
c[\mu_{0}]&-[\mu_{0}^{2}]c\\\ \varpi^{2}c&1+\varpi
c[\mu_{0}]\end{array}\right)\left(\begin{array}[]{cc}\varpi&[\mu_{1}]\\\
0&1\end{array}\right),1\right]$
and this equals
$\sum\limits_{\begin{subarray}{c}\mu\in
I_{2}\end{subarray}}\mu_{1}^{q-1-r+p^{l}}\left[\left(\begin{array}[]{cc}\varpi&[\mu_{0}]\\\
0&1\end{array}\right)\left(\begin{array}[]{cc}\varpi&[\mu_{1}-c_{0}\mu_{0}^{2}]\\\
0&1\end{array}\right)k,1\right]$
for $k\in I(1)$ by (9). Changing $\mu_{1}\rightarrow\mu_{1}+c_{0}\mu_{0}^{2}$,
we get
$\displaystyle\left(\begin{array}[]{cc}1&0\\\ \varpi
c&1\end{array}\right)s_{2}^{q-1-r+p^{l}}$
$\displaystyle=\sum\limits_{\begin{subarray}{c}\mu\in
I_{2}\end{subarray}}(\mu_{1}+c_{0}\mu_{0}^{2})^{q-1-r+p^{l}}\left[\left(\begin{array}[]{cc}\varpi^{2}&\mu\\\
0&1\end{array}\right),1\right]$
which we read modulo ${\rm Ker}~{}T_{1,2}$ and get
$\displaystyle s_{2}^{q-1-r+p^{l}}+\sum\limits_{\begin{subarray}{c}\mu\in
I_{2}\end{subarray}}{q-1-r+p^{l}\choose
q-1-r}(c_{0}\mu_{0}^{2})^{p^{l}}\mu_{1}^{q-1-r}\left[\left(\begin{array}[]{cc}\varpi^{2}&\mu\\\
0&1\end{array}\right),1\right]$
by Corollary 2.2 together with Lemma 3.5 (1), exactly as we have argued
before. Now this equals, modulo ${\rm Ker}~{}T_{1,2}$,
$\displaystyle s_{2}^{q-1-r+p^{l}}+\sum\limits_{\begin{subarray}{c}\mu_{0}\in
I_{1}\end{subarray}}(p-r_{l})c_{0}^{p^{l}}\mu_{0}^{2p^{l}}\left[\left(\begin{array}[]{cc}1&[\mu_{0}]\\\
0&1\end{array}\right)w,1\right]$
by Theorem 2.1 and Lemma 3.6 (1). By Lemma 3.5 (2), this vector belongs to
${\rm Ker}~{}T_{-1,0}$ (with the extra assumption that $3\leq r$ when $f=1$).
Thus, we have proved
$\left(\begin{array}[]{cc}1&0\\\ \varpi
c&1\end{array}\right)s_{2}^{q-1-r+p^{l}}\equiv s_{2}^{q-1-r+p^{l}}\mod({\rm
Ker}~{}T_{-1,0},{\rm Ker}~{}T_{1,2}).$
The proof for showing that
$\left(\begin{array}[]{cc}1+\varpi a&0\\\
0&1\end{array}\right)s_{2}^{q-1-r+p^{l}}-s_{2}^{q-1-r+p^{l}}\in({\rm
Ker}~{}T_{-1,0},{\rm Ker}~{}T_{1,2})$
is similar and therefore we skip it.
The argument for
$\displaystyle gt_{2}^{r+p^{l}}-t_{2}^{r+p^{l}}\in({\rm Ker}~{}T_{-1,0},{\rm
Ker}~{}T_{1,2})$
for all $g\in I(1)$ is similar to the one for $s_{2}^{q-1-r+p^{l}}$. Note that
corresponding to the case $3\leq r$ in the totally ramified case for
$s_{2}^{q-1-r+p^{l}}$, in the case of $t_{2}^{r+p^{l}}$ we will get $r\leq
p-4$. ∎
### 4.2. Linear independence
The following lemma gives the action of the Iwahori subgroup $I$ on the
$I(1)$-invariant vectors (cf. [Hen19, Lemma 3.6]).
###### Lemma 4.3.
Let $\left(\begin{array}[]{cc}a&b\\\ c&d\end{array}\right)\in I.$ Let
$s_{n}^{k}$ and $t_{n}^{s}$ be $I(1)$-invariants modulo $({\rm
Ker}~{}T_{-1,0},{\rm Ker}~{}T_{1,2}).$ Then they are $I$-eigenvectors and
those actions are given by
* (1)
$\left(\begin{array}[]{cc}a&b\\\ c&d\end{array}\right)\cdot
s_{n}^{k}=d^{r}(da^{-1})^{k}s_{n}^{k},$
* (2)
$\left(\begin{array}[]{cc}a&b\\\ c&d\end{array}\right)\cdot
t_{n}^{s}=a^{r}(da^{-1})^{s}t_{n}^{s}.$
###### Proof.
The proof is straightforward and we have already done it in the proof of the
second part of Corollary 3.2. ∎
###### Remark 8.
Lemmas 4.1, 4.2 and 4.3 remain true for $\beta s_{n}^{k}$ and $\beta
t_{n}^{s}$.
###### Proposition 4.4.
The set of vectors in $\mathcal{S}_{2}\cup\mathcal{T}_{2}$ of Theorem 1.2 are
linearly independent.
###### Proof.
Note that the vectors in $\mathcal{S}_{2}\cup\mathcal{T}_{2}$ consist of
vectors of the form
$s_{n}^{q-1-r+p^{l}},\beta s_{n}^{q-1-r+p^{l}},t_{n}^{r+p^{l}},\beta
t_{n}^{r+p^{l}}$
for $n\geq 2$ and $0\leq l\leq f-1$. These are invariant under $I(1)$ modulo
$({\rm Ker}~{}T_{-1,0},{\rm Ker}~{}T_{1,2})$ except for the case when both
$e=1$ and $f=1$ (cf. Lemmas 4.1, 4.2 and Remark 8).
For any vector $v\in{\rm ind}_{IZ}^{G}\chi_{r}$, note that $v$ and $\beta v$
cannot cancel each other (pictorially they are on two different sides of the
tree of ${\rm SL}_{2}(F)$). Therefore, it is enough to show that the set
$\\{s_{n}^{q-1-r+p^{l}},t_{n}^{r+p^{l}}\\}$, for $n\geq 2$ and $0\leq l\leq
f-1$, is linearly independent. Since $s_{n}^{q-1-r+p^{l}}$ and
$t_{n}^{r+p^{l}}$ have different $I$-eigenvalues, it is enough to show that
$\\{s_{n}^{q-1-r+p^{l}}\\}$ and $\\{t_{n}^{r+p^{l}}\\}$, for $n\geq 2$ and
$0\leq l\leq f-1$, are linearly independent.
We show that the vectors in
$\\{s_{n}^{q-1-r+p^{l}}\\}_{n\geq 2,0\leq l\leq f-1}$
are linearly independent, and the proof for $\\{t_{n}^{r+p^{l}}\\}$ is
similar. Suppose that
$\sum\limits_{\begin{subarray}{c}i=2\end{subarray}}^{n}c_{i}s_{i}^{q-1-r+p^{l}}\in({\rm
Ker}~{}T_{-1,0},{\rm Ker}~{}T_{1,2})$
where $c_{i}\in\overline{{\mathbb{F}}}_{p}$ and $n\in\mathbb{N}$. Since no
reduction is possible in the above expression and also these vectors obviously
cannot be in ${\rm Ker}~{}T_{-1,0}$, it follows that
$\sum\limits_{\begin{subarray}{c}i=2\end{subarray}}^{n}c_{i}s_{i}^{q-1-r+p^{l}}\in{\rm
Ker}~{}T_{1,2}.$
For $i\neq j$ with $2\leq i,j\leq n$, once again from the formula for
$T_{1,2}$, there cannot be any cancellation between
$T_{1,2}(c_{i}s_{i}^{q-1-r+p^{l}})$ and $T_{1,2}(c_{j}s_{j}^{q-1-r+p^{l}}),$
so we get
$c_{i}s_{i}^{q-1-r+p^{l}}\in{\rm Ker}~{}T_{1,2}$
for all $2\leq i\leq n.$ By Lemma 3.5 (1), it follows that $c_{i}=0$ for all
$2\leq i\leq n$. ∎
###### Remark 9.
It follows by eigenvalue considerations as in the proof of Proposition 4.4
that the set
$\mathcal{S}_{2}\cup\\{[{\rm Id},1],[\beta,1]\\}\cup\mathcal{T}_{2}$
is linearly independent.
### 4.3. Auxiliary lemmas
We will have to make use of the following elementary lemma [Hen19, Lemma 2.8].
###### Lemma 4.5.
Let $n\geq 1$ and $\phi:I_{n}\rightarrow\overline{\mathbb{F}}_{p}$ be any set
map. Then there exists a unique polynomial
$Q(x_{0},\dots,x_{n-1})\in\overline{\mathbb{F}}_{p}[x_{0},x_{1},\dots,x_{n-1}]$
in which degree of each variable is at most $q-1$ and
$\phi(\mu)=Q(\mu_{0},\mu_{1},\dots,\mu_{n-1})$ for all $\mu\in I_{n}.$
The next two lemmas are the first steps towards the proof of Theorem 1.2.
###### Lemma 4.6.
Let $\mu=[\mu_{0}]+[\mu_{1}]\varpi+\dots+[\mu_{n-1}]\varpi^{n-1}\in I_{n}$ and
$r=r_{0}+r_{1}p+\dots+r_{f-1}p^{f-1}$ with $0<r_{j}<p-1$ for all $0\leq j\leq
f-1.$ Let
$f_{n}=f_{n}^{\prime}+f_{n}^{\prime\prime}$
be such that
$f_{n}^{\prime}=\sum\limits_{\begin{subarray}{c}\mu\in
I_{n}\end{subarray}}a(\mu_{0},\mu_{1},\dots,\mu_{n-1})\left[\left(\begin{array}[]{cc}\varpi^{n}&\mu\\\
0&1\end{array}\right),1\right]$
and
$f_{n}^{\prime\prime}=\sum\limits_{\begin{subarray}{c}\mu\in
I_{n}\end{subarray}}b(\mu_{0},\mu_{1},\dots,\mu_{n-1})\left[\left(\begin{array}[]{cc}\varpi^{n-1}&[\mu]_{n-1}\\\
0&1\end{array}\right)\left(\begin{array}[]{cc}1&[\mu_{n-1}]\\\
0&1\end{array}\right)w,1\right],$
where $a(\mu_{0},\dots,\mu_{n-1})$ and $b(\mu_{0},\dots,\mu_{n-1})$ are
polynomials in $\mu_{0},\dots,\mu_{n-1}.$ Suppose
$\left(\begin{array}[]{cc}1&-\varpi^{n-1}\\\
0&1\end{array}\right)f_{n}-f_{n}\in({\rm Ker}~{}T_{-1,0},{\rm
Ker}~{}T_{1,2}).$
Then
1. (1)
the possible powers of $\mu_{n-1},$ say $k=k_{0}+k_{1}p+\dots+k_{f-1}p^{f-1}$,
in $a(\mu_{0},\dots,\mu_{n-1})$ will satisfy one of the following three
conditions:
1. (a)
there exists some $0\leq j^{\prime}\leq f-1$ such that
$k_{j^{\prime}}<p-1-r_{j^{\prime}},$
2. (b)
$k_{j}=p-1-r_{j}$ for all $0\leq j\leq f-1,$
3. (c)
$k_{j}=p-1-r_{j}$ for $j\neq l$ and $k_{l}=p-r_{l}$ for some $0\leq l\leq
f-1.$
2. (2)
the possible powers of $\mu_{n-1},$ say $k=k_{0}+k_{1}p+\dots+k_{f-1}p^{f-1}$,
in $b(\mu_{0},\dots,\mu_{n-1})$ will satisfy one of the following three
conditions:
1. (a)
there exists some $0\leq j^{\prime}\leq f-1$ such that such that
$k_{j^{\prime}}<r_{j^{\prime}},$
2. (b)
$k_{j}=r_{j}$ for all $0\leq j\leq f-1,$
3. (c)
$k_{j}=r_{j}$ for $j\neq l$ and $k_{l}=r_{l}+1$ for some $0\leq l\leq f-1.$
###### Proof of Lemma 4.6.
We will prove $(1)$ and the proof of $(2)$ is similar. Suppose $(1)$ does not
hold. Then there exists $k$ such that $k_{j}\geq p-1-r_{j}$ for all $0\leq
j\leq f-1$ with
$k_{j_{0}}>p-1-r_{j_{0}}\mbox{~{}for~{}some~{}}0\leq j_{0}\leq
f-1~{}\&~{}k\neq(p-r_{j_{0}})p^{j_{0}}+\sum\limits_{\begin{subarray}{c}j_{0}\neq
j=0\end{subarray}}^{f-1}(p-1-r_{j})p^{j}.$
Then either there exists $j_{1}$ with $j_{1}\neq j_{0}$ such that
$k_{j_{1}}>p-1-r_{j_{1}}$ or
$k=k_{j_{0}}p^{j_{0}}+\sum\limits_{\begin{subarray}{c}j_{0}\neq
j=0\end{subarray}}^{f-1}(p-1-r_{j})p^{j}$
with $k_{j_{0}}>p-r_{j_{0}}.$ Choose $k$ with the above property such that
there is no other monomial $\mu_{n-1}^{k^{\prime}}$ in
$a(\mu_{0},\dots,\mu_{n-1})$ with $k_{j}\leq k^{\prime}_{j}$ for all $0\leq
j\leq f-1$. Since a polynomial is of finite degree, such a $k$ exists. Let
$g=\left(\begin{array}[]{cc}1&-\varpi^{n-1}\\\ 0&1\end{array}\right).$
We have
$gf_{n}-f_{n}=(gf_{n}^{\prime}-f_{n}^{\prime})+(gf_{n}^{\prime\prime}-f_{n}^{\prime\prime})\in({\rm
Ker}~{}T_{-1,0},{\rm Ker}~{}T_{1,2}).$
Note that,
$\displaystyle gf_{n}^{\prime}-f_{n}^{\prime}$
$\displaystyle=\sum\limits_{\begin{subarray}{c}\mu\in
I_{n}\end{subarray}}\left[a([\mu]_{n-1},\mu_{n-1}+1)-a([\mu]_{n-1},\mu_{n-1})\right]\left[\left(\begin{array}[]{cc}\varpi^{n}&\mu\\\
0&1\end{array}\right),1\right]$
and
$\displaystyle gf_{n}^{\prime\prime}-f_{n}^{\prime\prime}$
$\displaystyle=\sum\limits_{\begin{subarray}{c}\mu\in
I_{n}\end{subarray}}\left[b([\mu]_{n-1},\mu_{n-1}+1)-b([\mu]_{n-1},\mu_{n-1})\right]\left[\left(\begin{array}[]{cc}\varpi^{n-1}&[\mu]_{n-1}\\\
0&1\end{array}\right)\left(\begin{array}[]{cc}1&[\mu_{n-1}]\\\
0&1\end{array}\right)w,1\right].$
Let
$\Delta a=a([\mu]_{n-1},\mu_{n-1}+1)-a([\mu]_{n-1},\mu_{n-1})$
considered as a polynomial in $\mu_{n-1}$ with coefficients in
$\overline{\mathbb{F}}_{p}[\mu_{0},\dots,\mu_{n-2}]$. By Theorem 2.1, we have
$(\mu_{n-1}+1)^{k}-\mu_{n-1}^{k}\equiv\sum\limits_{\begin{subarray}{c}i=0\end{subarray}}^{k-1}\prod_{j=0}^{f-1}{k_{j}\choose
i_{j}}\mu_{n-1}^{i}\mod p.$
Now if there exists $j_{1}$ with $j_{1}\neq j_{0}$ such that
$k_{j_{1}}>p-1-r_{j_{1}},$ take
$k^{\prime}=(k_{j_{1}}-1)p^{j_{1}}+\sum\limits_{\begin{subarray}{c}j_{1}\neq
j=0\end{subarray}}^{f-1}k_{j}p^{j}.$
The coefficient of $\mu_{n-1}^{k^{\prime}}$ in $\Delta a$ is
${k\choose k^{\prime}}={k_{j_{1}}\choose k_{j_{1}}-1}\not\equiv 0\mod p$
by Theorem 2.1 and Corollary 2.2. Note that the term involving
$\mu_{n-1}^{k^{\prime}}$ in $gf_{n}^{\prime}-f_{n}^{\prime}$ cannot get
cancelled by any other term in $gf_{n}-f_{n}$. Indeed, it cannot get cancelled
with any other term in $gf_{n}^{\prime}-f_{n}^{\prime}$ because of the choice
of $k$ and anyway no term in $gf_{n}^{\prime}-f_{n}^{\prime}$ can get
cancelled with a term in $gf_{n}^{\prime\prime}-f_{n}^{\prime\prime}$
(pictorially they represent edges of opposite orientation on the tree of ${\rm
SL}_{2}(F)$). So this term involving $\mu_{n-1}^{k^{\prime}}$ must be there in
$({\rm Ker}~{}T_{-1,0},{\rm Ker}~{}T_{1,2}),$ but then Lemma 3.5 (1) would
imply that there exists some $0\leq l\leq f-1$ such that
$k^{\prime}_{l}<p-1-r_{l},$ which contradicts our assumption. So $k$ must be
of the form
$k=k_{j_{0}}p^{j_{0}}+\sum\limits_{\begin{subarray}{c}j_{0}\neq
j=0\end{subarray}}^{f-1}(p-1-r_{j})p^{j}$
with $k_{j_{0}}>p-r_{j_{0}}.$ Taking
$k^{\prime}=(k_{j_{0}}-1)p^{j_{0}}+\sum\limits_{\begin{subarray}{c}j_{0}\neq
j=0\end{subarray}}^{f-1}(p-1-r_{j})p^{j},$
and using the same argument as in the previous case, we arrive at a
contradiction. ∎
###### Remark 10.
The idea of choosing $k$ as in Lemma 4.6 is already employed by Hendel in
[Hen19, Lemma 3.13].
Now we state one more lemma whose main idea of proof also comes from [Hen19,
Lemma 3.13]. In what follows, $B(t)$ denotes the ball of radius $m$ on the
tree of ${\rm SL}_{2}(F)$ with center at the vertex representing the trivial
coset $G/KZ$. Explicitly it consists of linear combinations of vectors of the
form
$B^{0}(t)=\left\\{[g_{n,\mu}^{0},1],\left[g_{n-1,[\mu]_{n-1}}^{0}\left(\begin{array}[]{cc}1&[\mu_{n-1}]\\\
0&1\end{array}\right)w,1\right]\right\\}_{n\leq t},$
and
$B^{1}(t)=\left\\{[g_{n-1,[\mu]_{n-1}}^{1}w,1],\left[g_{n-2,[\mu]_{n-2}}^{1}w\left(\begin{array}[]{cc}1&[\mu_{n-2}]\\\
0&1\end{array}\right)w,1\right]\right\\}_{n\leq t},$
where $\mu=[\mu_{0}]+[\mu_{1}]\varpi+\dots+[\mu_{n-1}]\varpi^{n-1}\in I_{n}$.
###### Lemma 4.7.
Let
$f_{n}^{\prime}=\sum\limits_{\begin{subarray}{c}\mu\in
I_{n}\end{subarray}}\sum\limits_{\begin{subarray}{c}l=0\end{subarray}}^{f-1}P_{l}([\mu]_{n-1})\mu_{n-1}^{q-1-r+p^{l}}\left[g^{0}_{n,\mu},1\right]$
and
$f_{n}^{\prime\prime}=\sum\limits_{\begin{subarray}{c}\mu\in
I_{n}\end{subarray}}\sum\limits_{\begin{subarray}{c}l=0\end{subarray}}^{f-1}Q_{l}([\mu]_{n-1})\mu_{n-1}^{r+p^{l}}\left[g^{0}_{n-1,[\mu]_{n-1}}\left(\begin{array}[]{cc}1&[\mu_{n-1}]\\\
0&1\end{array}\right)w,1\right],$
where $P_{l}([\mu]_{n-1})$ and $Q_{l}([\mu]_{n-1})$ are polynomials in
$\mu_{0},\dots,\mu_{n-2}$. Let $f_{n}=f_{n}^{\prime}+f_{n}^{\prime\prime}$.
Let $f=f_{n}+f^{\prime}$ be such that $f^{\prime}\in B(n-1)$ and
$\left(\begin{array}[]{cc}1&-\varpi^{n-m}\\\ 0&1\end{array}\right)f-f\in({\rm
Ker}~{}T_{-1,0},{\rm Ker}~{}T_{1,2}),$
for all $1\leq m\leq n-1.$ Then we have
$f_{n}^{\prime}=\sum\limits_{\begin{subarray}{c}\mu\in
I_{n}\end{subarray}}\sum\limits_{\begin{subarray}{c}l=0\end{subarray}}^{f-1}a_{l}\mu_{n-1}^{q-1-r+p^{l}}\left[g^{0}_{n,\mu},1\right]$
and
$f_{n}^{\prime\prime}=\sum\limits_{\begin{subarray}{c}\mu\in
I_{n}\end{subarray}}\sum\limits_{\begin{subarray}{c}l=0\end{subarray}}^{f-1}b_{l}\mu_{n-1}^{r+p^{l}}\left[g^{0}_{n-1,[\mu]_{n-1}}\left(\begin{array}[]{cc}1&[\mu_{n-1}]\\\
0&1\end{array}\right)w,1\right],$
where $a_{l}$ and $b_{l}$ are constants.
###### Proof of Lemma 4.7.
We do the proof only for $f_{n}^{\prime}$, as the case of
$f_{n}^{\prime\prime}$ is similar. The proof is by induction on $n$. Note that
$P_{l}([\mu]_{n-1})$ is independent of $\mu_{n-1}.$ Suppose it is independent
of $\mu_{n-1},\dots,\mu_{n-m+1}.$ Then
$f_{n}^{\prime}=\sum\limits_{\begin{subarray}{c}\mu\in
I_{n}\end{subarray}}\sum\limits_{\begin{subarray}{c}l=0\end{subarray}}^{f-1}P_{l}([\mu]_{n-m},\mu_{n-m})\mu_{n-1}^{q-1-r+p^{l}}\left[g^{0}_{n,\mu},1\right]$
We show that it is independent of $\mu_{n-m}$. It is given to us that
$\displaystyle\left(\begin{array}[]{cc}1&-\varpi^{n-m}\\\
0&1\end{array}\right)f-f$
$\displaystyle=\left[\left(\begin{array}[]{cc}1&-\varpi^{n-m}\\\
0&1\end{array}\right)f_{n}-f_{n}\right]+\left[\left(\begin{array}[]{cc}1&-\varpi^{n-m}\\\
0&1\end{array}\right)f^{\prime}-f^{\prime}\right]$ $\displaystyle\in({\rm
Ker}~{}T_{-1,0},{\rm Ker}~{}T_{1,2}).$
Now,
$\displaystyle\left(\begin{array}[]{cc}1&-\varpi^{n-m}\\\
0&1\end{array}\right)\left(\begin{array}[]{cc}\varpi^{n}&\displaystyle{\sum_{i=0}^{n-1}}[\mu_{i}]\varpi^{i}\\\
0&1\end{array}\right)$
$\displaystyle=\left(\begin{array}[]{cc}\varpi^{n}&[\mu_{0}]+\dots+[\mu_{n-1}]\varpi^{n-1}-\varpi^{n-m}\\\
0&1\end{array}\right)$
and this equals
$\displaystyle\left(\begin{array}[]{cc}\varpi^{n}&\displaystyle{\sum_{i=0}^{n-m-1}}[\mu_{i}]\varpi^{i}+[\mu_{n-m}-1]\varpi^{n-m}+[\mu^{\prime}_{n-m+1}]\varpi^{n-m+1}+\dots+[\mu^{\prime}_{n-1}]\varpi^{n-1}\\\
0&1\end{array}\right)$
where $\mu^{\prime}_{k}=\mu_{k}+c_{k}(\mu_{n-m},\dots,\mu_{n-2})$ for
$n-m+1\leq k\leq n-1.$
Note that the transformation
$\mu^{\prime}_{k}\mapsto\mu_{k}-c_{k}(\mu_{n-m},\dots,\mu_{n-2})$ does not
affect the variables $\mu_{k}$ for $n-m+1\leq k\leq n-1$ in
$P_{l}([\mu]_{n-1}),$ as it is independent of these variables. This
transformation together with $\mu_{n-m}\mapsto\mu_{n-m}+1$ gives
$\left(\begin{array}[]{cc}1&-\varpi^{n-m}\\\
0&1\end{array}\right)f_{n}^{\prime}=\sum\limits_{\begin{subarray}{c}\mu\in
I_{n}\end{subarray}}\sum\limits_{\begin{subarray}{c}l=0\end{subarray}}^{f-1}P_{l}([\mu]_{n-m},\mu_{n-m}+1)(\mu_{n-1}-c_{n-1})^{q-1-r+p^{l}}[g_{n,\mu}^{0},1].$
In the above expression, by $c_{n-1}$ we mean
$c_{n-1}(\mu_{n-m},\dots,\mu_{n-2})$. Now,
$\displaystyle\left(\begin{array}[]{cc}1&-\varpi^{n-m}\\\
0&1\end{array}\right)f_{n}^{\prime}-f_{n}^{\prime}$
$\displaystyle=\sum\limits_{\begin{subarray}{c}\mu\in
I_{n}\end{subarray}}\sum\limits_{\begin{subarray}{c}l=0\end{subarray}}^{f-1}\alpha(\mu,l)[g^{0}_{n,\mu},1],$
where
$\alpha(\mu,l)=\left[P_{l}([\mu]_{n-m},\mu_{n-m}+1)(\mu_{n-1}-c_{n-1})^{q-1-r+p^{l}}-P_{l}([\mu]_{n-m},\mu_{n-m})\mu_{n-1}^{q-1-r+p^{l}}\right].$
Thus,
$\displaystyle\left(\begin{array}[]{cc}1&-\varpi^{n-m}\\\
0&1\end{array}\right)f_{n}^{\prime}-f_{n}^{\prime}$
$\displaystyle=\sum\limits_{\begin{subarray}{c}\mu\in
I_{n}\end{subarray}}\sum\limits_{\begin{subarray}{c}l=0\end{subarray}}^{f-1}\left[P_{l}([\mu]_{n-m},\mu_{n-m}+1)-P_{l}([\mu]_{n-m},\mu_{n-m})\right]\mu_{n-1}^{q-1-r+p^{l}}[g^{0}_{n,\mu},1]$
$\displaystyle+\sum\limits_{\begin{subarray}{c}\mu\in
I_{n}\end{subarray}}\sum\limits_{\begin{subarray}{c}l=0\end{subarray}}^{f-1}\sum\limits_{\begin{subarray}{c}i=0\end{subarray}}^{q-1-r+p^{l}-1}\beta(\mu,l,i)[g^{0}_{n,\mu},1],$
where
$\beta(\mu,l,i)=P_{l}([\mu]_{n-m},\mu_{n-m}+1)(-1)^{i}{q-1-r+p^{l}\choose
i}(-c_{n-1})^{q-1-r+p^{l}-i}\mu_{n-1}^{i}.$
Now we read this modulo ${\rm Ker}~{}T_{1,2}$. Thus, we get
$\displaystyle\left(\begin{array}[]{cc}1&-\varpi^{n-m}\\\
0&1\end{array}\right)f_{n}^{\prime}-f_{n}^{\prime}$
$\displaystyle=\sum\limits_{\begin{subarray}{c}\mu\in
I_{n}\end{subarray}}\sum\limits_{\begin{subarray}{c}l=0\end{subarray}}^{f-1}\left[P_{l}([\mu]_{n-m},\mu_{n-m}+1)-P_{l}([\mu]_{n-m},\mu_{n-m})\right]\mu_{n-1}^{q-1-r+p^{l}}[g^{0}_{n,\mu},1]$
$\displaystyle+\sum\limits_{\begin{subarray}{c}\mu\in
I_{n}\end{subarray}}\sum\limits_{\begin{subarray}{c}l=0\end{subarray}}^{f-1}P_{l}([\mu]_{n-m},\mu_{n-m}+1){q-1-r+p^{l}\choose
q-1-r}(-c_{n-1})^{p^{l}}\mu_{n-1}^{q-1-r}[g^{0}_{n,\mu},1],$
by Corollary 2.2 and Lemma 3.5 (1), exactly as we have argued before in the
proof of Lemma 4.2. Now by Lemma 3.6 (1), it follows that, modulo $({\rm
Ker}~{}T_{-1,0},{\rm Ker}~{}T_{1,2})$, we have
$\displaystyle\left(\begin{array}[]{cc}1&-\varpi^{n-m}\\\
0&1\end{array}\right)f_{n}^{\prime}-f_{n}^{\prime}$
$\displaystyle=\sum\limits_{\begin{subarray}{c}\mu\in
I_{n}\end{subarray}}\sum\limits_{\begin{subarray}{c}l=0\end{subarray}}^{f-1}\left[P_{l}([\mu]_{n-m},\mu_{n-m}+1)-P_{l}([\mu]_{n-m},\mu_{n-m})\right]\mu_{n-1}^{q-1-r+p^{l}}[g^{0}_{n,\mu},1]+g_{n-1}$
where $g_{n-1}\in B(n-1)$. As $r_{l}\neq 0,$ by Lemmas 3.5 (1) and 3.6 (1) we
have
$\sum\limits_{\begin{subarray}{c}\mu\in
I_{n}\end{subarray}}\mu_{n-1}^{q-1-r+p^{l}}\left[g^{0}_{n,\mu},1\right]\notin({\rm
Ker}~{}T_{-1,0},{\rm Ker}~{}T_{1,2}).$
Also the term involving $\mu_{n-1}^{q-1-r+p^{l}}$ cannot get cancelled by any
other term in the expression
$\left(\begin{array}[]{cc}1&-\varpi^{n-m}\\\ 0&1\end{array}\right)f-f.$
So it follows that
$P_{l}([\mu]_{n-m},\mu_{n-m}+1)-P_{l}([\mu]_{n-m},\mu_{n-m})=0.$
Hence $P_{l}([\mu]_{n-1})$ is independent of $\mu_{n-m}.$ Therefore, by
induction $P_{l}([\mu]_{n-1})$ is a constant. ∎
### 4.4. Proof of Theorem 1.2
Clearly the vectors $\left[{\rm Id},1\right]$ and $\left[\beta,1\right]$ are
fixed by $I(1).$ By Lemmas 4.1 and 4.2 and Remark 8 the vectors in
$\mathcal{S}_{2}$ and $\mathcal{T}_{2}$ are $I(1)$-invariant modulo $({\rm
Ker}~{}T_{-1,0},{\rm Ker}~{}T_{1,2})$ except for the case when both $e=1$ and
$f=1.$ By Remark 9, the set $\mathcal{S}_{2}\cup\\{[{\rm
Id},1],[\beta,1]\\}\cup\mathcal{T}_{2}$ is linearly independent.
Now let $f\in{\rm ind}_{IZ}^{G}\chi_{r}$ be an $I(1)$-invariant of
$\pi_{r}=\frac{{\rm ind}_{IZ}^{G}\chi_{r}}{({\rm Ker}~{}T_{-1,0},{\rm
Ker}~{}T_{1,2})}.$
We write
$f=f^{0}+f^{1}$
where $f^{0}$ (resp. $f^{1}$) is a linear combination of vectors on the zero
side (resp. one side) of the tree of ${\rm SL}_{2}(F)$. By this, we mean
$f^{0}$ is a linear combination of vectors of the form
$[g_{n,\mu}^{0},1],\left[g_{n-1,[\mu]_{n-1}}^{0}\left(\begin{array}[]{cc}1&[\mu_{n-1}]\\\
0&1\end{array}\right)w,1\right]$
and $f^{1}$ is a linear combination of vectors of the form
$[g_{n-1,[\mu]_{n-1}}^{1}w,1],\left[g_{n-2,[\mu]_{n-2}}^{1}w\left(\begin{array}[]{cc}1&[\mu_{n-2}]\\\
0&1\end{array}\right)w,1\right].$
Then,
$gf^{i}-f^{i}\in({\rm Ker}~{}T_{-1,0},{\rm Ker}~{}T_{1,2}),$
for all $i\in\\{0,1\\}$ and $g\in I(1)$. Since $\beta f^{1}$ is a linear
combination of vectors on the zero side and $\beta$ normalizes $I(1)$, without
loss of generality, we may assume $f=f^{0}$. Write
$f=f_{n}+f^{\prime}$
with $f_{n}\neq 0$, $f^{\prime}\in B(n-1)$, for $n$ maximal. Now,
$f_{n}=\sum\limits_{\begin{subarray}{c}\mu\in
I_{n}\end{subarray}}a_{{\mu}}\left[g^{0}_{n,\mu},1\right]+\sum\limits_{\begin{subarray}{c}\mu\in
I_{n}\end{subarray}}b_{{\mu}}\left[g^{0}_{n-1,[\mu]_{n-1}}\left(\begin{array}[]{cc}1&[\mu_{n-1}]\\\
0&1\end{array}\right)w,1\right],$
where $\mu=[\mu_{0}]+[\mu_{1}]\varpi+\dots+[\mu_{n-1}]\varpi^{n-1}$ and
$a_{{\mu}},b_{{\mu}}\in\overline{\mathbb{F}}_{p}.$ By Lemma 4.5, the
coefficients $a_{{\mu}}$ and $b_{{\mu}}$ can be replaced by the polynomials
$a(\mu_{0},\dots,\mu_{n-1})$ and $b(\mu_{0},\dots,\mu_{n-1})$ respectively,
where each $\mu_{i}$ has maximum degree $q-1.$ Write
$f_{n}=f_{n}^{\prime}+f_{n}^{\prime\prime},$
where
$f_{n}^{\prime}=\sum\limits_{\begin{subarray}{c}\mu\in
I_{n}\end{subarray}}\sum\limits_{\begin{subarray}{c}i\end{subarray}}a(i_{0},i_{1},\dots,i_{n-1})\mu_{0}^{i_{0}}\dots\mu_{n-1}^{i_{n-1}}\left[g^{0}_{n,\mu},1\right],$
and
$f_{n}^{\prime\prime}=\sum\limits_{\begin{subarray}{c}\mu\in
I_{n}\end{subarray}}\sum\limits_{\begin{subarray}{c}j\end{subarray}}b(j_{0},j_{1},\dots,j_{n-1})\mu_{0}^{j_{0}}\dots\mu_{n-1}^{j_{n-1}}\left[g^{0}_{n-1,[\mu]_{n-1}}\left(\begin{array}[]{cc}1&[\mu_{n-1}]\\\
0&1\end{array}\right)w,1\right].$
Let
$g^{\prime}=\left(\begin{array}[]{cc}1&-\varpi^{n-1}\\\
0&1\end{array}\right)\in I(1).$
Since $f^{\prime}$ belongs in $B(n-1)$, it is easy to check that $g^{\prime}$
fixes $f^{\prime}$. This gives
$g^{\prime}f_{n}-f_{n}\in({\rm Ker}~{}T_{-1,0},{\rm Ker}~{}T_{1,2}).$
Now Lemma 3.5 (1) together with Lemma 4.6 (1) gives
$\displaystyle f_{n}^{\prime}$
$\displaystyle=\sum\limits_{\begin{subarray}{c}\mu\in
I_{n}\end{subarray}}\sum\limits_{\begin{subarray}{c}i\end{subarray}}a(i_{0},\dots,i_{n-2},q-1-r)\mu_{0}^{i_{0}}\dots\mu_{n-1}^{q-1-r}\left[g^{0}_{n,\mu},1\right]$
$\displaystyle+\sum\limits_{\begin{subarray}{c}\mu\in
I_{n}\end{subarray}}\sum\limits_{\begin{subarray}{c}l=0\end{subarray}}^{f-1}a_{l}([\mu]_{n-1})\mu_{n-1}^{q-1-r+p^{l}}\left[g^{0}_{n,\mu},1\right],$
which in turn implies that
$\displaystyle f_{n}^{\prime}-\sum\limits_{\begin{subarray}{c}\mu\in
I_{n}\end{subarray}}\sum\limits_{\begin{subarray}{c}l=0\end{subarray}}^{f-1}a_{l}([\mu]_{n-1})\mu_{n-1}^{q-1-r+p^{l}}\left[g^{0}_{n,\mu},1\right]$
$\displaystyle=\sum\limits_{\begin{subarray}{c}\mu_{0},\dots,\mu_{n-1}\end{subarray}}\sum\limits_{\begin{subarray}{c}i_{0},\dots,i_{n-2}\end{subarray}}a(i_{0},\dots,q-1-r)\mu_{0}^{i_{0}}\dots\mu_{n-1}^{q-1-r}\left[g^{0}_{n,\mu},1\right]$
which modulo ${\rm Ker}~{}T_{1,2}$ equals
$\displaystyle\sum\limits_{\begin{subarray}{c}\mu_{0},\dots,\mu_{n-2}\end{subarray}}\sum\limits_{\begin{subarray}{c}i_{0},\dots,i_{n-2}\end{subarray}}a(i_{0},\dots,q-1-r)\mu_{0}^{i_{0}}\dots\mu_{n-2}^{i_{n-2}}\left[g^{0}_{n-2,[\mu]_{n-2}}\left(\begin{array}[]{cc}1&[\mu_{n-2}]\\\
0&1\end{array}\right)w,1\right],$
by Lemma 3.6 (1). This vector belongs to $B(n-1)$ which we call
$g_{n-1}^{\prime}$. We get
$f_{n}^{\prime}=\sum\limits_{\begin{subarray}{c}\mu\in
I_{n}\end{subarray}}\sum\limits_{\begin{subarray}{c}l=0\end{subarray}}^{f-1}a_{l}([\mu]_{n-1})\mu_{n-1}^{q-1-r+p^{l}}\left[g^{0}_{n,\mu},1\right]+g_{n-1}^{\prime}.$
Similarly, working with $f_{n}^{\prime\prime},$ we get
$f_{n}^{\prime\prime}=\sum\limits_{\begin{subarray}{c}\mu\in
I_{n}\end{subarray}}\sum\limits_{\begin{subarray}{c}l=0\end{subarray}}^{f-1}b_{l}([\mu]_{n-1})\mu_{n-1}^{r+p^{l}}\left[g_{n-1,[\mu]_{n-1}}^{0}\left(\begin{array}[]{cc}1&[\mu_{n-1}]\\\
0&1\end{array}\right)w,1\right]+g_{n-1}^{\prime\prime}$
for some $g_{n-1}^{\prime\prime}\in B(n-1)$, by Lemmas 3.5 (2), 4.6 (2) and
3.6 (2).
For $1\leq m\leq n-1,$ we note that
$\left(\begin{array}[]{cc}1&-\varpi^{n-m}\\\ 0&1\end{array}\right)\in I(1).$
Using the condition
$\left(\begin{array}[]{cc}1&-\varpi^{n-m}\\\ 0&1\end{array}\right)f-f\in({\rm
Ker}~{}T_{-1,0},{\rm Ker}~{}T_{1,2}),$
by Lemma 4.7, we have
$f_{n}^{\prime}=\sum\limits_{\begin{subarray}{c}\mu\in
I_{n}\end{subarray}}\sum\limits_{\begin{subarray}{c}l=0\end{subarray}}^{f-1}a_{l,n}\mu_{n-1}^{q-1-r+p^{l}}\left[g^{0}_{n,\mu},1\right]+g_{n-1}^{\prime}$
and
$f_{n}^{\prime\prime}=\sum\limits_{\begin{subarray}{c}\mu\in
I_{n}\end{subarray}}\sum\limits_{\begin{subarray}{c}l=0\end{subarray}}^{f-1}b_{l,n}\mu_{n-1}^{r+p^{l}}\left[g_{n-1,[\mu]_{n-1}}^{0}\left(\begin{array}[]{cc}1&[\mu_{n-1}]\\\
0&1\end{array}\right)w,1\right]+g_{n-1}^{\prime\prime},$
where $a_{l}$ and $b_{l}$ are constants.
Hence $f_{n}$ takes the form
$f_{n}=\sum\limits_{\begin{subarray}{c}l=0\end{subarray}}^{f-1}a_{l,n}s_{n}^{q-1-r+p^{l}}+\sum\limits_{\begin{subarray}{c}l=0\end{subarray}}^{f-1}b_{l,n}t_{n}^{r+p^{l}}+g_{n-1},$
where
$g_{n-1}=g_{n-1}^{\prime}+g_{n-1}^{\prime\prime}\in B(n-1).$
Thus it follows that
$f-\sum\limits_{\begin{subarray}{c}l=0\end{subarray}}^{f-1}a_{l,n}s_{n}^{q-1-r+p^{l}}-\sum\limits_{\begin{subarray}{c}l=0\end{subarray}}^{f-1}b_{l,n}t_{n}^{r+p^{l}}=g_{n-1}+f^{\prime}$
is an $I(1)$-invariant vector modulo $({\rm Ker}~{}T_{-1,0},{\rm
Ker}~{}T_{1,2})$ in $B(n-1)$.
Applying this argument on vectors in $B(n-1)$ and repeating this process, we
get
$\displaystyle
f=\sum\limits_{\begin{subarray}{c}l=0\end{subarray}}^{f-1}a_{l,n}s_{n}^{q-1-r+p^{l}}+\sum\limits_{\begin{subarray}{c}l=0\end{subarray}}^{f-1}b_{l,n}t_{n}^{r+p^{l}}+\dots+\sum\limits_{\begin{subarray}{c}l=0\end{subarray}}^{f-1}a_{l,2}s_{2}^{q-1-r+p^{l}}+\sum\limits_{\begin{subarray}{c}l=0\end{subarray}}^{f-1}b_{l,2}t_{2}^{r+p^{l}}+f_{1},$
where $f_{1}$ is an $I(1)$-invariant in $B(1)$. Write
$f_{1}=f_{1}^{\prime}+f_{1}^{\prime\prime},$
where
$f_{1}^{\prime}=\sum\limits_{\begin{subarray}{c}\mu\in
I_{1}\end{subarray}}\sum\limits_{\begin{subarray}{c}i\end{subarray}}a_{i}\mu^{i}\left[g^{0}_{1,\mu},1\right],$
and
$f_{1}^{\prime\prime}=\sum\limits_{\begin{subarray}{c}\mu\in
I_{1}\end{subarray}}\sum\limits_{\begin{subarray}{c}j\end{subarray}}b_{j}\mu^{j}\left[\left(\begin{array}[]{cc}1&\mu\\\
0&1\end{array}\right)w,1\right].$
Using the action of
$u=\left(\begin{array}[]{cc}1&1\\\ 0&1\end{array}\right)$
on $f_{1}$, by Lemma 4.6 (1), the possible powers $i$ of $\mu$ in
$f_{1}^{\prime}$ will satisfy either $0\leq i\leq q-1-r$ or $i=q-1-r+p^{l}$
for some $0\leq l\leq f-1.$ If $i=q-1-r+p^{l},$ then
$\left(\begin{array}[]{cc}1&1\\\
0&1\end{array}\right)f_{1}^{\prime}-f_{1}^{\prime}=\sum\limits_{\begin{subarray}{c}\mu\in
I_{1}\end{subarray}}\mu^{q-1-r}\left[g^{0}_{1,\mu},1\right]\in({\rm
Ker}~{}T_{-1,0},{\rm Ker}~{}T_{1,2}).$
This, by Lemma 3.6 (1), gives $\left[\beta,1\right]\in({\rm
Ker}~{}T_{-1,0},{\rm Ker}~{}T_{1,2}),$ which is not possible. So we must have
$0\leq i\leq q-1-r.$ Then, by Lemma 3.5(1) and Lemma 3.6 (1), we have
$f_{1}^{\prime}=\left[\beta,1\right]\mod({\rm Ker}~{}T_{-1,0},{\rm
Ker}~{}T_{1,2}).$
Similarly, by Lemmas 3.5 (2) and 4.6 (2) and Lemma 3.6 (2) , we can show that
$f_{1}^{\prime\prime}=\left[{\rm Id},1\right]\mod({\rm Ker}~{}T_{-1,0},{\rm
Ker}~{}T_{1,2}).$
Thus, we have
$\displaystyle
f=\sum\limits_{\begin{subarray}{c}l=0\end{subarray}}^{f-1}a_{l,n}s_{n}^{q-1-r+p^{l}}+\sum\limits_{\begin{subarray}{c}l=0\end{subarray}}^{f-1}b_{l,n}t_{n}^{r+p^{l}}+\dots$
$\displaystyle+\sum\limits_{\begin{subarray}{c}l=0\end{subarray}}^{f-1}a_{l,2}s_{2}^{q-1-r+p^{l}}+\sum\limits_{\begin{subarray}{c}l=0\end{subarray}}^{f-1}b_{l,2}t_{2}^{r+p^{l}}+c\left[\beta,1\right]+d\left[{\rm
I}d,1\right].$
Now assume $e=1$ and $f=1$. Let $f\in{\rm ind}_{IZ}^{G}\chi_{r}$ be an
$I(1)$-invariant vector modulo $({\rm Ker}~{}T_{-1,0},{\rm Ker}~{}T_{1,2}).$
As in the previous case, we concentrate only on the zero side of the tree and
assume that $f=f^{0}.$ We write $f=f_{n}+f^{\prime}$ where $f_{n}\neq 0$ and
$f^{\prime}\in B(n-1).$ We further write
$f_{n}=f_{n}^{\prime}+f_{n}^{\prime\prime}$ where $f_{n}^{\prime}$ and
$f_{n}^{\prime\prime}$ are same as in the previous case. Following the steps
in the previous case, we have
$f_{n}^{\prime}=\sum\limits_{\begin{subarray}{c}\mu\in
I_{n}\end{subarray}}a_{0}\mu_{n-1}^{p-r}\left[g^{0}_{n,\mu},1\right]+g_{n-1}^{\prime},$
and
$f_{n}^{\prime\prime}=\sum\limits_{\begin{subarray}{c}\mu\in
I_{n}\end{subarray}}b_{0}\mu_{n-1}^{r+1}\left[g_{n-1,[\mu]_{n-1}}^{0}\left(\begin{array}[]{cc}1&[\mu_{n-1}]\\\
0&1\end{array}\right)w,1\right]+g_{n-1}^{\prime\prime},$
where $a_{0}$ and $b_{0}$ are constants and
$g_{n-1}^{\prime},g_{n-1}^{\prime\prime}\in B(n-1).$ Thus,
$f_{n}=\sum\limits_{\begin{subarray}{c}\mu\in
I_{n}\end{subarray}}a_{0}\mu_{n-1}^{p-r}\left[g^{0}_{n,\mu},1\right]+\sum\limits_{\begin{subarray}{c}\mu\in
I_{n}\end{subarray}}b_{0}\mu_{n-1}^{r+1}\left[g_{n-1,[\mu]_{n-1}}^{0}\left(\begin{array}[]{cc}1&[\mu_{n-1}]\\\
0&1\end{array}\right)w,1\right]+g_{n-1},$
where $g_{n-1}=g_{n-1}^{\prime}+g_{n-1}^{\prime\prime}\in B(n-1).$ Write
$f=f_{n}+f_{n-1}+f^{\prime}.$ We get
$\displaystyle\left(\begin{array}[]{cc}1&p^{n-2}\\\ 0&1\end{array}\right)f-f$
$\displaystyle=\left[\left(\begin{array}[]{cc}1&p^{n-2}\\\
0&1\end{array}\right)f_{n}-f_{n}\right]+\left[\left(\begin{array}[]{cc}1&p^{n-2}\\\
0&1\end{array}\right)f_{n-1}-f_{n-1}\right]$ $\displaystyle\in({\rm
Ker}~{}T_{-1,0},{\rm Ker}~{}T_{1,2}).$
For $e=1$, we have
$\displaystyle\left(\begin{array}[]{cc}1&p^{n-2}\\\
0&1\end{array}\right)f_{n}^{\prime}-f_{n}^{\prime}$
$\displaystyle=\sum\limits_{\begin{subarray}{c}\mu\in
I_{n}\end{subarray}}a_{0}\left[(\mu_{n-1}-(*))^{p-r}-\mu_{n-1}^{p-r}\right]\left[g^{0}_{n,\mu},1\right],$
where
$(*)=\sum\limits_{\begin{subarray}{c}s=1\end{subarray}}^{p-1}(-1)^{p-s}\frac{{p\choose
s}}{p}\mu_{n-2}^{s}.$
Then, by Lemmas 3.6 (1) and 3.5 (1), modulo $({\rm Ker}~{}T_{-1,0},{\rm
Ker}~{}T_{1,2})$, the above expression becomes
(11) $-\sum\limits_{\begin{subarray}{c}\mu\in
I_{n-1}\end{subarray}}a_{0}{p-r\choose
p-1-r}(*)\left[g^{0}_{n-2,[\mu]_{n-2}}\left(\begin{array}[]{cc}1&[\mu_{n-2}]\\\
0&1\end{array}\right)w,1\right].$
Writing $f_{n-1}=f_{n-1}^{\prime}+f_{n-1}^{\prime\prime}$, we have,
$\left(\begin{array}[]{cc}1&p^{n-2}\\\
0&1\end{array}\right)f_{n-1}-f_{n-1}=\left[\left(\begin{array}[]{cc}1&p^{n-2}\\\
0&1\end{array}\right)f_{n-1}^{\prime}-f_{n-1}^{\prime}\right]+\left[\left(\begin{array}[]{cc}1&p^{n-2}\\\
0&1\end{array}\right)f_{n-1}^{\prime\prime}-f_{n-1}^{\prime\prime}\right].$
No term in the first summand of the above equation can cancel a term in (11).
Also, by Lemma 4.6 (2), the possible powers, say $k,$ of $\mu_{n-2}$ in
$f_{n-1}^{\prime\prime}$ must satisfy either $0\leq k\leq r$ or $k=r+1.$ As
$r<p-1$, we have $\max(r+1)=p-1.$ So the maximum power of $\mu_{n-2}$ in the
second summand of the above equation is $p-2.$ In both the cases, the term
involving $\mu_{n-2}^{p-1}$ in (11) will not get cancelled. Since there is no
reduction, this term must be in ${\rm Ker}~{}T_{-1,0},$ which is not possible
by Lemma 3.5 (2). Thus we arrive at a contradiction. So $i_{n-1}$ can not be
$p-r.$ Thus one can always modify $f_{n}^{\prime}$ by a vector
$g_{n-1}^{\prime}$ in $B(n-1).$ Similarly, working with
$f_{n}^{\prime\prime},$ we can modify it by a vector $g_{n-1}^{\prime\prime}$
in $B(n-1).$ Thus $f_{n}$ is congruent to a vector $f_{n-1}$ in $B(n-1)$
modulo $({\rm Ker}~{}T_{-1,0},{\rm Ker}~{}T_{1,2})$ and hence by induction,
$f$ is congruent to a vector $f_{1}$ in $B(1)$ modulo $({\rm
Ker}~{}T_{-1,0},{\rm Ker}~{}T_{1,2}).$ Write
$f_{1}=f_{1}^{\prime}+f_{1}^{\prime\prime},$ where
$f_{1}^{\prime}=\sum\limits_{\begin{subarray}{c}i\end{subarray}}\sum\limits_{\begin{subarray}{c}\mu\in
I_{1}\end{subarray}}a_{i}\mu^{i}\left[g^{0}_{1,\mu},1\right],$
and
$f_{1}^{\prime\prime}=\sum\limits_{\begin{subarray}{c}j\end{subarray}}\sum\limits_{\begin{subarray}{c}\mu\in
I_{1}\end{subarray}}b_{j}\mu^{j}\left[\left(\begin{array}[]{cc}1&\mu\\\
0&1\end{array}\right)w,1\right].$
Considering the action of $\left(\begin{array}[]{cc}1&1\\\
0&1\end{array}\right)$ on $f_{1}$ as in the previous case, we have $0\leq
i\leq p-1-r$ and $0\leq j\leq r$, by Lemma 3.6 and Lemma 4.6. Then, by Lemma
3.6 and Lemma 3.5, modulo $({\rm Ker}~{}T_{-1,0},{\rm Ker}~{}T_{1,2}),$ we get
$f_{1}^{\prime}=\left[\beta,1\right]$ and $f_{1}^{\prime\prime}=\left[{\rm
Id},1\right].$ Thus we can conclude that
$f=c\left[{\rm Id},1\right]+d\left[\beta,1\right].$
This finishes the proof of Theorem 1.2.
### 4.5. A remark on $\pi_{r}$
We show that there is no isomorphism between
$\tau_{r}=\frac{{\rm ind}_{KZ}^{G}\sigma_{r}}{(T)}$
and
$\pi_{r}=\frac{{\rm ind}_{IZ}^{G}\chi_{r}}{({\rm Ker}~{}T_{-1,0},{\rm
Ker}~{}T_{1,2})}$
when $f\neq 1$; i.e., $F$ is not a totally ramified extension of
$\mathbb{Q}_{p}$ (cf. Remark 3).
Note that any $G$-linear isomorphism
$\varphi:\pi_{r}\rightarrow\tau_{r}$
must preserve $I(1)$-invariants and the corresponding $I$-eigenvalues.
Suppose $e=1,f\neq 1$; i.e., $F/\mathbb{Q}_{p}$ is unramified. In this case,
$s_{n}^{q-1-r+p^{l}}$, for $n\geq 2$, is an $I(1)$-invariant in $\pi_{r}$ such
that
$\left(\begin{array}[]{cc}a&b\\\ \varpi c&d\end{array}\right)\cdot
s_{n}^{q-1-r+p^{l}}=a^{r-p^{l}}d^{p^{l}}\cdot s_{n}^{q-1-r+p^{l}}$
by Lemma 4.3. By [Hen19, Theorem 1.2], a basis of the $I(1)$-invariants in
$\tau_{r}$ consists of the vectors
${\rm
Id}\otimes\bigotimes_{j=0}^{f-1}x_{j}^{r_{j}},\alpha\otimes\bigotimes_{j=0}^{f-1}y_{j}^{r_{j}},c_{n}^{p^{l}(r_{l}+1)},\beta
c_{n}^{p^{l}(r_{l}+1)}$
for $n\geq 1$, where
$c_{n}^{k}=\sum\limits_{\begin{subarray}{c}\mu\in
I_{n}\end{subarray}}\left(\begin{array}[]{cc}\varpi^{n}&\mu\\\
0&1\end{array}\right)\otimes\mu_{n-1}^{k}\bigotimes_{j=0}^{f-1}x_{j}^{r_{j}}.$
By [Hen19, Lemma 3.6],
$\left(\begin{array}[]{cc}a&b\\\ \varpi c&d\end{array}\right)\cdot
c_{n}^{k}=a^{r-2k}(ad)^{k}\cdot c_{n}^{k},$
and it follows that there is no $I(1)$-invariant vector in $\tau_{r}$ with
$I$-eigenvalue $a^{r-p^{l}}d^{p^{l}}$. Thus there is no vector in $\tau_{r}$
where $s_{n}^{q-1-r+p^{l}}$ can be mapped under $\varphi$. This gives a
contradiction.
Now, suppose $e>1,f>1$. In this case $t_{n}^{r+p^{l}}$, $n\geq 2$, is an
$I(1)$-invariant vector in $\pi_{r}$ with $I$-eigenvalue
$a^{q-1-p^{l}}d^{r+p^{l}}$, by Lemma 4.3. A basis of the $I(1)$-invariants in
$\tau_{r}$ consists of the vectors
${\rm
Id}\otimes\bigotimes_{j=0}^{f-1}x_{j}^{r_{j}},\alpha\otimes\bigotimes_{j=0}^{f-1}y_{j}^{r_{j}},c_{n}^{p^{l}(r_{l}+1)},\beta
c_{n}^{p^{l}(r_{l}+1)},d_{n}^{l},\beta d_{n}^{l},$
for $n\geq 1$, where
$d_{n}^{l}=\sum\limits_{\begin{subarray}{c}\mu\in
I_{n}\end{subarray}}\left(\begin{array}[]{cc}\varpi^{n}&\mu\\\
0&1\end{array}\right)\otimes\bigotimes_{l\neq j=0}^{f-1}x_{j}^{r_{j}}\otimes
x_{l}^{r_{l}-1}y_{l},$
by [Hen19, Theorem 1.2]. By [Hen19, Lemma 3.6],
$\left(\begin{array}[]{cc}a&b\\\ \varpi c&d\end{array}\right)\cdot
d_{n}^{l}=a^{r-2p^{l}}(ad)^{p^{l}}\cdot d_{n}^{l},$
and once again it can be checked that there is no $I(1)$-invariant vector in
$\tau_{r}$ with $I$-eigenvalue $a^{q-1-p^{l}}d^{r+p^{l}}$, where
$t_{n}^{r+p^{l}}$ can be mapped under $\phi$, giving a contradiction.
## Acknowledgements
The second author would like to thank Council of Scientific and Industrial
Research, Government of India (CSIR) and Industrial Research and Consultancy
Centre, IIT Bombay (IRCC) for financial support.
## References
* [AB13] U. K. Anandavardhanan and Gautam H. Borisagar, _On the $K(n)$-invariants of a supersingular representation of $GL_{2}(\mathbb{Q}_{p})$_, The legacy of Srinivasa Ramanujan, Ramanujan Math. Soc. Lect. Notes Ser., vol. 20, Ramanujan Math. Soc., Mysore, 2013, pp. 55–75. MR 3221302
* [AB15] by same author, _Iwahori-Hecke model for supersingular representations of ${\rm GL}_{2}(\mathbb{Q}_{p})$_, J. Algebra 423 (2015), 1–27. MR 3283706
* [BL94] L. Barthel and R. Livné, _Irreducible modular representations of ${\rm GL}_{2}$ of a local field_, Duke Math. J. 75 (1994), no. 2, 261–292. MR 1290194
* [BP12] Christophe Breuil and Vytautas Paškūnas, _Towards a modulo $p$ Langlands correspondence for ${\rm GL}_{2}$_, Mem. Amer. Math. Soc. 216 (2012), no. 1016, vi+114. MR 2931521
* [Bre03] Christophe Breuil, _Sur quelques représentations modulaires et $p$-adiques de ${\rm GL}_{2}(\mathbb{Q}_{p})$. I_, Compositio Math. 138 (2003), no. 2, 165–188. MR 2018825
* [Hen19] Yotam I. Hendel, _On the universal ${\rm mod}\,p$ supersingular quotients for ${\rm GL}_{2}(F)$ over $\overline{\mathbb{F}}_{p}$ for a general $F/\mathbb{Q}_{p}$_, J. Algebra 519 (2019), 1–38. MR 3873949
* [Sch11] Michael M. Schein, _An irreducibility criterion for supersingular $\mod\ p$ representations of ${\rm GL}_{2}(F)$ for totally ramified extensions $F$ of $\mathbb{Q}_{p}$_, Trans. Amer. Math. Soc. 363 (2011), no. 12, 6269–6289. MR 2833554
|
###### Abstract
In this article we consider the estimation of the log-normalization constant
associated to a class of continuous-time filtering models. In particular, we
consider ensemble Kalman-Bucy filter based estimates based upon several
nonlinear Kalman-Bucy diffusions. Based upon new conditional bias results for
the mean of the afore-mentioned methods, we analyze the empirical log-scale
normalization constants in terms of their $\mathbb{L}_{n}-$errors and
conditional bias. Depending on the type of nonlinear Kalman-Bucy diffusion, we
show that these are of order $(t^{1/2}/N^{1/2})+t/N$ or $1/N^{1/2}$
($\mathbb{L}_{n}-$errors) and of order $[t+t^{1/2}]/N$ or $1/N$ (conditional
bias), where $t$ is the time horizon and $N$ is the ensemble size. Finally, we
use these results for online static parameter estimation for above filtering
models and implement the methodology for both linear and nonlinear models.
Keywords: Kalman-Bucy filter, Riccati equations, nonlinear Markov processes.
Log-Normalization Constant Estimation using the Ensemble Kalman-Bucy Filter
with Application to High-Dimensional Models
BY DAN CRISAN1, PIERRE DEL MORAL2, AJAY JASRA3 & HAMZA RUZAYQAT3
1Department of Mathematics, Imperial College London, London, SW7 2AZ, UK.
E-Mail<EMAIL_ADDRESS>
2Center INRIA Bordeaux Sud-Ouest & Institut de Mathematiques de Bordeaux,
Bordeaux, 33405, FR. E-Mail<EMAIL_ADDRESS>
3Computer, Electrical and Mathematical Sciences and Engineering Division, King
Abdullah University of Science and Technology, Thuwal, 23955, KSA. E-Mail:
<EMAIL_ADDRESS><EMAIL_ADDRESS>
## 1 Introduction
The filtering problem concerns the recursive estimation of a partially
observed Markov process conditioned on a path of observations. It is found in
a wide class of real applications, including finance, applied mathematics and
engineering; see [2] for instance. This article focusses upon the computation
of the normalizing constant for certain classes of filtering problems, to be
introduced below. These normalizing constants can be of interest in statistics
and engineering for model selection and/or parameter estimation.
In this article we study linear and Gaussian models in continuous-time. It is
well-known that, for such models, under some minimal assumptions, the filter
is Gaussian, with mean satisfying the Kalman-Bucy (KB) equations and the
covariance matrix obeying a Riccati equation. In many cases of practical
interest, these latter equations may not be computable, for instance if one
does not have access to an entire trajectory of data. Instead one can use some
suitable approximations obtained via time-discretization methods. However,
even then, if the dimension $r_{1}$ of the state-vector is very large, the
numerical approximation of the KB and Riccati equations can become
computationally prohibitive, often with a computational effort of at least
$\mathcal{O}(r_{1}^{2})$ per update. The case where $r_{1}$ is large occurs in
many applications including ocean and atmosphere science (e.g. [1]) and oil
reservoir simulations (e.g. [17]). A rather elegant and successful solution,
in discrete-time, to this problem was developed in [10] in the guise of the
ensemble Kalman filter (EnKF), which can reduce the cost to
$\mathcal{O}(r_{1})$.
The EnKF can be interpreted as a mean-field particle approximation of a
conditional McKean-Vlasov diffusion (the K-B diffusion). This latter diffusion
shares the same law as the filter associated to the linear and Gaussian model
in continuous-time. Hence a possible alternative to recursively solving the
K-B and Riccati equations is to generate $N\in\mathbb{N}$ independent copies
from the K-B diffusion and use a simple Monte Carlo estimate for expectations
with respect to the filter. However, the diffusion process cannot be simulated
exactly, but can be approximated in a principled way by allowing the $N$
samples to interact; precise details are given in Section 2. The resulting
method, named the Ensemble Kalman-Bucy Filter (EnKBF), is by now rather well-
understood with several contributions on its convergence (as
$N\rightarrow\infty$) analysis; see for instance [3, 5, 8, 12, 19].
In this work we focus upon using several versions of EnKBF for an online
estimate of the normalization constant. In particular the contributions of
this work are:
1. 1.
New results on the conditional bias of the mean using the EnKBF
2. 2.
A derivation of an estimate of the normalization constant using the EnKBF.
3. 3.
A proof that the $\mathbb{L}_{n}-$error of the estimate on the log-scale is of
$\mathcal{O}\big{(}\sqrt{\frac{t}{N}}+\frac{t}{N}\big{)}$ or
$\mathcal{O}\big{(}\tfrac{1}{\sqrt{N}}\big{)}$, depending on the nonlinear
Kalman-Bucy diffusion and where $t$ is the time parameter.
4. 4.
A proof that the conditional bias of the estimate on the log-scale is of
$\mathcal{O}\big{(}\tfrac{t+\sqrt{t}}{N}\big{)}$ or
$\mathcal{O}\big{(}\frac{1}{N}\big{)}$, depending on the nonlinear Kalman-Bucy
diffusion.
5. 5.
A development of a method that uses this estimate to perform online static
parameter estimation.
The result in 1. is of independent interest, but is used directly in 4.. To
the best of our knowledge the estimate in 2. is new and can be computed with a
little extra computational cost over applying an EnKBF (under time
discretization). Whilst several authors have used the EnKF for normalization
constant estimation, e.g. [9], we have not seen the EnKBF version investigated
in the literature.
In addition to contribution 3. & 4., the results establish the decay of the
mean square error (MSE), for instance, of the log-normalizing constant
estimate as the time parameter increases. This rate is expected to occur in
practice, as we will see in simulations, and parallels other results found in
the literature for particle estimates of the normalization constant (e.g. such
as particle filters [7]). In relation to 5., if one assumes that the model of
interest has several unknown and time-homogeneous parameters, $\theta$ say,
then one is interested to estimate such parameters, for instance using
likelihood based methods. We show how our estimate can be leveraged for this
latter task. In this paper we use the simultaneous stochastic perturbation
stochastic approximation (SPSA) method [18] popular in engineering
applications. SPSA is based upon a finite difference estimator of the
gradient, w.r.t. $\theta$, of the log-normalizing constant. It constructs a
stochastic approximation scheme for the estimation of static parameters. We
are not aware of this method being used for the EnKBF. As it is a zero-order
optimization method, we expect to be computationally less expensive than
resorting to using other estimates of the gradient (of the log-normalizing
constant). It should be noted that our work focusses upon the standard or
‘vanilla’ EnKBF, the deterministic EnKBF [16] and deterministic transport type
EnKBFs [15]; other extensions are possible, for instance, to the feedback
particle filter.
This article is structured as follows. In Section 2 we provide details on the
class of filtering problems that we address as well as details on the ensemble
Kalman Bucy filter. The conditional bias of the mean associated to various
EnKBFs is also presented there. In Section 3 we discuss how the normalizing
constant estimate can be computed, as well as its $\mathbb{L}_{n}-$error and
conditional bias. The latter is supported by numerical results checking that
the order of convergence rate indeed holds even under naive Euler
discretization. In Section 4 we show how the normalizing constant estimate can
be used in online static parameter estimation problems.
## 2 Description of the Model and Algorithms
### 2.1 The Kalman-Bucy Filter
Let $(\Omega,\mathcal{F},\mathbb{P})$ be a probability space together with a
filtration $(\mathcal{G}_{t})_{t\geq 0}$ which satisfies the usual conditions.
On $(\Omega,\mathcal{F},\mathbb{P})$ we consider two $\mathcal{G}_{t}$-adapted
processes $X=\\{X_{t},\ t\geq 0\\}$ and $Y=\\{Y_{t},\ t\geq 0\\}$ that form a
time homogeneous linear-Gaussian filtering model of the following form
$\left\\{\begin{array}[]{rcl}dX_{t}&=&A\leavevmode\nobreak\
X_{t}\leavevmode\nobreak\ dt\leavevmode\nobreak\ +\leavevmode\nobreak\
R^{1/2}_{1}\leavevmode\nobreak\ dW_{t}\\\ dY_{t}&=&C\leavevmode\nobreak\
X_{t}\leavevmode\nobreak\ dt\leavevmode\nobreak\ +\leavevmode\nobreak\
R^{1/2}_{2}\leavevmode\nobreak\ dV_{t}.\end{array}\right.$ (1)
In the above display, $(W_{t},V_{t})$ is an $\mathcal{G}_{t}$-adapted
$(r_{1}+r_{2})$-dimensional Brownian motion, $X_{0}$ is an
$\mathcal{G}_{0}$-measurable $r_{1}$-valued Gaussian random vector with mean
and covariance matrix $(\mathbb{E}(X_{0}),P_{0})$ (independent of
$(W_{t},V_{t})$), the symmetric matrices $R^{1/2}_{1}$ and $R^{1/2}_{2}$ are
invertible, $A$ is a square $(r_{1}\times r_{1})$-matrix, $C$ is an
$(r_{2}\times r_{1})$-matrix, and $Y_{0}=0$. We let ${\cal
F}_{t}=\sigma\left(Y_{s},\leavevmode\nobreak\ s\leq t\right)$ be the
filtration generated by the observation process.
It is well-known that the conditional distribution $\eta_{t}$ of the signal
state $X_{t}$ given ${\cal F}_{t}$ is a $r_{1}$-dimensional Gaussian
distribution with a a mean and covariance matrix
$\widehat{X}_{t}:=\mathbb{E}(X_{t}\leavevmode\nobreak\ |\leavevmode\nobreak\
{\cal F}_{t})\quad\mbox{\rm and}\quad
P_{t}:=\mathbb{E}\left(\left(X_{t}-\mathbb{E}(X_{t}\leavevmode\nobreak\
|\leavevmode\nobreak\ {\cal
F}_{t})\right)\left(X_{t}-\mathbb{E}(X_{t}\leavevmode\nobreak\
|\leavevmode\nobreak\ {\cal F}_{t})\right)^{\prime}\right)$
given by the Kalman-Bucy and the Riccati equations
$\displaystyle d\widehat{X}_{t}$ $\displaystyle=$ $\displaystyle
A\leavevmode\nobreak\ \widehat{X}_{t}\leavevmode\nobreak\
dt+P_{t}\leavevmode\nobreak\ C^{\prime}R^{-1}_{2}\leavevmode\nobreak\
\left(dY_{t}-C\widehat{X}_{t}dt\right)$ (2) $\displaystyle\partial_{t}P_{t}$
$\displaystyle=$ $\displaystyle\mbox{\rm Ricc}(P_{t}).$ (3)
with the Riccati drift function from $\SS^{+}_{r_{1}}$ into $\SS_{r_{1}}$
(where $\SS^{+}_{r_{1}}$ (resp. $\SS_{r_{1}}$) is the collection of symmetric
and positive definite (resp. semi-definite) $r_{1}\times r_{1}$ matrices
defined for any $Q\in\SS^{+}_{r_{1}}$ ) by
$\mbox{\rm Ricc}(Q)=AQ+QA^{\prime}-QSQ+R\quad\mbox{\rm with}\quad
R=R_{1}\quad\mbox{\rm and}\quad S:=C^{\prime}R^{-1}_{2}C$ (4)
### 2.2 Nonlinear Kalman-Bucy Diffusions
We now consider three conditional nonlinear McKean-Vlasov type diffusion
processes
$\displaystyle d\overline{X}_{t}$ $\displaystyle=$ $\displaystyle
A\leavevmode\nobreak\ \overline{X}_{t}\leavevmode\nobreak\
dt\leavevmode\nobreak\ +\leavevmode\nobreak\ R^{1/2}_{1}\leavevmode\nobreak\
d\overline{W}_{t}+{\cal P}_{\eta_{t}}C^{\prime}R^{-1}_{2}\leavevmode\nobreak\
\left[dY_{t}-\left(C\overline{X}_{t}dt+R^{1/2}_{2}\leavevmode\nobreak\
d\overline{V}_{t}\right)\right]$ (5) $\displaystyle d\overline{X}_{t}$
$\displaystyle=$ $\displaystyle A\leavevmode\nobreak\
\overline{X}_{t}\leavevmode\nobreak\ dt\leavevmode\nobreak\
+\leavevmode\nobreak\ R^{1/2}_{1}\leavevmode\nobreak\ d\overline{W}_{t}+{\cal
P}_{\eta_{t}}C^{\prime}R^{-1}_{2}\leavevmode\nobreak\
\left[dY_{t}-\left(\frac{1}{2}C\left[\overline{X}_{t}+\eta_{t}(e)\right]dt\right)\right]$
(6) $\displaystyle d\overline{X}_{t}$ $\displaystyle=$ $\displaystyle
A\leavevmode\nobreak\ \overline{X}_{t}\leavevmode\nobreak\
dt\leavevmode\nobreak\ +\leavevmode\nobreak\ R_{1}{\cal
P}_{\eta_{t}}^{-1}\left(\overline{X}_{t}-\eta_{t}(e)\right)\leavevmode\nobreak\
dt+{\cal P}_{\eta_{t}}C^{\prime}R^{-1}_{2}\leavevmode\nobreak\
\left[dY_{t}-\left(\frac{1}{2}C\left[\overline{X}_{t}+\eta_{t}(e)\right]dt\right)\right]$
(7)
where $(\overline{W}_{t},\overline{V}_{t},\overline{X}_{0})$ are independent
copies of $(W_{t},V_{t},X_{0})$ (thus independent of the signal and the
observation path). In the above displayed formula ${\cal P}_{\eta_{t}}$ stands
for the covariance matrix
${\cal
P}_{\eta_{t}}=\eta_{t}\left[(e-\eta_{t}(e))(e-\eta_{t}(e))^{\prime}\right]\quad\mbox{\rm
with}\quad\eta_{t}:=\mbox{\rm Law}(\overline{X}_{t}\leavevmode\nobreak\
|\leavevmode\nobreak\ {\cal F}_{t})\quad\mbox{\rm and}\quad e(x):=x$ (8)
and we use the notation,
$\eta_{t}(f)=\int_{\mathbb{R}^{r_{1}}}f(x)\eta_{t}(dx)$ for
$f:\mathbb{R}^{r_{1}}\rightarrow\mathbb{R}^{r}$ that is $\eta_{t}-$integrable
(in particular, $r=r_{1}$ in (6) and (7) and $r=r_{1}^{2}$ in (8)). Any of
these probabilistic models will be commonly referred to as Kalman-Bucy
(nonlinear) diffusion processes. In the following we will denote by
${\mathcal{G}}_{t}$ the augmented filtration generated by $\overline{X}_{0}$
and the triplet of independent Brownian motions
$(Y_{t},\overline{W}_{t},\overline{V}_{t})$. The process (5) corresponds to
the vanilla type ensemble Kalman-Bucy filter that is typically used in the
literature. The process (6) is associated the so-called deterministic ensemble
Kalman-Bucy filter [16] and (7) is a deterministic transport-inspired equation
[15].
We have the following result that is considered, for instance, in [8]:
###### Lemma 2.1.
Let $\overline{X}_{t}$ be a process such that
$\mathbb{E}[\left|\overline{X}_{0}\right|^{2}]<\infty$ and that it satisfies
any one of (5)-(7). Then the conditional mean and the conditional covariance
matrix (given ${\mathcal{F}}_{t}$) of any of the nonlinear Kalman-Bucy
diffusions (5)-(7) satisfy equations (2) and (3), respectively.
This result enables us to approximate the mean and covariance associated to
the linear filtering problem in (1) by simulating $N$ i.i.d. samples from one
of the processes (5)-(7) exactly and then use the sample mean and sample
covariance to approximate the mean and covariance of the filter. Since this
exact simulation is seldom possible, we consider how one can generate
$N-$particle systems whose sample mean and sample covariance can be used to
replace the exact simulation.
###### Remark 2.1.
Let us observe that equations (5)-(7) have indeed a unique solution in the
class of processes $Z_{t}$ such that
$\mathbb{E}[\sup_{t\in[0,T]}\left|Z_{t}\right|^{2}]<\infty$. To show
existence, one considers the (linear version of the) equations (5)-(7) with
the conditional expectations $\eta_{t}(e)$ and
$\eta_{t}\left[(e-\eta_{t}(e))(e-\eta_{t}(e))^{\prime}\right]$ replaced by the
solution of the equations (2) and (3), respectively, that is,
$\displaystyle d\overline{X}_{t}$ $\displaystyle=$ $\displaystyle
A\leavevmode\nobreak\ \overline{X}_{t}\leavevmode\nobreak\
dt\leavevmode\nobreak\ +\leavevmode\nobreak\ R^{1/2}_{1}\leavevmode\nobreak\
d\overline{W}_{t}+P_{t}C^{\prime}R^{-1}_{2}\leavevmode\nobreak\
\left[dY_{t}-\left(C\overline{X}_{t}dt+R^{1/2}_{2}\leavevmode\nobreak\
d\overline{V}_{t}\right)\right]$ (9) $\displaystyle d\overline{X}_{t}$
$\displaystyle=$ $\displaystyle A\leavevmode\nobreak\
\overline{X}_{t}\leavevmode\nobreak\ dt\leavevmode\nobreak\
+\leavevmode\nobreak\ R^{1/2}_{1}\leavevmode\nobreak\
d\overline{W}_{t}+P_{t}C^{\prime}R^{-1}_{2}\leavevmode\nobreak\
\left[dY_{t}-\left(\frac{1}{2}C\left[\overline{X}_{t}+\eta_{t}(e)\right]dt\right)\right]$
(10) $\displaystyle d\overline{X}_{t}$ $\displaystyle=$ $\displaystyle
A\leavevmode\nobreak\ \overline{X}_{t}\leavevmode\nobreak\
dt\leavevmode\nobreak\ +\leavevmode\nobreak\
R_{1}P_{t}^{-1}\left(\overline{X}_{t}-\hat{X}_{t}\right)\leavevmode\nobreak\
dt+P_{t}C^{\prime}R^{-1}_{2}\leavevmode\nobreak\
\left[dY_{t}-\left(\frac{1}{2}C\left[\overline{X}_{t}+\hat{X}_{t}\right]dt\right)\right].$
(11)
Equations (9)-(11) have a unique solution (as they are linear) that indeed
satisfy the corresponding (nonlinear) equations (5)-(7). To show the
uniqueness of the solutions of equations (5)-(7), one observes first that any
solution $\eta_{t}$ of the (nonlinear) equations (5)-(7) has its conditional
expectations $\eta_{t}(e)$ and
$\eta_{t}\left[(e-\eta_{t}(e))(e-\eta_{t}(e))^{\prime}\right]$ uniquely
characterized by the equations (2) and (3). Therefore they satisfy the
corresponding linear versions of the equations (5)-(7), with the conditional
expectations $\eta_{t}(e)$ and
$\eta_{t}\left[(e-\eta_{t}(e))(e-\eta_{t}(e))^{\prime}\right]$ replaced by the
solution of the equations (2) and (3). In other words, they satisfy equations
(9)-(11) and therefore they are unique as the corresponding linear equations
(9)-(11) have a unique solution.
###### Remark 2.2.
If one modifies (1) to
$\left\\{\begin{array}[]{rcl}dX_{t}&=&f(X_{t})\leavevmode\nobreak\
dt\leavevmode\nobreak\ +\leavevmode\nobreak\ R^{1/2}_{1}\leavevmode\nobreak\
dW_{t}\\\ dY_{t}&=&C\leavevmode\nobreak\ X_{t}\leavevmode\nobreak\
dt\leavevmode\nobreak\ +\leavevmode\nobreak\ R^{1/2}_{2}\leavevmode\nobreak\
dV_{t}.\end{array}\right.$
for some non-linear function
$f:\mathbb{R}^{r_{1}}\rightarrow\mathbb{R}^{r_{1}}$, one can consider a
modification of any of (5)-(7). For instance in the case (5)
$d\overline{X}_{t}=f(\overline{X}_{t})\leavevmode\nobreak\
dt\leavevmode\nobreak\ +\leavevmode\nobreak\ R^{1/2}_{1}\leavevmode\nobreak\
d\overline{W}_{t}+{\cal P}_{\eta_{t}}C^{\prime}R^{-1}_{2}\leavevmode\nobreak\
\left[dY_{t}-\left(C\overline{X}_{t}dt+R^{1/2}_{2}\leavevmode\nobreak\
d\overline{V}_{t}\right)\right].$
We note however that any approximation of this process, as alluded to above,
does not typically provide an approximation of the non-linear filter.
Nonetheless it is considered in many works in the field.
### 2.3 Ensemble Kalman-Bucy Filters
We now consider three Ensemble Kalman-Bucy filters that correspond to the
mean-field particle interpretation of the nonlinear diffusion processes
(5)-(7). To be more precise we let
$(\overline{W}_{t}^{i},\overline{V}_{t}^{i},\xi_{0}^{i})_{1\leq i\leq N}$ be
$N$ independent copies of
$(\overline{W}_{t},\overline{V}_{t},\overline{X}_{0})$. In this notation, we
have the three McKean-Vlasov type interacting diffusion processes for
$i\in\\{1,\dots,N\\}$
$\displaystyle d\xi_{t}^{i}$ $\displaystyle=$ $\displaystyle
A\leavevmode\nobreak\
\xi_{t}^{i}dt+R_{1}^{1/2}d\overline{W}_{t}^{i}+p_{t}C^{\prime}R_{2}^{-1}\left[dY_{t}-\left(C\xi_{t}^{i}dt+R_{2}^{1/2}\leavevmode\nobreak\
d\overline{V}_{t}^{i}\right)\right]$ (12) $\displaystyle d\xi_{t}^{i}$
$\displaystyle=$ $\displaystyle A\leavevmode\nobreak\
\xi_{t}^{i}dt+R_{1}^{1/2}d\overline{W}_{t}^{i}+p_{t}C^{\prime}R_{2}^{-1}\left[dY_{t}-\left(\frac{1}{2}C\left[\xi_{t}^{i}+\eta_{t}^{N}(e)\right]dt\right)\right]$
(13) $\displaystyle d\xi_{t}^{i}$ $\displaystyle=$ $\displaystyle
A\leavevmode\nobreak\
\xi_{t}^{i}dt+R_{1}(p_{t})^{-1}\left(\xi_{t}^{i}-\eta_{t}^{N}(e)\right)\leavevmode\nobreak\
dt+p_{t}C^{\prime}R_{2}^{-1}\left[dY_{t}-\left(\frac{1}{2}C\left[\xi_{t}^{i}+\eta_{t}^{N}(e)\right]dt\right)\right]$
(14)
with the rescaled particle covariance matrices
$p_{t}:=\left(1-\frac{1}{N}\right)^{-1}\leavevmode\nobreak\
\mathcal{P}_{\eta_{t}^{N}}=\frac{1}{N-1}\sum_{1\leq i\leq
N}\left(\xi_{t}^{i}-m_{t}\right)\left(\xi_{t}^{i}-m_{t}\right)^{\prime}$ (15)
and the empirical measures
$\eta_{t}^{N}:=\frac{1}{N}\sum_{1\leq i\leq
N}\delta_{\xi_{t}^{i}}\quad\mbox{\rm and the sample mean}\quad
m_{t}:=\frac{1}{N}\sum_{1\leq i\leq N}\xi_{t}^{i}.$
Note that for (14) one needs $N\geq r_{1}$ for the almost sure invertibility
of $p_{t}$, but it is also sufficient to use the pseudo inverse instead of the
inverse.
These processes have been thoroughly studied in the literature and a review
can be found in [3]. We have the evolution equations for (12) and (13)
$\begin{array}[]{rcl}dm_{t}&=&\displaystyle A\leavevmode\nobreak\
m_{t}dt+p_{t}\leavevmode\nobreak\ C^{\prime}\Sigma^{-1}\leavevmode\nobreak\
\left(dY_{t}-Cm_{t}\leavevmode\nobreak\
dt\right)+\frac{1}{\sqrt{N+1}}\leavevmode\nobreak\ d\overline{M}_{t}\\\ &&\\\
dp_{t}&=&\displaystyle\mbox{\rm Ricc}(p_{t})\leavevmode\nobreak\
dt+\frac{1}{\sqrt{N}}\leavevmode\nobreak\ dM_{t}\end{array}$
with a triplet of pairwise orthogonal martingales $M_{t}$ and
$\overline{M}_{t}$ and the innovation martingale
$\widehat{M}_{t}:=Y_{t}-C\int_{0}^{t}\widehat{X}_{s}ds.$
Note that for any $t\geq 0$, the absolute moments of $\overline{M}_{t}$ and
$M_{t}$ are uniformly, w.r.t. $N$, upper-bounded. We remark that the
mathematical theory for (14) in the linear-Gaussian setting reverts to that of
the standard Kalman-Bucy filter as detailed in [5].
###### Remark 2.3.
Returning to the context of Remark 2.2, one could, for instance, run the
following version of (12) for $i\in\\{1,\dots,N\\}$
$d\xi_{t}^{i}=f(\xi_{t}^{i})dt+R_{1}^{1/2}d\overline{W}_{t}^{i}+p_{t}^{N}C^{\prime}R_{2}^{-1}\left[dY_{t}-\left(C\xi_{t}^{i}dt+R_{2}^{1/2}\leavevmode\nobreak\
d\overline{V}_{t}^{i}\right)\right]$
again, noting that this will typically not produce a consistent approximation
of the non-linear filter, as would be the case for (12) with the model (1).
Below we denote $\|\cdot\|$ as the $L_{2}-$norm for vectors. For a square
matrix $B$ say, $B_{\textrm{sym}}=\tfrac{1}{2}(B+B^{\prime})$ and $\mu(B)$
being the largest eigenvalue of $B_{\textrm{sym}}$. In the cases of (12) and
(13), [8] consider the following assumption (recall (4)): we have $\mu(S)>0$
with
$S=\mu(S)Id$ (16)
where $Id$ is the $r_{1}\times r_{1}$ identity matrix (in general we use $Id$
to denote the identity matrix of ‘appropriate dimension’, appropriate
depending on the context). [8] prove the following time-uniform convergence
theorem for the mean.
###### Theorem 2.1.
Consider the cases of (12) and (13). Assume that (16) holds and that
$\mu(A)<0$. Then for any $n\geq 1$ and $N$ sufficiently large, we have that
$\sup_{t\geq
0}\mathbb{E}[\|m_{t}-\widehat{X}_{t}\|^{n}]^{\tfrac{1}{n}}\leq\frac{\mathsf{C}(n)}{\sqrt{N}}$
where $\mathsf{C}(n)<\infty$ is a constant that does not depend upon $N$.
We denote by $P_{\infty}$ as the solution to $\mbox{\rm Ricc}(P_{\infty})=0$.
Set $\mu(A-P_{\infty}S)$ as the largest eigenvalue of
$\tfrac{1}{2}(A-P_{\infty}S+(A-P_{\infty}S)^{\prime})$. For the case of (13),
the following result, presented in [3], can be shown.
###### Theorem 2.2.
Consider the case of (13). Assume that $\mu(A-P_{\infty}S)<0$ and
$S\in\mathbb{S}_{r_{1}}^{+}$. Then for any $n\geq 1$, $N\geq 2$ there exists a
$t(n,N)>0$ with $t(n,N)\rightarrow\infty$ as $N\rightarrow\infty$, such that
for any $t\in[0,t(n,N)]$
$\mathbb{E}[\|m_{t}-\widehat{X}_{t}\|^{n}]^{\tfrac{1}{n}}\leq\frac{\mathsf{C}}{\sqrt{N}}$
where $\mathsf{C}<\infty$ is a constant that does not depend upon $N$.
$t(n,N)$ is characterized in [6].
We note that $t(n,N)$ is $\mathcal{O}(\log(N))$; see [6, Theorem 2.1] when
$\epsilon=1/\sqrt{N}$ ($\epsilon$ is as [6, Theorem 2.1]).
For the case of (14) we have the following result. Below we cite [3, eq.
(2.7)] and it should be noted that the notation of that article differs
slightly to this one, so we mean of course that the equation holds in the
notation of this article. Also note that we use the inverse of $p_{t}^{N}$,
but as noted above this need not be the case and then the constraint on $N$
below is not required.
###### Theorem 2.3.
Consider the case of (14). Assume that [3, eq. (2.7)] holds. Then for any
$n\geq 1$ there exists a $\mathsf{C}<\infty$, $\kappa\in(0,1)$ such that for
any $N\geq r_{1}$ and $t\geq 0$ we have that
$\mathbb{E}[\|m_{t}-\widehat{X}_{t}\|^{n}]^{\tfrac{1}{n}}\leq\frac{\mathsf{C}\kappa^{t}}{\sqrt{N}}.$
### 2.4 Conditional Bias Result
We set
$\widehat{m}_{t}:=\mathbb{E}(m_{t}\leavevmode\nobreak\ |\leavevmode\nobreak\
{\cal F}_{t})\qquad\widehat{p}_{t}:=\mathbb{E}(p_{t}\leavevmode\nobreak\
|\leavevmode\nobreak\ {\cal F}_{t}).$
We consider bounds on
$\mathbb{E}\left(\|\widehat{m}_{t}-\widehat{X}_{t}\|^{n}\right)^{1/n}.$
These bounds do not currently exist in the literature and will allow us to
investigate our new estimator for the normalization constant, later on in the
article. In the case (14) we have (see e.g. [5])
$\widehat{m}_{t}=\widehat{X}_{t}\quad\mbox{\rm and}\quad\widehat{p}_{t}=P_{t}$
(17)
so we focus only on (12) and (13).
Using the second order Taylor-type expansions presented in [4] when the pair
$(A,R_{1}^{1/2})$ is stabilizable and $(A,S^{1/2})$ is detectable we have the
uniform almost sure bias and for any $n\geq 1$, the $\mathbb{L}_{n}$-mean
error estimates
$0\leq P_{t}-\widehat{p}_{t}\leq\mathsf{C}_{1}/N\quad\mbox{\rm
and}\quad\mathbb{E}\left(\|P_{t}-p_{t}\|^{n}\leavevmode\nobreak\
|\leavevmode\nobreak\ {\cal F}_{t}\right)^{1/n}\leq\mathsf{C}_{2}(n)/\sqrt{N}$
(18)
as soon as $N$ is chosen sufficiently large, for some deterministic constants
$\mathsf{C}_{1},\mathsf{C}_{2}(n)$ that do not depend on the time horizon or
$N$ (but $\mathsf{C}_{2}(n)$ does depend upon $n$). Whenever $\mu(A)<0$,
Theorem 2.1 yields the rather crude estimate
$\mathbb{E}\left(\|\widehat{m}_{t}-\widehat{X}_{t}\|^{n}\right)^{1/n}=\mathbb{E}\left(\|\mathbb{E}\left(m_{t}-\widehat{X}_{t}\leavevmode\nobreak\
|\leavevmode\nobreak\ {\cal
F}_{t}\right)\|^{n}\right)^{1/n}\leq\mathbb{E}\left(\|m_{t}-\widehat{X}_{t}\|^{n}\right)^{1/n}\leq\frac{\mathsf{C}(n)}{\sqrt{N}}.$
(19)
Our objective is to improve the estimate (19). We observe that
$\begin{array}[]{l}d(m_{t}-\widehat{X}_{t})\\\ \\\ =\displaystyle
A\leavevmode\nobreak\
(m_{t}-\widehat{X}_{t})dt+(p_{t}-P_{t})\leavevmode\nobreak\
C^{\prime}\Sigma^{-1}\leavevmode\nobreak\
\left(dY_{t}-Cm_{t}\leavevmode\nobreak\ dt\right)\\\ \\\ \displaystyle\hskip
85.35826pt+P_{t}\leavevmode\nobreak\ C^{\prime}\Sigma^{-1}\leavevmode\nobreak\
\left(dY_{t}-Cm_{t}\leavevmode\nobreak\ dt\right)-P_{t}\leavevmode\nobreak\
C^{\prime}\Sigma^{-1}\leavevmode\nobreak\
\left(dY_{t}-C\widehat{X}_{t}\leavevmode\nobreak\
dt\right)+\frac{1}{\sqrt{N+1}}\leavevmode\nobreak\
d\overline{M}_{t}\end{array}$ (20)
which yields
$\begin{array}[]{l}d(\widehat{m}_{t}-\widehat{X}_{t})\\\ \\\
=\displaystyle(A-P_{t}S)\leavevmode\nobreak\
(\widehat{m}_{t}-\widehat{X}_{t})dt+(\widehat{p}_{t}-P_{t})\leavevmode\nobreak\
C^{\prime}\Sigma^{-1}\leavevmode\nobreak\
\left(dY_{t}-C\widehat{X}_{t}\leavevmode\nobreak\
dt\right)+\mathbb{E}\left((P_{t}-p_{t})\leavevmode\nobreak\
S(m_{t}-\widehat{X}_{t})\leavevmode\nobreak\ |\leavevmode\nobreak\ {\cal
F}_{t}\right)\leavevmode\nobreak\ dt.\end{array}$ (21)
Equation (21) is deduced by conditioning both terms in (20) with respect to
${\cal F}_{t}$, exchanging the (stochastic) integration with the conditional
expectation for all the terms on the right hand side of (20). To justify this
Fubini-like exchange one can use, for example, Lemma 3.21 in [2]. One also
needs to use the fact that, for arbitrary $0\leq s\leq t$,
$\widehat{m}_{s}=\mathbb{E}(m_{s}\leavevmode\nobreak\ |\leavevmode\nobreak\
{\cal F}_{s})=\mathbb{E}(m_{s}\leavevmode\nobreak\ |\leavevmode\nobreak\ {\cal
F}_{t})$ and $\widehat{p}_{s}:=\mathbb{E}(p_{s}\leavevmode\nobreak\
|\leavevmode\nobreak\ {\cal F}_{s})=\mathbb{E}(p_{s}\leavevmode\nobreak\
|\leavevmode\nobreak\ {\cal F}_{t})$. To justify this, one can use, for
example, Proposition 3.15 in [2]. We note that both Proposition 3.15 and Lemma
3.21 hold true when the conditioning is done with respect the equivalent
probability measure $\tilde{\mathbb{P}}$ under which $Y$ is a Brownian motion.
However since $(\overline{W}_{t},\overline{V}_{t},\overline{X}_{0})$ are
independent of $Y$ under $\mathbb{P}$ they remain independent of $Y$ under
$\tilde{\mathbb{P}}$ and therefore conditioning the processes $\xi^{i}$ under
$\tilde{\mathbb{P}}$ is the same as conditioning them under $\mathbb{P}$. Note
that argument this does _not_ apply to the original signal $X$.
Let ${\cal E}_{s,t}$ be the exponential semigroup (or the state transition
matrix) associated with the smooth flow of matrices $t\mapsto(A-P_{t}S)$
defined for any $s\leq t$ by the forward and backward differential equations,
$\partial_{t}\,{\cal E}_{s,t}=(A-P_{t}S)\,{\cal E}_{s,t}\quad\mbox{\rm
and}\quad\partial_{s}\,{\cal E}_{s,t}=-{\cal E}_{s,t}\,(A-P_{s}S)$
with ${\cal E}_{s,s}=Id$. Equivalently in terms of the matrices ${\cal
E}_{t}:={\cal E}_{0,t}$ we have ${\cal E}_{s,t}={\cal E}_{t}{\cal
E}_{s}^{-1}$. Under some observability and controllability conditions, we have
that the drift matrices $(A-P_{t}S)$ delivers a uniformly stable time varying
linear system in the sense that
$\|{\cal E}_{s,t}\|\leq c\leavevmode\nobreak\ e^{-\lambda\leavevmode\nobreak\
(t-s)}\quad\mbox{\rm for some}\quad\lambda>0.$ (22)
See for instance Eq. (2.13) in the review article [3]. In this notation,
recalling that $\widehat{m}_{0}=\widehat{X}_{0}$ we have the bias formulae
$\widehat{m}_{t}-\widehat{X}_{t}=\int_{0}^{t}{\cal
E}_{s,t}\leavevmode\nobreak\
\underbrace{(\widehat{p}_{s}-P_{s})}_{\in[-c_{1}/N,0]\leavevmode\nobreak\
a.e.}\leavevmode\nobreak\ C^{\prime}\Sigma^{-1}\leavevmode\nobreak\
\left(dY_{s}-C\widehat{X}_{s}\leavevmode\nobreak\ dt\right)+\int_{0}^{t}{\cal
E}_{s,t}\leavevmode\nobreak\ \mathbb{E}\left((P_{s}-p_{s})\leavevmode\nobreak\
S(m_{s}-\widehat{X}_{s})\leavevmode\nobreak\ |\leavevmode\nobreak\ {\cal
F}_{s}\right)\leavevmode\nobreak\ ds.$
Note that $(\widehat{p}_{s}-P_{s})\in[-c_{1}/N,0]$ a.e. can be justified via
(18). Combining the Burkholder-Davis-Gundy inequality with the almost sure and
uniform bias estimate (18) and the exponential decay (22) we obtain the
uniform estimate
$\begin{array}[]{l}\displaystyle\mathbb{E}\left(\|\int_{0}^{t}{\cal
E}_{s,t}\leavevmode\nobreak\ (\widehat{p}_{s}-P_{s})\leavevmode\nobreak\
C^{\prime}\Sigma^{-1}\leavevmode\nobreak\
\left(dY_{s}-C\widehat{X}_{s}\leavevmode\nobreak\
dt\right)\|^{n}\right)^{1/n}\displaystyle\leq\frac{\mathsf{C}(n)}{N}\end{array}$
for some deterministic constants $\mathsf{C}(n)$ that do not depend upon $t$
nor $N$. Conversely, combining Hölder’s inequality with the uniform variance
estimate (18) we have the almost sure inequality
$\mathbb{E}\left(\|\mathbb{E}\left((P_{s}-p_{s})\leavevmode\nobreak\
S(m_{s}-\widehat{X}_{s})\leavevmode\nobreak\ |\leavevmode\nobreak\ {\cal
F}_{s}\right)\|^{n}\right)^{1/n}\leq\frac{\mathsf{C}(n)}{N}\leavevmode\nobreak\
$
for some deterministic constants $\mathsf{C}(n)$ that do not depend upon $t$
nor $N$. Applying the generalized Minkowski inequality we have thus proved the
following theorem.
###### Theorem 2.4.
Consider the cases of (12) and (13). Assume that (16) holds and that
$\mu(A)<0$. Then for any $n\geq 1$ and $N$ sufficiently large, we have
$\sup_{t\geq
0}\mathbb{E}\left(\|\widehat{m}_{t}-\widehat{X}_{t}\|^{n}\right)^{1/n}\leq\frac{\mathsf{C}(n)}{N}$
where $\mathsf{C}(n)<\infty$ is a constant that does not depend upon $N$.
## 3 Computing the Normalizing Constant
### 3.1 Estimation
We define
$Z_{t}:=\frac{{\mathcal{L}}_{X_{0:t},Y_{0:t}}}{{\mathcal{L}}_{X_{0:t},W_{0:t}}}$
to be the density of ${\mathcal{L}}_{X_{0:t},Y_{0:t}}$, the law of the process
$(X,Y)$ and that of ${\mathcal{L}}_{X_{0:t},W_{0:t}}$, the law of the process
$(X,W)$. That is,
${\mathbb{E}}[f(X_{0:t})g(Y_{0:t})]={\mathbb{E}}[f(X_{0:t})g(W_{0:t})Z_{t}(X,Y)].$
One can show that (see Exercise 3.14 pp 56 in [2])
$Z_{t}(X,Y)=\exp{\left[\int_{0}^{t}\left[\langle
CX_{s},R_{2}^{-1}dY_{s}\rangle-\frac{1}{2}\langle
X_{s},SX_{s}\rangle\leavevmode\nobreak\ ds\right]\right]}.$
Following a standard approach (see, e.g., Chapter 3 in [2]), we introduce a
probability measure $\tilde{\mathbb{P}}$ by specifying its Radon–Nikodym
derivative with respect to $\mathbb{P}$ to be given by $Z_{t}(X,Y)^{-1}$, i.e.
$\left.\frac{{\mathrm{d}}\tilde{\mathbb{P}}}{{\mathrm{d}}\mathbb{P}}\right|_{\mathcal{G}_{t}}=Z_{t}.$
Under $\tilde{\mathbb{P}}$, the observation process $Y$ is a scaled Brownian
motion independent of $X$; additionally the law of the signal process $X$
under $\tilde{\mathbb{P}}$ is the same as its law under $\mathbb{P}$111The
proof at this statement is an immediate application of Girsanov’s theorem and
follows in a similar manner with the proof of Proposition 3.13 in [2]..
Moreover, for every $f$ defined on the signal path space, we have the Bayes’
formula
$\mathbb{E}\left(f((X_{s})_{s\in[0,t]})\leavevmode\nobreak\
|\leavevmode\nobreak\ {\cal
F}_{t}\right)=\frac{\tilde{\mathbb{E}}\left(f((X_{s})_{s\in[0,t]})\leavevmode\nobreak\
Z_{t}(X,Y)\leavevmode\nobreak\ |\leavevmode\nobreak\ {\cal
F}_{t}\right)}{\tilde{\mathbb{E}}\left(Z_{t}(X,Y)\leavevmode\nobreak\
|\leavevmode\nobreak\ {\cal F}_{t}\right)}.$ (23)
where $\tilde{\mathbb{E}}()$ means expectation with respect to
$\tilde{\mathbb{P}}$222 Formula (23) is commonly known as the Kallianpur-
Striebel formula. For a proof, see, e.g., Prop 3.16 in [2].. Since, under
$\tilde{\mathbb{P}}$, $X$ and $Y$ are independent we can interpret the
numerator $\tilde{\mathbb{E}}\left(f((X_{s})_{s\in[0,t]})\leavevmode\nobreak\
Z_{t}(X,Y)\leavevmode\nobreak\ |\leavevmode\nobreak\ {\cal F}_{t}\right)$ and
$\tilde{\mathbb{E}}\left(Z_{t}(X,Y)\leavevmode\nobreak\ |\leavevmode\nobreak\
{\cal F}_{t}\right)$ as integrals with respect to the law of the signal, where
the integrant has the observation process path fixed $Y$.333This
interpretation can be made rigorous, see e.g. the robust representation
formulation of the filtering problem described in Chapter 5 in [2].. We can
therefore, let $\overline{Z}_{t}(Y)$ be the likelihood function defined by
$\overline{Z}_{t}(Y):=\mathbb{E}_{Y}\left(Z_{t}(X,Y)\right),$
where $\mathbb{E}_{Y}\left(\mbox{\LARGE.}\right)$ stands for the expectation
w.r.t. the signal process when the observation is fixed and independent of the
signal. In this notation, the Bayes’ formula (23) takes the form
$\mathbb{E}\left(f((X_{s})_{s\in[0,t]})\leavevmode\nobreak\
|\leavevmode\nobreak\ {\cal
F}_{t}\right)=\frac{\mathbb{E}_{Y}\left(f((X_{s})_{s\in[0,t]})\leavevmode\nobreak\
Z_{t}(X,Y)\right)}{\mathbb{E}_{Y}\left(Z_{t}(X,Y)\right)}.$
We have
$\displaystyle dZ_{t}(X,Y)$ $\displaystyle=$ $\displaystyle\left[\langle
CX_{t},R_{2}^{-1}dY_{t}\rangle-\frac{1}{2}\langle
X_{t},SX_{t}\rangle\leavevmode\nobreak\
dt\right]Z_{t}+\frac{1}{2}\leavevmode\nobreak\ \langle
X_{t},SX_{t}\rangle\leavevmode\nobreak\ Z_{t}\leavevmode\nobreak\ dt$
$\displaystyle=$ $\displaystyle Z_{t}(X,Y)\leavevmode\nobreak\ \langle
CX_{t},R_{2}^{-1}dY_{t}\rangle$
We check this claim using the fact that
$\displaystyle d\langle CX_{t},R_{2}^{-1}dY_{t}\rangle d\langle
CX_{t},R_{2}^{-1}dY_{t}\rangle$ $\displaystyle=$
$\displaystyle\sum_{k,l}(CX_{t})(k)(R^{-1/2}_{2}dV_{t})(k)(CX_{t})(l)(R^{-1/2}_{2}dV_{t})(l)$
$\displaystyle=$
$\displaystyle\sum_{k,l,k^{\prime},l^{\prime}}(CX_{t})(k)\leavevmode\nobreak\
R^{-1/2}_{1}(k,k^{\prime})dV_{t}(k^{\prime})(CX_{t})(l)R_{2}^{-1/2}(l,l^{\prime})dV_{t}(l^{\prime})$
$\displaystyle=$
$\displaystyle\sum_{k,l,k^{\prime}}(CX_{t})(k)\leavevmode\nobreak\
R_{2}^{-1}(k,l)(CX_{t})(l)dt=\langle X_{t},SX_{t}\rangle\leavevmode\nobreak\
dt$
This implies that
$\overline{Z}_{t}(Y)=1+\int_{0}^{t}\leavevmode\nobreak\
\overline{Z}_{s}(Y)\leavevmode\nobreak\
\overline{Z}_{s}(Y)^{-1}\mathbb{E}_{Y}\left(Z_{s}(X,Y)\leavevmode\nobreak\
\langle
CX_{s},R_{2}^{-1}dY_{s}\rangle\right)=1+\int_{0}^{t}\leavevmode\nobreak\
\overline{Z}_{s}(Y)\leavevmode\nobreak\ \langle
C\widehat{X}_{s},R_{2}^{-1}dY_{s}\rangle$
from which we conclude that
$\overline{Z}_{t}(Y)=\exp{\left[\int_{0}^{t}\left[\leavevmode\nobreak\ \langle
C\widehat{X}_{s},R_{2}^{-1}dY_{s}\rangle-\frac{1}{2}\langle\widehat{X}_{s},S\widehat{X}_{s}\rangle\leavevmode\nobreak\
ds\right]\right]}$
or equivalently
$d\overline{Z}_{t}(Y)=\overline{Z}_{t}(Y)\leavevmode\nobreak\ \langle
C\widehat{X}_{t},R_{2}^{-1}dY_{t}\rangle.$
This suggests the estimator:
$\overline{Z}_{t}^{N}(Y)=\exp{\left[\int_{0}^{t}\left[\leavevmode\nobreak\
\langle Cm_{s},R_{2}^{-1}dY_{s}\rangle-\frac{1}{2}\langle
m_{s},Sm_{s}\rangle\leavevmode\nobreak\ ds\right]\right]}.$ (24)
which satisfies the equation
$d\overline{Z}_{t}(Y)=\overline{Z}_{t}(Y)\leavevmode\nobreak\ \langle
Cm_{t},R_{2}^{-1}dY_{t}\rangle.$ (25)
We remark that any of the three processes (12)-(14) can be used to compute
$\overline{Z}_{t}^{N}(Y)$.
### 3.2 Justification of the Estimator
We now seek to justify the estimator $\overline{Z}_{t}^{N}(Y)$. We have
$\begin{array}[]{l}\overline{Z}_{t}(X,Y)\\\ \\\
:=\overline{Z}_{t}(Y)^{-1}Z_{t}(X,Y)\\\ \\\
=\exp{\left[\int_{0}^{t}\left[\langle
C(X_{s}-\widehat{X}_{s}),R_{2}^{-1}(dY_{s}-C\widehat{X}_{s}ds)\rangle-\left[\frac{1}{2}\langle
X_{s},SX_{s}\rangle-\frac{1}{2}\langle\widehat{X}_{s},S\widehat{X}_{s}\rangle-\langle
X_{s}-\widehat{X}_{s},S\widehat{X}_{s}\rangle\right]\leavevmode\nobreak\
ds\right]\right]}\\\ \\\ =\exp{\left[\int_{0}^{t}\left[\langle
C(X_{s}-\widehat{X}_{s}),R_{2}^{-1}(dY_{s}-C\widehat{X}_{s}ds)\rangle-\frac{1}{2}\langle(X_{s}-\widehat{X}_{s}),S(X_{s}-\widehat{X}_{s})\rangle\leavevmode\nobreak\
ds\right]\right]}\end{array}$
from which we conclude that
$d\overline{Z}_{t}(X,Y)=\overline{Z}_{t}(X,Y)\leavevmode\nobreak\ \langle
C(X_{s}-\widehat{X}_{s}),R_{2}^{-1}(dY_{s}-C\widehat{X}_{s}ds)\rangle.$
This already implies that
$\mathbb{E}\left[\overline{Z}_{t}(X,Y)\leavevmode\nobreak\
|\leavevmode\nobreak\ {\cal
F}_{t}\right]=1\Longleftrightarrow\mathbb{E}\left[Z_{t}(X,Y)\leavevmode\nobreak\
|\leavevmode\nobreak\ {\cal F}_{t}\right]=\overline{Z}_{t}(Y).$
Now, we have
$\begin{array}[]{l}\overline{Z}_{t}^{N}(Y)\overline{Z}_{t}^{-1}(Y)\\\ \\\
=\exp{\left[\int_{0}^{t}\left[\leavevmode\nobreak\ \langle
C(m_{s}-\widehat{X}_{s}),R_{2}^{-1}dY_{s}\rangle-\frac{1}{2}\left[\langle
m_{s},Sm_{s}\rangle-\langle\widehat{X}_{s},S\widehat{X}_{s}\rangle\right]\leavevmode\nobreak\
ds\right]\right]}\\\ \\\ =\exp{\left[\int_{0}^{t}\left[\leavevmode\nobreak\
\langle
C(m_{s}-\widehat{X}_{s}),R_{2}^{-1}dY_{s}\rangle-\frac{1}{2}\langle(m_{s}-\widehat{X}_{s}),S(m_{s}-\widehat{X}_{s})\rangle\leavevmode\nobreak\
ds\right]\right]\nobreak\leavevmode}\\\ \\\
\times\exp{\left[\int_{0}^{t}\left[\leavevmode\nobreak\
\frac{1}{2}\langle(m_{s}-\widehat{X}_{s}),S(m_{s}-\widehat{X}_{s})\rangle-\frac{1}{2}\left[\langle
m_{s},Sm_{s}\rangle-\langle\widehat{X}_{s},S\widehat{X}_{s}\rangle\leavevmode\nobreak\
ds\right]\right]\right]}\\\ \\\
=\exp{\left[\int_{0}^{t}\left[\leavevmode\nobreak\ \langle
C(m_{s}-\widehat{X}_{s}),R_{2}^{-1}dY_{s}\rangle-\frac{1}{2}\langle(m_{s}-\widehat{X}_{s}),S(m_{s}-\widehat{X}_{s})\rangle\leavevmode\nobreak\
ds\right]\right]}\times\exp{\left[\int_{0}^{t}\langle\widehat{X}_{s}-m_{s},S\widehat{X}_{s}\rangle
ds\right]}\\\ \\\ =\exp{\left[\int_{0}^{t}\left[\leavevmode\nobreak\ \langle
C(m_{s}-\widehat{X}_{s}),R_{2}^{-1}dY_{s}\rangle-\frac{1}{2}\langle(m_{s}-\widehat{X}_{s}),S(m_{s}-\widehat{X}_{s})\rangle\leavevmode\nobreak\
ds\right]\right]}\times\exp{\left[-\int_{0}^{t}\langle
C(m_{s}-\widehat{X}_{s}),R^{-1}C\widehat{X}_{s}\rangle ds\right]}\\\ \\\
=\exp{\left[\int_{0}^{t}\left[\leavevmode\nobreak\ \langle
C(m_{s}-\widehat{X}_{s}),R_{2}^{-1}(dY_{s}-C\widehat{X}_{s}ds)\rangle-\frac{1}{2}\langle(m_{s}-\widehat{X}_{s}),S(m_{s}-\widehat{X}_{s})\rangle\leavevmode\nobreak\
ds\right]\right]}\end{array}$
Observe that $dY_{s}-C\widehat{X}_{s}ds$ is an ${\cal F}_{s}$-martingale
increment. We also have
$d(\overline{Z}_{t}^{N}(Y)\overline{Z}_{t}^{-1}(Y))=\overline{Z}_{t}^{N}(Y)\overline{Z}_{t}^{-1}(Y)\leavevmode\nobreak\
\langle
C(m_{t}-\widehat{X}_{t}),R_{2}^{-1}(dY_{t}-C\widehat{X}_{t}dt)\rangle.$ (26)
Now to conclude, it follows that that
$\overline{Z}_{t}^{N}(Y)\overline{Z}_{t}^{-1}(Y)$ is a positive local
martingale and therefore a supermartingale, hence
$\mathbb{E}(\overline{Z}_{t}^{N}(Y)\overline{Z}_{t}^{-1}(Y))\leq\mathbb{E}(\overline{Z}_{0}^{N}(Y)\overline{Z}_{0}^{-1}(Y))=1$
for all $t\geq 0$. To justify the martingale property of
$\overline{Z}_{t}^{N}(Y)\overline{Z}_{t}^{-1}(Y)$ one can use, for example, an
argument based on Corollary 5.14 in [13]. As a result of the above
calculations we can deduce that $\overline{Z}_{t}^{N}(Y)$ is in some sense
well-defined, but a biased estimator of the normalization constant. Therefore,
we will focus on the logarithm of the normalization constant, as it is this
latter quantity that is used in practical algorithms. We begin by
investigating the conditional bias.
### 3.3 Conditional Bias
We now consider the estimate of the logarithm of the normalizing constant
$U_{t}^{N}(Y):=\log(\overline{Z}_{t}^{N}(Y))$
with the notation $U_{t}(Y)=\log(\overline{Z}_{t}^{N}(Y))$.
Set $\widehat{U}_{t}(Y):=\mathbb{E}(U_{t}^{N}(Y)\leavevmode\nobreak\
|\leavevmode\nobreak\ {\cal F}_{t})$. Then, using (26) we have
$\begin{array}[]{l}U_{t}^{N}(Y)-U_{t}(Y)\\\ \\\
=\log(\overline{Z}_{t}^{N}(Y)/\overline{Z}_{t}(Y))\\\ \\\
\displaystyle=\int_{0}^{t}\left[\leavevmode\nobreak\ \langle
C(m_{s}-\widehat{X}_{s}),R_{2}^{-1}(dY_{s}-C\widehat{X}_{s}ds)\rangle-\frac{1}{2}\langle(m_{s}-\widehat{X}_{s}),S(m_{s}-\widehat{X}_{s})\rangle\leavevmode\nobreak\
ds\right].\end{array}$
This yields the bias formula
$\begin{array}[]{l}\widehat{U}_{t}(Y)-U_{t}(Y)\\\ \\\
\displaystyle=\int_{0}^{t}\left[\leavevmode\nobreak\ \langle
C(\widehat{m}_{s}-\widehat{X}_{s}),R_{2}^{-1}(dY_{s}-C\widehat{X}_{s}ds)\rangle-\frac{1}{2}\leavevmode\nobreak\
\mathbb{E}\left(\langle(m_{s}-\widehat{X}_{s}),S(m_{s}-\widehat{X}_{s})\rangle\leavevmode\nobreak\
|\leavevmode\nobreak\ {\cal F}_{s}\right)\leavevmode\nobreak\
ds\right].\end{array}$ (27)
Taking the expectation in the above display, when (16) holds and $\mu(A)<0$
and applying Theorem 2.1 we readily check that
$0\leq\mathbb{E}\left(U_{t}(Y)\right)-\mathbb{E}\left(\widehat{U}_{t}(Y)\right)=\frac{\mu(S)}{2}\leavevmode\nobreak\
\int_{0}^{t}\mathbb{E}\left(\|m_{s}-\widehat{X}_{s}\|^{2}\right)\leavevmode\nobreak\
ds\leq\frac{\mathsf{C}t}{N}$
which is the bias, but is not of particular interest. So we will focus on the
conditional bias
$\mathbb{E}\left(\left[\widehat{U}_{t}(Y)-U_{t}(Y)\right]^{n}\right)^{1/n}$.
In the case (14), recall (17). In this context, using the generalized
Minkowski inequality and under the assumptions of Theorem 2.3 we find that
$\sup_{t\geq
0}{\mathbb{E}\left(\left|\widehat{U}_{t}(Y)-U_{t}(Y)\right]^{n}\right|^{1/n}}\leq\frac{\mathsf{C}(n)}{N}$
where $\mathsf{C}(n)<\infty$ is a constant that does not depend upon $N$.
For the cases of (12) and (13), we start with the fact that the conditional
bias in (27) is decomposed into two terms
$\alpha_{t}:=\int_{0}^{t}\langle
C(\widehat{m}_{t}-\widehat{X}_{s}),R_{2}^{-1}(dY_{s}-C\widehat{X}_{s}ds)\rangle\quad\mbox{\rm
and}\quad\beta_{t}:=\frac{1}{2}\leavevmode\nobreak\
\int_{0}^{t}\mathbb{E}\left(\langle(m_{s}-\widehat{X}_{s}),S(m_{s}-\widehat{X}_{s})\rangle\leavevmode\nobreak\
|\leavevmode\nobreak\ {\cal F}_{s}\right)\leavevmode\nobreak\ ds.$
Using the uniform estimates presented in Section 2.3 we have
$\mathbb{E}\left(\left|\beta_{t}\right|^{n}\right)^{1/n}\leq\frac{\mathsf{C}(n)t}{N}$
for some deterministic constants $\mathsf{C}(n)$ that do not depend on the
time horizon. On the other hand, combining the Burkholder-Davis-Gundy
inequality with the uniform bias estimate (19) we have the rather crude
estimate
$\mathbb{E}\left(\left|\alpha_{t}\right|^{n}\right)\leq\mathsf{C}(n)\left(\frac{t}{N}\right)^{n/2}$
(28)
for some deterministic constants $\mathsf{C}(n)$ that do not depend on the
time horizon. Arguing as for the proof of Theorem 2.4, we check that
$\mathbb{E}\left(\left|\alpha_{t}\right|^{n}\right)\leq\mathsf{C}(n)\left(\frac{\sqrt{t}}{N}\right)^{n}.$
Observe that the above improves the crude estimate stated in (28). This yields
the following corollary.
###### Corollary 3.1.
Consider the cases of (12) and (13). Assume that (16) holds and that
$\mu(A)<0$. Then for any $n\geq 1$, $t\geq 0$ and $N$ sufficiently large, we
have that
$\mathbb{E}\left(\left|\widehat{U}_{t}(Y)-U_{t}(Y)\right|^{n}\right)^{1/n}\leq\frac{\mathsf{C}(n)(t+\sqrt{t})}{N}$
where $\mathsf{C}(n)<\infty$ is a constant that does not depend upon $t$ or
$N$.
### 3.4 $\mathbb{L}_{n}-$Error
We now consider the $\mathbb{L}_{n}-$error of the estimate of the log-of the
normalizing constant $U_{t}^{N}(Y)=\log(\overline{Z}_{t}^{N}(Y)).$
In the case of (12) and (13) we have the following.
###### Proposition 3.1.
Consider the cases of (12) and (13). Assume that (16) holds and that
$\mu(A)<0$. Then for any $n\geq 1$, $t\geq 0$ and $N$ sufficiently large, we
have that
$\mathbb{E}\left(\left|U_{t}^{N}(Y)-U_{t}(Y)\right|^{n}\right)^{1/n}\leq\mathsf{C}(n)\Big{(}\sqrt{\frac{t}{N}}+\frac{t}{N}\Big{)}$
where $\mathsf{C}(n)<\infty$ is a constant that does not depend upon $t$ or
$N$.
###### Proof.
We have that
$U_{t}^{N}(Y)-U_{t}(Y)=\int_{0}^{t}\Big{[}\langle
C(m_{s}-\widehat{X}_{s}),R_{2}^{-1}(dY_{s}-C\widehat{X}_{s}ds)\rangle-\frac{1}{2}\langle(m_{s}-\widehat{X}_{s}),S(m_{s}-\widehat{X}_{s})\rangle\leavevmode\nobreak\
ds\Big{]}.$
So one can apply the Minkowski inequality to yield that
$\mathbb{E}\left(\left|U_{t}^{N}(Y)-U_{t}(Y)\right|^{n}\right)^{1/n}\leq$
$\mathbb{E}\left(\Big{|}\int_{0}^{t}\langle
C(m_{s}-\widehat{X}_{s}),R_{2}^{-1}(dY_{s}-C\widehat{X}_{s}ds)\rangle
ds\Big{|}^{n}\right)^{1/n}+\frac{1}{2}\leavevmode\nobreak\
\mathbb{E}\left(\Big{|}\int_{0}^{t}\langle(m_{s}-\widehat{X}_{s}),S(m_{s}-\widehat{X}_{s})\rangle
ds\Big{|}^{n}\right)^{1/n}.$ (29)
For the left term on the R.H.S. of (29) one can use the Burkholder-Gundy-Davis
inequality along with Theorem 2.1 to yield that
$\mathbb{E}\left(\Big{|}\int_{0}^{t}\langle
C(m_{s}-\widehat{X}_{s}),R_{2}^{-1}(dY_{s}-C\widehat{X}_{s}ds)\rangle
ds\Big{|}^{n}\right)^{1/n}\leq\frac{\mathsf{C}(n)t^{\frac{1}{2}}}{\sqrt{N}}$
for some $\mathsf{C}<\infty$ a constant that does not depend upon $t$ or $N$.
For the right term on the R.H.S. of (29) one can use the generalized Minkowski
inequality and Theorem 2.1 to give
$\mathbb{E}\left(\Big{|}\int_{0}^{t}\langle(m_{s}-\widehat{X}_{s}),S(m_{s}-\widehat{X}_{s})\rangle
ds\Big{|}^{n}\right)^{1/n}\leq\frac{\mathsf{C}(n)t}{N}$
for some $\mathsf{C}(n)<\infty$ a constant that does not depend upon $t$ or
$N$. The proof is now easily concluded. ∎
In the case of (13) we can refine further:
###### Proposition 3.2.
Consider the case of (13). Assume that $\mu(A-P_{\infty}S)<0$ and
$S\in\mathbb{S}_{r_{1}}^{+}$. Then for any $N\geq 2$ there exists a $t(N)>0$
with $t(N)\rightarrow\infty$ as $N\rightarrow\infty$, such that for any
$t\in[0,t(N)]$
$\mathbb{E}\left(\left|U_{t}^{N}(Y)-U_{t}(Y)\right|^{n}\right)^{1/n}\leq\mathsf{C}(n)\Big{(}\sqrt{\frac{t}{N}}+\frac{t}{N}\Big{)}$
where $\mathsf{C}(n)<\infty$ is a constant that does not depend upon $t$ or
$N$.
###### Proof.
The proof is essentially as Proposition 3.1 except one should use Theorem 2.2
instead of Theorem 2.1. ∎
In the case of (14) we have the following, whose proof is again similar to the
above results.
###### Proposition 3.3.
Consider the case of (14) . Assume that [3, eq. (2.7)] holds. Then for any
$t\geq 0$ and $N\geq r_{1}$, we have that
$\mathbb{E}\left(\left|U_{t}^{N}(Y)-U_{t}(Y)\right|^{n}\right)^{1/n}\leq\frac{\mathsf{C}(n)}{\sqrt{N}}$
where $\mathsf{C}(n)<\infty$ is a constant that does not depend upon $t$ or
$N$.
Both Proposition 3.1 and Proposition 3.2 establish that one can estimate the
log of the normalization constant using the ensemble Kalman-Bucy type filters,
with a mean square error (for instance) that grows at most linearly in time.
Proposition 3.3 provides a uniform in time error, mostly following from the
deterministic nature of the algorithm and the dependence on standard Kalman-
Bucy theory.
One can say more, when considering the average estimation error in over a
window of time $w$ as we now state. We restrict ourselves to the mean square
error below. In the cases of both (12) and (13) we have the time-uniform
upper-bound.
###### Corollary 3.2.
Consider the case of (12) and (13). Assume that (16) holds and that
$\mu(A)<0$. Then for any $t>0$, $0<w<t$ and $N$ sufficiently large, we have
that
$\mathbb{E}\Bigg{[}\Bigg{(}\frac{1}{w}\Bigg{\\{}\log\Bigg{(}\frac{\overline{Z}_{t}^{N}(Y)}{\overline{Z}_{t-w}^{N}(Y)}\Bigg{)}-\log\Bigg{(}\frac{\overline{Z}_{t}(Y)}{\overline{Z}_{t-w}(Y)}\Bigg{)}\Bigg{\\}}\Bigg{)}^{2}\Bigg{]}\leq\frac{\mathsf{C}}{N}$
where $\mathsf{C}<\infty$ is a constant that does not depend upon $t$, $w$ or
$N$.
In the case of (13) we refine further and have the following time-uniform
upper-bound.
###### Corollary 3.3.
Consider the cases of (13). Assume that $\mu(A-P_{\infty}S)<0$ and
$S\in\mathbb{S}_{r_{1}}^{+}$. Then for any $N\geq 2$ there exists a $t(N)>0$
with $t(N)\rightarrow\infty$ as $N\rightarrow\infty$, such that for any
$t\in(0,t(N)]$, $0<w<t$
$\mathbb{E}\Bigg{[}\Bigg{(}\frac{1}{w}\Bigg{\\{}\log\Bigg{(}\frac{\overline{Z}_{t}^{N}(Y)}{\overline{Z}_{t-w}^{N}(Y)}\Bigg{)}-\log\Bigg{(}\frac{\overline{Z}_{t}(Y)}{\overline{Z}_{t-w}(Y)}\Bigg{)}\Bigg{\\}}\Bigg{)}^{2}\Bigg{]}\leq\frac{\mathsf{C}}{N}$
where $\mathsf{C}<\infty$ is a constant that does not depend upon $t$, $w$ or
$N$.
Note that, in Corollary 3.3 as we require $t\in(0,t(N)]$ and
$t(N)=\mathcal{O}(\log(N))$ this may not be as useful as Corollary 3.2, but
the assumption of a stable system in Corollary 3.2 is much stronger than the
hypotheses in Corollary 3.3; see [3] for a discussion.
### 3.5 Simulation Results
For $(i,k,L)\in\\{1,\cdots,N\\}\times\mathbb{N}_{0}\times\mathbb{N}_{0}$, let
$\Delta_{L}=2^{-L}$ and consider the Euler-discretization of (12)-(14):
$\begin{array}[]{lccl}(\textbf{F1})&\xi_{(k+1)\Delta_{L}}^{i}&=&\xi_{k\Delta_{L}}^{i}+A\xi_{k\Delta_{L}}^{i}\Delta_{L}+R_{1}^{1/2}\big{\\{}\overline{W}_{(k+1)\Delta_{L}}^{i}-\overline{W}_{k\Delta_{L}}^{i}\big{\\}}+\\\
&&&p_{k\Delta_{L}}C^{\prime}R_{2}^{-1}\Big{(}\big{\\{}Y_{(k+1)\Delta_{L}}-Y_{k\Delta_{L}}\big{\\}}-\Big{[}C\xi_{k\Delta_{L}}^{i}\Delta_{L}+R_{2}^{1/2}\big{\\{}\overline{V}_{(k+1)\Delta_{L}}^{i}-\overline{V}_{k\Delta_{L}}^{i}\big{\\}}\Big{]}\Big{)}\\\
(\textbf{F2})&\xi_{(k+1)\Delta_{L}}^{i}&=&\xi_{k\Delta_{L}}^{i}+A\xi_{k\Delta_{L}}^{i}\Delta_{L}+R_{1}^{1/2}\big{\\{}\overline{W}_{(k+1)\Delta_{L}}^{i}-\overline{W}_{k\Delta_{L}}^{i}\big{\\}}+\\\
&&&p_{k\Delta_{L}}C^{\prime}R_{2}^{-1}\left(\big{\\{}Y_{(k+1)\Delta_{L}}-Y_{k\Delta_{L}}\big{\\}}-C\left(\dfrac{\xi_{k\Delta_{L}}^{i}+m_{k\Delta_{L}}}{2}\right)\Delta_{L}\right)\\\
(\textbf{F3})&\xi_{(k+1)\Delta_{L}}^{i}&=&\xi_{k\Delta_{L}}^{i}+A\xi_{k\Delta_{L}}^{i}\Delta_{L}+R_{1}\left(p_{k\Delta_{L}}\right)^{-1}\left(\xi_{k\Delta_{L}}^{i}-m_{k\Delta_{L}}\right)\Delta_{L}+\\\
&&&p_{k\Delta_{L}}C^{\prime}R_{2}^{-1}\left(\big{\\{}Y_{(k+1)\Delta_{L}}-Y_{k\Delta_{L}}\big{\\}}-C\left(\dfrac{\xi_{k\Delta_{L}}^{i}+m_{k\Delta_{L}}}{2}\right)\Delta_{L}\right)\end{array}$
(30)
and the discretization of $\overline{Z}_{T}^{N}(Y)$:
$\displaystyle\overline{Z}_{T}^{N,L}(Y)=\exp\left\\{\sum_{k=0}^{T/\Delta_{L}-1}\left\langle
m_{k\Delta_{L}},C^{\prime}R_{2}^{-1}\big{[}Y_{(k+1)\Delta_{L}}-Y_{k\Delta_{L}}\big{]}\right\rangle-\frac{1}{2}\left\langle
m_{k\Delta_{L}},S\,m_{k\Delta_{L}}\right\rangle\Delta_{L}\right\\}.$ (31)
For the purpose of showing that the mean square error of the estimate of the
log of the normalization constant in the cases F1 & F2 is of
$\mathcal{O}(\frac{t}{N})$ and in the case F3 is of
$\mathcal{O}(\frac{1}{N})$, we take $r_{1}=r_{2}=1$, $A=-2$, $R_{1}^{-1/2}=1$,
$R_{2}^{-1/2}=2$, $C$ a uniform random number in $(0,1]$ and
$\xi_{0}^{i}\overset{i.i.d.}{\sim}\mathcal{N}(0.5,0.2)$ (normal distribution
mean 0.5 and variance 0.2). In Tables 1 \- 2 and Figures 1 \- 2, we show that
the rate of the MSE of the estimate in (31) for the cases F1 & F2 is of
$\mathcal{O}(\frac{t}{N})$, even though we have used a naive time
discretization for our results. In Table 3 and Figure 3 we show that the rate
in the F3 case is of $\mathcal{O}(1/N)$.
| $N=1000$ | $N=500$ | $N=250$
---|---|---|---
t | MSE | MSE$/(t/N)$ | MSE | MSE$/(t/N)$ | MSE | MSE$/(t/N)$
50 | 2.194E-04 | 4.4E-03 | 5.404E-04 | 5.4E-03 | 8.903E-04 | 4.5E-03
100 | 5.483E-04 | 5.5E-03 | 1.442E-03 | 7.2E-03 | 3.082E-03 | 7.7E-03
200 | 1.085E-03 | 5.4E-03 | 2.901E-03 | 7.3E-03 | 4.442E-03 | 5.6E-03
400 | 2.283E-03 | 5.7E-03 | 5.047E-03 | 6.3E-03 | 1.017E-02 | 6.4E-03
800 | 4.894E-03 | 6.1E-03 | 8.579E-03 | 5.4E-03 | 1.985E-02 | 6.2E-03
1600 | 9.716E-03 | 6.1E-03 | 2.039E-02 | 6.4E-03 | 2.904E-02 | 4.5E-03
3200 | 1.974E-02 | 6.2E-03 | 3.504E-02 | 5.5E-03 | 7.835E-02 | 6.1E-03
6400 | 3.571E-02 | 5.6E-03 | 7.087E-02 | 5.5E-03 | 1.599E-01 | 6.2E-03
Table 1: The mean square error (MSE) and MSE$/(t/N)$ for $N\in\\{250,500,1000\\}$ in the F1 case. | $N=1000$ | $N=500$ | $N=250$
---|---|---|---
t | MSE | MSE$/(t/N)$ | MSE | MSE$/(t/N)$ | MSE | MSE$/(t/N)$
50 | 1.137E-03 | 5.7E-03 | 5.127E-04 | 5.1E-03 | 3.317E-04 | 6.6E-03
100 | 2.157E-03 | 5.4E-03 | 1.128E-03 | 5.6E-03 | 8.892E-04 | 8.9E-03
200 | 4.601E-03 | 5.8E-03 | 2.708E-03 | 6.8E-03 | 1.260E-03 | 6.3E-03
400 | 9.777E-03 | 6.1E-03 | 5.250E-03 | 6.6E-03 | 2.436E-03 | 6.1E-03
800 | 1.979E-02 | 6.2E-03 | 9.949E-03 | 6.2E-03 | 4.639E-03 | 5.8E-03
1600 | 4.901E-02 | 7.7E-03 | 1.595E-02 | 5.0E-03 | 8.895E-03 | 5.6E-03
3200 | 9.593E-02 | 7.5E-03 | 3.001E-02 | 4.7E-03 | 1.925E-02 | 6.0E-03
6400 | 1.744E-01 | 6.8E-03 | 5.415E-02 | 4.2E-03 | 3.635E-02 | 5.7E-03
Table 2: The MSE and MSE$/(t/N)$ for $N\in\\{250,500,1000\\}$ in the F2 case. $N$ | MSE | MSE$\times N$
---|---|---
50 | 1.073E-05 | 5.4E-04
100 | 4.951E-06 | 5.0E-04
200 | 2.867E-06 | 5.7E-04
400 | 1.492E-06 | 6.0E-04
800 | 8.157E-07 | 6.5E-04
1600 | 3.119E-07 | 5.0E-04
3200 | 1.773E-07 | 5.7E-04
6400 | 6.344E-08 | 4.1E-04
Table 3: The MSE and MSE$\times N$ for $t=100$ in the F3 case. Figure 1: The
mean square error associated to the EnKBF in the F1 case. This plots the MSE
against the time parameter for $N\in\\{250,500,1000\\}$. MSE $\approx$ 5.9E-03
$\left(\frac{t}{N}\right)$, $\approx$ 6.1E-03 $\left(\frac{t}{N}\right)$,
$\approx$ 5.6E-03 $\left(\frac{t}{N}\right)$, for $N=250,500,1000$,
respectively. Figure 2: The mean square error associated to the EnKBF in the
F2 case. This plots the MSE against the time parameter for
$N\in\\{250,500,1000\\}$. MSE $\approx$ 6.4E-03 $\left(\frac{t}{N}\right)$,
$\approx$ 5.5E-03 $\left(\frac{t}{N}\right)$, $\approx$ 6.4E-03
$\left(\frac{t}{N}\right)$, for $N=250,500,1000$, respectively. Figure 3: The
mean square error associated to the EnKBF in the F3 case. This plots the MSE
against the ensemble size $N$ for fixed time $t=100$. MSE $\approx$
5.4E-04$/N$.
## 4 Application to Static Parameter Estimation
### 4.1 Approach
We now assume that there are a collection of unknown parameters,
$\theta\in\Theta\subseteq\mathbb{R}^{d_{\theta}}$, associated to the model
(1). For instance $\theta$ could consist of some or all of the elements of
$A,C,R_{1},R_{2}$. We will then include an additional subscript $\theta$ in
each of the mathematical objects that have been introduced in the previous two
sections. As an example, we would write
$Z_{t,\theta}(X,Y)=\exp{\left[\int_{0}^{t}\left[\langle
CX_{s},R_{2}^{-1}dY_{s}\rangle-\frac{1}{2}\langle
X_{s},SX_{s}\rangle\right]\leavevmode\nobreak\ ds\right]}.$
For $0<s<t$ we introduce the notation
$\overline{Z}_{s,t,\theta}(Y):=\frac{\overline{Z}_{t,\theta}(Y)}{\overline{Z}_{s,\theta}(Y)}$
with the obvious extension to
$\overline{Z}_{s,t,\theta}^{N}(Y):=\overline{Z}_{t,\theta}^{N}(Y)/\overline{Z}_{s,\theta}^{N}(Y)$.
The basic idea of our approach is to consider the recursive estimation of
$\theta$ on the basis of the arriving observations. In particular, for
notational convenience, we shall consider a method which will update our
current estimate of $\theta$ at unit times. Our objective is to follow a
recursive maximum likelihood (RML) method (e.g. [11]) which is based upon the
following update at any time $t\in\mathbb{N}$
$\theta_{t}=\theta_{t-1}+\kappa_{t}\nabla_{\theta}\log\Big{(}\overline{Z}_{t-1,t,\theta}(Y)\Big{)}\Big{|}_{\theta=\theta_{t-1}}$
where $\\{\kappa_{t}\\}_{t\in\mathbb{N}}$ is a sequence of real numbers with
$\kappa_{t}>0$, $\sum_{t\in\mathbb{N}}\kappa_{t}=\infty$,
$\sum_{t\in\mathbb{N}}\kappa_{t}^{2}<\infty$. Computing the gradient of
$\overline{Z}_{t-1,t,\theta}(Y)$ can be computationally expensive, so our
approach is to use a type of finite difference estimator via SPSA. We note
that a classical finite difference estimator would require $2d_{\theta}$
evaluations of $\overline{Z}_{t-1,t,\theta}(Y)$, whereas the SPSA method keeps
this evaluation down to 2; as computational cost is a concern, we prefer this
afore-mentioned method.
Our approach is the following, noting that we will use the argument $(k)$ to
denote the $k^{th}-$element of a vector.
1. 1.
Initialization: Set an initial $\theta_{0}\in\Theta$ and choose two step sizes
$\\{\kappa_{t}\\}_{t\in\mathbb{N}}$ and $\\{\nu_{t}\\}_{t\in\mathbb{N}}$ such
that $\kappa_{t}>0$, $\kappa_{t},\nu_{t}\rightarrow 0$,
$\sum_{t\in\mathbb{N}}\kappa_{t}=\infty$,
$\sum_{t\in\mathbb{N}}\frac{\kappa_{t}^{2}}{\nu_{t}^{2}}<\infty$. Set $t=1$
and generate i.i.d. the initial ensemble from $\eta_{0,\theta_{0}}$.
2. 2.
Iteration:
* •
For $k\in\\{1,\dots,d_{\theta}\\}$, independently, sample $\Delta_{t}(k)$ from
a Bernoulli distribution with success probability $1/2$ and support
$\\{-1,1\\}$.
* •
Set $\theta_{t-1}^{+}=\theta_{t-1}+\nu_{t}\Delta_{t}$,
$\theta_{t-1}^{+}=\theta_{t-1}-\nu_{t}\Delta_{t}$.
* •
Using the EnKBF, with samples simulated under $\theta_{t-1}$, generate
estimates $\overline{Z}_{t-1,t,\theta_{t-1}^{+}}^{N}(Y)$ and
$\overline{Z}_{t-1,t,\theta_{t-1}^{-}}^{N}(Y)$.
* •
Set, for $k\in\\{1,\dots,d_{\theta}\\}$,
$\theta_{t}(k)=\theta_{t-1}(k)+\kappa_{t}\frac{1}{2\nu_{t}\Delta_{t}(k)}\log\Bigg{(}\frac{\overline{Z}_{t-1,t,\theta_{t-1}^{+}}^{N}(Y)}{\overline{Z}_{t-1,t,\theta_{t-1}^{-}}^{N}(Y)}\Bigg{)}.$
* •
Run the EnKBF from $t-1$ (starting with samples simulated under
$\theta_{t-1}$) up-to time $t$ using the parameter $\theta_{t}$. Set $t=t+1$
and return to 2..
We note that in practice, one must run a time-discretization of the EnKBF,
rather than the EnKBF itself. We remark also that the use of SPSA for
parameter estimation associated to hidden Markov models has been considered
previously, for instance in [14].
### 4.2 Simulation Results
We use the algorithm described in the previous section along with the Euler-
discretization in (30) to estimate the parameters in three different models.
In particular, we show that the algorithm works in both linear and non-linear
models. In all three models, the data is generated from the true parameters.
#### 4.2.1 Linear Gaussian Model
In the first example, we take $A=\theta_{1}Id$, $R_{1}^{-1/2}=\theta_{2}R$,
where
$\displaystyle R=\begin{bmatrix}1&0.5\\\ 0.5&1&0.5\\\ &0.5&\ddots&\ddots\\\
&&\ddots&\ddots&0.5\\\ &&&0.5&1\end{bmatrix}$
$C=\alpha_{1}(r_{1},r_{2})C^{*}$, where $C^{*}$ is a uniform random matrix and
$\alpha_{1}(r_{1},r_{2})$ is a constant, $R_{2}^{-1/2}=\alpha_{2}(r_{2})Id$,
where $\alpha_{2}(r_{2})$ is a constant. In Figures 4 \- 9, we show the
results for the parameters estimation of $\theta_{1}$ and $\theta_{2}$ in the
cases $r_{1}=r_{2}=2$ and $r_{1}=r_{2}=100$. In all cases we take $N=100$
except in the case when $r_{1}=r_{2}=100$ in F3, where we take $N=200$ to
assure the invertibility of $p_{t}^{N}$, otherwise, the condition number of
$p_{t}^{N}$ will be huge. The discretization level is $L=8$ in all cases. The
initial state is $X_{0}\sim\mathcal{N}(4\mathbf{1}_{r_{1}},Id)$, where
$\mathbf{1}_{r_{1}}$ is a vector of 1’s in $\mathbb{R}^{r_{1}}$.
The results display that in a reasonable case, one can estimate low-
dimensional parameters using RML via SPSA. We now consider a few nonlinear
models.
(a) (b)
Figure 4: The blue curves along with their average (in red) are trajectories
from the execution of the algorithm in Section 4.1 for the estimation of
$(\theta_{1},\theta_{2})$ in the case F1 with $r_{1}=r_{2}=2$. The initial
values of the parameters are $(-1,2)$. The green horizontal lines represent
the true parameter values $(\theta_{1}^{*},\theta_{2}^{*})=(-2,1)$. We take
$\nu_{t}=t^{-0.1}$ and $\kappa_{t}=0.09$ when $t\leq 400$ and
$\kappa_{t}=3\times t^{-0.601}$ otherwise.
(a) (b)
Figure 5: The blue curves along with their average (in red) are trajectories
from the execution of the algorithm in Section 4.1 for the estimation of
$(\theta_{1},\theta_{2})$ in the case F1 with $r_{1}=r_{2}=100$. The initial
values of the parameters are $(-1,2)$. The green horizontal lines represent
the true parameter values $(\theta_{1}^{*},\theta_{2}^{*})=(-2,1)$. We take
$\nu_{t}=t^{-0.1}$ and $\kappa_{t}=0.1$ when $t\leq 400$ and
$\kappa_{t}=3\times t^{-0.601}$ otherwise.
(a) (b)
Figure 6: The blue curves along with their average (in red) are trajectories
from the execution of the algorithm in Section 4.1 for the estimation of
$(\theta_{1},\theta_{2})$ in the case F2 with $r_{1}=r_{2}=2$. The initial
values of the parameters are $(-1,2)$. The green horizontal lines represent
the true parameter values $(\theta_{1}^{*},\theta_{2}^{*})=(-2,1)$. We take
$\nu_{t}=t^{-0.1}$ and $\kappa_{t}=0.09$ when $t\leq 300$ and
$\kappa_{t}=t^{-0.7}$ otherwise.
(a) (b)
Figure 7: The blue curves along with their average (in red) are trajectories
from the execution of the algorithm in Section 4.1 for the estimation of
$(\theta_{1},\theta_{2})$ in the case F2 with $r_{1}=r_{2}=100$. The initial
values of the parameters are $(-1,2)$. The green horizontal lines represent
the true parameter values $(\theta_{1}^{*},\theta_{2}^{*})=(-2,1)$. We take
$\nu_{t}=t^{-0.1}$ and $\kappa_{t}=0.08$ when $t\leq 400$ and
$\kappa_{t}=3\times t^{-0.64}$ otherwise.
(a) (b)
Figure 8: The blue curves along with their average (in red) are trajectories
from the execution of the algorithm in Section 4.1 for the estimation of
$(\theta_{1},\theta_{2})$ in the case F3 with $r_{1}=r_{2}=2$. The initial
values of the parameters are $(-1,2.2)$. The green horizontal lines represent
the true parameter values $(\theta_{1}^{*},\theta_{2}^{*})=(-2,1)$. We take
$\nu_{t}=t^{-0.1}$ and $\kappa_{t}=0.09$ when $t\leq 100$ and
$\kappa_{t}=t^{-0.8}$ otherwise.
(a) (b)
Figure 9: The blue curves along with their average (in red) are trajectories
from the execution of the algorithm in Section 4.1 for the estimation of
$(\theta_{1},\theta_{2})$ in the case F3 with $r_{1}=r_{2}=100$. The initial
values of the parameters are $(-1,2)$. The green horizontal lines represent
the true parameter values $(\theta_{1}^{*},\theta_{2}^{*})=(-2,1)$. We take
$\nu_{t}=t^{-0.1}$ and $\kappa_{t}=0.09$ when $t\leq 200$ and
$\kappa_{t}=2\times t^{-0.601}$ otherwise.
#### 4.2.2 Lorenz’ 63 Model
We now consider the following nonlinear model with $r_{1}=r_{2}=3$, which is a
simplified mathematical model for atmospheric convection.
$\begin{array}[]{rcl}dX_{t}&=&f(X_{t})\leavevmode\nobreak\
dt\leavevmode\nobreak\ +\leavevmode\nobreak\ R^{1/2}_{1}\leavevmode\nobreak\
dW_{t}\\\ dY_{t}&=&C\leavevmode\nobreak\ X_{t}\leavevmode\nobreak\
dt\leavevmode\nobreak\ +\leavevmode\nobreak\ R^{1/2}_{2}\leavevmode\nobreak\
dV_{t},\end{array}$
where
$\displaystyle f_{1}(X_{t})$ $\displaystyle=\theta_{1}(X_{t}(2)-X_{t}(1)),$
$\displaystyle f_{2}(X_{t})$
$\displaystyle=\theta_{2}X_{t}(1)-X_{t}(2)-X_{t}(1)X_{t}(3),$ $\displaystyle
f_{3}(X_{t})$ $\displaystyle=X_{t}(1)X_{t}(2)-\theta_{3}X_{t}(3),$
where $X_{t}(i)$ is the $i^{th}$ component of $X_{t}$. We also have
$R_{1}^{1/2}=Id$,
$\displaystyle C_{ij}=\left\\{\begin{array}[]{cl}\frac{1}{2}&\text{if }i=j\\\
\frac{1}{2}&\text{if }i=j-1\\\ 0&\text{otherwise }\end{array}\right.,\qquad
i,j\in\\{1,2,3\\},$
and
$(R_{2}^{1/2})_{ij}=2\,q\left(\frac{2}{5}\min\\{|i-j|,r_{2}-|i-j|\\}\right)$,
$i,j\in\\{1,2,3\\}$, where
$\displaystyle
q(x)=\left\\{\begin{array}[]{cl}1-\frac{3}{2}x+\frac{1}{2}x^{3}&\text{if
}0\leq x\leq 1\\\ 0&\text{otherwise }\end{array}\right..$
In Figures 10 \- 12, we show the results for the parameters estimation of
$\theta=(\theta_{1},\theta_{2},\theta_{3})$ in the F1, F2 and F3 cases. The
ensemble size is $N=100$ and the discretization level is $L=8$ in all cases.
The initial state is $X_{0}\sim\mathcal{N}(\mathbf{1}_{r_{1}},\frac{1}{2}Id)$,
where $\mathbf{1}_{r_{1}}$ is a vector of 1’s in $\mathbb{R}^{r_{1}}$.
(a) (b) (c)
Figure 10: The blue curves along with their average (in red) are trajectories
from the execution of the algorithm in Section 4.1 for the estimation of
$(\theta_{1},\theta_{2},\theta_{3})$ in the case F1. The initial values of the
parameters are $(7.5,26.7,6.5)$. The green horizontal lines represent the true
parameter values
$(\theta_{1}^{*},\theta_{2}^{*},\theta_{3}^{*})=(10,28,\frac{8}{3})$. We take
$\nu_{t}=t^{-0.2}$ an $\kappa_{t}=0.0314$ when $t\leq 100$ and
$\kappa_{t}=t^{-0.71}$ otherwise.
(a) (b) (c)
Figure 11: The blue curves along with their average (in red) are trajectories
from the execution of the algorithm in Section 4.1 for the estimation of
$(\theta_{1},\theta_{2},\theta_{3})$ in the case F2. The initial values of the
parameters are $(7.5,26.7,6.5)$. The green horizontal lines represent the true
parameter values
$(\theta_{1}^{*},\theta_{2}^{*},\theta_{3}^{*})=(10,28,\frac{8}{3})$. We take
$\nu_{t}=t^{-0.2}$ and $\kappa_{t}=0.0139$ when $t\leq 300$ and
$\kappa_{t}=t^{-0.75}$ otherwise.
(a) (b) (c)
Figure 12: The blue curves along with their average (in red) are trajectories
from the execution of the algorithm in Section 4.1 for the estimation of
$(\theta_{1},\theta_{2},\theta_{3})$ in the case F3. The initial values of the
parameters are $(7.5,26.7,6.5)$. The green horizontal lines represent the true
parameter values
$(\theta_{1}^{*},\theta_{2}^{*},\theta_{3}^{*})=(10,28,\frac{8}{3})$. We take
$\nu_{t}=t^{-0.2}$ and $\kappa_{t}=0.0314$ when $t\leq 100$ and
$\kappa_{t}=t^{-0.71}$ otherwise.
#### 4.2.3 Lorenz’ 96 Model
Finally, we consider the following nonlinear model, Lorenz’ 96, with
$r_{1}=r_{2}=40$. The solution of this model has a chaotic behavior and it
describes the evolution of a scalar quantity on a circle of latitude.
$\begin{array}[]{rcl}dX_{t}&=&f(X_{t})\leavevmode\nobreak\
dt\leavevmode\nobreak\ +\leavevmode\nobreak\ R^{1/2}_{1}\leavevmode\nobreak\
dW_{t}\\\ dY_{t}&=&X_{t}\leavevmode\nobreak\ dt\leavevmode\nobreak\
+\leavevmode\nobreak\ R^{1/2}_{2}\leavevmode\nobreak\ dV_{t},\end{array}$
where
$\displaystyle f_{i}(X_{t})$
$\displaystyle=(X_{t}(i+1)-X_{t}(i-2))X_{t}(i-1)-X_{t}(i)+\theta$
where $X_{t}(i)$ is the $i^{th}$ component of $X_{t}$, and it is assumed that
$X_{t}(-1)=X_{t}(r_{1}-1)$, $X_{t}(0)=X_{t}(r_{1})$ and
$X_{t}(r_{1}+1)=X_{t}(1)$. We also have $R_{1}^{1/2}=\sqrt{2}Id$ and
$R_{2}^{1/2}=\frac{1}{2}Id$. $\theta$ here represents the external force in
the system, while $(X_{t}(i+1)-X_{t}(i-2))X_{t}(i-1)$ is the advection-like
term and $-X_{t}(i)$ is the damping term. In Figure 13, we show the results
for the parameters estimation of $\theta$ in the F1, F2 and F3 cases. The
ensemble size is $N=100$ and the discretization level is $L=8$ in all cases.
In F1 & F2 cases, $X_{t}$ is initialized as follows: $X_{0}(1)=8.01$ and
$X_{0}(k)=8$ for $1<k\leq 40$. In F3, to avoid having the matrix $p_{0}^{N}$
equal to zero, we take $X_{0}\sim\mathcal{N}(8\mathbf{1}_{r_{1}},0.05Id)$.
(a) (b) (c)
Figure 13: The blue curves along with their average (in red) are trajectories
from the execution of the algorithm in Section 4.1 for the estimation of
$\theta$ in the case F1 (left), F2 (middle) and F3 (right). The initial value
of $\theta$ is 10. The green horizontal lines represent the true parameter
value $\theta^{*}=8$. We take $\nu_{t}=t^{-0.1}$ and $\kappa_{t}=0.0314$ when
$t\leq 50$ and $\kappa_{t}=t^{-0.75}$ otherwise.
#### Acknowldegements
AJ & HR were supported by KAUST baseline funding. DC has been partially
supported by EU project STUOD - DLV-856408.
## References
* [1] Allen, J. I., Eknes, M. & Evensen, G. (2002). An Ensemble Kalman filter with a complex marine ecosystem model: Hindcasting phytoplankton in the Cretan Sea. _Ann. Geo-phys._ , 20, 1–13.
* [2] Bain, A. & Crisan, D. (2009).Fundamentals of Stochastic Filtering. Springer, New York.
* [3] Bishop, A. N. & Del Moral, P. (2020). On the mathematical theory of ensemble (linear-gaussian) Kalman–Bucy filtering. arXiv preprint arXiv:2006.08843.
* [4] Bishop, A. N. & Del Moral, P. (2020). On the stability of matrix-valued Riccati diffusions. _Electron. J. Probab._ , 24, 1–40.
* [5] Bishop, A. N. & Del Moral, P. (2017). On the stability of Kalman–Bucy diffusion processes. SIAM J. Control Optim., 55, 4015-4047.
* [6] Bishop, A. N. & Del Moral, P. (2019). Stability Properties of Systems of Linear Stochastic Differential Equations with Random Coefficients. SIAM J. Control Optim., 57, 1023–1042.
* [7] Cerou, F. Del Moral, P & Guyader, A. (2011) A non-asymptotic variance theorem for un-normalized Feynman- Kac particle models. _Ann. Inst. Henri Poincare_ , 4, 629–649.
* [8] Del Moral, P. & Tugaut, J. (2018). On the stability and the uniform propagation of chaos properties of ensemble Kalman–Bucy filters. Ann. Appl. Probab., 28, 790–850.
* [9] Drovandi, C., Everitt, R., Golightly, A. & Prangle, D. (2019). Ensemble MCMC: Accelerating Pseudo-Marginal MCMC for State Space Models using the Ensemble Kalman Filter. arXiv preprint arXiv:1906.02014v2.
* [10] Evensen, G. (1994). Sequential data assimilation with a non-linear quasi-geostrophic model using Monte Carlo methods to forecast error statistics. _J. Geophys. Res._ , 99, 10143–10162.
* [11] Le Gland, F. & Mevel, M. (1997). Recursive identification in hidden Markov models. _Proc. 36th IEEE Conf. Dec. Contr._ , 3468-3473.
* [12] Le Gland, F., Monbet, V. & Tran, V.-D. (2011). Large sample asymptotics for the ensemble Kalman filter. In _The Oxford Handbook of Nonlinear Filtering_ , Oxford Univ. Press, Oxford, 598–631.
* [13] Karatzas, I. & Shreve, S. (1991). _Brownian Motion and Stochastic Calculus_. Second edition, Springer: New York.
* [14] Poyiadjis, G, Singh, S.S. & Doucet, A. (2006) Gradient-free maximum likelihood parameter estimation with particle filters. _Proc Amer. Control Conf._ , 6–9.
* [15] Reich, S. & Cotter C. (2013). Ensemble filter techniques for intermittent data assimilation. In _Large Scale Inverse Problems: Computational Methods and Applications in the Earth Sciences_ (eds: M. Cullen, M.A. Freitag, S. Kindermann, R. Scheichl), 91–134. De Gruyter Publishers.
* [16] Sakov, P. & Oke, P. R. (2008). A deterministic formulation of the ensemble Kalman filter: an alternative to ensemble square root filters. _Tellus A._ , 60, 361–371.
* [17] Skjervheim, J -A., Evensen, G., Aanonsen, S. I., Ruud, B. O. & Johansen, T. A. (2005). Incorporating 4D seismic data in reservoir simulation models using ensemble Kalman filter. _SPE_ , 12, 95789.
* [18] Spall, J. C. (1992) Multivariate stochastic approximation using a simultaneous perturbation gradient approximation. _IEEE Trans. Auto. Control._ 37, 332-341.
* [19] Tong, X. T., Majda, A. J. & Kelly, D. (2016). Nonlinear stability and ergodicity of ensemble based Kalman filters. _Nonlinearity_ , 29, 657–691.
|
Machine learning with limited data Doctor of Philosophy Biological Sciences
iFlyTek-Surrey Joint Research Centre on Artificial Intelligence, UK
August 27, 2024
###### Abstract
Abstract Thanks to the availability of powerful computing resources, big data
and deep learning algorithms, we have made great progress on computer vision
in the last few years. Computer vision systems begin to surpass humans in some
tasks, such as object recognition, object detection, face recognition and pose
estimation. Lots of computer vision algorithms have been deployed to real
world applications and started to improve our life quality. However, big data
and labels are not always available. Sometimes we only have very limited
labeled data, such as medical images which requires experts to label them. In
this paper, we study few shot image classification, in which we only have very
few labeled data.
Machine learning with little data is a big challenge. To tackle this
challenge, we propose two methods and test their effectiveness thoroughly. One
method is to augment image features by mixing the style of these images. The
second method is applying spatial attention to explore the relations between
patches of images.
We also find that domain shift is a critical issue in few shot learning when
the training domain and testing domain are different. So we propose a more
realistic cross-domain few-shot learning with unlabeled data setting, in which
some unlabeled data is available in the target domain. We propose two methods
in this setting. Our first method transfers the style information of the
unlabeled target dataset to the samples in the source dataset and trains a
model with stylized images and original images. Our second method proposes a
unified framework to fully utilize all the data. Both of our methods surpass
the baseline method by a large margin.
Key words: few shot learning, image classification, domain shift
###### Contents
1. 0 Introduction
1. 1 Introduction
2. 2 Structure of the Confirmation Report
3. 3 Novel Work Undertaken
2. 1 Literature review
1. 1 Definition
2. 2 Relation with other types of machine learning
3. 3 Overview of few shot learning
3. 2 Methods for Few Shot Learning
1. 1 Style mix for few shot learning
1. 1 Background
2. 2 Approach
3. 3 Experiments
2. 2 Spatial attention for few shot learning
1. 1 Background
2. 2 Approach
3. 3 Experiments
3. 3 Problems of few shot learning methods and remaining challenges
4. 3 Cross-domain Few shot learning with unlabeled data
1. 1 Background
1. 1 Domain shift in few shot learning
2. 2 Cross domain few shot learning
2. 2 Cross domain few shot learning with unlabeled data
1. 1 Definition
2. 2 Related work
3. 3 Style transfer for cross-domain few-shot learning with unlabeled data
1. 1 Background
2. 2 Approach
3. 3 Experiments
4. 4 Discussion
4. 4 Contrastive learning for cross-domain few-shot learning with unlabeled data
1. 1 Background
2. 2 Approach
3. 3 Experiment
4. 4 Discussion
5. 4 Future directions and open challenges
###### List of Figures
1. 1 data augmentation ([10])
2. 2 style transfer ([27])
3. 3 TSNE plots of features
4. 4 Attention ([2])
5. 1 Domain shift in few shot learning ([7])
6. 2 style transfer ([29])
7. 3 stylized images ([21])
8. 4 DomainNet ([37])
###### List of Tables
1. 1 Results of style mix
2. 2 Result of style mix in different places
3. 3 Hyperparameter tuning for $\alpha$ and p
4. 4 Results of style mix when inserting style mix layers during testing
5. 5 Results of spatial attention
6. 1 Results of style transfer
7. 2 Results of contrastive learning
## Chapter 0 Introduction
### 1 Introduction
Nowadays, it’s well known that deep learning is powerful. We use it to solve
lots of problems in computer vision, natural language processing, and speech
processing. However, we also know that deep learning models have a tremendous
number of parameters and consume enormous amount of data. For example, an
image recognition model, AlexNet([32]), was trained on ImageNet([11]), which
has millions of images. BERT([12]), with 340 million parameters, used
BooksCorpus (800M words) (Zhu et al.,2015) and English Wikipedia (2,500M
words). GPT-3 ([3]), a recently developed natural language model, has 175
billion parameters. Thus a large number of memory cards, storage drives, and
GPUs are consumed. One consequence is that training these large models
consumes lots of energy and produces tons of carbon dioxide, which is not
environmentally friendly.
This poses a huge challenge for us. Then can we learn a model with little
data? For example, what if you are asked to recognize dogs and cats with only
one image labeled as dog and one image labeled as cat? From a statistical
point of view, this is almost impossible. Rules can only be derived from a
large amount of samples. But we all know human beings have remarkable ability
to learn from little data. For instance, suppose you are given an image of an
apple, then you can tell apples of slightly different shape and different
color and texture from oranges even though you only see one example of apples.
To address this problem, in recent years, some researchers start to study this
few shot learning problem, which is learning with only a very limited amount
of data. In this setting, we are allowed to transfer knowledge from another
additional dataset and test our trained models on a small dataset. Most few-
shot learning methods are also meta learning methods, in which the algorithm
tries to learn to learn the model. So it’s called learning to learn, which is
a bi-level learning. There are also some non meta-learning based methods, one
of which is transfer learning. Using transfer learning, we train a model on
the additional dataset and then fine tune it on our target dataset which only
has limited data.
Most of these methods can be categorized as 4 different types of methods:
optimization based method, metric learning based methods, data augmentation
based methods and parameter generation based methods (or black box / model-
based methods). Optimization-based method is the most intuitive meta learning
method in which the model tries to optimize the inner optimization problem.
The inner optimization tasks are subtasks. By doing this, we hope the model
can learn some meta knowledge for all sub-tasks and then generalize to unseen
tasks during testing. Metric learning based methods try to find a feature
space where all sub-tasks can perform well, i.e., images in the same class are
closer in the feature space while images in different classes stay further.
Data augmentation based methods deal with the data deficiency by learning to
augment the data during meta-training on the meta-training dataset and
augmenting the novel dataset during meta-testing. The Parameter generation
methods, also called black box or model-based methods, try to generate
parameters of sub-tasks from a meta-learner (the learner in the outside
level). During training, the model takes samples from a task as input and
saves knowledge in the activation states (parameters) in the model. During
testing, the model outputs parameters for a task conditioned on the labeled
training data and makes predictions using these parameters.
We introduce 2 methods from different perspectives. Our first style mix method
borrows ideas from the literature of image style transfer and then tries to
augment the data in the feature space by mixing the style from different
images. Doing so, we hope to mitigate the problem of scarce data in few shot
learning. Inspired by ([51]), we also try to use attention mechanisms to
address few shot image classification. In the method, we let the query images
explicitly attend to important parts of images in the support set. The spatial
relations among different parts of image are utilized explicitly, which is not
the case in ResNet ([25]) used by most other methods.
We carefully design and conduct our experiments and thoroughly verify our
ideas above. But the improvement is not impressive enough. We then find out
the reason why: most of meta learning methods don’t work well for few shot
classification. We do lots of experiments and gather lots of evidence from
recently published papers. A paper [Rapid learning or feature reuse] claims
that feature reuse is the reason why most optimization-based methods work, by
conducting intensive experiments. We also find that pretraining the feature
extractor is vital to achieve high accuracy for most methods. Further meta
learning only gives 1 or 2 percent of accuracy improvement. Another factor we
found is non-episodic training, such as Baseline++ ([7]) and SimpleShot
([52]), can achieve almost the same accuracy as episodic training, which is
the common practise in few shot learning. Additionally, deeper models can also
achieve better accuracies ([7]). And if there is a domain shift between the
training set and the testing set, the accuracy will drop significantly ([7]).
All these evidences show that the current meta-learning methods don’t work
well for few shot image classification and can’t help much. Later on, after we
found this phenomenon, a series of papers, including ([45]; [13]) were
published, verifying this idea. All of these show that meta learning reached
its bottleneck and we need some fundamental and theoretical breakthroughs. And
it may take several decades to see these breakthroughs.
We then shift research focus to related problems, cross domain few shot
learning and cross-domain few shot learning with unlabeled data. We observe
that models trained on base dataset perform badly when there is a big domain
shift between the base dataset and the novel dataset. Thus it’s important to
study cross-domain few shot learning, in which we have one or multiple
datasets from different domains as training datasets and we will test our
model on another different domain. But this is a too challenging task since
during training we are not able to touch the target testing domain. So a new
and more realistic setting for few shot learning is proposed and studied. We
provide some unlabeled target data for training. In most cases, unlabeled data
are easy to obtain. For example, we can have tons of chest X-ray images
collected from hospitals when they do body checks every day for patients. But
collecting labels (canner or not) is not easy since they require medical
expertise, which is usually only owned by doctors. So we propose a new
setting, in which there is some unlabeled data in the target domain.
We introduce two methods we propose to tackle this problem. Our first method
transfers domain information to the source labeled dataset by style transfer
and then trains a model using the stylized images and the images in the
original source dataset. We also propose a pseudo-label method to solve this
problem: clustering for all unlabeled data and then co-trained the model using
both labled source domain data, clustered target domain data and unclustered
target data with a contrastive loss. We achieve big accuracy improvements
compared to our baseline for these two methods. We also propose another
research idea, which combines domain translation based methods and pseudo-
label learning methods. We are implementing this and hopefully we get more
improvements.
In the future,we hope to keep exploring cross-domain few shot learning with or
without unlabeled data. Also we plan to study compositional and causal models,
which are critical abilities when there are only very little data. We hope
that by developing compostial and causal models, we can really get rid of the
need of big data and build extremely efficient models. Also compositial and
causal learning are the ways to general artificial intelligence, in my
opinion.
### 2 Structure of the Confirmation Report
In chapter 2, we give a detailed survey about this field. We first give a
formal definition of few shot learning. Then we show a taxonomy and an
overview of previous methods in literature. Details of important papers in
this field are also discussed. In the end, we point out problems and
challenges for researchers.
In chapter 3, we introduce two algorithms we propose and experiment results
and analysis of these two methods. Since we only have very limited data, we
first propose a method to increase features. We then show how it works and
experiment results of it. Inspired by ([51]),we propose a method which fully
exploits the spatial relations between the query image and gallery images. We
present the experimental results of this method.
In chapter 4, we show the necessities to study cross-domain few shot learning
and present a new machine learning setting, cross-domain few shot learning
with unlabelled data. We claim this is a more realistic and important setting.
Several methods are proposed and show excellent results. We will give some
preliminary results and show our future directions.
Finally, we discuss open challenges in this field and our future directions in
chapter 6. Cross-domain few shot learning needs to be further explored. Models
with compositional and causal learning should be proposed to address the data
scarcity problem in few-shot learning.
### 3 Novel Work Undertaken
We propose 2 new methods for the conventional few shot learning and
systematically experimental their effectiveness.
We point out the problem in the development of meta learning when no one in
the research community noticed this.
We propose a new and more realistic cross domain few shot learning setting,
which is cross-domain few shot learning with unlabeled data. We develop two
methods in this setting and design and conduct experiments for it. Our methods
surpass the baseline methods with a large margin.
## Chapter 1 Literature review
### 1 Definition
In this chapter we give a detailed introduction to few shot learning. Most of
the time, few shot learning means few shot classification. We start from a
formal definition.
Conventional supervised machine learning In the conventional supervised
machine learning, we have a dataset $D=\\{X_{j},Y_{j}\\}_{j=1..N}$, where
$\\{X_{j},Y_{j}\\}$ is a pair of a data point and it’s label. In the image
classification problem, we $X_{j}$ is an image and $Y_{j}$ is a category. D
can be split into a training set $D_{train}$ and a testing set $D_{test}$. The
loss function can be cross-entropy loss. Then we optimize the loss with
respect to the parameter theta:
$\theta^{*}=\underset{\theta}{argmin}\;L(D;\theta)$
Few shot learning In few shot learning, our dataset D is very small (less than
30 samples per class in classification problem). In most cases, we only study
the case where there are only less than 10 samples per class. This is an ill-
posed problem and almost impossible to achieve high performance without
utilizing another dataset based on our current understanding of machine
learning. So we are allowed to use another dataset in the hope that we can
learn some meta-knowledge from it and transfer it to the small dataset. The
additional dataset is called the auxiliary dataset or the base dataset.
Most methods in literature address few shot learning problems as meta learning
problems. Here, let’s formally define meta learning: We have a dataset D,
splitted to a training set (called meta-training set or base dataset)
$D_{base}$ and testing set (called meta-testing set or novel dataset)
$D_{novel}$. We sample lots of few shot learning tasks from D and assume all
tasks have a task distribution $p(\tau)$. During training (also called meta-
training), we sample some few shot learning tasks [$D_{train}$, $D_{test}$]
from $D_{base}$. $D_{train}$ and $D_{test}$ are training and testing sets of a
few shot learning tasks. They are also called support sets and query sets
which are different from meta-training and meta-testing sets. Our learning
goal is to minimize a loss function $L(D_{base},w)$. L measures the
performance of the model on a series of tasks. $\omega$ is the parameter of
the model, also called meta-parameters which represent meta-knowledge from all
sampled tasks:
$\omega^{*}=\underset{\omega}{argmin}\;E_{\tau\sim p(\tau)}(D;\omega)$
During the meta-testing (also called fine-tuning in some papers) stage, we use
the meta-knowledge learned during meta-training to learn a model for each task
sampled from $D_{novel}$. $\theta$ is the parameter for each few shot learning
task:
$\theta^{*}=\underset{\theta}{argmax}\;log\;p(\theta|\omega^{*},D^{train(i)}_{novel})$
Finally, we evaluate our algorithm on $D_{novel}^{test(i)}$. Note that
$D_{novel}^{train(i)}$ and $D_{novel}^{test(i)}$ are the support and query set
of the $i^{th}$ task sampled from the novel dataset.
Meta learning tries to learn meta-knowledge from a series of few short tasks
sampled from the auxiliary dataset and hope the model can perform well to
novel tasks which have different classes. In conventional machine learning, we
assume there is a data distribution. In meta learning, we assume there is a
task distribution. All tasks are sampled from the same task distribution, so
if the model fits the tasks sampled from the auxiliary dataset well, it can
also generalize to tasks which have new categories. The new categories are
never seen by the model during meta-training, so performing well during meta-
testing requires generalization.
The above definition is from a task-distribution point of view. There is also
another point of view: bi-level optimization point of view ([26]), especially
for optimization based methods. This kind of method treats learning to
learning as two levels of optimization problem: in the inner loop, you
optimize the model for the specific few shot learning problem and get optimal
parameters $\theta$ while in the outer loop you optimize the mata-model for
all few shot tasks to get the optimal meta-parameters $\omega$. Note that
$\theta$ is conditional on the outer-loop parameters $\omega$. In all
optimization based papers, the most famous one is MAML([17]). In MAML, meta-
parameter $\omega$ is the initial value of $\theta$.
$\omega^{*}=\underset{\omega}{argmin}\sum_{i=1}^{M}L^{meta}(\theta^{*(i)}(\omega)),\omega,D^{test(i)}_{base})$
$s.t.\;\theta^{*(i)}(w)=\underset{\theta}{argmin}\;L^{task}(\theta|\omega,D^{train(i)}_{base})$
Relation between few shot learning and meta-learning Meta-learning is just one
way to deal with few shot learning and it treats few shot learning as a ‘data
sample’. There are also other ways to tackle few shot learning, such as
transfer learning, such as GPT-3([3]).
### 2 Relation with other types of machine learning
One/Zero shot learning When the number of samples in each class is one, it’s
called one shot learning. When it’s zero, it’s called zero shot learning. One
shot learning is always studied in few shot learning while zero shot learning
is not the same. In few shot learning, we leverage the auxiliary dataset while
in zero shot learning, the model relies on attributes or semantic information
of those classes to transfer some supervision signals and make learning
feasible ([53]). There is a chance to borrow some methods from zero-shot
learning for Few-shot learning.
Transfer learning Transfer learning tries to transfer knowledge learned from
another dataset to the target dataset. In the current setting of few shot
classification, it can be viewed as a special case of transfer learning, we
transfer knowledge from base dataset to novel dataset.
Multitask learning multitask learning aims to learn several related tasks
together. This is a more efficient way of learning so that we can save time
and computing resources. In addition, this should also be one feature of
general artificial intelligence. Meta learning can also be viewed as a special
case of multitask learning in which all tasks are sampled from the same task
distribution, which means all tasks share the same statistics. Also during
testing, it’s tested on novel tasks. It’s not the case for regular multitask
learning.
Semi-supervised learning Semi-supervised learning has both labeled data and
unlabeled data while in the classic few shot learning, we only have very
limited labeled data. The amount of both labeled and unlabeled data, which
semi-supervised data can use, is much more than than few shot learning. There
is an intersection for these two types of learning: semi-supervised few shot
learning.
### 3 Overview of few shot learning
In this section, we will discuss the taxonomy of few shot learning methods and
methods under each category. There are four types of methods: optimization
based methods, metric learning based methods, parameter generation based
methods (or black box / model-based methods), and data augmentation based
methods.
Optimization based methods These methods treat inner-level tasks as
optimization problems and try to extract meta-knowledge for inner tasks
([26]). This can also be viewed as learning good global initial points for
inner-level tasks so that in the inner loop, after a few steps of gradient
descent, the model can quickly adapt to an unseen task. Rapid learning is what
we need in mata learning and also avoids overfitting. The most famous one is
MAML ([17]), in which in the inner loop, you update the model parameters theta
a few steps using $D_{base}^{train}$ and in the outer loop, you update the
original model, before being updated by the inner loop, using loss computed on
$D_{base}^{test}$. Because MAML differentiates through the optimization
process, we need to compute second order derivatives. Then the vanilla MAML
algorithm has a great computing cost when there are more than one steps of
gradient descent in the inner loop. To mitigate this problem, the author also
proposes an approximation method, FOMAML, which drops the second order
derivatives and only keeps the first order derivatives. ([40]) further
proposes a method which only depends on the solution of the inner loop
optimization not the path of it, which enables the inner model to update
infinite steps without worrying about the computation cost. ([36]) presents a
method, called Reptile, which keeps sampling tasks and then trains the model
on them with a few steps of gradient descent and then moves the meta-
parameters towards the updated parameters. LSTM meta learner ([41]) models the
meta-learner as a LSTM. It takes sampled tasks as input and keeps updating the
model for T rounds. Their contribution is representing the optimization of the
inner task as evolution of LSTM cell states. MetaOptNet ([33]) uses linear
classifiers and SVMs for the base leaner in the inner loop. Since it has a
closed form solution, it doesn’t have the high computation cost problem as
MAML.
Metric learning based methods try to ‘learn to compare’ (compare samples from
query set to the samples from the support set) and obtain a metric function
under which data samples in the query and support set stay closer if they are
from the same class and stay further if they are from different classes.
$P_{\theta}(y|x,S)=\sum_{(x_{i},y_{i})\in S}k_{\theta}(x,x_{i})y_{i}$
The probability of label y of sample x in the query set is the weighted sum of
all the labels in the support set and these weights are generated by a kernel
function meaning the distance between the sample x and another sample x’ in
the support set. ([31]) first use a siamese neural network to deal with this
problem. They solve this task as a verification problem in which a pair of
images go through the same embedding network and the L1 distance is optimized
to be small if they are from the same class. Matching Networks ([50]), shows
the kernel function can be an attention kernel which is the normalized softmax
of cosine similarities between samples from the query and the support set. The
feature embedding functions for the support and query set are LSTMs. Instead
of fixed metric functions, such as L1, L2 or cosine distance , Relation
Network ([44]) shows that learned metric functions parametrized by neural
networks are also possible. Prototypical Networks ([43]) is a simple metric
learning method in which the kernel function is the L2 distance between the
query sample and the centroids of all features of each class.
Model based methods don’t have an explicit form of the base learner. The meta
learner gets samples of each task as input and encodes the information in the
internal states of the model. Based on these states, the model makes
predictions for the query set. ([42]) use external memory for encoding and
retrieving task information. Because the model is only exposed to the label of
$x_{t}$ at time t+1, the model is forced to predict the label of $x_{t}$ given
data before time t. ([34]) uses temporal convolution to gather information for
past experience (tasks) and soft attention to pinpoint specific information
for the current tasks for prediction. Meta networks ([35]) has meta learner
and base learner. The meta-learner is trained to generate parameters (called
fast weights in the paper) using LSTM for itself and for the base learner.
([22]) proposes to leverage an attention base classification weight generator
to generate weights for those classes in the base learner. The base learner is
a cosine distance classifier. Different from other papers, the author claims
that their methods not only perform well on novel classes but also on base
classes. Imprinted weights ([38]) has the similar idea that the weight of
novel classes are generated using the normalized averaged features of samples
in the support set.
Data augmentation based methods mitigate the data scarcity problem by
augmenting the data, which is an intuitive strategy. Sometimes, this method
can be integrated with other types of methods, such as metric learning based
methods. Some of these methods learn a data generator using the base dataset
and generate data for novel classes during meta-testing. ([24]) aims to
augment the data using appearance variance in the base dataset. Style transfer
can also be leveraged to augment the data for few shot learning ([1]). [54]
integrates data generators with meta learning methods directly in one
framework. ([56]) uses a pretrained saliency model to segment the foreground
and background and then combine these foregrounds and backgrounds to get new
images. ([8]) has an image deformation sub-network which takes a pair of
images as input and one for keeping the content and another for diversifying
the deformations. The output images will be used to train the model.
## Chapter 2 Methods for Few Shot Learning
In this chapter we present two methods we propose for few shot learning from
different perspectives. And then we show experiments we conduct and
conclusions we draw from these experiments. Our first method, style mix,
borrowing ideas from style transfer, augments the data in the feature space.
Then we present a method exploring the relationship between tasks during
training. We apply knowledge distillation techniques for two tasks sharing the
same label space and domain adaptation methods for two tasks having disjoint
label space. Inspired by the attention mechanism proposed by Non-local neural
networks ([51]), we explore this idea in few shot learning: ask the images in
the query set to attend to pixels in the images in the support set. We will
introduce each of them in the following sections.
### 1 Style mix for few shot learning
#### 1 Background
Data augmentation is widely used in machine learning and deep learning. When
you have only very little data, overfitting is almost unavoidable. To increase
the dataset without labeling, the easiest way is to augment the dataset.
Common augmentation techniques including geometric and color transformation,
such as relation, random cropping, random easing, horizontal and vertical
flipping, noise injection, as shown in figure 3.1. These are basic techniques
and later researchers also developed more sophisticaled techniques, such as
feature space augmentation, adversarial training, and GAN data augmentation.
Recently some automatically learned augmentation methods are proposed, such as
AutoAugment ([9]) which search optimal combinations of augmentation techniques
as hyper-parameter searching. The principle of data augmentation is to
preserve the semantic identity (labels) while modifying other aspects using
some human known prior information. For example, we know rotating images will
not alter the label of the objects in the image. Some methods apply too much
transformation and the resulting images make less sense to humans.
Figure 1: data augmentation ([10])
Mixup is a regulation technique to overcome overfitting in deep learning,
first proposed by ([57]). As shown in the formulas,
$\tilde{y}=\lambda
y_{i}+(1-\lambda)y_{j},\;where\;y_{i},y_{j}\;are\;raw\;input\;vectors$ (1)
$\tilde{x}=\lambda
x_{i}+(1-\lambda)x_{j},\;where\;x_{i},x_{j}\;are\;raw\;input\;vectors$ (2)
we get the linear combination of a pair of samples and their corresponding
labels. Then $\tilde{x}$ and $\tilde{y}$ are used as training samples. This
regularization forces models to favor those with linear behaviors of samples.
Although it is quite simple, this has been proved to be very effective. Later,
in ([49]), instead of using linear combination of samples, it uses linear
combinations of hidden representation of samples. They show that this can
improve representation by obtaining smoother decision boundaries and fully
utilizing hidden representation which contains high level information.
Style transfer is a task in image processing in which the artistic styles of
images are required to change but the content should be kept, as shown in
figure 3.2. ([19]) first studies how to use CNN to deal with this problem. In
this work, it shows the feature response of a pretrained convolutional neural
network can be used as content and the feature summary statistics contains
style information. Then by recombining content and style information, we can
create new images with different styles. ([15]) shows that a conditional
install normalization layer can be used to transfer the styles:
$CIN(x,s)=\gamma^{s}(\frac{x-\mu(x)}{\sigma(x)})+\beta^{s}$ (3)
in which the feature x is first normalized by its mean and standard deviation.
And then two parameters gamma and beta are learned for the style s. Then
([29]) modifies this, and uses the mean and standard deviation of the style
image directly without learning those parameters:
$AdaIN(x,y)=\gamma(y)(\frac{x-\mu(x)}{\sigma(x)})+\mu(y)$ (4)
in which $\sigma$ and $\mu$ are computed across spatial locations of the style
image. With this modification, we can replace the style of image x with the
style of image y. This method can produce images of arbitrary styles in real
time, so it’s much better compared to the previous method.
Figure 2: style transfer ([27])
#### 2 Approach
Inspired by Adain ([29]) and manifold mixup ([49]), we propose a method to
increase the diversity of the features for few shot learning. Because we have
very limited data in shot learning, it’s important to augment the data. We
propose to mix styles of different images in the same task and in two
different tasks. In the first case, during meta learning, for each image, we
randomly select another image. Then we compute the mean and standard deviation
for these two images across the spatial locations:
$\sigma_{nc}(x)=\sqrt{\frac{1}{HW}\sum_{h=1}^{H}\sum_{w=1}^{W}(x_{nchw}-\mu_{nc}(x))^{2}+\epsilon}$
(5) $\mu_{nc}(x)=\frac{1}{HW}\sum_{h=1}^{H}\sum_{w=1}^{W}x_{nchw}$ (6)
in which H, W are the height and width of the feature map x. After we get the
mean, $\mu$, and standard deviation, $\sigma$ for the two images x and y, we
mix the style (styled information is contained in the mean and standard
deviation) of them:
$\mu=\lambda*\mu_{x}+(1-\lambda)*\mu_{y}$ (7)
$\sigma=\lambda*\sigma_{x}+(1-\lambda)*\sigma_{y}$ (8)
$\tilde{x}=\sigma(\frac{x-\mu(x)}{\sigma(x)})+\mu$ (9)
in which $\lambda$ is the mixing coefficient, sampled from a Beta
distribution, which is proposed in ([57]). Notice in the formula 3.9, we first
get the normalized feature without style info and then we scale it with a
mixed standard deviation and shift it with the mixed mean. After this
operation, the feature will have the mixed style of the images. In this way we
hope we get features of diverse styles and we can overcome the data scarcity
problem in few-shot learning. The algorithm is shown in Algorithm 1.
Result: Classification accuracy
while _Traning_ do
Sample a task T of size n;
while _$n_{image}$ <n_ do
Sample two images $x$ and $y$ from T;
Get feature maps for $x$ and $y$, f(x) and f(y);
Obtain $\mu_{x}$ and $\mu_{y}$ using formula 3.5 and 3.6;
Obtain the feature map $\tilde{x}$ with mixed style of image x and y using
formula 3.7, 3.8 and 3.9;
$n_{image}$ = $n_{image}$ +1
Calcuate the few shot classfication loss and accuracy using $\tilde{x}$ ;
Update the model
Algorithm 1 style mix
#### 3 Experiments
Dataset We benmark our algorithm on MiniImageNet, which is widely used as a
benchmark in few-shot learning. MiniImagenet is a subset of ImageNet ([11]),
which has 100 classes in total and 600 images per class. 1000 classes are
divided into 64 classes for training, 16 classes for validation and 20 classes
for testing. This is a small data but has enough clases and images per class.
Thus, it is a perfect testbed for a few-shot learning algorithm. we sample
5-way 1-shot (5 clases and 1 image per class in the query set and 15 images in
the support set) tasks and 5-way 5-shot (5 classes and 5 images per class in
the query set and 15 images in the support set) tasks for training and
testing. During testing ,we average our result over 2000 episodes, while
earlier work only use 600 episodes, which has high variance and is not
trustable.
Backbone We adopt ResNet12 ([25]) here for the feature extractor and
Prototypical networks ([43]) for few-shot learning. Earlier work uses Conv4
(Convolutional neural works with only 4 layers) and later on, Resnet12,
Resnet18 and wideresnet28-10 are used. Conv4 is too small, has underfit the
dataset easily and WideResnet28-10 has too many parameters. We compare
Resnet12, Resnet18 and WideResnet and found Resnet12 is enough.
Style mix implementation We implement a stylemix layer in which we perform
style mixing within each episode or two episodes. This layer can be inserted
after each ResNet block and we tested several combinations.
Training our training has two steps: pretraining on all 64 clases and episodic
training using prototypical networks with style mix. PreTraining is widely
used in few-shot learning and has been proven to be vital for the final
performance.
Result We give the averaged classification accuracy on the testing set over
2000 episodes in table 3.1.
method | 5-way 5-shot accuracy
---|---
Baseline | 78.34
ProtoNet | 79.26
Style mix (after 1st Resnet blcok) | 78.62
Table 1: Results of style mix
In this table, our method achieves comparable results with Protonet. We
conduct thorough hyperparameter tuning and detailed analysis trying to figure
out why it doesn’t work.
Hyperparameter tuning First we study where we should put the style mix layers
and how many style mix layers we should put in the feature extractor. There
are 3 possible locations to insert the style mix layer and we also try
inserting multiple style mix layers. The result is shown in table 3.2. We can
see that the deeper and more the style mix layers are, the worse performance
we get.
method | 5-way 5-shot accuracy
---|---
Baseline | 78.34
ProtoNet | 79.26
Style mix (after 1st Resnet blcok) | 78.62
Style mix (after 2st Resnet blcok) | 78.47
Style mix (after 3st Resnet blcok) | 77.67
Style mix (after all 4 Resnet blcoks) | 74.35
Table 2: Result of style mix in different places
Then we tune two hyperparameters in our method (table 3.3), $\alpha$ and p.
$\alpha$ is a parameter in the Beta distribution, which controls the shape of
the distribution. When $\alpha$ gets bigger, we mix the style more. p is the
probability of mixing. When p is larger, we mix the styles more. One
conclusion can be drawn that no matter how we fine tune these parameters,
style mix can not beat the baseline and ProtoNet.
| $\alpha=0.1$ | $\alpha=0.1$ | $\alpha=0.1$ | $\alpha=0.1$
---|---|---|---|---
p=0 | | 79.26 | |
p=0.2 | | 79.37 | |
p=0.5 | 78.87 | 79.26 | 79.37 | 78.60
p=0.8 | | 78.97 | |
Table 3: Hyperparameter tuning for $\alpha$ and p
Analysis Then we perform some analysis trying to figure out how our method
doesn’t work well as we expect. We insert style mix layers only during testing
to see the effect and them. As we can see from table 3.4, the performance gets
worse when the style mix layers are inserted deeper. From the below t-SNE
embedding of feature maps (figure 3.3), we show that features start to mix
together when there is too much style mixing. So the discriminative power is
reduced. That explain why the stylization doesn’t work.
Insert styleMix during testing ONLY using trained ProtoNet, 5-shot,
MiniImageNet
---
No styleMix | After 1st block | After 2nd block | After 3rd block | After 4th block
79.26 | 74.69 | 71.46 | 66.23 | 60.61
Table 4: Results of style mix when inserting style mix layers during testing
Figure 3: TSNE plots of features
Fig 1 (right top): no StyleMix layer. Fig 2 (left bottom): 1 StyleMix after
1st ResNet block. Fig 3 (right bottom): 4 styleMix layers after each ResNet
block. Mixing probability is 1.
### 2 Spatial attention for few shot learning
#### 1 Background
Attention is all you need ([48]) first explores attention mechanism in deep
learning. Attention mechanism tries to mimic the attention mechanism in
bilgraph, in which we only focus on the most important parts. For instance, if
we look at the figure 3.4, most of us only focus or visual attention on the
dog head. Then BERT ([12]) and GPT ([3]) show it’s superior performance in
natural language processing. Non-local neural networks is the first work
applying attention mechanism in computer vision, in which a spatial attention
on feature maps is designed and shows excellent performance in image and video
processing tasks. Then vision transformer ([14]) and i-GPT ([5]) shows
transformers with attention can even beat Convolutional neural networks when a
large amount of data is available.
Figure 4: Attention ([2])
Here we briefly introduce the most basic self-attention mechanism. It tries to
improve the representation of each element in a set by aggregating information
from other elements in the same set based on the similarity with other
elements:
$Attention(Q,K,V)=softmax(\frac{QK^{T}}{\sqrt{d_{k}}})V$ (10)
The input of a transformer is a set of Q (query), K (key) and V(value). Query,
key and value are concepts from retrieval systems. Q is the element we
aggregate information for. Key and value are paired. Value is what we look for
but it’s often easier to retrieve value by key. In the above formula, we first
compute the scaled dot product of Q and K. After applying a softmax function,
the attention weight is normalized and it can be viewed as the similarity
between the query and the key. based on similarities, we aggregate values. The
final output contains information from other elements in the same set so it’s
called self attention. And in self-attention, Q=K=V.
Our work is also inspired by DeepEMD ([55]) in which image patches are used to
computing the matching cost. So the relation between image patches can be
utilized.
#### 2 Approach
We develop several variants of our method applying spatial attention in few-
shot learning. We start our introduction from the most basic one. We first
apply self attention to feature maps of images in the same class. 5 feature
maps of the shape NCHW from the same class are concatenated to the shape
NC(5H)W. Then we apply self-attention to it. Each element of the feature maps
will aggregate information from other elements of the feature maps. Doing so
we hope the output feature map becomes more information and then we get better
prototypes (averaged feature maps) for late classification. we also apply
self-attention to 5 prototypes so that our final feature maps become more
separable. The algorithm is shown in Algorithm 2.
Inspired by DeepEMD ([55]), we concatenate the feature maps of each query
image and 5 prototypes and then apply self-attention. In this way, we hope to
fully utilize the spatial correspondence among all locations of the feature
map of the query image and 5 prototypes. This is different from the original
prototypical networks ([43]) and here we focus more on local information
rather than global information and local information are richer.
Result: Classification accuracy
while _Traning_ do
Sample a s-way 5-shot task T of size;
while _$n_{class}$ <5_ do
Get five images $x_{1}$, $x_{2}$, … $x_{5}$ for the class $n_{c}lass$ from T;
Get feature maps for $x_{1}$, $x_{2}$, … $x_{5}$, $f(x_{1})$, $f(x_{2})$, …
$f(x_{5})$;
concatenate the feature maps $f(x_{1})$, $f(x_{2})$, … $f(x_{5})$ along the
asix H;
Applay self attention on the concatenated feature map f(x) using the formula
3.10;
Split the feature map after the attention to five feature maps
$f(x_{1})^{\prime}$, $f(x_{2})^{\prime}$… $f(x_{5})^{\prime}$;
$n_{class}$ = $n_{class}$ +1;
Obtain five prototypes by averaging five feature maps in each class;
Apply attention on five prototypes;
Calcuate the few shot classfication loss and accuracy using augmented
prototypes after the attention ;
Update the model
Algorithm 2 Spatial attention
#### 3 Experiments
We verify our idea on the same MiniImagenet 5-shot taks. Protonet is our
baseline. We use ResNet12 as the feature extractor as it’s large enough to fit
the dataset compared to Conv4 and less computationally intensive compared to
WideResnet-28-10. We implement an attention layer and insert it into different
places of the feature extractor, which are layers after each residual block.
The result is shown in table 3.5. We can see that no matter where the
attention is inserted, the performance doesn’t change. We will discuss the
reason in the following section.
method | 5-way 5-shot accuracy
---|---
ProtoNet | 79.97
Spatial attention after the 2nd residual block) | 79.59
Spatial attention after the 3rd residual block) | 79.55
Spatial attention after the 4th residual block) | 79.51
Spatial attention after 4 residual blocks) | 80.02
Table 5: Results of spatial attention
### 3 Problems of few shot learning methods and remaining challenges
Since this paper ([39]) was published, we realize that we have reached a
bottleneck for all few-shot learning methods. This paper gives a conclusion
based on a series of experiments that MAML works is not due to it can rapidly
adapt to new tasks and it’s because it learned good features from the base
dataset which is useful during testing for the novel dataset. Baseline++ ([7])
shows that a simple cosine classifier can outperform all the star-of-the-art
methods. ([45]) also claims that meta learning doesn’t work well for few shot
image classification and all performance boost is from the good feature
embedding. ([13]) and SimpleShot([52]) also show the similar phenomena:
feature embedding is the key in few shot image classification. All of this
mean that we haven’t got enough progress even though lots of fancy meta
learning methods are proposed. The area reaches its bottleneck and we won’t
get any real progress until there is a theoretical and funmentai breakthrough.
This could explain why our earlier methods don’t show significant improvement.
It’s hard to get any progress in the area so we transfer our focus to other
settings.
## Chapter 3 Cross-domain Few shot learning with unlabeled data
In this chapter we introduce the domain shift which exists in few-shot
learning and two settings of cross-domain few-shot learning. Domain shift is
studied in domain adaptation, domain generalization and it has been observed
that the accuracy will drop significantly if there exists a domain shift
between the base dataset and the novel dataset. To address this problem, we
introduce cross-domain few shot learning. This is a more realistic setting as
the domain shift is everywhere in our real life. Also, because we know from
the previous section that most meta-learning methods just learn useful
features, the accuracy will drop if the training and testing distribution
doesn’t match. So cross-domain few shot learning becomes an important research
problem.
But without touching the target domain where testing happens, the problems
become extremely difficult. In most cases, it’s easy to collect some unlabeled
data. So cross-domain few shot learning with unlabeled data becomes a more
realistic and feasible problem. Studying this problem will have more
applications in real life. We will give both the formal definition and related
work of cross-domain few shot learning and cross-domain few shot learning with
domain shift.
Then we will introduce two methods we propose for cross-domain few shot
learning with unlabeled data. Detailed and through experiment will be shown in
later sections.
### 1 Background
#### 1 Domain shift in few shot learning
Domain shift means the mismatch between two distributions, which is well
studied in domain adaptation and domain generalization. Domain adaptation
tries to transfer knowledge from one source domain to another domain but these
two domains share the same label space. Domain generalization tries to train a
model on some source domains and generalize to another unseen domain. The
difference between domain adaptation and domain generalization is that in
domain generalization the model never sees the target domain during training.
In baseline++ ([7]), the author shows that the classification accuracy will
drop if the domain shift between training and testing increases as shown in
the figure 4.1. In this paper the author conduct several experiments:
evaluating several few-shot learning methods, such as baseline, baseline++,
MachingNet, ProtoNet, MAML and RelationNet on CUB->CUB, MiniImageNet,
MiniImageNet->CUB. MiniImageNet->CUB means training on MiniImageNet but
testing on CUB, which is the cross-domain few shot learning. Domain shift
affects the classification accuracy so much and it is so common in real life,
so we decide to dive into cross-domain few-shot learning in the next section.
Figure 1: Domain shift in few shot learning ([7])
#### 2 Cross domain few shot learning
Definition We already discuss domain shift in the above section. Here let’s
give a formal definition for cross-domain few shot learning: In cross-domain
few shot learning, we have one or several source dataset (base dataset) which
has $C1...C_{n}$ classes and one target dataset (novel dataset) which contains
$C_{n+1}...C_{n+m}$ classes. So the source dataset does not contain classes
from the novel dataset and vice versa. During training only source dataset can
be used and during testing the target dataset is used.
We can see from the above definition that this setting requires generalization
abilities which means that the model needs to perform well not only on the
base dataset but also on the novel dataset.
Related work There is very little work been done on this topic. Feature-wise
transformation ([46]) is one of the first work on this. It tries to augment
image features using affine transformation to simulate image distributions
from different domains during training. The hyperparameters of the affine
transformation are learned using a learning-to-learn approach. In this paper,
MiniImageNet, CUB, Cars, Places and Plantae are used as datasets, which are
all natural image datasets. This means that the domain shift is not big. Then
in ([23]), the author proposes another dataset, which includes non-natural
images, such as satellite and chest X-ray images. This makes the benchmark
more realistic. The author also proposes some baseline methods in the papers.
But this setting is too challenging as it requires the model to get good
performance on unseen domain and unseen classes. So instead we put our focus
on another new setting where there are some unlabeled data in the target
domain. In this case the model can gather information about the target domain
and bridge the domain gap between the base dataset and the target dataset.
### 2 Cross domain few shot learning with unlabeled data
We propose a new setting which is not studied by other researchers before:
Cross-domain few-shot learning with unlabeled data. we add unlabeled data for
the target domain for two reasons: unlabeled data are easy to collect in most
cases and unlabeled data can bridge the domain gap between the source dataset
and the target domain.
#### 1 Definition
We have a labeled dataset from domain S (base dataset) which has classes
$C_{1}...C_{n}$ and another unlabeled dataset from another domain T containing
classes $C_{n+1}...C_{n+m}$ for training. Testing data is from domain T but
containing classes $C{n+m}...C{m+p}$. Note that the two later datasets are
from the same domain and all 3 datasets have disjoint label space.
#### 2 Related work
Since this setting is proposed by us, there is no work that has been done by
others. Related areas include heterogeneous domain adaptation where the source
and target domain don’t share the same label space, unsupervised meta learning
where the base dataset doesn’t have labels and unsupervised domain adaptation
for person/object re-identification.
Heterogeneous domain adaptation is also less studied also. ([18]) adopts a
deep clustering method for this problem. ([4]) proposes a method using a
shared space between the source and target domain. It trains the model with an
unsupervised factorisation loss and a graph-based loss.
Unsupervised meta learning Recently some researchers start to study this
problem. UMTRA ([30]) is an unsupervised meta-learning method with tasks
constructed by random sampling and augmentation. CACTUs([28]) uses clustering
to automatically construct tasks for unsupervised meta-learning.
UDA-Re-ID is called unsupervised domain adaptation for person/object
identification. In this setting, they don’t have class labels but have person
ids which can be viewed as fine-grained class labels. The person IDs in the
target and source domain are disjoint. There are two types of methods in this
field, domain translation based methods ,such as using CycleGan ([58]) to
translate images from the target domain to the source domain and then train a
model with translated images. Another type of methods is pseudo-label based
method in which clustering techniques are used to extract pseudo-labels for
the unlabeled target data and then unlabeled data with pseudo-labels is used
to train a model.
### 3 Style transfer for cross-domain few-shot learning with unlabeled data
We propose a method using style transfer. We first feed the images in the
unlabeled target dataset as style images and imges in the source dataset as
content images to a style transfer model. Then we obtain stylized images which
have the same style as the unlabeled target dataset but have same semantic
content as the source dataset. We train the stylized images and the original
images together. We will show our methods obtain superior results compared to
baselines.
#### 1 Background
Style transfer is introduced in section 3.1.1 so we will not introduce it
here. ([21]) shows that Convolutional neural networks trained on ImageNet are
heavily biased toward texture instead of shape, which is not preferred because
texture information is not the essential feature of a class. If we change the
texture of an object in an image, it should belong to the same class. But the
shape is different. Then the authors try to remove texture information and
obtain a dataset with only shape information using a style transfer model.
Then they use the stylized ImageNet to train a model. Results show that the
method is more robust to texture variance and environment changes.
#### 2 Approach
Our method is inspired by ([21]). In our case, we have a different problem:
cross-domain few-shot learning with unlabeled data. How to utilize the source
dataset and the unlabeled target dataset is a question. We come up with a
method which transfers the style of the source dataset to the style of the
unlabeled target dataset by feeding images in the unlabeled target dataset as
style images and images in source dataset as content images to a style
transfer model ([29]). The style transfer model is shown in the figure 4.2.
The stylized image looks like images in figure 4.3 if painting images are used
as style images and natural images are used as content images.
Figure 2: style transfer ([29]) Figure 3: stylized images ([21])
Why is style transfer related to this problem? We find style information carry
domain information and lots of style transfer methods share the same ideas as
domain adaptation methods. By transferring styles, we actually transfer the
domain information and bridge the domain gap between the source dataset and
the target dataset.
After style transfer we train a ProtoNet ([43]) with both the source dataset
images and the stylized images. Testing is conducted on the target testing
target dataset. The algorithm is shown in algorithm 3.
Result: Classification accuracy
Download the style transfer model from ([29]);
N = total number of images;
while _$n_{img}$ < N_ do
Sample a image $x$ from the source dataset and another image $y$ from the
unlabeled dataset;
Feed $x$ as the content image and $y$ as the style image to the model and
obtain a stylized image $z$;
$n_{img}$ = $n_{img}$ +1;
Train a ProtoNet with all original images and stylized images
Algorithm 3 Style transfer
#### 3 Experiments
Dataset We build our own dataset using DomainNet ([37]). The original dataset
contains 6 domains: Real, clipart, painting, sketch, infograph and quickdraw
and 345 classes. A quick view is in figure 4.4. Instead of using the whole
dataset, we select only a subset of the dataset. Since infograph is too noisy
and quickdraw contains too little information, we only use four domains: real,
sketch, clipart and sketch. Real is used as the source dataset and 3 others
are used as target dataset. Similar to MiniImageNet, we have 64 labeled
classes as the base dataset and 16 classes for the unlabeled target dataset
and 20 classes as the testing target dataset.
Figure 4: DomainNet ([37])
Result We still use the same dataset, MiniImagenet to evaluate our algorithm.
ResNet12 is used as the feature extractor, the same as before. we also fine-
tune the hyperparameter, style transfer coefficient, which means the degree of
style transfer. We find we get the best performance when the style transfer
coefficient is 1.0. Our result is shown in table 4.1, in which we have two
baselines: ‘Backbone trained on real without unlabeled data’ and ‘ADDA’.
‘Backbone trained on real without unlabeled data’ means we only train a
ResNet12 using the labeled source dataset without using the unlabeled target
dataset and test it on the testing dataset using mean centroid classifier.
‘ADDA’ is a method in ([47]) and we apply it for this problem. The table shows
that our stylization method outperforms all the baseline methods by a large
margin, especially on sketch, on which the improvement is 5 percent.
| clipart | painting | sketch | average
---|---|---|---|---
ADDA | 59.03 | 55.48 | 53.71 | 56.07
Stylization | 67.42 | 60.45 | 58.62 | 62.16
Table 1: Results of style transfer
#### 4 Discussion
Our stylization method has a clear motivation and also shows greater accuracy
improvement compared to baseline methods. There is only one defect: the style
transfer model used models weights trained on ImageNet, which means that our
methods take advantage of another dataset. This is a little bit controversial
because we can claim that we only used the style information and the paper,
([21]), also uses the same method. In the future, we can try to avoid using
model weights trained on ImageNet.
### 4 Contrastive learning for cross-domain few-shot learning with unlabeled
data
We also propose a method using contrastive learning, in which we use a
clustering method, DBSCAN, to get pseudo-labels and update our model using a
loss combining all images in the labeled dataset, clustered images in the
unlabeled target dataset and unclustered images. We will show that our method
obtains surprising results compared to baseline methods.
#### 1 Background
Clustering is an unsupervised learning method to group samples, which stay
close according to some distance measure, together as a cluster when we don’t
have labels for it. Common methods are k-means, DBSCAN ([16]) and Hierarchical
clustering. DBSCAN is a density-based method where samples in the high density
region are grouped together and points in the low density region are marked as
outliers.
Contrastive learning is now widely used in self-supervised learning. In
contrastive learning we try to find positive and negative samples and force
the model to obtain a small distance between a reference sample and its
positive samples and big distance between the reference and negative sample.
For example, in SimCLR ([6]), augmented images of the same image are treated
as positive samples while other images are treated as negative samples. Using
this distance supervision signal we can learn meaningful representation even
though we may not have labels. Recent self-supervised methods are going to
surpass the performance of supervised methods.
#### 2 Approach
Our method is inspired by ([20]). The original method is designed for object
re-identification and we re-purpose it for our setting. First we use the
clustering algorithm DBSCAN (k-means can also be applied) to cluster the
unlabeled target domain data. Then we have 3 sets of samples: labeled source
dataset, clustered unlabeled target dataset and unclustered samples in the
unlabeled target dataset. So methods in UDA Re-ID discard those unclustered
data but we think it’s a waste of data so we propose a unified framework to
fully utilize all the data we have:
$L_{f}=-log\frac{exp(<f,z^{+}>/\tau)}{\sum_{k=1}^{n^{s}}exp(<f,w_{k}>/\tau)+\sum_{k=1}^{n_{c}^{t}}exp(<f,c_{k}>/\tau)+\sum_{k=1}^{n_{o}^{t}}exp(<f,v_{k}>/\tau)}$
(1)
This is our loss function in which f is a feature vector of an image either
from the source dataset, clustered unlabeled target dataset or uncluster
outliers. $z^{+}$ indicates a positive prototype corresponding to $f$. $w_{k}$
is the centroid of all features in the class $k$, $c_{k}$ is the centroid of
all features in the cluster $k$ and $v_{k}$ is the $k^{th}$ outlier feature.
If f is from the source dataset, then $z^{+}=w_{k}$ is the centroid feature
vector of class k that f belongs to; if f is from the clustered dataset, then
$z^{+}=c_{k}$ is the centroid feature vector of the cluster $k$ that f belongs
to and if $f$ is from the unclusted outliers then $z^{+}=v_{k}$ is the outlier
feature that f belongs to.
Our algorithm alternates between updating the model using the above
comparative learning loss and producing pseudo-labels using clustering
methods. Detailed description is show in algorithm 4.
Result: Classification accuracy
Let N be the number of iterations for updating the model after each
clustering;
while _Traing_ do
Obtain pseudo-labels for clustered samples and unclustered samples using
DBSCAN algorithm with the unlabeled dataset;
$iter$ = 0; while _$iter$ <N_ do
$iter$ = $iter$ \+ 1 ;
Updating the model using the loss function in formula 1.1 with all labeled
samples in the source dataset, clustered samples with their pseudo-labels and
unclustered samples;
Algorithm 4 Contrastive learning
#### 3 Experiment
We first benchmark our algorithm on the same dataset as the dataset in 4.3.
Resnet12 is used as the feature extractor. we update the model 400 iterations
after each step of clustering. As clustering takes a long time (several
minutes), we should update our model for more iteration after each clustering.
Our result is shown in table 4.2:
| clipart | painting | sketch | average
---|---|---|---|---
Backbone trained on real without unlabeled data | 65.17 | 59.13 | 53.62 | 59.31
ADDA | 59.03 | 55.48 | 53.71 | 56.07
Stylization | 67.42 | 60.45 | 58.62 | 62.16
Contrastive learning | 66.18 | 58.60 | 56.38 | 60.39
Contrastive learning (enlarged unlabeled dataset) | 69.78 | 60.94 | 58.76 | 63.16
Table 2: Results of contrastive learning
We can see from the table our method surpasses the baseline on most target
dataset but the margin is not big enough. So we suspect that the unlabeled
dataset is not big enough. The original unlabeled target dataset has only
about 2-3k images, much smaller than the source dataset 20-40 k images.
Increase the size of the unlabeled dataset We increase the size of unlabeled
dataset for two reasons: First, in most cases, unlabeled dataset are easy to
build because no labels are needed and we don’t need experts to label them.
Second, most un-supervised learning methods consume more samples compared to
supervised learning methods. By increasing the size of the unlabeled dataset,
we have more room for these algorithms. Our result is shown in table 4.2 after
the size of unlabeled dataset is increased. From the table we can see that our
method shows a big improvement on accuracies compared to our baselines, which
means our method is very effective for this setting.
#### 4 Discussion
This work is still ongoing and there are lots of things to be done. For
example, the current clustering is still very slow, especially when the
unlabeled dataset is very big. We need to find a faster clustering algorithm
or reduce the number of clustering. Also we have another idea: combining
domain-translation based methods and pseudo-based methods, which will take
advantage of both methods and has a great potential to fully use the data and
surpass all the methods.
## Chapter 4 Future directions and open challenges
In the next future years of my Phd research, there are several directions I
want to explore:
Cross-domain few shot learning with unlabeled data It is still worth
exploring. This is a more realistic setting and we should carefully study this
problem and develop more algorithms for it. New research ideas can be drawn
from UDA ReID or self-supervised learning.
Compositional learning for few-shot learning Current few-shot learning methods
still heavily rely on learned features from the base dataset which does not
and will not solve the few-shot learning. In my opinion, if we can find
primitive features and then compose features from them. The data we need will
be very little. The tasks can also be composed of primitive tasks if we can
find the relations between different tasks.
Causal inference for few shot learning Causal inference is always an
interesting topic which tries to find causal factors behind data. But it’s so
different from machine learning and it has not been well studied by computer
vision research yet. If we can fully utilize causal inference models, then we
will be closer to artificial general intelligence and the data we need to
train a model will also be much less.
Analysis of common practice in few shot learning Lots of common practice are
only based on intuitions, which means they may be wrong practice. To move the
research in this field forward, we need to carefully verify them by conducting
thorough experiments or deriving theorems. For instance, why should we adopt
episodic training? Why does pretraining help?
Continual few shot learning and open set learning Continual learning means the
model needs to learn a series of tasks continuously and the model can’t forget
previous tasks. This is an important feature of artificial general
intelligence. Continual few shot learning is a excent testbed as few-shot
learning is only conducted on small dataset. Open set learning means the model
needs to deal with unseen classes during testing. This is related to meta
learning and more realistic compared to the classic supervised learning.
To sum up, we need to address several challenges before we can achieve
excellent performance on few-shot learning without the auxiliary dataset and
become closer to artificial general intelligence.
## References
* [1] Antreas Antoniou, Amos Storkey and Harrison Edwards “Data Augmentation Generative Adversarial Networks” In _arXiv preprint arXiv:1711.04340_ , 2017
* [2] “Attention? Attention!” (Accessed on 10/24/2020), https://lilianweng.github.io/lil-log/2018/06/24/attention-attention.html
* [3] Tom B Brown et al. “Language models are few-shot learners” In _arXiv preprint arXiv:2005.14165_ , 2020
* [4] Xiaobin Chang, Yongxin Yang, Tao Xiang and Timothy M. Hospedales “Disjoint Label Space Transfer Learning with Common Factorised Space” In _arXiv preprint arXiv:1812.02605_ , 2018
* [5] Mark Chen et al. “Generative Pretraining from Pixels” In _Proceedings of the 37th International Conference on Machine Learning_ , 2020
* [6] Ting Chen, Simon Kornblith, Mohammad Norouzi and Geoffrey Hinton “A Simple Framework for Contrastive Learning of Visual Representations” In _arXiv preprint arXiv:2002.05709_ , 2020
* [7] Wei-Yu Chen et al. “A Closer Look at Few-shot Classification” In _arXiv preprint arXiv:1904.04232_ , 2019
* [8] Zitian Chen et al. “Image Deformation Meta-Networks for One-Shot Learning” In _arXiv preprint arXiv:1905.11641_ , 2019
* [9] Ekin D. Cubuk et al. “AutoAugment: Learning Augmentation Policies from Data” In _arXiv preprint arXiv:1805.09501_ , 2018
* [10] “Data Augmentation: How to use Deep Learning when you have Limited Data” (Accessed on 10/23/2020), https://www.kdnuggets.com/2018/05/data-augmentation-deep-learning-limited-data.html
* [11] J. Deng et al. “ImageNet: A Large-Scale Hierarchical Image Database” In _CVPR09_ , 2009
* [12] Jacob Devlin, Ming-Wei Chang, Kenton Lee and Kristina Toutanova “Bert: Pre-training of deep bidirectional transformers for language understanding” In _arXiv preprint arXiv:1810.04805_ , 2018
* [13] Guneet S. Dhillon, Pratik Chaudhari, Avinash Ravichandran and Stefano Soatto “A Baseline for Few-Shot Image Classification” In _arXiv preprint arXiv:1909.02729_ , 2019
* [14] Alexey Dosovitskiy et al. “An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale” In _arXiv preprint arXiv:2010.11929_ , 2020
* [15] Vincent Dumoulin, Jonathon Shlens and Manjunath Kudlur “A Learned Representation For Artistic Style” In _arXiv preprint arXiv:1610.07629_ , 2016
* [16] Martin Ester, Hans-Peter Kriegel, Jörg Sander and Xiaowei Xu “A density-based algorithm for discovering clusters in large spatial databases with noise.” In _Kdd_ 96.34, 1996, pp. 226–231
* [17] Chelsea Finn, Pieter Abbeel and Sergey Levine “Model-Agnostic Meta-Learning for Fast Adaptation of Deep Networks” In _arXiv preprint arXiv:1703.03400_ , 2017
* [18] Boyan Gao, Yongxin Yang, Henry Gouk and Timothy M Hospedales “Deep Clustering for Domain Adaptation” In _ICASSP 2020-2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)_ , 2020, pp. 4247–4251 IEEE
* [19] Leon A. Gatys, Alexander S. Ecker and Matthias Bethge “A Neural Algorithm of Artistic Style” In _arXiv preprint arXiv:1508.06576_ , 2015
* [20] Yixiao Ge et al. “Self-paced Contrastive Learning with Hybrid Memory for Domain Adaptive Object Re-ID” In _arXiv preprint arXiv:2006.02713_ , 2020
* [21] Robert Geirhos et al. “ImageNet-trained CNNs are biased towards texture; increasing shape bias improves accuracy and robustness” In _arXiv preprint arXiv:1811.12231_ , 2018
* [22] Spyros Gidaris and Nikos Komodakis “Dynamic Few-Shot Visual Learning without Forgetting” In _arXiv preprint arXiv:1804.09458_ , 2018
* [23] Yunhui Guo et al. “A Broader Study of Cross-Domain Few-Shot Learning” In _arXiv preprint arXiv:1912.07200_ , 2019
* [24] Bharath Hariharan and Ross Girshick “Low-shot Visual Recognition by Shrinking and Hallucinating Features” In _arXiv preprint arXiv:1606.02819_ , 2016
* [25] Kaiming He, Xiangyu Zhang, Shaoqing Ren and Jian Sun “Deep Residual Learning for Image Recognition” In _arXiv preprint arXiv:1512.03385_ , 2015
* [26] Timothy Hospedales, Antreas Antoniou, Paul Micaelli and Amos Storkey “Meta-Learning in Neural Networks: A Survey” In _arXiv preprint arXiv:2004.05439_ , 2020
* [27] “How to style transfer your own images - GoDataDriven” (Accessed on 10/23/2020), https://godatadriven.com/blog/how-to-style-transfer-your-own-images/
* [28] Kyle Hsu, Sergey Levine and Chelsea Finn “Unsupervised Learning via Meta-Learning” In _arXiv preprint arXiv:1810.02334_ , 2018
* [29] Xun Huang and Serge Belongie “Arbitrary Style Transfer in Real-time with Adaptive Instance Normalization” In _arXiv preprint arXiv:1703.06868_ , 2017
* [30] Siavash Khodadadeh, Ladislau Bölöni and Mubarak Shah “Unsupervised Meta-Learning For Few-Shot Image Classification” In _arXiv preprint arXiv:1811.11819_ , 2018
* [31] Gregory Koch “Siamese neural networks for one-shot image recognition”, 2015
* [32] Alex Krizhevsky, Ilya Sutskever and Geoffrey E Hinton “ImageNet Classification with Deep Convolutional Neural Networks” In _Advances in Neural Information Processing Systems 25_ Curran Associates, Inc., 2012, pp. 1097–1105 URL: http://papers.nips.cc/paper/4824-imagenet-classification-with-deep-convolutional-neural-networks.pdf
* [33] Kwonjoon Lee, Subhransu Maji, Avinash Ravichandran and Stefano Soatto “Meta-Learning with Differentiable Convex Optimization” In _arXiv preprint arXiv:1904.03758_ , 2019
* [34] Nikhil Mishra, Mostafa Rohaninejad, Xi Chen and Pieter Abbeel “A Simple Neural Attentive Meta-Learner” In _arXiv preprint arXiv:1707.03141_ , 2017
* [35] Tsendsuren Munkhdalai and Hong Yu “Meta Networks” In _arXiv preprint arXiv:1703.00837_ , 2017
* [36] Alex Nichol, Joshua Achiam and John Schulman “On First-Order Meta-Learning Algorithms” In _arXiv preprint arXiv:1803.02999_ , 2018
* [37] Xingchao Peng et al. “Moment Matching for Multi-Source Domain Adaptation” In _arXiv preprint arXiv:1812.01754_ , 2018
* [38] Hang Qi, Matthew Brown and David G. Lowe “Low-Shot Learning with Imprinted Weights” In _arXiv preprint arXiv:1712.07136_ , 2017
* [39] Aniruddh Raghu, Maithra Raghu, Samy Bengio and Oriol Vinyals “Rapid Learning or Feature Reuse? Towards Understanding the Effectiveness of MAML” In _arXiv preprint arXiv:1909.09157_ , 2019
* [40] Aravind Rajeswaran, Chelsea Finn, Sham Kakade and Sergey Levine “Meta-Learning with Implicit Gradients” In _arXiv preprint arXiv:1909.04630_ , 2019
* [41] Sachin Ravi and Hugo Larochelle “Optimization as a model for few-shot learning”, 2016
* [42] Adam Santoro et al. “Meta-learning with memory-augmented neural networks” In _International conference on machine learning_ , 2016, pp. 1842–1850
* [43] Jake Snell, Kevin Swersky and Richard S. Zemel “Prototypical Networks for Few-shot Learning” In _arXiv preprint arXiv:1703.05175_ , 2017
* [44] Flood Sung et al. “Learning to Compare: Relation Network for Few-Shot Learning” In _arXiv preprint arXiv:1711.06025_ , 2017
* [45] Yonglong Tian et al. “Rethinking Few-Shot Image Classification: a Good Embedding Is All You Need?” In _arXiv preprint arXiv:2003.11539_ , 2020
* [46] Hung-Yu Tseng, Hsin-Ying Lee, Jia-Bin Huang and Ming-Hsuan Yang “Cross-Domain Few-Shot Classification via Learned Feature-Wise Transformation” In _arXiv preprint arXiv:2001.08735_ , 2020
* [47] Eric Tzeng, Judy Hoffman, Kate Saenko and Trevor Darrell “Adversarial Discriminative Domain Adaptation” In _arXiv preprint arXiv:1702.05464_ , 2017
* [48] Ashish Vaswani et al. “Attention Is All You Need” In _arXiv preprint arXiv:1706.03762_ , 2017
* [49] Vikas Verma et al. “Manifold Mixup: Better Representations by Interpolating Hidden States” In _arXiv preprint arXiv:1806.05236_ , 2018
* [50] Oriol Vinyals et al. “Matching Networks for One Shot Learning” In _arXiv preprint arXiv:1606.04080_ , 2016
* [51] Xiaolong Wang, Ross Girshick, Abhinav Gupta and Kaiming He “Non-local neural networks” In _Proceedings of the IEEE conference on computer vision and pattern recognition_ , 2018, pp. 7794–7803
* [52] Yan Wang, Wei-Lun Chao, Kilian Q. Weinberger and Laurens Maaten “SimpleShot: Revisiting Nearest-Neighbor Classification for Few-Shot Learning” In _arXiv preprint arXiv:1911.04623_ , 2019
* [53] Yaqing Wang, Quanming Yao, James Kwok and Lionel M. Ni “Generalizing from a Few Examples: A Survey on Few-Shot Learning” In _arXiv preprint arXiv:1904.05046_ , 2019
* [54] Yu-Xiong Wang, Ross Girshick, Martial Hebert and Bharath Hariharan “Low-Shot Learning from Imaginary Data” In _arXiv preprint arXiv:1801.05401_ , 2018
* [55] Chi Zhang, Yujun Cai, Guosheng Lin and Chunhua Shen “DeepEMD: Differentiable Earth Mover’s Distance for Few-Shot Learning” In _arXiv preprint arXiv:2003.06777_ , 2020
* [56] Hongguang Zhang, Jing Zhang and Piotr Koniusz “Few-Shot Learning via Saliency-guided Hallucination of Samples” In _arXiv preprint arXiv:1904.03472_ , 2019
* [57] Hongyi Zhang, Moustapha Cisse, Yann N. Dauphin and David Lopez-Paz “mixup: Beyond Empirical Risk Minimization” In _arXiv preprint arXiv:1710.09412_ , 2017
* [58] Jun-Yan Zhu, Taesung Park, Phillip Isola and Alexei A. Efros “Unpaired Image-to-Image Translation using Cycle-Consistent Adversarial Networks” In _arXiv preprint arXiv:1703.10593_ , 2017
|
# Dimerization in quantum spin chains with $O(n)$ symmetry
Jakob E. Björnberg Department of Mathematics, Chalmers University of
Technology and the University of Gothenburg, Sweden<EMAIL_ADDRESS>,
Peter Mühlbacher Department of Mathematics, University of Warwick, Coventry,
CV4 7AL, United Kingdom<EMAIL_ADDRESS>, Bruno Nachtergaele
Department of Mathematics and Center for Quantum Mathematics and Physics
University of California, Davis
Davis, CA 95616, USA<EMAIL_ADDRESS>and Daniel Ueltschi Department of
Mathematics, University of Warwick, Coventry, CV4 7AL, United Kingdom
<EMAIL_ADDRESS>
###### Abstract.
We consider quantum spins with $S\geq 1$, and two-body interactions with
$O(2S+1)$ symmetry. We discuss the ground state phase diagram of the one-
dimensional system. We give a rigorous proof of dimerization for an open
region of the phase diagram, for $S$ sufficiently large. We also prove the
existence of a gap for excitations.
###### 1991 Mathematics Subject Classification:
82B10, 82B20, 82B26
© 2021 by the authors. This paper may be reproduced, in its entirety, for non-
commercial purposes.
###### Contents
1. 1 Introduction
1. 1.1 A family of quantum spin chains with $O(n)$-invariant interactions
2. 1.2 Ground state phase diagram for general $n\geq 3$
3. 1.3 The $S=1$ model ($n=3$)
4. 1.4 Our result about dimerization
5. 1.5 Gap for excitations
2. 2 Graphical representation for $O(n)$ models
3. 3 The contour model
1. 3.1 Contours
2. 3.2 Domains and admissibility of contours
3. 3.3 Decomposition of $H(\omega)$
4. 4 Proof of dimerization
1. 4.1 Setting of the cluster expansion
2. 4.2 Cluster expansion for the partition function
3. 4.3 Dimerization
4. 4.4 Proof of exponential decay of correlations
5. 5 Proof of the spectral gap
6. A The interaction $uT+vP$ when $n$ is even
## 1\. Introduction
Over the course of almost a century of studying quantum spin chains,
physicists and mathematicians have uncovered a wide variety of interesting
physical phenomena and in the process invented an impressive arsenal of new
mathematical techniques and structures. Nevertheless, our understanding of
these simplest of quantum many-body systems is still far from complete. For
many models of interest we have only partial information about the ground
state phase diagram, the nature of the phase transitions, and the spectrum of
excitations. We consider here a family of spin systems with two-body
interactions, where interactions are translation invariant and $O(2S+1)$
invariant. We investigate the ground state phase diagram, looking for ground
states that possess less symmetry than the interactions. Our main result is a
rigorous proof of dimerization (where translation invariance is broken) in a
region of the phase diagram with $S$ large enough (Theorem 1.1). We also prove
exponential clustering (Theorem 1.2) and the existence of a gap (Theorem 1.4).
The family of models is introduced in Section 1.1; the phase diagram for
general $S\geq 1$ is described in Section 1.2; the case $S=1$ has received a
lot of attention and we discuss it explicitly in Section 1.3; our result about
dimerization is stated in Section 1.4.
The $O(n)$ models have a graphical representation which we describe in Section
2. We use it to define a “contour model” in Section 3 where contours are shown
to have small weights. This allows to use the method of cluster expansion and
prove dimerization in Section 4.
### 1.1. A family of quantum spin chains with $O(n)$-invariant interactions
We consider a family of quantum spin chains consisting of $2\ell$ spins of
magnitude $S$ defined by a nearest-neighbor Hamiltonian $H_{\ell}$ acting on
the Hilbert space $\mathcal{H}_{\ell}=({\mathbb{C}}^{n})^{\otimes 2\ell}$,
with $n=2S+1\geq 2$, of the form
$H_{\ell}=\sum_{x=-\ell+1}^{\ell-1}h_{x,x+1},$ (1.1)
where $h_{x,x+1}$ denotes a copy of $h=h^{*}\in M_{n}({\mathbb{C}})\otimes
M_{n}({\mathbb{C}})$ acting on the nearest neighbor pair at sites $x$ and
$x+1$.
We are interested in the family of interactions
$h=uT+vQ,\quad u,v\in{\mathbb{R}},$ (1.2)
where $T$ is the transposition operator defined by
$T(\phi\otimes\varphi)=\varphi\otimes\phi$, for
$\phi,\varphi\in{\mathbb{C}}^{n}$, and $Q$ is the orthogonal projection onto
the one-dimensional subspace of ${\mathbb{C}}^{n}\otimes{\mathbb{C}}^{n}$
spanned by a vector of the form
$\psi=\frac{1}{\sqrt{n}}\sum_{\alpha=1}^{n}e_{\alpha}\otimes e_{\alpha},$
(1.3)
for some orthornormal basis $\\{e_{\alpha}|\alpha=1,\ldots,n\\}$ of
${\mathbb{C}}^{n}$.
The spectrum of $h$ is easy to find. $T$ has the eigenvalues $1$ and $-1$,
corresponding to the symmetric and antisymmetric subspaces of
${\mathbb{C}}^{n}\otimes{\mathbb{C}}^{n}$, whose dimensions are $n(n+1)/2$ and
$n(n-1)/2$, respectively. Since $\psi$ is symmetric, the eigenvalues of $h$
are $u+v,u,-u$.
Let $R$ be a linear transformation represented by an orthogonal matrix in the
basis $\\{e_{\alpha}\\}$, meaning $\langle e_{\alpha},RR^{\rm
T}e_{\beta}\rangle=\delta_{\alpha\beta}$. This amounts to defining a specific
representation of $O(n)$ on the system under consideration. It is then
straightforward to check $(R\otimes R)\psi=\psi$. It follows that $R\otimes R$
commutes with $Q=|\psi\rangle\langle\psi|$. Since $T$ also commutes with
$R\otimes R$, the Hamiltonians with interaction $h$ given in (1.2) have a
local $O(n)$ symmetry. This family of models is in fact, up to a trivial
additive constant, the most general translation-invariant nearest neighbor
Hamiltonian for spins of dimension $n$ and with a translation-invariant local
$O(n)$ symmetry.
To make contact with previous results in the literature, it is useful to note
a couple of equivalent forms of the spin chains we consider. First, for
integer values of $S$, that is odd dimensions $n$, consider the orthonormal
basis $\\{e_{\alpha}\\}$, relabeled by $\alpha=-S,\ldots,S$, and related to
the standard eigenbasis of the third spin matrix $S^{(3)}$, satisfying
$S^{(3)}|\alpha\rangle=\alpha|\alpha\rangle$, as follows: for $\alpha=0$ take
$e_{0}={\rm i}^{S}|0\rangle$, and for $\alpha>0$ define
$e_{\alpha}=\frac{{\rm
i}^{S-\alpha}}{\sqrt{2}}\bigl{(}|\alpha\rangle+|-\alpha\rangle\bigr{)},\quad
e_{-\alpha}=\frac{{\rm
i}^{S-\alpha+1}}{\sqrt{2}}\bigl{(}|\alpha\rangle-|-\alpha\rangle\bigr{)}.$
(1.4)
Then, we have
$\psi=\phi:=\frac{1}{\sqrt{n}}\sum_{\alpha=-S}^{S}(-1)^{S-\alpha}|\alpha,-\alpha\rangle,$
(1.5)
which is the $SU(2)$ singlet vector in the standard spin basis. The
transposition operator $T$ is of course not affected by any translation-
invariant local basis change. Therefore, for odd $n$, and with a simple change
of basis, the family of interactions (1.2) is seen to be equivalent to
$\tilde{h}=uT+vP,\quad u,v\in{\mathbb{R}},$ (1.6)
where $P$ is the orthogonal projection onto the singlet state $\phi$.
The case of even $n$ is different. Interactions $h$ and $\tilde{h}$ are not
unitarily equivalent. But the model with interaction $\tilde{h}$ is
nonetheless interesting and we discuss it in Appendix A. We also prove
dimerization and a gap in this case, see Theorem 1.3 and Theorem 1.4.
For $n\geq 2$, $u=0$, and $v=-1$, this is the much studied $-P^{(0)}$ spin
chain [7, 2, 13, 5, 20, 19, 4].
### 1.2. Ground state phase diagram for general $n\geq 3$
We start with the phase diagram for arbitrary $n\geq 3$ and discuss the
special case $n=3$ in Section 1.3. The ground state phase diagram of the spin
chain with nearest-neighbor interactions $h_{x,x+1}=uT_{x,x+1}+vQ_{x,x+1}$ is
depicted in Fig. 1. It can be broadly divided into four domains.
$u$$v$AA’BB’CReshetikhin$v=-\frac{2n}{n-2}u$ferromagneticdimerizationincommensuratephase
correlationsMatrix-product state(s)$v=-2u$ Figure 1. Ground state phase
diagram for the chain with nearest-neighbor interactions $uT+vQ$ for $n\geq
3$. Our main result, Theorem 1.1, is a proof of dimerization in an open region
around the point B’.
The domain formed by the quadrant $u\leq 0,v\geq 0$ (blue region in Fig. 1) is
ferromagnetic. There are many ground states and they minimize $h_{x,x+1}$ for
all $x$; that is, they are frustration-free. The ground state energy per bond
is equal to $u$. Indeed, let $\varphi=\sum_{\alpha}c_{\alpha}e_{\alpha}$ with
$\sum_{\alpha}|c_{\alpha}|^{2}=1$. It is clear that
$|\varphi\otimes\varphi\rangle$ is eigenstate of $T$ with eigenvalue 1;
further, we have
$\langle\varphi\otimes\varphi|Q|\varphi\otimes\varphi\rangle=\frac{1}{n}\Bigl{|}\sum_{\alpha}c_{\alpha}^{2}\Bigr{|}^{2}.$
(1.7)
The latter is zero when $\sum_{\alpha}c_{\alpha}^{2}=0$. Since $Q$ is a
projector, such a state is eigenstate with eigenvalue 0. Notice that the state
$R\varphi$ also satisfies this condition, for all orthogonal transformation
$R$.
The product state $\otimes_{x=-\ell+1}^{\ell}\varphi$ is then a ground state
of $h_{x,x+1}$ with eigenvalue $u$, for all $x$. In addition to these product
states, we can obviously take linear combinations.
The next domain is the arc-circle between $(u,v)=(-1,0)$ and the “Reshetikhin
point” with $v=-\frac{2n}{n-2}u$ (yellow region in Fig. 1), which features
dimerization. In order to see that dimerization is plausible as soon as $v<0$,
let
$\varphi_{x,x+1}=\sqrt{1-{\varepsilon}^{2}}\,|S,S\rangle+\frac{{\varepsilon}}{\sqrt{n-1}}\sum_{\alpha=-S}^{S-1}|\alpha,\alpha\rangle.$
(1.8)
and consider the (partially) dimerized state
$\varphi_{-\ell+1,-\ell+2}\otimes\varphi_{-\ell+3,-\ell+4}\otimes\dots$. For
${\varepsilon}=0$, this is a product state, but for ${\varepsilon}\neq 0$ it
is not. Roughly half the edges, namely the edges $x,x+1$ with
$x=-\ell+1,-\ell+3,\dotsc$, are dimerized and their energy is
$\langle\varphi_{x,x+1}|uT_{x,x+1}+vQ_{x,x+1}|\varphi_{x,x+1}\rangle=u+\frac{v}{n}(1+2\sqrt{n-1}{\varepsilon})+O({\varepsilon}^{2}).$
(1.9)
The non-dimerized edges contribute
$\langle\varphi_{x-1,x}\otimes\varphi_{x+1,x+2}|uT_{x,x+1}+vQ_{x,x+1}|\varphi_{x-1,x}\otimes\varphi_{x+1,x+2}\rangle=u+\frac{v}{n}+O({\varepsilon}^{2}).$
(1.10)
The average energy per bond of the state $\varphi_{x,x+1}$ is then
$u+\frac{v}{n}+v\frac{\sqrt{n-1}}{n}{\varepsilon}$, up to
$O({\varepsilon}^{2})$ corrections. When $v<0$ the optimal product states have
energy $u+\frac{v}{n}$ (using (1.7) with $\sum_{\alpha}c^{2}_{\alpha}=1$), so
the partially dimerized state (1.8) has lower energy when ${\varepsilon}$ is
positive and small.
Our main result is that dimerization does occur in an open domain around the
point B’, provided $n$ is sufficiently large, see Theorem 1.1. This extends
the results of [19, 4], valid at the point B’.
Then comes the domain formed by the arc-circle between the Reshetikhin point
$v=-\frac{2n}{n-2}u$ and $(u,v)=(1,0)$ (red region in Fig. 1). For $n$ odd a
unique translation-invariant ground state is expected.
This domain contains several interesting special cases. The direction
$(u=1,v=0)$ is the the $SU(n)$ generalization of the spin-1/2 Bethe-ansatz
solvable Heisenberg model studied by Sutherland and others [24]. The direction
$v=-\frac{2n}{n-2}u$ was solved by Reshetikhin [22] (this generalizes the
Takhtajan–Babujian model for $n=3$). These models are gapless. The direction
$v=-2u$ is a frustration free point and the ground states are given matrix-
product states. For odd $n$, these are generalizations of the AKLT model. The
ground state for the infinite chain is unique and is in the Haldane phase. For
even $n$, there are two matrix-product ground states that break the
translation invariance of the chain down to period 2 [28].
The final domain is the quadrant $u,v>0$. The ground states are expected to
have slow decaying correlations with incommensurate phase correlations. That
is, spin-spin correlations between sites 0 and $x$ are expected to be of the
form $|x|^{-r}\cos(\omega|x|)$ for $|x|$ large, and where $r,\omega$ depend on
the parameters $u,v$ [11].
It is perhaps worth mentioning that the phase diagram for spatial dimensions
other than 1 is quite different. Dimerization is not expected. Instead, the
system displays various forms of magnetic long-range orders (ferromagnetic,
spin nematic, Néel, …). See [30] for results about magnetic ordering for all
$n\geq 2$ and for parameters that correspond to the dimerized phase here.
### 1.3. The $S=1$ model ($n=3$)
For $n=3$, the family of models is equivalent to the familiar spin-1 chain
with bilinear and biquadratic interactions. The latter is most often
parametrized by an angle $\phi$ as follows:
$\cos\phi\,\vec{S}_{x}\cdot\vec{S}_{x+1}+\sin\phi\,(\vec{S}_{x}\cdot\vec{S}_{x+1})^{2}=3(\sin\phi-\cos\phi)P+\cos\phi\,T+\sin\phi\,I.$
(1.11)
We can apply the change of basis that is the inverse of Eq. (1.4), namely
$|0\rangle=-{\rm i}\,e_{0},\quad|1\rangle=\tfrac{1}{\sqrt{2}}(e_{1}-{\rm
i}\,e_{-1}),\quad|-1\rangle=\tfrac{1}{\sqrt{2}}(e_{1}+{\rm i}\,e_{-1}).$
(1.12)
Then the interaction is given by (1.11) but with the operator $Q$ instead of
$P$.
$\cos\phi$$\sin\phi$ASutherlandA’BB’CTakhtajan-BabujianAKLT
($\tan\phi=\frac{1}{3}$)ferromagneticdimerizationincommensuratephase
correlations Figure 2. Ground state phase diagram for the $S=1$ chain with
nearest-neighbor interactions
$\cos\phi\vec{S}_{x}\cdot\vec{S}_{x+1}+\sin\phi(\vec{S}_{x}\cdot\vec{S}_{x+1})^{2}$.
The domains and the points are the same as those in Fig. 1.
The ground state phase diagram with parameter $\phi$ is depicted in Fig. 2.
The domains and the points are the same as in Fig. 1. The ferromagnetic domain
corresponds to $\phi\in(\frac{\pi}{2},\frac{5\pi}{4})$, and the model is
frustration-free in this range. Among the ground states, there is a family of
product states that shows that the $O(3)$ symmetry of the Hamiltonian is
spontaneously broken. As a consequence, the Goldstone Theorem [15] implies
that there are gapless excitations above the ground state in this region. The
dimerization domain is $\phi\in(\frac{5\pi}{4},\frac{7\pi}{4})$. The next
domain is $\phi\in(-\frac{\pi}{4},\frac{\pi}{4})$ with unique, translation-
invariant ground states. Finally, the domain
$\phi\in(\frac{\pi}{4},\frac{\pi}{2})$ is expected to display states with slow
decay of correlations, with incommensurate phase.
There are several points where exact and/or rigorous information is available:
(i) $\phi\in[0,\pi/2]$ with $\tan\phi=1/3$, it is the spin-1 AKLT chain [3]
with interaction $\tilde{h}$ given by the orthogonal projection on the spin-2
states. In the thermodynamic limit, it has a unique ground state of Matrix
Product form with a non-vanishing spectral gap and exact exponential decay of
correlations; (ii) the two points with $\tan\phi=1$, A and A’ in Fig. 2, have
$SU(3)$ symmetry and are often referred to as the Sutherland model [24]. An
exact solution for the ground state at $\phi=-3\pi/4$ is gapless and highly
degenerate, while for $\phi=\pi/4$ is believed to be a unique critical state
with gapless excitations; (iii) the point $\phi=-\pi/4$ is the Bethe-ansatz
solvable Takhtajan–Babujian model [25, 6], which is also gapless; (iv) the
point $\phi=-\pi/2$, is the $-P^{(0)}$ spin-1 chain, already mentioned above.
Aizenman, Duminil-Copin, and Warzel proved that it has two dimerized
(2-periodic) ground states with exponential decay of correlations [4]; all
evidence indicates that these states are gapped.
Let us briefly comment on higher spatial dimensions. Dimerization is not
expected. Various rigorous results about magnetic long-range order have been
established: for $\phi=0$ [9]; for $\phi\gtrsim\frac{5\pi}{4}$ [26, 30]; and
for $\phi\lesssim 0$ [16]. Recently, the model on the complete graph has been
studied by Ryan using methods based on the Brauer algebra [23], which plays a
role in the representation theory of the orthogonal groups analogous to that
of the symmetric group for the general linear groups.
### 1.4. Our result about dimerization
Let us introduce the operators $L^{\alpha,\alpha^{\prime}}$,
$1\leq\alpha<\alpha^{\prime}\leq n$, that are generators of the Lie algebra
$\mathfrak{o}(n)$:
$L^{\alpha,\alpha^{\prime}}=|\alpha\rangle\langle\alpha^{\prime}|-|\alpha^{\prime}\rangle\langle\alpha|.$
(1.13)
And for $x\in\\{-\ell+1,\dots,\ell\\}$, let $L_{x}^{\alpha,\alpha^{\prime}}$
be the operator in $\mathcal{H}_{\ell}$ that acts as
$L^{\alpha,\alpha^{\prime}}$ at the site $x$, and as the identity elsewhere.
###### Theorem 1.1.
There exist constants $n_{0},u_{0},c>0$ (independent of $\ell$) such that for
$n>n_{0}$ and $|u|<u_{0}$, we have that for all
$1\leq\alpha<\alpha^{\prime}\leq n$,
$\begin{split}&\lim_{\beta\to\infty}\Bigl{[}\langle
L_{0}^{\alpha,\alpha^{\prime}}L_{1}^{\alpha,\alpha^{\prime}}\rangle_{\ell,\beta,u}-\langle
L_{-1}^{\alpha,\alpha^{\prime}}L_{0}^{\alpha,\alpha^{\prime}}\rangle_{\ell,\beta,u}\Bigr{]}>c\qquad\text{for
all $\ell$ odd;}\\\ &\lim_{\beta\to\infty}\Bigl{[}\langle
L_{0}^{\alpha,\alpha^{\prime}}L_{1}^{\alpha,\alpha^{\prime}}\rangle_{\ell,\beta,u}-\langle
L_{-1}^{\alpha,\alpha^{\prime}}L_{0}^{\alpha,\alpha^{\prime}}\rangle_{\ell,\beta,u}\Bigr{]}<-c\qquad\text{for
all $\ell$ even.}\end{split}$
Figure 3. Illustration for dimerization. Depending on whether $\ell$ is even
or odd, the site $x=0$ is more entangled with its left or its right neighbor.
Theorem 1.1 establishes the existence of at least two distinct infinite-volume
ground states, close to the point B’ of the phase diagram (see Fig. 3). Notice
that the same result holds if we replace the operators
$L_{0}^{\alpha,\alpha^{\prime}}L_{1}^{\alpha,\alpha^{\prime}}$ with spin
operators $S_{0}^{(3)}S_{1}^{(3)}$, diagonal in the basis $\\{e_{\alpha}\\}$.
We expect that there are exactly two extremal ground states, precisely given
by limits $\ell\to\infty$ along odd or even integers. We also expect that, if
we take the chain to be $\\{-\ell,-\ell+1,\dots,\ell\\}$, the corresponding
infinite-volume ground state is equal to the average of the two extremal
states.
The next result shows that the ground state retains the $O(n)$ symmetry of the
system, that there is no magnetic long range order. This is indeed an
attribute of dimerisation.
###### Theorem 1.2.
There exist constants $n_{0},u_{0},c_{1},c_{2},C>0$ (independent of $\ell$)
such that for $n>n_{0}$ and $|u|<u_{0}$, we have
$\lim_{\beta\to\infty}\bigl{|}\langle L_{x}^{\alpha,\alpha^{\prime}}\,{\rm
e}^{-tH_{\ell}}\,L_{y}^{\alpha,\alpha^{\prime}}\,{\rm
e}^{tH_{\ell}}\,\rangle_{\ell,\beta,u}\bigr{|}\leq C\,{\rm
e}^{-c_{1}|x-y|-c_{2}|t|}\,$
for all $\ell\in{\mathbb{N}}$, all $x,y\in\\{-\ell+1,\dots,\ell\\}$, all
$1\leq\alpha<\alpha^{\prime}\leq n$, and all $t\in{\mathbb{R}}$.
Dimerization has been established in [19, 4] at the point B’ in the phase
diagrams of Figs 1 and 2. The earlier result [19] uses the loop representation
of [5] combined with a Peierls argument; it holds for $S\geq 8$ (or $n\geq
17$). The second result, due to Aizenman, Duminil-Copin and Warzel, remarkably
holds for all $S\geq 1$ ($n\geq 3$), i.e., for all values of $S$ (or $n$)
where dimerization is expected. It uses the loop representation and random
cluster representation of [5] as well as recent results for the two-
dimensional random cluster model [8, 21].
Away from the point B’ these methods do not apply. In this article we use the
loop representation of [30], which combines those of [27, 5], in order to get
a contour model; see Theorem 2.1. The loop representation involves a
probability measure for $u,v\leq 0$ only; it involves a signed measure
otherwise. This is described in Section 2. For large $n$, typical
configurations involve many loops, that are short loops located on all the
dimerized edges. We define contours to be excitations with respect to this
background. It is possible to obtain a contour model with piecewise compatible
contours, that is suitable for a cluster expansion (Section 3). This method is
robust regarding signs and it allows to intrude in the region with positive
parameter $u$. Proving that the expansion converges is difficult, since the
cost of excitations is entropic rather than energetic. This is done in Section
4. This allows to establish dimerization in the loop model, see Theorem 4.7.
It is equivalent to Theorem 1.1, thus proving our main result. Theorem 1.2 is
proved in Subsection 4.4.
The interaction that is responsible for dimerization is the operator
$Q_{x,x+1}$ and we prove that dimerization is stable under perturbations of
this interaction by $uT_{x,x+1}$, with $|u|$ sufficiently small. It should be
possible to prove stability under more general perturbations that are not
necessarily invariant under the group $O(n)$. Since the unperturbed model is
not frustration-free, this does not follow from the recent result about the
stability of gapped phases with discrete symmetry breaking in [18], which
requires the frustration-free property. For translation-invariant
perturbations by $O(n)$ invariant next-nearest neighbor or further terms, the
methods of this paper should generalize in a straightforward manner.
We now discuss the case of the spin chain with Hamiltonian
$\tilde{H}_{\ell}=\sum_{x=-\ell+1}^{\ell-1}\bigl{(}uT_{x,x+1}+vP_{x,x+1}\bigr{)},$
(1.14)
where $P$ is projection onto the singlet state (recall (1.6)). We have a
similar result about dimerization. Let $S^{(i)}$, $i=1,2,3$, be the spin
operators that are the generators of the $SU(2)$ symmetry group for
$\tilde{H}_{\ell}$. In the basis $|\alpha\rangle$ where $P$ is the projection
onto the vector $\phi$ in (1.5), we can choose $S^{(3)}$ such that
$S^{(3)}|\alpha\rangle=\alpha|\alpha\rangle$. Let
$\langle
S_{x}^{(i)}S_{y}^{(i)}\rangle_{\ell,\beta,u}^{\tilde{}}=\frac{1}{{\operatorname{Tr\,}}\,{\rm
e}^{-\beta\tilde{H}_{\ell}}\,}{\operatorname{Tr\,}}S_{x}^{(i)}S_{y}^{(i)}\,{\rm
e}^{-\beta\tilde{H}_{\ell}}\,.$ (1.15)
###### Theorem 1.3.
Let $v=-1$, and $i\in\\{1,2,3\\}$. There exist constants $n_{0},u_{0},c>0$
(independent of $\ell$) such that for $n>n_{0}$ and $|u|<u_{0}$, we have
$\begin{split}&\lim_{\beta\to\infty}\Bigl{[}\langle
S_{0}^{(i)}S_{1}^{(i)}\rangle_{\ell,\beta,u}^{\tilde{}}-\langle
S_{-1}^{(i)}S_{0}^{(i)}\rangle_{\ell,\beta,u}^{\tilde{}}\Bigr{]}>c\qquad\text{for
all $\ell$ odd;}\\\ &\lim_{\beta\to\infty}\Bigl{[}\langle
S_{0}^{(i)}S_{1}^{(i)}\rangle_{\ell,\beta,u}^{\tilde{}}-\langle
S_{-1}^{(i)}S_{0}^{(i)}\rangle_{\ell,\beta,u}^{\tilde{}}\Bigr{]}<-c\qquad\text{for
all $\ell$ even.}\end{split}$
When $n$ is odd this theorem is equivalent to Theorem 1.1, as the correlations
of spin operators are the same as correlations of operators
$L_{x,y}^{\alpha,\alpha^{\prime}}$, up to some factors. In the case where $n$
is even, this is no longer the case and the proof needs to be adapted; the
modifications are described in Appendix A.
### 1.5. Gap for excitations
Let $E_{0}^{(\ell)}<E_{1}^{(\ell)}<\dots$ be the eigenvalues of $H_{\ell}$,
and $\tilde{E}_{0}^{(\ell)}<\tilde{E}_{1}^{(\ell)}<\dots$ be the eigenvalues
of $\tilde{H}_{\ell}$. The gaps are defined as
$\begin{split}\Delta^{(\ell)}=E_{1}^{(\ell)}-E_{0}^{(\ell)},\\\
\tilde{\Delta}^{(\ell)}=\tilde{E}_{1}^{(\ell)}-\tilde{E}_{0}^{(\ell)}.\end{split}$
(1.16)
The gaps are obviously positive but the question is whether they are so
uniformly in $\ell$.
###### Theorem 1.4.
There exist constants $n_{0},u_{0},c>0$ (independent of $\ell$) such that for
$n>n_{0}$ and $|u|<u_{0}$, we have
* (a)
The multiplicities of $E_{0}^{(\ell)}$ and $\tilde{E}_{0}^{(\ell)}$ are equal
to 1. (That is, ground states are unique.)
* (b)
$\Delta^{(\ell)}\geq c$ and $\tilde{\Delta}^{(\ell)}\geq c$ for all $\ell$.
Recall that the chain is $\\{-\ell+1,\dots,\ell\\}$ and it always contains an
even number of sites. Our theorem does not cover the chains with odd numbers
of sites, although we expect the corresponding Hamiltonians to be gapped as
well.
The spatial exponential decay proved in Theorem 1.2 is also a consequence of
Theorem 1.4, due to the Exponential Clustering Theorem (see the simultaneous
articles [10, 17]). Our proof here is motivated by [12]. For the model
$H_{\ell}$ it can be found in Section 5. It relies on a loop and contour
representation, and on cluster expansions, as for the proof of dimerization.
The modifications for $\tilde{H}_{\ell}$ are discussed in the appendix.
## 2\. Graphical representation for $O(n)$ models
Consider the one-dimensional graph consisting of the $2\ell$ vertices
$V_{\ell}\mathrel{\mathop{:}}=\\{-\ell+1,\dots,\ell\\}$ and the edges
$E_{\ell}\mathrel{\mathop{:}}=\big{\\{}(x,x+1):-\ell+1\leq
x\leq\ell-1\big{\\}}$. Fix $\beta>0$. To each vertex and edge of this graph we
associate a periodic time interval $T_{\beta}=(-\beta,\beta)_{\text{per}}$ to
obtain a set of _space-time vertices_
$\overline{V}_{\ell,\beta}\mathrel{\mathop{:}}=V_{\ell}\times T_{\beta}$ as
well as a set of _space-time edges_
$\overline{E}_{\ell,\beta}\mathrel{\mathop{:}}=E_{\ell}\times T_{\beta}$.
By a _configuration_ $\omega$ we mean a finite subset of
$\overline{E}_{\ell,\beta}$, each point of $\omega$ receiving a _mark_ or .
The points of $\omega$ will collectively be called _links_ , those marked
being referred to as _crosses_ and those marked as _double-bars_. We write
$\omega=(\omega_{\mathchoice{\vbox{\hbox{\includegraphics[width=5.60002pt]{dbar.pdf}}}}{\vbox{\hbox{\includegraphics[width=5.60002pt]{dbar.pdf}}}}{\vbox{\hbox{\includegraphics[width=4.20004pt]{dbar.pdf}}}}{\vbox{\hbox{\includegraphics[width=3.5pt]{dbar.pdf}}}}},\omega_{\mathchoice{\vbox{\hbox{\includegraphics[width=5.60002pt]{cross.pdf}}}}{\vbox{\hbox{\includegraphics[width=5.60002pt]{cross.pdf}}}}{\vbox{\hbox{\includegraphics[width=4.20004pt]{cross.pdf}}}}{\vbox{\hbox{\includegraphics[width=3.5pt]{cross.pdf}}}}})$
and denote the set of all such (link) configurations $\Omega_{\ell,\beta}$.
To every configuration $\omega\in\Omega_{\ell,\beta}$ corresponds a set of
_loops_ ; see Fig. 4 for an illustration. A loop $l$ is a closed, injective
trajectory
$\displaystyle[0,L]_{\text{per}}$ $\displaystyle\to\overline{V}_{\ell}$
$\displaystyle t$ $\displaystyle\mapsto l(t)=(v(t),T(t)),$
such that $x(t)$ is piecewise constant and $T^{\prime}(t)\in\\{\pm 1\\}$. We
call $L\equiv|l|$ the _length_ of $l$, that is the smallest $L>0$ in the above
equation. A jump occurs at $t\in[0,L]$ provided that $\\{x(t-),x(t+)\\}\times
T(t)$ contains a link. We have $T^{\prime}(t+)=-T^{\prime}(t-)$ in case that
link is a double bar and $T^{\prime}(t+)=T^{\prime}(t-)$ in case it is a
cross. We identify loops with identical support and we occasionally abuse
notation and identify a loop with the set of links it traverses. The number of
loops in a configuration $\omega$ is denoted $\mathcal{L}(\omega)$. The number
of links in a configuration $\omega$ is denoted by $\\#\omega$. Similarly the
number of double bars is denoted by
$\\#\omega_{\mathchoice{\vbox{\hbox{\includegraphics[width=5.60002pt]{dbar.pdf}}}}{\vbox{\hbox{\includegraphics[width=5.60002pt]{dbar.pdf}}}}{\vbox{\hbox{\includegraphics[width=4.20004pt]{dbar.pdf}}}}{\vbox{\hbox{\includegraphics[width=3.5pt]{dbar.pdf}}}}}$
and the number of crosses is denoted by
$\\#\omega_{\mathchoice{\vbox{\hbox{\includegraphics[width=5.60002pt]{cross.pdf}}}}{\vbox{\hbox{\includegraphics[width=5.60002pt]{cross.pdf}}}}{\vbox{\hbox{\includegraphics[width=4.20004pt]{cross.pdf}}}}{\vbox{\hbox{\includegraphics[width=3.5pt]{cross.pdf}}}}}$.
For $u\in\mathbb{R}$, we define the following _signed measure_ on the set
$\Omega_{\ell,\beta}$ of link configurations $\omega$:
${\rm
d}\bar{\rho}_{u}(\omega)=u^{\\#\omega_{\mathchoice{\vbox{\hbox{\includegraphics[width=4.00002pt]{cross.pdf}}}}{\vbox{\hbox{\includegraphics[width=4.00002pt]{cross.pdf}}}}{\vbox{\hbox{\includegraphics[width=3.00003pt]{cross.pdf}}}}{\vbox{\hbox{\includegraphics[width=2.5pt]{cross.pdf}}}}}}{\rm
d}^{\otimes\\#\omega}x,$ (2.1)
where ${\rm d}x$ is the Lebesgue measure on $\overline{E}_{\ell,\beta}$. We
also introduce the following normalized measure $\rho_{u}$, satisfying
$\rho_{u}(\Omega_{\ell,\beta})=1$:
${\rm d}\rho_{u}(\omega)=\,{\rm e}^{-(1+u)2\beta|E_{\ell}|}\,{\rm
d}\bar{\rho}_{u}(\omega)$ (2.2)
If $u$ is positive, the measure $\rho_{u}$ is a positive measure and hence a
probability measure; in fact, under this measure $\omega$ has the distribution
of a Poisson point process with intensity $u$ for crosses and intensity $1$
for double-bars . But we also allow small, negative $u$. Let
$Z_{\ell,\beta,n,u}\mathrel{\mathop{:}}=\int_{\Omega_{\ell,\beta}}{\rm
d}\rho_{u}(\omega)\,n^{\mathcal{L}(\omega)-\\#\omega_{\mathchoice{\vbox{\hbox{\includegraphics[width=4.00002pt]{dbar.pdf}}}}{\vbox{\hbox{\includegraphics[width=4.00002pt]{dbar.pdf}}}}{\vbox{\hbox{\includegraphics[width=3.00003pt]{dbar.pdf}}}}{\vbox{\hbox{\includegraphics[width=2.5pt]{dbar.pdf}}}}}}.$
(2.3)
This loop model is equivalent to the quantum spin system, and the next result
is an instance of this equivalence. The equivalence goes back to Tóth [27] and
Aizenman–Nachtergaele [5] for special choices of the parameters; the general
case of the interaction (1.2) is due to [30]. Note that it holds for arbitrary
finite graphs, not only for chains.
We write $x\not\leftrightarrow y$ to characterize the set of configurations
$\omega$ where $(x,0)$ and $(y,0)$ belong to distinct loops;
$x\overset{+}{\longleftrightarrow}y$ where the top of $(x,0)$ is connected to
the bottom of $(y,0)$; and $x\overset{-}{\longleftrightarrow}y$ where the top
of $(x,0)$ is connected to the top of $(y,0)$ (see [30, Fig. 2] for an
illustration).
###### Theorem 2.1.
For the Hamiltonian (1.1) with $h_{x,x+1}=-uT_{x,x+1}-Q_{x,x+1}$, we have that
* (a)
$\displaystyle{\operatorname{Tr\,}}\,{\rm e}^{-2\beta H_{\ell}}\,=\,{\rm
e}^{2\beta(1+u)|E_{\ell}|}\,Z_{\ell,\beta,n,u}$.
* (b)
For all $1\leq\alpha<\alpha^{\prime}\leq n$, we have
${\operatorname{Tr\,}}L_{x}^{\alpha,\alpha^{\prime}}L_{y}^{\alpha,\alpha^{\prime}}\,{\rm
e}^{-2\beta H_{\ell}}\,=\tfrac{2}{n}\,{\rm
e}^{2\beta(1+u)|E_{\ell}|}\,\int_{\Omega_{\ell,\beta}}{\rm
d}\rho_{u}(\omega)\,n^{\mathcal{L}(\omega)-\\#\omega_{\mathchoice{\vbox{\hbox{\includegraphics[width=4.00002pt]{dbar.pdf}}}}{\vbox{\hbox{\includegraphics[width=4.00002pt]{dbar.pdf}}}}{\vbox{\hbox{\includegraphics[width=3.00003pt]{dbar.pdf}}}}{\vbox{\hbox{\includegraphics[width=2.5pt]{dbar.pdf}}}}}}\bigl{(}{\mathchoice{\rm
1\mskip-4.0mul}{\rm 1\mskip-4.0mul}{\rm 1\mskip-4.5mul}{\rm
1\mskip-5.0mul}}[x\overset{-}{\longleftrightarrow}y]-{\mathchoice{\rm
1\mskip-4.0mul}{\rm 1\mskip-4.0mul}{\rm 1\mskip-4.5mul}{\rm
1\mskip-5.0mul}}[x\overset{+}{\longleftrightarrow}y]\bigr{)}.$
The sign of the parameter $u$ in the definition of the interaction has indeed
changed; but the theorem holds for arbitrary real (or even complex)
parameters. Theorem 2.1 can also be formulated for the interaction
$h_{x,x+1}=-uT_{x,x+1}-vQ_{x,x+1}$, by inserting the factor
$v^{\\#\omega_{\mathchoice{\vbox{\hbox{\includegraphics[width=4.00002pt]{dbar.pdf}}}}{\vbox{\hbox{\includegraphics[width=4.00002pt]{dbar.pdf}}}}{\vbox{\hbox{\includegraphics[width=3.00003pt]{dbar.pdf}}}}{\vbox{\hbox{\includegraphics[width=2.5pt]{dbar.pdf}}}}}}$
inside the integrals.
###### Proof.
The proof of (a) can be found in [30, Theorem 3.2] and (b) is similar, so we
only sketch it here. Let $\Sigma(\omega)$ be the set of “space-time spin
configurations” that are constant along the loops (so that
$|\Sigma(\omega)|=n^{\mathcal{L}(\omega)}$). By a standard Feynman-Kac
expansion, we get
${\operatorname{Tr\,}}\,{\rm e}^{-2\beta H_{\ell}}\,=\,{\rm
e}^{2\beta(1+u)|E_{\ell}|}\,\int_{\Omega_{\ell,\beta}}{\rm
d}\rho_{u}(\omega)\,n^{-\\#\tilde{\omega}_{\mathchoice{\vbox{\hbox{\includegraphics[width=4.00002pt]{dbar.pdf}}}}{\vbox{\hbox{\includegraphics[width=4.00002pt]{dbar.pdf}}}}{\vbox{\hbox{\includegraphics[width=3.00003pt]{dbar.pdf}}}}{\vbox{\hbox{\includegraphics[width=2.5pt]{dbar.pdf}}}}}}\sum_{\sigma\in\Sigma(\omega)}1.$
(2.4)
We recognize the partition function in (2.3), so we get (a).
For (b) we need a modified set of space-time spin configurations where the
spin value must jump from $\alpha$ to $\alpha^{\prime}$, or from
$\alpha^{\prime}$ to $\alpha$, at the points $(x,0)$ and $(y,0)$. Let
$\Sigma_{x,y}^{\alpha,\alpha^{\prime}}$ be this set. We then have
${\operatorname{Tr\,}}_{x}^{\alpha,\alpha^{\prime}}L_{y}^{\alpha,\alpha^{\prime}}\,{\rm
e}^{-2\beta H_{\ell}}\,=\,{\rm
e}^{2\beta(1+u)|E_{\ell}|}\,\int_{\Omega_{\ell,\beta}}{\rm
d}\rho_{u}(\omega)\,n^{-\\#\tilde{\omega}_{\mathchoice{\vbox{\hbox{\includegraphics[width=4.00002pt]{dbar.pdf}}}}{\vbox{\hbox{\includegraphics[width=4.00002pt]{dbar.pdf}}}}{\vbox{\hbox{\includegraphics[width=3.00003pt]{dbar.pdf}}}}{\vbox{\hbox{\includegraphics[width=2.5pt]{dbar.pdf}}}}}}\\\
\sum_{\sigma\in\Sigma_{x,y}^{\alpha,\alpha^{\prime}}(\omega)}\langle\sigma_{x,0+}|L_{x}^{\alpha,\alpha^{\prime}}|\sigma_{x,0-}\rangle\,\langle\sigma_{y,0+}|L_{y}^{\alpha,\alpha^{\prime}}|\sigma_{y,0-}\rangle.$
(2.5)
It is necessary that $(x,0)$ and $(y,0)$ belong to the same loop in order to
get a nonzero contribution. Further, we have
$\langle\sigma_{x,0+}|L_{x}^{\alpha,\alpha^{\prime}}|\sigma_{x,0-}\rangle\,\langle\sigma_{y,0+}|L_{y}^{\alpha,\alpha^{\prime}}|\sigma_{y,0-}\rangle=\begin{cases}-1&\text{if
}x\overset{+}{\longleftrightarrow}y,\\\ +1&\text{if
}x\overset{-}{\longleftrightarrow}y.\end{cases}$ (2.6)
Since
$|\Sigma_{x,y}^{\alpha,\alpha^{\prime}}(\omega)|=\frac{2}{n}n^{\mathcal{L}(\omega)}$,
we get (b). ∎
From now on and to the end of this article we work with the loop model.
###### Remark 2.2 (Intuition).
It is helpful to think of $\rho_{u}$ as an a-priori measure on a gas of loops,
and rewrite the integrand
$n^{\mathcal{L}(\omega)-\\#\omega_{\mathchoice{\vbox{\hbox{\includegraphics[width=4.00002pt]{dbar.pdf}}}}{\vbox{\hbox{\includegraphics[width=4.00002pt]{dbar.pdf}}}}{\vbox{\hbox{\includegraphics[width=3.00003pt]{dbar.pdf}}}}{\vbox{\hbox{\includegraphics[width=2.5pt]{dbar.pdf}}}}}}$
as $\,{\rm e}^{-(\log n)H(\omega)}\,$, with ‘Hamiltonian’
$-H(\omega)\mathrel{\mathop{:}}=\mathcal{L}(\omega)-\\#\omega_{\mathchoice{\vbox{\hbox{\includegraphics[width=5.60002pt]{dbar.pdf}}}}{\vbox{\hbox{\includegraphics[width=5.60002pt]{dbar.pdf}}}}{\vbox{\hbox{\includegraphics[width=4.20004pt]{dbar.pdf}}}}{\vbox{\hbox{\includegraphics[width=3.5pt]{dbar.pdf}}}}},$
(2.7)
and inverse temperature $\log n$. Thinking of $n$ as large, the Laplace
principle tells us that ‘typical’ configurations should maximise
$n^{\mathcal{L}(\omega)-\\#\omega_{\mathchoice{\vbox{\hbox{\includegraphics[width=4.00002pt]{dbar.pdf}}}}{\vbox{\hbox{\includegraphics[width=4.00002pt]{dbar.pdf}}}}{\vbox{\hbox{\includegraphics[width=3.00003pt]{dbar.pdf}}}}{\vbox{\hbox{\includegraphics[width=2.5pt]{dbar.pdf}}}}}}$.
Our goal is to write $Z_{\ell,\beta,n,u}$ as a dominant contribution from such
maximizers, and some excitations.
We end this section with the following remark about working with a signed
measure. Since the (possibly signed) measure $\rho_{u}$ is closely related to
the probability measure $\rho_{1}$, it is easy to see that any event $A$
satisfying $\rho_{1}(A)=0$ also has zero measure under $\rho_{u}$. In fact, we
have the following slightly stronger property:
###### Lemma 2.3.
If $A$ is an event such that $\rho_{1}(A)=0$ and
$f:\Omega_{\ell,\beta}\to\mathbb{R}$ is a $\rho_{1}$-integrable function, then
for any $u\in\mathbb{R}$ we have that
$\int_{A}{\rm d}\rho_{u}(\omega)f(\omega)=0.$ (2.8)
###### Proof.
Using (2.1) and (2.2) it is easy to see that
$\left|\int_{A}{\rm d}\rho_{u}(\omega)f(\omega)\right|\leq C\int_{A}{\rm
d}\rho_{1}(\omega)|f(\omega)|=0,$ (2.9)
for some finite constant $C$ depending only on $u,\ell,\beta$. ∎
As a consequence, we may assume that crosses and double-bars occur at
different times, also when $u<0$ and the measure $\rho_{u}$ carries signs. We
implicitly used this property when defining loops.
## 3\. The contour model
### 3.1. Contours
We classify loops as follows, see Fig. 4. A loop is _contractible_ if it can
be continuously deformed to a point and _winding_ otherwise. Not all loops are
contractible since our time interval $T_{\beta}$ is periodic. A loop is _long_
if it visits three or more distinct vertices or if it is winding; it is
_short_ otherwise.
Figure 4. A configuration $\omega$ consisting of three short loops (green,
brown, purple), and three long loops (red, blue, orange) two of which are
winding loops (blue, orange).
We define a _canonical orientation_ of the space-time vertices
$\overline{V}_{\ell,\beta}$, using the directions up ($\uparrow$) and down
($\downarrow$), by orienting the leftmost space-time vertex
$\\{-\ell+1\\}\times T_{\beta}$ _down_ $\downarrow$ and requiring that
neighbouring space-time vertices have opposite orientations; see Fig. 5. We
write $V^{\uparrow}_{\ell}\mathrel{\mathop{:}}=\\{x\in V_{\ell}:x+\ell\mbox{
is even}\\}$ for the set of vertices with up-orientation, and
$V^{\downarrow}_{\ell}\mathrel{\mathop{:}}=\\{x\in V_{\ell}:x+\ell\mbox{ is
odd}\\}$ for the set of vertices with down-orientation, and introduce the
following subsets of the edge-set $E_{\ell}$:
$\begin{split}E_{\ell}^{+}&\mathrel{\mathop{:}}=\big{\\{}(x,x+1)\in
E_{\ell}:x\in V^{\downarrow}_{\ell},x+1\in V^{\uparrow}_{\ell}\big{\\}},\\\
E^{-}_{\ell}&\mathrel{\mathop{:}}=E_{\ell}\setminus
E^{+}_{\ell}=\big{\\{}(x,x+1)\in E_{\ell}:x\in V^{\uparrow}_{\ell},x+1\in
V^{\downarrow}_{\ell}\big{\\}}.\end{split}$ (3.1)
We define $\overline{E}_{\ell,\beta}^{+}$ and $\overline{E}_{\ell,\beta}^{-}$,
as well as $\overline{V}^{\uparrow}_{\ell,\beta}$ and
$\overline{V}^{\downarrow}_{\ell,\beta}$, analogously.
Figure 5. (a) The canonical orientation of $\overline{V}_{\ell,\beta}$ with
the set $E^{+}_{\ell}$ highlighted red. (b) a configuration $\omega$ with many
short loops; these are positively oriented under the canonical orientation.
These definitions are motivated as follows. We expect that ‘typical’
configurations $\omega$ contain many short loops. To maximize the number of
short loops one places only double-bars in $\overline{E}^{+}_{\ell}$, as in
Fig. 5 (b). The canonical orientation is chosen so that all the short loops in
such a configuration are positively oriented (i.e. counter-clockwise). The
canonical orientation will be useful in classifying the excitations away from
such ‘typical’ $\omega$. Also note that if the origin $0$ belongs to a short,
positively oriented loop, then we have $0\leftrightarrow 1$ for $\ell$ odd and
$0\leftrightarrow-1$ for $\ell$ even. To prove our main result Theorem 1.1 we
will essentially argue that the origin is likely to belong to a short,
positively oriented loop.
Given a loop $l$ in a configuration $\omega$, we define a _segment_ of $l$ as
a trajectory of $l$ between two times $0\leq s_{1}<s_{2}\leq L(l)$ when $l$
passes through height $\beta$. That is to say,
$l(s_{1})=(v_{1},\beta),l(s_{2})=(v_{2},\beta)$ for some $v_{i}\in V_{\ell}$,
while $l$ does not pass through height $\beta$ in times $t\in(s_{1},s_{2})$.
We say that a segment is _spanning_ if for every $t\in T_{\beta}$ there exists
a $v=v(t)\in V_{\ell}$ such that the segment traverses $(v,t)$. Note that a
spanning segment is not necessarily part of a winding loop. See Fig. 6.
Figure 6. The leftmost and rightmost loops are winding loops with one spanning
segment each. The loop in the middle is contractible. There are four spanning
segments in total.
###### Definition 3.1 (Contours).
We say that two loops are _connected_ if they share a link or both are
winding. A _contour_ is then a maximally connected set of long loops.
###### Remark 3.2.
For later reference, we note here that any cross which is traversed by some
loop in a contour is necessarily traversed both ways by the contour; see Fig.
7.
Figure 7. A cross is traversed by a contour $\gamma$. If the red loop visits
a third vertex, it is a long loop; otherwise it must be a winding loop. In
both cases, it is actually part of $\gamma$.
A contour which contains at least one winding loop will be called a _winding
contour_. See Fig. 8.
Figure 8. Two contours: One winding contour, consisting of two winding loops,
and one consisting of four long, but contractible loops.
We need a notion of _interior_ of a contour, and for this it is useful to
regard our configuration $\omega$ as living in the bi-infinite cylinder
$C_{\beta}={\mathbb{R}}\times T_{\beta}$. More precisely, given $\omega$ we
consider the subset $\overline{\omega}$ of $C_{\beta}$ obtained as the union
of (i) $\overline{V}_{\beta}$ embedded in $C_{\beta}$ in the natural way, and
(ii) the links of $\omega$ embedded as straight line segments connecting
adjacent points of $\overline{V}_{\beta}$. Note that, in the embedding
$\overline{\omega}$, crosses and double-bars are embedded in the same way. For
a loop $l$ of $\omega$, define its _support_ $S(l)$ as the subset of
$\overline{\omega}$ traced out by $l$, meaning the union of the vertical and
horizontal line segments of $\overline{\omega}$ corresponding to the intervals
of $\overline{V}_{\beta}$ and the links of $\omega$ traversed by $l$. For a
contour $\gamma$ of $\omega$ we then make the following definitions.
* •
The _support_ $S(\gamma)$ is the union of the supports $S(l)$ of the loops $l$
belonging to $\gamma$. Note that $S(\gamma)$ is a closed subset of
$C_{\beta}$.
* •
The _exterior_ $E(\gamma)$ is the union of the unbounded connected components
of $C_{\beta}\setminus S(\gamma)$. Note that $E(\gamma)$ is open.
* •
The _interior_
$I(\gamma)\mathrel{\mathop{:}}={C_{\beta}\setminus\overline{E(\gamma)}}$. Note
that $I(\gamma)$ is an open set.
* •
The _boundary_ $B(\gamma)\mathrel{\mathop{:}}=\overline{E(\gamma)}\setminus
E(\gamma)$ which is a closed set.
* •
The (vertical) length $|\gamma|$ of a contour as the sum of the (vertical)
lengths of its loops, $|\gamma|\mathrel{\mathop{:}}=\sum_{l\in\gamma}|l|$.
These notions are illustrated in Figs 9–11.
Having defined $I(\gamma)$ as a subset of the cylinder $C_{\beta}$, we may
also regard $I(\gamma)$ (or more precisely, its closure
$\overline{I(\gamma)}$) as a subset of $\overline{E}_{\ell,\beta}$ by
identifying a point $(x,x+1)\times\\{t\\}\in\overline{E}_{\ell,\beta}$ with
the closed line-segment from $(x,t)$ to $(x+1,t)$ in $C_{\beta}$. Similarly,
$S(\gamma)$ and $B(\gamma)$ may be regarded as subsets of
$\overline{V}_{\ell,\beta}\cup\omega$. We freely switch between these points
of view.
Figure 9. A configuration $\omega$ with three contours highlighted green, blue
and red. The green contour consists of two winding loops. Figure 10. The
corresponding embedding $\overline{\omega}\subseteq
C_{\beta}=T_{\beta}\times\mathbb{R}$, with the supports $S(\gamma)$ of the
contours highlighted with the corresponding colors. Figure 11. The interiors
of the corresponding contours with the boundaries $B(\gamma)$ receiving the
canonical orientation. The green and blue contours are of positive type
(interiors $I(\gamma)$ on the left) while the red contour is of negative type
(interior on the right).
Fixing a contour $\gamma$, note that the boundary $B(\gamma)$ consists of a
collection of closed curves and horizontal line segments (of length 1). We use
the canonical orientation of $\overline{V}_{\ell,\beta}$ to orient each
vertical segment of $B(\gamma)$. It is not hard to see that this gives a
consistent orientation of all the closed curves constituting $B(\gamma)$.
(This follows from Remark 3.2.) Recall the standard notion of a _positively
oriented curve_ as one whose interior is always on the left.
###### Definition 3.3 (Type of a contour).
We say that the contour $\gamma$ is of _positive type_ if the canonical
orientation of $B(\gamma)$ is _positive_ in the sense that $I(\gamma)$ is on
the left of each closed curve of $B(\gamma)$. Otherwise we say that $\gamma$
is of _negative type_ (being of negative type is equivalent to the interior
being on the right).
###### Remark 3.4.
Suppose that $\omega\in\Omega_{\ell,\beta}$ is such that a given point
$\bar{v}\in\overline{V}_{\ell,\beta}$ is not on or inside any contour, that is
to say
$\bar{v}\in\bigcap_{\gamma\in\Gamma(\omega)}E(\gamma),$ (3.2)
where $E(\gamma)$ is the exterior of $\gamma$ defined above. Then we have that
$\bar{v}$ is on a _positively oriented_ short loop. Indeed, this is related to
the fact that all external contours are of positive type, see Lemma 3.6.
### 3.2. Domains and admissibility of contours
We now introduce several notations and definitions pertaining to contours and
how they relate to each other. First, given $\omega\in\Omega_{\ell,\beta}$ we
define $\Gamma(\omega)=\\{\gamma_{1},\dotsc,\gamma_{k}\\}$ as the set of
contours in the configuration $\omega$. Here, and in what follows, a contour
may be identified with the set of links it traverses. The collection of all
possible contours will be denoted
$X_{\ell,\beta}=\bigcup_{\omega\in\Omega_{\ell,\beta}}\Gamma(\omega)$, and we
write $X^{+}_{\ell,\beta}\subseteq X_{\ell,\beta}$ for the collection of
_positive-type_ contours. We write
$\mathfrak{X}_{\ell,\beta}=\bigcup_{k\geq
0}\binom{X_{\ell,\beta}}{k}\quad\mbox{and}\quad\mathfrak{X}^{+}_{\ell,\beta}=\bigcup_{k\geq
0}\binom{X^{+}_{\ell,\beta}}{k}$ (3.3)
for the set of finite collections of contours, respectively positive-type
contours. Elements of $\mathfrak{X}_{\ell,\beta}$ and of
$\mathfrak{X}^{+}_{\ell,\beta}$ will usually be denoted by $\Gamma$. It is
important to note that far from every such set $\Gamma$ of contours can be
obtained as $\Gamma(\omega)$ for some $\omega\in\Omega_{\ell,\beta}$; in fact,
we will devote some effort to identifying criteria under which such an
$\omega$ does indeed exist. We say that $\Gamma\in\mathfrak{X}_{\ell,\beta}$
is _admissible_ if $\Gamma=\Gamma(\omega)$ for some
$\omega\in\Omega_{\ell,\beta}$, and write
$\mathfrak{A}_{\ell,\beta}=\Gamma(\Omega_{\ell,\beta})$ for the collection of
admissible sets of contours.
Recall that the interior $I(\gamma)$ of a contour $\gamma$ is by definition an
open subset of the cylinder $C_{\beta}$. Also recall that we regard
$\overline{E}_{\ell,\beta}$ as a closed subset of $C_{\beta}$ by identifying a
point $(x,x+1)\times\\{t\\}\in\overline{E}_{\ell,\beta}$ with the closed line-
segment from $(x,t)$ to $(x+1,t)$. We now define the _(interior) domains_ of
$\gamma$ as follows.
###### Definition 3.5.
A domain $D$ of $\gamma$ is a subset of
$\overline{E}_{\ell,\beta}\cap{I(\gamma)}$ which, when regarded as a subset of
$C_{\beta}$ as above, is connected, satisfies $D\cap S(\gamma)=\varnothing$,
and is maximal with these properties.
We define the _type_ of a domain in a similar way to the type of a contour.
Namely, we orient the (topological) boundary of $D$ consistenly with the
canonical orientation of $\overline{V}_{\ell,\beta}$ and say that $D$ is of
_positive type_ if this is a positive orientation (interior on the left), and
of _negative type_ otherwise. See Fig. 12.
Figure 12. A contour $\gamma$ of positive type, containing three domains
$D_{1},D_{2},D_{3}$. Domains $D_{1}$ and $D_{3}$ are of negative type, while
$D_{2}$ is of positive type.
Given two contours $\gamma$ and $\gamma^{\prime}$, we say that $\gamma$ is a
_descendant_ of $\gamma^{\prime}$, writing $\gamma\prec\gamma^{\prime}$, if
$S(\gamma)\subseteq D$ for some domain $D$ of $\gamma^{\prime}$. Given
$\Gamma\in\mathfrak{X}_{\ell,\beta}$ and $\gamma,\gamma^{\prime}\in\Gamma$, we
say that $\gamma$ is an _immediate descendant of $\gamma^{\prime}$ in
$\Gamma$_ if $\gamma\prec\gamma^{\prime}$ and there is no
$\overline{\gamma}\in\Gamma$ satisfying both $\gamma\prec\overline{\gamma}$
and $\overline{\gamma}\prec\gamma^{\prime}$. It is important to note that the
notion of being an _immediate_ descendant depends not only on the two contours
$\gamma$ and $\gamma^{\prime}$ but on the set $\Gamma$; in other words,
immediate descendancy cannot be checked in a pairwise manner. If
$\gamma\in\Gamma$ is not the descendant of any other contour
$\gamma^{\prime}\in\Gamma$ then we say that $\gamma$ is an _external_ contour;
this notion is also dependent on the set $\Gamma$.
Note that the unique (if it exists) winding contour is always external since a
winding loop cannot be in the interior of any contractible loop.
###### Lemma 3.6.
Fix $\Gamma\in\mathfrak{X}_{\ell,\beta}$. Then $\Gamma$ is admissible, i.e.
$\Gamma\in\mathfrak{A}_{\ell,\beta}$, if and only if the following hold:
1. (1)
all external contours in $\Gamma$ are of positive type;
2. (2)
for any pair of distinct contours $\gamma,\gamma^{\prime}\in\Gamma$ we have
that either
$\overline{I(\gamma)}\cap\overline{I(\gamma^{\prime})}=\varnothing$ or
$\gamma\prec\gamma^{\prime}$ or $\gamma^{\prime}\prec\gamma$;
3. (3)
if $\gamma$ is an immediate descendant of $\gamma^{\prime}$, in a domain $D$
of $\gamma^{\prime}$, then the types of $\gamma$ and of $D$ coincide;
4. (4)
there exists at most one winding contour $\gamma\in\Gamma$.
###### Proof.
It is easy to see that the four conditions above hold for any admissible
$\Gamma=\Gamma(\omega)$. To show the converse, we construct an explicit
$\omega\in\Omega_{\ell,\beta}$ with $\Gamma(\omega)=\Gamma$. Starting from the
empty configuration $\omega_{0}=\varnothing\in\Omega_{\ell,\beta}$, add all
links of all external contours and then place a double bar at height 0, say,
on each $e\in E_{\ell}^{+}$ that does not have any link on it. This defines
$\omega_{1}$ such that $\Gamma(\omega_{1})$ is precisely the set of external
contours of $\Gamma$. Next, add the links of all contours which are immediate
descendants of external contours. This does not create any new long loops
apart from those in these contours because their types coincide with those of
the domains they are in. Iterate this procedure until there are no more
contours left to add. ∎
An important prerequisite for applying a cluster expansion is to be able to
verify the admissibility of a set of contours in a pairwise manner. As
indicated above, and in the light of Lemma 3.6, this is not directly possible
since the notion of being an immediate descendant depends on the whole set
$\Gamma$. We get around this issue by introducing a notion of _compatibility_
which applies to sets of positive-type contours
$\Gamma\in\mathfrak{X}^{+}_{\ell,\beta}$, and which can be checked in a
pairwise manner. We then show that there is a bijective correspondence between
admissible and compatible sets of contours.
The bijective correspondence referred to above involves _shifting_ contours
and rests on the simple observation that if $\gamma$ is a negative-type
contour, then $\gamma^{\prime}=\gamma+(1,0)$ (i.e. $\gamma$ translated to the
right one unit) is a positive-type contour.
Figure 13. A contour $\gamma$ and its two appropriately shifted domains
(shaded areas). The lower one was not moved since it already was positive
type. The upper one was shifted one column to the right. If a
$\gamma^{\prime}\in X_{\ell,\beta}$ gets placed inside it,
$S^{-1}(\Gamma=(\gamma,\gamma^{\prime}))$ will return an admissible collection
of contours.
Given a positive-type contour $\gamma\in X^{+}_{\ell,\beta}$ with domains
$D_{1}(\gamma),\dotsc,D_{k}(\gamma)\subseteq I(\gamma)$, we define the
_appropriately shifted domains_ $D_{i}^{+}(\gamma)$ of $\gamma$ by
$D_{i}^{+}(\gamma)=\left\\{\begin{array}[]{ll}D_{i}(\gamma),&\mbox{if
}D_{i}(\gamma)\mbox{ is of positive type},\\\
D_{i}(\gamma)+(1,0),&\mbox{otherwise}.\end{array}\right.$ (3.4)
Note that while $D_{i}^{+}(\gamma)\subseteq\overline{I(\gamma)}$, a shifted
domain may intersect the boundary $B(\gamma)$. See Fig. 13.
###### Definition 3.7.
Given two positive-type contours $\gamma,\gamma^{\prime}\in
X^{+}_{\ell,\beta}$, we say that $\gamma$ and $\gamma^{\prime}$ are
_compatible_ if one of the following hold:
1. (1)
$\overline{I(\gamma)}\cap\overline{I(\gamma^{\prime})}=\varnothing$, or
2. (2)
$S(\gamma)\subseteq D_{i}^{+}(\gamma^{\prime})$ for some $i$, or
3. (3)
$S(\gamma^{\prime})\subseteq D_{i}^{+}(\gamma)$ for some $i$, or
4. (4)
at least one of $\gamma$ or $\gamma^{\prime}$ is not a winding contour.
We define
$\delta(\gamma,\gamma^{\prime})=\begin{cases}1&\text{if
$\gamma,\gamma^{\prime}$ are compatible}\\\ 0&\text{otherwise.}\end{cases}$
(3.5)
Finally, we let
$\mathfrak{C}^{+}_{\ell,\beta}\subseteq\mathfrak{X}^{+}_{\ell,\beta}$ to be
the collection of all pairwise compatible sets of positive-type contours; that
is,
$\Gamma=\\{\gamma_{1},\dotsc,\gamma_{k}\\}\in\mathfrak{X}^{+}_{\ell,\beta}$
belongs to $\mathfrak{C}^{+}_{\ell,\beta}$ if $\prod_{1\leq i<j\leq
k}\delta(\gamma_{i},\gamma_{j})=1$.
A compatible set $\Gamma$ is generally itself not admissible, since compatible
contours may overlap but admissible contours may not. Intuitively, one obtains
an admissible set of contours from a compatible set by ‘shifting back’ the
appropriately shifted domains $D_{i}^{+}(\gamma)$ and the contours they
contain. For nested contours, the shift is performed iteratively; see Fig. 14.
Figure 14. (a) A compatible set of contours
$\Gamma=\\{\gamma_{1},\gamma_{2},\gamma_{3},\gamma_{4},\gamma_{5}\\}$. We
have, for example, $\sigma_{\Gamma}(\gamma_{3})=2$ since $\gamma_{3}$ lies in
a shifted domain of both $\gamma_{1}$ and $\gamma_{2}$, while
$\sigma_{\Gamma}(\gamma_{4})=1$ since it lies in a shifted domain of
$\gamma_{1}$ only. (b) The admissible set $\Sigma(\Gamma)$.
More formally, define the _shift_
$\Sigma:\mathfrak{C}^{+}_{\ell,\beta}\to\mathfrak{X}_{\ell,\beta}$ as follows.
First, given $\Gamma\in\mathfrak{C}^{+}_{\ell,\beta}$ and $\gamma\in\Gamma$,
write $\sigma_{\Gamma}(\gamma)$ for the number of contours
$\gamma^{\prime}\in\Gamma\setminus\\{\gamma\\}$ such that $\gamma\subseteq
D^{+}_{i}(\gamma^{\prime})\neq D_{i}(\gamma^{\prime})$. This represents the
number of times $\gamma$ is shifted to the right in order to obtain the
compatible set $\Gamma$ from an admissible set of contours. We define
$\Sigma(\Gamma)=\\{\gamma-(\sigma_{\Gamma}(\gamma),0):\gamma\in\Gamma\\}.$
(3.6)
###### Lemma 3.8.
The shift $\Sigma$ is a bijection from $\mathfrak{C}^{+}_{\ell,\beta}$, the
collection of compatible sets of contours, to $\mathfrak{A}_{\ell,\beta}$, the
collection of admissible sets of contours.
###### Proof.
It is easy to construct an inverse $\Sigma^{-1}$ of $\Sigma$ on
$\mathfrak{A}_{\ell,\beta}$, as follows. Given
$\Gamma\in\mathfrak{A}_{\ell,\beta}$, start with an external contour $\gamma$
(which is of positive type by Lemma 3.6) and form its appropriately shifted
domains $D_{i}^{+}(\gamma)$. In doing so, shift also the descendants of
$\gamma$ along with their domains. Note that all the immediate descendants of
$\gamma$ are then mapped to positive type contours. Then iteratively continue
this procedure for the (shifted) immediate descendants of $\gamma$. The
resulting set $\Sigma^{-1}(\Gamma)$ then satisfies Definition 3.7.
It remains to show that $\Sigma(\Gamma)\in\mathfrak{A}_{\ell,\beta}$ for all
$\Gamma\in\mathfrak{C}^{+}_{\ell,\beta}$, i.e. that $\Sigma(\Gamma)$ satisfies
Lemma 3.6. Compatibility ensures that there is at most one winding contour. It
is clear that external contours are of positive type since they are not
shifted. For $\gamma,\gamma^{\prime}$ with disjoint interiors, this property
is preserved by $\Sigma$; if $\gamma\subseteq D_{i}^{+}(\gamma^{\prime})$ then
the shifting ensures that the images of $\gamma,\gamma^{\prime}$ under
$\Sigma$ satisfy $\gamma\prec\gamma^{\prime}$, while the relative amounts by
which the contours are shifted ensures that the types of immediate descendants
in $\Sigma(\Gamma)$ coincide with the types of the relevant domains. ∎
We close this subsection with a simple lemma about counting the amount of
‘available space’ for short loops in a configuration $\omega$, in terms of the
lengths of the contours. For $\Gamma\in\mathfrak{A}_{\ell,\beta}$, we define
the _free set_ $F(\Gamma)\subseteq\overline{E}_{\ell,\beta}$ as the space-time
edges where we can add links without modifying the contours in $\Gamma$ or
creating new ones.
###### Lemma 3.9.
Let $\Gamma\in\mathfrak{A}_{\ell,\beta}$ be an admissible set of contours.
Then
$|F(\Gamma)|=|\overline{E}_{\ell,\beta}^{+}|-\tfrac{1}{2}{\textstyle\sum_{\gamma\in\Gamma}|\gamma|.}$
(3.7)
###### Proof.
We need to show that
$2|\overline{E}_{\ell,\beta}^{+}|=2|F(\Gamma)|+{\textstyle\sum_{\gamma\in\Gamma}|\gamma|.}$
Note that $2|\overline{E}_{\ell,\beta}^{+}|=|\overline{V}_{\ell,\beta}|$ and
that $2|F(\Gamma)|$ equals the total length of all the short loops. But any
point in $\overline{V}_{\ell,\beta}$ lies either on a contour or on a short
loop, thus
$|\overline{V}_{\ell,\beta}|=2|F(\Gamma)|+{\textstyle\sum_{\gamma\in\Gamma}|\gamma|},$
as required. ∎
### 3.3. Decomposition of $H(\omega)$
Recall from (2.7) the quantity
$-H(\omega)=\mathcal{L}(\omega)-|\omega_{\mathchoice{\vbox{\hbox{\includegraphics[width=5.60002pt]{dbar.pdf}}}}{\vbox{\hbox{\includegraphics[width=5.60002pt]{dbar.pdf}}}}{\vbox{\hbox{\includegraphics[width=4.20004pt]{dbar.pdf}}}}{\vbox{\hbox{\includegraphics[width=3.5pt]{dbar.pdf}}}}}|$.
We now show that $H(\omega)$ can be decomposed as a sum over contours and we
prove bounds on the summands. To this end, for a loop $l$ let $\mathcal{T}(l)$
denote the number of _turns_ that $l$ makes; symbolically
$\mathcal{T}(l)=\\#\mathchoice{\vbox{\hbox{\includegraphics[width=8.00003pt]{dbar-
top.pdf}}}}{\vbox{\hbox{\includegraphics[width=8.00003pt]{dbar-
top.pdf}}}}{\vbox{\hbox{\includegraphics[width=6.00006pt]{dbar-
top.pdf}}}}{\vbox{\hbox{\includegraphics[width=5.0pt]{dbar-
top.pdf}}}}+\\#\mathchoice{\vbox{\hbox{\includegraphics[width=8.00003pt]{dbar-
bot.pdf}}}}{\vbox{\hbox{\includegraphics[width=8.00003pt]{dbar-
bot.pdf}}}}{\vbox{\hbox{\includegraphics[width=6.00006pt]{dbar-
bot.pdf}}}}{\vbox{\hbox{\includegraphics[width=5.0pt]{dbar-bot.pdf}}}}$. For a
contour $\gamma$, write $\mathcal{T}(\gamma)$ for the total number of U-turns
of all loops in $\gamma$. Next define the function
$h:X_{\ell,\beta}\to\mathbb{Z}$ by
$h(\gamma)=\mathcal{L}(\gamma)-\tfrac{1}{2}\mathcal{T}(\gamma)$ (3.8)
where $\mathcal{L}(\gamma)$ denotes the number of loops in the contour
$\gamma$.
###### Lemma 3.10.
For $\omega\in\Omega_{\ell,\beta}$ with contours $\Gamma=\Gamma(\omega)$ we
have $-H(\omega)=\sum_{\gamma\in\Gamma}h(\gamma)$.
###### Proof.
Since every double-bar of $\omega$ accounts for exactly two turns (of either
one or two loops), we have
$-H(\omega)=\sum_{l}\big{(}1-\tfrac{1}{2}\mathcal{T}(l)\big{)}$ (3.9)
where the sum is over all loops $l$ in the configuration $\omega$. The result
now follows from the observation that short, non-winding loops make exactly
two turns. ∎
Write
$\\#\gamma_{\mathchoice{\vbox{\hbox{\includegraphics[width=5.60002pt]{dbar.pdf}}}}{\vbox{\hbox{\includegraphics[width=5.60002pt]{dbar.pdf}}}}{\vbox{\hbox{\includegraphics[width=4.20004pt]{dbar.pdf}}}}{\vbox{\hbox{\includegraphics[width=3.5pt]{dbar.pdf}}}}}$
for the number of double-bars visited by $\gamma$ and
$\\#\gamma_{\mathchoice{\vbox{\hbox{\includegraphics[width=5.60002pt]{cross.pdf}}}}{\vbox{\hbox{\includegraphics[width=5.60002pt]{cross.pdf}}}}{\vbox{\hbox{\includegraphics[width=4.20004pt]{cross.pdf}}}}{\vbox{\hbox{\includegraphics[width=3.5pt]{cross.pdf}}}}}$
for the number of crosses.
###### Lemma 3.11.
For contours $\gamma$ without crosses, the function
$h:X_{\ell,\beta}\to\mathbb{Z}$ satisfies
$h(\gamma)\leq-\tfrac{1}{3}\\#\gamma+2\ell\hbox{\rm
1\kern-2.70004ptI}\\{\gamma\text{ has a spanning segment}\\}.$ (3.10)
Note that the constant $-\tfrac{1}{3}$ is tight for the smallest non-winding
contours with six double-bars and no crosses, while for larger contours the
constant may be taken closer to $-\tfrac{1}{2}$. As to the indicator function,
we will see that contours containing spanning segments become very rare
asymptotically.
###### Proof.
Write $\mathcal{W}(l)$ for the number of winding segments in $l$ and
$\mathcal{W}(\gamma)=\sum_{l\in\gamma}\mathcal{W}(l)$. We claim that it
suffices to show that $h(\gamma)\leq r(\gamma)$ where
$r(\gamma)=-\tfrac{1}{3}\mathcal{T}(\gamma)+\mathcal{W}(\gamma).$ (3.11)
Indeed, $r(\gamma)$ is bounded above by the right-hand-side of (3.10) for the
following reasons:
* •
double-bars visited twice by $\gamma$ count twice in $\mathcal{T}(\gamma)$ but
only once in
$\\#\gamma=\\#\gamma_{\mathchoice{\vbox{\hbox{\includegraphics[width=5.60002pt]{dbar.pdf}}}}{\vbox{\hbox{\includegraphics[width=5.60002pt]{dbar.pdf}}}}{\vbox{\hbox{\includegraphics[width=4.20004pt]{dbar.pdf}}}}{\vbox{\hbox{\includegraphics[width=3.5pt]{dbar.pdf}}}}}$,
while those visited once by $\gamma$ count once in both, meaning that
$\mathcal{T}(\gamma)\geq\\#\gamma$;
* •
$\mathcal{W}(\gamma)\leq 2\ell\hbox{\rm 1\kern-2.70004ptI}\\{\gamma\text{ has
a spanning segment}\\}$ since each point of the form
$(x,0)\in\overline{V}_{\ell,\beta}$ is visited by at most one winding segment.
Next, the claimed inequality $h(\gamma)\leq r(\gamma)$ is equivalent to:
$\mathcal{T}(\gamma)+6\mathcal{W}(\gamma)\geq 6\mathcal{L}(\gamma).$ (3.12)
To establish (3.12), first note that both sides are additive over loops. Thus
it suffices to show that any long or winding loop $l$ satisfies
$\mathcal{T}(l)+6\mathcal{W}(l)\geq 6.$ (3.13)
If $\mathcal{W}(l)\geq 1$ this is clear, hence we may assume that the loop is
non-winding. A long, non-winding loop which traverses only double-bars
necessarily makes at least 6 turns, see Fig. 15. This proves (3.13) and hence
the claim. ∎
###### Lemma 3.12.
For all contours $\gamma\in X_{\ell,\beta}$ we have
$h(\gamma)\leq 2\ell\hbox{\rm 1\kern-2.70004ptI}\\{\gamma\text{ has a spanning
segment}\\}.$ (3.14)
In particular $h(\gamma)\leq 0$ for all non-winding contours.
###### Proof.
Note that $h(\gamma)$ is additive over loops $l\in\gamma$. So since there can
be at most $2\ell$ spanning segments, it suffices to show for every loop that
$1-{1\over
2}(\\#\mathchoice{\vbox{\hbox{\includegraphics[width=8.00003pt]{dbar-
top.pdf}}}}{\vbox{\hbox{\includegraphics[width=8.00003pt]{dbar-
top.pdf}}}}{\vbox{\hbox{\includegraphics[width=6.00006pt]{dbar-
top.pdf}}}}{\vbox{\hbox{\includegraphics[width=5.0pt]{dbar-
top.pdf}}}}+\\#\mathchoice{\vbox{\hbox{\includegraphics[width=8.00003pt]{dbar-
bot.pdf}}}}{\vbox{\hbox{\includegraphics[width=8.00003pt]{dbar-
bot.pdf}}}}{\vbox{\hbox{\includegraphics[width=6.00006pt]{dbar-
bot.pdf}}}}{\vbox{\hbox{\includegraphics[width=5.0pt]{dbar-
bot.pdf}}}})\leq\hbox{\rm 1\kern-2.70004ptI}\\{l\text{ has a spanning
segment}\\}$. This is clearly true. ∎
###### Lemma 3.13.
For $\gamma\in X_{\ell,\beta}$, all $u$ with $|u|\leq 1$, and all $\kappa>0$,
we have
$n^{h(\gamma)}|u|^{\\#\gamma_{\mathchoice{\vbox{\hbox{\includegraphics[width=4.00002pt]{cross.pdf}}}}{\vbox{\hbox{\includegraphics[width=4.00002pt]{cross.pdf}}}}{\vbox{\hbox{\includegraphics[width=3.00003pt]{cross.pdf}}}}{\vbox{\hbox{\includegraphics[width=2.5pt]{cross.pdf}}}}}}\leq\min(n,|u|^{-\kappa/2})^{-(\frac{1}{3}-\kappa)\\#\gamma}n^{2\ell{\scriptsize\hbox{\rm
1\kern-1.89003ptI}}\\{\gamma\text{ has a spanning segment}\\}}.$ (3.15)
###### Proof.
If $\gamma$ has no spanning segment and
$\\#\gamma_{\mathchoice{\vbox{\hbox{\includegraphics[width=5.60002pt]{cross.pdf}}}}{\vbox{\hbox{\includegraphics[width=5.60002pt]{cross.pdf}}}}{\vbox{\hbox{\includegraphics[width=4.20004pt]{cross.pdf}}}}{\vbox{\hbox{\includegraphics[width=3.5pt]{cross.pdf}}}}}\geq{\kappa\over
2}(\frac{1}{3}-\kappa)\\#\gamma$, the claim follows from $h(\gamma)\leq 0$
(Lemma 3.12). Now consider the case where $\gamma$ has no spanning segment and
$\\#\gamma_{\mathchoice{\vbox{\hbox{\includegraphics[width=5.60002pt]{cross.pdf}}}}{\vbox{\hbox{\includegraphics[width=5.60002pt]{cross.pdf}}}}{\vbox{\hbox{\includegraphics[width=4.20004pt]{cross.pdf}}}}{\vbox{\hbox{\includegraphics[width=3.5pt]{cross.pdf}}}}}\leq{\kappa\over
2}(\frac{1}{3}-\kappa)\\#\gamma$. If $\\#\gamma<({\kappa\over 2}({1\over
3}-\kappa))^{-1}$, then
$\\#\gamma_{\mathchoice{\vbox{\hbox{\includegraphics[width=5.60002pt]{cross.pdf}}}}{\vbox{\hbox{\includegraphics[width=5.60002pt]{cross.pdf}}}}{\vbox{\hbox{\includegraphics[width=4.20004pt]{cross.pdf}}}}{\vbox{\hbox{\includegraphics[width=3.5pt]{cross.pdf}}}}}=0$
and thus we may apply Lemma 3.11 to get the desired bound in this case. So
assume now that $1\leq{\kappa\over 2}({1\over 3}-\kappa)\\#\gamma$ (and still
that $\gamma$ has no spanning segment and
$\\#\gamma_{\mathchoice{\vbox{\hbox{\includegraphics[width=5.60002pt]{cross.pdf}}}}{\vbox{\hbox{\includegraphics[width=5.60002pt]{cross.pdf}}}}{\vbox{\hbox{\includegraphics[width=4.20004pt]{cross.pdf}}}}{\vbox{\hbox{\includegraphics[width=3.5pt]{cross.pdf}}}}}\leq{\kappa\over
2}(\frac{1}{3}-\kappa)\\#\gamma$).
Let $\Gamma$ denote the collection of contours and small loops obtained by
removing all crosses from $\gamma$, and let $m$ denote the number of small
loops in $\Gamma$. Since the removal of a cross can only create at most one
more loop, we have that
$m\leq\\#\gamma_{\mathchoice{\vbox{\hbox{\includegraphics[width=5.60002pt]{cross.pdf}}}}{\vbox{\hbox{\includegraphics[width=5.60002pt]{cross.pdf}}}}{\vbox{\hbox{\includegraphics[width=4.20004pt]{cross.pdf}}}}{\vbox{\hbox{\includegraphics[width=3.5pt]{cross.pdf}}}}}+1\leq{\kappa\over
2}(\frac{1}{3}-\kappa)\\#\gamma+1\leq\kappa(\frac{1}{3}-\kappa)\\#\gamma$, and
that
$h(\gamma)\leq
h(\Gamma)+\\#\gamma_{\mathchoice{\vbox{\hbox{\includegraphics[width=5.60002pt]{cross.pdf}}}}{\vbox{\hbox{\includegraphics[width=5.60002pt]{cross.pdf}}}}{\vbox{\hbox{\includegraphics[width=4.20004pt]{cross.pdf}}}}{\vbox{\hbox{\includegraphics[width=3.5pt]{cross.pdf}}}}}\leq
h(\Gamma)+\frac{\kappa}{2}(\tfrac{1}{3}-\kappa)\\#\gamma.$ (3.16)
Since every short loop uses at most two double bars, the number of double bars
belonging to contours of $\Gamma$ is at least
$\\#\gamma_{\mathchoice{\vbox{\hbox{\includegraphics[width=5.60002pt]{dbar.pdf}}}}{\vbox{\hbox{\includegraphics[width=5.60002pt]{dbar.pdf}}}}{\vbox{\hbox{\includegraphics[width=4.20004pt]{dbar.pdf}}}}{\vbox{\hbox{\includegraphics[width=3.5pt]{dbar.pdf}}}}}-2m\geq\bigl{(}1-{\kappa\over
2}(\tfrac{1}{3}-\kappa)\bigr{)}\\#\gamma-2\kappa(\tfrac{1}{3}-\kappa)\\#\gamma=\bigl{(}1-\tfrac{5}{2}\kappa(\tfrac{1}{3}-\kappa)\bigr{)}\\#\gamma.$
(3.17)
Applying Lemma 3.11, this allows us to conclude that
$h(\Gamma)\leq-\tfrac{1}{3}\big{(}1-\tfrac{5}{2}\kappa(\tfrac{1}{3}-\kappa)\big{)}\\#\gamma.$
(3.18)
Combining (3.16) and (3.18) we conclude that
$h(\gamma)\leq-\big{[}\tfrac{1}{3}-3\kappa(\tfrac{1}{3}-\kappa)\big{]}\\#\gamma\leq-(\tfrac{1}{3}-\kappa)\\#\gamma.$
(3.19)
It remains to show the claim for $\gamma$ with a spanning segment. To this
end, consider $\omega$ obtained as follows: Denote by $\omega_{0}$ a
configuration of links such that its set of contours
$\Gamma(\omega_{0})=\\{\gamma\\}$. Now add $|E^{+}_{\ell}|=\ell$ double bars
at the same height, exactly one per column in $E^{+}_{\ell}$. Denote this
configuration by $\omega$ and note that $\Gamma(\omega)$ does not contain any
winding contours. Observe that the number of crosses is unchanged and we added
$\ell$ links, hence changing the number of loops by at most $\ell$. Using
these observations and Lemma 3.10 we thus get
$h(\gamma)=-H(\omega_{0})=-H(\omega)+\mathcal{E}=\sum_{\gamma^{\prime}\in\Gamma(\omega)}h(\gamma^{\prime})+\mathcal{E},$
(3.20)
where $\mathcal{E}$ is an error that is bounded by $|\mathcal{E}|\leq 2\ell$
and all $\gamma^{\prime}\in\Gamma(\omega)$ are non-winding so that the
previous bounds apply. This concludes the proof. ∎
Figure 15. A long, non-winding loop makes at least 6 turns.
## 4\. Proof of dimerization
### 4.1. Setting of the cluster expansion
We summarize the main results of the method of cluster expansion as we need
it. The following setting and theorem was proposed in [29], extending the
results of [14] to the continuous setting and general repulsive interactions.
Let $\Gamma$ be a measurable space, $\eta$ a complex measure on $\Gamma$ such
that $|\eta|(\Gamma)<\infty$, where $|\eta|$ is the total variation (absolute
value) of $\eta$. Let $\zeta$ be a symmetric function
$\Gamma\times\Gamma\to{\mathbb{C}}$ such that
$|1+\zeta(\gamma,\gamma^{\prime})|\leq 1$ for all
$\gamma,\gamma^{\prime}\in\Gamma$. Define the partition function $Z$ by
$Z=\sum_{k\geq 0}\frac{1}{k!}\int{\rm d}\eta(\gamma_{1})\dots\int{\rm
d}\eta(\gamma_{k})\prod_{1\leq i<j\leq
k}\big{(}1+\zeta(\gamma_{i},\gamma_{j})\bigr{)}.$ (4.1)
Finally, define the cluster function
$\varphi(\gamma_{1},\dots,\gamma_{k})=\begin{cases}1&\text{if }k=1,\\\
\frac{1}{k!}\sum_{G}\prod_{\\{i,j\\}\in
G}\zeta(\gamma_{i},\gamma_{j})&\text{otherwise,}\end{cases}$ (4.2)
where the sum is over connected graphs of $k$ elements, and the product is
over the edges of $G$. Then we have the following expressions and estimates.
###### Theorem 4.1.
Assume that there exist functions $a,b:\Gamma\to[0,\infty)$ such that for all
$\gamma\in\Gamma$, we have the following Kotecký-Preiss criterion
$\int{\rm d}|\eta|(\gamma^{\prime})|\zeta(\gamma,\gamma^{\prime})|\,{\rm
e}^{a(\gamma^{\prime})+b(\gamma^{\prime})}\,\leq a(\gamma).$ (4.3)
(Also, assume that $\int{\rm d}|\eta|(\gamma)\,{\rm
e}^{a(\gamma)+b(\gamma)}\,<\infty$.) Then we have the following.
* (a)
The partition function is equal to
$Z=\exp\biggl{\\{}\sum_{k\geq 1}\int{\rm d}\eta(\gamma_{1})\dots\int{\rm
d}\eta(\gamma_{k})\;\varphi(\gamma_{1},\dots,\gamma_{k})\biggr{\\}},$
where the combined sum and integral converges absolutely.
* (b)
For all $\gamma_{1}\in\Gamma$,
$1+\sum_{k\geq 2}k\int{\rm d}|\eta|(\gamma_{2})|\dots\int{\rm
d}|\eta|(\gamma_{k})\Bigl{(}\sum_{i=1}^{k}|\zeta(\gamma,\gamma_{i})|\Bigr{)}|\varphi(\gamma_{1},\dots,\gamma_{k})|\,{\rm
e}^{b(\gamma_{1})+\dots+b(\gamma_{k})}\,\leq\,{\rm e}^{a(\gamma_{1})}\,.$
* (c)
For all $\gamma\in\Gamma$,
$\sum_{k\geq 1}\int{\rm d}|\eta|(\gamma_{1})|\dots\int{\rm
d}|\eta|(\gamma_{k})\Bigl{(}\sum_{i=1}^{k}|\zeta(\gamma,\gamma_{i})|\Bigr{)}|\varphi(\gamma_{1},\dots,\gamma_{k})|\,{\rm
e}^{b(\gamma_{1})+\dots+b(\gamma_{k})}\,\leq a(\gamma).$
This theorem can be found in [29], see Theorems 1 and 3 there, as well as Eqs
(18) and (19). Notice that the term $b(\gamma)$ is not usually part of the
Kotecký–Preiss criterion and is not needed for convergence of the cluster
expansion. But it gives better estimates, see (b) and (c) above, which are
most helpful in proving exponential decay.
### 4.2. Cluster expansion for the partition function
Let us return to our loop model. We start with the partition function (2.3),
namely
$Z_{\ell,\beta,n,u}=\,{\rm
e}^{-(1+u)|\overline{E}_{\ell,\beta}|}\,\int_{\Omega_{\ell,\beta}}{\rm
d}\bar{\rho}_{u}(\omega)n^{\mathcal{L}(\omega)-\\#\omega_{\mathchoice{\vbox{\hbox{\includegraphics[width=4.00002pt]{dbar.pdf}}}}{\vbox{\hbox{\includegraphics[width=4.00002pt]{dbar.pdf}}}}{\vbox{\hbox{\includegraphics[width=3.00003pt]{dbar.pdf}}}}{\vbox{\hbox{\includegraphics[width=2.5pt]{dbar.pdf}}}}}}$
(4.4)
where ${\rm
d}\bar{\rho}_{u}(\omega)=u^{\\#\omega_{\mathchoice{\vbox{\hbox{\includegraphics[width=4.00002pt]{cross.pdf}}}}{\vbox{\hbox{\includegraphics[width=4.00002pt]{cross.pdf}}}}{\vbox{\hbox{\includegraphics[width=3.00003pt]{cross.pdf}}}}{\vbox{\hbox{\includegraphics[width=2.5pt]{cross.pdf}}}}}}{\rm
d}^{\otimes\\#\omega}x$ is given in (2.1). Since we identify contours
$\gamma\in X_{\ell,\beta}$ with the links they are made up of, ${\rm
d}\bar{\rho}_{u}(\gamma)$ is also well defined. We define
$\tilde{w}(\gamma)\mathrel{\mathop{:}}=\,{\rm
e}^{-(1+u)\frac{1}{2}|\gamma|}\,n^{h(\gamma)}u^{\\#\gamma_{\mathchoice{\vbox{\hbox{\includegraphics[width=4.00002pt]{cross.pdf}}}}{\vbox{\hbox{\includegraphics[width=4.00002pt]{cross.pdf}}}}{\vbox{\hbox{\includegraphics[width=3.00003pt]{cross.pdf}}}}{\vbox{\hbox{\includegraphics[width=2.5pt]{cross.pdf}}}}}},$
(4.5)
where $h(\gamma)$ is defined in (3.8). Let
$\mathcal{L}(\Gamma)=\\{l:\exists\gamma\in\Gamma:l\in\gamma\\}$ be the set of
loops in a (not necessarily admissible) collection of contours $\Gamma$. Let
$Y_{\ell,\beta}\subseteq X_{\ell,\beta}$ the set of contours $\gamma$ (not
necessarily admissible) consisting of two adjacent winding loops not
traversing any links and let
$\mathcal{Y}_{\ell,\beta}\mathrel{\mathop{:}}=\\{\Gamma\subseteq
Y_{\ell,\beta}:\gamma\cap\gamma^{\prime}=\emptyset,\;\forall\gamma\neq\gamma^{\prime}\in\Gamma\\}$.
Now let
$w(\gamma)\mathrel{\mathop{:}}=\sum_{\tilde{\gamma}\in
g(\gamma)}\tilde{w}(\tilde{\gamma})(-\,{\rm
e}^{-2\beta}\,)^{\\#\gamma\setminus\tilde{\gamma}\over 2},$ (4.6)
where
$g(\gamma)=\\{\tilde{\gamma}\subseteq\gamma\mid\exists\Gamma^{\prime}\in\mathcal{Y}_{\ell,\beta}:\gamma=\tilde{\gamma}\cup\mathcal{L}(\Gamma^{\prime})\\}$
(4.7)
is the set of contours $\tilde{\gamma}$ such that $\gamma$ can be obtained by
adding pairs of adjacent, winding loops not traversing any links (those that
come from having an “empty good column”) to $\tilde{\gamma}$ and
$\\#\gamma\setminus\tilde{\gamma}$ denotes the number of loops that are in
$\gamma$, but not in $\tilde{\gamma}$ – necessarily an even number by the
definition of $\mathcal{Y}_{\ell,\beta}$.
Note that for $\gamma\in X_{\ell,\beta}^{\rm{nw}}$ we have
$g(\gamma)=\\{\gamma\\}$, so $w(\gamma)=\tilde{w}(\gamma)$.
###### Proposition 4.2.
We have, for any $u\in\mathbb{R}$,
$Z_{\ell,\beta,n,u}=\,{\rm e}^{-(1+u)|\overline{E}_{\ell}^{-}|}\,\sum_{k\geq
0}\frac{1}{k!}\int_{X^{+}_{\ell}}{\rm
d}\bar{\rho}_{1}(\gamma_{1})\dots\int_{X^{+}_{\ell}}{\rm
d}\bar{\rho}_{1}(\gamma_{k})\Bigl{(}\prod_{i=1}^{k}w(\gamma_{i})\Bigr{)}\prod_{1\leq
i<j\leq k}\delta(\gamma_{i},\gamma_{j}).$
###### Proof.
First note that $d\bar{\rho}_{u}$ factorises, i.e. for
$\omega_{1},\omega_{2}\in\Omega_{\ell,\beta}$ sharing no links, we have
$d\bar{\rho}_{u}(\omega_{1}\cup\omega_{2})=d\bar{\rho}_{u}(\omega_{1})d\bar{\rho}_{u}(\omega_{2})$.
In particular, for any admissible set
$\\{\gamma^{1},\dotsc,\gamma^{k}\\}\in\mathfrak{A}_{\ell,\beta}$ of contours,
${\rm d}\bar{\rho}_{u}(\gamma^{1}\cup\dotsc\cup\gamma^{k})=\prod_{i=1}^{k}{\rm
d}\bar{\rho}_{u}(\gamma^{i}).$ (4.8)
Now let $\Gamma_{0}\in\mathfrak{C}^{+}_{\ell,\beta}$ denote a fixed set of
_compatible positive-type_ contours and let
$\mathcal{A}(\Gamma_{0})=\\{\omega\in\Omega_{\ell,\beta}:\Gamma(\omega)=\Sigma(\Gamma_{0})\\}$
(4.9)
denote the set of link-configurations $\omega$ that induce the set of contours
$\Gamma_{0}$ without adding any new contours. By considering the admissible
set $\Gamma(\omega)$ and its shift
$\Gamma_{0}=\Sigma^{-1}(\Gamma(\omega))=\\{\gamma^{1},\dotsc,\gamma^{k}\\}$,
we conclude that
$\begin{split}Z_{\ell,\beta,n,u}=\,{\rm
e}^{-(1+u)|\overline{E}_{\ell}|}\,\sum_{k\geq
0}&\frac{1}{k!}\int_{X^{+}_{\ell,\beta}}{\rm
d}\bar{\rho}_{1}(\gamma^{1})u^{\\#\gamma^{1}_{\mathchoice{\vbox{\hbox{\includegraphics[width=4.00002pt]{cross.pdf}}}}{\vbox{\hbox{\includegraphics[width=4.00002pt]{cross.pdf}}}}{\vbox{\hbox{\includegraphics[width=3.00003pt]{cross.pdf}}}}{\vbox{\hbox{\includegraphics[width=2.5pt]{cross.pdf}}}}}}\dots\int_{X^{+}_{\ell,\beta}}{\rm
d}\bar{\rho}_{1}(\gamma^{k})u^{\\#\gamma^{k}_{\mathchoice{\vbox{\hbox{\includegraphics[width=4.00002pt]{cross.pdf}}}}{\vbox{\hbox{\includegraphics[width=4.00002pt]{cross.pdf}}}}{\vbox{\hbox{\includegraphics[width=3.00003pt]{cross.pdf}}}}{\vbox{\hbox{\includegraphics[width=2.5pt]{cross.pdf}}}}}}\\\
&\prod_{1\leq i<j\leq
k}\delta(\gamma^{i},\gamma^{j})\int_{\mathcal{A}(\Gamma_{0})}{\rm
d}\bar{\rho}_{u}\big{(}\omega\setminus\Sigma(\Gamma_{0})\big{)}n^{\mathcal{L}(\omega)-\\#\omega_{\mathchoice{\vbox{\hbox{\includegraphics[width=4.00002pt]{dbar.pdf}}}}{\vbox{\hbox{\includegraphics[width=4.00002pt]{dbar.pdf}}}}{\vbox{\hbox{\includegraphics[width=3.00003pt]{dbar.pdf}}}}{\vbox{\hbox{\includegraphics[width=2.5pt]{dbar.pdf}}}}}}.\end{split}$
(4.10)
We also used Remark 3.2, which tells us that crosses are never ‘shared’
between distinct contours or between a contour and a short loop. (Note that in
the last integral, we have the measure $\bar{\rho}_{u}$ rather than
$\bar{\rho}_{1}$ as in the other integrals.)
Next, applying Lemma 3.10 to write
$\mathcal{L}(\omega)-\\#\omega_{\mathchoice{\vbox{\hbox{\includegraphics[width=5.60002pt]{dbar.pdf}}}}{\vbox{\hbox{\includegraphics[width=5.60002pt]{dbar.pdf}}}}{\vbox{\hbox{\includegraphics[width=4.20004pt]{dbar.pdf}}}}{\vbox{\hbox{\includegraphics[width=3.5pt]{dbar.pdf}}}}}=\sum_{i=1}^{k}h(\gamma^{i})$,
we obtain
$\int_{\mathcal{A}(\Gamma_{0})}{\rm
d}\bar{\rho}_{u}\big{(}\omega\setminus\Sigma(\Gamma_{0})\big{)}n^{\mathcal{L}(\omega)-\\#\omega_{\mathchoice{\vbox{\hbox{\includegraphics[width=4.00002pt]{dbar.pdf}}}}{\vbox{\hbox{\includegraphics[width=4.00002pt]{dbar.pdf}}}}{\vbox{\hbox{\includegraphics[width=3.00003pt]{dbar.pdf}}}}{\vbox{\hbox{\includegraphics[width=2.5pt]{dbar.pdf}}}}}}=\prod_{i=1}^{k}n^{h(\gamma^{i})}\int_{\mathcal{A}(\Gamma_{0})}{\rm
d}\bar{\rho}_{u}\big{(}\omega\setminus\Sigma(\Gamma_{0})\big{)}.$ (4.11)
For $\Gamma\in\mathfrak{C}^{+}_{\ell,\beta}$ denote by $\mathfrak{W}(\Gamma)$
the set of columns of $\bar{E}_{\ell,\beta}$ where adding a double bar at any
height would not change the set of contours $\Sigma(\Gamma)$. Using Lemma 3.9
and recalling that $|T_{\beta}|=2\beta$, we get
$\begin{split}\int_{\mathcal{A}(\Gamma_{0})}{\rm
d}\bar{\rho}_{u}\big{(}\omega\setminus\Sigma(\Gamma_{0})\big{)}&=\,{\rm
e}^{(1+u)|F(\Gamma(\omega))|}\,\big{(}\,{\rm e}^{-|T_{\beta}|}\,(\,{\rm
e}^{|T_{\beta}|}\,-1)\big{)}^{|\mathfrak{W}(\Gamma_{0})|}\\\ &=\,{\rm
e}^{(1+u)|\overline{E}_{\ell}^{+}|}\,\Bigl{(}\prod_{\gamma\in\Gamma_{0}}\,{\rm
e}^{-(1+u)\frac{1}{2}|\gamma|}\,\Bigr{)}(1-\,{\rm
e}^{-2\beta}\,)^{|\mathfrak{W}(\Gamma_{0})|}.\end{split}$ (4.12)
Denote by $\gamma_{w}(\Gamma)$ the unique winding contour in $\Gamma$, if it
exists, and $\emptyset$ otherwise. For all
$\Gamma\in\mathfrak{C}^{+}_{\ell,\beta}$ we have
$\begin{split}(1-\,{\rm
e}^{-2\beta}\,)^{|\mathfrak{W}(\Gamma)|}&=\sum_{W\subseteq\mathfrak{W}(\Gamma)}\prod_{w\in
W}(-\,{\rm e}^{-2\beta}\,)\\\
&=\sum_{\Gamma^{\prime}\in\mathcal{Y}_{\ell,\beta}}\prod_{\gamma^{\prime}\in\Gamma^{\prime}}(-\,{\rm
e}^{-2\beta}\,)\times\big{(}\hbox{\rm
1\kern-2.70004ptI}\big{\\{}\gamma^{\prime}_{w}\in\mathcal{A}_{\ell,\beta}\big{\\}}\prod_{\gamma\in\Gamma\setminus\\{\gamma_{w}(\Gamma)\\}}\delta(\gamma,\gamma^{\prime}_{w})\big{)}\\\
&=\int_{\mathfrak{X}_{\ell,\beta}}{\rm
d}\bar{\rho}_{1}(\Gamma^{\prime})\bigg{[}\prod_{\gamma^{\prime}\in\Gamma^{\prime}}(-\,{\rm
e}^{-2\beta}\,\hbox{\rm 1\kern-2.70004ptI}\\{\gamma^{\prime}\in
Y_{\ell,\beta}\\})\bigg{]}\\\ &\qquad\times\hbox{\rm
1\kern-2.70004ptI}\big{\\{}\gamma^{\prime}_{w}\in\mathcal{A}_{\ell,\beta}\big{\\}}\prod_{\gamma\in\Gamma\setminus\\{\gamma_{w}(\Gamma)\\}}\delta(\gamma,\gamma^{\prime}_{w})\prod_{\gamma,\gamma^{\prime}\in\Gamma^{\prime}}\hbox{\rm
1\kern-2.70004ptI}\\{\gamma\cap\gamma^{\prime}=\emptyset\\},\end{split}$
(4.13)
where
$\gamma^{\prime}_{w}\equiv\gamma^{\prime}_{w}(\Gamma,\Gamma^{\prime})\mathrel{\mathop{:}}=\gamma_{w}(\Gamma)\cup\mathcal{L}(\Gamma^{\prime})$.
For the last equation we used that $\bar{\rho}_{1}$ is just the counting
measure on subsets of $Y_{\ell,\beta}$ since these loops do not traverse any
links. Intuitively this amounts to summing over “admissible extensions”
$\Gamma^{\prime}$ of some given set of contours $\Gamma$ and assigning a
different weight to these extensions. But instead of integrating over one set
of contours and then another set of contours that are treated differently, we
might also integrate over one set of contours and then decide which weight to
give to each part of the contour. More rigorously, we combine Eqs (4.10),
(4.11), (4.12) and (4.13) to get
$\begin{split}Z_{\ell,\beta,n,u}\,{\rm
e}^{(1+u)|\overline{E}_{\ell}^{-}|}\,&=\int_{\mathfrak{C}^{+}_{\ell,\beta}}{\rm
d}\bar{\rho}_{1}(\Gamma)\bigg{(}\prod_{\gamma\in\Gamma}\tilde{w}(\gamma)\bigg{)}(1-\,{\rm
e}^{-2\beta}\,)^{|\mathfrak{W}(\Gamma)|}\\\
&=\int_{\mathfrak{C}^{+}_{\ell,\beta}}{\rm
d}\bar{\rho}_{1}(\Gamma)\bigg{(}\prod_{\gamma\in\Gamma}\tilde{w}(\gamma)\bigg{)}\int_{\mathfrak{X}_{\ell,\beta}}{\rm
d}\bar{\rho}_{1}(\Gamma^{\prime})\bigg{[}\prod_{\gamma^{\prime}\in\Gamma^{\prime}}(-\,{\rm
e}^{-2\beta}\,\hbox{\rm 1\kern-2.70004ptI}\\{\gamma^{\prime}\in
Y_{\ell,\beta}\\})\bigg{]}\\\ &\qquad\times\hbox{\rm
1\kern-2.70004ptI}\big{\\{}\gamma^{\prime}_{w}\in\mathcal{A}_{\ell,\beta}\big{\\}}\prod_{\gamma\in\Gamma\setminus\\{\gamma_{w}(\Gamma)\\}}\delta(\gamma,\gamma^{\prime}_{w})\prod_{\gamma,\gamma^{\prime}\in\Gamma^{\prime}}\hbox{\rm
1\kern-2.70004ptI}\\{\gamma\cap\gamma^{\prime}=\emptyset\\}\\\
&=\int_{\mathfrak{C}^{+}_{\ell,\beta}}{\rm
d}\bar{\rho}_{1}(\Gamma)\prod_{\gamma\in\Gamma}\sum_{\tilde{\gamma}\in
g(\gamma)}\tilde{w}(\tilde{\gamma})(-\,{\rm
e}^{-2\beta}\,)^{\\#\gamma\setminus\tilde{\gamma}\over 2}.\end{split}$ (4.14)
∎
In what follows we estimate integrals over contours $\gamma$ which intersect a
given point or interval. Winding and non-winding contours are treated
separately; we will actually see that winding contours play a very limited
role for $\beta$ large. We write $X_{\ell,\beta}^{\mathrm{w}}\subseteq
X_{\ell,\beta}$ and $X_{\ell,\beta}^{\mathrm{nw}}\subseteq X_{\ell,\beta}$ for
the sets of winding and non-winding contours, respectively. We also write
$X_{\ell,\beta}^{\mathrm{w}}(k)\subseteq X_{\ell,\beta}$ for the set of
winding contours which traverse exactly $k$ links, and we write
$X_{\ell,\beta}^{\mathrm{nw}}(\bar{v},k)\subseteq X_{\ell,\beta}$ for the set
of non-winding contours $\gamma$, which traverse $k$ links and which visit the
point $\bar{v}\in\overline{V}_{\ell}$. Similarly, if
$I=[(v,s),(v,t)]\subseteq\overline{V}_{\ell,\beta}$ is an interval we write
$X_{\ell,\beta}^{\mathrm{nw}}(I,k)$ for the set of non-winding contours
$\gamma$, which traverse $k$ links and which intersect $I$.
###### Lemma 4.3.
Fix any $\ell\in{\mathbb{N}},\beta>0$ $c>0$ and points
$(v,s),(v,t)\in\overline{V}_{\ell}$ with $s<t$. Write $I=[(v,s),(v,t)]$. Then
we have
$\displaystyle\int_{X_{\ell,\beta}^{\mathrm{nw}}(I,k)}{\rm
d}\bar{\rho}_{1}(\gamma)\,{\rm e}^{-c|\gamma|}\,\leq
8^{k-1}c^{-k}\big{(}1+c|I|\big{)},$ (4.15)
$\displaystyle\int_{X_{\ell,\beta}^{\mathrm{w}}(k)}{\rm
d}\bar{\rho}_{1}(\gamma)\,{\rm
e}^{-c|\gamma|}\,\leq(2\ell+1)^{2\ell+2}(k+1)^{2\ell}8^{k-2}c^{-k}.$ (4.16)
###### Proof.
Let us start with the case of non-winding contours and the case when
$I=\\{\bar{v}\\}$ contains only one point. Elements $\gamma\in
X_{\ell,\beta}^{\mathrm{nw}}(\bar{v},k)$ may be encoded using tuples
$(t_{1},\dots,t_{k},l_{1},\dots,l_{k-1})\in{\mathbb{R}}_{+}^{k}\times(\\{\mathchoice{\vbox{\hbox{\includegraphics[width=8.00003pt]{cross.pdf}}}}{\vbox{\hbox{\includegraphics[width=8.00003pt]{cross.pdf}}}}{\vbox{\hbox{\includegraphics[width=6.00006pt]{cross.pdf}}}}{\vbox{\hbox{\includegraphics[width=5.0pt]{cross.pdf}}}},\mathchoice{\vbox{\hbox{\includegraphics[width=8.00003pt]{dbar.pdf}}}}{\vbox{\hbox{\includegraphics[width=8.00003pt]{dbar.pdf}}}}{\vbox{\hbox{\includegraphics[width=6.00006pt]{dbar.pdf}}}}{\vbox{\hbox{\includegraphics[width=5.0pt]{dbar.pdf}}}}\\}\times\\{\mathtt{L},\mathtt{R}\\}\times\\{\mathtt{1},\mathtt{2}\\})^{k-1}$,
as follows.
* •
Consider a walker started at $\bar{v}$ and travelling upwards until it first
encounters the endpoint of a link; store the vertical distance traversed as
$t_{1}$.
* •
This link can go to the left, $\mathtt{L}$, or to the right $\mathtt{R}$; it
can be a double bar or a cross ; and it can be traversed by loops in
$\gamma$ once, $\mathtt{1}$, or twice, $\mathtt{2}$. Store this information as
$l_{1}$.
* •
Having crossed the link, our walker follows $\gamma$ and records vertical
distances until _previously unexplored_ links as $t_{i}$ and information about
those links as $l_{i}$, as before.
* •
If a loop is closed and there are still links that are traversed twice by
loops in $\gamma$, but have only been visited once by our walker, the walker
continues walking from such a link and recording $t_{i}$ and $l_{i}$ as before
(we fix some arbitrary rule for selecting the link and the direction of
travel).
* •
This procedure is iterated until the entire contour has been traversed.
See Fig. 16 for an illustration.
Figure 16. Illustration of $t_{1},\dots,t_{k},l_{1},\dots,l_{k-1}$. Note, for
example, that $t_{5}$ is not merely the distance between the fourth and fifth
links, and that $t_{6}$ does not readily admit an interpretation as distance
between links at all.
Noting that $t_{1}+\dots+t_{k}\leq|\gamma|$ and that the number of options for
$\\{l_{i}\\}_{i=1}^{k-1}$ is bounded by $8^{k-1}$, we get
$\int_{X_{\ell,\beta}^{\mathrm{nw}}(\bar{v},k)}{\rm
d}\bar{\rho}_{1}(\gamma)\,{\rm e}^{-c|\gamma|}\,\leq
8^{k-1}\int_{0}^{\infty}dt_{1}\dots\int_{0}^{\infty}{\rm d}t_{k}\;\,{\rm
e}^{-c(t_{1}+\dots+t_{k})}\,=8^{k-1}c^{-k}.$ (4.17)
Next, we may apply a similar argument to obtain that, for ${\varepsilon}>0$
small enough,
$\int_{X_{\ell,\beta}^{\mathrm{nw}}((v,t),k)\setminus
X_{\ell,\beta}^{\mathrm{nw}}((v,t+{\varepsilon}),k)}{\rm
d}\bar{\rho}_{1}(\gamma)\,{\rm e}^{-c|\gamma|}\,\leq
8^{k-1}c^{-k+1}\tfrac{1-\,{\rm e}^{-{\varepsilon}c}\,}{c}\leq
8^{k-1}c^{-k+1}{\varepsilon}.$ (4.18)
Indeed, for a contour $\gamma$ which visits $(v,t)$ but not
$(v,t+{\varepsilon})$, we must have $t_{1}\leq{\varepsilon}$ in the encoding
above, and replacing the integral over $t_{1}\in[0,\infty)$ with an integral
over $t_{1}\in[0,{\varepsilon}]$ gives the claim. Next, to deduce (4.15) from
(4.17) and (4.18), we argue as follows. If $\gamma$ visits $I=[(v,s),(v,t)]$,
then either $\gamma$ contains the endpoint $(v,t)$, or there are $r\in(s,t)$
and ${\varepsilon}>0$ such that $(v,r)\in\gamma$ but
$(v,r+{\varepsilon})\not\in\gamma$. Using (4.17), the first possibility
accounts for the first term $8^{k-1}c^{-k}$ in (4.15). The other possibility
accounts for the second term, which one may, for example, see by using a fine
dyadic discretization of the interval $I$ and passing to the limit using
(4.18) and the monotone convergence theorem.
For winding contours $\gamma$, recall that they consist of $r_{1}\leq 2\ell+1$
winding loops with a finite number of contractible long loops attached to at
least one of them. In particular there are at most $r\leq r_{1}\leq 2\ell+1$
winding loops that do not share a link and are not connected via a sequence of
long, but contractible loops. Let us denote these by
$\gamma_{1},\dots,\gamma_{r}$, and the numbers of links they each visit by
$k_{1},\dots,k_{r}$, respectively, where $k=\sum_{i=1}^{r}k_{i}$. There are
$r$ vertices $v_{1},\dots,v_{r}\in V_{\ell}$ such that $\gamma_{i}$ visits
$(v_{i},0)$. Summing over the possibilities for $r$, $v_{1},\dots,v_{r}$, as
well as $k_{1},\dots,k_{r}$, and applying the argument for (4.17) to each
$\gamma_{i}$, we obtain
$\begin{split}\int_{X_{\ell,\beta}^{\mathrm{w}}(k)}&{\rm
d}\bar{\rho}_{1}(\gamma)\,{\rm
e}^{-c|\gamma|}\,\leq\sum_{r=1}^{2\ell+1}{2\ell+1\choose
r}\sum_{k_{1},\dots,k_{r}\geq 0\atop k_{1}+\dots+k_{r}=k}8^{k-2}c^{-k}\\\
&\leq(2\ell+1)(2\ell+1)^{2\ell+1}\max_{1\leq r\leq
2\ell+1}\Bigl{|}\Bigl{\\{}(k_{1},\dots,k_{r})\in{\mathbb{N}}^{r}:\sum_{i}k_{i}=k\Bigr{\\}}\Bigr{|}8^{k-2}c^{-k}\\\
&\leq(2\ell+1)^{2\ell+2}(k+1)^{2\ell}8^{k-2}c^{-k}.\end{split}$ (4.19)
∎
In order to ensure the convergence of the cluster expansion, we need to check
that interactions between contours are small so as to satisfy the Kotecký-
Preiss criterion in Eq. (4.3). For $a_{1},a_{2},b_{1},b_{2}\geq 0$, let us
introduce
$a(\gamma)=a_{1}|\gamma|+a_{2}\\#\gamma,\qquad
b(\gamma)=b_{1}|\gamma|+b_{2}\\#\gamma,$ (4.20)
where $\\#\gamma$ denotes the number of links visited by $\gamma$. Then we
have the following bound.
###### Lemma 4.4 (Kotecký–Preiss criterion).
Let $w(\gamma)$ be as in (4.6). Then there exist $n_{0}$, $u_{0}$, $a_{1}$,
$a_{2}$, $b_{1}$, $b_{2}>0$ (independent of $\ell,n,u$), and
$\beta_{0}(\ell,n)$, such that for $n>n_{0}$, $|u|<u_{0}$, and
$\beta>\beta_{0}(\ell,n)$, we have for any $\ell$ and any $\gamma_{0}\in
X^{+}_{\ell,\beta}$ that
$\int_{X^{+}_{\ell,\beta}}{\rm d}\bar{\rho}_{1}(\gamma)|w(\gamma)|\,{\rm
e}^{a(\gamma)+b(\gamma)}\,(1-\delta(\gamma,\gamma_{0}))\leq a(\gamma_{0}).$
(4.21)
###### Proof.
Let us alleviate the notation by introducing
$\bar{w}(\gamma)\mathrel{\mathop{:}}=|w(\gamma)|\,{\rm
e}^{a(\gamma)+b(\gamma)}\,.$ (4.22)
We use Lemma 3.13 with $\kappa$ such that $\frac{1}{3}-\kappa=\frac{1}{4}$ and
with $u_{0}$ such that $u_{0}^{-\kappa/2}=n_{0}$, and we set
$c_{1}=\tfrac{1-u_{0}}{2}-a_{1}-b_{1},\qquad c_{2}=\tfrac{1}{4}\log
n_{0}-a_{2}-b_{2}.$ (4.23)
Clearly,
$\int_{X^{+}_{\ell,\beta}}{\rm
d}\bar{\rho}_{1}(\gamma)\bar{w}(\gamma)(1-\delta(\gamma,\gamma_{0}))\leq\int_{X^{\mathrm{w}}_{\ell,\beta}}{\rm
d}\bar{\rho}_{1}(\gamma)\bar{w}(\gamma)+\int_{X^{\mathrm{nw}}_{\ell,\beta}}{\rm
d}\bar{\rho}_{1}(\gamma)\bar{w}(\gamma)(1-\delta(\gamma,\gamma_{0})).$ (4.24)
Let us first consider the contribution of winding contours. Notice that
$\hbox{\rm 1\kern-2.70004ptI}\\{\gamma\text{ has a spanning
segment}\\}\leq|\gamma|/(2\beta).$ (4.25)
For $\gamma\in X_{\ell,\beta}^{\mathrm{w}}$, note that
$|w(\gamma)|\leq|g(\gamma)|\tilde{w}_{u=-u_{0}}(\gamma)$ where $g(\gamma)$ is
given in (4.7). Here $|g(\gamma)|\leq 2^{|E_{\ell}|}$, which is some constant
depending only on $\ell$. Hence, also for $\gamma\in X^{\rm w}_{\ell,\beta}$
we get $\bar{w}(\gamma)\leq\,{\rm e}^{-c_{1}|\gamma|}\,\,{\rm
e}^{-c_{2}\\#\gamma}\,$ for $\beta\equiv\beta(\ell,n)$ large enough.
Using Lemma 4.3 and the fact that a winding contour $\gamma$ satisfies
$|\gamma|\geq\tfrac{1}{2}|\gamma|+\beta$, the first term on the right in
(4.24) satisfies
$\begin{split}\int_{X^{\mathrm{w}}_{\ell,\beta}}{\rm
d}\bar{\rho}_{1}(\gamma)\bar{w}(\gamma)&\leq\,{\rm
e}^{-c_{1}\beta}\,\sum_{k\geq 0}\,{\rm
e}^{-c_{2}k}\,\int_{X^{\mathrm{w}}_{\ell,\beta}(k)}{\rm
d}\bar{\rho}_{1}(\gamma)\,{\rm e}^{-\tfrac{c_{1}}{2}|\gamma|}\,\\\
&\leq(2\ell+1)^{2\ell+2}\,{\rm e}^{-c_{1}\beta}\,\sum_{k\geq
0}(k+1)^{2\ell}\big{(}\tfrac{16}{c_{1}}\,{\rm e}^{-c_{2}}\,\big{)}^{k}\leq
c(\ell,n)\,{\rm e}^{-c^{\prime}\beta}\,,\end{split}$ (4.26)
with some absolute constant $c^{\prime}>{1\over 4}$ and $c(\ell,n)<\infty$ for
$n_{0}$ sufficiently large such that the geometric series converges. In
particular (4.26) gets arbitrarily small for $\beta,n_{0}$ large enough.
We now turn to non-winding contours. We have
$\begin{split}\bar{w}(\gamma)&=\,{\rm e}^{-{1+u\over
2}|\gamma|+(a_{1}+b_{1})|\gamma|}\,n^{h(\gamma)}|u|^{\\#\gamma_{\mathchoice{\vbox{\hbox{\includegraphics[width=4.00002pt]{cross.pdf}}}}{\vbox{\hbox{\includegraphics[width=4.00002pt]{cross.pdf}}}}{\vbox{\hbox{\includegraphics[width=3.00003pt]{cross.pdf}}}}{\vbox{\hbox{\includegraphics[width=2.5pt]{cross.pdf}}}}}}\,{\rm
e}^{(a_{2}+b_{2})\\#\gamma}\,\\\ &\leq\,{\rm
e}^{-(\frac{1-u_{0}}{2}-a_{1}-b_{1})|\gamma|}\,n_{0}^{-\frac{1}{4}\\#\gamma}\,{\rm
e}^{(a_{2}+b_{2})\\#\gamma}\,\\\ &=\,{\rm e}^{-c_{1}|\gamma|}\,\,{\rm
e}^{-c_{2}\\#\gamma}\,.\end{split}$ (4.27)
Note that if $\delta(\gamma,\gamma_{0})=0$ then $\gamma$ and $\gamma_{0}$
intersect somewhere on $\overline{V}_{\ell,\beta}$. We may decompose the
subset of $\overline{V}_{\ell,\beta}$ visited by $\gamma_{0}$ as a union of
closed intervals $I_{1},\dotsc,I_{m}$ where $m\leq\\#\gamma_{0}$ is the number
of links of $\gamma_{0}$. Noting also that a non-winding contour $\gamma$ has
at least 5 links, we obtain from Lemma 4.3 that
$\begin{split}\int_{X^{\mathrm{nw}}_{\ell,\beta}}{\rm
d}\bar{\rho}_{1}(\gamma)\bar{w}(\gamma)(1-\delta(\gamma,\gamma_{0}))&\leq\sum_{j=1}^{m}\sum_{k\geq
5}\,{\rm e}^{-c_{2}k}\,\int_{X^{\mathrm{nw}}_{\ell,\beta}(I_{j},k)}{\rm
d}\bar{\rho}_{1}(\gamma)\,{\rm e}^{-c_{1}|\gamma|}\,\\\
&\leq\sum_{j=1}^{m}\sum_{k\geq 5}\,{\rm
e}^{-c_{2}k}\,8^{k-1}c_{1}^{-k}\big{(}1+c_{1}|I_{j}|\big{)}\\\
&\leq\big{(}\tfrac{1}{8}\\#\gamma_{0}+\tfrac{c_{1}}{8}|\gamma_{0}|\big{)}\sum_{k\geq
5}\big{(}\tfrac{8}{c_{1}}\,{\rm e}^{-c_{2}}\,\big{)}^{k}.\end{split}$ (4.28)
The Lemma 4.4 holds true provided that $\beta$ is large enough, and
$\tfrac{1}{8}\sum_{k\geq 5}\big{(}\tfrac{8}{c_{1}}\,{\rm
e}^{-c_{2}}\,\big{)}^{k}\leq a_{2},\qquad\tfrac{c_{1}}{8}\sum_{k\geq
5}\big{(}\tfrac{8}{c_{1}}\,{\rm e}^{-c_{2}}\,\big{)}^{k}\leq a_{1}.$ (4.29)
Both conditions are fulfilled for $n_{0}$ (and therefore $c_{2}$) large
enough. ∎
We will also need an estimate on the integral of contours that contain or
surround a given point.
###### Corollary 4.5.
For any ${\varepsilon}>0$, there exists
$n_{0},u_{0},a_{1},a_{2},b_{1},b_{2}>0$ (independent of $\ell,n,u$) such that
for $n>n_{0}$, $|u|<u_{0}$, $\beta>\beta_{0}(\ell,n)$, we have
$\int_{X^{+}_{\ell,\beta}}{\rm d}\bar{\rho}_{1}(\gamma)|w(\gamma)|\,{\rm
e}^{a(\gamma)+b(\gamma)}\,{\mathchoice{\rm 1\mskip-4.0mul}{\rm
1\mskip-4.0mul}{\rm 1\mskip-4.5mul}{\rm
1\mskip-5.0mul}}\\{(0,0)\in\overline{I(\gamma)}\\}\leq{\varepsilon}.$ (4.30)
###### Proof.
We proceed as in Eq. (4.24), so that it suffices to bound the contribution
from winding and non-winding contours separately. Recall the definition of
$\bar{w}$ in (4.22). Using Eq. (4.26), we can make the contribution from
winding contours arbitrarily small, say ${\varepsilon}/2$, by choosing
$\beta\equiv\beta(\ell,n)$ sufficiently large, i.e. we have
$\int_{X^{+}_{\ell,\beta}}{\rm
d}\bar{\rho}_{1}(\gamma)\bar{w}(\gamma){\mathchoice{\rm 1\mskip-4.0mul}{\rm
1\mskip-4.0mul}{\rm 1\mskip-4.5mul}{\rm
1\mskip-5.0mul}}\\{(0,0)\in\overline{I(\gamma)}\\}\leq\int_{X^{\text{nw}}_{\ell,\beta}}{\rm
d}\bar{\rho}_{1}(\gamma)\bar{w}(\gamma)\hbox{\rm
1\kern-2.70004ptI}\\{(0,0)\in\overline{I(\gamma)}\\}+{{\varepsilon}\over 2}.$
(4.31)
If $\gamma$ is non-winding and has $k$ links, then it must pass by a site at
time 0 at distance less than $k/2$ from $(0,0)$. Thus we have the bound
$\int_{X^{\text{nw}}_{\ell,\beta}}{\rm
d}\bar{\rho}_{1}(\gamma)\bar{w}(\gamma){\mathchoice{\rm 1\mskip-4.0mul}{\rm
1\mskip-4.0mul}{\rm 1\mskip-4.5mul}{\rm
1\mskip-5.0mul}}\\{(0,0)\in\overline{I(\gamma)}\\}\leq\int_{X^{\text{nw}}_{\ell,\beta}}{\rm
d}\bar{\rho}_{1}(\gamma)\bar{w}(\gamma)\\#\gamma\;{\mathchoice{\rm
1\mskip-4.0mul}{\rm 1\mskip-4.0mul}{\rm 1\mskip-4.5mul}{\rm
1\mskip-5.0mul}}\\{(0,0)\in\gamma)\\}.$ (4.32)
This can be shown to be arbitrarily small, say less than ${\varepsilon}/2$,
when $n$ and $\beta$ are large, as in the previous lemma. ∎
Let $\mathcal{C}_{k}$ denote the set of connected (undirected) graphs with
vertex set $\\{1,\dotsc,k\\}$ and define
$\varphi(\gamma_{1},\dotsc,\gamma_{k})=\left\\{\begin{array}[]{ll}1,&\mbox{if
}k=1,\\\ \tfrac{1}{k!}\sum_{G\in\mathcal{C}_{k}}\prod_{ij\in
G}(\delta(\gamma_{i},\gamma_{j})-1),&\mbox{if }k\geq 2,\end{array}\right.$
(4.33)
where the product in the second line is over the edges of $G$. The following
is the main consequence of Theorem 4.1, which holds because of Lemma 4.4.
###### Proposition 4.6 (Cluster expansion of the partition function).
For parameters as in Lemma 4.4, the following sum converges absolutely:
$\Phi_{\ell,\beta}:=\sum_{m\geq 1}\int_{X^{+}_{\ell,\beta}}{\rm
d}\bar{\rho}_{1}(\gamma_{1})\dots\int_{X^{+}_{\ell,\beta}}{\rm
d}\bar{\rho}_{1}(\gamma_{k})\Bigl{(}\prod_{i=1}^{k}w(\gamma_{i})\Bigr{)}\varphi(\gamma_{1},\dotsc,\gamma_{m}),$
(4.34)
and we have that
$\,{\rm
e}^{(1+u)|\overline{E}^{-}_{\ell,\beta}|}\,Z_{\ell,\beta,n,u}=\exp\big{(}\Phi_{\ell,\beta}\big{)}.$
(4.35)
Notice that $\Phi_{\ell,\beta}$ depends on $n$ and $u$ as well.
### 4.3. Dimerization
Let us introduce the (signed) measure $\mu_{\ell,\beta,n,u}$ such that the
integral of a function $f:\Omega_{\ell,\beta}\to{\mathbb{C}}$ is given by
$\mu_{\ell,\beta,n,u}(f)=\frac{1}{Z_{\ell,\beta,n,u}}\int_{\Omega_{\ell,\beta}}{\rm
d}\rho_{u}(\omega)\,n^{\mathcal{L}(\omega)-\\#\omega_{\mathchoice{\vbox{\hbox{\includegraphics[width=4.00002pt]{dbar.pdf}}}}{\vbox{\hbox{\includegraphics[width=4.00002pt]{dbar.pdf}}}}{\vbox{\hbox{\includegraphics[width=3.00003pt]{dbar.pdf}}}}{\vbox{\hbox{\includegraphics[width=2.5pt]{dbar.pdf}}}}}}f(\omega).$
(4.36)
The next theorem can be understood as dimerization in the loop model. Together
with Theorem 2.1 it implies Theorem 1.1.
###### Theorem 4.7.
For any $c>0$, there exist $n_{0},u_{0}>0$ such that for all $n>n_{0}$ and
$|u|<u_{0}$, we have
* (a)
For all $\ell$ even:
$\begin{split}&\liminf_{\beta\to\infty}\mu_{\ell,\beta,n,u}(0\overset{-}{\longleftrightarrow}-1)>1-c,\quad\text{and}\quad\limsup_{\beta\to\infty}|\mu_{\ell,\beta,n,u}(0\overset{+}{\longleftrightarrow}-1)|<c;\\\
&\limsup_{\beta\to\infty}|\mu_{\ell,\beta,n,u}(0\overset{+}{\longleftrightarrow}1)|<c,\quad\text{and}\quad\limsup_{\beta\to\infty}|\mu_{\ell,\beta,n,u}(0\overset{-}{\longleftrightarrow}1)|<c.\end{split}$
(4.37)
* (b)
For all $\ell$ odd:
$\begin{split}&\liminf_{\beta\to\infty}\mu_{\ell,\beta,n,u}(0\overset{-}{\longleftrightarrow}1)>1-c,\quad\text{and}\quad\limsup_{\beta\to\infty}|\mu_{\ell,\beta,n,u}(0\overset{+}{\longleftrightarrow}1)|<c;\\\
&\limsup_{\beta\to\infty}|\mu_{\ell,\beta,n,u}(0\overset{+}{\longleftrightarrow}-1)|<c,\quad\text{and}\quad\limsup_{\beta\to\infty}|\mu_{\ell,\beta,n,u}(0\overset{-}{\longleftrightarrow}-1)|<c.\end{split}$
(4.38)
Notice that the limits $\beta\to\infty$ actually exist; this could be
established using the correspondence with the quantum spin system, where
convergence is clear. This is less visible in the loop model, though, hence
the use of $\limsup$ and $\liminf$ so we do not need to prove it.
###### Proof.
Assume without loss of generality that $\ell$ is odd. The case of $\ell$ even
works similarly. Let $\mathcal{O}$ (for _outside_) denote the event that
$(0,0)$ is not on or inside any contour, that is
$\mathcal{O}=\Big{\\{}\omega\in\Omega_{\ell,\beta}:(0,0)\in\bigcap_{\gamma\in\Gamma(\omega)}E(\gamma)\Big{\\}}.$
(4.39)
By Remark 3.4 and the fact that $\mu_{\ell,\beta,n,u}(1)=1$,
$\mu_{\ell,\beta,n,u}(0\overset{-}{\longleftrightarrow}1)=1-\mu_{\ell,\beta,n,u}({\mathchoice{\rm
1\mskip-4.0mul}{\rm 1\mskip-4.0mul}{\rm 1\mskip-4.5mul}{\rm
1\mskip-5.0mul}}\\{0\overset{-}{\longleftrightarrow}1\\}^{\rm
c})=1-\mu_{\ell,\beta,n,u}(\hbox{\rm
1\kern-2.70004ptI}_{\mathcal{O}^{c}}{\mathchoice{\rm 1\mskip-4.0mul}{\rm
1\mskip-4.0mul}{\rm 1\mskip-4.5mul}{\rm
1\mskip-5.0mul}}\\{0\overset{-}{\longleftrightarrow}1\\}^{\rm c}).$ (4.40)
We also have
$\begin{split}&\mu_{\ell,\beta,n,u}(0\overset{+}{\longleftrightarrow}1)=\mu_{\ell,\beta,n,u}(\hbox{\rm
1\kern-2.70004ptI}_{\mathcal{O}^{c}}{\mathchoice{\rm 1\mskip-4.0mul}{\rm
1\mskip-4.0mul}{\rm 1\mskip-4.5mul}{\rm
1\mskip-5.0mul}}\\{0\overset{+}{\longleftrightarrow}1\\}^{\rm c}),\\\
&\mu_{\ell,\beta,n,u}(0\overset{\pm}{\longleftrightarrow}-1)=\mu_{\ell,\beta,n,u}(\hbox{\rm
1\kern-2.70004ptI}_{\mathcal{O}^{c}}{\mathchoice{\rm 1\mskip-4.0mul}{\rm
1\mskip-4.0mul}{\rm 1\mskip-4.5mul}{\rm
1\mskip-5.0mul}}\\{0\overset{\pm}{\longleftrightarrow}-1\\}^{\rm
c}).\end{split}$ (4.41)
Then Theorem 4.7 follows from the next lemma (Lemma 4.8), with the function
$f$ being an indicator function. ∎
Given a set of compatible contours $\Gamma\in\mathfrak{C}^{+}_{\ell,\beta}$
and its “shifted version” $\Sigma(\Gamma)\in\mathfrak{A}_{\ell,\beta}$, we
identify their contours in the natural way; i.e. for every $\gamma\in\Gamma$
there exists a unique $\Sigma(\gamma;\Gamma)\in\Sigma(\Gamma)$ that is
obtained by shifting $\gamma$. Every $\Gamma\in\mathfrak{C}^{+}_{\ell,\beta}$
can now be uniquely decomposed into
$\Gamma=\Gamma_{0}\dot{\cup}\Gamma\setminus\Gamma_{0}$ with
$\Gamma_{0}\mathrel{\mathop{:}}=\\{\gamma\in\Gamma\mid(0,0)\in\overline{I(\Sigma(\gamma;\Gamma))}\\},$
(4.42)
where $I(\gamma)$ is the interior of $\gamma$, defined in Section 3. This
decomposition will be useful in the proof of the following lemma.
###### Lemma 4.8.
Let $g:\Omega_{\ell,\beta}\to{\mathbb{R}}$ be a function that, for every
$\omega$, only depends on the contours
$\gamma\in\Gamma(\omega)\in\mathfrak{A}_{\ell,\beta}$ that surround or contain
$(0,0)$. Assume that $|g|\leq 1$. Then for every ${\varepsilon}>0$, there
exists $n_{0}\in{\mathbb{N}},u_{0}>0$ such that for all $\ell$, $n>n_{0}$,
$|u|<u_{0}$, and $\beta\equiv\beta(\ell,n)$ large enough,
$|\mu_{\ell,\beta,n,u}(\hbox{\rm
1\kern-2.70004ptI}_{\mathcal{O}^{c}}g)|<{\varepsilon}.$ (4.43)
###### Proof.
We have
$\mu_{\ell,\beta,n,u}(\hbox{\rm
1\kern-2.70004ptI}_{\mathcal{O}^{c}}g)=\frac{Z_{\ell,\beta,n,u}[\mathcal{O}^{c};g]}{Z_{\ell,\beta,n,u}},$
(4.44)
where
$\begin{split}Z_{\ell,\beta,n,u}[\mathcal{O}^{c};g]&=\,{\rm
e}^{-(1+u)|\overline{E}_{\ell,\beta}^{-}|}\,\int_{\mathfrak{C}^{+}_{\ell,\beta}\setminus\\{\emptyset\\}}{\rm
d}\bar{\rho}_{1}(\Gamma_{0})g(\Gamma_{0})\Big{(}\prod_{\gamma\in\Gamma_{0}}w(\gamma)\Big{)}\hbox{\rm
1\kern-2.70004ptI}\Bigl{\\{}(0,0)\in\overline{I(\Sigma(\gamma;\Gamma_{0}))}\;\forall\gamma\in\Gamma_{0}\Bigr{\\}}\\\
&\times\sum_{m\geq 0}{1\over m!}\int_{X_{\ell,\beta}^{+}}{\rm
d}\bar{\rho}_{1}(\gamma_{1})\dots\int_{X_{\ell,\beta}^{+}}{\rm
d}\bar{\rho}_{1}(\gamma_{m})\Bigl{(}\prod_{i=1}^{m}w_{\Gamma_{0}}(\gamma_{i})\Bigr{)}\prod_{1\leq
i<j\leq m}\delta(\gamma_{i},\gamma_{j}).\end{split}$ (4.45)
Notice that the contour weights in the last line depend on $\Gamma_{0}$ and
are defined as
$w_{\Gamma_{0}}(\gamma)\mathrel{\mathop{:}}=w(\gamma)\hbox{\rm
1\kern-2.70004ptI}\\{(0,0)\notin\overline{I(\Sigma(\gamma;\Gamma_{0}))}\\}\prod_{\gamma_{0}\in\Gamma_{0}}\delta(\gamma,\gamma_{0}).$
(4.46)
Intuitively, we first integrate over all contours surrounding $(0,0)$ (after
shifting, they are called $\Gamma_{0}$) and then we integrate out the
remaining contours that are compatible with $\Gamma_{0}$.
The second line of Eq. (4.45) has the structure of a partition function. Since
$|w_{\Gamma_{0}}(\gamma_{i})|\leq|w(\gamma_{i})|$, Lemma 4.4 holds for the
modified weights too, and therefore also the suitable modification of
Proposition 4.6. This is then equal to
$\exp\big{(}\Phi_{\ell,\beta}(\Gamma_{0})\big{)}$ where
$\Phi_{\ell,\beta}(\Gamma_{0})\mathrel{\mathop{:}}=\sum_{m\geq
1}\int_{X^{+}_{\ell,\beta}}{\rm
d}\bar{\rho}_{1}(\gamma_{1})\dots\int_{X^{+}_{\ell,\beta}}{\rm
d}\bar{\rho}_{1}(\gamma_{m})\Bigl{(}\prod_{i=1}^{m}w_{\Gamma_{0}}(\gamma_{i})\Bigr{)}\varphi(\gamma_{1},\dotsc,\gamma_{m}).$
(4.47)
Notice that the sum in $\Phi_{\ell,\beta}(\Gamma_{0})$ converges absolutely.
Then
$\mu(\hbox{\rm
1\kern-2.70004ptI}_{\mathcal{O}^{c}}g)=\int_{\mathfrak{C}^{+}_{\ell,\beta}\setminus\\{\emptyset\\}}{\rm
d}\bar{\rho}_{1}(\Gamma_{0})g(\Gamma_{0})\Big{(}\prod_{\gamma\in\Gamma_{0}}w(\gamma)\Big{)}\\\
\hbox{\rm
1\kern-2.70004ptI}\bigl{\\{}(0,0)\in\overline{I(\Sigma(\gamma;\Gamma_{0}))}\;\forall\gamma\in\Gamma_{0}\bigr{\\}}\exp\big{\\{}\Phi_{\ell,\beta}(\Gamma_{0})-\Phi_{\ell,\beta}\big{\\}}$
(4.48)
Let $\delta_{0}(\gamma,\Gamma_{0})\mathrel{\mathop{:}}=\hbox{\rm
1\kern-2.70004ptI}\\{(0,0)\notin\overline{I(\Sigma(\gamma;\Gamma_{0}))}\\}\prod_{\gamma_{0}\in\Gamma_{0}}\delta(\gamma,\gamma_{0})$
be the indicator function for $\gamma$ being a contour that is compatible with
$\Gamma_{0}$ and should not be part of $\Gamma_{0}$. Then
$\Phi_{\ell,\beta}-\Phi_{\ell,\beta}(\Gamma_{0})=\sum_{m\geq
1}\int_{X^{+}_{\ell,\beta}}{\rm
d}\bar{\rho}_{1}(\gamma_{1})\dots\int_{X^{+}_{\ell,\beta}}{\rm
d}\bar{\rho}_{1}(\gamma_{m})\Big{(}\prod_{i=1}^{m}w(\gamma_{i})\Big{)}\varphi(\gamma_{1},\dots,\gamma_{m})\\\
\hbox{\rm 1\kern-2.70004ptI}\bigl{\\{}\exists i\leq
m:\delta_{0}(\gamma_{i},\Gamma_{0})=0\bigr{\\}}.$ (4.49)
We bound these “corrections coming from contours not in $\Gamma_{0}$” as
follows:
$\begin{split}&|\Phi_{\ell,\beta}-\Phi_{\ell,\beta}(\Gamma_{0})|\\\
&\leq\sum_{m\geq 1}\int_{X^{+}_{\ell,\beta}}{\rm
d}\bar{\rho}_{1}(\gamma_{1})\dots\int_{X^{+}_{\ell,\beta}}{\rm
d}\bar{\rho}_{1}(\gamma_{m})\Big{(}\prod_{i=1}^{m}|w(\gamma_{i})|\Big{)}|\varphi(\gamma_{1},\dots,\gamma_{m})|\hbox{\rm
1\kern-2.70004ptI}\\{\exists i\leq m:\delta_{0}(\gamma_{i},\Gamma_{0})=0\\}\\\
&\leq\int_{X^{+}_{\ell,\beta}}{\rm
d}\bar{\rho}_{1}(\gamma_{1})|w(\gamma_{1})|(1-\delta_{0}(\gamma_{1},\Gamma_{0}))\biggl{(}1+\sum_{m\geq
2}m\int_{X^{+}_{\ell,\beta}}{\rm
d}\bar{\rho}_{1}(\gamma_{2})\dots\int_{X^{+}_{\ell,\beta}}{\rm
d}\bar{\rho}_{1}(\gamma_{m})|\varphi(\gamma_{1},\dots,\gamma_{m})|\biggr{)}\\\
&\leq\int_{X^{+}_{\ell,\beta}}{\rm
d}\bar{\rho}_{1}(\gamma_{1})|w(\gamma_{1})|\,{\rm
e}^{a(\gamma_{1})}\,(1-\delta_{0}(\gamma_{1},\Gamma_{0})).\end{split}$ (4.50)
The last inequality follows from Theorem 4.1 (b). It is easy to see from the
definition of $\delta_{0}$ that
$1-\delta_{0}(\gamma_{1},\Gamma_{0})\leq\sum_{\gamma_{0}\in\Gamma_{0}}(1-\delta(\gamma_{0},\gamma_{1}))+\hbox{\rm
1\kern-2.70004ptI}\\{(0,0)\in\overline{I(\Sigma(\gamma_{1};\Gamma_{0}))}\\}$.
Thus
$|\Phi_{\ell,\beta}-\Phi_{\ell,\beta}(\Gamma_{0})|\leq\sum_{\gamma_{0}\in\Gamma_{0}}a(\gamma_{0})+\int_{X^{+}_{\ell,\beta}}{\rm
d}\bar{\rho}_{1}(\gamma_{1})|w(\gamma_{1})|\,{\rm
e}^{a(\gamma_{1})}\,\hbox{\rm
1\kern-2.70004ptI}\\{(0,0)\in\overline{I(\gamma_{1})}\\}.$ (4.51)
We used Lemma 4.4 for the first summand and translation invariance for the
second summand. Using Corollary 4.5, we can bound the second summand by
${\varepsilon}$, arbitrarily (and uniformly in $\ell,n,u$) small, for
$n_{0},\beta$ large enough.
Plugging these bounds back into (4.48) and using $|g|\leq 1$, we get
$\begin{split}\big{|}\mu_{\ell,\beta,n,u}&(\hbox{\rm
1\kern-2.70004ptI}_{\mathcal{O}^{c}}g)\big{|}\leq\,{\rm
e}^{{\varepsilon}}\,\int_{\mathfrak{C}^{+}_{\ell,\beta}\setminus\\{\emptyset\\}}{\rm
d}\bar{\rho}_{1}(\Gamma_{0})\prod_{\gamma_{0}\in\Gamma_{0}}\big{(}|w(\gamma_{0})|\,{\rm
e}^{a(\gamma_{0})}\,\big{)}\hbox{\rm
1\kern-2.70004ptI}\\{(0,0)\in\overline{I(\Sigma(\gamma_{0};\Gamma_{0}))}\;\forall\gamma_{0}\in\Gamma_{0}\\}\\\
&\leq\,{\rm e}^{{\varepsilon}}\,\sum_{m\geq 1}\int_{X_{\ell,\beta}}{\rm
d}\bar{\rho}_{1}(\gamma_{1})\bar{w}(\gamma_{1})\hbox{\rm
1\kern-2.70004ptI}\\{(0,0)\in\overline{I(\gamma_{1})}\\}\;\dots\int_{X_{\ell,\beta}}{\rm
d}\bar{\rho}_{1}(\gamma_{m})\bar{w}(\gamma_{m})\hbox{\rm
1\kern-2.70004ptI}\\{(0,0)\in\overline{I(\gamma_{m})}\\}\\\ &\leq 2\,{\rm
e}^{{\varepsilon}}\,\sum_{m\geq 1}{\varepsilon}^{m}=2\,{\rm
e}^{{\varepsilon}}\,\bigg{(}{{\varepsilon}\over
1-{\varepsilon}}\bigg{)}.\end{split}$ (4.52)
For the second inequality we use the fact that all admissible
$\Sigma(\Gamma_{0})\in\mathfrak{A}_{\ell,\beta}$ such that
$(0,0)\in\overline{I(\Sigma(\gamma_{0};\Gamma_{0}))}\;\forall\gamma_{0}\in\Gamma_{0}$
can be written uniquely as
$\\{\gamma_{1},\dots,\gamma_{m}\\}\in\mathfrak{A}_{\ell,\beta}$ with
$(0,0)\in\overline{I(\gamma_{m})}$ and $\gamma_{i}$ surrounding $\gamma_{j}$
whenever $i<j$. In particular all $\gamma_{i}$ must surround $(0,0)$. The last
inequality follows from applying Corollary 4.5 for each of the nested
integrals, with a factor of $2$ since we integrate over $X_{\ell,\beta}$
instead of $X_{\ell,\beta}^{+}$.
Since ${\varepsilon}$ can be made arbitrarily small by taking $n_{0},\beta$
large, we get the lemma. ∎
### 4.4. Proof of exponential decay of correlations
We now turn to exponential decay. Theorem 1.2 is an immediate consequence of
the following result about loop correlations.
###### Theorem 4.9.
There exists an $n_{0},u_{0},C,c_{1},c_{2}>0$ (independent of $\ell,n,u$) such
that for $n>n_{0},|u|<u_{0}$, we have
$\left.\begin{array}[]{ll}\big{|}\mu_{\ell,\beta,n,u}\big{(}(x,s)\leftrightarrow(y,t)\big{)}\big{|}\\\
\big{|}\mu_{\ell,\beta,n,u}\big{(}(x,s)\ext@arrow
9999{\arrowfill@\leftarrow\relbar\rightarrow}{}{+}(y,t)\big{)}\big{|}\\\
\big{|}\mu_{\ell,\beta,n,u}\big{(}(x,s)\ext@arrow
9999{\arrowfill@\leftarrow\relbar\rightarrow}{}{-}(y,t)\big{)}\big{|}\end{array}\right\\}\leq
C\,{\rm e}^{-c_{1}|x-y|-c_{2}|s-t|}\,.$ (4.53)
for all $\ell\in{\mathbb{N}}$, all $x,y\in\\{-\ell+1,\dots,\ell\\}$, and all
$s,t\in{\mathbb{R}}$.
###### Proof.
All three bounds can be proved in the same way; here we only discuss the first
one. We closely follow the proof of Lemma 4.8 and assume $x=s=0$ for
notational convenience.
It turns out that the proof for $|x-y|\leq 1$ present uninformative technical
difficulties. On the other hand, exponential decay can be easily proved in the
equivalent quantum model by expanding the trace in the basis of eigenvectors
of the Hamiltonian and by using the existence of a spectral gap (which is
proved in the next section).
So it is enough to consider here $|x-y|>1$. This allows us to write $\hbox{\rm
1\kern-2.70004ptI}\\{(x,s)\leftrightarrow(y,t)\\}(\omega)=\hbox{\rm
1\kern-2.70004ptI}_{\mathcal{O}^{c}}(\omega)g(\omega)$ with a function $g$
such that $|g|\leq 1$ and that only depends on the contours
$\gamma\in\Gamma(\omega)\in\mathfrak{A}_{\ell,\beta}$ that surround or contain
$(x,s)$. (Recall that we assumed $(x,s)$ to be the origin $(0,0)$ — otherwise
one would simply have to redefine $\mathcal{O}$ and $\Gamma_{0}$ to depend on
$(x,s)$.) We proceed as in Lemma 4.8 to get
$\mu_{\ell,\beta,n,u}\big{(}(x,s)\leftrightarrow(y,t)\big{)}=\frac{Z_{\ell,\beta,n,u}[\mathcal{O}^{c};g]}{Z_{\ell,\beta,n,u}},$
(4.54)
where $Z_{\ell,\beta,n,u}[\mathcal{O}^{c};g]$ is given as in Eq. (4.45).
Proceeding exactly the same way, we get the analogue of Eq. (4.52), namely
$\begin{split}&\,{\rm
e}^{c_{1}|x-y|+c_{2}|s-t|}\,|\mu_{\ell,\beta,n,u}(\hbox{\rm
1\kern-2.70004ptI}_{\mathcal{O}^{c}}g)|\leq\,{\rm
e}^{c_{1}|x-y|+c_{2}|s-t|}\,\,{\rm
e}^{{\varepsilon}}\,\int_{\mathfrak{C}^{+}_{\ell,\beta}\setminus\\{\emptyset\\}}{\rm
d}\bar{\rho}_{1}(\Gamma_{0})g(\Gamma_{0})\prod_{\gamma_{0}\in\Gamma_{0}}\big{(}|w(\gamma_{0})|\,{\rm
e}^{a(\gamma_{0})}\,\big{)}\\\ &\hskip 227.62204pt\times\hbox{\rm
1\kern-2.70004ptI}\\{(x,s)\in\overline{I(\Sigma(\gamma_{0};\Gamma_{0}))}\forall\gamma_{0}\in\Gamma_{0}\\}\\\
&\qquad\leq\,{\rm
e}^{{\varepsilon}}\,\int_{\mathfrak{C}^{+}_{\ell,\beta}\setminus\\{\emptyset\\}}{\rm
d}\bar{\rho}_{1}(\Gamma_{0})\prod_{\gamma_{0}\in\Gamma_{0}}\big{(}|w(\gamma_{0})|\,{\rm
e}^{a(\gamma_{0})+b(\gamma_{0})}\,\big{)}\hbox{\rm
1\kern-2.70004ptI}\\{(x,s)\in\overline{I(\Sigma(\gamma_{0};\Gamma_{0}))}\forall\gamma_{0}\in\Gamma_{0}\\}\\\
&\qquad\leq 2\,{\rm e}^{{\varepsilon}}\,\left({{\varepsilon}\over
1-{\varepsilon}}\right).\end{split}$ (4.55)
Here we chose ${\varepsilon}\in(0,1)$ to be a constant, independent of all
other parameters $n,\beta,\ell,u$.
For all $\gamma\in\Gamma_{0}$ such that $g(\Gamma_{0})\neq 0$ (hence
$g(\Gamma_{0})=1$) we have $|\gamma|\geq 2|s-t|$ and $\\#\gamma\geq 2|x-y|$.
Recall the function $b$ of Lemma 4.4. Choosing $c_{1}=2b_{1},c_{2}=2b_{2}$, we
get the second inequality. Corollary 4.5 then allows us to proceed as in Eq.
(4.52), which gives the last inequality. ∎
## 5\. Proof of the spectral gap
We follow the method of Kennedy and Tasaki [12] and show that the method of
cluster expansion can be used to prove the existence of a positive spectral
gap. Indeed, it implies the validity of the following lemma (recall that
$Z_{\ell,\beta}={\operatorname{Tr\,}}\,{\rm e}^{-2\beta H_{\ell}}\,$).
###### Lemma 5.1.
There exists $n_{0},u_{0},c>0$ (independent of $\ell,\beta,n,u$) and
$C_{\ell}$ (independent of $\beta,n,u$) such that for all $n\geq n_{0}$ and
$|u|\leq u_{0}$, we have for all $\beta\geq\frac{1}{2}$ that
$\bigl{|}E_{0}^{(\ell)}+\tfrac{1}{2\beta}\log Z_{\ell,\beta}\bigr{|}\leq
C_{\ell}\,{\rm e}^{-\beta c}\,.$
###### Proof.
We check that, for all $1\leq\beta<\beta^{\prime}$, we have
$\bigl{|}\tfrac{1}{2\beta}\Phi_{\ell\beta}-\tfrac{1}{2\beta^{\prime}}\Phi_{\ell\beta^{\prime}}\bigr{|}\leq
C_{\ell}\,{\rm e}^{-\beta c}\,.$ (5.1)
Since $-\frac{1}{2\beta}\log
Z_{\ell,\beta}=2\ell(1+u)-\frac{1}{2\beta}\log\Phi_{\ell,\beta}$, we get the
lemma by taking the limit $\beta^{\prime}\to\infty$.
Let ${\mathfrak{C}}_{\beta}^{+}$ denote the set of clusters in $V_{\ell}\times
T_{\beta}$, i.e. the sequences of contours
$\Gamma=(\gamma_{1},\dots,\gamma_{k})$, $\gamma_{i}\in X_{\ell,\beta}^{+}$,
such that $\varphi(\Gamma)\neq 0$. For $t\in T_{\beta}$, let ${\mathchoice{\rm
1\mskip-4.0mul}{\rm 1\mskip-4.0mul}{\rm 1\mskip-4.5mul}{\rm
1\mskip-5.0mul}}_{t}(\Gamma)$ be the indicator that the cluster $\Gamma$
crosses the line $V_{\ell}\times\\{t\\}$. Let
${\mathfrak{L}}(\Gamma)\in[0,2\beta]$ be the vertical length of the cluster
$\Gamma$:
${\mathfrak{L}}(\Gamma)=\int_{-\beta}^{\beta}{\mathchoice{\rm
1\mskip-4.0mul}{\rm 1\mskip-4.0mul}{\rm 1\mskip-4.5mul}{\rm
1\mskip-5.0mul}}_{t}(\Gamma)\,{\rm d}t.$ (5.2)
Then, from the cluster expansion of Proposition 4.6,
$\begin{split}\frac{1}{2\beta}\Phi_{\ell,\beta}&=\frac{1}{2\beta}\int_{{\mathfrak{C}}_{\beta}^{+}}{\rm
d}\bar{\rho}_{1}(\Gamma)w(\Gamma)\varphi(\Gamma)=\frac{1}{2\beta}\int_{{\mathfrak{C}}_{\beta}^{+}}{\rm
d}\bar{\rho}_{1}(\Gamma)\int_{-\beta}^{\beta}{\rm d}t\frac{{\mathchoice{\rm
1\mskip-4.0mul}{\rm 1\mskip-4.0mul}{\rm 1\mskip-4.5mul}{\rm
1\mskip-5.0mul}}_{t}(\Gamma)}{{\mathfrak{L}}(\Gamma)}w(\Gamma)\varphi(\Gamma)\\\
&=\frac{1}{2\beta}\int_{-\beta}^{\beta}{\rm
d}t\int_{{\mathfrak{C}}_{\beta}^{+}}{\rm
d}\bar{\rho}_{1}(\Gamma)\frac{w(\Gamma)\varphi(\Gamma)}{{\mathfrak{L}}(\Gamma)}{\mathchoice{\rm
1\mskip-4.0mul}{\rm 1\mskip-4.0mul}{\rm 1\mskip-4.5mul}{\rm
1\mskip-5.0mul}}_{t}(\Gamma)=\int_{{\mathfrak{C}}_{\beta}^{+}}{\rm
d}\bar{\rho}_{1}(\Gamma)\frac{w(\Gamma)\varphi(\Gamma)}{{\mathfrak{L}}(\Gamma)}{\mathchoice{\rm
1\mskip-4.0mul}{\rm 1\mskip-4.0mul}{\rm 1\mskip-4.5mul}{\rm
1\mskip-5.0mul}}_{0}(\Gamma).\end{split}$ (5.3)
We used Fubini’s theorem to exchange the integrals, and time translation
invariance in the last step. The last expression is convenient to cancel terms
for different $\beta$s; for $\beta<\beta^{\prime}$, we have
$\begin{split}\frac{1}{2\beta}\Phi_{\ell,\beta}&-\frac{1}{2\beta^{\prime}}\Phi_{\ell,\beta^{\prime}}=\int_{{\mathfrak{C}}_{\beta}^{+}}{\rm
d}\bar{\rho}_{1}(\Gamma)\frac{w(\Gamma)\varphi(\Gamma)}{{\mathfrak{L}}(\Gamma)}{\mathchoice{\rm
1\mskip-4.0mul}{\rm 1\mskip-4.0mul}{\rm 1\mskip-4.5mul}{\rm
1\mskip-5.0mul}}_{0}(\Gamma)-\int_{{\mathfrak{C}}_{\beta^{\prime}}^{+}}{\rm
d}\bar{\rho}_{1}(\Gamma)\frac{w(\Gamma)\varphi(\Gamma)}{{\mathfrak{L}}(\Gamma)}{\mathchoice{\rm
1\mskip-4.0mul}{\rm 1\mskip-4.0mul}{\rm 1\mskip-4.5mul}{\rm
1\mskip-5.0mul}}_{0}(\Gamma)\\\ &=\int_{{\mathfrak{C}}_{\beta}^{+}}{\rm
d}\bar{\rho}_{1}(\Gamma)\frac{w(\Gamma)\varphi(\Gamma)}{{\mathfrak{L}}(\Gamma)}{\mathchoice{\rm
1\mskip-4.0mul}{\rm 1\mskip-4.0mul}{\rm 1\mskip-4.5mul}{\rm
1\mskip-5.0mul}}_{0}(\Gamma){\mathchoice{\rm 1\mskip-4.0mul}{\rm
1\mskip-4.0mul}{\rm 1\mskip-4.5mul}{\rm
1\mskip-5.0mul}}_{{\mathfrak{L}}(\Gamma)=2\beta}-\int_{{\mathfrak{C}}_{\beta^{\prime}}^{+}}{\rm
d}\bar{\rho}_{1}(\Gamma)\frac{w(\Gamma)\varphi(\Gamma)}{{\mathfrak{L}}(\Gamma)}{\mathchoice{\rm
1\mskip-4.0mul}{\rm 1\mskip-4.0mul}{\rm 1\mskip-4.5mul}{\rm
1\mskip-5.0mul}}_{0}(\Gamma)1_{{\mathfrak{L}}(\Gamma)\geq 2\beta}.\end{split}$
(5.4)
Indeed, the contribution of clusters with ${\mathfrak{L}}(\Gamma)<2\beta$ has
precisely canceled.
Figure 17. The contour $\gamma^{(\eta)}$ used in the proof of Lemma 5.1.
We can use estimates from cluster expansions in order to bound the terms
above. For $\eta>0$, let us introduce the contour $\gamma^{(\eta)}$ that
surrounds the horizontal axis at time 0, as shown in Fig. 17. Its vertical
length goes to 0 as $\eta\to 0$. We have
$a(\gamma^{(\eta)})=2(2\ell-1)a_{1}+4(\ell+1)\eta a_{2}$; and if
$\eta<2\beta$, we have
${\mathchoice{\rm 1\mskip-4.0mul}{\rm 1\mskip-4.0mul}{\rm 1\mskip-4.5mul}{\rm
1\mskip-5.0mul}}_{0}(\Gamma)\leq
1-\delta(\gamma^{(\eta)},\Gamma)\leq\sum_{\gamma\in\Gamma}\bigl{(}1-\delta(\gamma^{(\eta)},\gamma)\bigr{)}.$
(5.5)
Using $1\leq 2\beta\leq\sum_{\gamma\in\Gamma}|\gamma|$, which holds for
contours such that $1_{{\mathfrak{L}}(\Gamma)\geq 2\beta}=1$, we have, using
Lemma 4.4 and the estimate in Theorem 4.1 (c) that
$\begin{split}\,{\rm
e}^{2b_{2}\beta}\,\biggl{|}\int_{{\mathfrak{C}}_{\beta}^{+}}{\rm
d}\bar{\rho}_{1}(\Gamma)&\frac{w(\Gamma)\varphi(\Gamma)}{{\mathfrak{L}}(\Gamma)}{\mathchoice{\rm
1\mskip-4.0mul}{\rm 1\mskip-4.0mul}{\rm 1\mskip-4.5mul}{\rm
1\mskip-5.0mul}}_{0}(\Gamma){\mathchoice{\rm 1\mskip-4.0mul}{\rm
1\mskip-4.0mul}{\rm 1\mskip-4.5mul}{\rm
1\mskip-5.0mul}}_{{\mathfrak{L}}(\Gamma)=2\beta}\biggr{|}\\\
&\leq\int_{{\mathfrak{C}}_{\beta}^{+}}{\rm
d}\bar{\rho}_{1}(\Gamma)|w(\Gamma)|\,{\rm
e}^{b_{2}\sum_{\gamma\in\Gamma}|\gamma|}\,|\varphi(\Gamma)|\sum_{\gamma\in\Gamma}\bigl{(}1-\delta(\gamma^{(\eta)},\gamma)\bigr{)}\\\
&\leq 2(2\ell-1)a_{1}+4(\ell+1)\eta a_{2}.\end{split}$ (5.6)
The other term in the right side of (5.4) can be estimated in the same way,
giving the same bound. This proves Lemma 5.1 with $C_{\ell}=4(2\ell-1)a_{1}$
(we can take $\eta\to 0$) and $c=2b_{2}$. ∎
###### Proof of Theorem 1.4.
With $m_{i}^{(\ell)}$ the multiplicity of the eigenvalue $E_{i}^{(\ell)}$
(satisfying $\sum_{i\geq 0}m_{i}^{(\ell)}=n^{2\ell}$), we have
$Z_{\ell,\beta}=\sum_{i\geq 0}m_{i}^{(\ell)}\,{\rm e}^{-2\beta
E_{i}^{(\ell)}}\,=m_{0}^{(\ell)}\,{\rm e}^{-2\beta
E_{0}^{(\ell)}}\,\biggl{(}1+\sum_{i\geq
1}\tfrac{m_{i}^{(\ell)}}{m_{0}^{(\ell)}}\,\,{\rm
e}^{-2\beta(E_{i}^{(\ell)}-E_{0}^{(\ell)})}\,\biggr{)}.$ (5.7)
Thus
$-\tfrac{1}{2\beta}\log Z_{\ell,\beta}=E_{0}^{(\ell)}-\tfrac{1}{2\beta}\log
m_{0}^{(\ell)}-\tfrac{1}{2\beta}R(\ell,\beta),$ (5.8)
where
$R(\ell,\beta)=\log\biggl{(}1+\sum_{i\geq
1}\tfrac{m_{i}^{(\ell)}}{m_{0}^{(\ell)}}\,\,{\rm
e}^{-2\beta(E_{i}^{(\ell)}-E_{0}^{(\ell)})}\,\biggr{)}\geq\log\biggl{(}1+\tfrac{m_{1}^{(\ell)}}{m_{0}^{(\ell)}}\,\,{\rm
e}^{-2\beta\Delta^{(\ell)}}\,\biggr{)}\geq\,{\rm
e}^{-3\beta\Delta^{(\ell)}}\,,$ (5.9)
for $\beta$ large enough (depending on $\ell$). On the other hand, Lemma 5.1
implies that
$-\tfrac{1}{2\beta}\log Z_{\ell,\beta}=E_{0}^{(\ell)}+R^{\prime}(\ell,\beta),$
(5.10)
where $|R^{\prime}(\ell,\beta)|\leq C_{\ell}\,{\rm e}^{-\beta c}\,$. We then
have
$-\tfrac{1}{2\beta}\log
m_{0}^{(\ell)}-\tfrac{1}{2\beta}R(\ell,\beta)=R^{\prime}(\ell,\beta).$ (5.11)
Using the bound $R(\ell,\beta)\leq n^{2\ell}\,{\rm
e}^{-2\beta\Delta^{(\ell)}}\,$ where $\Delta^{(\ell)}>0$, and looking at the
asymptotic $\beta\to\infty$, we see that $m_{0}^{(\ell)}=1$. Next, using
(5.9), we get
$\tfrac{1}{2\beta}\,{\rm e}^{-3\beta\Delta^{(\ell)}}\,\leq C_{\ell}\,{\rm
e}^{-\beta c}\,$ (5.12)
for all $\beta$ sufficiently large; this implies that
$\Delta^{(\ell)}\geq\frac{1}{3}c$, uniformly in $\ell,n,u$. ∎
## Appendix A The interaction $uT+vP$ when $n$ is even
For $n$ odd, the interactions $uT+vP$ and $uT+vQ$ are related by the unitary
transformation of Eq. (1.4). This holds for models defined on arbitrary graphs
or lattices.
We now discuss the case of $n$ even. As we shall see, we need to restrict
ourselves to bipartite graphs (of which the chain is of course an example). We
work with the $S^{(3)}$-eigenbasis $e_{\alpha}:=|\alpha\rangle$ with
$\alpha=-S,-S+1,\dots,S$. To begin, we define a unitary $V$ by setting
$V|\alpha\rangle=(-1)^{S-\alpha}|-\alpha\rangle.$ (A.1)
With $\psi$ the vector of (1.3) and $\phi$ the vector of (1.5), we have
$\phi=(\hbox{\rm 1\kern-2.70004ptI}\otimes V)\psi.$ (A.2)
Therefore, since $P$ is the projection onto $\phi$,
$P=(\hbox{\rm 1\kern-2.70004ptI}\otimes V)Q(\hbox{\rm
1\kern-2.70004ptI}\otimes V^{*}).$ (A.3)
Since $T\psi=\psi$ and $T\phi=-\phi$, we have $TQT=Q$ and $TPT=P$. Using these
properties we find
$(V\otimes\hbox{\rm 1\kern-2.70004ptI})Q(V^{*}\otimes\hbox{\rm
1\kern-2.70004ptI})=(V\otimes\hbox{\rm
1\kern-2.70004ptI})TQT(V^{*}\otimes\hbox{\rm 1\kern-2.70004ptI})=T(\hbox{\rm
1\kern-2.70004ptI}\otimes V)Q(\hbox{\rm 1\kern-2.70004ptI}\otimes V^{*})T=P.$
(A.4)
Both models are translation-invariant although the unitary that relates them
is not:
$(V\otimes\hbox{\rm 1\kern-2.70004ptI}\otimes V\cdots\otimes\hbox{\rm
1\kern-2.70004ptI})\left[\sum_{x=-\ell+1}^{\ell-1}Q_{x,x+1}\right](V^{*}\otimes\hbox{\rm
1\kern-2.70004ptI}\otimes V^{*}\cdots\otimes\hbox{\rm
1\kern-2.70004ptI})=\sum_{x=-\ell+1}^{\ell-1}P_{x,x+1}.$ (A.5)
Let $\tilde{T}$ be the transformation of the operator $T$. We have
$\tilde{T}=(\hbox{\rm 1\kern-2.70004ptI}\otimes V)T(\hbox{\rm
1\kern-2.70004ptI}\otimes V^{*})=(\hbox{\rm 1\kern-2.70004ptI}\otimes
V)(V^{*}\otimes\hbox{\rm 1\kern-2.70004ptI})T=-(V\otimes V)T.$ (A.6)
Let us summarize the above considerations by the following proposition. We
define the new Hamiltonian
$H_{\ell}^{\prime}=\sum_{x=-\ell+1}^{\ell-1}\bigl{(}u\tilde{T}_{x,x+1}+vQ_{x,x+1}\bigr{)}$.
###### Proposition A.1.
For $n$ even, the interaction $uT+vP$ is unitarily equivalent with
$u\tilde{T}+vQ$. The Hamiltonian $\tilde{H}_{\ell}$ defined in (1.14) is
unitarily equivalent to $H_{\ell}^{\prime}$.
Notice that, when $u=0$, the $Q$-model and the $P$-model are unitarily
equivalent for all $n$. The proposition is stated for chains, but it clearly
holds for arbitrary bipartite graphs.
Next, we derive a loop representation for the model $H_{\ell}^{\prime}$.
###### Proposition A.2.
There exists a function $s(l)$ from the set of loops to $\pm 1$ such that for
all $n\geq 2$,
* (a)
$\displaystyle{\operatorname{Tr\,}}\,{\rm e}^{-2\beta
H_{\ell}^{\prime}}\,=\,{\rm e}^{2\beta(1+u)|E_{\ell}|}\,\int{\rm
d}\rho_{u}(\omega)n^{\mathcal{L}(\omega)-|\omega_{\mathchoice{\vbox{\hbox{\includegraphics[width=4.00002pt]{dbar.pdf}}}}{\vbox{\hbox{\includegraphics[width=4.00002pt]{dbar.pdf}}}}{\vbox{\hbox{\includegraphics[width=3.00003pt]{dbar.pdf}}}}{\vbox{\hbox{\includegraphics[width=2.5pt]{dbar.pdf}}}}}|}\prod_{\text{loop
$l$ in $\omega$}}s(l)$.
* (b)
For $i=1,2,3$,
${\operatorname{Tr\,}}S_{x}^{(i)}S_{y}^{(i)}\,{\rm e}^{-2\beta
H_{\ell}^{\prime}}\,=\tfrac{n^{2}-1}{12}\,{\rm
e}^{2\beta(1+u)|E_{\ell}|}\,\int_{\Omega_{\ell,\beta}}{\rm
d}\rho_{u}(\omega)\,n^{\mathcal{L}(\omega)-|\omega_{\mathchoice{\vbox{\hbox{\includegraphics[width=4.00002pt]{dbar.pdf}}}}{\vbox{\hbox{\includegraphics[width=4.00002pt]{dbar.pdf}}}}{\vbox{\hbox{\includegraphics[width=3.00003pt]{dbar.pdf}}}}{\vbox{\hbox{\includegraphics[width=2.5pt]{dbar.pdf}}}}}|}\\\
\times\bigl{(}{\mathchoice{\rm 1\mskip-4.0mul}{\rm 1\mskip-4.0mul}{\rm
1\mskip-4.5mul}{\rm
1\mskip-5.0mul}}[x\overset{+}{\longleftrightarrow}y]-{\mathchoice{\rm
1\mskip-4.0mul}{\rm 1\mskip-4.0mul}{\rm 1\mskip-4.5mul}{\rm
1\mskip-5.0mul}}[x\overset{-}{\longleftrightarrow}y]\bigr{)}\prod_{\text{loop
$l$ in $\omega$}}s(l).$
Proposition A.2 is stated for chains, but it actually holds for arbitrary
bipartite graphs (unlike Theorem 2.1 which holds for all finite graphs). For
odd $n$ the signs $s(l)$ are all equal to $+1$.
###### Proof.
First, observe that the number of crosses along the trajectory of a loop, is
even (here, if a cross is traversed twice in a loop, it counts as two).
Indeed, the total number of crosses and double-bars along the trajectory is
even because the graph is bipartite; and the number of double-bars is even
because the number of changes in vertical direction is even; so the number of
crosses is also even.
The expansion of the operator $\,{\rm e}^{-2\beta H_{\ell}^{\prime}}\,$ can be
made in terms of configurations $\omega$, and of “space-time spin
configurations” (see [30]). The space-time spin configurations that are
compatible with $\omega$ have the property that their value on a loop is
$\pm\alpha$ for some $\alpha=-S,\dots,S$, the changes of signs occurring when
traversing crosses (and any such choice results in a possible space-time spin
configuration because the number of crosses along a loop trajectory is even).
Proceeding as in Theorem 2.1, we find that
${\operatorname{Tr\,}}\,{\rm e}^{-2\beta H_{\ell}^{\prime}}\,=\,{\rm
e}^{2\beta(1+u)|E_{\ell}|}\,\int{\rm
d}\rho_{u}(\omega)n^{\mathcal{L}(\omega)-|\omega_{\mathchoice{\vbox{\hbox{\includegraphics[width=4.00002pt]{dbar.pdf}}}}{\vbox{\hbox{\includegraphics[width=4.00002pt]{dbar.pdf}}}}{\vbox{\hbox{\includegraphics[width=3.00003pt]{dbar.pdf}}}}{\vbox{\hbox{\includegraphics[width=2.5pt]{dbar.pdf}}}}}|}{\boldsymbol{s}}(\omega),$
(A.7)
where ${\boldsymbol{s}}(\omega)$ is an overall sign:
${\boldsymbol{s}}(\omega)=\pm 1$. Notice that, since $\tilde{T}$ involves a
minus sign, there is no need to change the sign of $u$ in the interaction as
in Theorem 2.1.
Figure 18. Signs arising when traversing crosses. Left: the crosses are
separated by an even number of double bars which yields the factor
$(-1)^{S-\alpha}(-1)^{S+\alpha}=-1$. Right: the crosses are separated by an
odd number of double bars which yields the factor 1.
The signs are due to the action of operators $V$. We can collect the signs for
each loop individually. Consider two successive crosses. If the vertical
direction is the same (which is the case if there is an even number of double-
bars between them), we get the factor
$(-1)^{S-\alpha}(-1)^{S+\alpha}=(-1)^{2S}=-1.$ (A.8)
If the vertical direction is opposite (which is the case if there is an odd
number of double-bars between them), the factor is
$(-1)^{S-\alpha}(-1)^{S-\alpha}=1.$ (A.9)
This is illustrated in Fig. 18. The value of $s(l)$ is the product of these
factors. Notice that the sign does not depend on the value of $\alpha$ in the
loop. This proves item (a) of the proposition.
The spin correlations are the same for all $i=1,2,3$ by symmetry and it is
enough to consider $i=3$. This is identical to [30, Theorem 3.5 (a)] except
for the signs (the claim there was restricted to odd $n$ where $s(l)=+1$).
Using space-time spin configurations, we have
${\operatorname{Tr\,}}S_{x}^{(i)}S_{y}^{(i)}\,{\rm e}^{-2\beta
H_{\ell}^{\prime}}\,=\,{\rm
e}^{2\beta(1+u)|E_{\ell}|}\,\int_{\Omega_{\ell,\beta}}{\rm
d}\rho_{u}(\omega)\,n^{-|\omega_{\mathchoice{\vbox{\hbox{\includegraphics[width=4.00002pt]{dbar.pdf}}}}{\vbox{\hbox{\includegraphics[width=4.00002pt]{dbar.pdf}}}}{\vbox{\hbox{\includegraphics[width=3.00003pt]{dbar.pdf}}}}{\vbox{\hbox{\includegraphics[width=2.5pt]{dbar.pdf}}}}}|}\Bigl{(}\prod_{\text{loop
$l$ in
$\omega$}}s(l)\Bigr{)}\sum_{\sigma\in\Sigma(\omega)}\sigma_{x,0}\sigma_{y,0}.$
(A.10)
We used the fact that the signs do not depend on the spin values of the loops.
If $(x,0)$ and $(y,0)$ do not belong to the same loop, the sum over $\sigma$
is zero. If $(x,0)$ and $(y,0)$ belong to the same loop and the connection is
$x\overset{+}{\longleftrightarrow}y$, then $\sigma_{x,0}=\sigma_{y,0}$ and the
sum gives $\frac{3}{S}(S+1)n=\frac{1}{12}(n^{2}-1)n$. If the connection is
$x\overset{-}{\longleftrightarrow}y$, then $\sigma_{x,0}=-\sigma_{y,0}$ and we
get minus the same factor. This gives the identity (b). ∎
We can now prove Theorem 1.3.
###### Proof of Theorem 1.3.
By Proposition A.1, the claims of Theorem 1.3 are equivalent to proving
dimerization in the model with Hamiltonian $H_{\ell}^{\prime}$. We use the
loop representation of Proposition A.2. We can then retrace the steps of the
proof of Theorem 4.7. In doing so, note that all short loops $l$ have
$s(l)=+1$, while long or winding loops have $s(l)=\pm 1$. We incorporate the
latter factors in the weights $w(\gamma)$ of the contours, see (4.6).
Therefore, the only difference is that the weights of contours have possibly
other signs. All bounds are the same, though, and the cluster expansion gives
the same result. ∎
The proof of the gap for $\tilde{H}_{\ell}$ is exactly the same as the proof
for $H_{\ell}$ described in Section 5.
Acknowledgements: We are grateful to Vojkan Jakšić and the Centre de
Recherches Mathématiques of Montreal for hosting us during the thematic
semester “Mathematical challenges in many-body physics and quantum
information”, with support from the Simons Foundation through the Simons–CRM
scholar-in-residence program. We also thank the referees for useful comments.
JEB gratefully acknowledges support from _Vetenskapsrådet_ grants 2015-0519
and 2019-04185 as well as _Ruth och Nils-Erik Stenbäcks stiftelse_.
BN is supported in part by the National Science Foundation under grant
DMS-1813149.
## References
* [1]
* [2] I. Affleck, _Exact results on the dimerisation transition in $su(n)$ antiferromagnetic chains_, J. Phys.: Condens. Matter 2, 405–415 (1990)
* [3] I. Affleck, T. Kennedy, E.H. Lieb, H. Tasaki, Valence bond ground states in isotropic quantum antiferromagnets, Comm. Math. Phys. 115, 477-528 (1988)
* [4] M Aizenman, H. Duminil-Copin, S. Warzel, Dimerization and Néel order in different quantum spin chains through a shared loop representation, Ann. Henri Poincaré 21, 2737–2774 (2020)
* [5] M. Aizenman, B. Nachtergaele, Geometric aspects of quantum spin states, Comm. Math. Phys. 164, 17–63 (1994)
* [6] H. Babujian, Exact solution of the one-dimensional isotropic Heisenberg chain with arbitrary spins S, Phys. Lett. 90A, 479–482 (1982)
* [7] M.N. Barber, M.T. Batchelor, _Spectrum of the biquadratic spin-1 antiferromagnetic chain_ , Phys. Rev. B 40, 4621–4626 (1989)
* [8] H. Duminil-Copin, J.-H. Li, I. Manolescu, Universality for the random-cluster model on isoradial graphs, Electron. J. Probab. 23, 1–70 (2018)
* [9] F.J. Dyson, E.H. Lieb, B. Simon, Phase transitions in quantum spin systems with isotropic and nonisotropic interactions, J. Statist. Phys. 18, 335–383 (1978)
* [10] M.B. Hastings, T. Koma, Spectral gap and exponential decay of correlations, Commun. Math. Phys. 265, 781–804 (2006)
* [11] C. Itoi, M.-H. Kato, Extended massless phase and the Haldane phase in a spin-1 isotropic antiferromagnetic chain, Phys. Rev. B 55, 8295–8303 (1997)
* [12] T. Kennedy, H. Tasaki, Hidden symmetry breaking and the Haldane phase in $S=1$ quantum spin chains, Commun. Math. Phys. 147, 431–484 (1992).
* [13] A. Klümper, _The spectra of $q$-state vertex models and related antiferromagnetic quantum spin chains the spectra of q-state vertex models and related antiferromagnetic quantum spin chains_, J. Phys. A: Math. Gen. 23, 809–823 (1990)
* [14] R. Kotecký, D. Preiss, Cluster expansion for abstract polymer models, Commun. Math. Phys. 103, 491–498 (1986)
* [15] L.J. Landau, J. F. Perez, W.F. Wreszinski, Energy gap, clustering, and the Goldstone theorem in statistical mechanics, J. Stat. Phys. 26, 755–766 (1981)
* [16] B. Lees, _Existence of Néel order in the S=1 bilinear-biquadratic Heisenberg model via random loops_ , Commun. Math. Phys. 347, 83–101 (2016)
* [17] B. Nachtergaele, R. Sims, Lieb-Robinson Bounds and the Exponential Clustering Theorem, Commun. Math. Phys. 265, 119–130 (2006)
* [18] B. Nachtergaele, R. Sims, A. Young, Quasi-Locality Bounds for Quantum Lattice Systems. Part II. Perturbations of Frustration-Free Spin Models with Gapped Ground States, arXiv:2010.15337 (2020).
* [19] B. Nachtergaele, D. Ueltschi, A direct proof of dimerization in a family of $SU(n)$-invariant quantum spin chains, Lett. Math. Phys. 107, 1629–1647 (2017)
* [20] R.I. Nepomechie, R.A. Pimenta, _Universal Bethe ansatz solution for the Temperley-Lieb spin chain_ , Nucl. Phys. B 910, 910–928 (2016)
* [21] G. Ray, Y. Spinka, A short proof of the discontinuity of phase transition in the planar random-cluster model with q ¿ 4, Commun. Math. Phys. 378, 1977–1988 (2020)
* [22] N.Yu. Reshetikhin, A method of functional equations in the theory of exactly solvable quantum systems, Lett. Math. Phys. 7, 205–213 (1983)
* [23] K. Ryan, The free energy of a class of spin 1/2 and 1 quantum spin systems on the complete graph, arXiv:2011.07007 (2020)
* [24] B. Sutherland, Model for a multicomponent quantum system, Phys. Rev. B 12, 3795–3805 (1975)
* [25] L.A. Takhtajan, The picture of low-lying excitations in the isotropic Heisenberg chain of arbitrary spins, Phys. Lett. 87A, 479–482 (1982)
* [26] K. Tanaka, A. Tanaka, T. Idokagi, Long-range order in the ground state of the S=1 isotropic bilinear-biquadratic exchange Hamiltonian, J. Phys. A 34, 8767–8780 (2001)
* [27] B. Tóth, Improved lower bound on the thermodynamic pressure of the spin $1/2$ Heisenberg ferromagnet, Lett. Math. Phys. 28, 75–84 (1993)
* [28] H.-H. Tu, G.-M. Zhang, T. Xiang, Class of exactly solvable $SO(n)$ symmetric spin chains with matrix product ground states, Phys. Rev. B, 78, 094404 (2008)
* [29] D. Ueltschi, Cluster expansions and correlation functions, Moscow Math J. vol 4 no 2, p. 511–522 (2004)
* [30] D. Ueltschi, Random loop representations for quantum spin systems, J. Math. Phys. 54, 083301, 1–40 (2013)
|
# Hiding Behind Machines:
When Blame Is Shifted to Artificial Agents
Till Feier TUM School of Governance, TU Munich, Richard-Wagner-Straße 1, 80333
Munich, Germany<EMAIL_ADDRESS>Jan Gogoll Bavarian Institute for Digital
Transformation / TU Munich, Gabelsbergerstr. 4, 80333 Munich, Germany,
<EMAIL_ADDRESS>Matthias Uhl ZD.B Junior Research Group “Ethics of
Digitization”, TUM School of Governance, TU Munich, Richard-Wagner-Straße 1,
80333 Munich, Germany<EMAIL_ADDRESS>
###### Abstract
The transfer of tasks with sometimes far-reaching moral implications to
autonomous systems raises a number of ethical questions. In addition to
fundamental questions about the moral agency of these systems, behavioral
issues arise. This article focuses on the responsibility of agents who decide
on our behalf. We investigate the empirically accessible question of whether
the production of moral outcomes by an agent is systematically judged
differently when the agent is artificial and not human. The results of a
laboratory experiment suggest that decision-makers can actually rid themselves
of guilt more easily by delegating to machines than by delegating to other
people. Our results imply that the availability of artificial agents could
provide stronger incentives for decision makers to delegate morally sensitive
decisions.
> “Once the rockets are up, who cares where they come down? That’s not my
> department” says Wernher von Braun.
> – Tom Lehrer
## 1 Introduction
In 2017, the airline AirBerlin, Lufthansa’s biggest competitor for German
domestic flights, went bankrupt. Lufthansa passengers soon noticed a
significant spike in ticket prices. The resulting backlash was enormous and
subsequently led to an investigation by Germany’s Federal Cartel Office (FCO).
Lufthansa was quick to blame its automated booking system, which had allegedly
responded to a spike in demand. This did little to quell the public outrage
and FCO president Andreas Mundt famously said that “companies can’t hide
behind algorithms” (Busse, 2017). Our article tests the empirical content of
this normative statement.
Algorithms have become an integral part of society and are responsible for
more tasks than ever before (Manyika et al., 2017). This is not only pushing
the frontiers of technology but also challenging our traditional concepts of
guilt and responsibility. The diffusion of responsibility between a multitude
of actors has long been a trademark of modern institutions. The challenge of
locating responsibility within a complex system may even become exponentially
more difficult when moral agency is distributed between human operators and
autonomous agents (Nissenbaum, 1996). The fact that even the designers of
algorithms cannot fully explain their decisions (e.g., black box models in
machine learning) adds to this complexity (Rudin, 2019).
Normative ethics raise a fundamental question with respect to the increasing
use of artificial agents in decision-making. In which sense can artificial
agents also be moral agents and therefore be responsible for their actions?
Philosophers and engineers started pondering this question decades ago when
computers were merely functioning as calculators (Moor, 1979). There is still
little consensus on the matter and the idea of “moral machines” remains under
debate (Allen and Wallach, 2012). In any case, the philosophical discussion is
predominantly concerned with normative aspects and the question of who ought
to be held responsible or blamed if a machine brings about moral evil.
Although these normative questions are of obvious importance, the behavioral
impact that the interaction with artificial agents has on the operator’s
conduct is not well understood either. It seems crucial, however, for research
on behavioral consequences to feed back into the ethical debate. Human
operators might get trapped between the increasing capabilities and autonomy
of artificial agents on one hand and our rigid moral understanding of guilt
and responsibility on the other. Operators may end up merely filling
responsibility gaps in various systems instead of actually being in charge.
They might also over-utilize artificial agents even when they are not the
ideal choice for a given task, because they wish to use machines as scapegoats
(Danaher, 2016).
This study investigates whether delegators will be able to successfully shift
blame by delegating tasks to artificial agents. Our investigation focuses on
whether people judge delegations to human and artificial agents differently in
light of given outcomes. We find no differences between judgments toward human
and artificial agents in the event of good outcomes. This means that the
morally beneficial delegation to an artificial agent was considered neither
better nor worse than the beneficial delegation to a human agent. However, if
a bad outcome occurred, delegators fared significantly better if a machine
agent caused the failure. Interestingly, participants did not seem to
anticipate this pattern as we did not find significant differences regarding
delegation decisions themselves. In fact, decisions involving human and
artificial agents seem driven by the expected utility of the delegation for
the affected party.
Our article proceeds as follows. In Section 2, we will derive our research
question. In section 3, we will outline the experimental setup to test it. We
will discuss our results in Section 4 and conclude in Section 5.
## 2 Background and Research Question
A substantial amount of literature is available on the parameters that
influence automation use in teams of human supervisors and the machines at
their disposal (Dzindolet et al., 2002). In recent years, a number of studies
investigated how those parameters change once decisions have moral undertones,
i.e., an impact on third parties. Goldbach et al. (2019) found people to be
hesitant to delegate decisions to algorithms that affect both the decision-
maker herself as well as third parties. Similarly, a study by Niszczota and
Kaszás (2020) suggests that algorithm aversion extends to the financial
sector, and that people especially prefer human over artificial agents when it
comes to making financial decisions with moral undertones. In a laboratory
study, Gogoll and Uhl (2018) identified a strong aversion against delegating
morally relevant tasks to algorithms. It seems that people were less willing
to delegate tasks to machines if those decisions imposed monetary
externalities on third parties. While their study checked for “perceived
utility” of the artificial agent and the trust in that agent, they were unable
to determine the exact causes of the profound algorithm aversion.
Other studies also support the idea that a lack of trust is unlikely to be the
cause of aversion towards machine use. If anything, there seems to be over-
reliance and over-trust in machines, even if the lives of people are at stake
(Robinette et al., 2016). This suggests that there have to be other causes for
machine aversion in decisions with moral implications. We hypothesize that
responsibility is a key concept in understanding this phenomenon. Perceived
responsibility is already an important research topic in relation to machine
use, probably most prominently in the context of automated driving (Hevelke
and Nida-Rümelin, 2015). However, little attention has been paid to the
question of how the introduction of machine agents might affect the blame and
praise that people ascribe to the delegator in light of a given outcome.
Understanding how the use of artificial agents is judged would be an important
first step in gaining a deeper understanding of how their availability might
influence people’s motivation to delegate morally sensible tasks. This is
closely linked to the idea that avoiding blame or shirking responsibility can
be pivotal factors in delegation decisions.
Strategies of blame avoidance and responsibility shifting have long been
discussed in the political sciences (Weaver, 1986). Instances of so-called
blame games can frequently be observed in political systems with regard to
policy making and implementation in the European Union (Heinkelmann-Wild and
Zangl, 2020) or between officials from different levels of government in the
United States (Maestas et al., 2008). Unsurprisingly, similar strategies can
also be observed in the private sector, for instance, when it comes to
upholding employment standards within franchise networks (Hardy, 2019). The
idea that responsibility shifting can in fact be a pivotal factor in
delegation decisions plays an especially prominent role in public choice
theory (Fiorina, 1986).
Experimental evidence of this phenomenon comes from Fischbacher et al. (2008),
who showed that responsibility attribution can sometimes be effectively
shifted and that this constitutes a powerful motive for decision-makers. Other
experiments provide evidence that this is true even if the agent in question
is effectively powerless (Hill, 2015), or the delegation by the principal
eliminated the possibility of a fair outcome (Oexl and Grossman, 2013). But
does this also hold true for artificial agents?
Popular culture places a strong emphasis on human responsibility despite an
empirical decline of human control in many areas (Elish and Hwang, 2015). The
concept of “algorithmic outrage asymmetry” suggests that people are less
morally outraged by algorithmic wrongdoing than by human wrongdoing, for
instance, in cases of discrimination by age, race or gender (Bigman et al.,
2020). If, however, people seek to attribute blame, some argue that it will
not be placed on algorithms but that humans could emerge as “moral crumple
zones.” In human-machine teams, humans would then have to take on blame even
for accidents outside of their control (Elish, 2019).
While some argue that humans will bear all of the responsibility and none of
the control when working with machines, others think that machines are
perfectly suited to be used as scapegoats. So far very limited empirical
evidence has been provided to back either position. Strobel and Kirchkamp
(2017) investigated whether choices in a dictator game and perceived guilt
change when players can share responsibility with machines. The authors report
that perceived responsibility and guilt did not vary significantly when
comparing purely human teams with human-machine teams. They write that people
tend to make fewer selfish decisions when partnered with machines, though that
effect was insignificant.
So while responsibility shifting has long been established as an integral part
in delegation decisions, empirical evidence of whether the introduction of
artificial agents is rendering this motive mute or even more important is
lacking. This constitutes a serious research gap, the implications of which
exceed academic relevance. A better understanding of responsibility shifting
to artificial agents could explain over and underreliance on machines and
profoundly influence legal decision-making regarding automation and
digitization. For instance, administrators are struggling to provide
governance strategies for automated vehicles, because of the ambiguity with
respect to liability (Taeihagh and Lim, 2019). A better understanding of how
people actually attribute responsibility is likely to help create guidelines
that are not only more effective but also more likely to gain consensus.
Deeper insights into the phenomenon might also help to shield human operators
from unjust recrimination. As mentioned above, our classic understanding of
human-machine teams has cemented a focus on human responsibility despite a
decline of human control in various areas (Elish and Hwang, 2015). This is
especially troubling since responsibility has proven to be be an important
factor in understanding and predicting punishment patterns. (for example (Fehr
and Schmidt, 1999; Bolton and Ockenfels, 2000)) It seems that human operators
are in fact in harms way and might become scapegoats in cases of technical
failure.
In contrast, machine use might emerge as a strategy for self-exculpation in
critical situations. This could have detrimental effects if it leads to an
overuse of machines - a bleak prospect given the growing capabilities of
algorithms and the potential harm this implies for workers and consumers. In
sum, there are plenty of reasons to investigate attributions of blame and
praise in the context of automation.
To shed light on this problem, we will test the following conjecture in a
laboratory experiment.
Conjecture: Delegators are rewarded differently for delegating other-regarding
tasks to artificial as compared to human agents.
On one hand, the delegation of a task that could carry severe consequences for
a third party to an unmonitored machine might be considered careless and
result in punishment or defamation. On the other hand, principals might be
exculpated entirely since people deem the failure of machines to be even more
outside of their control than when delegating to another human. Either way, a
better understanding of public reservations regarding the introduction of
novel technology is likely to prove useful in future moral and legal
considerations. The experiment designed to test the above conjecture will be
outlined in the following chapter.
Additionally, we will explore whether the effect of perceived utility on
delegation decisions varies depending on the agent’s artificial or human
nature. Based on a definition by Dzindolet et al. (2002), we define the
perceived utility of employing an artificial agent as the difference between
the perceived reliability of an automated device and the perceived reliability
of manual control. Furthermore, we will include risk attitudes into our
analysis (O’Donoghue and Somerville, 2018).
## 3 Design
The experiment consisted of (1) a logic task, (2) a delegation decision, (3)
an evaluation of the delegation decision, (4) a self-assessment of one’s
performance and elicitation of risk attitudes. Subjects received a 40 ECU
show-up fee increased by the outcome of the delegation (successful or not),
plus (minus) their reward (their punishment) for their decision to delegate or
not and the bonus for the accuracy of their self-assessment and the lottery as
payment for the experiment. An experimental currency unit was used (ECU) with
the exchange rate of 10 ECU for 1 EUR.
Subjects received their instructions on-screen and were fully informed about
the rules of the game. They were randomly matched according to perfect
stranger matching, i.e., no two participants interacted more than once during
the experiment. The experimental manipulation consisted of changing the nature
of the agent to which a task could be delegated: it was either another human
participant or an artificial agent (see 3.2). Table 1 gives an overview over
the four stages of the experiment.
Stage | Human (Machine) Treatment
---|---
(1) Solve logic task | Participants solve a series of logic puzzles
(2) Make delegation decision | Participants decide whether to delegate to another human (a machine) or to have their own work count
(3) Evaluate delegation decision | Participants reward or punish decision to delegate or not
(4) Self-assess & choose lotteries | Participants guess how many errors they made in logic task and reveal their risk attitude by choosing between lotteries
Table 1: Overview of experimental stages
### 3.1 Logic Task
The main task of the experiment was a logic puzzle. We asked participants to
complete ten puzzles within a five-minute time frame. Each puzzle consisted of
a sequence of three patterns and a placeholder for the missing fourth pattern.
The right answer had to be derived from the given sequence and selected from a
set of four alternatives. To identify the right answer, participants always
had to focus on the circles, while ignoring differently shaped symbols and any
colors (see Figure 1). Because we assumed that delegation decisions would
heavily depend on the agent’s perceived capabilities, we chose a task
involving visual perception to foster participants’ intuition that the
algorithm could err. Participants had to complete ten of the tasks listed
above before they could continue to the second stage of the experiment.
Figure 1: Example of a logic task
### 3.2 Delegation Decision
After the logic task had been completed, participants were randomly assigned
to one of two treatments: the “human” or the “machine” treatment. Only at this
point did subjects learn about the nature of the agent to whom they could
delegate the task.
To give participants an idea of their potential delegatee’s capability, they
were shown representations of the agent’s performance. In the human treatment,
the $n$ participants were shown a histogram displaying the performance of the
other $(n-1)$ participants based on the actual results of the running session.
In the machine treatment, participants were shown a histogram with information
about how often the algorithm failed to give correct answers in $n$ trial
runs. The algorithm was programmed such that it mirrored the performance of
the human participants in the room. The performance of the artificial agent
was therefore as good as that of the participants in the respective session.
This process ensured that the delegation decisions were based on the agent’s
nature instead of any assumptions about differing capabilities.
Participants were then asked to make the delegation decision. They chose
whether their own performance or that of their human or artifical agent
(depending on the treatment) would determine the third party’s payoff. The
delegation decision would thus not affect the payment of the delegator but
that of another participant. Conditional on participants’ decisions to
delegate or not, one of the ten solutions to the ten puzzles was randomly
chosen from the answers of their agent or their own answers. If the selected
puzzle was solved correctly, the third party received an additional payoff. If
it was solved incorrectly, the third party did not receive an additional
payoff.
Figure 2: Delegation decision for human and machine treatment
### 3.3 Evaluation of the Delegation
Participants were now given the opportunity to punish or reward another
participant’s decision to delegate or not. Participants did not know the
actual decision of the participant that they evaluated, nor whether this
decision had resulted in a good or bad outcome for themselves. They were asked
to alter the payment of the participant upwards or downwards by at most 40
ECU. Reward and punishment choices were contingent on the two possible
decisions to delegate or not and the two possible outcomes of success or
failure. Thus, in any case, four choices had to be made. Only the one that
applied to the actual decision of the evaluated participant combined with
actual outcome did finally apply (Selten, 1967). Table 2 depicts the table
that subjects saw on their screen.
Lower or increase subjects payoff given that: | Amount
---|---
Subject used own work – outcome: success | *enter amount*
Subject used own work – outcome: failure | *enter amount*
Subject delegated (machine/human) – outcome: success | *enter amount*
Subject delegated (machine/human) – outcome: failure | *enter amount*
Table 2: Decision to increase or decrease the delegator’s pay-off
### 3.4 Self-Assessment and Risk Attitudes
Subsequently, participants were asked to assess their own performance by
estimating how many mistakes they had made during the logic task in the
experiment’s first part. This estimate was incentivized by an additional
payment of 50 ECU if they guess correctly. This procedure allowed us to
analyze the effect of self-assessments on delegation decisions. As is standard
in incentivized economic experiments, the experiment was concluded by an
elicitation of participants’ risk attitudes.111We used the procedure
introduced by Holt and Laury (2002). The task is based on ten choices between
pairs of lotteries. The potentional payoffs for the safe lottery range from
€2.5 to €1.6 and are therefore always less extreme than those for the risky
lottery that range from €4.35 to €0.10. For the first pair of lotteries, the
probability of the high payoff is equally low in both lotteries but equally
increases for both lotteries with each new pair. A participant should switch
to the risky lottery once the probability for the high payoff is sufficiently
high according to his or her personal risk attitude. The later a participant
switches to the risky lottery, the more risk averse he or she is.
## 4 Results
The experiment took place in a major German university between February and
May 2019. A total of 149 subjects participated in six sessions, 43% were
female and the average age was 23.08 years (sd = 3.83). Participants received
a show-up fee of €4.00 and could earn additional money in the experiment. A
session lasted about 45 minutes and the average payment was about €13.50 per
participant. The experiment was programmed in z-Tree (Fischbacher, 2007) and
subjects were recruited via ORSEE (Greiner et al., 2004). Data analysis was
conducted using Python’s numpy, scipy, and statsmodels.api libraries. The
preprocessed data set and the code are available
online.222https://doi.org/10.5281/zenodo.4446581
Remember that after the delegation decision, all participants were asked to
evaluate the decision of the participant who was responsible for their own
pay-off. Responsibility arose from the decision to either delegate (to a human
or machine) or rely on one’s own work (in both treatments). Notice again that
participants were informed whether they were randomly assigned to the human or
machine treatment, but not whether the other participant had actually
delegated or not. They were therefore asked to judge the decision with respect
to the four possible outcomes according to the so-called strategy method (see
Section 3.3). This means that subjects could increase (or deduct) the payment
of their responsible participant by an integer between 0 and 40 for the cases
of her (1) having brought about a good outcome herself or (2) having brought
about a bad outcome herself. Furthermore, they were increasing (or deducting)
the payment for the case of her (3) having delegated to an agent causing a
good outcome on her behalf or (4) having delegated to an agent causing a bad
outcome on her behalf.
To test our conjecture that delegators are rewarded differently for delegating
other-regarding tasks to artificial as opposed to human agents, we contrast
evaluations between both treatments.
Let us first consider the good-outcome case. In the human treatment,
participants that did not delegate and caused the good outcome themselves
were, on average, rewarded 20.28 ECU (sd = 20.85). Participants that delegated
to another human who then brought about the good outcome on their behalf were
rewarded 20.94 ECU (sd = 21.34). This difference is insignificant (p = 0.720,
paired t-test).
In the machine treatment, participants that did not delegate and caused the
good outcome themselves were on average rewarded 23.78 ECU (sd = 20.14).
Participants that delegated to a machine were rewarded 27.00 ECU (sd = 16.71).
This difference is again insignificant (p = 0.106, paired t-test).
Result 1: Delegators did not lose any recognition for a good outcome if it was
caused by either their human or their artificial agent.
Let us now consider the bad-outcome case. In the human treatment, participants
that did not delegate and caused the bad outcome themselves were on average
rewarded 7.29 ECU (sd = 22.59 ECU). Participants that delegated to another
human were on average rewarded with 8.26 ECU (sd = 21.56 ECU). This difference
is insignificant (p = 0.559, paired t-test).
In the machine treatment, participants that did not delegate and caused the
bad outcome themselves were on average rewarded 8.53 ECU (sd = 26.52).
Participants that delegated to a machine were on average rewarded 12.96 ECU
(sd = 24.44). This difference is significant (p = 0.041, paired t-test).
In the machine treatment, delegators earned higher rewards if a machine agent
caused the failure. This confirms our conjecture stated in Section 2.
Principals are rewarded differently for delegating tasks depending on the
nature of the agent. More specifically, they fare better if a machine agent
caused a bad outcome than if they did so themselves.
Result 2: Delegators did not successfully shift blame for a bad outcome if
their human agent caused it instead of themselves. They did, however,
successfully shift blame for a bad outcome if their artificial agent caused it
instead of themselves.
Figure 3 illustrates this asymmetry in termsn of rewards between the human and
the machine treatment for the bad-outcome case.
Figure 3: Reward decisions if the bad outcome prevails.
To see whether participants exploited this incentive to delegate to machine
agents that does not exist for human agents, we compare the proportions of
delegators in both treatments. In the human treatment, 35 out of 76 (46.1%)
delegated the task to another human participant, whereas 41 out of of 73
(56.2%) delegated to the machine agent in the machine treatment. The
difference in the proportions of delegators is insignificant (p = 0.28?, Chi-
Square Test for Independence). The similar proportions of delegators in both
treatments suggests that the decision to delegate to the agent is not
primarily driven by strategic concerns of avoiding blame in the event of a bad
outcome.
This is corroborated by the logistic regression reported in Table 3. The
propensity to delegate (1 = yes) is positively predicted by participants’
self-assessment of the number of errors they believe they have committed in
the logic task (p < 0.001). “Machine” captures whether the agent is human (0)
or artificial (1), which does not significantly influence this result. The
effect to which self-assessment predicts the delegation decision is robust to
controlling for risk attitude, although participants who are more risk averse
are also more likely to delegate (p < 0.017).
Dep. Variable: | Del | No. Observations: | 149
---|---|---|---
Model: | GLM | Df Residuals: | 145
Model Family: | Binomial | Df Model: | 3
| coef | std err | z | P$>$$|$z$|$ | [0.025 | 0.975]
---|---|---|---|---|---|---
Intercept | -3.2254 | 0.795 | -4.055 | 0.000 | -4.784 | -1.666
Self_Assessment | 0.4092 | 0.105 | 3.906 | 0.000 | 0.204 | 0.615
Machine | 0.3228 | 0.365 | 0.885 | 0.376 | -0.392 | 1.038
Risk | 0.1806 | 0.076 | 2.383 | 0.017 | 0.032 | 0.329
Table 3: Generalized Linear Model Regression Results
It thus appears that the expected utility of delegating was the driving factor
for participants’ decisions to do so or not regardless of the nature of their
agent. Those who were less confident regarding their own performance were more
likely to delegate the task. Their estimated number of errors, which they
stated in an incentivized self-assessment, had a significant impact on their
delegation decision.
As Table 4 shows, the self-assessment of participants is also predicted by the
actual number of a subject’s errors in the logic task. Whether they have a
human or an artificial agent at their avail did not influence their self-
assessment. Also, subjects’ risk attitudes have no impact on their self
assessment. These findings indicate that participants have a realistic
impression about their own performance. Subjects who performed better in the
logic task were accordingly less likely to delegate.
Dep. Variable: | Self_Assessment | R-squared: | 0.269
---|---|---|---
Model: | OLS | Adj. R-squared: | 0.253
Method: | Least Squares | F-statistic: | 17.74
No. Observations: | 149 | |
| coef | std err | t | P$>$$|$t$|$ | [0.025 | 0.975]
---|---|---|---|---|---|---
Intercept | 3.2012 | 0.474 | 6.749 | 0.000 | 2.264 | 4.139
Machine | 0.4598 | 0.276 | 1.666 | 0.098 | -0.086 | 1.005
Error | 0.3955 | 0.059 | 6.741 | 0.000 | 0.280 | 0.512
Risk | -0.0298 | 0.055 | -0.539 | 0.591 | -0.139 | 0.079
Table 4: Linear Regression – Influences on Self-assessment
## 5 Conclusion
Our findings indicate that delegators may be judged with more leniency in the
event of a bad outcome if they delegate tasks to artificial agents instead of
human agents. It does therefore stand to reason that machine agents can be
successfully used to avoid ostracism. The FCO president’s statement quoted at
the beginning, i.e. that companies cannot hide behind algorithms, might not
hold true empirically. Companies might well capitalize the effective shift of
responsibility to algorithms if they fail. This is all the more true if they
do not suffer any comparable loss of prestige as a result of the delegation in
the event of success, as our data also suggest. The fact that the delegator in
a between-subjects design receives a discharge if the agent is artificial but
not if the agent is human at least suggests that the corresponding judgment is
not reflective, but that it is a subtle behavioral tendency. If this is true,
the discharge is not ethically desired by the evaluator.
Our results might indicate the importance of institutional solutions that hold
companies and ultimately individuals liable for the moral wrongs that their
artificial agents bring about. These institutions grow even more important if
they have to compensate for consumers’ behavioral reluctance to attribute
blame in such cases. A deeper inquiry into the reluctance to punish that we
observe seems warranted by the idea that extrinsic social motivation is an
important factor in moral decision-making (Cappelen et al., 2017). The ability
to use artificial agents as a smoke screen might encourage various decision
makers to engage in more activities that are considered undesirable by
stakeholders. The reassuring fact that delegators in this experiment did not
exploit the effective release from blame does not imply that others will not-
especially if they see through this behavioral tendency.
Our experiment is subject to limitations. One is its rather explorative
nature, which is owed to the lack of a well-established theoretical framework:
Scientists have only recently started to address the influence of algorithms
on human decision-making empirically. To the best of our knowledge, the
projects by Gogoll and Uhl (2018) and Strobel and Kirchkamp (2017) cited above
are the only experiments dedicated to the issue so far. Furthermore, the
stylized setting used in our experiment limits the scope of the conclusion
that we can draw. This concern is best explained in relation to internal and
external validity. Internal validity means that the experiment is designed in
such a way that it warrants conclusions about the behavior of its participants
inside of the laboratory, for instance, by keeping relevant factors constant
(ceteris paribus) or omitting irrelevant influences (ceteris absentibus).
Experiments are externally valid if their design produces findings that are
informative about behavior outside of the laboratory (Guala, 2002). Some
authors posit an inverse relationship between internal and external validity
(Guala et al., 2005; Loewenstein, 1999). Our findings provide a first
indication that the nature of the agent has an effect on the reward and
punishment that the delegator receives. The phenomenon we observed would have
to be replicated in other contexts inside and outside of the laboratory to
eliminate the possibility of it being a mere artifact.
We believe that our findings represent an early step in understanding
delegation decisions in a domain that is gaining relevance, i.e., the use of
artificial agents in moral decision-making. More generally, our experimental
results illustrate the necessity of investigating the interaction between
humans and machines in behavioral settings. It is insufficient to rely on
ethicists’ armchair arguments and on surveys that study laymen’s intuitions
regarding the ethical implications of algorithms, because ethically relevant
implications may arise as unintended results from the interactions between
people and machines. The emerging phenomena might then be difficult to
anticipate and sometimes even counter-intuitive. Further research is urgently
needed to create a more conclusive picture of the subtle factors that drive
our behavior when cooperating (and competing) with machines.
## References
* Allen and Wallach (2012) Allen, C. and W. Wallach (2012). Moral machines: contradiction in terms or abdication of human responsibility. Robot Ethics: The Ethical and Social Implications of Robotics, MIT Press, Cambridge (MA), 55–68.
* Bardsley et al. (2020) Bardsley, N., R. Cubitt, G. Loomes, P. Moffatt, C. Starmer, and R. Sugden (2020). Experimental economics: Rethinking the rules. Princeton University Press.
* Bénabou and Tirole (2002) Bénabou, R. and J. Tirole (2002). Self-confidence and personal motivation. The quarterly journal of economics 117(3), 871–915.
* Bigman et al. (2020) Bigman, Y., K. Gray, A. Waytz, M. Arnestad, and D. Wilson (2020). Algorithmic discrimination causes less moral outrage than human discrimination.
* Bolton and Ockenfels (2000) Bolton, G. E. and A. Ockenfels (2000). Erc: A theory of equity, reciprocity, and competition. American economic review 90(1), 166–193.
* Busse (2017) Busse, C. (2017). Bundeskartellamt rügt lufthansa. https://www.sueddeutsche.de/wirtschaft/nach-air-berlin-pleite-bundeskartellamt-ruegt-lufthansa-1.3806188.
* Cappelen et al. (2017) Cappelen, A. W., T. Halvorsen, E. Ø. Sørensen, and B. Tungodden (2017). Face-saving or fair-minded: What motivates moral behavior? Journal of the European Economic Association 15(3), 540–557.
* Danaher (2016) Danaher, J. (2016). Robots, law and the retribution gap. Ethics and Information Technology 18(4), 299–309.
* Dzindolet et al. (2002) Dzindolet, M. T., L. G. Pierce, H. P. Beck, and L. A. Dawe (2002). The perceived utility of human and automated aids in a visual detection task. Human Factors 44(1), 79–94.
* Elish and Hwang (2015) Elish, M. and T. Hwang (2015). Praise the machine! punish the human! the contradictory history of accountability in automated aviation. Punish the Human.
* Elish (2019) Elish, M. C. (2019). Moral crumple zones: Cautionary tales in human-robot interaction. Engaging Science, Technology, and Society 5, 40–60.
* Fehr and Schmidt (1999) Fehr, E. and K. M. Schmidt (1999). A theory of fairness, competition, and cooperation. The quarterly journal of economics 114(3), 817–868.
* Fiorina (1986) Fiorina, M. P. (1986). Legislator uncertainty, legislative control, and the delegation of legislative power. JL Econ. & Org. 2, 33.
* Fischbacher (2007) Fischbacher, U. (2007). z-tree: Zurich toolbox for ready-made economic experiments. Experimental economics 10(2), 171–178.
* Fischbacher et al. (2008) Fischbacher, U. et al. (2008). Shifting the blame: On delegation and responsibility. Technical report, Institute for Empirical Research in Economics-University of Zurich.
* Gogoll and Uhl (2018) Gogoll, J. and M. Uhl (2018). Rage against the machine: Automation in the moral domain. Journal of Behavioral and Experimental Economics 74, 97–103.
* Goldbach et al. (2019) Goldbach, C., D. Kayar, T. Pitz, and J. Sickmann (2019). Transferring decisions to an algorithm: A simple route choice experiment. Transportation research part F: traffic psychology and behaviour 65, 402–417.
* Greiner et al. (2004) Greiner, B. et al. (2004). The online recruitment system orsee 2.0-a guide for the organization of experiments in economics. University of Cologne, Working paper series in economics 10(23), 63–104.
* Guala (2002) Guala, F. (2002). On the scope of experiments in economics: comments on siakantaris. Cambridge Journal of economics 26(2), 261–267.
* Guala et al. (2005) Guala, F. et al. (2005). The methodology of experimental economics. Cambridge University Press.
* Hardy (2019) Hardy, T. (2019). Shifting risk and shirking responsibility? the challenge of upholding employment standards regulation within franchise networks.
* Heinkelmann-Wild and Zangl (2020) Heinkelmann-Wild, T. and B. Zangl (2020). Multilevel blame games: Blame-shifting in the european union. Governance 33(4), 953–969.
* Hevelke and Nida-Rümelin (2015) Hevelke, A. and J. Nida-Rümelin (2015). Responsibility for crashes of autonomous vehicles: an ethical analysis. Science and engineering ethics 21(3), 619–630.
* Hill (2015) Hill, A. (2015). Does delegation undermine accountability? experimental evidence on the relationship between blame shifting and control. Journal of Empirical Legal Studies 12(2), 311–339.
* Holt and Laury (2002) Holt, C. A. and S. K. Laury (2002). Risk aversion and incentive effects. American economic review 92(5), 1644–1655.
* Konow (2000) Konow, J. (2000). Fair shares: Accountability and cognitive dissonance in allocation decisions. American economic review 90(4), 1072–1091.
* Loewenstein (1999) Loewenstein, G. (1999). Experimental economics from the vantage-point of behavioural economics. The Economic Journal 109(453), F25–F34.
* Maestas et al. (2008) Maestas, C. D., L. R. Atkeson, T. Croom, and L. A. Bryant (2008). Shifting the blame: Federalism, media, and public assignment of blame following hurricane katrina. Publius: The Journal of Federalism 38(4), 609–632.
* Manyika et al. (2017) Manyika, J., M. Chui, M. Miremadi, J. Bughin, M. George, P. Willmott, and M. Dewhurst (2017). A future that works: Automation, employment and productivity. McKinsey.
* Moor (1979) Moor, J. H. (1979). Are there decisions computers should never make. Nature and System 1(4), 217–229.
* Nissenbaum (1996) Nissenbaum, H. (1996). Accountability in a computerized society. Science and engineering ethics 2(1), 25–42.
* Niszczota and Kaszás (2020) Niszczota, P. and D. Kaszás (2020). Robo-fund aversion: People prefer it when humans and not computers make investment decisions with moral undertones.
* O’Donoghue and Somerville (2018) O’Donoghue, T. and J. Somerville (2018). Modeling risk aversion in economics. Journal of Economic Perspectives 32(2), 91–114.
* Oexl and Grossman (2013) Oexl, R. and Z. J. Grossman (2013). Shifting the blame to a powerless intermediary. Experimental Economics 16(3), 306–312.
* Robinette et al. (2016) Robinette, P., W. Li, R. Allen, A. M. Howard, and A. R. Wagner (2016). Overtrust of robots in emergency evacuation scenarios. In 2016 11th ACM/IEEE International Conference on Human-Robot Interaction (HRI), pp. 101–108. IEEE.
* Rudin (2019) Rudin, C. (2019). Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead. Nature Machine Intelligence 1(5), 206–215.
* Selten (1967) Selten, R. (1967). Die strategiemethode zur erforschung des eingeschränkt rationalen verhaltens im rahmen eines oligopolexperiments, s. 136–168. Beiträge zur experimentellen Wirtschafts-Forschung. Tübingen: JCB Mohr.
* Strobel and Kirchkamp (2017) Strobel, C. and O. Kirchkamp (2017). Sharing responsibility with a machine.
* Taeihagh and Lim (2019) Taeihagh, A. and H. S. M. Lim (2019). Governing autonomous vehicles: emerging responses for safety, liability, privacy, cybersecurity, and industry risks. Transport Reviews 39(1), 103–128.
* Weaver (1986) Weaver, R. K. (1986). The politics of blame avoidance. Journal of public policy, 371–398.
|
# Fluid dynamics in the warp drive spacetime geometry
Osvaldo L. Santos-Pereira<EMAIL_ADDRESS>Physics Institute, Universidade
Federal do Rio de Janeiro, Brazil Everton M. C. Abreu<EMAIL_ADDRESS>Physics Department, Universidade Federal Rural do Rio de Janeiro, Seropédica,
Brazil Physics Department, Universidade Federal de Juiz de Fora, Brazil
Graduate Program in Applied Physics, Physics Institute, Universidade Federal
do Rio de Janeiro, Brazil Marcelo B. Ribeiro<EMAIL_ADDRESS>Physics
Institute, Universidade Federal do Rio de Janeiro, Brazil Graduate Program in
Applied Physics, Physics Institute, Universidade Federal do Rio de Janeiro,
Brazil Valongo Observatory, Universidade Federal do Rio de Janeiro, Brazil
(August 27, 2024)
###### Abstract
The Alcubierre warp drive metric is a spacetime geometry featuring a spacetime
distortion, called warp bubble, where a massive particle inside it acquires
global superluminal velocities, or warp speeds. This work presents solutions
of the Einstein equations for the Alcubierre metric having fluid matter as
gravity source. The energy-momentum tensor considered two fluid contents, the
perfect fluid and the parametrized perfect fluid (PPF), a tentative more
flexible model whose aim is to explore the possibilities of warp drive
solutions with positive matter density content. Santos-Pereira et al. (2020)
already showed that the Alcubierre metric having dust as source connects this
geometry to the Burgers equation, which describes shock waves moving through
an inviscid fluid, but led the solutions back to vacuum. The same happened for
two out of four solutions subcases for the perfect fluid. Other solutions for
the perfect fluid indicate the possibility of warp drive with positive matter
density, but at the cost of a complex solution for the warp drive regulating
function. Regarding the PPF, solutions were also obtained indicating that warp
speeds could be created with positive matter density. Weak, dominant, strong
and null energy conditions were calculated for all studied subcases, being
satisfied for the perfect fluid and creating constraints in the PPF quantities
such that positive matter density is also possible for creating a warp bubble.
Summing up all results, energy-momentum tensors describing more complex forms
of matter, or field, distributions generate solutions for the Einstein
equations with the warp drive metric where negative matter density might not
be a strict precondition for attaining warp speeds.
warp drive, cosmic fluid, Burgers equation, shock waves, Alcubierre geometry
###### pacs:
04.20.Gz; 04.90.+e; 47.40.-x
## 1 Introduction
It is well known that in general relativity particles can travel globally at
superluminal speeds, whereas locally they cannot surpass the light speed. The
warp drive spacetime geometry advanced by Alcubierre Alcubierre1994 uses
these physical properties to propel material particles at superluminal speeds.
It creates a limited spacetime distortion, called warp bubble, such that the
spacetime is contracted in front of it and expanded behind the bubble as it
moves along a geodesic. This warp drive metric is such that a particle trapped
inside this bubble would locally move at subluminal speeds, whereas the bubble
with the particle inside acquires global superluminal velocities, or warp
speeds. In the seminal paper, Alcubierre also concluded that the warp metric
would imply the violation of the energy conditions since it appeared that
negative energy density would be required for the creation of the warp bubble.
Since this original work many authors have contributed in our understanding of
the theoretical details of the Alcubierre warp drive metric and possible
feasibility of matter particles acquiring warp speeds. Ford and Roman
FordRoman1996 applied quantum inequalities to calculate the amount of
negative energy required to transport particles at superluminal speeds. They
concluded that such energy requirements would be huge, so the amount of
negative energy density necessary for the practical construction of a warp
bubble would be impossible to achieve. Pfenning and Ford Pfenning1997 also
used quantum inequalities to calculate the necessary bubble parameters and
energy for the warp drive viability, reaching at an enormous amount of energy,
ten orders of magnitude greater than the mass-energy of the entire visible
Universe, also negative. Hiscock hiscock computed the vacuum energy-momentum
tensor (EMT) of a reduced two dimensional quantized scalar field of the warp
drive spacetime. He showed that in this reduced context that the EMT diverges
if the apparent velocity of the bubble is greater than the speed of light.
Such divergence is connected to the construction of an horizon in this two
dimensional spacetime. Due to the semiclassical effects, the superluminal
travel via warp drive might unfeasible. For example, to an observer within the
warp drive bubble, the backward and forward walls look like the horizon of a
white hole and of a black hole, respectively, resulting with a Hawking
radiation.
The issue of superluminal speeds of massive particles traveling faster than
photons has also been studied by Krasnikov Krasnikov1998 , who argued that
this would not be possible if some conjectures for globally hyperbolic
spacetimes are made. He described some spacetime topologies and their
respective need of the tachyon existence for the occurrence of travel at warp
speeds. This author also advanced a peculiar spacetime where superluminal
travel would be possible without tachyons, named as the Krasnikov tube by
Everett and Roman EveretRoman1997 , who generalized the metric designed by
Krasnikov by proposing a tube in the direction of the particle’s path
providing a connection between Earth and a distant star. Inside the tube the
spacetime is flat and the lightcones are opened out in order to allow the one
direction superluminal travel. For the Krasnikov tube to work they showed that
huge quantities of negative energy density would also be necessary. Since the
tube does not possess closed timelike curves, it would be theoretically
possible to design a two way non-overlapping tube system such that it would
work as a time machine. In addition, the EMT is positive in some regions. Both
the metric and the obtained EMT were thoroughly analyzed in Refs. Lobo2002 ;
Lobo2003 .
A relevant contribution to warp drive theory was made by van de Broeck
Broeck1999 , who demonstrated that a small modification of the original
Alcubierre geometry greatly diminishes to a few solar masses the total
negative energy necessary for the creation of the warp bubble distortion, a
result that led him to hypothesize that other geometrical modifications of
this type could further reduce in a dramatic fashion the amount of energy
necessary to create a warp drive bubble. Natario Natario2002 designed a new
warp drive concept with zero expansion by using spherical coordinates and the
$X$-axis as the polar one. Lobo and Visser LoboVisser2004b ; LoboVisser2004
discussed that the center of the warp bubble, as proposed by Alcubierre, need
to be massless (see also White2003 ; White2011 ). A linearized model for both
approaches was introduced and it was demonstrated that for small speeds the
amassed negative energy inside the warp field is a robust fraction of the
particle’s mass inside the center of the warp bubble. Lee and Cleaver cleaver1
; cleaver2 have looked at how external radiation might affect the Alcubierre
warp bubble, possibly making it energetically unsustainable, and how a
proposed warp field interferometer could not detect spacetime distortions.
Mattingly et al. cleaver3 ; cleaver4 discussed curvature invariants in the
Natario and Alcubierre warp drives.
In a previous paper nos we have considered some of these issues, but from a
different angle. Since the Alcubierre metric was not advanced as a solution of
the Einstein equations, as it was originally proposed simply as an ad hoc
geometry designed to create a spacetime distortion such that a massive
particle inside it travels at warp speeds, whereas locally it never exceeds
the light speed, the basic question was then the possible types of matter-
energy sources capable of creating a warp bubble. To answer this question the
Einstein equations have to be solved with some form of EMT as source. The
simplest one to start with is incoherent matter. Following this line of
investigation, we showed that the dust solutions of the Einstein equations for
the warp drive metric implied in vacuum, that is, such distribution is
incapable of creating a warp bubble, nevertheless, the Burgers equation
appeared as part of the solution of the Einstein equations. In addition, since
the Burgers equation describes shock waves moving in an inviscid fluid, it was
also found that at these shock waves behave as plane waves Trefethen2001 ;
Evans2010 ; Forsyth1906 ; Bateman1915 ; Burgers1948 .
In this paper we generalize the results obtained in Ref. nos by following the
next logical step, that is, investigating perfect fluid as EMT source for the
Alcubierre metric. We also propose a slightly generalized perfect fluid EMT,
called here parametrized perfect fluid (PPF), in order to produce a
tentatively more flexible model such that the pressure may have different
parameter values. The aim is to see if more flexible EMT distributions could
relax the original requirement that warp speeds could only be achieved by
means of negative matter density.
For the perfect fluid EMT solutions we found that two out of four subcases
turn out to be the dust solution of Ref. nos where both the matter density
and pressure vanish, but the Burgers equation also appears as a result of the
solutions of the Einstein equations turcos . Two other subcases, however,
indicate that warp speeds are possible with positive matter density, but at
the cost of a complex solution for the warp metric regulating function. Weak,
dominant, strong and null energy conditions were calculate for both EMTs and
all perfect fluid solutions satisfy them. In the case of the PPF, two out of
four solutions give rise to a nonlinear equation of state linking various
pressures to the matter density. Other solutions produced results where a
nonvanishing pressure occurs with a vanishing matter density, condition
considered unphysical and then dismissed. The solutions also produced
parameters and equations of state related to pressure and inequalities that
satisfy all the energy conditions. These results indicate that energy-momentum
tensors describing more complex forms of matter distributions generate
solutions for the Einstein equations with the warp drive metric where negative
matter density might not be a strict precondition.
The plan for the paper is as follows. In Section 2 we briefly review the
Alcubierre warp drive theory and present the relevant equations and all
nonzero components of the Einstein tensor for the warp drive metric. In
Section 3 the Einstein equations are then solved and solutions presented for
the warp drive metric having a perfect fluid gravity source. In Section 4 the
non-zero components of the Einstein tensor in the warp drive geometry are
written in terms of the PPF EMT. Solutions for this more flexible EMT are also
obtained and studied in all subcases. Section 5 presents the EMT divergence of
both the perfect fluid and PPF, whereas Section 6 discusses the energy
conditions for the two types of EMTs. Section 7 provides further discussions
on the results presented in the previous sections, and Section 8 depicts our
conclusions.
## 2 Einstein Equations
We shall start this section by brief reviewing the Alcubierre warp drive
metric. Subsequently, the nonzero components of the Einstein tensor of this
metric will also be explicitly shown. The expressions presented in this
section form the basic set of equations required for the next sections.
### 2.1 The Alcubierre warp drive geometry
The start up geometry advanced in Ref. Alcubierre1994 may be written as
follows,
${ds}^{2}=-\left(\alpha^{2}-\beta_{i}\beta^{i}\right)\,dt^{2}+2\beta_{i}\,dx^{i}\,dt+\gamma_{ij}\,dx^{i}\,dx^{j}\,\,,$
(2.1)
where $d\tau$ is the proper time lapse, $\alpha$ is the lapse function,
$\beta^{i}$ is the spacelike shift vector and $\gamma_{ij}$ is the spatial
metric for the hypersurfaces.111Throughout this paper Greek indices will range
from 0 to 3, whereas the Latin ones indicate the spacelike hypersurfaces and
will range from 1 to 3. The lapse function $\alpha$ and the shift vector
$\beta^{i}$ are functions to be determined, whereas $\gamma_{ij}$ is a
positive-definite metric on each of the spacelike hypersurfaces, for all
values of time, a feature that makes the spacetime globally hyperbolic
Alcubierre2012 ; DeWitt1979 .
Alcubierre Alcubierre1994 assumed the following particular parameter choices
for Eq. (2.1),
$\displaystyle\alpha$ $\displaystyle=1,$ (2.2) $\displaystyle\beta^{1}$
$\displaystyle=-v_{s}(t)f\big{[}r_{s}(t)\big{]},$ (2.3)
$\displaystyle\beta^{2}$ $\displaystyle=\beta^{3}=0,$ (2.4)
$\displaystyle\gamma_{ij}$ $\displaystyle=\delta_{ij}.$ (2.5)
Hence, the Alcubierre warp drive metric is given by,
$ds^{2}=-\left[1-v_{s}(t)^{2}f(r_{s})^{2}\right]dt^{2}-v_{s}(t)f(r_{s})\,dx\,dt+dx^{2}+dy^{2}+dz^{2}\,\,,$
(2.6)
where $v_{s}(t)$ is the velocity of the center of the bubble moving along the
curve $x_{s}(t)$. This is given by the following expression,
$v_{s}(t)=\frac{dx_{s}(t)}{dt}\,\,.$ (2.7)
The function $f(r_{s})$ is the warp drive regulating function. It describes
the shape of the warp bubble, which is given by the following expression
Alcubierre1994 ,
$f(r_{s})=\frac{\tanh\left[\sigma(r_{s}+R)\right]-\tanh\left[\sigma(r_{s}-R)\right]}{2\tanh(\sigma
R)}\,\,,$ (2.8)
where $\sigma$ and $R$ are parameters to be determined. The variable
$r_{s}(t)$ defines the distance from the center of the bubble $[x_{s}(t),0,0]$
to a generic point $(x,y,z)$ on the surface of the bubble, given by the
following equation,
$r_{s}(t)=\sqrt{\left[x-x_{s}(t)\right]^{2}+y^{2}+z^{2}}.$ (2.9)
From the above one can see that the motion is one-dimensional, since the
$x$-coordinate is the only one perturbed by the function $x_{s}(t)$.
### 2.2 Einstein tensor components
Let us now adopt Alcubierre’s original notation by assuming
$\beta=-\beta^{1}=v_{s}(t)f(r_{s})$ (2.10)
in Eq. (2.3). Then, the nonzero components of the Einstein tensor for the
metric (2.6) are given by the following expressions:
$\displaystyle G_{00}$ $\displaystyle=$
$\displaystyle-\frac{1}{4}(1+3\beta^{2})\left[\left(\frac{\partial\beta}{\partial
y}\right)^{2}+\left(\frac{\partial\beta}{\partial
z}\right)^{2}\right]-\beta\left(\frac{\partial^{2}\beta}{\partial
y^{2}}+\frac{\partial^{2}\beta}{\partial z^{2}}\right),$ (2.11) $\displaystyle
G_{01}$ $\displaystyle=$
$\displaystyle\frac{3}{4}\beta\left[\left(\frac{\partial\beta}{\partial
y}\right)^{2}+\left(\frac{\partial\beta}{\partial
z}\right)^{2}\right]+\frac{1}{2}\left(\frac{\partial^{2}\beta}{\partial
y^{2}}+\frac{\partial^{2}\beta}{\partial z^{2}}\right),$ (2.12) $\displaystyle
G_{02}$ $\displaystyle=$
$\displaystyle-\frac{1}{2}\frac{\partial^{2}\beta}{\partial x\partial
y}-\frac{\beta}{2}\left(2\frac{\partial\beta}{\partial
y}\,\frac{\partial\beta}{\partial x}+\beta\frac{\partial^{2}\beta}{\partial
x\partial y}+\frac{\partial^{2}\beta}{\partial t\partial y}\right),$ (2.13)
$\displaystyle G_{03}$ $\displaystyle=$
$\displaystyle-\frac{1}{2}\frac{\partial^{2}\beta}{\partial x\partial
z}-\frac{\beta}{2}\left(2\frac{\partial\beta}{\partial
z}\,\frac{\partial\beta}{\partial x}+\beta\frac{\partial^{2}\beta}{\partial
x\partial z}+\frac{\partial^{2}\beta}{\partial t\partial z}\right),$ (2.14)
$\displaystyle G_{11}$ $\displaystyle=$
$\displaystyle-\frac{3}{4}\left[\left(\frac{\partial\beta}{\partial
y}\right)^{2}+\left(\frac{\partial\beta}{\partial z}\right)^{2}\right],$
(2.15) $\displaystyle G_{12}$ $\displaystyle=$
$\displaystyle\frac{1}{2}\left(2\frac{\partial\beta}{\partial
y}\,\frac{\partial\beta}{\partial x}+\beta\frac{\partial^{2}\beta}{\partial
x\partial y}+\frac{\partial^{2}\beta}{\partial t\partial y}\right),$ (2.16)
$\displaystyle G_{13}$ $\displaystyle=$
$\displaystyle\frac{1}{2}\left(2\frac{\partial\beta}{\partial
z}\,\frac{\partial\beta}{\partial x}+\beta\frac{\partial^{2}\beta}{\partial
x\partial z}+\frac{\partial^{2}\beta}{\partial t\partial z}\right),$ (2.17)
$\displaystyle G_{23}$ $\displaystyle=$
$\displaystyle\frac{1}{2}\frac{\partial\beta}{\partial
z}\,\frac{\partial\beta}{\partial y},$ (2.18) $\displaystyle G_{22}$
$\displaystyle=$ $\displaystyle-\left[\frac{\partial^{2}\beta}{\partial
t\partial x}+\beta\frac{\partial^{2}\beta}{\partial
x^{2}}+\left(\frac{\partial\beta}{\partial
x}\right)^{2}\right]-\frac{1}{4}\left[\left(\frac{\partial\beta}{\partial
y}\right)^{2}-\left(\frac{\partial\beta}{\partial z}\right)^{2}\right],$
(2.19) $\displaystyle G_{33}$ $\displaystyle=$
$\displaystyle-\left[\frac{\partial^{2}\beta}{\partial t\partial
x}+\beta\frac{\partial^{2}\beta}{\partial
x^{2}}+\left(\frac{\partial\beta}{\partial
x}\right)^{2}\right]+\frac{1}{4}\left[\left(\frac{\partial\beta}{\partial
y}\right)^{2}-\left(\frac{\partial\beta}{\partial z}\right)^{2}\right]\,\,.$
(2.20)
## 3 Perfect fluid
Besides incoherent matter, or dust, already studied in Ref. nos , the simplest
matter-energy distribution to be considered as gravity source for the possible
creation of a warp bubble, and then warp speeds, is the perfect fluid. Hence,
this section will discuss matter content solutions of the Einstein equations
considering a perfect fluid matter source EMT for the Alcubierre metric.
### 3.1 Perfect fluid content solutions
The EMT for a perfect fluid may be written as follows,
$T_{\alpha\beta}=\left(\mu+p\right)\,u_{\alpha}u_{\beta}+p\,g_{\alpha\beta},$
(3.1)
where $\mu$ is the matter density, $p$ is the fluid pressure,
$g_{\alpha\beta}$ is the metric tensor and $u_{\alpha}$ is the 4-velocity of
an observer inside the fluid. Perfect fluids have no shear stress, rotation,
heat conduction or viscosity, nevertheless this ideal fluid provides a more
complex matter content than simple dust Schutz2009 , allowing us to study if a
warp bubble can be created with this gravity source and how the respective
gravity field equations solutions can be understood.
For the metric (2.6) the perfect fluid EMT assumes the following form,
$T_{\alpha\beta}=\begin{pmatrix}\mu+\beta^{2}p&-\beta p&0&0\\\ -\beta
p&p&0&0\\\ 0&0&p&0\\\ 0&0&0&p\end{pmatrix}\,.$ (3.2)
Let us now use Eqs. (2.11) to (2.20) with the EMT above in the Einstein
equations. Substituting components $G_{11}=8\pi T_{11}$ and $G_{01}=8\pi
T_{01}$ into $G_{00}=8\pi T_{00}$, after some algebra and simplifications we
may write the following expression,
$T_{00}+2\beta T_{01}+\frac{1}{3}(3\beta^{2}-1)T_{11}=0\,\,.$ (3.3)
Substituting the values for the EMT components $T_{00}=\mu+\beta^{2}p$,
$T_{01}=-\beta p$, $T_{11}=p$ Eq. (3.3) results in the following expression,
$p=3\mu.$ (3.4)
This is an equation of state for the Alcubierre metric having a perfect fluid
gravity source EMT.
The component $G_{23}$ is zero since $T_{23}=0$. This case leads us to either
$\partial\beta/\partial y$, or $\partial\beta/\partial z$, or both, equal to
zero. Let us now analyze these possibilities and its consequences.
Case 1: $\bm{\left[\displaystyle\frac{\partial\beta}{\partial z}=0\right]}$
As $\beta$ does not depend on $z$, the Einstein tensor components $G_{13}$,
$G_{23}$ and $G_{03}$ are identically zero. Substituting this case into
$G_{11}=8\pi T_{11}$, where $G_{11}$ is given by Eq. (2.15) and $T_{11}=p$, it
follows immediately the result below,
$\frac{3}{4}\left(\frac{\partial\beta}{\partial y}\right)^{2}=-8\pi p.$ (3.5)
Substituting $(\partial\beta/\partial y)^{2}$ above into $G_{01}=8\pi T_{01}$,
where $G_{01}$ is given by Eq. (2.12) and $T_{01}=-\beta p$ from Eq. (3.2), as
well as $G_{12}=8\pi T_{12}$ and $G_{02}=8\pi T_{02}$, where $T_{12}=0$ and
$T_{02}=0$, the Einstein equations are reduced to the following equations,
$\displaystyle p=3\mu,$ (3.6)
$\displaystyle\left(\frac{\partial\beta}{\partial
y}\right)^{2}=-\frac{32}{3}\pi p=-\,32\pi\,\mu,$ (3.7)
$\displaystyle\frac{\partial^{2}\beta}{\partial y^{2}}=0,$ (3.8)
$\displaystyle\frac{\partial^{2}\beta}{\partial x\partial y}=0,$ (3.9)
$\displaystyle 2\frac{\partial\beta}{\partial
y}\,\frac{\partial\beta}{\partial x}+\frac{\partial^{2}\beta}{\partial
t\partial y}=0,$ (3.10) $\displaystyle\frac{\partial^{2}\beta}{\partial
t\partial x}+\beta\frac{\partial^{2}\beta}{\partial
x^{2}}+\left(\frac{\partial\beta}{\partial x}\right)^{2}=-\frac{64}{3}\pi
p\,=\,-\,64\pi\mu,$ (3.11) $\displaystyle\frac{\partial^{2}\beta}{\partial
t\partial x}+\beta\frac{\partial^{2}\beta}{\partial
x^{2}}+\left(\frac{\partial\beta}{\partial x}\right)^{2}=-\frac{128}{3}\pi
p\,=\,-\,128\pi\mu\,\,.$ (3.12)
Eq. (3.7) implies that $\partial\beta/\partial y$ must be constant, since the
pressure $p$ is assumed constant. Eq. (3.8) also shows that $\beta$ must be a
linear function of the $y$-coordinate, which means that $\beta$ must have a
possible additional dependence of arbitrary functions on $t$ and $x$. Both
expressions in Eqs. (3.11) and (3.12) constitute the same homogeneous partial
differential equation, but with different inhomogeneous parts, so the solution
of the inhomogeneous equation is not unique, unless the pressure $p$ is zero.
Then, considering these points and Eqs. (3.7) and (3.10), it follows that,
$\frac{\partial\beta}{\partial y}\,\frac{\partial\beta}{\partial x}=0,$ (3.13)
which means that either of these partial derivatives, or both, vanish. Let us
discuss both possibilities below.
Case 1a: $\bm{\left[\displaystyle\frac{\partial\beta}{\partial
z}=0\,\,\,\,\text{and}\,\,\,\,\frac{\partial\beta}{\partial x}=0\right]}$
For this case the set of partial differential equations from Eqs. (3.6) to
(3.12) simplify to,
$\displaystyle p=3\mu,$ (3.14) $\displaystyle\frac{\partial\beta}{\partial
y}=\pm\sqrt{-32\pi\mu}.$ (3.15)
The above equations mean that the matter density $\mu$ must be negative or
zero for a non complex solution of the Einstein equations, and $\beta$ must be
a function of both $t$ and $x$ coordinates only. The equation above is readily
integrated, yielding
$\beta(t,y)=\pm\sqrt{-32\pi\mu}\,y+g(t)\,\,,$ (3.16)
where $g(t)$ is a function to be determined by the boundary conditions.
Case 1b: $\bm{\left[\displaystyle\frac{\partial\beta}{\partial
z}=0\,\,\,\,\text{and}\,\,\,\,\frac{\partial\beta}{\partial y}=0\right]}$
In this case the pressures vanishes, since $\partial\beta/\partial y=0$, and
the set of partial differential equations (3.6) to (3.12) simplify to,
$\displaystyle p=3\mu=0,$ (3.17)
$\displaystyle\frac{\partial^{2}\beta}{\partial t\partial
x}+\beta\frac{\partial^{2}\beta}{\partial
x^{2}}+\left(\frac{\partial\beta}{\partial x}\right)^{2}=0.$ (3.18)
Therefore, for $p=0$ the equation of state $p=3\mu$ implies zero matter
density as well, which reduces the solution to the dust case and then vacuum.
This also leads to the appearance of shock waves as plane waves, since
$\beta=\beta(x,t)$ and the field equations are reduced to the Burgers
equation, as studied in Ref. nos . This is the case because Eq. (3.18) can be
written in the following form,
$\frac{\partial\beta}{\partial t}+\frac{1}{2}\frac{\partial}{\partial
x}(\beta^{2})=h(t).$ (3.19)
Here $h(t)$ is a generic function to be determined by boundary conditions. In
its homogeneous form, where $h(t)=0$, it takes the conservation form of the
inviscid Burgers equation.
$\frac{\partial\beta}{\partial t}+\frac{1}{2}\frac{\partial}{\partial
x}(\beta^{2})=0\,\,.$ (3.20)
See Ref. nos for details of the Burgers equation in this context.
Case 2: $\bm{\left[\displaystyle\frac{\partial\beta}{\partial y}=0\right]}$
As $\beta$ does not depend on the $y$-coordinate, it is easy to see that
$G_{12}$, $G_{23}$ and $G_{02}$ are identically zero. In addition, considering
this case into $G_{11}=8\pi T_{11}$, it follows immediately that,
$-\frac{3}{4}\left(\frac{\partial\beta}{\partial z}\right)^{2}=8\pi p.$ (3.21)
Substituting Eq. (3.21) in $G_{01}=8\pi T_{01}$, where $G_{01}$ is given by
Eq. (2.12), $T_{01}=-\beta p$ from Eq. (3.2), and inserting the component
$G_{13}=8\pi T_{13}$ into the component $G_{03}=8\pi T_{03}$, where
$T_{13}=T_{03}=0$, after some algebra we reach at the following expressions,
$\displaystyle p=3\mu,$ (3.22)
$\displaystyle\left(\frac{\partial\beta}{\partial
z}\right)^{2}=-\frac{32}{3}\pi p,$ (3.23)
$\displaystyle\frac{\partial^{2}\beta}{\partial z^{2}}=0,$ (3.24)
$\displaystyle\frac{\partial^{2}\beta}{\partial x\partial z}=0,$ (3.25)
$\displaystyle 2\frac{\partial\beta}{\partial
z}\,\frac{\partial\beta}{\partial x}+\frac{\partial^{2}\beta}{\partial
t\partial z}=0,$ (3.26) $\displaystyle\frac{\partial^{2}\beta}{\partial
t\partial x}+\beta\frac{\partial^{2}\beta}{\partial
x^{2}}+\left(\frac{\partial\beta}{\partial x}\right)^{2}=-\frac{64}{3}\pi p,$
(3.27) $\displaystyle\frac{\partial^{2}\beta}{\partial t\partial
x}+\beta\frac{\partial^{2}\beta}{\partial
x^{2}}+\left(\frac{\partial\beta}{\partial x}\right)^{2}=-\frac{128}{3}\pi p.$
(3.28)
Eq. (3.24) shows that $\beta$ is a linear function with respect to the
$z$-coordinate. Eq. (3.23) implies that $\partial\beta/\partial z$ must be
constant since the pressure $p$ is assumed to be a constant. This means that
all second partial derivatives of $\partial\beta/\partial z$ must vanish. Eqs.
(3.27) and (3.28) are the same homogeneous partial differential equation, but
both right hand side of theirs are different. Hence, the solution of the non-
homogeneous equation is not unique, unless the pressure $p$ is zero.
Considering Eq. (3.26), this case also unfolds in two possibilities, since,
$\frac{\partial\beta}{\partial z}\,\frac{\partial\beta}{\partial x}=0,$ (3.29)
and either or both are zero. Let us now analyze each subcase.
Case 2a: $\bm{\displaystyle\left[\frac{\partial\beta}{\partial y}=0\
\text{and}\ \frac{\partial\beta}{\partial x}=0\right]}$
For this case Eqs. (3.22) to (3.28) yield,
$\displaystyle p=3\mu\,,$ (3.30) $\displaystyle\frac{\partial\beta}{\partial
z}=\pm\,\sqrt{-32\pi\mu}\,,$ (3.31)
$\displaystyle\frac{\partial\beta}{\partial z}=\pm\,\sqrt{\pm 96\pi\mu}\,.$
(3.32)
Eq. (3.31) means that the matter density $\mu$ must be negative or zero for a
non complex solution of the Einstein equations. Eq. (3.32) allows for possible
positive matter density. In addition, $\beta$ has its dependence reduced to
$\beta=\beta(z,t)$. If the matter density $\mu$ is assumed constant, then the
above expressions can be integrated, yielding
$\displaystyle\beta(z,t)=\pm\sqrt{-32\pi\mu}\,z+\bar{g}(t)\,,$ (3.33)
$\displaystyle\beta(z,t)=\pm\sqrt{\pm 96\pi\mu}\,z+\bar{h}(t)\,.$ (3.34)
where $\bar{g}(t)$ and $\bar{h}(t)$ are arbitrary functions to be determined
by boundary conditions.
Case 2b $\bm{\displaystyle\left[\frac{\partial\beta}{\partial y}=0\
\text{and}\ \frac{\partial\beta}{\partial z}=0\right]}$
For this subcase Eq. (3.23) implies zero pressure, and the set of partial
differential equations (3.22) to (3.28) simplify to,
$\displaystyle p=3\mu=0,$ (3.35)
$\displaystyle\frac{\partial^{2}\beta}{\partial t\partial
x}+\beta\frac{\partial^{2}\beta}{\partial
x^{2}}+\left(\frac{\partial\beta}{\partial x}\right)^{2}=0\,\,,$ (3.36)
where the last equation is the result of the only nonzero Einstein tensor
components $G_{22}$ and $G_{33}$. These results are the same as Case 1b above,
that is, the dust solution for the Alcubierre warp drive metric that results
in the Burgers equation (3.19) and its inviscid form given by Eq. (3.20), as
well as shock waves as plane waves nos .
$\square$
Table 1 summarizes all cases and their respective results of the Einstein
equations with the Alcubierre warp drive metric having a perfect fluid matter
content as gravity source.
Case | Condition | Results
---|---|---
$1)\ \displaystyle{\frac{\partial\beta}{\partial z}=0}$ | $1a)\ \displaystyle{\frac{\partial\beta}{\partial x}=0}$ | $\begin{array}[]{ll}p=3\mu\\\\[6.0pt] \beta=\beta(y,t)\\\\[6.0pt] \displaystyle{\frac{\partial\beta}{\partial y}=\pm\sqrt{-32\pi\mu}}\\\\[8.0pt] \end{array}$
$1b)\ \displaystyle{\frac{\partial\beta}{\partial y}=0}$ | $\begin{array}[]{ll}p=3\mu=0\\\\[6.0pt] \beta=\beta(x,t)\\\\[6.0pt] \displaystyle{\frac{\partial\beta}{\partial t}+\frac{1}{2}\frac{\partial}{\partial x}(\beta^{2})=h(t)}\\\\[8.0pt] \rightarrow\mbox{this is the dust solution of Ref.\ \cite[cite]{\@@bibref{Authors Phrase1YearPhrase2}{nos}{\@@citephrase{(}}{\@@citephrase{)}}}}\\\\[4.0pt] \end{array}$
$2)\ \displaystyle{\frac{\partial\beta}{\partial y}=0}$ | $2a)\ \displaystyle{\frac{\partial\beta}{\partial x}=0}$ | $\begin{array}[]{ll}p=3\mu\\\\[6.0pt] \beta=\beta(z,t)\\\\[6.0pt] \displaystyle{\frac{\partial\beta}{\partial z}=\pm\,\sqrt{-32\pi\mu}}\\\\[8.0pt] \displaystyle{\frac{\partial\beta}{\partial z}=\,\pm\sqrt{\pm\,96\pi\mu}}\\\\[8.0pt] \end{array}$
$2b)\ \displaystyle{\frac{\partial\beta}{\partial z}=0}$ | $\begin{array}[]{ll}p=3\mu=0\\\\[6.0pt] \beta=\beta(x,t)\\\\[6.0pt] \displaystyle{\frac{\partial\beta}{\partial t}+\frac{1}{2}\frac{\partial}{\partial x}(\beta^{2})=h(t)}\\\\[6.0pt] \rightarrow\mbox{this is the dust solution of Ref.\ \cite[cite]{\@@bibref{Authors Phrase1YearPhrase2}{nos}{\@@citephrase{(}}{\@@citephrase{)}}}}\\\\[4.0pt] \end{array}$
Table 1: Summary of all solutions of the Einstein equations with the
Alcubierre warp drive metric having perfect fluid EMT as mass-energy source.
### 3.2 Discussion
Cases 1b and 2b are simply the dust case already studied in Ref. nos ,
apparently being unable to generate a warp bubble since this is a vacuum
solution, although it connects the warp metric to the Burgers equation and
then to shock waves in the form of plane waves.
Cases 1a and 2a share the same equation of state $p=3\mu$, but the coordinate
dependencies are different, since $\beta=\beta(y,t)$ and $\beta=\beta(z,t)$,
respectively. For $\beta$ to be a real valued function the matter density
$\mu$ must be negative in Case 1a. From Eqs. (3.16) and (3.33) we have assumed
a constant matter density, which means a straightforward integration, but
$\mu$ can be also a function of both $t$ and $y$ coordinates in the Case 1a,
or a function of both $t$ and $z$ coordinates in the Case 2a.
Inasmuch as the matter density must be negative for real solutions, one could
think of defining the total mass-energy density as follows,
$\displaystyle\mu(t,x^{j})=\mu^{+}+a(t,x^{j})\mu^{-}\leq 0,$ (3.37)
$\displaystyle x^{j}=y,z,$
where $\mu^{+}$ is the positive portion of the matter density of the perfect
fluid and $\mu^{-}$ its negative portion that would allow the warp bubble to
exist. $a(t,x^{j})$ would be a regulating function that depends on both time
$t$ and space $x^{j}$ coordinates, being related to the shape and location of
the bubble. Remembering that $x^{j}=y$ for the Case 1a and $x^{j}=z$ for the
Case 2a, since $\mu(t,x^{j})$ must be negative there would be a restriction
for the positive and negative portions of the matter density in Eq. (3.37).
It might be argued that there is no problem in complex solutions for $\beta$,
but in the warp drive scenario $\beta=v_{s}(t)f(r_{s})$ determines both the
velocity and the shape of the bubble, so either the velocity $v_{s}(t)$ or the
regulating function $f(r_{s})$ of the bubble shape must be complex. A complex
velocity has no physical meaning, but a complex regulating function could be
acceptable if we only consider its real part. Thus, in such situation the
formation of a warp bubble capable of generating warp speeds could still be
possible in the presence of a perfect fluid positive matter density EMT as
gravity source.
Nevertheless, caution is required here because if the result of integrating
$\beta$ turns out to be purely imaginary it is not clear what the bubble shape
being represented by an imaginary function means. Therefore, in principle it
seems reasonable to start with $\beta$ being a real function and the warp
bubble requiring a perfect fluid with negative mass-energy density for warp
speeds to be physically viable. But, it seems to us that this point remains
open to debate.
Considering the results above it is apparent that a perfect fluid EMT
generated a more complex set of solutions of the Einstein equations than the
dust one, then it is conceivable that viable warp speeds are also possible in
more complex EMTs. One such possibility will be discussed next.
## 4 Parametrized perfect fluid
Let us propose a generalization of the perfect fluid EMT having seven
quantities, namely the mass-energy density $\mu$, the $\beta$ function and
five different pressures $A,B,C,D$ and $p$. The last quantity $D$ is a
momentum density parameter. In the perfect fluid of Eq. (3.2) the pressure
denoted by $p$ are all the same in the EMT, a constraint that has been relaxed
here. Let us call the perfect fluid generalization with the quantities above
as the parametrized perfect fluid (PPF). Its respective EMT may be written as
below,
$T_{\alpha\sigma}=\begin{pmatrix}\mu+\beta^{2}p&-\beta D&0&0\\\ -\beta
D&A&0&0\\\ 0&0&B&0\\\ 0&0&0&C\end{pmatrix}\,.$ (4.1)
The quantities $A,B,C,D$ and $p$ will not be assumed as constants, but rather
as functions of the spacetime coordinates $(t,x,y,z)$.
This is clearly a more flexible EMT than the perfect fluid, and it is being
proposed here as a tentative model in order to explore the consequences of
more complex EMTs in terms of generating possible positive matter solutions of
the Einstein equations with the warp drive metric without the caveats of the
perfect fluid solutions discussed above. It is a tentative proposal for a more
flexible, or toy, model for the possible creation of warp bubbles, and then
warp speeds. Section 7.2 provides more details on the physics of this specific
fluid proposal.
The nonzero components of the Einstein equations for the PPF are given by the
following expressions,
$\displaystyle-\frac{1}{4}(1+3\beta^{2})\left[\left(\frac{\partial\beta}{\partial
y}\right)^{2}+\left(\frac{\partial\beta}{\partial
z}\right)^{2}\right]-\beta\left(\frac{\partial^{2}\beta}{\partial
y^{2}}+\frac{\partial^{2}\beta}{\partial z^{2}}\right)$ $\displaystyle=$
$\displaystyle 8\pi(\mu+\beta^{2}p),$ (4.2)
$\displaystyle\frac{3}{4}\beta\left[\left(\frac{\partial\beta}{\partial
y}\right)^{2}+\left(\frac{\partial\beta}{\partial
z}\right)^{2}\right]+\frac{1}{2}\left(\frac{\partial^{2}\beta}{\partial
y^{2}}+\frac{\partial^{2}\beta}{\partial z^{2}}\right)$ $\displaystyle=$
$\displaystyle-8\pi\beta D,$ (4.3)
$\displaystyle-\frac{1}{2}\frac{\partial^{2}\beta}{\partial x\partial
y}-\frac{\beta}{2}\left(2\frac{\partial\beta}{\partial
y}\,\frac{\partial\beta}{\partial x}+\beta\frac{\partial^{2}\beta}{\partial
x\partial y}+\frac{\partial^{2}\beta}{\partial t\partial y}\right)$
$\displaystyle=$ $\displaystyle 0,$ (4.4)
$\displaystyle-\frac{1}{2}\frac{\partial^{2}\beta}{\partial x\partial
z}-\frac{\beta}{2}\left(2\frac{\partial\beta}{\partial
z}\,\frac{\partial\beta}{\partial x}+\beta\frac{\partial^{2}\beta}{\partial
x\partial z}+\frac{\partial^{2}\beta}{\partial t\partial z}\right)$
$\displaystyle=$ $\displaystyle 0,$ (4.5)
$\displaystyle-\frac{3}{4}\left[\left(\frac{\partial\beta}{\partial
y}\right)^{2}+\left(\frac{\partial\beta}{\partial z}\right)^{2}\right]$
$\displaystyle=$ $\displaystyle 8\pi A,$ (4.6)
$\displaystyle\frac{1}{2}\left(2\frac{\partial\beta}{\partial
y}\,\frac{\partial\beta}{\partial x}+\beta\frac{\partial^{2}\beta}{\partial
x\partial y}+\frac{\partial^{2}\beta}{\partial t\partial y}\right)$
$\displaystyle=$ $\displaystyle 0,$ (4.7)
$\displaystyle\frac{1}{2}\left(2\frac{\partial\beta}{\partial
z}\,\frac{\partial\beta}{\partial x}+\beta\frac{\partial^{2}\beta}{\partial
x\partial z}+\frac{\partial^{2}\beta}{\partial t\partial z}\right)$
$\displaystyle=$ $\displaystyle 0,$ (4.8)
$\displaystyle\frac{1}{2}\frac{\partial\beta}{\partial
z}\,\frac{\partial\beta}{\partial y}$ $\displaystyle=$ $\displaystyle 0,$
(4.9) $\displaystyle-\left[\frac{\partial^{2}\beta}{\partial t\partial
x}+\beta\frac{\partial^{2}\beta}{\partial
x^{2}}+\left(\frac{\partial\beta}{\partial
x}\right)^{2}\right]-\frac{1}{4}\left[\left(\frac{\partial\beta}{\partial
y}\right)^{2}-\left(\frac{\partial\beta}{\partial z}\right)^{2}\right]$
$\displaystyle=$ $\displaystyle 8\pi B,$ (4.10)
$\displaystyle-\left[\frac{\partial^{2}\beta}{\partial t\partial
x}+\beta\frac{\partial^{2}\beta}{\partial
x^{2}}+\left(\frac{\partial\beta}{\partial
x}\right)^{2}\right]+\frac{1}{4}\left[\left(\frac{\partial\beta}{\partial
y}\right)^{2}-\left(\frac{\partial\beta}{\partial z}\right)^{2}\right]$
$\displaystyle=$ $\displaystyle 8\pi C.$ (4.11)
Substituting Eqs. (4.6) and (4.3) into Eq. (4.2) it follows immediately that,
$\mu=\beta^{2}(2D-A-p)+\frac{A}{3}\,\,.$ (4.12)
This expression shows that the fluid density not only depends on the pressure
components $A$ and $p$, but also on the momentum component $D$ and the warp
bubble, since it varies with the shift vector $\beta$ and, hence, the bubble
movement. So, the bubble shape modifies the fluid density in a local way, a
result that may imply an analogy with classical fluid dynamics and shock waves
in fluids with global velocity greater than the speed of sound in that medium.
Applying this analogy to the warp drive, it may mean that the warp bubble
plays the role of a shock wave in a fluid that moves with apparent velocity
greater than the speed of light for an outside observer far away from the
bubble, this being a result of the nonlinearity of the Einstein equations. The
bubble modifies the fluid density which then causes the bubble motion. This
classical relativistic fluid analogy may be a physical argument for a
mechanism which accounts for the great amount of energy necessary for the
feasibility of the warp drive.
After some algebra on the set of Einstein equations above they can be
rewritten as below,
$\displaystyle\beta^{2}(2D-A-p)+\frac{A}{3}$ $\displaystyle=$
$\displaystyle\mu,$ (4.13) $\displaystyle\frac{\partial^{2}\beta}{\partial
x\partial y}$ $\displaystyle=$ $\displaystyle 0,$ (4.14)
$\displaystyle\frac{\partial^{2}\beta}{\partial x\partial z}$ $\displaystyle=$
$\displaystyle 0,$ (4.15) $\displaystyle\left(\frac{\partial\beta}{\partial
y}\right)^{2}+\left(\frac{\partial\beta}{\partial z}\right)^{2}$
$\displaystyle=$ $\displaystyle-\frac{32}{3}\pi A,$ (4.16)
$\displaystyle\left(\frac{\partial\beta}{\partial
y}\right)^{2}-\left(\frac{\partial\beta}{\partial z}\right)^{2}$
$\displaystyle=$ $\displaystyle 16\pi(C-B),$ (4.17) $\displaystyle
2\frac{\partial\beta}{\partial y}\,\frac{\partial\beta}{\partial
x}+\frac{\partial^{2}\beta}{\partial t\partial y}$ $\displaystyle=$
$\displaystyle 0,$ (4.18) $\displaystyle 2\frac{\partial\beta}{\partial
z}\,\frac{\partial\beta}{\partial x}+\frac{\partial^{2}\beta}{\partial
t\partial z}$ $\displaystyle=$ $\displaystyle 0,$ (4.19)
$\displaystyle\frac{\partial\beta}{\partial z}\,\frac{\partial\beta}{\partial
y}$ $\displaystyle=$ $\displaystyle 0,$ (4.20)
$\displaystyle\frac{\partial}{\partial x}\left[\frac{\partial\beta}{\partial
t}+\frac{1}{2}\frac{\partial}{\partial x}(\beta^{2})\right]$ $\displaystyle=$
$\displaystyle-32\pi(B+C),$ (4.21)
$\displaystyle\frac{\partial^{2}\beta}{\partial
y^{2}}+\frac{\partial^{2}\beta}{\partial z^{2}}$ $\displaystyle=$
$\displaystyle 16\pi\beta(A-D).$ (4.22)
Eq. (4.20) shows that the solutions for the set of differential equations
above have similar alternative cases as in the perfect fluid solutions, that
is, either $\partial\beta/\partial z=0$ and/or $\partial\beta/\partial y=0$.
Both situations and their respective subcases will be discussed next.
Case 1: $\bm{\left[\displaystyle\frac{\partial\beta}{\partial z}=0\right]}$
The set of equations (4.13) to (4.22) simplify to,
$\displaystyle\mu=\beta^{2}(2D-A-p)+\frac{A}{3},$ (4.23)
$\displaystyle\frac{\partial^{2}\beta}{\partial x\partial y}=0,$ (4.24)
$\displaystyle\left(\frac{\partial\beta}{\partial
y}\right)^{2}=-\frac{32}{3}\pi A=16\pi(C-B),$ (4.25) $\displaystyle
2\frac{\partial\beta}{\partial y}\,\frac{\partial\beta}{\partial
x}+\frac{\partial^{2}\beta}{\partial t\partial y}=0,$ (4.26)
$\displaystyle\frac{\partial}{\partial x}\left[\frac{\partial\beta}{\partial
t}+\frac{1}{2}\frac{\partial}{\partial x}(\beta^{2})\right]=-32\pi(B+C).$
(4.27) $\displaystyle\frac{\partial^{2}\beta}{\partial
y^{2}}=16\pi\beta(A-D).$ (4.28)
From Eq. (4.25) a relation between the pressures $A,B$ and $C$ are
straightforward,
$B=C+\frac{2}{3}A.$ (4.29)
From Eqs. (4.24) and (4.25) it is easy to verify that $A$ and $C-B$ do not
depend on the $x$-coordinate. In addition, for real solutions $A$ must be
negative, assuming only real values that are equal to or smaller than zero.
Differentiating Eq. (4.25) with respect to $x$ yields,
$2\frac{\partial\beta}{\partial y}\frac{\partial^{2}\beta}{\partial x\partial
y}=0\,\Longrightarrow\frac{\partial\beta}{\partial
y}=0\quad\mbox{and/or}\quad\frac{\partial^{2}\beta}{\partial x\partial y}=0,$
(4.30)
and inserting the result $\partial\beta/\partial y=0$ into Eq. (4.26) it
follows that,
$\frac{\partial^{2}\beta}{\partial t\partial y}=0,$ (4.31)
The result $\partial\beta/\partial y=0$ means that $A=D$ from Eq. (4.28) and
$C=B$ from Eq. (4.25). Hence, the set of equations from Eq. (4.23) to Eq.
(4.27) simplify for $\partial\beta/\partial y=0$,
$\displaystyle\mu=\beta^{2}(D-p)+\frac{D}{3}\,=\,\beta^{2}(A-p)+\frac{A}{3},$
(4.32) $\displaystyle\frac{\partial}{\partial
x}\left[\frac{\partial\beta}{\partial t}+\frac{1}{2}\frac{\partial}{\partial
x}(\beta^{2})\right]=-64\pi B\,=-64\pi C.$ (4.33)
From the result $\partial^{2}\beta/\partial x\partial y=0$ of Eq. (4.30) the
set of equations (4.23) to (4.28) is recovered. Therefore, Eqs. (4.30) show
that this case separates itself in two conditions, either
$\partial\beta/\partial x=0$ or $\partial\beta/\partial y=0$. Next we analyze
both conditions.
Case 1a: $\bm{\left[\displaystyle\frac{\partial\beta}{\partial
z}=0\,\,\,\,\text{and}\,\,\,\,\frac{\partial\beta}{\partial x}=0\right]}$
Setting Eq.(4.27) equal to zero then $B=-C$, and the set of equations (4.23)
to (4.28) simplify to,
$\displaystyle\mu$ $\displaystyle=$
$\displaystyle\beta^{2}(2D-A-p)+\frac{A}{3},$ (4.34) $\displaystyle B$
$\displaystyle=$ $\displaystyle-C=\frac{1}{3}A,$ (4.35)
$\displaystyle\left(\frac{\partial\beta}{\partial y}\right)^{2}$
$\displaystyle=$ $\displaystyle 32\pi C,$ (4.36)
$\displaystyle\frac{\partial^{2}\beta}{\partial y^{2}}$ $\displaystyle=$
$\displaystyle 16\pi\beta(A-D).$ (4.37)
For this case, there is an equation of state given by Eq. (4.34), $\beta$ is a
function of time and $y$ coordinates and must be found by solving both Eqs.
(4.36) and (4.37) in terms of the pressures $A,C$ and $D$. Note that Eq.
(4.35) relates the pressures $A,B$ and $C$. The EMT for this case may be
written as follows,
$T_{\alpha\sigma}=\begin{pmatrix}\beta^{2}(2D-A)+A/3&-\beta D&0&0\\\ -\beta
D&A&0&0\\\ 0&0&A/3&0\\\ 0&0&0&-A/3\end{pmatrix}\,.$ (4.38)
One should also note that for the $T_{00}$ component from Eq. (4.38) to be of
positive value the following inequality must hold,
$\beta^{2}>\frac{A}{3(A-2D)}.$ (4.39)
Case 1b: $\bm{\left[\displaystyle\frac{\partial\beta}{\partial
z}=0\,\,\,\,\text{and}\,\,\,\,\frac{\partial\beta}{\partial y}=0\right]}$
Since Eq. (4.25) is equal to zero, then $B=C$. Similarly for Eq. (4.28) it is
clear that $A=D$. Consequently, the set of equations (4.23) to (4.28) simplify
to,
$\displaystyle\mu=\beta^{2}(2D-A-p)+\frac{A}{3},$ (4.40) $\displaystyle
B=C+\frac{2}{3}A,$ (4.41) $\displaystyle B=C,$ (4.42) $\displaystyle A=D,$
(4.43) $\displaystyle\frac{\partial}{\partial
x}\left[\frac{\partial\beta}{\partial t}+\frac{1}{2}\frac{\partial}{\partial
x}(\beta^{2})\right]=-64\pi B.$ (4.44)
But, from Eq. (4.25) we have that $A=0$ because $B=C$. So, from Eq. (4.32) we
have that $\mu=-\beta^{2}p$, and one is left with a non homogeneous Burgers
equation. This case reduces the above set of equations to
$\displaystyle\mu=-\beta^{2}p,$ (4.45) $\displaystyle B=C,$ (4.46)
$\displaystyle\frac{\partial}{\partial x}\left[\frac{\partial\beta}{\partial
t}+\frac{1}{2}\frac{\partial}{\partial x}(\beta^{2})\right]=-64\pi B.$ (4.47)
Eq. (4.45) represents an equation of state between matter density $\mu$ and
the pressure $p$. The pressures $B$ and $C$ are functions of the spacetime
coordinates $(t,x,y,z)$ and from Eq. (4.46) they are equal. Eq. (4.47) is a
non homogeneous Burgers equation, since its right hand side cannot be readily
integrated. It might mean that there is no conservation law that can describe
the warp drive for the PPF EMT. The only possible way leading to conservation
law is for $B$ being a constant, which then allows a straightforward
integration. However, since this is not the case, namely, all the pressures
from the EMT are not, necessarily, constant functions, it is necessary to
determine this functions through the boundary conditions. The EMT for the Case
1b case then yields,
$T_{\alpha\sigma}=\begin{pmatrix}0&0&0&0\\\ 0&0&0&0\\\ 0&0&B&0\\\
0&0&0&B\end{pmatrix}\,.$ (4.48)
The only non vanishing components of the PPF EMT above are $T_{22}$ and
$T_{33}$. This case recovers the perfect fluid Case 1b, that is, the dust EMTs
when one chooses $p=B=0$. For $B\neq 0$, one would have to solve Eq. (4.47) to
determine how the bubble moves in this type of fluid spacetime. Again,
negative matter density emerges from the Einstein equation solutions for this
specific choice of EMT.
Nevertheless, this is a rather peculiar EMT, since the $T_{00}$ component is
zero, but the equation of state (4.45) remains and only two diagonal terms are
not zero. In addition, it is contradictory with the perfect fluid solution
because the PPF reduces to the perfect fluid under the condition
$p=A=B=C=D,$ (4.49)
but in this solution $A=0$, but $B\not=0$. So, we discard this case as
unphysical.
Case 2: $\bm{\left[\displaystyle\frac{\partial\beta}{\partial y}=0\right]}$
The set of equations (4.13) to (4.22) simplify to,
$\displaystyle\mu=\beta^{2}(2D-A-p)+\frac{A}{3},$ (4.50)
$\displaystyle\frac{\partial^{2}\beta}{\partial x\partial z}=0,$ (4.51)
$\displaystyle\left(\frac{\partial\beta}{\partial
z}\right)^{2}=-\frac{32}{3}\pi A=-\,16\pi(B-C),$ (4.52) $\displaystyle
2\frac{\partial\beta}{\partial z}\,\frac{\partial\beta}{\partial
x}+\frac{\partial^{2}\beta}{\partial t\partial z}=0,$ (4.53)
$\displaystyle\frac{\partial}{\partial x}\left[\frac{\partial\beta}{\partial
t}+\frac{1}{2}\frac{\partial}{\partial x}(\beta^{2})\right]=-32\pi(B+C)\,\,,$
(4.54) $\displaystyle\frac{\partial^{2}\beta}{\partial
z^{2}}=16\pi\beta(A-D).$ (4.55)
Eq. (4.52) straightforwardly implies the following relationship between
pressures $A,B$ and $C$,
$B=C+\frac{2}{3}A.$ (4.56)
Eqs. (4.51) and (4.52) indicate that $A$ and $B-C$ do not depend on the
$x$-coordinate. In addition, for real solutions $A$ must be negative or zero.
Finally, Eq. (4.51) shows that that this case also unfolds in two subcases:
$\partial\beta/\partial z=0$ and/or $\partial\beta/\partial x=0$.
Case 2a: $\bm{\left[\displaystyle\frac{\partial\beta}{\partial
y}=0\,\,\,\,\text{and}\,\,\,\,\frac{\partial\beta}{\partial x}=0\right]}$
Clearly $B=-C$ as a consequence of Eq. (4.54). Then, the set of equations Eqs.
(4.50) to (4.55) simplifies to,
$\displaystyle\mu=\beta^{2}(2D-A-p)+\frac{A}{3},$ (4.57) $\displaystyle
B=-C=\frac{A}{3},$ (4.58) $\displaystyle\left(\frac{\partial\beta}{\partial
z}\right)^{2}=32\pi C,$ (4.59) $\displaystyle\frac{\partial^{2}\beta}{\partial
z^{2}}=16\pi\beta(A-D).$ (4.60)
The EMT for this case is given by the following expression,
$T_{\alpha\sigma}=\begin{pmatrix}\beta^{2}(2D-A)+A/3&-\beta D&0&0\\\ -\beta
D&A&0&0\\\ 0&0&-A/3&0\\\ 0&0&0&A/3\end{pmatrix},$ (4.61)
which is almost equal to Case 1a (Eq. 4.38), apart from the change of signs in
the components $T_{22}$ and $T_{33}$. Both have the same $T_{00}$ component
and hence, the same equation of state, and the condition for $T_{00}$ to be of
positive is given by,
$\beta^{2}>\frac{A}{3(A-2D)}.$ (4.62)
Besides, this expressions determines another inequality that $T_{00}$ must
follow to be positive,
$A-2D>0.$ (4.63)
Case 2b: $\bm{\left[\displaystyle\frac{\partial\beta}{\partial
y}=0\,\,\,\,\text{and}\,\,\,\,\frac{\partial\beta}{\partial z}=0\right]}$
The set of equations Eqs. (4.50) to (4.55) immediately simplify to,
$\displaystyle\mu=\beta^{2}(2D-A-p)+\frac{A}{3},$ (4.64) $\displaystyle
B=C-\frac{2}{3}A,$ (4.65) $\displaystyle B=C,$ (4.66) $\displaystyle A=D=0,$
(4.67) $\displaystyle\frac{\partial}{\partial
x}\left[\frac{\partial\beta}{\partial t}+\frac{1}{2}\frac{\partial}{\partial
x}(\beta^{2})\right]=-64\pi B\,=-64\pi C\,\,.$ (4.68)
Therefore $\mu=-\beta^{2}p$, a non homogeneous Burgers equation is also
present, and the expressions above are reduced to
$\displaystyle\mu=-\beta^{2}p,$ (4.69) $\displaystyle B=C,$ (4.70)
$\displaystyle\frac{\partial}{\partial x}\left[\frac{\partial\beta}{\partial
t}+\frac{1}{2}\frac{\partial}{\partial x}(\beta^{2})\right]=-64\pi B,$ (4.71)
which are the same results as obtained in Case 1b, as well as the EMT, that
is,
$T_{\alpha\sigma}=\begin{pmatrix}0&0&0&0\\\ 0&0&0&0\\\ 0&0&B&0\\\
0&0&0&B\end{pmatrix}\,.$ (4.72)
$\square$
Again, because this solution cannot fulfill the requirement given by Eq.
(4.49), in fact it contradicts it since one cannot have both $A=0$ and
$B\not=0$, this solution is, similarly to Case 1b, dismissed as unphysical.
Table 2 summarizes the Einstein equations solutions obtained as a result of
the PPF EMT in the Alcubierre warp drive metric.
Case | Conditions | Results
---|---|---
$1)\ \displaystyle{\frac{\partial\beta}{\partial z}=0}$ | $1a)\ \displaystyle{\frac{\partial\beta}{\partial x}=0}$ | $\begin{array}[]{ll}\\\\[-10.0pt] \displaystyle{\mu=\beta^{2}(2D-A-p)+\frac{A}{3}}\\\\[7.0pt] \beta=\beta(t,y)\\\\[7.0pt] \displaystyle{B=-C=\frac{A}{3}}\\\\[7.0pt] \displaystyle{\left(\frac{\partial\beta}{\partial y}\right)^{2}=32\pi C},\\\\[10.0pt] \displaystyle{\frac{\partial^{2}\beta}{\partial y^{2}}=16\pi\beta(A-D)}\\\\[7.0pt] \end{array}$
$1b)\ \displaystyle{\frac{\partial\beta}{\partial y}=0}$ | $\begin{array}[]{ll}\\\\[-10.0pt] \displaystyle{\mu=-\beta^{2}p}\\\\[7.0pt] \beta=\beta(t,x)\\\\[7.0pt] \displaystyle{B=C}\\\\[5.0pt] \displaystyle{A=D=0}\\\\[7.0pt] \displaystyle{\frac{\partial}{\partial x}\left[\frac{\partial\beta}{\partial t}+\frac{1}{2}\frac{\partial}{\partial x}(\beta^{2})\right]=-64\pi B}\\\\[8.0pt] \rightarrow\mbox{solution {dismissed} as unphysical}\\\ \end{array}$
$2)\ \displaystyle{\frac{\partial\beta}{\partial y}=0}$ | $2a)\ \displaystyle{\frac{\partial\beta}{\partial x}=0}$ | $\begin{array}[]{ll}\\\\[-10.0pt] \displaystyle{\mu=\beta^{2}(2D-A-p)+\frac{A}{3}}\\\\[7.0pt] \beta=\beta(t,z)\\\\[7.0pt] \displaystyle{B=-C=\frac{A}{3}}\\\\[7.0pt] \displaystyle{\left(\frac{\partial\beta}{\partial z}\right)^{2}=32\pi C}\\\\[9.0pt] \displaystyle{\frac{\partial^{2}\beta}{\partial z^{2}}=16\pi\beta(A-D)}\\\\[7.0pt] \end{array}$
$2b)\ \displaystyle{\frac{\partial\beta}{\partial z}=0}$ | $\begin{array}[]{ll}\\\\[-10.0pt] \displaystyle{\mu=-\beta^{2}p}\\\\[7.0pt] \beta=\beta(t,x)\\\\[7.0pt] \displaystyle{B=C}\\\\[5.0pt] \displaystyle{A=D=0}\\\\[7.0pt] \displaystyle{\frac{\partial}{\partial x}\left[\frac{\partial\beta}{\partial t}+\frac{1}{2}\frac{\partial}{\partial x}(\beta^{2})\right]=-64\pi B}\\\\[8.0pt] \rightarrow\mbox{solution {dismissed} as unphysical}\end{array}$
Table 2: Summary of all solutions of the Einstein equations for the Alcubierre
warp drive metric having the parametrized perfect fluid (PPF) EMT. The cases,
conditions and their respective designations are the same as in Table 1
### 4.1 Discussion
As in the perfect fluid situations, Cases 1b and 2b are the same for the PPF,
producing the same equations of state, coordinate dependency for the $\beta$
function and a non-homogeneous Burgers equation. As stated above, only for a
constant pressure $B$ that a conservation law could possibly emerge from the
Burgers equation (4.71) that is common to both cases. Otherwise this
expression cannot be readily integrated unless further boundary conditions are
met (see Section 5.3 below). Nevertheless, because the solutions of these two
cases cannot satisfy the requirement established by Eq. (4.49) for reducing
the PPF to the perfect fluid, they are dismissed as unphysical.
Regarding Cases 1a and 2a, the only physically plausible solutions remaining
for the PPF EMT, they must obey the same inequalities (4.62) and (4.63) so
that the $T_{00}$ component be positive in both Eqs. (4.38) and (4.61). The
function $\beta$ depends on the $y$ coordinate in the former case and the $z$
one in the later, and, respectively, are a result of the integration of the
following differential equations,
$\beta\frac{\partial\beta}{\partial y}=\frac{C}{A-D},$ (4.73)
$\beta\frac{\partial\beta}{\partial z}=\frac{C}{A-D}.$ (4.74)
However, the quantities $A$, $C$ and $D$ are function of the coordinates, thus
integrating the expressions above require further conditions, such as the ones
respectively discussed in Sections 5.1 and 5.2 below.
#### 4.1.1 Equations of state
The perfect fluid solutions have equations of state given by (Ref. EllisElst,
, p. 14),
$p=p(\mu)=(\gamma-1)\mu,\;\;\;\dot{\gamma}=\frac{\mathrm{d}\gamma}{\mathrm{d}t}=0,$
(4.75)
Ordinary fluids can be approximated with $1\leq\gamma\leq 2$, where incoherent
matter, or dust, corresponds to $\gamma=1$, and radiation to
$\gamma=\frac{4}{3}$.
In the analyzes above we found that for the perfect fluid content solution
Cases 1a and 2a for the warp drive metric the equation of state is given by
$p=3\mu$ (see Table 1), which corresponds to $\gamma=4$. However, they do not
allow a simple expressions for the equation of state similar to Eq. (4.75) for
the PPF solutions 1a and 2a (see Table 2). The PPF content solution for the
Cases 1b and 2b were dismissed as unphysical,
## 5 EMT Divergence of the perfect fluid and PPF
In this section we will investigate the associated conservation laws to the
Einstein equations under a warp drive spacetime by means of the usual
condition that the EMT divergence must be zero for both the perfect fluid and
PPF. We shall start with the EMT for the PPF because from its very definition
the perfect fluid can be recovered by setting the equality (4.49).
For the fluids discussed here, setting ${T^{\alpha\sigma}}_{;\sigma}=0$ in the
EMT (4.1) results in the following expressions,
$\displaystyle-\frac{\partial\beta}{\partial
x}(D+\mu)-\frac{\partial\mu}{\partial t}-\beta\left[\frac{\partial D}{\partial
x}+\frac{\partial\mu}{\partial x}+\frac{\partial\beta}{\partial
t}(2p+A-3D)\right]$ $\displaystyle+\beta^{2}\left[\frac{\partial D}{\partial
t}-\frac{\partial p}{\partial t}+3\frac{\partial\beta}{\partial
x}(D-p)\right]+\beta^{3}\left(\frac{\partial D}{\partial x}-\frac{\partial
p}{\partial x}\right)$ $\displaystyle=0,$ (5.1) $\displaystyle\frac{\partial
A}{\partial x}+\frac{\partial\beta}{\partial
t}(D-A)+\beta\left[3\frac{\partial\beta}{\partial x}(D-A)+\frac{\partial
D}{\partial t}-\frac{\partial A}{\partial
t}\right]+\beta^{2}\left(\frac{\partial D}{\partial x}-\frac{\partial
A}{\partial x}\right)$ $\displaystyle=0,$ (5.2) $\displaystyle\frac{\partial
B}{\partial y}+\beta\frac{\partial\beta}{\partial y}(D-A)$ $\displaystyle=0,$
(5.3) $\displaystyle\frac{\partial C}{\partial
z}+\beta\frac{\partial\beta}{\partial z}(D-A)$ $\displaystyle=0.$ (5.4)
### 5.1 Case 1a: $\bm{\left[\displaystyle\frac{\partial\beta}{\partial
z}=0\right.}$ and $\bm{\left.\displaystyle\frac{\partial\beta}{\partial
x}=0\right]}$
Eqs. (5.1) to Eq. (5.4) are reduced to the ones below,
$\displaystyle-\frac{\partial\mu}{\partial t}-\beta\left[\frac{\partial
D}{\partial x}+\frac{\partial\mu}{\partial x}+\frac{\partial\beta}{\partial
t}(2p+A-3D)\right]+\beta^{2}\left(\frac{\partial D}{\partial t}-\frac{\partial
p}{\partial t}\right)$ $\displaystyle+\beta^{3}\left(\frac{\partial
D}{\partial x}-\frac{\partial p}{\partial x}\right)$ $\displaystyle=0,$ (5.5)
$\displaystyle\frac{\partial A}{\partial x}+\frac{\partial\beta}{\partial
t}(D-A)+\beta\left(\frac{\partial D}{\partial t}-\frac{\partial A}{\partial
t}\right)+\beta^{2}\left(\frac{\partial D}{\partial x}-\frac{\partial
A}{\partial x}\right)$ $\displaystyle=0,$ (5.6) $\displaystyle\frac{\partial
B}{\partial y}+\beta\frac{\partial\beta}{\partial y}(D-A)$ $\displaystyle=0,$
(5.7) $\displaystyle\frac{\partial C}{\partial z}$ $\displaystyle=0.$ (5.8)
The results above concern the PPF. These four equations together with the five
ones shown in the respective results of Table 2 means an overdetermined system
for the six unknowns $\mu,p,A,B,C$, $D$.
The perfect fluid is recovered by setting Eq. (4.49). Hence, Eqs. (5.5) to
(5.8) become,
$\displaystyle-\frac{\partial\mu}{\partial t}-\beta\left(\frac{\partial
p}{\partial x}+\frac{\partial\mu}{\partial x}\right)$ $\displaystyle=0,$ (5.9)
$\displaystyle\frac{\partial p}{\partial x}$ $\displaystyle=0,$ (5.10)
$\displaystyle\frac{\partial p}{\partial y}$ $\displaystyle=0,$ (5.11)
$\displaystyle\frac{\partial p}{\partial z}$ $\displaystyle=0,$ (5.12)
and, thus, the pressure does not depend on the spatial coordinates. In
addition, Eq. (5.9) reduces to the expression below,
$\frac{\partial\mu}{\partial t}+\beta\frac{\partial\mu}{\partial x}=0,$ (5.13)
which is the continuity equation, where $\mu$ plays the role of the fluid
density and $\beta$ is the flow velocity vector field. It is worth mentioning
that for a constant density the fluid has incompressible flow, then all the
partial derivatives of $\beta$ in terms of the spatial coordinates vanish and
the flow velocity vector field has null divergence, this being a classical
fluid dynamics scenario, and the local volume dilation rate is zero.
### 5.2 Case 2a: $\bm{\left[\displaystyle\frac{\partial\beta}{\partial
y}=0\right.}$ and $\bm{\left.\displaystyle\frac{\partial\beta}{\partial
x}=0\right]}$
Equations (5.1) to (5.4) simplify to the following expressions:
$\displaystyle-\frac{\partial\mu}{\partial t}-\beta\left[\frac{\partial
D}{\partial x}+\frac{\partial\mu}{\partial x}+\frac{\partial\beta}{\partial
t}(2p+A-3D)\right]$ $\displaystyle+\beta^{2}\left(\frac{\partial D}{\partial
t}-\frac{\partial p}{\partial t}\right)+\beta^{3}\left(\frac{\partial
D}{\partial x}-\frac{\partial p}{\partial x}\right)$ $\displaystyle=0,$ (5.14)
$\displaystyle\frac{\partial A}{\partial x}+\frac{\partial\beta}{\partial
t}(D-A)+\beta\left(\frac{\partial D}{\partial t}-\frac{\partial A}{\partial
t}\right)+\beta^{2}\left(\frac{\partial D}{\partial x}-\frac{\partial
A}{\partial x}\right)$ $\displaystyle=0,$ (5.15) $\displaystyle\frac{\partial
B}{\partial y}$ $\displaystyle=0,$ (5.16) $\displaystyle\frac{\partial
C}{\partial z}+\beta\frac{\partial\beta}{\partial z}(D-A)$ $\displaystyle=0.$
(5.17)
Assuming Eq. (4.49) the perfect fluid is recovered, resulting in the already
discussed Eqs. (5.9) to (5.12), as well as the continuity equation (5.13).
### 5.3 Cases 1b and 2b:
$\bm{\left[\displaystyle\frac{\partial\beta}{\partial y}=0\right.}$ and
$\bm{\left.\displaystyle\frac{\partial\beta}{\partial z}=0\right]}$
It follows from Table 1 that conditions 1b and 2b for the perfect fluid imply
$\mu=p=0$, which also means $A=B=C=D=0$. Then, Eqs. (5.1) to (5.4) of the null
EMT divergence are are immediately satisfied for the perfect fluid.
Regarding the PPF, we have already seen that in these cases the solutions were
dismissed as unphysical. Nonetheless, it is worth analyzing the resulting
expressions to show that they lead to trivial cases or to the dust solution
already studied in Ref. nos .
Table 2 shows the following solutions for the PPF EMT considering Cases 1b and
2b: $A=D=0$, $B=C$, and $\mu=-\beta^{2}p$. Hence, Eqs. (5.1) to (5.4) are
reduced to the following expressions,
$\displaystyle-\frac{\partial\mu}{\partial
t}-\beta\left(\frac{\partial\mu}{\partial x}+2p\frac{\partial\beta}{\partial
t}\right)-\beta^{2}\frac{\partial p}{\partial t}-\beta^{3}\frac{\partial
p}{\partial x}$ $\displaystyle=0,$ (5.18) $\displaystyle\frac{\partial
B}{\partial y}$ $\displaystyle=0,$ (5.19) $\displaystyle\frac{\partial
C}{\partial z}=\frac{\partial B}{\partial z}$ $\displaystyle=0.$ (5.20)
Eqs. (5.19) and (5.20) imply that $B$ does not depend on both the $y$ and $z$
coordinates. Inserting the state equation $\mu=-\beta^{2}p$ (see Table 2) into
Eq. (5.18) yields
$p\beta^{2}\frac{\partial\beta}{\partial x}=0.$ (5.21)
This expression can be separated in three subcases: $\beta=0$, $p=0$,
$\partial\beta/\partial x=0$. According to Eq. (4.71) the last subcase implies
$B=0$ and, hence, there is no Burgers equation. Let us discuss below the
consequences of each of these subcases.
#### 5.3.1 Subcase $\left[\,\beta=0\,\right]$
This solution reduces the warp drive metric (2.6) into a Minkowski flat
spacetime. There is no warp bubble and, so, no warp drive.
#### 5.3.2 Subcase $\left[\,B=0\,\right]$
Remembering Eqs. (4.48) and (4.72) this means that all EMT components are
zero, even though, the equation of state $\mu=-\beta^{2}p$ remains. But, since
there is no EMT both $\mu$ and $p$ must vanish, hence this case reduces itself
to the dust solution nos .
#### 5.3.3 Subcase $\left[\,p=0\,\right]$
Considering the equation of state $\mu=-\beta^{2}p$ then the matter density is
also zero. The EMT for the PPF assumes the form given by Eq. (4.72), which
implies in pressure without matter density, an unphysical situation that
should either be dismissed or assumed to be just the dust solution.
## 6 Energy conditions
The energy conditions are well known inequalities discussed by Hawking and
Ellis HawkingEllis1973 that may be applied to the matter content in physical
systems as boundary conditions and to test if the energy of such systems
follows positive constraint values.
This section aims at obtaining these conditions for the Alcubierre warp drive
geometry considering the previously discussed EMTs for both the perfect fluid
and PPF. Our focus will be on the main classical inequalities, namely, the
weak, dominant, strong and null energy conditions. Our analysis starts with
the PPF EMT defined in Eq. (4.1) in order to constrain its quantities so that
the inequalities are satisfied for each of the four just named conditions.
Then the same analysis for the perfect fluid is performed by reducing the
results to the choice set by Eq. (4.49), but considering only the physically
plausible cases.
### 6.1 Weak Energy Condition (WEC)
The WEC requires that at each point of the spacetime the EMT must obey the
following inequality,
$T_{\alpha\sigma}\,u^{\alpha}u^{\sigma}\geq 0.$ (6.1)
This is valid for any timelike vector $\textbf{u}\,(u_{\alpha}u^{\alpha}<0)$
as well as any null vector $\textbf{k}\,(k_{\alpha}k^{\alpha}=0$; see Section
6.4 below). For an observer with a unit tangent vector v at some point of the
spacetime, the local energy density measured by any observer is non-negative
HawkingEllis1973 . Considering the Eulerian (normal) observers from
Alcubierre1994 with 4-velocity given by the expressions below
$u^{\alpha}=(1,\beta,0,0)\,,\ \ u_{\alpha}=(-1,0,0,0),$ (6.2)
and the EMT, given by Eq. (4.1), computing the expression
$T_{\alpha\sigma}\,u^{\alpha}u^{\sigma}$ for the WEC yields,
$\displaystyle T_{\alpha\sigma}\,u^{\alpha}u^{\sigma}$
$\displaystyle=T_{00}\,u^{0}u^{0}+2T_{01}\,u^{0}u^{1}+T_{11}\,u^{1}u^{1}$
$\displaystyle=\mu+\beta^{2}(p-2D+A)\,.$ (6.3)
Let us now calculate the WEC for all perfect fluid and PPF Einstein equations
solutions obtained in the previous sections as referred to in Table 2.
#### 6.1.1 Cases 1a and 2a
Let us start by considering the PPF. Substituting the equation of state (4.12)
into Eq. (6.3) yields,
$T_{\alpha\sigma}\,u^{\alpha}u^{\sigma}=\frac{A}{3}\geq 0.$ (6.4)
So, the pressure $A$ must positive. Considering that both cases led to Eq.
(4.58), the pressure $C$ must be negative. Besides, for the state equation
(4.12) the density $\mu$ becomes positive if
$A+p\leq 2D\,\,.$ (6.5)
This inequality implies that it is possible for the fluid density to account
for both negative and positive values in a local way depending on how the
momentum and pressure components relate to each other in the spacetime.
Remembering Eq. (4.49), the resulting inequality (6.4) indicates that the WEC
is satisfied with the perfect fluid EMT for a positive pressure $p$.
Considering Eqs. (3.30) and (3.31) that would mean a complex solution for
$\beta$, possibility already envisaged in Section 3.2. Note that this is a
general result for the perfect fluid with a cross term solution.
#### 6.1.2 Cases 1b and 2b
It has been shown above that in these two cases the PPF solutions result in
$A=D=0$ and $\mu=-\beta^{2}p$, which substituted in Eq. (6.3) yield
$T_{\alpha\sigma}\,u^{\alpha}u^{\sigma}=0$. Hence, the WEC is not violated for
the PPF.
For the perfect fluid, Table 1 shows that in these two cases the perfect fluid
solutions reduce to the dust content solution for the warp drive metric, which
trivially satisfies the WEC (Ref. nos, , p. 6).
### 6.2 Dominant Energy Condition (DEC)
The DEC states that for every timelike vector $u_{a}$ the following inequality
must be satisfied,
$T^{\alpha\beta}\,u_{\alpha}u_{\beta}\geq 0,\quad\text{and}\quad
F^{\alpha}F_{\alpha}\leq 0,$ (6.6)
where $F^{\alpha}=T^{\alpha\beta}u_{\beta}$ is a non-spacelike vector. This
condition means that for any observer the local energy density is non-negative
and the local energy flow vector is non-spacelike. In any orthonormal basis
the energy dominates the other components of the EMT,
$T^{00}\geq|T^{ab}|,\ \text{for each}\ a,b.$ (6.7)
Hawking and Ellis HawkingEllis1973 suggested that this condition must hold
for all known forms of matter and that it should be the case in all
situations.
Evaluating this condition for PPF, the first DEC inequality is given by the
following expression,
$T^{\alpha\beta}\,u_{\alpha}u_{\beta}=T^{00}=\beta^{2}(A-2D+p)+\mu\geq 0,$
(6.8)
whereas the second inequality $F^{\alpha}F_{\alpha}\leq 0$ yields,
$\displaystyle(T^{\alpha\beta}\,u_{\beta})(T_{\alpha\beta}\,u^{\beta})=$
$\displaystyle-\mu^{2}-A^{2}\beta^{4}-\beta^{4}p^{2}+A^{2}\beta^{2}-(4\beta^{4}-\beta^{2})D^{2}$
$\displaystyle+2(2A\beta^{4}-A\beta^{2})D-2(A\beta^{2}-2\beta^{2}D)\mu$
$\displaystyle-2(A\beta^{4}-2\beta^{4}D+\beta^{2}\mu)p\leq 0.$ (6.9)
As before, let us analyze next the solutions of the Einstein equations
considering each group of subcases as referred to in Table 2.
#### 6.2.1 Cases 1a and 2a
The two important results for these two cases are Eqs. (4.57) and (4.58).
Substituting them into the two DEC inequalities above result in the following
expressions,
$\displaystyle T^{\alpha\beta}\,u_{\alpha}u_{\beta}$
$\displaystyle=\frac{A}{3}\geq 0,$ (6.10)
$\displaystyle(T^{\alpha\beta}\,u_{\beta})(T_{\alpha\beta}\,u^{\beta})$
$\displaystyle=\left[(A-D)\beta+\frac{A}{3}\right]\left[(A-D)\beta-\frac{A}{3}\right]\leq
0.$ (6.11)
Eq. (6.10) is equal to Eq. (6.4), so this dominant energy condition is the
same as the weak one. Eq. (6.11) separates in two inequalities,
$\displaystyle-\frac{A}{3\left(D-A\right)}\leq\beta\leq\frac{A}{3\left(D-A\right)},$
(6.12)
$\displaystyle-\frac{A}{3\left(A-D\right)}\leq\beta\leq\frac{A}{3\left(A-D\right)},$
(6.13)
which can be rewritten as
$\left|\beta\right|\leq\frac{1}{3}\left|\frac{A}{D-A}\right|.$ (6.14)
In addition to the inequalities above, Eq. (6.11) also provides the following
solutions,
$\beta=\pm\frac{A}{3(A-D)}.$ (6.15)
Equation (6.14) shows that $\beta$ is upper bounded, but since both $A$ and
$D$ have no restraints, $\beta$ can be greater than unity. Remembering that
$\beta=v_{s}(t)f[r_{s}(t)]$, the dominant energy condition is not violated for
cases where the apparent warp bubble speed is greater than the speed of light.
This result also depends on the relation between the pressure component $A$
and the momentum component $D$ as can be seen in Eqs. (6.12) and (6.13). As an
example suppose that $A=D+1$ in Eq. (6.12), so that
$-\frac{D+1}{3}\leq v_{s}(t)f[r_{s}(t)]\leq\frac{D+1}{3}.$ (6.16)
The symmetry on the negative and positive values that $\beta$ may assume means
the direction in which the warp bubble is moving on the $x$-axis on the
spatial hyper surface. Note that, depending on the value of $D$, the warp
bubble may assume the speed $v_{s}(t)$ greater than the speed of light.
Regarding the perfect fluid, choosing as in Eq. (4.49) the DEC is given by the
following expressions,
$\displaystyle T^{\alpha\beta}\,u_{\alpha}u_{\beta}=\frac{p}{3}$
$\displaystyle\geq 0,$ (6.17)
$\displaystyle(T^{\alpha\beta}\,u_{\beta})(T_{\alpha\beta}\,u^{\beta})=p^{2}$
$\displaystyle\geq 0,$ (6.18)
which means that for the DEC to be satisfied a positively valued matter
density is enough (see Table 1), although this also means a complex result for
$\beta$ (see Section 3.2).
#### 6.2.2 Cases 1b and 2b
In these cases the PPF solutions, although considered unphysical, produced
$B=C$, $A=D=0$ and $\mu=-\beta^{2}p$ (see Table 2). So, as shown by the
expressions below the DEC is immediately satisfied,
$\displaystyle T^{\alpha\beta}\,u_{\alpha}u_{\beta}=T^{00}=0,$ (6.19)
$\displaystyle(T^{\alpha\beta}\,u_{\beta})(T_{\alpha\beta}\,u^{\beta})=0\,\,,$
(6.20)
The perfect fluid is the dust solution for both cases (see Table 1), which is
a vacuum solution and satisfies the inequalities above trivially.
### 6.3 Strong Energy Condition (SEC)
For any timelike vector $u^{\alpha}$ the EMT must obey the following
inequality for the SEC be valid,
$\left(T_{\alpha\beta}-\frac{1}{2}T\,g_{\alpha\beta}\right)u^{\alpha}u^{\beta}\geq
0,$ (6.21)
This requirement is stronger than the WEC and only makes sense in general
relativistic framework because this theory is governed by the Einstein
equations. These conditions imply that gravity is always attractive.
To obtain expression for the SEC let us start with the scalar
$T=g^{{}_{\alpha\beta}}T_{\alpha\beta}$, where $g^{\alpha\beta}$ is given by
the Alcubierre warp drive metric and $T_{\alpha\beta}$ is the EMT for the PPF
(Eq. 4.1). The result is shown below,
$T=g^{{}_{\alpha\beta}}T_{\alpha\beta}=-\mu+A+B+C+\beta^{2}(2D-p-A).$ (6.22)
Substituting the equation of state (4.12) into the expression above yields,
$T=\frac{2}{3}A+B+C.$ (6.23)
As shown in Section 6.1 above, Eq. (4.12) implies that
$T_{\alpha\sigma}u^{\alpha}u^{\sigma}=A/3$ for $u^{\sigma}=(1,\beta,0,0)$.
Inasmuch as $g_{\alpha\sigma}u^{\alpha}u^{\sigma}=-1$, then the SEC are given
by the following expression,
$\left(T_{\alpha\sigma}-\frac{1}{2}T\,g_{\alpha\sigma}\right)u^{\alpha}u^{\sigma}=\frac{2A}{3}+\frac{(B+C)}{2}\geq
0.$ (6.24)
The particular cases, as summarized in Table 2, can now be discussed.
#### 6.3.1 Cases 1a and 2a
These cases resulted in the constraint $B=-\,C=A/3$. Eq. (6.24) may then be
rewritten in the form below,
$\left(T_{\alpha\sigma}-\frac{1}{2}T\,g_{\alpha\sigma}\right)u^{\alpha}u^{\sigma}=\frac{2A}{3}=2B\geq
0\,\,.$ (6.25)
Clearly, for the SEC be satisfied it is only necessary that the pressure
component $A$ must be non negative.
Regarding the perfect fluid its respective SEC reads as,
$\left(T_{\alpha\sigma}-\frac{1}{2}T\,g_{\alpha\sigma}\right)u^{\alpha}u^{\sigma}=\frac{2p}{3}\geq
0.$ (6.26)
So, the SEC can be satisfied by the perfect fluid in these subcases by
requiring a positive pressure $p$.
#### 6.3.2 Cases 1b and 2b
In these cases the PPF solutions stated that $A=D=0$ and $B=C$. Hence, the SEC
satisfied under the constraint below,
$\left(T_{\alpha\sigma}-\frac{1}{2}T\,g_{\alpha\sigma}\right)u^{\alpha}u^{\sigma}=B\geq
0,$ (6.27)
despite providing unphysical solutions. The perfect fluid solutions are just
the dust solution that is a vacuum solution nos , so satisfying the SEC
trivially.
### 6.4 Null Energy Condition (NEC)
The SEC and WEC are satisfied in the limit of the null observers with
4-velocities k. To satisfy the NEC the EMT must follow the inequality below,
$T_{\alpha\sigma}\,k^{\alpha}k^{\sigma}\geq 0,\ \text{for any null vector}\
k^{\alpha}.$ (6.28)
To calculate the NEC let us suppose the following null vector below,
$k^{\alpha}=(a,b,0,0),$ (6.29)
where the relation between $a$ and $b$ can be obtained by imposing the
condition
$k_{\alpha}k^{\alpha}=0,$ (6.30)
that leads to
$a^{2}\beta^{2}-2ab\beta-a^{2}+b^{2}=0,$ (6.31)
whose roots for $a/b$ yield,
$\frac{a}{b}=\frac{1}{\beta+1}\ \ \text{and}\ \
\frac{a}{b}=\frac{1}{\beta-1}.$ (6.32)
Considering the results above, the NEC may be written as follows,
$T_{\alpha\sigma}\,k^{\alpha}k^{\sigma}=T_{00}k^{0}k^{0}+2T_{01}k^{0}k^{1}+T_{11}k^{1}k^{1},$
(6.33)
where $T_{\alpha\sigma}$ is the PPF EMT of Eq. (4.1). So, the null condition
is given by the following expression,
$T_{\alpha\sigma}\,k^{\alpha}k^{\sigma}=a^{2}\beta^{2}p-2ab\beta
D+b^{2}A+a^{2}\mu\geq 0,$ (6.34)
Substituting the two roots in Eqs. (6.32) into Eq. (6.34) results in
expressions below,
$\displaystyle\frac{b^{2}}{(\beta+1)^{2}}\,\left[\mu-\beta^{2}(2D-A-p)+2\beta(A-D)+A\right]\geq
0,$ (6.35)
$\displaystyle\frac{b^{2}}{(\beta-1)^{2}}\,\left[\mu-\beta^{2}(2D-A-p)-2\beta(A-D)+A\right]\geq
0.$ (6.36)
Next, as previously, we shall proceed to investigate first the Cases 1a and 2a
and then 1b and 2b ones, as defined in Tables 1 and 2, for the PPF and the
perfect fluid.
#### 6.4.1 Cases 1a and 2a
Considering the PPF first, the equation of state (4.57) is a solution of both
cases (see Table 2), so the resulting NEC given by Eqs. (6.35) and (6.36)
simplify to the following expressions,
$\displaystyle\frac{b^{2}}{(\beta+1)^{2}}\,\left[\frac{4A}{3}+2\beta(A-D)\right]\geq
0,$ (6.37)
$\displaystyle\frac{b^{2}}{(\beta-1)^{2}}\,\left[\frac{4A}{3}-2\beta(A-D)\right]\geq
0,$ (6.38)
whose solutions for $\beta$ yield the NEC for the PPF EMT, as below,
$\displaystyle\beta\geq\frac{2}{3}\frac{A}{D-A}\ \ \text{and}\ \ \beta\neq-1,$
(6.39) $\displaystyle\beta\leq\frac{2}{3}\frac{A}{A-D}\ \ \text{and}\ \
\beta\neq 1.$ (6.40)
Remembering that $\beta=v_{s}(t)f(r_{s})$, then Eq. (6.39) states that the
warp bubble can not assume the negative sign of the speed of light, but it may
assume values above it. To verify that, we have to choose $A=D-1>2/3$ and
$\beta>1$. Eq. (6.40) states that $\beta$ cannot assume the exact value of the
speed of light and it is bounded by a superior value which can be greater than
the speed of light. To have this, one just have to choose for example
$A=D+1>2/3$.
For the perfect fluid, according to Eq. (4.49) all pressures are equal. Hence,
substituting both values for $a/b$ found in Eq. (6.34) into Eqs. (6.35) and
(6.36) they then simplify to expressions below,
$\displaystyle\frac{b^{2}}{(\beta+1)^{2}}\,\left(\mu+p\right)\geq 0,$ (6.41)
$\displaystyle\frac{b^{2}}{(\beta-1)^{2}}\,\left(\mu+p\right)\geq 0.$ (6.42)
These results mean that for the NEC is satisfied for the perfect fluid
provided that
$\mu+p\geq 0.$ (6.43)
Remembering that $p=3\mu$ for the perfect fluid in theses two cases, then the
NEC is fulfilled if the matter density does not have negative values.
#### 6.4.2 Cases 1b and 2b
These cases, although set as unphysical, we already concluded that the PPF
relates its pressures by the following expressions: $A=D=0$ and $B=C$.
Therefore, Eqs. (6.35) and (6.36) yield,
$\displaystyle\frac{b^{2}}{(\beta+1)^{2}}\,\left(\mu+\beta^{2}p\right)\geq 0,$
(6.44)
$\displaystyle\frac{b^{2}}{(\beta-1)^{2}}\,\left(\mu+\beta^{2}p\right)\geq 0.$
(6.45)
So, the NEC establishes that $\mu+\beta^{2}p\geq 0$, which is readily
satisfied since in these cases we already concluded that $\mu=-\beta^{2}p$,
provided that $\beta\neq-1$ for the former and $\beta\neq 1$ for the latter.
Regarding the perfect fluid, these cases reduce to the dust solution studied
in nos , which is a vacuum solution and, therefore, immediately satisfies the
NEC.
Table 3 summarizes the results of all energy conditions for the PPF and
perfect fluid EMTs with the Alcubierre warp drive metric.
Cases (refer to Tables 1 and 2) | Energy Conditions | Results
---|---|---
1a and 2a | Weak | $\begin{array}[]{ll}\\\\[-6.0pt] \text{Perfect fluid:}\ \displaystyle{p\geq 0}\\\\[8.0pt] \text{PPF:}\ \displaystyle{A\geq 0}\\\\[8.0pt] \end{array}$
Dominant | $\begin{array}[]{ll}\\\\[-6.0pt] \text{Perfect Fluid:}\ \displaystyle{p\geq 0}\\\\[8.0pt] \text{PPF:}\ \displaystyle{A\geq 0}\ \ \text{and}\ \ \displaystyle{\left|\beta\right|\leq\frac{1}{3}\left|\frac{A}{D-A}\right|}\\\\[12.0pt] \end{array}$
Strong | $\begin{array}[]{ll}\\\\[-6.0pt] \text{Perfect fluid}:\ \displaystyle{p}\geq 0\\\\[8.0pt] \text{PPF}:\ \displaystyle{A}\geq 0\\\\[8.0pt] \end{array}$
Null | $\begin{array}[]{ll}\\\\[-6.0pt] \text{Perfect fluid}:\ \displaystyle{\mu+p\geq 0}\\\\[8.0pt] \text{PPF}:\ \displaystyle{\beta\geq\frac{2}{3}\frac{A}{D-A}\ \ \text{and}\ \ \beta\neq-1}\\\\[8.0pt] \text{PPF}:\ \displaystyle{\beta\leq\frac{2}{3}\frac{A}{A-D}\ \ \text{and}\ \ \beta\neq 1}\\\\[8.0pt] \end{array}$
1b and 2b | Weak | immediately satisfied for perfect fluid and PPF.
Dominant | immediately satisfied for perfect fluid and PPF.
Strong | $\begin{array}[]{ll}\\\\[-6.0pt] \text{Perfect fluid: immediately satisfied.}\\\\[8.0pt] \text{PPF:}\ B\geq 0\\\\[2.0pt] \end{array}$
Null | immediately satisfied for perfect fluid and PPF.
Table 3: Summary of all energy conditions results for the perfect fluid and
PPF EMTs with the Alcubierre warp drive spacetime geometry.
## 7 Further discussions
This section discusses some points, and raises others, all related to the
physics of the warp drive as suggested by the results presented in the
previous sections. It aims at offering some thoughts that may be important in
fostering further understanding on how a superluminal travel can be achieved.
### 7.1 Regulating Function and the Burgers Equation
The regulating function (2.8) describes the shape of the warp bubble, but it
is not uniquely determined. However, the integration the Einstein equations in
both the perfect fluid and the PPF led to the appearance of generic functions
in the dynamic equations which may end up connected to the regulating
function, a situation that adds to its nonuniqueness. This means that
physically feasible superluminal speeds will require the specification of
these generic functions, possibly by boundary conditions. We shall show below
an example of this situation using the dust solution.
Considering Eqs. (2.3) and (2.10) the partial derivative of the Burgers
equation (3.19) yields,
$\frac{\partial\beta}{\partial
t}=f\frac{d^{2}x_{s}}{dt^{2}}+v_{s}\frac{\partial f}{\partial
r_{s}}\frac{\partial r_{s}}{\partial t},$ (7.1)
where the simplified notation $f=f\left[r_{s}(t)\right]$ and $v_{s}=v_{s}(t)$
was adopted. Since,
$\frac{\partial\beta}{\partial x}=v_{s}\frac{\partial f}{\partial
r_{s}}\frac{\partial r_{s}}{\partial x},$ (7.2)
the Burgers equation becomes,
$f\frac{d^{2}x_{s}}{dt^{2}}+v_{s}\frac{\partial f}{\partial
r_{s}}\frac{\partial r_{s}}{\partial t}+v_{s}^{2}f\frac{\partial f}{\partial
r_{s}}\frac{\partial r_{s}}{\partial x}=h(t).$ (7.3)
The partial derivative of Eq. (2.8) may be written as follows,
$\frac{\partial f}{\partial r_{s}}=\frac{\sigma}{2\tanh{(\sigma
R)}}\left\\{\operatorname{sech}^{2}[\sigma(r_{s}+R)]-\operatorname{sech}^{2}[\sigma(r_{s}-R)]\right\\}.$
(7.4)
The partial derivatives of $r_{s}$ yield,
$\displaystyle\frac{\partial r_{s}}{\partial t}$
$\displaystyle=\pm\frac{dx_{s}}{dt}=\pm v_{s}(t),$ (7.5)
$\displaystyle\frac{\partial r_{s}}{\partial x}$ $\displaystyle=\pm 1,$ (7.6)
and remembering that in the dust case $\beta=\beta(x,t)$, then $r_{s}(t)$ is
given by (see Eq. 2.9),
$r_{s}(t)=\sqrt{\left[x-x_{s}(t)\right]^{2}}=\lvert x-x_{s}(t)\rvert.$ (7.7)
Considering the expressions above, Eq. (7.3) may be rewritten as follows,
$f\frac{d^{2}x_{s}}{dt^{2}}\pm v_{s}^{2}\frac{\sigma F}{2\tanh{(\sigma R)}}\pm
v_{s}^{2}f\frac{\sigma F}{2\tanh{(\sigma R)}}=h(t),$ (7.8)
where
$F(r_{s})\equiv\operatorname{sech}^{2}[\sigma(r_{s}+R)]-\operatorname{sech}^{2}[\sigma(r_{s}-R)].$
(7.9)
Remembering that the regulating function $f[r_{s}(t)]$, as defined by Eq.
(2.8), can be approximated by a top hat function when $\sigma\gg R$ (see Ref.
nos , §2.1), in this limit Eq. (7.8) takes the following form inside the
bubble,
$\frac{d^{2}x_{s}}{dt^{2}}=h(t),$ (7.10)
where $f=1$. Outside the bubble $f=0$ and then $h(t)=0$, which means plane
shock waves described by the inviscid Burgers equation nos .
Hence, the nonuniqueness of the shift vector $\beta=v_{s}(r_{s})f[r_{s}(t)]$
arises from the fact that not only the function $h(t)$ is arbitrary, but also
because the regulating function $f[r_{s}(t)]$ only requires a top hat behavior
with null values outside the bubble. So, any well behaved function that
respects such constraints may be part of a solution of the Burgers equation.
Moreover, from the energy conditions calculated for the PPF, as summarized in
Table 3, one can see that $\beta$ plays a fundamental role in Cases 1a and 2a
for the null and dominant energy conditions to be satisfied, which adds
further physical constraints to its behavior. So, it is clear that the
nonlinearity of the Einstein equations imply that the generic functions
appearing in their integration become entangled with the regulating function
in a non trivial manner.
To extend this analysis to the perfect fluid and PPF contents require further
constraints on the shift vector, something which at this stage would be done
in an entirely arbitrary manner. Perhaps in the future that can be done under
more physically plausible reasoning.
### 7.2 Anisotropic Fluids
The PPF proposed in Section 4 aimed at offering an alternative EMT for solving
the Einstein equations endowed with the warp drive metric. As we shall see
below, The PPF can actually be seen as an anisotropic fluid with heat flux
CarlEckart1940 .
In general relativity the energy momentum tensor $T_{\mu\nu}$ represents the
source of energy and momentum, where $T_{00}$ is the flow of energy across a
surface of constant time (energy density), $T_{0i}$ is the energy flux across
a surface in the $i$ direction (constant $x^{i}$), $T_{i0}$ is the momentum
density and $T_{ij}$ is the momentum flux. If we choose a comoving frame of
reference that moves with the same velocity as the fluid this means that
particles in this fluid will have zero velocity and the flux of energy will be
only through the flux of heat, and the momentum flux will be via some sort of
dissipative phenomena such as viscosity, thermal radiation or even some sort
of electromagnetic type of radiation.
The general stress-energy tensor of a relativistic fluid can be written in the
form below CarlEckart1940 ; Pimentel2016 ,
$T^{\alpha\beta}=\mu
u^{\alpha}u^{\beta}+ph^{\alpha\beta}+u^{\alpha}q^{\beta}+u^{\beta}q^{\alpha}+\pi^{\alpha\beta},$
(7.11)
where
$h_{\alpha\beta}=g_{\alpha\beta}+u_{a}\,u_{b}\,,$ (7.12)
projects tensors onto hypersurfaces orthogonal to $u^{\alpha}$, $\mu$ is the
matter density, $p$ is the fluid static pressure, $q^{\alpha}$ is the heat
flux vector and $\pi^{\alpha\beta}$ is the viscous shear tensor. The world
lines of the fluid elements are the integral curves of the four-velocity
vector $u^{\alpha}$. The heat flux vector and viscous shear tensor are
transverse to the world lines, that is,
$q_{a}\,u^{a}=0,\;\;\pi_{ab}\,u^{b}=0.$ (7.13)
In terms of coordinates we can write the energy momentum tensor for a general
fluid as
$T_{\alpha\beta}=\begin{pmatrix}\varepsilon&q_{a}\\\
q_{b}&\pi_{ab}\end{pmatrix},$ (7.14)
where $q_{a}$ is the three-vector heat flux vector, $\varepsilon$ is a scalar
function and $\pi_{ab}$ is a three by three matrix viscous stress tensor,
which is symmetric and traceless. Both $q_{a}$ and $\pi_{ab}$ have
respectively three and five linearly independent components. Together with the
density $\mu$ and the static pressure $p$, this makes a total of ten linearly
independent components which is the number of linearly independent components
in a four-dimensional symmetric rank two tensor. We noticed that the Einstein
tensor components are highly non-linear for the warp drive metric and the off
diagonal terms require those free parameters for a non over determined
solution. For a non curved metric, i.e., the Minkowski metric
$\eta_{\alpha\beta}$, the energy momentum tensor for a perfect fluid with
anisotropic pressures can be written as
$T_{\alpha\beta}=\begin{pmatrix}\mu&0&0&0\\\ 0&p_{x}&0&0\\\ 0&0&p_{y}&0\\\
0&0&0&p_{z}\end{pmatrix}.$ (7.15)
Isotropic static pressure means that $p_{x}=p_{y}=p_{z}=p$. The perfect fluid
has no heat flux or dissipative phenomena, then
$(q^{\alpha}=0,\pi^{\alpha\beta}=0)$. This special case with dust content is
the well-known EMT, that is,
$T^{\alpha\beta}=\mu
u^{\alpha}u^{\beta}+ph^{\alpha\beta}=(\mu+p)u^{\alpha}u^{\beta}+pg^{\alpha\beta}\,.$
(7.16)
The PPF proposed in Section 4 has the matrix form given by Eq. (4.1), which
may be broken down as the sum of a perfect fluid in a warp drive background as
given by Eq. (3.2), and a dissipative fluid with the heat flux four-vector
given by, $q_{\alpha}=-\frac{1}{2}(q_{0},q_{1},0,0)$. Hence,
$q_{\alpha}u_{\beta}+q_{\beta}u_{\alpha}=\begin{pmatrix}q_{0}&q_{1}&0&0\\\
q_{1}&0&0&0\\\ 0&0&0&0\\\ 0&0&0&0\end{pmatrix},$ (7.17)
since the four-velocity for the moving frame is $u_{\alpha}=(-1,0,0,0)$. The
isotropic term is given by,
$\pi_{\alpha\beta}=\begin{pmatrix}\pi_{00}&0&0&0\\\ 0&\pi_{01}&0&0\\\
0&0&\pi_{02}&0\\\ 0&0&0&\pi_{03}\end{pmatrix},$ (7.18)
and
$\mu+\beta^{2}p\to(\mu+\pi_{00}+q_{0})+\beta^{2}p\,.$ (7.19)
So, the four parameters of the PPF are as follows,
$A=\pi_{01}+p\,,$ (7.20) $B=\pi_{02}+p\,,$ (7.21) $C=\pi_{03}+p\,,$ (7.22)
$D=p-\frac{q_{1}}{\beta}.$ (7.23)
The tensor $\pi_{\alpha\beta}$ must be traceless, giving us one more equation
to solve for the free parameters (anisotropic pressures)
$\pi_{00}+\pi_{01}+\pi_{02}+\pi_{03}=0$ (7.24)
From the above one can see that the warp drive metric endowed with the PPF
allows the study of fluid anisotropy coupled with possible dissipative effects
that could lead to a warp drive bubble. Perfect fluids are well known to be
part of solutions for the Einstein equations, this being the case of the
standard FLRW cosmological model that accounts for the expansion, isotropy and
spatial homogeneity of the universe. On the other hand, anisotropic imperfect
fluids offer a more complex source of gravitational effects, presenting
dissipative processes, shear and bulk viscous pressures, interaction between
fluids, radiation processes, electromagnetic interaction and even collision
between particles, charged or not. These fluids are even known to avoid the
big bang singularity in cosmological models murphy ; klimek ; msc ; p1 ; p2 .
MacCallum MacCallum1979 discussed various ways of generating anisotropy such
as the presence of electromagnetic fields, the presence of viscous terms and
the anisotropic stresses due to the anisotropic expansion of a cloud of
collisionless particles. Another way to account for viscosity, heat and energy
flux is the interaction of two or more fluids Letelier1980 ; Bayin1982 .
### 7.3 Other aspects of warp drive physics
The points presented above regarding the physical feasibility of superluminal
travel with the Alcubierre spacetime geometry by no means exhaust this
discussion. Several issues remain open, with each of them deserving separate
studies that are beyond the scope of this paper, since in here we focused on
the basic properties of the Einstein equations’ solutions of the warp drive
metric with fluid content. With respect to the open issues, one can point out
the amount of mass-energy density, exotic or not, necessary for the
feasibility for the warp drive in the context of both the perfect fluid and
the PPF, as well as quantum effects and the question of stability or
instability in our solutions. These issues deserve further investigations and
are the subject of ongoing research.
## 8 Conclusions
In this work we have analyzed the Einstein equations for the Alcubierre warp
drive metric having as gravity source two types of energy-momentum tensors
(EMT) for fluid, namely the perfect fluid and the parametrized perfect fluid
(PPF). The latter is defined by allowing the EMT pressure components of the
perfect fluid to be different from one another and dependent on all
coordinates.
After obtaining the components of the Einstein tensor for the warp drive
metric we calculated the dynamic equations for both fluids by solving the
respective Einstein equations. Solutions were found in the form of various
equations of state, and, by further imposing the null divergence for the EMTs,
new constraints were also found for the various variables. The weak, dominant,
strong and null energy conditions were also calculated, which implied further
constraints upon the free quantities for the EMTs. For the perfect fluid these
were on the matter density $\mu$ and pressure $p$. For the PPF that occurred
on the pressure components $p,A,B,C$, the matter density for the fluid $\mu$
and the momentum component $D$.
We found two main groups of solutions subcases possessing different conditions
for each EMT. For the perfect fluid, one solution may be interpreted as
requiring that the warp bubble can only be viable with negative matter
density. The alternative interpretation is that the warp bubble is possible
with positive matter density, but in this case the regulating function
$f(r,s)$, which shapes the warp bubble, becomes a complex function. This comes
from results allowing the possibility that the function
$\beta=v_{s}(t)f(r_{s})$ may have complex solutions once the matter density is
positive. Other results in both fluids reinforce our earlier finding in Ref.
nos that the warp bubble necessary for generating superluminal velocities, or
warp speeds, can be interpreted as a shock wave from classical fluid dynamics
theory.
We named four cases arising from the solutions of the Einstein equations: 1a,
1b, 2a and 2b. However, the Cases 1a and 2a are very similar, or equal, to
each other, the same happening to Cases 1b and 2b. For this reason they were
grouped together in the tables that summarize all results.
Specifically, Cases 1b and 2b for the perfect fluid reduced the solutions to
the ones found for dust content already studied in Ref. nos , that is, a
vacuum solution unable to create a warp bubble, but which connects the warp
metric to the inviscid Burgers equation, also yielding $\beta$ as a function
of $t$ and $x$ coordinates. The null EMT divergence is satisfied and a
continuity equation was also found (Eq. 5.13).
Cases 1b and 2b for the PPF resulted in a equation of state of the form
$\mu=-\beta^{2}p$, coordinate dependency of the $\beta$ function became
$\beta=\beta(x,t)$ and a non-homogeneous Burgers equation (4.71) emerged.
However, the pressures were constrained in such a way that this PPF EMT
solution for the warp drive was dismissed as unphysical.
Cases 1a and 2a for the perfect fluid produced an equation of state relating
pressure and matter density given by $p=3\mu$. The $\beta$ function
dependencies became $\beta=\beta(y,t)$ in the former case and
$\beta=\beta(z,t)$ in the latter, and both produced a differential equation
for $\beta$ that either requires negative density for $\beta$ to be a real
valued function, or a positive matter density which then leads to a complex
solution for $\beta$, which in turn leads to a complex regulating function
$f(r_{s})$ whose possible real part would then be related to a physically
viable warp bubble.
Cases 1a and 2a for the PPF resulted in an equation of state relating almost
all quantities, in the form $\mu=\beta^{2}(2D-A-p)+A/3$ which is valid for
both cases. Coordinate dependency on $\beta$ became, respectively, resulted in
$\beta=\beta(y,t)$ and $\beta=\beta(z,t)$. The function $\beta$ is also
governed by the first order differential equations (4.73) and (4.74),
respectively.
The null EMT divergence $T{{}^{\alpha\sigma}}_{;\sigma}=0$ were calculated,
producing further sets of very nonlinear differential equations constraining
all quantities in the PPF which could be used, in principle, to determine all
pressures and matter density in this fluid. For the perfect fluid, Cases 1a
and 2a are reduced to a continuity equation (5.13) including the function
$\beta$, which is interpreted as playing the role of the flow velocity vector
field. Cases 1b and 2b reduced the PPF to either the trivial condition of
Minkowski flat spacetime with no warp drive, or an EMT with all components
being zero, that is a vacuum case.
It has already been seen in Ref. nos that the Burgers equation appears
connected to the dust solution of the warp drive metric, which is in fact a
vacuum solution. Cases 1b and 2b of the perfect fluid became reduced to the
dust solution, and a non-homogeneous form of the Burgers equation appears in
these respective cases for the PPF, although the whole solutions in these
cases were dismissed as unphysical. The solutions that presented themselves as
the most plausible ones for creating warp speeds, 1a and 2a for both the
perfect fluid and PPF, do not generate a Burgers equation.
The weak, dominant, strong and null energy conditions were also studied in the
context of the perfect fluid and PPF energy-momentum tensors for a warp drive
metric. The resulting expressions were found to satisfy all conditions in the
perfect fluid EMT. Regarding the PPF, specific expressions constraining its
EMT quantities were obtained in order to satisfy these energy conditions, but
they do not necessarily lead to the conclusion that negative matter density is
always necessary for viable warp speeds, particularly because in the PPF the
pressure $A$ must assume values equal to zero or positive.
Summing up, the results of this paper indicate that warp speeds might be
physically viable in the context of positive matter density as some solutions
of the Einstein equations for both fluids keep open this possibility.
Nevertheless, such a situation creates the additional issue in the perfect
fluid context concerning the meaning of a possible complex regulating function
in the warp metric, a result that may be interpreted as a caveat, or major
stumbling block. Such difficulty does not appear to happen in the PPF
scenario, although this fluid was considered here mainly as a hypothetical
model whose aim was to investigate whether or not new possibilities arise in
the solutions of the Einstein field equations considering more complex energy-
momentum tensors having the Alcubierre warp drive metric. On this front it
seems then that the initial conclusions about the unphysical nature of warp
drive, or the impossibility of generating warp speeds, may not be not as
stringent as initially thought, or, perhaps, not valid at all.
## Acknowledgments
We are grateful to the referees for useful comments. E.M.C.A. thanks CNPq
(Conselho Nacional de Desenvolvimento Científico e Tecnológico), Brazilian
scientific support federal agency, for partial financial support, grants
numbers 406894/2018-3 and 302155/2015-5.
## References
* (1) M. Alcubierre, The warp drive: hyper-fast travel within general relativity, Class. Quantum Grav. 11 (1994) L73, arXiv:gr-qc/0009013.
* (2) L.H. Ford and T.A. Roman, Quantum Field Theory Constrains Traversable Wormhole Geometries, Phys. Rev. D 53 (1996) 5496, arXiv:gr-qc/9510071.
* (3) M.J. Pfenning and L.H. Ford, The unphysical nature of Warp Drive, Class. Quantum Grav. 14 (1997) 1743, arXiv:gr-qc/9702026.
* (4) W.A. Hiscock, Quantum effects in the Alcubierre warp-drive spacetime, Class. Q. Grav. 14 (1997) L183.
* (5) S.V. Krasnikov, Hyperfast Interstellar Travel in General Relativity, Phys. Rev. D 57 (1998) 4760, arXiv:gr-qc/9511068.
* (6) A. Everett and T.A. Roman, A Superluminal Subway: The Krasnikov Tube, Phys. Rev. D 56 (1997) 2100, arXiv:gr-qc/9702049.
* (7) F.S.N. Lobo and P. Crawford, Weak Energy Condition Violation and Superluminal Travel. In: L. Fernández-Jambrina and L.M. González-Romero (eds) Current Trends in Relativistic Astrophysics. Lecture Notes in Physics, vol. 617, (2003). Springer, Berlin, Heidelberg, arXiv:gr-qc/0204038.
* (8) F.S.N. Lobo and P. Crawford, Weak Energy Condition Violation and Superluminal Travel, Lect. Notes Phys. 617 (2003) 277, arXiv:gr-qc/0204038.
* (9) C. Van Den Broeck, A warp drive with more reasonable total energy, Class. Quantum Grav. 16 (1999) 3973, arXiv:gr-qc/9905084.
* (10) J. Natario, Warp Drive With Zero Expansion, Class. Quantum Grav. 19 (2002) 1157, arXiv: gr-qc/0110086.
* (11) F.S.N. Lobo and M. Visser, Fundamental limitations on ‘warp drive’ spacetimes, Class. Quantum Grav. 21 (2004) 5871, arXiv:gr-qc/0406083.
* (12) F.S.N. Lobo and M. Visser, Linearized warp drive and the energy conditions, arXiv:gr-qc/0412065.
* (13) H.G. White, A Discussion of Space-Time Metric Engineering, Gen. Relat. Grav. 35 (2003) 2025.
* (14) H.G. White, Warp Field Mechanics 101, J. Brit. Interplanetary Society 66 (2011) 242.
* (15) J. Lee and G. Cleaver, Effects of External Radiation on an Alcubierre Warp Bubble, Physics Essays 29 (2016) 201
* (16) J. Lee and G. Cleaver, The Inability of the White-Juday Warp Field Interferometer to Spectrally Resolve Spacetime Distortions, Int. J. Modern Phys.: Advances in Theory and Application 2 (2017) 35; arXiv:1407.7772
* (17) B. Mattingly, A. Kar, M. Gorban, W. Julius, C. Watson, M.D. Ali, A. Baas, C. Elmore, J. Lee, B. Shakerin, E. Davis and G. Cleaver, Curvature Invariants for the Accelerating Natario Warp Drive, Particles, 3 (2020) 642, arXiv:2008.03366
* (18) B. Mattingly, A. Kar, M. Gorban, W. Julius, C.K. Watson, M. Ali, A. Baas, C. Elmore, J.S. Lee, B. Shakerin, E.W. Davis and G.B. Cleaver, Curvature Invariants for the Alcubierre and Natário Warp Drives, Universe, 7 (2021) 21, arXiv:2010.13693v2
* (19) O.L. Santos-Pereira, E.M.C. Abreu and M.B. Ribeiro, Dust content content solutions for the Alcubierre warp drive spacetime. Eur. Phys. J. C, 80 (2020) 786, arXiv:2008.06560
* (20) B. Schutz, A First Course in General Relativity 2nd Edition, Cambridge University Press, 2009.
* (21) M. Alcubierre, Introduction to 3+1 Numerical Relativity (International Series of Monographs on Physics). Oxford University Press, 2012.
* (22) B.S. DeWitt, in General Relativity: An Einstein Centenary Survey, edited by Hawking, S.W. and Israel, W., Cambridge University Press, Cambridge, 1980.
* (23) S.W. Hawking and G.F.R. Ellis, The Large Scale Structure of Space-Time, Cambridge University Press, 1973.
* (24) L.N. Trefethen and K. Embree (editors), unpublished, 2001; https://people.maths.ox.ac.uk/trefethen/pdectb.html (accessed 14 October 2018).
* (25) L.C. Evans, Partial Differential Equations, Graduate Studies in Mathematics, vol. 19, pp. 749, American Mathematical Society, 2010.
* (26) A.R. Forsyth, Theory of differential equations. Cambridge, University Press, 1906.
* (27) H. Bateman, Some recent researches on the motion of fluids, Mont. Weat. Rev. 43 (1915) 163.
* (28) J.M. Burgers, A mathematical model illustrating the theory of turbulence. Adv. Appl. Mech. 1 (1948) 171.
* (29) T. Ceylan and B. Okutmuştur, Finite volume approximation of the relativistic Burgers equation on a Schwarzschild(anti-)de Sitter spacetime, Turkish J. Math. 41 (2017) 1027; The relativistic Burgers equation on a de Sitter spacetime. Derivation and finite volume approximation, Int. J. Pure. Math. 2 (2015) 20; P.G. LeFloch, H. Makhlof and B. Okutmuştur, Relativistic Burgers equation on curved spacetime. Derivation and finite volume approximation, arXiv: 1206.3018.
* (30) G. Ellis, H. van Elst, Cargese Lectures 1998: Cosmological Models. NATO Adv. Study Inst. Ser. C. Math. Phys. Sci. 541 (1999) 1–116, arXiv:gr-qc/9812046.
* (31) C. Eckart, The Thermodynamics of Irreversible Processes. III. Relativistic Theory of the Simple Fluid, Phys. Rev. 58 (1940), 919.
* (32) O.M. Pimentel, F.D. Lora-Clavijo and G.A. Gonzalez, The energy-momentum tensor for a dissipative fluid in general relativity. Gen. Relat. Grav. 48 (2016) 124.
* (33) G.L. Murphy, Big-Bang Model without Singularities, Phys. Rev. D 8 (1973) 4231.
* (34) Z. Klimek, Some Thermodynamic Features of Cosmological Models with Viscosity, Acta Cosmologica 2 (1974) 49.
* (35) M.B. Ribeiro, Cosmological Models with Dissipative Fluids, Master’s thesis, Observatório Nacional, Rio de Janeiro, 1987 (in Portuguese). doi:10.13140/RG.2.1.1184.0724
* (36) M.B. Ribeiro and A.K. Sanyal, Bianchi VI0 Viscous Fluid Cosmology with Magnetic Field, J. Math. Phys. 28 (1987) 657.
* (37) A. Banerjee, M.B. Ribeiro and N.O. Santos, Anisotropic Viscous Fluid Cosmological Model, Astrophys. Space Sci. 136 (1987) 331.
* (38) M.A.H. MacCallum, in: General Relativity: An Einstein Centenary Survey, ed. S.W. Hawking and W. Israel, Cambridge University Press, 1979.
* (39) P.S. Letelier, Anisotropic fluids with two-perfect-fluid components, Phys. Rev. D 22 (1980) 807.
* (40) S.S. Bayin, Anisotropic fluid spheres in general relativity, Phys. Rev. D 26 (1982) 1262.
|
# VOTE400(Voide Of The Elderly 400 Hours): A Speech Dataset to Study Voice
Interface for Elderly-Care
Minsu Jang1, Sangwon Seo2, Dohyung Kim1, Jaeyeon Lee1 and Jaehong Kim1 and
Jun-Hwan Ahn2 1Minsu Jang, Dohyung Kim, Jaeyeon Lee and Jaehong Kim are with
Electronics and Telecommunications Research Institute, Daejeon-si, South Korea
minsu at etri.re.kr2Sangwon Seo and Jun-Hwan Ahn is with MINDs Lab Inc.,
Kyungki-do, South Korea asdn9353 at mindslab.ai
###### Abstract
This paper introduces a large-scale Korean speech dataset, called VOTE400,
that can be used for analyzing and recognizing voices of the elderly people.
The dataset includes about 300 hours of continuous dialog speech and 100 hours
of read speech, both recorded by the elderly people aged 65 years or over. A
preliminary experiment showed that speech recognition system trained with
VOTE400 can outperform conventional systems in speech recognition of elderly
people’s voice. This work is a multi-organizational effort led by ETRI and
MINDs Lab Inc. for the purpose of advancing the speech recognition performance
of the elderly-care robots.
## I INTRODUCTION
Voice interface is the most intuitive, comfortable and universal interface for
interacting with service robots. Recent advancement of commercial cloud-based
speech-to-text (STT) services allowed devising a voice interface for service
robots a very simple process of integrating a service API for STT into the
robot SW system.
While these commercial systems work very well with adults in the ages of
between 20 and 60, it easily fails with voices from older adults aged 65 years
or over. It is known that speech signals from older adults bring about
difficulties for automated speech recognition as they tend to be imprecise in
consonant pronunciation, include tremors, and have slower articulations [1].
In the need to develop a speech recognition system that are specialized to the
speech signals from older adults, we built a speech dataset by collecting
large-scale dialogue and read speech from older adults. The result of our
effort is 400 hours of Korean speech data which we named as ’VOTE400 (Voice Of
The Elderly 400 Hours) and open-sourced for any non-commercial research
projects (https://ai4robot.github.io/mindslab-etri-vote400).
## II Dataset Description
### II-A Dataset Collection
For recruiting and collecting voice data from older adults, we got assistance
from a Korean governmental office, called Dok-Geo-No-In-Jong-Haap-Ji-Won
Center (Ji-Won Center: http://www.1661-2129.or.kr/index.html) devoted to the
support of older adults living alone. With the support from the Ji-Won Center,
we could collect a large-scale dialog speech and read speech from a number of
older adults across various regions of South Korea.
#### II-A1 Dialog Speech
To collect spontaneous speech data from older adults, we could utilize a
support program of Ji-Won Center called Saa-Raang-It-Gi where social workers
regularly visit elderly people’s homes for consulting on health-related issues
and relieving loneliness. After explaining about the data collection
experiment and getting consent of participation from the elderly,
conversations between a social worker and an elderly were recorded using a
smartphone.
The recordings from these program sessions were sent to Ji-Won Center and a
screening process was performed to remove every dialogue involving sensitive
personal information. Then, a quality assurance process was followed to filter
out speech segments incomprehensible by human listener due to imprecise
pronunciation or significant noise.
#### II-A2 Read Speech
To amend the relatively low-quality of the dialog speech dataset, we launched
another data collection process to acquire read speech from older adults. We
built and utilized in the process a dedicated speech collection system, where
a tablet-based client program presents a sentence to read to an elderly user;
makes a recording and sends it to a server; where the recording is inspected
to be accepted or not. In total, the number of unique sentences chosen to be
read by participants was 2,250. These sentences were selected by considering
how often these could be casually uttered by older adults in daily lives.
TABLE I: Raw Data Collection of Dialog Speech Region(R) | No. Participants | Len. (hrs)
---|---|---
Seoul-si(SE) | 620 | 122
Busan-si(PS) | 242 | 90
Daegu-si(DG) | 202 | 33
Gwangju-si(GJ) | 179 | 63
Daejeon-si(DJ) | 275 | 66
Ulsan-si(WS) | 80 | 28
Goyang-si(GG) | 335 | 69
Gangwon-do(GW) | 178 | 45
Chungcheongbuk-do(CB) | 252 | 92
Chungcheongnam-do(CN) | 317 | 46
Jeollanam-do(JN) | 323 | 103
Gyeongsangbuk-do(GB) | 378 | 116
Total | 3,381 | 873
### II-B Dialog Speech Data
The total number of elderly participants is 3,381 and the total length of
recordings is 873 hours. This is the result of collective efforts by regional
senior citizens welfare institutes collaborating with the Ji-Won Center. Table
I shows the regional distributions of all the participants and the length of
recordings per region.
After the screening and the QA process mentioned in the previous subsection,
we finalized 300 hours of dialog speech to be included in the VOTE400 dataset.
Transcription for every final speech data was done by human annotators. In
VOTE400, we provide for a recording session in a WAV file. The audio format of
the WAV file is as shown in table II. Every WAV file is accompanied by a
transcription text file encoded in ISO-8859. The transcription does not
include audio-text alignment information.
TABLE II: VOTE400 Dialog Speech Audio Format Property | Value
---|---
Format. | PCM
Format Settings | Little/Signed
Codec ID | 1
Bit Rate Mode | Constant
Bit Rate. | 256
Channel(s) | 1
Sampling Rate | 16 kHz
Bit Depth | 16 bits
The file name of each recording follows the pattern of
<P-ID>_<G>_<A>_<R>_<DT>, where P-ID is a unique participant ID; G is a gender
value (F for female, M for male), A is a age value, R is a regional code, and
DT is the data-time of the recording session.
Participants and speech audio statistics for VOTE400 dialog dataset are shown
in table III and table IV.
TABLE III: Demographics of VOTE400 Dialog Speech Region(R) | No. Participants | Age ($\mu/\sigma$)
---|---|---
Seoul-si(SE) | 251(F:210,M:41) | 78.98/5.13
Daegu-si(DG) | 108(F:95,M:13) | 80.33/6.08
Gyoungki-do(GG) | 110(F:83,M:27) | 80.17/5.41
Chungcheongnam-do(CN) | 6(F:6,M:0) | 77.00/3.69
Jeollanam-do(JN) | 70(F:56,M:14) | 80.76/4.90
Busan-si(PS) | 160(F:137,M:23) | 78.70/5.51
Daejeon-si(DJ) | 96(F:72,M:24) | 78.81/5.24
Gangwon-do(GW) | 109(F:94,M:15) | 80.07/5.50
Gyeongsangbuk-do(GB) | 98(F:95,M:3) | 80.87/4.48
Gwangju-si(GJ) | 87(F:70,M:17) | 79.39/5.77
Chungcheongbuk-do(CB) | 17(F:17,M:0) | 80.47/5.51
Ulsan-si(WS) | 58(F:49,M:9) | 76.97/4.48
Total | 1,170(F:984,M:186) | 79.47/5.37
TABLE IV: Speech Audio Statistics for VOTE400 Dialog Speech Region(R) | Len.(secs) | Len.($\mu/\sigma$)
---|---|---
Seoul-si(SE) | 151,010 | 601.63/239.83
Daegu-si(DG) | 60,740 | 562.42/228.14
Gyoungki-do(GG) | 107,935 | 981.23/357.19
Chungcheongnam-do(CN) | 5,193 | 865.62/293.98
Jeollanam-do(JN) | 81,767 | 1,168.10/294.85
Busan-si(PS) | 200,207 | 1,251.30/255.85
Gangwon-do(GW) | 95,420 | 875.42/158.18
Daejeon-si(DJ) | 123,138 | 1,282.70/293.83
Gyeongsangbuk-do(GB) | 71,175 | 726.28/308.80
Gwangju-si(GJ) | 92,699 | 1,065.52/276.53
Chungcheongbuk-do(CB) | 20,135 | 1,184.41/309.54
Ulsan-si(WS) | 70,754 | 1,219.90/254.43
Total | 1,080,179 | 923.23/380.17
### II-C Read Speech Data
The total number of elderly participants is 104 and the total length of
recordings is 100 hours. Table VI shows the statistics of VOTE400 read speech
data.
Audio format of the VOTE400 read speech data is as shown in V, which is
slightly different from the format of the dialog speech data.
TABLE V: VOTE400 Read Speech Audio Format Property | Value
---|---
Format. | PCM
Format Settings | Little/Signed
Codec ID | 1
Bit Rate Mode | Constant
Bit Rate. | 705.6 kb/s
Channel(s) | 1
Sampling Rate | 44.1 kHz
Bit Depth | 16 bits
TABLE VI: Regional Distributions of VOTE400 Read Dataset Region(G) | No. Persons | No. Sent. | Len.($\mu/\sigma$)
---|---|---|---
Gyeongsangnam-do(GB) | 20 | 22,575 | 3.18/1.38
Seoul-si(SE) | 18 | 19,220 | 3.31/1.49
Jeollanam-do(JN) | 21 | 21,393 | 3.36/1.52
Daegu-si(DG) | 25 | 26,950 | 3.60/1.87
Gangwon-do(GW) | 20 | 21,676 | 2.73/1.12
Total | 104 | 111,814 | 3.25/1.54
The file name of read speech data follows the pattern of
PID_<P-ID>_<DATE>_<SENTENCE-NO>_<R>, where P-ID is a unique participant ID,
DATA is the date of recording, SENTENCE-NO is a serial number put to each of
the recorded sentences, and R is the region code as shown in table VI. Each
WAV file contains a single sentence, accompanied by a transcription text file
encoded in EUC-KR.
Though the number of sentences chosen and presented to the participants was
originally 2,250, the final total number of unique sentences in VOTE400 read
speech data is 7,832, due to mistakes and slight variations in real utterances
by older adults.
## III Preliminary Experiment
We conducted a preliminary experiment by training a MINDs Lab Inc.’s
proprietary baseline speech recognizer(M), which is based on LSTM
architecture, and estimating the STT accuracy using VOTE400. After fine-tuning
the baseline with 50 hours each of dialog speech data and read speech data of
VOTE400, a simple test with 100 sentences from different regions was performed
and the results are as shown in table VII, along with the results when the
sentences were tested on a commercial cloud-based STT engine(C).
TABLE VII: STT Performance Test Results with VOTE400 Region(G) | Gender | Acc. M (%) | Acc. C (%)
---|---|---|---
SE | M | 90 | 90
SE | F | 90 | 80
GW | M | 80 | 90
GW | F | 90 | 80
DG | M | 70 | 80
DG | F | 90 | 80
GN | M | 90 | 80
GN | F | 80 | 80
JN | M | 70 | 50
JN | F | 80 | 60
## IV Summary
We described a Korean speech dataset VOTE400 which is collected entirely from
older adults of more than 75 years old. VOTE400 contains 300 hours of dialogue
speech data and 100 hours of read speech data, with proficient varieties in
gender and regions. To our knowledge, VOTE400 is by far one of the largest
voice datasets that is oriented to voices of the elderly. We hope that this
dataset will be useful to study older adult’s voice features and realize voice
technologies that work sufficiently well in elderly-care robotics.
## ACKNOWLEDGMENT
This work was supported by the Institute of Information communi-cations
Technology Planning Evaluation(IITP) grant funded by the Koreagovernment(MSIT)
(No. 2017-0-00162, Development of Human-care RobotTechnology for Aging
Society)
## References
* [1] Vacher, M., Aman, F., Rossato, S. and Portet, F., 2015, August. Development of automatic speech recognition techniques for elderly home support: Applications and challenges. In International Conference on Human Aspects of IT for the Aged Population (pp. 341-353). Springer, Cham.
|
# Improved Wall-Normal Derivative Formulae for Anisotropic Adaptive Simplex-
Element Grids
Hiroaki Nishikawa
National Institute of Aerospace, Hampton, VA 23666, USA Research Fellow
(hiro@nianet.org), 100 Exploration Way, Hampton, VA 23666 USA, Associate
Fellow AIAA
###### Abstract
In this paper, we explore methods for computing wall-normal derivatives used
for calculating wall skin friction and heat transfer over a solid wall in
unstructured simplex-element (triangular/tetrahedral) grids generated by
anisotropic grid adaptation. Simplex-element grids are considered as efficient
and suitable for automatic grid generation and adaptation, but present a
challenge to accurately predict wall-normal derivatives. For example, wall-
normal derivatives computed by a simple finite-difference approximation, as
typically done in practical fluid-dynamics simulation codes, are often
contaminated with numerical noise. To address this issue, we propose an
improved method based on a common step-length for the finite-difference
approximation, which is otherwise random due to grid irregularity and thus
expected to smooth the wall-normal derivative distribution over a boundary.
Also, we consider using least-squares gradients to compute the wall-normal
derivatives and discuss their possible improvements. Numerical results show
that the improved methods greatly reduce the noise in the wall-normal
derivatives for irregular simplex-element grids.
## 1 Introduction
Towards fully automated computational fluid dynamics (CFD) simulations, there
is an increasing interest in developing robust and accurate discretizations
and solvers for unstructured simplex-element grids with anisotropic grid
adaptation
[WhiteNishikawaBaurle_scitech2020,Kleb_etal_aiaa2019-2948,Nishikawa_scitech2020,nishikawa_centroid:JCP2020,uns_grid_adaptation_aiaa2018-1103,alauzet-
loseille-decade-aniso-adapt-cfd]. While fully-automated CFD simulations have
recently been successfully demonstrated for practical three-dimensional
problems [Kleb_etal_aiaa2019-2948], improvements are still desired in the
discretization method especially in terms of accuracy. For example,
computations of derivatives such as the viscous/heat fluxes and the vorticity
are known to suffer from numerical noise on highly irregular and skewed
simplex-element grids [liu_nishikawa_aiaa2016-3969]. Impact of numerical noise
may be reduced in integrated quantities (e.g., drag coefficient), but when
accuracy is required for a pointwise quantity, e.g., heating rate at a
stagnation point, numerical noise can be a serious problem, causing a large
error and a great degree of uncertainty in the prediction. In order to develop
a reliable automated CFD solver, it is necessary to be able to accurately
predict pointwise derivatives, especially at wall boundaries, on irregular
anisotropic simplex-element grids.
There have been efforts to improve derivative accuracy on unstructured grids:
derivative variables introduced in target equations
[nishikawa_hyperbolic_poisson:jcp2020,liu_nishikawa_aiaa2016-3969], improved
least-squares methods
[NovelGradStencil:CF2018,SozerChristophCetin:AIAA2014,ShimaKitamuraHaga_AIAAJ2013],
and implicit gradient methods [WangRenPanLi:JCP2017,nishikawa_igg:JCP2018].
All these methods are promising for improving derivative accuracy, but they
are not specifically designed to accurately estimate derivatives at a boundary
and thus further improvements may be achieved by developing a more specific
technique. In this work, we focus on wall-normal derivatives and explore
various methods for accurately estimating the wall-normal derivatives relevant
to the viscous stresses and heat fluxes at a wall, as a post-processing step
for a given numerical solution on adaptive simplex-element grids. A popular
technique used in practical CFD codes for computing the wall-normal
derivatives is a one-sided finite-difference formula applied in the wall-
normal direction with the numerical solution in a nearby node/cell and a
boundary condition value. For example, it provides an approximation to
$\partial u/\partial n$ that dominates the viscous stress at a wall, where $u$
is a wall-parallel velocity components and $n$ is the coordinate normal to the
wall. It can provide an accurate approximation on structured grids, but will
introduce numerical noise on irregular simplex-element grids
[WhiteNishikawaBaurle_scitech2020]. We conjecture that the numerical noise in
the derivative is due to the irregularity of the spacing used in the finite-
difference formula. That is, the distance between the center of a cell
adjacent to a boundary and a wall-boundary face centroid varies along a
boundary in adaptive simplex-element grids. Then, we propose to keep the same
distance for the finite-difference formula applied at all wall-boundary faces.
Numerical experiments show that the modified finite-difference method greatly
improves the skin friction distribution over a flat plate in a laminar flow,
where the distance is determined as $\eta=0.5$ at all boundary faces, where
$\eta$ is a normalized coordinate in the Blasius boundary-layer solution. In
order to extend this technique to more general flows over complex geometries,
it is necessary to develop a method for estimating a universal distance for
the finite-difference formula.
Another post-processing technique can be devised based on gradients computed
by a least-squares method at nodes or cells. Following a recent study
[WhiteNishikawaBaurle_scitech2020], where heating rate prediction over a flat
plate in a hypersonic turbulent flow has been shown to be greatly improved by
the face-averaged nodal-gradient (F-ANG) approach applied to the finite-volume
discretization [WhiteNishikawaBaurle_scitech2020,NishikawaWhite_FANG:jcp2020],
we employ the F-ANG approach in this study, but only consider the post-
processing technique for estimating the wall-normal derivatives for a given
numerical solution. Discussions on the impact of the gradient approach (used
in the discretization) on wall heating prediciton are given in a companion
paper [NishikawaWhite_scitech2021-xxxx]. In the F-ANG method, gradients are
computed and stored at nodes while numerical solutions are stored at cells.
Therefore, the wall-normal derivatives are already available at wall nodes.
However, these derivatives can be noisy for irregular anisotropic simplex-
element grids. A simple method for reducing the noise would be to average the
nodal gradients, and there are two possible techniques: face and cell
averages. We compare these averaged gradients for estimating the wall-normal
derivatives. Numerical results indicate that the cell-averaged nodal-gradients
can greatly reduce the noise in the wall-normal derivatives, but not as much
as the improved finite-difference formula.
## 2 Finite-Volume Discretization
We consider a cell-centered finite-volume discretization for the Navier-Stokes
equations integrated over a computational cell $j$ (see Figure 1) with the
midpoint rule:
$\displaystyle{\bf Res}_{j}=\sum_{k\in\\{k_{j}\\}}{\bf\Phi}_{jk}A_{jk},$ (2.1)
where $\\{k_{j}\\}$ is a set of neighbors of the cell $j$, $A_{jk}$ is the
length of the face across $j$ and $k$, and ${\bf\Phi}_{jk}$ is a numerical
flux. In this work, the Roe [Roe_JCP_1981] flux is used for the inviscid
terms, and the alpha-damping flux [nishikawa:AIAA2010] for the viscous terms.
The solution values are stored in the primitive variables ${\bf w}=(\rho,{\bf
v},T)$, at cells as point values at the centroid [nishikawa_centroid:JCP2020],
where $\rho$ is the density, ${\bf v}$ is the velocity vector, and $T$ is the
temperature. Second-order accuracy is achieved by the linearly-exact flux
quadrature (i.e., the midpoint rule) and the numerical flux evaluated with
linearly reconstructed solutions. In this work, we focus on the F-ANG approach
[WhiteNishikawaBaurle_scitech2020,NishikawaWhite_FANG:jcp2020], which has been
shown to yield significantly more accurate wall-derivatives than a
conventional finite-volume scheme, but still involves mild oscillations in the
derivatives on irregular simplex-element grids. In the F-ANG approach, we
compute the solution gradients at nodes using solution values stored at cells
by a linear least-squares method. Then, the nodal gradients are used in the
linear reconstruction, averaged over the nodes of a face:
$\displaystyle{\bf w}_{L}={\bf w}_{j}+\nabla{\bf w}_{f}\cdot\Delta{\bf
x}_{jm},\quad{\bf w}_{R}={\bf w}_{k}+\nabla{\bf w}_{f}\cdot\Delta{\bf
x}_{km},$ (2.2)
where $\nabla{\bf w}_{f}$ denotes the face-averaged gradient:
$\displaystyle\nabla{\bf w}_{f}=\frac{1}{2}\left(\nabla{\bf
w}_{jk}^{\ell}+\nabla{\bf w}_{jk}^{r}\right),$ (2.3)
where the superscripts $\ell$ and $r$ denote the left and right nodes of the
face $[j,k]$ (see Figure 1). This approach has various advantages especially
for simplex-element grids: e.g., no interpartition is required for gradients,
the residual stencil is greatly reduced. See
[WhiteNishikawaBaurle_scitech2020,NishikawaWhite_FANG:jcp2020] for further
details.
Figure 1: Stencil for cell-centered finite-volume discretization.
## 3 Methods for Computing Wall-Normal Derivatives
In this section, we describe methods for computing a wall-normal derivative,
taking the velocity derivative $\partial u/\partial n$ as an example, where
$u$ is the wall-parallel component of the velocity and $n$ denotes the
coordinate normal to the wall. We assume that the value at a wall $u_{b}$ is
known, e.g., $u_{b}=0$ for a no-slip condition. The wall-normal derivative
will be computed at a boundary node or at a boundary face, depending on the
method.
Figure 2: Geometry of an irregular triangular grid over a boundary. Black
filled circles are boundary nodes; smaller circles are the centroids of
triangles adjacent to the boundary.
### 3.1 Nodal gradients
The wall-normal derivative is computed at a boundary node $i$ by projecting
the nodal gradient $\nabla u_{i}$ in the direction of wall normal (see Figure
2):
$\displaystyle\left.\frac{\partial u}{\partial n}\right|_{i}=\nabla
u_{i}\cdot\hat{\bf n}_{i},$ (3.1)
where $\hat{\bf n}_{i}$ denotes a unit vector at the node $i$ normal to the
wall pointing towards the interior domain. This is perhaps the simplest method
since the nodal gradients are already available in the F-ANG method.
### 3.2 Face-averaged nodal gradients
The wall-normal derivative is computed at a face $j$ with a face-averaged
gradient:
$\displaystyle\left.\frac{\partial u}{\partial
n}\right|_{j}=\overline{\nabla}u_{j}^{face}\cdot\hat{\bf n}_{j},$ (3.2)
where $\overline{\nabla}u_{j}^{face}$ is the average of the nodal gradients
over the boundary face $j$. In Figure 2, $\overline{\nabla}u_{j}^{face}$ is
obtained by averaging the nodal gradients at the left and right nodes of the
boundary face $j$.
### 3.3 Cell-averaged nodal gradients
The wall-normal derivative is computed at a face $j$ with a cell-averaged
gradient:
$\displaystyle\left.\frac{\partial u}{\partial
n}\right|_{j}=\overline{\nabla}u_{j}^{cell}\cdot\hat{\bf n}_{j},$ (3.3)
where $\overline{\nabla}u_{j}^{cell}$ is the average of the nodal gradients
over a cell adjacent to the face $j$. In Figure 2,
$\overline{\nabla}u_{j}^{cell}$ is obtained by averaging the nodal gradients
at the three nodes of the triangular cell $c$.
(a) Finite-difference I.
(b) Finite-difference II.
(c) Finite-difference III.
Figure 3: Geometries for the finite-difference wall-normal derivative
formulae.
### 3.4 Finite-difference I
The wall-normal derivative can be computed at a face $j$ by the following
finite-difference formula:
$\displaystyle\left.\frac{\partial u}{\partial
n}\right|_{j}=\frac{u_{c}-u_{b}}{({\bf x}_{c}-{\bf x}_{j})\cdot\hat{\bf
n}_{j}},$ (3.4)
where $u_{c}$ is a solution value stored at the cell $c$ adjacent to the
boundary face $j$, $u_{b}$ is a known boundary value (e.g., from a boundary
condition), ${\bf x}_{c}$ is the centroid position of the cell $c$, ${\bf
x}_{j}$ is the centroid of the boundary face $j$. See Figure 3(a). This
formula estimates the derivative at a point where the cell-center ${\bf
x}_{c}$ is projected on the boundary, not at ${\bf x}_{j}$.
### 3.5 Finite-difference II
The wall-normal derivative can be estimated at the face centroid by
$\displaystyle\left.\frac{\partial u}{\partial
n}\right|_{j}=\frac{u_{c^{\prime}}-u_{b}}{({\bf x}_{c}-{\bf
x}_{j})\cdot\hat{\bf n}_{j}},$ (3.5)
where $u_{c^{\prime}}$ is a solution extrapolated from the cell center ${\bf
x}_{c}$ to the point right above the face centroid (see Figure 3(b)):
$\displaystyle u_{c^{\prime}}=u_{c}+\nabla u_{c}\cdot({\bf
x}_{c^{\prime}}-{\bf x}_{c}),\quad{\bf x}_{c^{\prime}}={\bf x}_{j}+[({\bf
x}_{c}-{\bf x}_{j})\cdot\hat{\bf n}_{j}]\hat{\bf n}_{j}.$ (3.6)
This approximation is exact for linear functions at the face centroid.
### 3.6 Finite-difference III
In the previous methods, the finite-difference formula is applied with varying
step-length as the distance between the cell center and the face center in the
normal direction changes randomly on irregular simplex-element grids. Hence,
the methods are strongly dependent on the grid geometry, which is not suitable
for adaptive simplex-element grids. To minimize the dependence on the grid
geometry, we generalize the previous formula as
$\displaystyle\left.\frac{\partial u}{\partial
n}\right|_{j}=\frac{u_{p}-u_{b}}{\overline{h}},$ (3.7)
where $u_{p}$ is a solution extrapolated from the cell center ${\bf x}_{c}$ to
the point at a certain distance $\overline{h}$ from the face center in the
wall normal direction (see Figure 3(c)):
$\displaystyle u_{p}=u_{c}+\nabla u_{c}\cdot({\bf x}_{p}-{\bf
x}_{c}),\quad{\bf x}_{p}={\bf x}_{j}+\overline{h}\,\hat{\bf n}_{j},$ (3.8)
where $\overline{h}$ is a global constant or varies along a wall but constant
in a transformed normal coordinate. This approximation is also exact for
linear functions at the face centroid.
This formula does not depend strongly on the cell centroids of the adjacent
cells. Whether in the physical or transformed coordinate, the above formula
keeps the same step-length of the finite-difference formula over a wall and
thus is expected to produce a smoother derivative (and its error)
distribution.
In this work, we investigate the effects of the use of a common step-length
$\overline{h}$ and explore estimation techniques of $\overline{h}$ for various
viscous-flow problems on adaptive grids. Here, we provide preliminary results
for a laminar flat-plate computation with the height $\overline{h}$ determined
as described below.
### 3.7 Finite-difference with face-area-weighted centroids
Finite-difference III requires a common length to be defined. For a laminar
flow over a flat plate, a normalized wall-normal distance, as we will describe
later, has been found to serve well for subsonic flows. However, such an
estimate is not applicable to high-speed flows and practical turbulent flows
over complex geometries. As a practical technique, we consider the face-area-
weighted centroid [nishikawa_centroid:JCP2020], which could reduce the
centroid-height variation along a boundary. The resulting method is similar to
the finite-difference II but ${\bf x}_{c^{\prime}}$ is replaced by the face-
area-weighted centroid:
$\displaystyle{\bf
x}_{c^{\prime}}=\frac{\displaystyle\sum_{k\in\\{k_{j}\\}}\hat{A}_{jk}^{p}{\bf
x}_{jk}}{\displaystyle\sum_{k\in\\{k_{j}\\}}\hat{A}_{jk}^{p}},\quad\hat{A}_{jk}=\frac{{A}_{jk}}{\displaystyle\max_{k\in\\{k_{j}\\}}{A}_{jk}},$
(3.9)
where${A}_{jk}$ is the area (length in two dimensions) of the face across $j$
and $k$, ${\bf x}_{jk}$ is the face centroid, and $p=2$.
## 4 Numerical Results
### 4.1 Two-dimensional laminar flat plate at $M_{\infty}=0.15$,
$Re_{\infty}=10^{6}$
For a laminar flat plate, we consider the normalized wall-normal coordinate
$\eta$:
$\displaystyle\eta=\frac{y}{x}\sqrt{Re_{x}},\quad
Re_{x}=\frac{\rho_{\infty}U_{\infty}x}{\mu_{\infty}},$ (4.1)
where $x$ is the distance from the leading edge, $\rho_{\infty}$,
$U_{\infty}$, $\mu_{\infty}$ are free stream density, flow speed, and
viscosity. For the finite-difference III, we fix the height to be $\eta=0.5$,
which is translated into the $y$-coordinate as
$\displaystyle\overline{h}=\eta\sqrt{\frac{x_{j}}{Re_{\infty}}},$ (4.2)
where $\eta=0.5$, $x_{j}$ is the $x$-coordinate of the face centroid, and
$Re_{\infty}$ is the free stream Reynolds number per unit length. Note that we
compute the wall-normal velocity derivative at boundary face centroids and
thus $x_{j}$ will not be zero. As can be seen from the formula, the height
$\overline{h}$ changes in the physical coordinate parabolically but remains
constant in the normalized wall-normal coordinate $\eta$. In the results, this
method is denoted by FD-III($\eta=0.5$). For a more general flow problem, this
approach may not be suitable. Therefore, we also consider using a fixed length
over the wall in the physical space:
$\displaystyle\overline{h}=\eta\sqrt{{Re_{\infty}}},$ (4.3)
which corresponds to using the estimated distance with $\eta=0.5$ at
$x_{j}=1.0$ everywhere. In this case, $\overline{h}$ is a global constant.
An adaptive grid was generated by the open-source tool refine
[park_darmofal:AIAA2008-917] developed by Mike Park at NASA Langley Research
Center and available at https://github.com/nasa/refine. Adaptation was
performed with the multiscale metric technique to control interpolation error
in L2-norm [ugawg-aiaa-verification] with Mach Hessian computed at nodes (by a
least-squares method) from the numerical solution obtained at cells with the
F-ANG finite-volume solver for the Navier-Stokes equations. The resulting
adaptive grid is shown in Figures 4(a) and 4(b). As can be seen, it is a fully
irregular grid towards the flat plate located at the bottom in $x\in[0,2]$.
Note that we already have a numerical solution on this grid as shown by Mach
contours in Figure 4(b). Then, the skin friction is computed at the flat plate
as
$\displaystyle C_{fx}=(\tau_{xx},\tau_{xy})\cdot\hat{\bf
n}\frac{2}{M_{\infty}^{2}}=\mu\left(\frac{2}{3}\frac{\partial u}{\partial
x}-\frac{1}{3}\frac{\partial v}{\partial y},\frac{\partial u}{\partial
y}+\frac{\partial v}{\partial x}\right)\cdot\hat{\bf
n}\frac{2}{M_{\infty}^{2}}=\mu\left(\frac{\partial u}{\partial
y}+\frac{\partial v}{\partial x}\right)\frac{2}{M_{\infty}^{2}},$ (4.4)
where $u$ and $v$ are the $x$\- and $y$-components of the velocity, the factor
$1/M_{\infty}^{2}$ comes from the nondimensionalization (i.e., the velocity is
nondimensionaliezd by the free stream speed of sound), and $\hat{\bf n}$ is
either $\hat{\bf n}_{i}$ or $\hat{\bf n}_{j}$, depending on where the
gradients are computed. The term $\partial v/\partial x$ is ignored in the
finite-difference methods, but retained in the LSQ gradient methods.
Results are shown in Figures 4(c)-4(i), where numerical skin-friction
distributions are compared with that of the Blasius solution. All numerical
skin-friction distributions are close to the theoretical distribution on
average, but significant differences are observed in the noise level. First,
the FD-I and FD-II methods produce noisy skin-friction distributions as shown
in Figures 4(c) and 4(d). On the other hand, as expected, the proposed methods
produce much smoother distributions as shown in Figures 4(e) and 4(f). Note
that the choice of $\overline{h}$ has no major impact on the skin-friction
distribution: Figure 4(e) for $\eta=0.5$, and Figure 4(f) for a global
constant corresponding to $\eta=0.5$ and $x=1$. These results indicate that it
is the uniformness of the step-length, whether in $\eta$ or $y$, that affects
the smoothness of the skin friction distribution. This is encouraging since
the method can be extended to more general problems by developing a method for
estimating a single value of the step-length $\overline{h}$ in a given
physical problem.
For the LSQ gradient methods, the direct use of the nodal gradients leads to a
noisy skin-friction distribution as shown in Figure 4(g). A slightly smoother
distribution is obtained with the face-averaged nodal gradients as shown in
Figure 4(h). The smoothest distribution is obtained with the cell-averaged
nodal gradients as shown in Figure 4(i). The results indicate that averaging
with more nodal gradients leads to a smoother distribution, but not as smooth
as those obtained with the finite-difference III. Note that the gradients are
linearly exact no mater how they are averaged, and the averaging operator acts
as a kind of filtering to remove high-frequency contents from the underlying
gradient approximation.
(a) Adaptive grid (3457 nodes).
(b) Zoomed-in view of the grid (with a scaled $y$-coordinate) and Mach
contours.
(c) FD-I.
(d) FD-II.
(e) FD-III ($\eta=0.5$).
(f) FD-III ($\eta=0.5$, $x=1$).
(g) Nodal gradients.
(h) Face-averaged gradients.
(i) Cell-averaged gradients.
Figure 4: Results for a laminar flow over a flat plate at $M_{\infty}=0.15$
and $Re_{\infty}=10^{6}$ on an adaptive triangular grid. The flat plate is
located at $y=0$ in $x\in[0,2]$. The skin friction coefficient $C_{fx}$ is
plotted at wall face centers, except for the figure (g).
## 5 Concluding Remarks
We have proposed improved wall-normal derivative formulae for adaptive
anisotropic simplex-element grids, based on the simple idea of unifying the
finite-difference distance over a wall. The improved formulae have been
demonstrated for a laminar flow over a flat plate to reduce the amount of
numerical noise in the skin friction distribution. In the future work, these
formulae will have to be extended to three dimensions and to turbulent flows.
## Acknowledgments
This work was supported by the Hypersonic Technology Project, through the
Hypersonic Airbreathing Propulsion Branch of the NASA Langley Research Center,
under Contract No. 80LARC17C0004. The author would like to thank Mike Park
(NASA Langley Geometry Laboratory) for his assistance with the use of refine.
## References
* [1] White, J., Nishikawa, H., and Baurle, R., “A 3-D Nodal-Averaged Gradient Approach for Unstructured-grid Cell-centered Finite-volume Methods for Application to Turbulent Hypersonic Flow,” SciTech 2020 Forum, AIAA Paper 2020-0652, Orlando, FL, 2020.
* [2] Kleb, W. L., Park, M. A., Wood, W. A., Bibb, K. L., Thompson, K. B., and Gomez, R. J., “Sketch-to-Solution: An Exploration of Viscous CFD with Automatic Grids,” 24th AIAA Computational Fluid Dynamics Conference, AIAA Paper 2019-2948, Dallas, TX, 2019.
* [3] Nishikawa, H., “A Face-Area-Weighted Centroid Formula for Reducing Grid Skewness and Improving Convergence of Edge-Based Solver on Highly-Skewed Simplex Grids,” SciTech 2020 Forum, AIAA Paper 2020-1786, Orlando, FL, 2020.
* [4] Nishikawa, H., “A Face-Area-Weighted Centroid Formula for Finite-Volume Method That Improves Skewness and Convergence on Triangular Grids,” J. Comput. Phys., Vol. 401, 2020, pp. 109001.
* [5] Park, M. A., Barral, N., Ibanez, D., Kamenetskiy, D. S., Krakos, J. A., Michal, T. R., and Loseille, A., “Unstructured Grid Adaptation and Solver Technology for Turbulent Flows,” 56th AIAA Aerospace Sciences Meeting, AIAA Paper 2018-1103, Kissimmee, Florida, 2018.
* [6] Alauzet, F. and Loseille, A., “A Decade of Progress on Anisotropic Mesh Adaptation for Computational Fluid Dynamics,” Computer-Aided Design, March 2015, pp. 13–39.
* [7] Liu, Y. and Nishikawa, H., “Third-Order Inviscid and Second-Order Hyperbolic Navier-Stokes Solvers for Three-Dimensional Inviscid and Viscous Flows,” 46th AIAA Fluid Dynamics Conference, AIAA Paper 2016-3969, Washington, D.C., 2016.
* [8] Nishikawa, H., “A hyperbolic Poisson solver for tetrahedral grids,” J. Comput. Phys., Vol. 409, May 2020, pp. 109358.
* [9] Xiong, M., Deng, X., Gao, X., Dong, Y., Xu, C., and Wang, Z., “A Novel Stencil Selection Method for the Gradient Reconstruction on Unstructured Grid Based on OpenFOAM,” Comput. Fluids, Vol. 172, 2018, pp. 426–442.
* [10] Sozer, E., Brehm, C., and Kiris, C. C., “Gradient Calculation Methods on Arbitrary Polyhedral Unstructured Meshes for Cell-Centered CFD Solvers,” Proc. of 52nd AIAA Aerospace Sciences Meeting, AIAA Paper 2014-1440, National Harbor, Maryland, 2014.
* [11] Shima, E., Kitamura, K., and Haga, T., “Green-Gauss/Weighted-Least-Squares Hybrid Gradient Reconstruction for Arbitrary Polyhedra Unstructured Grids,” AIAA J., Vol. 51, No. 11, 2013, pp. 2740–2747.
* [12] Wang, Q., Ren, Y.-X., Pan, J., and Li, W., “Compact High Order Finite Volume Method on Unstructured Grids III: Variational Reconstruction,” J. Comput. Phys., Vol. 337, 2017, pp. 1–26.
* [13] Nishikawa, H., “From Hyperbolic Diffusion Scheme to Gradient Method: Implicit Green-Gauss Gradients for Unstructured Grids,” J. Comput. Phys., Vol. 372, 2018, pp. 126–160.
* [14] Nishikawan, H. and White, J. A., “An efficient cell-centered finite-volume method with face-averaged nodal-gradients for triangular grids,” J. Comput. Phys., Vol. 411, 2020, pp. 109423.
* [15] White, J. A. and Nishikawa, H., “Computation of the Wall Heat Flux for a Hypersonic Flow Using Tetrahedral Grids,” AIAA SciTech, 2021, to be published as AIAA paper.
* [16] Roe, P. L., “Approximate Riemann Solvers, Parameter Vectors, and Difference Schemes,” J. Comput. Phys., Vol. 43, 1981, pp. 357–372.
* [17] Nishikawa, H., “Beyond Interface Gradient: A General Principle for Constructing Diffusion Schemes,” Proc. of 40th AIAA Fluid Dynamics Conference and Exhibit, AIAA Paper 2010-5093, Chicago, 2010.
* [18] Park, M. A. and Darmofal, D. L., “Parallel Anisotropic Tetrahedral Adaptation,” Proc. of 46th AIAA Aerospace Sciences Meeting and Exhibit, AIAA Paper 2008-917, Reno, Nevada, 2008.
* [19] Park, M. A., Balan, A., Anderson, W. K., Galbraith, M. C., Caplan, P. C., Carson, H. A., Michal, T., Krakos, J. A., Kamenetskiy, D. S., Loseille, A., Alauzet, F., Frazza, L., and Barral, N., “Verification of Unstructured Grid Adaptation Components,” AIAA Paper 2019–1723, 2019.
|
# Utilizing uncertainty estimation in deep learning segmentation of
fluorescence microscopy images with missing markers
###### Abstract
Fluorescence microscopy images contain several channels, each indicating a
marker staining the sample. Since many different marker combinations are
utilized in practice, it has been challenging to apply deep learning based
segmentation models, which expect a predefined channel combination for all
training samples as well as at inference for future application. Recent work
circumvents this problem using a modality attention approach to be effective
across any possible marker combination. However, for combinations that do not
exist in a labeled training dataset, one cannot have any estimation of
potential segmentation quality if that combination is encountered during
inference. Without this, not only one lacks quality assurance but one also
does not know where to put any additional imaging and labeling effort. We
herein propose a method to estimate segmentation quality on unlabeled images
by ($i$) estimating both aleatoric and epistemic uncertainties of
convolutional neural networks for image segmentation, and ($ii$) training a
Random Forest model for the interpretation of uncertainty features via
regression to their corresponding segmentation metrics. Additionally, we
demonstrate that including these uncertainty measures during training can
provide an improvement on segmentation performance.
## 1 Introduction
Recent progress in Convolutional Neural Networks (CNNs) for image segmentation
has enabled many possibilities in biomedical tissues analysis using
fluorescence microscopy (FM) [1], which images fluorescent dyes (called
_markers_) that target different cells or anatomical networks in biological
specimens [2]. Different markers are combined and registered as image
channels, which are for our purposes equivalent to modalities in other imaging
techniques such as with different sequences in MRI. However, the maximum
number of markers in a FM sample is limited due to their spectral overlap,
thus requiring different combinations for each biological study. Furthermore,
the lack of immunostaining consistency due to sample preparation difficulties,
together with limited sample availability, often leads to datasets of images
with missing markers, which have been challenging for traditional CNNs. To
address this, a modality sampling and attention approach named _Marker
Sampling & Marker Excite_ (_MS-ME_) was proposed in [3], which allows for
training and inference with a single model on datasets with heterogeneous
marker combinations. Although _MS-ME_ permits predictions on marker
combinations unseen during training, one does not know in advance what
segmentation quality to expect for such unseen combinations without any
existing labeled data. Such quality estimators would be very valuable not only
in predicting potential future performance, but also in deciding where to
invest additional imaging and labeling effort.
Uncertainty estimation in deep neural networks with different Bayesian
approximations was shown to improve model predictions, either by explaining
this within the loss function or by aggregating predictions from ensembles
[4]. Such estimations have been successfully applied to medical image
segmentation [5]. _Epistemic_ uncertainty accounts for lack of confidence in
the parameters of a model, i.e. uncertainty that can be reduced with
additional labeled data. This can be estimated using so-called Dropout layers
[6] both at training and inference time, known as Monte Carlo (MC) dropouts
[7, 8]. _Aleatoric_ uncertainty captures the inherent noise in the
observations, i.e. uncertainty that cannot be reduced with additional data;
and it can be estimated with the inclusion of a stochastic loss as proposed in
[4]. We herein study the use of the above uncertainty estimation tools to
design a method for the estimation of segmentation quality in the above-
described problem setting of heterogeneous FM marker combinations.
## 2 Methods
As illustrated in Fig. 1, we extend _MS-ME_ to also estimate aleatoric and
epistemic uncertainties, using the summary statistics of which we train a
regression Random Forest (RF) to predict quantitative segmentation outcomes.
Fig. 1: Illustration of our proposed framework for predicting segmentation
quality with missing markers.
### 2.1 Learning from images with missing markers
FM images are formed by different combinations of markers $m_{G}$, with
$G\subseteq\\{1,\dots,K\\}$, and $K$ the number of possible markers
represented as channels in an image $x\in\mathbb{R}^{h\times w\times|G|}$. We
denote the combination as the successive indexes of the markers it contains,
e.g. $m_{24}$ is a combination of markers 2 and 4. The challenge of different
marker combinations in FM images for both in training and testing was
addressed in [3] using _MS-ME_. Marker sampling (MS) refers to MC dropout of
modalities at training time, to make testing generalizable to different
availability of markers. For Marker Excite (ME), a feature-wise attention
module that has 2 fully-connected layers and a one-hot encoded input of
available modalities for the sample is added at different layers of a UNet
[1]. For a detailed network structure and implementation details, see [3].
### 2.2 Uncertainty estimation in CNN-based segmentation
Different uncertainties are estimated following the frameworks proposed in [4,
5] and included within a _MS-ME_ model $f$ that is applied on an input image
$x$ to predict its corresponding segmentation $\hat{y}=\textit{softmax}(z)$,
where $z$ are the logits resulting from the model: $z=f(x)$.
To estimate epistemic uncertainty $u_{\mathrm{e}}$ of $x$, MC Dropout is
employed at different layers of a CNN $f_{\mathrm{e}(p)}$ with probability $p$
both at training and inference. Since the output of the network is stochastic,
$\hat{y}$ and $u_{\mathrm{e}}$ are estimated respectively as the mean and
standard deviation (SD) of $T$ samples from $f_{\mathrm{e}(p)}(x)$. When
explicitly stated, we add MC Dropout only after the last layer as proposed in
[7].
Aleatoric uncertainty $u_{\mathrm{a}}$ is calculated by explicitly adding a
predictive variance $u_{\mathrm{a}}$ to our model output, i.e.
$[z,\,u_{a}]=f_{\mathrm{a}}(x)$ . This model $f_{\mathrm{a}}$ is trained with
a stochastic cross-entropy loss that adds a noise component
$\epsilon\sim\mathcal{N}(0,\,I)$ to multiple ($T$) model predictions, so that
the loss evaluates their mean:
$\hat{y}^{\text{training}}=\frac{1}{T}\sum_{t}^{T}\textit{softmax}(z+u_{a}\epsilon_{t})$
At inference, both $\hat{y}$ and $u_{a}$ can be obtained without sampling.
Both techniques are simultaneously included in a single _combined_ model
$f_{\mathrm{e+a}(p)}$ that is employed to separately estimate both $u_{e}$ and
$u_{a}$. The number of prediction samples is set to $T=50$ in our experiments.
### 2.3 Predicting segmentation quality from uncertainty
Different uncertainty measures only provide visual cues about prediction
errors. We herein, however, aim to obtain a quantitative estimation of the
segmentation quality $q$. To this end, we propose different regression models
$g$ to obtain quality predictions $\hat{q}=g(u)$ from uncertainties $u$
obtained from $f(x)$ as described above. We train $g$ on $u$ extracted from
all possible marker combinations within the validation set of the segmentation
task, and compare $\hat{q}$ with the ground truth $q$ extracted from their
corresponding segmentations. For comparison, herein we use the $F1$-score for
binary classification, i.e.:
$q=\frac{2\left|y\cap\operatorname*{argmax}(\hat{y})\right|}{|y|+|\operatorname*{argmax}(\hat{y})|}\
.$
We subsequently evaluate $\hat{q}$ on the test set using a Root Mean Squared
Error (RMSE) metric with respect to $q$ across all possible marker
combinations. We employ 4-fold cross-validation on the same data split as the
segmentation task.
A first approach tested is to train an additional CNN $g_{\mathrm{CNN}}$ to
predict $\hat{q}=g_{\mathrm{CNN}}(\\{u_{e},u_{a}\\})$. We design a simple
regression architecture with a very small number of parameters to avoid
overfitting to the limited size of the validation set. By denoting 2D
convolutional layers with $l$ nodes and $3\times 3$ kernels as $\text{C}(l)$,
2D max pooling layers with $3\times 3$ kernel as MP, and fully connected
layers with $l$ nodes as $\text{FC}(l)$, we use the following CNN:
$\text{C}(4)\shortrightarrow\\!\text{MP}\\!\shortrightarrow\\!\text{C}(8)\shortrightarrow\\!\text{MP}\\!\shortrightarrow\\!\text{C}(16)\shortrightarrow\\!\text{MP}\\!\shortrightarrow\\!\text{C}(32)\shortrightarrow\\!\text{MP}\\!\shortrightarrow\\!\text{C}(64)\shortrightarrow\\!\text{FC}(128)\\!\shortrightarrow\\!\text{FC}(1)$.
We add a ReLu activation after every convolutional or fully connected layer,
and use a batch size of 2. We train for 100 epochs with $L2$ loss and Adam
optimizer [9] with a learning rate of $10^{-3}$.
Since different uncertainty maps may contain error-related information that
qualitatively correlate with $q$, we hypothesize that traditional machine-
learning models that have far lower degrees-of-freedom by utilizing hand-
crafted globally informative features are potentially more robust to
overfitting to the limited data available for this task. To this end, we
alternatively train a RF model $g_{\mathrm{RF}}$ with 128 trees, using mean
squared error as split criterion. As features we use: uncertainty map
percentiles (99 values from 1st to 99th), its cumulative histogram (in 13 bins
from values 0.05 to 0.65), its first four statistical moments, and a one-hot
vector indicating which among all possible marker combinations is used. We
compare three approaches: $g_{\mathrm{RF}}(u_{e})$ with only epistemic
features, $g_{\mathrm{RF}}(u_{a})$ with only aleatoric, and
$g_{\mathrm{RF}}(\\{u_{e},u_{a}\\})$ with both.
## 3 Results and Discussion
### 3.1 Dataset and use of markers
We employ the FM dataset of bone marrow vasculature described in [10] with the
experimental settings in [2]. The dataset contains 8 samples decomposed into
different 2D patches (a total of 230), each with $K=5$ markers. We use the
annotated class _sinusoids_ , which amounts to 11.41% of the pixels, the rest
being considered background. The samples are divided into 5 for training, 1
for validation, and 2 for testing. Since the segmentation quality vastly
differs across marker combinations, we evaluate a relative segmentation
$F1$-score with respect to a reference model on the test set. This score
compares the pair-wise $F1$-scores across 4 cross-validation steps and all
possible combinations (31) of the 5 available markers. Since it is not
feasible to study all possible marker combinations in the training set, 7
training scenarios are proposed. In the first one, all 5 markers are available
in all 5 training samples. In the rest, test cases #1 to #6 given in [2],
where different markers are artificially ablated on each of the samples, are
adopted herein for comparability with the baseline.
### 3.2 Utilizing uncertainties in segmentation
We herein study how the segmentation $F1$-score is affected in our missing
markers framework when accounting for the presented uncertainty methods in the
_MS-ME_ architecture. We compare the performance to a reference _MS-ME_
without uncertainty estimation across all 31 possible marker combinations, 4
cross-validation steps, and 7 missing marker scenarios. The results in Fig.
2(a) show that in the estimation of $u_{e}$, it is best to use MC dropout in
all convolutional layers with $p=0.2$ ($f_{\mathrm{e}}(p=0.2)$), which is also
superior to baseline _MS-ME_. We assess that the $F1$-score superiority is not
merely due to the use of Dropout by comparing it in Fig. 2(b) to a _MS-ME_
model with standard Dropout layers (not sampling at inference) with the same
probability, which is significantly worse than _MS-ME_ and
$f_{\mathrm{e}}(p=0.2)$.
Fig. 2: Segmentation performance of different proposed models $f$ evaluated
w.r.t. baseline _MS-ME_ (n=868). (a) Effect of different $p$ in
$f_{\mathrm{e}}(p)$. (b) Comparison with conventional Dropout at training. (c)
Comparison of the proposed uncertainty-based models. Symmetrical logarithmic
scale (linear between -2 and 2, logarithmic elsewhere) is used. Significance
is shown between different models (*0pt -.4ex 0pt .5ex$|$—–*0pt -.4ex 0pt
.5ex$|$), or w.r.t. the reference _MS-ME_ (—–), in red when a proposed model
performs worse.
Using the stochastic loss for the estimation of $u_{a}$ with $f_{\mathrm{a}}$
leads to an inferior $F1$-score as seen in Fig. 2(c). But as discussed in the
next part, such information is desirable regardless of the negative results.
Thus, we employ $f_{\mathrm{e+a}}(p=0.2)$ that allows for prediction of both
$u_{e}$ and $u_{e}$ while providing the best $F1$-score together with
$f_{\mathrm{e}}(p=0.2)$.
### 3.3 Predicting segmentation quality for the unseen
In addition to the superior $F1$-score achieved with the use of
$f_{\mathrm{e+a}}(p=0.2)$, the simultaneous estimation of $u_{e}$ and $u_{a}$
allows to visually inspect potential mistakes in the CNN prediction. Here, we
use such uncertainty maps to estimate the $F1$-score $q$ of segmented images.
We evaluate our proposed methods on the training setting denominated case #6,
for which patches have marker combinations $m_{135}$, $m_{124}$, $m_{35}$,
$m_{23}$, or $m_{45}$ depending on which of the 5 training samples they belong
to. This setting contains a variety of markers for each sample that depicts
scenarios usually found in practise, and it allows to evaluate the predicted
$q$ in markers not available for any of its training samples.
We show in Fig. 3(a) the RMSE results for the $F1$-score prediction with the
methods presented above. Using $g_{\mathrm{CNN}}$ leads to worse RMSE results
than any of the $g_{\mathrm{RF}}$ methods, which can be explained by the
tendency of CNNs to overfit in such a small training set, and that the
simplicity of the task asks for the efficient use of predefined explanatory
features where RF methods excel. $g_{\mathrm{RF}}(\\{u_{e},u_{a}\\})$ is seen
to be superior in RMSE to both $g_{\mathrm{RF}}(u_{e})$ and
$g_{\mathrm{RF}}(u_{a})$. This observation can be attributed to the fact that
$u_{e}$ and $u_{a}$ capture information about distinct segmentation errors
that both help in the regression of a more accurate $F1$-score value. Although
further improvements could be achieved with deep ensembles, which have been
reported to estimate more accurate uncertainties than MC dropout [11], they
are computationally very expensive, which would annul a main advantage of our
framework of training a single model.
We subsequently employ $g_{\mathrm{RF}}(\\{u_{e},u_{a}\\})$ to estimate
$F1$-score for both seen and unseen marker combinations in our training
scenario, which is observed in Fig. 3(b) to closely follow ($R^{2}$$=$$0.98$)
the ground-truth predictions on average across cross-validation steps. Thus,
despite potentially high standard deviations (and high RMSE) per individual
patch predictions, we can still make accurate overall predictions about the
expected quality of a marker combination, including those that are unseen.
Fig. 3: Prediction of $F1$-score from uncertainty maps on the test set. (a)
RMSE of different models (n=124). (b) Comparison of predicted $F1$-score using
$g_{\mathrm{RF}}(\\{u_{e},u_{a}\\})$ to ground truth for each of the 31 marker
combinations shown as point (mean) and bars (SD) (n=4).
We also visually assess the individual effects of $u_{e}$ and $u_{a}$ in Fig.
4. $u_{a}$ is observed to focus on the edges of the vessels, where
discrepancies in annotations often exist and thus fit the definition of
aleatoric uncertainty of capturing noise inherent in the observations.
Meanwhile, $u_{e}$ captures mistakenly segmented vessels which appear in
$m_{4}$ but are not sinusoids. Such errors are related to lack of information
in the model, and may be corrected by adding labeled training data containing
patterns similar to those highlighted by $u_{e}$.
Fig. 4: Visual example of a prediction and their uncertainties for the test
set marker combination $m_{34}$ in training case #6. The error image shows
both false positives and negatives. The $F1$-score is predicted using
$g_{\mathrm{RF}}(\\{u_{e},u_{a}\\})$.
## 4 Conclusion
With the proposed framework for inspection of segmentation uncertainties with
missing FM markers, the advantages are twofold: First, accounting for
epistemic and aleatoric uncertainties simultaneously produces segmentation
results superior to the state-of-the-art _MS-ME_ model. Second, with a
uncertainty feature based regressor, we can estimate segmentation quality for
any possible marker combination, whether it was seen during training or not.
Thus, in practise we can quantitatively evaluate how suitable a trained model
is for future samples stained with a previously unseen marker combination,
i.e. to discard them or preferentially annotate them, e.g., in an active
learning framework. Furthermore, comparison of different uncertainties may
indicate whether to focus more on labeled data or on improving image
acquisition quality. Note that the proposed methods are applicable also to
other multi-modality frameworks, such as MRI.
## 5 Compliance with Ethical Standards
This research was conducted on a dataset described in [10, 3], where the
ethical compliance for sample preparation was described.
## 6 Acknowledgments
Funding was provided by Hasler Foundation, Swiss National Science Foundation
(SNSF), Swiss Cancer Research Foundation, and Julius-Müller Foundation.
## References
* [1] Thorsten Falk et al., “U-net: deep learning for cell counting, detection, and morphometry,” Nature methods, vol. 16, no. 1, pp. 67–70, 2019.
* [2] Alvaro Gomariz et al., “Imaging and spatial analysis of hematopoietic stem cell niches,” Annals of the New York Academy of Sciences, vol. 1466, no. 1, pp. 5–16, 2020.
* [3] Alvaro Gomariz et al., “Modality attention and sampling enables deep learning with heterogeneous marker combinations in fluorescence microscopy,” arXiv preprint arXiv:2008.12380, 2020.
* [4] Alex Kendall and Yarin Gal, “What uncertainties do we need in bayesian deep learning for computer vision?,” in NeurIPS, 2017.
* [5] Yuta Hiasa et al., “Automated muscle segmentation from clinical ct using bayesian u-net for personalized musculoskeletal modeling,” IEEE Transactions on Medical Imaging, vol. 39, no. 4, pp. 1030–1040, 2019.
* [6] Nitish Srivastava et al., “Dropout: a simple way to prevent neural networks from overfitting,” The journal of machine learning research, vol. 15, no. 1, pp. 1929–1958, 2014.
* [7] Yarin Gal and Zoubin Ghahramani, “Dropout as a bayesian approximation: Representing model uncertainty in deep learning,” in ICML, 2016.
* [8] Yarin Gal and Zoubin Ghahramani, “Bayesian convolutional neural networks with bernoulli approximate variational inference,” in ICLR workshop track, 2016.
* [9] Diederik P Kingma and Jimmy Ba, “Adam: A method for stochastic optimization,” in ICLR, 2014.
* [10] Alvaro Gomariz et al., “Quantitative spatial analysis of haematopoiesis-regulating stromal cells in the bone marrow microenvironment by 3d microscopy,” Nature communications, vol. 9, no. 1, pp. 1–15, 2018.
* [11] Stanislav Fort, Huiyi Hu, and Balaji Lakshminarayanan, “Deep ensembles: A loss landscape perspective,” 2020.
|
# Deep Neural Networks for Active Wave Breaking Classification
Caio E. Stringari France Energies Marines, Plouzané, 29280, France
<EMAIL_ADDRESS>Pedro V. Guimarães France
Energies Marines, Plouzané, 29280, France PPGOceano, Federal University of
Santa Catarina, Florianópolis, 88040-900, Brazil these authors contributed
equally to this work Jean-François Filipot France Energies Marines,
Plouzané, 29280, France these authors contributed equally to this work
Fabien Leckler France Energies Marines, Plouzané, 29280, France Rui Duarte
France Energies Marines, Plouzané, 29280, France
###### Abstract
Wave breaking is an important process for energy dissipation in the open ocean
and coastal seas. It drives beach morphodynamics, controls air-sea
interactions, determines when ship and offshore structure operations can occur
safely, and influences on the retrieval of ocean properties from satellites.
Still, wave breaking lacks a proper physical understanding mainly due to
scarce observational field data. Consequently, new methods and data are
required to improve our current understanding of this process. In this paper
we present a novel machine learning method to detect active wave breaking,
that is, waves that are actively generating visible bubble entrainment in
video imagery data. The present method is based on classical machine learning
and deep learning techniques and is made freely available to the community
alongside this publication. The results indicate that our best performing
model had a balanced classification accuracy score of $\approx$ 90% when
classifying active wave breaking in the test dataset. An example of a direct
application of the method includes a statistical description of geometrical
and kinematic properties of breaking waves. We expect that the present method
and the associated dataset will be crucial for future research related to wave
breaking in several areas of research, which include but are not limited to:
improving operational forecast models, developing risk assessment and coastal
management tools, and refining the retrieval of remotely sensed ocean
properties.
## Introduction
Wave breaking is one of the most challenging water wave phenomena to
investigate. Despite nearly two centuries of research, the current practical
description of the process is still mostly empirical [1, 2, 3, 4]. Precise
wave breaking modelling requires explicit knowledge of the wave phase speed
and the fluid velocity distribution on the wave crest which can only be done
by numerically solving the Navier-Stokes equations in a framework that
currently is too computationally expensive for practical applications[5].
Consequently, current large-scale state-of-the-art wave models rely on
statistical approaches (mostly the spectral approach) to represent the waves
[6, 7, 8]. This type of models parameterizes wave breaking as a function of
known parameters such as the wind speed [6], the local wave height to water
depth ratio [7] or semi-empirical breaking wave height probability
distributions [2, 9]. Due to a lack of observed data, the constants involved
in these models have been derived from limited datasets that may not
adequately represent the natural environment. For example, a recent study has
shown that popular surf zone parametric wave breaking models incorrectly
represented the fraction of broken waves in their formulations with errors
>50%, despite these models being able to adequately represent surf zone energy
dissipation possibly due to parameter tuning [10]. This paper aims to start
addressing the data unavailability issue by providing a reliable and
reproducible method to detect and track waves that are actively breaking in
video imagery data.
Wave breaking directly affects several environmental phenomena. For example,
wave breaking is considered as the main driver for air-sea exchanges[11],
being often described as a function of Phillips’ [12] $\Lambda(c)dc$
parameter. This parameter is defined as the average total length per unit
surface area of breaking fronts that have velocities in the range $c$ to
$c+dc$ and its moments are assumed to correspond to different phenomena such
as whitecap coverage (second moment), rate of air entrainment (third moment),
and momentum flux (fourth moment). As previously identified [13], different
interpretations of how data is processed to obtain $\Lambda(c)dc$ resulted in
differences of 175% in the first moment and 300% in the fifth moment of
$\Lambda(c)$ [14, 13]. This issue does not seem to limit the applicability of
Phillips’ framework. A recent study, for example, used a wave breaking
parameterization fully based on $\Lambda(c)dc$ to derived whitecap coverage
off the coast of California[15]. To the authors’ knowledge, however, no study
has assessed the impact that errors in wave breaking detection has on measured
$\Lambda(c)$ distributions and how such errors would impact on the conclusions
that have been drawn from these models (for example, the assumption that
$\Lambda(c)$ is a simple function of wind speed[11]). More importantly, in the
case which new observations diverge from Phillips’ original theory (as we will
later show in this paper), new and improved wave breaking models will
certainly follow.
Further, breaking waves have a direct impact on the remote sensing of the
ocean surface from satellites, aircraft, or platform mounted instruments. Wave
breaking has been observed to significantly modulate measurements of
backscatter off radars with large increases in backscatter being directly
correlated with breaking waves[16]. Given that part of the data that will be
used here (see Methods for details) coincides with the location of previous
wave backscatter studies[16], data derived from our method could to lead to
explicit formulations correlating wave breaking and radar backscatter. Such
formulations could, in the future, lead to more reliable satellite-derived
scatterometer data, which currently ignores [17] or relies on empirical wave
breaking models[18]. The foam generated by breaking waves at the sea surface
also influences on the measurement of microwave brightness temperature
signatures for moderate to high wind speeds [19, 20]. To date, no study has,
however, systematically quantified such temperature signatures for large field
datasets and most of our knowledge is empirical. Such limitation could be
addressed by extending wave breaking detection and tracking methods to
infrared wavelengths, which is directly correlated to water temperature[21].
Advancements in this regard could lead to improvements in global climate
models given that the inclusion of wave-generated heat transfer has shown to
reduce cold biases in such models for non-breaking waves[22]. Breaking waves
are expected to transfer more heat than non-breaking waves[23] and, when
considered, could improve climate models even further.
With the recent explosion in the usage of machine learning, particularly deep
neural networks, it was only a matter of time before these techniques were
adapted to wave research. Recent studies have shown that deep neural networks
can accurately be used to classify different types of surf zone breakers [24],
obtain wave heights from video data [25], track waves in the surf zone[26] and
the shoreline position [27] and, when applied to pressure transducer data,
distinguish between broken and unbroken waves [10]. In this paper, we describe
a robust and extensible method to identify and track breaking waves in video
imagery data using a combination of classic machine learning algorithms and
deep learning. More importantly, we make the present dataset and code base
fully available for future researchers who can either use it directly, re-
train the models adding more training data, or expand the present method to
other cases (for example, object segmentation). The development and
standardization of a general framework to detect wave breaking in video
imagery data should help to provide the wave breaking statistics database that
is currently needed to assess, or further develop, our current understanding
of several processes related to breaking waves. Ultimately, the efforts
initiated here aim to lead to improvements in several areas of research such
as wave modelling and forecasting, wave energy tracking and harvesting [28],
ocean-atmosphere interaction [15], remote sensing of the ocean[16], and safety
at sea and coasts[29].
## Method
### Model Definition
In this study, we have developed a novel method to detect active wave breaking
in video imagery data. Similarly to previous studies, we exploit the fact that
breaking wave crests generate characteristic white foam patches from bubble
entrainment [30, 31]. Differently from the vast majority of previous methods,
here we make a clear distinction between active wave breaking (that is,
visible foam being actively generated by wave breaking) from passive foam and
to all other non-wave breaking instances. To the authors’ knowledge, only the
methods of Mironov & Dulov[30] and Kleiss & Melville[32] have previously
attempted to distinguish between active wave breaking and passive foam. Both
their methods start similarly to ours (see next paragraph) by identifying
bright pixels in a given series of images via thresholding. Next,
interconnected pixels are connected in space and time using a nearest neighbor
search (Mironov & Dulov[30]) or by extracting connected contours (Kleiss &
Melville[32]). In both previous methods, the separation between active and
passive wave breaking is then done by empirically defining foam expansion
direction and velocity limits. The present method aims to remove empirical
steps in active wave breaking detection by using data-driven methods and deep
neural networks instead of experiment-specific parameters.
Our method starts by naïvely identifying wave breaking candidates in a given
input image. This was done by identifying bright (that is, white) pixels in
the image using the local thresholding algorithm available from the OpenCV[33]
library with its default parameters. Interconnected regions of bright pixels
in the image were then clustered using DBSCAN [34] and a minimum enclosing
ellipse [35] was fitted to each identified cluster. The DBSCAN step requires
the user to define a minimum number of pixels to be clustered (default value
is 10 for all data used here) and a maximum distance allowed between pixels
(default value of 5 for all data used here). This step resulted in large
quantities of bright pixel patches being detected. At this stage, it was not
possible to determine whether a given identified cluster of pixels represented
active wave breaking, passive foam, or any other instance of bright pixels
(for example, birds, boats, coastal structures, sun glints or white clouds).
Generically, this step can be thought of a feature extraction step and can be
replaced by any other equivalent approach (for example, image feature
extractors). To avoid exponential memory consumption growth generated by
DBSCAN, the input image may be subdivided into blocks of regular size. This
step was easily parallelized with performance increasing linearly with number
of computing threads.
The second step of the method consisted of training a deep convolutional
neural network to distinguish between active wave breaking and passive foam
(and all other non-desired occurrences). This step can be understood as a
binary classification step. Figure 1 shows a schematic representation of the
present method. From the original images, randomly selected subsets of size
256x256 pixels centered on the fitted ellipses were extracted and manually
labelled as either active wave breaking (label 1, or positive) or otherwise
(label 0, or negative). The present training dataset was generated using raw
video imagery data from Guimaraẽs et al. [36] and consisted of 19000 training
and 1300 testing images (see Table 1). The training dataset is further split
into 80% training and 20% validation data[37]. The validation dataset is
reserved for future hyper-parameter fine-tuning. Note that the present dataset
has a class imbalance of $\approx$90% towards the negative label, that is, for
each active wave breaking sample (label 1, or positive) there are nine
instances that were not active wave breaking (label 0, or negative). Data
augmentation [38] (rotation, vertical and horizontal flips, and zoom) was
employed during training to increase the variety of samples of the positive
label.
Five state-of-the-art neural network architectures, or backbones (VGG16 [39],
ResNet50V2 [40], InceptionResNetV2 [41], MobileNetV2 [42] and EfficientNetB5
[43]), were implemented and can be chosen by end users (see our Github code
repository https://github.com/caiostringari/deepwaves for usage guidance). All
of these backbones make use of convolutional layers to gradually extract
information from the input images. The main differences between them are how
many layers they have, the size of the convolution windows, how data is
normalized in each layer and how each layer connects to each other (or to
previous layers in the stack). For example, VGG16 is 16 layers deep, uses 3x3
convolution windows and has no normalization. InceptionResNetV2 is 164 layers
deep, uses a mix of 5x5, 3x3 and 1x1 convolution windows and uses batch
normalization[44] to help avoiding overfitting. Residual Networks (namely,
ResNet50V2) not only connect adjacent layers but also take into account errors
(residuals) from previous layers. In general, more modern architectures
(namely, EfficientNet) are wider (mainly by having parallel convolution
windows) as well as much deeper than older architectures (namely, VGG16). The
final top layers of the network were fixed for all backbones and consisted of
flattening the last convolutional layer, two fully-connected layers with 50%
dropout [45], and a final classification layer with sigmoid activation.
The optimization step (training) was done using the Adam [46] implementation
of the stochastic gradient descent method [47] and binary cross-entropy was
used as the loss function. Note that this step must be computed using a
graphics processing unit (GPU) in order to achieve feasible computation times.
The models took from three (VGG16) to twelve (EffientNetB5) hours to train
using a NVIDIA GTX 1080 GPU with a batch size of sixty-four images. After the
neural networks were sufficiently trained, the best performing network (VGG16,
see below) was used to classify all the naïvely identified wave breaking
candidates and only the events classified as active wave breaking were kept
for further analyses. Although VGG16 was chosen for presenting the results,
the performance of the other architectures is nearly identical to VGG16 on
real-world applications. Finally, note that user-tunable parameters for the
neural networks (for example, learning rate and neuron activation thresholds)
were kept unchanged from their default values in the TensorFlow library[48].
This implicates that more aggressive parameter optimization could improve the
results presented here even further. For sake of brevity, we refer the reader
to the project’s code repository for guidance on how to select hyper-
parameters.
Figure 1: Schematic representation of a deep convolutional neural network. The
input layer consists of an active wave breaking candidate and has shape
256x256x3 (image height, image width and number of RGB channels). In the case
of a grey-scale image, the single grey channel was triplicated. The red dashed
rectangle in the analyzed image shows the region of interest which, in this
particular case, was equal to the stereo video reconstruction area. The
intermediary convolutional layers (or backbones) are interchangeable with
different options available (see text for details). The function of the
convolutional layers is to extract features from the input image by using
convolutions (3x3 in this example) and max pooling[49] (that is, selecting the
brightest pixel in a given window). The last convolutional layer is flattened
(that is, turned into a one-dimensional vector) and is connected to two fully-
connected (that is, multi-layer perceptron-like) layers and one final
classification layer. A 50% dropout (that is, random selection of neurons in a
layer) is applied after each fully connected layer. The final classification
layer has one unit and uses sigmoid activation with a threshold of 0.5. Table
1: Data characterization summary table. $f$ is the sampling frequency in
Hertz, $D$ is the duration of the experiment in seconds, $H_{s}$ is the
significant wave height computed from the wave spectrum, $T_{p1}$ and $T_{p2}$
are, respectively, the peak wave period of the first and second spectral
partitions computed following Hanson & Jensen[50], $D_{p}$ is the peak wave
direction, and $U_{10}$ is the wind speed measured or converted (denoted by
the $*$) to a height of 10m above the sea surface using Large & Pond’s[51]
formula. Tr. S. and Ts. S. are the sample size of the train and test datasets,
respectively.
Location | Date and Time | $f$ $[Hz]$ | D. $[min]$ | $H_{s}$ [m] | $T_{p1}$ $[s]$ | $T_{p2}$ $[s]$ | $U_{10}$ $[ms^{-1}]$ | $D_{p}$ | Tr. S. | Ts. S.
---|---|---|---|---|---|---|---|---|---|---
Black Sea | 2011/10/01 14:18 | 12 | 07 | 0.3 | 6.20 | 3.10 | 10.7* | WSW | 1000 | 100
Black Sea | 2011/10/04 09:38 | 12 | 20 | 0.36 | 6.10 | 2.63 | 10.1* | WSW | 1000 | 100
Black Sea | 2011/10/04 11:07 | 12 | 30 | 0.45 | 6.10 | 3.16 | 12.2* | WSW | 1000 | 100
Black Sea | 2011/10/04 13:30 | 12 | 30 | 0.55 | 6.60 | 3.71 | 12.9* | WSW | 1000 | 100
Black Sea | 2013/09/22 13:00 | 10 | 15 | 0.66 | 4.30 | 30 | 8.7* | E | 1000 | 100
Black Sea | 2013/09/25 12:15 | 12 | 15 | 0.41 | 4.10 | 1.20 | 6.1* | N | 1000 | 100
Black Sea | 2013/09/30 10:20 | 12 | 15 | 0.65 | 5.70 | 1.80 | 15.2* | N | 1000 | 100
Adriatic Sea | 2014/03/27 09:10 | 12 | 60 | 1.36 | 5.02 | - | 9.9 | ENE | 2000 | 100
Adriatic Sea | 2015/03/05 10:35 | 12 | 60 | 1.33 | 6.10 | 5.02 | 9.0 | ENE | 2000 | 100
Yellow Sea | 2017/05/13 05:00 | 10 | 05 | 1.93 | 5.20 | 4.02 | 13.4 | NW | 2000 | 100
La Jument | 2017/12/15 14:20 | 10 | 30 | 5.88 | 10.00 | 6.80 | 13.3 | W | 2000 | 100
La Jument | 2018/01/03 09:39 | 10 | 30 | 10.03 | 12.80 | 9.30 | 17.9 | W | 2000 | 100
La Jument | 2018/01/04 11:43 | 10 | 30 | 7.52 | 11.10 | - | 14.7 | W | 2000 | 100
| | | | | | | | Total: | 19000 | 1300
The last step of the method consisted of grouping the active wave breaking
events in space and time (note that at this stage bright pixels are only
grouped in space). Time-discrete wave breaking events had their bounding
ellipse region filled in pixel space ($i,j$) and were stacked in time ($t$)
resulting in a point-cloud-like three-dimensional ($i,j,t$) structure. The
DBSCAN[34] algorithm was then used to cluster these data and obtain uniquely
labelled clusters. The two parameters for the DBSCAN algorithm, that is, the
minimum number of samples in each cluster ($n_{min}$) and the minimum distance
allowed between two samples ($eps$), were set to be equals to the minimum
number of points inside a fitted ellipse among all ellipses ($n_{min}$) and
equals to the sampling frequency in Hertz ($eps$). These values for $n_{min}$
and $eps$ were constant among the analyzed datasets. Note that this final step
can be replaced by any other density-based clustering algorithm or other more
sophisticated algorithms such as SORT[52] (which is also available on the
project’s code repository but not was used here).
Note that up to the clustering step, all calculations were done in pixel
domain and "time" was obtained from sequential frame numbers. To convert
between pixel and metric coordinates, the grids constructed using the stereo-
video dataset available from Guimarães et al. [36] were used in the present
study. If stereo video data are not available, such conversions could be done
knowing the camera position in real-world (that is, metric) coordinates, which
is usually done using surveyed ground control points[53, 54]. Conversion from
sequential frame numbers to time (in seconds) can be done by knowing the
sample rate (in frames per second, that is, Hertz) for the data. The total
amount of computational time required to process 20 minutes of raw video
recorded at 10Hz with an image size of 5 megapixels is approximately two hours
on a modern six-core computer (assuming that a pre-trained neural network is
available). Much shorter processing times are achievable for smaller images
sizes and higher number of computation threads.
### Evaluation Metrics
Due to the imbalanced characteristic of the dataset, the classification
accuracy score in isolation is not an appropriated metric to evaluated the
present test dataset. For instance, a classifier that guesses that all labels
are negative (that is, 0) would automatically obtain a high score ($\approx$
90%). To properly assess the performance of the classifier, more robust
metrics were defined. These metrics are defined as follows:
* •
True Positives ($TF$) and True Negatives ($TN$) are samples that were
correctly classified. False Positives ($FP$) or Type I errors are false
samples that were incorrectly classified as true, and False Negatives ($FN$)
or Type II errors are true samples that were classified as false.
* •
Accuracy is the percentage of examples correctly classified considering both
classes, that is, Accuracy=$\frac{TP+TN}{T+N}$, where $T$ and $N$ are the
total number of positive and negative samples, respectively.
* •
Precision is the percentage of predicted positives that were correctly
classified, that is, Precision=$\frac{TP}{TP+FP}$.
* •
Recall is the percentage of actual positives that were correctly classified,
that is, Recall=$\frac{TP}{TP+FN}$
* •
Area under the curve ($AUC$) is the area defined by plotting the FP rate
against the TP rate (also referred to as receiver operating characteristic
curve). This metric indicates the probability of a classifier ranking a random
positive sample higher than a random negative sample.
All metrics described above were monitored during the training process and
were individually assessed to rank model performance. The training curves
shown in Figure 2-a were used to assess when overfitting, that is, decreases
in the loss function value for the training dataset that were not reflected in
the validation dataset, started to occur. Training epochs after overfitting
started were discarded. Here we favor the $AUC$ curves as shown in Figure 2-b
to indicate better performing models because AUC is a more robust metric than
the classification score and at the same time presents a smooth evolution with
training epochs. Finally, a confusion matrix (or table of confusion) as shown
in Figure 2-c was plotted for each model to assess Type I and Type II errors.
## Results
### Classifier performance
From the analysis of all training curves, confusion matrices, and from Table
2, the best performing backbone architecture during training was ResNet50V2 by
a considerable margin ($AUC$=0.989). These results however, did not translate
to the validation dataset ($AUC$=0.873). Considering only the validation data,
VGG16 was the best performing backbone with $AUC$=0.946. Considering only the
test dataset, the best performing model was also VGG16 ($AUC$=0.855). Overall,
VGG16 was selected as the best performing model and the results presented in
the next sections will use this backbone. Other evaluation metrics such as the
accuracy score, precision, and recall closely followed the evolution of AUC
with training epochs (compare Figures 2-a and b, for example). In general, as
the loss value decreased, the number of false positives decreased, which made
the precision and recall to increase. This behavior was consistent for all
models. Given that different architectures may perform better for other
datasets and further optimization could change the model ranking presented
here, all pre-trained models and their training metrics evolution are made
available in the project’s code repository. Finally, it is worth mentioning
that larger models (for example, VGG19, ResNET152 and EfficientNetB7) could
achieve better results but this was not attempted there due to hardware
limitations (that is, these models required more memory than what was
available on the NVIDIA GTX 1080 GPU used in this study).
Figure 2: Examples of training curves and confusion matrix for the best overall performing model (VGG16). a) Loss function value for training and validation data. b) $AUC$ value for training and validation data. In a) and b) the hatched area indicates the epochs after which the model started to overfit, the thick colored lines show smoothed loss or $AUC$ values (average at every 10 epochs), and the transparent lines show raw loss or $AUC$ values. c) confusion matrix calculated using test data only. Table 2: Evaluation metrics for all tested backbone architectures. Refer to the main manuscript text for definition of evaluation metrics. The best performing models are shown in boldface. Results are sorted by AUC. Model | Accuracy | TP | FP | TN | FN | Precision | Recall | AUC
---|---|---|---|---|---|---|---|---
Train
ResNetV250 | 0.97 | 1414 | 198 | 13978 | 280 | 0.877 | 0.835 | 0.989
VGG16 | 0.93 | 855 | 273 | 13911 | 831 | 0.758 | 0.507 | 0.943
InceptionResnetV2 | 0.927 | 886 | 359 | 13823 | 802 | 0.712 | 0.525 | 0.932
EfficientNetB5 | 0.772 | 1403 | 3346 | 10920 | 297 | 0.295 | 0.825 | 0.874
MobileNet | 0.904 | 436 | 268 | 13916 | 1250 | 0.619 | 0.259 | 0.848
Validation
VGG16 | 0.932 | 221 | 65 | 3478 | 204 | 0.773 | 0.52 | 0.946
InceptionResnetV2 | 0.921 | 190 | 81 | 3466 | 231 | 0.701 | 0.451 | 0.93
EfficientNetB5 | 0.809 | 353 | 687 | 2856 | 72 | 0.339 | 0.831 | 0.897
MobileNet | 0.908 | 123 | 64 | 3479 | 302 | 0.658 | 0.289 | 0.878
ResNetV250 | 0.919 | 197 | 97 | 3450 | 224 | 0.67 | 0.468 | 0.873
Test
VGG16 | 0.876 | 106 | 80 | 945 | 69 | 0.57 | 0.606 | 0.855
ResNetV250 | 0.881 | 95 | 63 | 962 | 80 | 0.601 | 0.543 | 0.843
InceptionResnetV2 | 0.882 | 91 | 57 | 968 | 84 | 0.615 | 0.52 | 0.839
EfficientNetB5 | 0.873 | 88 | 65 | 960 | 87 | 0.575 | 0.503 | 0.827
MobileNet | 0.875 | 30 | 5 | 1020 | 145 | 0.857 | 0.171 | 0.768
### Test Case with Real-world Data
Figure 3 shows the results of the application of the best performing model
architecture (VGG16) on La Jument (2018/01/03 09:39) and Black Sea (2011/10/04
09:38) data. Visual inspection of these results confirmed the ability of the
neural network to correctly classify the naïvely identified wave breaking
candidates and only keep instances that represented active wave breaking.
Moreover, the same neural network was able to correctly classify active wave
breaking events for the rogue waves seen at La Jument[29] and for the small
wind-generated breakers seen in the Black Sea data. This result highlights the
ability of the neural network to generalize well on the dataset, which is a
difficult result to achieve. From the analysis of the training curves, the
averaged classification error (accounting for the imbalance in the data)
should be of the order of $\approx$10%, which to the authors’ knowledge it was
not assessed by other wave breaking detection methods. We strongly recommend
that future research should report active wave breaking detection errors when
used to assess or further develop models that depend on wave breaking data.
Figure 3: Example of the application of the method. a) Results of the naïve
wave breaking detection (thresholding + DBSCAN) for La Jument data (03/01/2018
09:39). Note the great amount of passive foam being detected as active wave
breaking. b) Results of active wave breaking detection using VGG16 as backbone
for La Jument data (03/01/2018 09:39). Note the significant reduction in the
amount of passive foam being detected. In both plots, number of clusters
refers to the number of clusters identified by DBSCAN. The red dashed
rectangle indicates the region of interest which, in these examples, was the
same as the stereo-video reconstruction area. c) Results of the naïve wave
breaking detection (thresholding + DBSCAN) for Black sea data (04/10/2011
09:38). d) Results of active wave breaking detection using VGG16 as backbone
for Black Sea data (04/10/2011 09:38). Animations of these results are
available at https://github.com/caiostringari/deepwaves. In this particular
example, the image was subdivided into blocks of 256x256 pixels for
processing. Note that identical results were seen using other architectures
other than VGG16 to classify these data.
### Comparison with Mironov & Dulov (2008)
To highlight the improvements obtained by the present method, this section
presents a comparison between Mironov & Dulov’s (2008)[30] automatic active
wave breaking detection method and the present method. To perform this task,
data from the 2013 Black Sea experiments in Table 1 that had previously been
classified and investigated in detail by Guimarães[55] were used. All the
active breaking events (that is, considering all 900s of data) detected using
Mironov & Dulov’s (2008)[30] were manually classified as true or false and
compared to the labels predicted by our method. To the authors’ knowledge,
these data are the only currently available data that has been classified by
both methods as well as manually. Figure 4 shows the result of the compassion
between models. On average, our method has relatively $\approx$50% less error
then Mironov & Dulov’s (2008) method with an averaged absolute reduction in
error of $\approx$15%. The results in Figure 4 are also consistent with the
results seen in Table 2 which showed that our model had errors in the order of
$\approx$15% when considering the validation and test datasets. Note that all
tested neural network architectures performed very similarly with only a
slight advantage for InceptionResnetV2. It is also worth mentioning that
Mironov & Dulov’s (2008) method was designed and optimized to work
specifically with data from the Black Sea and it is not guaranteed that it
will generalize to other datasets. Our method, on the contrary, has been shown
to generalize well for very distinct datasets (see Figure 3, for example) and
with further optimization could achieve even better performance.
Figure 4: Comparison between averaged classification error for active wave
breaking detection between Mironov & Dulov’s (2008)[30] method and all the
neural network architectures implemented here. The error bars represent one
standard deviation from the mean. The data used for this comparison are from
the 2013 Black Sea experiments described in Table 1.
### Wave Breaking Statistics
In this section we briefly present examples of wave breaking statistics that
can be directly derived from the classified wave breaking data. For brevity,
the analysis is limited to data from the Black Sea because our classifier
performed best at this location (classification errors < 10% considering both
2011 and 2013 experiments). Five quantities will be analysed: the wave
breaking duration ($T_{br}$), the wave breaking area ($A_{br}$), the major
($a$) and minor ($b$) axis of the fitted ellipses (representative of the wave
breaking lengths at their maximum during the active part of the wave
breaking), and Phillips’ distribution $\Lambda(c)dc$. These quantities were
obtained directly from space-time clustered wave breaking events with the
exception of the cumulative wave breaking area ($A_{br}$) which was calculated
from the projections of pixels clustered in the first step of the method to
metric coordinates. The results of this analyses are shown in Figure 5.
The wave breaking duration ($T_{br}$ normalized by wave peak period ($T_{p}$),
Figure 5-a) roughly followed a shifted Gamma probability density function
(PDF) and had a mean value of 0.12 and most frequent value (mode) of 0.13.
This result shows that the active part of the wave breaking process happens
very quickly. The wave breaking area ($A_{br}$, Figure 5-b) closely followed a
Pareto distribution which indicates that large wave breaking events are
relatively rare in the data. The ratio between the major and minor axis of the
fitted ellipses ($a/b$, Figure 5-c) followed a Beta PDF and had a mean of 2.5
and mode of 1.9, which indicates that the ellipses’ major axis is
approximately double the size of the minor axis. Assuming a negligible wave
front angle, the wave breaking area scaling relation from Duncan [56]
($A_{br}/b^{2}$, Figure 5-d) also followed a Beta PDF and had mean of 0.1 and
mode of 0.8, which is consistent with the previously reported value[56]
(0.1$\pm$0.01). The wave breaking area (Figure 5-e) showed a quadratic
increase with wave breaking event duration, which is trivial but seems not to
have been previously directly shown before. Finally, Figure 5-f shows the
$\Lambda(c)dc$ distributions obtained using our method and considering that
the ellipse major axis is representative of the wave breaking length. The
observed distributions greatly deviated from the theoretical $c^{-6}$
relation[12]. Note that all PDF fits shown here were statistically significant
at the 95% confidence level using the two-tailed Kolmogorov–Smirnov test. See
Table 3 for the description of the parameters of the PDFs presented in this
section and the Discussion section for the contextualization of these results
and the possible implications that they have for future research.
Figure 5: Examples of statistical properties of breaking waves that can be directly obtained from the proposed method. a) Probability distribution of the wave breaking duration ($T_{br}$) normalized by wave peak period ($T_{p}$). The blue line shows the Gamma fit to the data, the purple dashed line shows the mean value (0.13) and the orange dashed line shows the mode value (0.12). b) Probability distribution of the wave breaking area ($A_{br}$) normalized by wavelength ($\frac{g}{2\pi}T_{p}^{2}$). The blue line shows the Pareto fit to the data. c) Probability distribution of the ratio between major and minor axis ($a/b$). The blue line shows the Beta fit to the data, the purple dashed line shows the mean value (2.4) and the orange dashed line shows the mode value (1.8). d) Probability distribution of the wave breaking area scaling parameter ($A_{br}/b^{2}$). The blue line shows the Beta fit to the data, the purple dashed line shows the mean value (0.1), the orange dashed line shows the mode value (0.08), and the red line shows Duncan’s constant value (0.11). e) Evolution of the wave breaking area over time. The black markers show the observed values, the error bars show values binned at 0.15s intervals, the blue line shows the quadratic regression to the data, the blue swath shows the 95% confidence interval for the regression, and $r^{2}$ is the correlation coefficient. f) Phillips[12] $\Lambda(c)dc$ distributions assuming that the wave breaking speed is the same as the phase speed of the carrier wave (that is, $c=c_{br}$) grouped by wind speed. The black dashed line shows the $c^{-6}$ theoretical decay, the colored markers show the average crest length binned at 0.1$ms^{-1}$ wave speed intervals, and the colored lines show a running average with a window size of 10 bins. Table 3: Fitted parameters for the distributions seen in Figure 5 . All fits were done using Scipy[57] and, consequently, the values reported here follow Scipy’s conventions. | Distribution | Parameter 1 | Parameter 2 | Location | Scale
---|---|---|---|---|---
Figure 5-a | Gamma | 3.001 | - | 0.053 | 0.0260
Figure 5-b | Pareto | 4.339 | - | -0.007 | 0.007
Figure 5-c | Beta | 6.117 | 73.948 | 0.207 | 30.440
Figure 5-d | Beta | 6.389 | 40.470 | -0.760 | 24.438
## Discussion
We have presented a new method to detect active wave breaking in video imagery
data that is robust and easily transferable to future studies. Overall, VGG16
was the best performing architecture which is a surprising result given that
VGG is a considerably older architecture that has been superseded by more
recent architectures such as ResNets and EfficientNets [43]. Also
surprisingly, EfficientNet was one of the worst performers considering the
test dataset despite the state-of-the-art results that this architecture
achieved in recent years[43]. One explanation could be that given VGG16 is the
model with the highest number of parameters (despite being the shallowest
network), it adapted better to the characteristics of the present dataset.
Another explanation could be that EfficientNet is currently overfit on the
ImageNet[58] dataset that was used to train this network hence its high
reported classification scores. Another aspect to take into consideration is
the speed in which the models can predict on new data (inference). While VGG16
was the best performing model, its size slows down inference time. In this
regard, MobileNetV2 offers real-time inference speed at 10Hz for small image
sizes (for example $512\times 512$px). Consequently, a model based on
MobileNetV2 could be used to detect active wave breaking in an operational
fashion in, for example, offshore floating wind turbines that are susceptible
to wave breaking impacts and used to adjust anchor and cable settings in real-
time.
From the analysis of the application of the method on real-world data, it was
visually observed that small active wave breaking events were not always
detected, particularly for the Black Sea data. There are two possible
explanations for this error. The simplest explanation could be that this type
of events is under-represented in the training dataset. The solution for this
issue consists of adding more data representative of these instances to the
training dataset. The second possibility is that the image size becomes too
small in the deeper layers of the network which makes it impossible for the
network to learn these events (see below for further discussion). A solution
for this issue could be to increase the input image size but this was not
attempted here due to hardware constraints (that is, memory limitation, as
discussed above).
Neural networks have been historically seen as black boxes in which only the
final classification outputs are relevant. There has been, however, an
increase in interest to understanding how neural networks are learning. One
technique that is of particular interest for the present paper is the concept
of Gradient-weighted Class Activation Mapping (Grad-CAM)[59]. Briefly, this
technique shows which regions of a convolutional layer are more important for
the results obtained by the classifier. Figure 6 shows the results of Grad-CAM
applied to examples for all unique locations from Table 1 considering VGG16’s
last convolutional layer. When considering only actively breaking waves
(Figure 6-a to d) it is evident that VGG16 closely mimicked how a human would
classify these data, that is, it directly searched for the regions of the
image that corresponded to active wave breaking. In the case of passive foam
(Figure 6-e to h), VGG16 seemed to use larger portions of the image but at the
same time focused on the flocculent foam seen in the images as a human
classifier would do. In general, these results show that our model truly
learned how to classify the images and is not merely guessing the labels.
Figure 6: Results of Grad-CAM [59] for all unique experiment locations
described in Table 1 applied to actively breaking waves (top row) and to
passive foam (bottom row). a) Actively breaking wave example recorded at
Adriatic Sea (2015/03/05 10:35). b) Actively breaking wave example recorded at
the Black Sea (2011/10/04 11:07). c) Actively breaking wave example recorded
at La Jument (2018/01/03 09:39). d) Actively breaking wave example recorded at
the Yellow Sea (2017/05/13 05:00). e) Passive foam example recorded at
Adriatic Sea (2015/03/05 10:35). f) Passive foam example recorded at the Black
Sea (2011/10/04 11:07). g) Passive foam example recorded at La Jument
(2018/01/03 09:39). h) Passive foam example recorded at the (2017/05/13
05:00). In all panels, the color scale indicates the weights of the class
activation map with brighter colors showing regions of the image which the
neural network used to classify each particular sample.
The promising results presented here indicate that the current method should
be extended to an object detection framework. Such a framework would eliminate
the need for the image thresholding and DBSCAN steps. This implementation
could be done by using a strongly-supervised architecture such as UNet [60],
or by using a weakly-supervised method derived from Grad-CAM, for example. As
a final recommendation regarding the machine learning aspect of this paper, we
strongly encourage future researchers to add samples to the training dataset
that matches their specific research needs instead of blindly applying the
provided pre-trained models.
We have also presented examples of wave breaking statistics that can be
obtained using the proposed method. In general, the patterns observed here
agreed with previously reported scaling factors that support the idea of wave
breaking self-similarity. For example, Figure 5-f directly showed that the
scaling parameter $A_{br}/b^{2}$ approaches the constant 0.1 value from
Duncan’s [56] laboratory experiments. Another variable that showed signs of a
self-similar behavior was the wave breaking area ($A_{br}$) which was very
well described by a Pareto distribution and presented a steady quadratic
increase with wave breaking duration ($T_{br}$). Extensive research [11, 15]
has been grounded on the assumption that wave breaking is self-similar, but
inconsistencies with this approach have been reported before [14]. Contrarily
to other studies [11, 31], however, the $\Lambda(c)$ distributions obtained
here did not closely match the theoretical $c^{-6}$ for a sea-state in
equilibrium. As reported before [13], these differences may be due to the fact
that here we only considered actively breaking waves for our analysis whereas
other studies seem to track the speeds of both actively breaking waves and
passive foam [32, 31]. Another possibility is that other phenomena not
accounted by Phillips’[12] theory (for example, wave modulation[61]) play
important role on wave breaking. A future publication that investigates the
mechanisms related to the wave breaking kinematics using the method and data
obtained here will soon follow.
## Conclusion
We described a novel method to detect and classify actively breaking waves in
video imagery data. Our method achieved promising results when assessed using
several different metrics. Further, by analyzing the deeper layers of our
neural network, we showed that the model mimicked how a human classifier would
perform a similar classification task. As an application of the method, we
presented wave breaking statistics and breaking wave crest length
distributions. Our method can thus be useful for the investigation of several
ocean-atmosphere interaction processes and should, in the future, lead to
advancements in operational wave forecast models, gas-exchange models, and
general safety at seas. Finally, we strongly recommend that future research
should focus on standardized methods to study wave breaking so that a
consistent dataset can be generated and used freely and unambiguously by the
community.
## References
* [1] Battjes, J. A. & Janssen, J. Energy loss and set-up due to breaking of random waves. _Coastal Engineering_ 32, 569–587 (1978).
* [2] Thornton, E. B. & Guza, R. T. Transformation of Wave Height Distribution. _Journal of Geophysical Research_ 88, 5925–5938 (1983).
* [3] Banner, M. L., Babanin, A. V. & Young, I. R. Breaking Probability for Dominant Waves on the Sea Surface. _Journal of Physical Oceanography_ 30, 3145–3160, DOI: 10.1175/1520-0485(2000)030<3145:BPFDWO>2.0.CO;2 (2000).
* [4] Banner, M. L., Gemmrich, J. R. & Farmer, D. M. Multiscale measurements of ocean wave breaking probability. _Journal of Physical Oceanography_ 32, 3364–3375, DOI: 10.1175/1520-0485(2002)032<3364:MMOOWB>2.0.CO;2 (2002).
* [5] Cavaleri, B. Y. L. Wave Modeling: Where to Go in the Future. _Bulletin of American Meteorological Society_ 207–2014, DOI: 10.1175/BAMS-87-2-207 (2006).
* [6] The WAVEWATCH Development Group (WW3DG). User manual and system documentation of WAVEWATCH III R version 6.07. Tech. Rep., NOAA/NWS/NCEP/MMAB, College Park, MD, USA (2019).
* [7] Booij, N., Ris, R. C. & Holthuijsen, L. H. A third-generation wave model for coastal regions 1 . Model description and validation. _Journal of Geophysical Research_ 104, 7649–7666 (1999).
* [8] The WANDI Group. The WAN Model - A Third Generation Ocean Wave Prediction Model. _Journal of Physical Oceanography_ 18, 1775–1810 (1988).
* [9] Filipot, J. F., Ardhuin, F. & Babanin, A. V. A unified deep-to-shallow water wave-breaking probability parameterization. _Journal of Geophysical Research: Oceans_ 115, 1–15, DOI: 10.1029/2009JC005448 (2010).
* [10] Stringari, C. E. & Power, H. E. The Fraction of Broken Waves in Natural Surf Zones. _Journal of Geophysical Research: Oceans_ 124, 1–27, DOI: 10.1029/2019JC015213 (2019). arXiv:1904.06821v1.
* [11] Melville, W. K. & Matusov, P. Distribution of breaking waves at the ocean surface. _Nature_ 417, 58–63, DOI: 10.1038/417058a (2002).
* [12] Phillips, O. M. Spectral and statistical properties of the equilibrium range in wind-generated gravity waves. _Journal of Fluid Mechanics_ 156, 505–531, DOI: 10.1017/S0022112085002221 (1985).
* [13] Banner, M. L., Zappa, C. J. & Gemmrich, J. R. A note on the Phillips spectral framework for ocean whitecaps. _Journal of Physical Oceanography_ 44, 1727–1734, DOI: 10.1175/JPO-D-13-0126.1 (2014).
* [14] Gemmrich, J. R., Banner, M. L. & Garrett, C. Spectrally resolved energy dissipation rate and momentum flux of breaking waves. _Journal of Physical Oceanography_ 38, 1296–1312, DOI: 10.1175/2007JPO3762.1 (2008).
* [15] Romero, L. Distribution of Surface Wave Breaking Fronts. _Geophysical Research Letters_ 46, 10463–10474, DOI: 10.1029/2019GL083408 (2019).
* [16] Yurovsky, Y. Y., Kudryavtsev, V. N., Chapron, B. & Grodsky, S. A. Modulation of Ka-Band Doppler Radar Signals Backscattered from the Sea Surface. _IEEE Transactions on Geoscience and Remote Sensing_ 56, 2931–2948, DOI: 10.1109/TGRS.2017.2787459 (2018).
* [17] Huang, L., Liu, B., Li, X., Zhang, Z. & Yu, W. Technical evaluation of Sentinel-1 IW mode cross-pol radar backscattering from the ocean surface in moderate wind condition. _Remote Sensing_ 9, 1–21, DOI: 10.3390/rs9080854 (2017).
* [18] Hwang, P. A., Zhang, B., Toporkov, J. V. & Perrie, W. Comparison of composite Bragg theory and quad-polarization radar backscatter from RADARSAT-2: With applications to wave breaking and high wind retrieval. _Journal of Geophysical Research: Oceans_ 115, 1–12, DOI: 10.1029/2009JC005995 (2010).
* [19] Monahan, E. C. Oceanic whitecaps: Sea surface features detectable via satellite that are indicators of the magnitude of the air-sea gas transfer coefficient. _Journal of Earth System Science_ 111, DOI: 10.1007/BF02701977 (2002).
* [20] Reul, N. & Chapron, B. A model of sea-foam thickness distribution for passive microwave remote sensing applications. _Journal of Geophysical Research C: Oceans_ 108, 19–1, DOI: 10.1029/2003jc001887 (2003).
* [21] Carini, R. J., Chickadel, C. C., Jessup, A. T. & Thompson, J. Estimating wave energy dissipation in the surf zone using thermal infrared imagery. _Journal of Geophysical Research: Oceans_ 120, 3937–3957, DOI: 10.1002/2014JC010561.Received (2015).
* [22] Wang, S. _et al._ Improving the Upper-Ocean Temperature in an Ocean Climate Model (FESOM 1.4): Shortwave Penetration Versus Mixing Induced by Nonbreaking Surface Waves. _Journal of Advances in Modeling Earth Systems_ 11, 545–557, DOI: 10.1029/2018MS001494 (2019).
* [23] Komori, S. _et al._ Laboratory measurements of heat transfer and drag coefficients at extremely high wind speeds. _Journal of Physical Oceanography_ 48, 959–974, DOI: 10.1175/JPO-D-17-0243.1 (2018).
* [24] Buscombe, D. & Carini, R. J. A data-driven approach to classifying wave breaking in infrared imagery. _Remote Sensing_ 11, 1–10, DOI: 10.3390/RS11070859 (2019).
* [25] Buscombe, D., Carini, R. J., Harrison, S. R., Chickadel, C. C. & Warrick, J. A. Optical wave gauging using deep neural networks. _Coastal Engineering_ 155, 103593, DOI: 10.1016/j.coastaleng.2019.103593 (2020).
* [26] Kim, J., Kim, J., Kim, T., Huh, D. & Caires, S. Wave-tracking in the surf zone using coastal video imagery with deep neural networks. _Atmosphere_ 11, 1–13, DOI: 10.3390/atmos11030304 (2020).
* [27] Bieman, J. P. D., Ridder, M. P. D. & Gent, M. R. A. V. Deep learning video analysis as measurement technique in physical models. _Coastal Engineering_ 158, 103689, DOI: 10.1016/j.coastaleng.2020.103689 (2020).
* [28] Zheng, C. W., Chen, Y. G., Zhan, C. & Wang, Q. Source Tracing of the Swell Energy: A Case Study of the Pacific Ocean. _IEEE Access_ 7, 139264–139275, DOI: 10.1109/ACCESS.2019.2943903 (2019).
* [29] Filipot, J.-F. _et al._ La Jument lighthouse: a real-scale laboratory for the study of giant waves and their loading on marine structures. _Philosophical Transactions of the Royal Society A: Mathematical, Physical and Engineering Sciences_ 377, 20190008, DOI: 10.1098/rsta.2019.0008 (2019).
* [30] Mironov, A. S. & Dulov, V. A. Detection of wave breaking using sea surface video records. _Measurement Science and Technology_ 19, DOI: 10.1088/0957-0233/19/1/015405 (2008).
* [31] Sutherland, P. & Melville, W. K. Field measurements and scaling of ocean surface wave-breaking statistics. _Geophysical Research Letters_ 40, 3074–3079, DOI: 10.1002/grl.50584 (2013).
* [32] Kleiss, J. M. & Melville, W. K. Observations of wave breaking kinematics in fetch-limited seas. _Journal of Physical Oceanography_ 40, 2575–2604, DOI: 10.1175/2010JPO4383.1 (2010).
* [33] Bradski, G. The OpenCV Library. _Dr. Dobb’s Journal of Software Tools_ (2000).
* [34] Ester, M., Kriegel, H.-P., Sander, J. & Xu, X. Density-Based Clustering Methods. _Comprehensive Chemometrics_ 2, 635–654, DOI: 10.1016/B978-044452701-1.00067-3 (1996). 10.1.1.71.1980.
* [35] Moshtagh, N. Minimum Volume Enclosing Ellipsoids. _Convex Optimization_ 111, 112–118 (2005).
* [36] Guimaraes, P. V. _et al._ A Data Set of Sea Surface Stereo Images to Resolve Space-Time Wave Fields. _Scientific Data_ 7, 1–12, DOI: 10.12770/af599f42-2770-4d6d-8209-13f40e2c292f (2020).
* [37] Hastie, T., Tibshirani, R. & Friedman, J. _The Elements of Statistical Learning_ (Springer International Publishing, 2009), second edi edn.
* [38] Shorten, C. & Khoshgoftaar, T. M. A survey on Image Data Augmentation for Deep Learning. _Journal of Big Data_ 6, DOI: 10.1186/s40537-019-0197-0 (2019).
* [39] Simonyan, K. & Zisserman, A. Very deep convolutional networks for large-scale image recognition. _3rd International Conference on Learning Representations, ICLR 2015 - Conference Track Proceedings_ 1–14 (2015). arXiv:1409.1556v6.
* [40] He, K., Zhang, X., Ren, S. & Sun, J. Identity mappings in deep residual networks. _Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)_ 9908 LNCS, 630–645, DOI: 10.1007/978-3-319-46493-0_38 (2016). 1603.05027.
* [41] Szegedy, C., Ioffe, S., Vanhoucke, V. & Alemi, A. A. Inception-v4, inception-ResNet and the impact of residual connections on learning. _31st AAAI Conference on Artificial Intelligence, AAAI 2017_ 4278–4284 (2017). 1602.07261.
* [42] Howard, A. G. _et al._ MobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications (2017). 1704.04861.
* [43] Tan, M. & Le, Q. V. EfficientNet: Rethinking model scaling for convolutional neural networks. _36th International Conference on Machine Learning, ICML 2019_ 2019-June, 10691–10700 (2019). 1905.11946.
* [44] Ioffe, S. & Szegedy, C. Batch normalization: Accelerating deep network training by reducing internal covariate shift. _32nd International Conference on Machine Learning, ICML 2015_ 1, 448–456 (2015). 1502.03167.
* [45] Srivastava, N., Hinton, G., Krizhevsky, A., Sutskever, I. & Salakhutdinov, R. Dropout: A Simple Way to Prevent Neural Networks from Overfitting. _Journal of Machine Learning Research_ 15, 1929–1958, DOI: 10.1109/ICAEES.2016.7888100 (2014).
* [46] Kingma, D. P. & Ba, J. Adam: A Method for Stochastic Optimization. In _3rd International Conference for Learning Representations_ , 1–15, DOI: http://doi.acm.org.ezproxy.lib.ucf.edu/10.1145/1830483.1830503 (San Diego, California, 2014). 1412.6980.
* [47] Ruder, S. An overview of gradient descent optimization algorithms (2016). 1609.04747.
* [48] Abadi, M. _et al._ TensorFlow: Large-scale machine learning on heterogeneous systems (2015). Software available from tensorflow.org.
* [49] Nagi, J. _et al._ Max-Pooling Convolutional Neural Networks for Vision-based Hand Gesture Recognition. In _2011 IEEE International Conference on Signal and Image Processing Applications, ICSIPA 2011_ , 342–347 (2011).
* [50] Hanson, J. L. & Jensen, R. Wave system diagnostics for numerical wave models. _8 th International Workshop on Wave Hindcasting and Forecasting, Oahu, Hawaii, November_ (2004).
* [51] Large, W. G. & Pond, S. Open Ocean Momentum Flux Measurements in Moderate to Strong Winds. _Journal of Physical Oceanography_ DOI: 10.1175/1520-0485(1981)0112.0.CO;2 (1981).
* [52] Bewley, A., Ge, Z., Ott, L., Ramos, F. & Upcroft, B. Simple online and realtime tracking. _Proceedings - International Conference on Image Processing, ICIP_ 2016-August, 3464–3468, DOI: 10.1109/ICIP.2016.7533003 (2016). 1602.00763.
* [53] Holman, R. A. & Stanley, J. The history and technical capabilities of Argus. _Coastal Engineering_ 54, 477–491, DOI: 10.1016/j.coastaleng.2007.01.003 (2007).
* [54] Schwendeman, M., Thomson, J. & Gemmrich, J. R. Wave breaking dissipation in a Young Wind Sea. _Journal of Physical Oceanography_ 44, 104–127, DOI: 10.1175/JPO-D-12-0237.1 (2014).
* [55] Guimaraes, P. V. _Sea surface and energy dissipation_. Ph.D. thesis, Universitè de Bretagne Loire (2018).
* [56] Duncan, J. H. An Experimental Investigation of Breaking Waves Produced by a Towed Hydrofoil. _Proceedings of the Royal Society A: Mathematical, Physical and Engineering Sciences_ 377, 331–348, DOI: 10.1098/rspa.1981.0127 (1981).
* [57] Virtanen, P. _et al._ SciPy 1.0: Fundamental Algorithms for Scientific Computing in Python. _Nature Methods_ DOI: https://doi.org/10.1038/s41592-019-0686-2 (2020).
* [58] Russakovsky, O. _et al._ ImageNet Large Scale Visual Recognition Challenge. _International Journal of Computer Vision_ 115, 211–252, DOI: 10.1007/s11263-015-0816-y (2015). 1409.0575.
* [59] Selvaraju, R. R. _et al._ Grad-CAM: Visual Explanations from Deep Networks via Gradient-Based Localization. _International Journal of Computer Vision_ 128, 336–359, DOI: 10.1007/s11263-019-01228-7 (2020). 1610.02391.
* [60] Ronneberger, O., P.Fischer & Brox, T. U-net: Convolutional networks for biomedical image segmentation. In _Medical Image Computing and Computer-Assisted Intervention (MICCAI)_ , vol. 9351 of _LNCS_ , 234–241 (Springer, 2015). (available on arXiv:1505.04597 [cs.CV]).
* [61] Donelan, M. A., Haus, B. K., Plant, W. J. & Troianowski, O. Modulation of short wind waves by long waves. _Journal of Geophysical Research: Oceans_ 115, 1–12, DOI: 10.1029/2009JC005794 (2010).
## Acknowledgements
This work benefited from France Energies Marines and State financing managed
by the National Research Agency under the Investments for the Future program
bearing the reference numbers ANR-10-IED-0006-14 and ANR-10-IEED-0006-26 for
the projects DiME and CARAVELE.
## Author contributions statement
C.E.S. idealized and implemented the method, C.E.S. drafted the manuscript,
P.V.G., F.L., J.F.F. and R. D. conducted the field experiments, C.E.S. and
P.V.G. analyzed the data, J.F. provided funding. All authors reviewed the
manuscript.
## Additional information
Data, pre-trained networks and auxiliary programs necessary to reproduce the
results in this paper are available at:
https://github.com/caiostringari/deepwaves.
## Competing Interests Statement
The authors declare no known competing interests.
|
# Kähler geometry of quiver varieties and machine learning
George Jeffreys Department of Mathematics and Statistics, Boston University,
111 Cummington Mall, Boston MA 02215, USA<EMAIL_ADDRESS>and Siu-Cheong Lau
Department of Mathematics and Statistics, Boston University, 111 Cummington
Mall, Boston MA 02215, USA<EMAIL_ADDRESS>
###### Abstract.
We develop an algebro-geometric formulation for neural networks in machine
learning using the moduli space of framed quiver representations. We find
natural Hermitian metrics on the universal bundles over the moduli which are
compatible with the GIT quotient construction by the general linear group, and
show that their Ricci curvatures give a Kähler metric on the moduli. Moreover,
we use toric moment maps to construct activation functions, and prove the
universal approximation theorem for the multi-variable activation function
constructed from the complex projective space.
## 1\. Introduction
Machine learning by artificial neural networks has made exciting developments
and has been applied to many branches of science in recent years.
Mathematically, stochastic gradient flow over a matrix space (or called the
weight space) is the central tool. The non-convex nature of the cost function
has made the problem very interesting. Current research has focused on
different types of stochastic gradient flows and finding new types of
networks, which have brought great improvements of computational efficiency.
In geometry and physics, the applications of gradient flow and Morse theory
have a long history and have brought numerous fundamental breakthroughs. For
instance, the gradient flow of the Yang-Mills functional is used to find
Hermitian Yang-Mills connections, whose existence in a stable holomorphic
vector bundle is proved by Donaldson [Don85] and Uhlenbeck-Yau [UY86]. The
celebrated Ricci flow found by Hamilton [Ham82], which is a crucial tool to
solve the three-dimensional Poincaré conjecture, is essentially a gradient
flow [Per02, Per03]. Its Kähler analog has been an important tool in finding
Kähler-Einstein metrics on Fano manifolds [Yau96, Tia97, Don12, CSW18, CDS15a,
CDS15b, CDS15c]. In these works, GIT quotients and finite-dimensional models
have provided important motivations and guidelines [Don99]. Hamiltonian Floer
theory [Flo89], which is essentially Morse theory on the loop space, was
invented to solve the Arnold conjecture [FHS95, Ono95, FO99]. Various versions
of Floer theory have been crucial ingredients in the study of mirror symmetry.
In this paper, we would like to develop a foundational _algebro-geometric
formulation_ for neural networks in machine learning. The theory of _quiver
representations_ , which is a well-developed branch of mathematics motivated
from Lie theory and has been an important tool in mathematical physics, will
be well suited for this purpose.
A quiver representation assigns to a directed graph $Q$ a bunch of vector
spaces for the vertices and a bunch of linear maps for the arrows. Such a
construction is in common with neural networks. However, in order to use
quiver theory to formulate machine-learning neural networks, there are two
main differences between these two subjects that needs to be addressed.
1. (1)
Compactness of moduli space. A moduli space of quiver representations [Kin94]
is defined by identifying _isomorphic_ quiver representations using GIT
quotients. As a result, the moduli space is _compact_ when the quiver has no
oriented cycle. On the other hand, the matrix space used in neural networks is
non-compact. In machine learning, isomorphic quiver representations may
correspond to physically different input or output information and in general
cannot be identified.
2. (2)
Non-linearity. _Activation functions_ , which are non-linear maps on the
vector spaces over the vertices, serve as a crucial ingredient to achieve
machine learning of non-linear functions. Such non-linearity jumps out of the
category of quiver representations. This is also related to the first point
above. Namely, such non-linear maps are not necessarily equivariant under the
group of automorphisms of quiver representations.
For the first point, we shall use framed quiver representations, which were
first found by Nakajima [Nak94] in the study of affine Lie algebras. A framed
representation assigns to each vertex a vector space together with a choice of
‘framing’ (for instance it is a basis in the simplest situation). In the
applications considered here, such a decoration makes sure that isomorphic
framed quiver representations correspond to the same physical state. Note that
framed quiver moduli $\mathcal{M}$ are also compact when $Q$ has no oriented
cycle. _Compactness_ is one of the main advantages of our algebro-geometric
formulation, which makes sure the convergence of a gradient flow.
In this formulation, the weight matrices are encoded as morphisms between the
_universal vector bundles_ (over the framed quiver moduli) associated to the
vertices. The data flow is encoded by sections of the universal bundles, which
are sent from one to another bundles by the morphisms associated to the arrows
of $Q$. The cost function, and hence its gradient flow, is defined on the
framed quiver moduli $\mathcal{M}$.
In particular, the _critical points and the gradient flow are controlled by
the topology of $\mathcal{M}$_ (for instance, the Morse inequalities). The
topology of a framed quiver moduli is well-understood by the work of Reineke
[Rei08] when $Q$ has no oriented cycle. $\mathcal{M}$ is an iterated Grassmann
bundle, and its Poincaré polynomial is a product of that of the Grassmannians.
For the purpose of gradient flow, one needs to choose a _Kähler metric_ on
$\mathcal{M}$, and also Hermitian metrics on the universal vector bundles. As
a result, we have found metrics that are defined by explicit beautiful
formulae. These metrics are not just $U_{\vec{d}}$-equivariant so that they
descend to symplectic quotients, but are also
$\mathrm{GL}_{\vec{d}}$-equivariant and hence compatible with the GIT
construction of $\mathcal{M}$. Moreover, they are compatible with the iterated
Grassmann structure found by Reineke. In application, such metrics would
simplify the actual computational algorithm over the quiver moduli. They are
summarized as follows.
###### Theorem 1.1 (Combining Theorem 3.7, 3.15,3.18).
Let $Q$ be an arbitrary quiver. Fix a vertex $i\in Q_{0}$. Let $\rho$ be the
row vector whose entries are $V_{\gamma}e^{\left(t(\gamma)\right)}$, where
$\gamma$ is any path whose head $h(\gamma)$ is $i$ (including the trivial
path), $t(\gamma)$ denotes its tail, and
$V_{\gamma}\in\mathrm{Hom}(\mathbb{C}^{d_{t(\gamma)}},\mathbb{C}^{d_{h(\gamma)}})$
is the representing matrix of $\gamma$. Then
$(\rho_{i}\rho_{i}^{*})^{-1}=\left(\sum_{h(\gamma)=i}\left(V_{\gamma}e^{\left(t(\gamma)\right)}\right)\left(V_{\gamma}e^{\left(t(\gamma)\right)}\right)^{*}\right)^{-1}$
is $\mathrm{GL}_{\vec{d}}$-equivariant, and it descends to a metric on the
universal bundle $\mathcal{V}_{i}$ over a certain domain of convergence
$\mathcal{M}^{\circ}$.
When $Q$ has no oriented cycle, $\mathcal{M}^{\circ}=\mathcal{M}$. Moreover,
the Ricci curvature of the induced metric on $\bigotimes_{i\in
Q_{0}}\mathcal{V}_{i}$ gives a Kähler metric on $\mathcal{M}$.
The precise definition of $M^{\circ}$ is given in Section 3.3.
In this paper, we focus on the framed quiver moduli defined over complex
numbers. In actual applications, we can also restrict to real coefficients.
Then the above formula defines a bundle metric over
$\mathcal{M}_{\mathbb{R}}$, and the Ricci curvature gives a Riemannian metric
on $\mathcal{M}_{\mathbb{R}}$.
Now let us address the second point. Namely, we need to introduce _non-
linearity_ in addition to the usual theory of quiver representations. By
definition, morphisms between universal vector bundles over $\mathcal{M}$ are
linear along fibers. They correspond to weight matrices in neural networks. To
introduce non-linearity, we shall treat the universal bundles as fiber bundles
and construct suitable _fiber-bundle maps_ that play the role of activation
functions.
One of the commonly-used activation functions is
$\frac{e^{2x}}{1+e^{2x}}:\mathbb{R}\to(0,1).$
We observe that this function also appears in the base of the symplectic
trivialization of the open dense toric orbit of $\mathbb{P}^{1}$ as a toric
variety:
$(\mathbb{C}^{\times},\omega_{\mathbb{C}\mathbb{P}^{1}})\cong((0,1)\times\mathbb{S}^{1},\omega_{\mathrm{std}}),$
or lifted to the universal cover:
$(\mathbb{C},\pi^{*}\omega_{\mathbb{C}\mathbb{P}^{1}})\cong((0,1)\times\mathbb{R},\omega_{\mathrm{std}}).$
Here, $\omega_{\mathbb{C}\mathbb{P}^{1}}$ denotes the Fubini-Study metric of
$\mathbb{C}\mathbb{P}^{1}$, that is, the standard area form of the unit
sphere; $\omega_{\mathrm{std}}=dx\wedge dy$ is the standard symplectic form.
Similarly, another activation function
$\frac{z}{\sqrt{1+|z|^{2}}}:\mathbb{C}\to\mathbb{C}$
also arises as a symplectic trivialization:
$(\mathbb{C},\omega_{\mathbb{C}\mathbb{P}^{1}})\cong(\\{w\in\mathbb{C}:|w|<1\\},\omega_{\mathbb{C}}).$
Motivated from these observations, we consider
$\sigma(x)=\left(\frac{e^{2x_{i}}}{1+\sum_{j=1}^{n}e^{2x_{j}}}\right)_{i=1}^{n}:\mathbb{R}^{n}\to\Delta$
and
$\psi(\vec{z})=\left(\frac{z_{i}}{\sqrt{1+\|\vec{z}\|^{2}}}\right)_{i=1}^{n}:\mathbb{C}^{n}\to\\{\vec{w}\in\mathbb{C}^{n}:\|\vec{w}\|<1\\}$
as multi-variable activation functions, where $\Delta$ denotes the standard
simplex with vertices $0$ and $\epsilon_{1},\ldots,\epsilon_{n}$, the standard
basis of $\mathbb{R}^{n}$. They arise from symplectomorphisms
$(\mathbb{C}^{n},\omega_{\mathbb{C}\mathbb{P}^{n}})\cong(\\{\vec{w}\in\mathbb{C}^{n}:\|\vec{w}\|<1\,\omega_{\mathbb{C}^{n}}\\})$.
More generally, these come from moment maps of _toric varieties_ [Gui94,
Abr98]. We note that $\psi$ has an advantage of being $U(n)$-equivariant.
The _universal approximation theorem_ (see for instance [Cyb89, Pet99, MM92,
Pin99]) provides a theoretical foundation for the success of neural networks.
In existing literature, the theorem was proved for single-variable activation
functions.
In this paper, we prove the universal approximation theorem for the above
multivariable function $\sigma=\sigma_{\mathbb{R}^{n}}$. Note that $\sigma$ is
the softmax function
$\left(\frac{e^{2x_{i}}}{\sum_{j=0}^{n}e^{2x_{j}}}\right)_{i=0}^{n}$
restricted to the hyperplane $x_{0}=0$ and composed with the projection along
$x_{0}$-direction. We shall restrict to real coefficients in this theorem.
###### Theorem 1.2 (same as Theorem 5.2).
Let $K$ be a compact set of $\mathbb{R}^{d_{1}}$, and
$f:K\to\mathbb{R}^{d_{3}}$ a continuous function. For any $\epsilon>0$, there
exists $d_{2}>0$ and $W_{1}\in\operatorname{Mat}(d_{2},d_{1})$,
$W_{2}\in\operatorname{Mat}(d_{3},d_{2})$, $b\in\mathbb{R}^{d_{2}}$ such that
$\|f^{U}_{W_{1},W_{2},b}-f\|_{L^{2}(K)}<\epsilon$. Here,
$f^{U}_{W_{1},W_{2},b}(x)=W_{2}\cdot\sigma_{\mathbb{R}^{d_{1}}}(W_{1}\cdot
x+b)$ is the function coming from the $A_{3}$ quiver.
The $A_{3}$ quiver corresponds to the feed-forward network with one input
layer, one middle layer and one output layer. See Figure 1.
${a}$${b}$${c}$$\scriptstyle{\alpha}$$\scriptstyle{\beta}$ Figure 1. The
$A_{3}$ quiver.
The above theorem is proved by using the tropical limit of the toric manifold
$\mathbb{P}^{n}$, and a geometric object that we call a centered polyhedral
web, which is an analog of a tropical variety in an integral affine manifold.
Since we do not have integral structure in the context here, we need to invent
this new notion.
In above, we have focus on explaining non-linearity for a single vector space.
We shall _globalize them as non-linear fiber-bundle morphisms_ for the
universal bundles over $\mathcal{M}$. This can be achieved with the help of
Hermitian metrics on the universal bundles, so that the Fubini-Study metric on
$\mathbb{P}^{n}$ can be globalized as a fiberwise symplectic structure on
projective bundles over $\mathcal{M}$. Actually, the globalization from a
single framing vector space $V$ to the universal fiber bundle over
$\mathcal{M}$ _works for any continuous function_ $V\to V$ (and in particular
for a symplectomorphism from $V$ to its image). Combining the ingredients
explained above together, we can construct a gradient flow over the framed
quiver moduli to achieve machine learning. The detail is given in Section 4.
Such an algebro-geometric formulation has several _advantages_. First, the
gradient flow under consideration runs in a _compact_ manifold. This ensures
the existence of absolute extrema, convergence of the flow, and upper bound
for the norm of the gradient vector field. Second, because of compactness, the
flow is constrained by _topology_ of the manifold due to Morse theory. See
Section 4.4. Finally, the moduli space has extra _symmetry_ coming from
framing. If we use activation functions that respect this symmetry (for
instance $\psi$ above enjoys $U(n)$-equivariance), we can perform dimension
reduction which improves the effectiveness of the network. (See Proposition
4.16 and 4.17.)
In summary, from this point of view, the success of neural network is resulted
from the interplay between algebraic morphisms and (transcendental)
symplectomorphisms. Interestingly, such an interplay is also an important
feature that occurs in the study of complete integrable systems and mirror
symmetry for toric manifolds and flag varieties, see for instance [Gui94,
FLTZ12, Abo06, CLL12, NNU10, HKL18].
### Some related works
The relation between neural network and quiver representation was investigated
in the recent paper [AJ20]. Their work considered the quotient space by
$(\mathbb{C}^{\times})^{N}$ of pairs $(W,f)$, where $W$ is a quiver
representation of $Q$ with the dimension vector $\vec{1}$, and $f$ associates
each vertex a function $\mathbb{C}\to\mathbb{C}$ (playing the role of an
activation function). Moreover, in dimension $\vec{1}$ (which is a typical
case for machine learning), they invented an interesting way of encoding the
data flow as a quiver representation. (In our work, the data flow is given as
sections of universal bundles over the quiver moduli.)
The approach and the goal of this paper is rather different. We aim at
formulating machine learning as a gradient flow over a compact quiver moduli.
In [AJ20], quiver representations were used in encoding the data in the
network; however, the machine learning process was independent of the quiver
moduli. Moreover, ‘double-framing’ was used, and the corresponding moduli
space is non-compact.
The map $(W,f)\mapsto W$ gives an infinite-dimensional fiber bundle over the
quiver moduli $\mathcal{M}$, whose fibers are the spaces of choices of
activation functions. In a typical program of machine learning, the activation
functions are fixed during the optimization process. In order to formulate the
program as a gradient flow over the compact moduli $\mathcal{M}$, we found a
non-trivial way by equipping intermediate vertices with additional framings
and metrics, so that we can lift $f$ to be a well-defined fiber-bundle map
over $\mathcal{M}$. Note that $f$ is not equivariant under the group action of
$(\mathbb{C}^{\times})^{N}$ ($\mathrm{GL}_{\vec{d}}$ in the higher rank case).
Such a lifting is an important non-trivial step.
Furthermore, we have dealt with representations of general rank $\vec{d}$, and
a class of activation functions coming from toric symplectomorphisms.
Different functions (on the same domain and target) are obtained if we deform
the toric Kähler metric. To also optimize the activation functions during the
learning process (see also [GGL19]), we may consider a gradient flow on
$\mathcal{M}\times\mathcal{K}$ where $\mathcal{K}$ denotes the moduli of toric
Kähler metrics in the same class. By the celebrated works of [Don99, Sem92],
$\mathcal{K}$ is an infinite-dimensional negatively curved symmetric space.
Recently, there is a rising interest of applying geometric techniques to the
study of neural networks. For instance, in the works [GBH18, CYRL19],
hyperbolic spaces are applied to machine learning in graphs and achieved great
performance.
Moreover, the applications of symmetry and group equivariance in neural
networks were studied and developed in [CW16, CGW19, CGKW18, CWKW19, CAW+19,
dHCW20]. Overall, these works aim at capturing symmetry of the input data and
designing networks that are adapted to such symmetry. Moreover, homogeneous
spaces (in place of vector spaces) have been employed in layers of
convolutional neural networks.
In comparison, our paper aims at revealing the geometric nature of neural
networks and build a connection with algebraic geometry. The resulting framed
quiver moduli, which has interesting topology and metrics, is the main
geometric object of interest. Furthermore, we study activation functions that
respects the ‘intrinsic symmetry’ over the quiver moduli, which can provide a
more effective algorithm by dimension reduction.
In the reverse direction, there are interesting applications of machine
learning in frontier geometry and physics. For instance, [HY20] used machine
learning to solve problems in computing graph Laplacians, such as recognizing
graph Ricci-flatness and predicting the spectral gap. In physics, [HSTT18a,
HSTT18b] used deep learning to study AdS/CFT correspondence by discretizing
the equation of motion. Since we have formulated machine learning using quiver
representations, it will be interesting to find direct relations between
quiver gauge theory and these problems that can be attacked via machine
learning.
### Organization of this paper
In Section 2, we will take a quick review on quiver representations and their
moduli spaces. In Section 3, we will construct nice Hermitian metrics on
universal bundles over the moduli. For readers who are mainly interested in
machine learning, Section 3 can be skipped for the first reading. Then we give
an algebro-geometric formulation of neural network using quiver
representations in Section 4. In Section 5, we prove the universal
approximation theorem for the multivariable activation function $\sigma$.
## Acknowledgment
We are grateful to Marco Antonio Armenta for informing us about the work
[AJ20] and the further useful discussions. We express our gratitude to Shing-
Tung Yau for his generous encouragement. The work of S.C. Lau in this paper is
partially supported by the Simons collaboration grant.
## 2\. Review of framed quiver moduli
Let $Q$ be a directed graph. Denote by $Q_{0},Q_{1}$ the set of vertices and
arrows respectively. A quiver representation $V$ with dimension vector
$\vec{d}\in\mathbb{Z}_{\geq 0}^{Q_{0}}$ associates each arrow $a$ with a
matrix $V(a)$ of size $d_{h(a)}\times d_{t(a)}$ (where $h(a),t(a)$ denote the
head and tail vertices of $a$ respectively). The set of complex quiver
representations with dimension $\vec{d}$ form a vector space denoted by
$R_{\vec{d}}(Q)$. The moduli space of quiver representations is a GIT quotient
of $R_{\vec{d}}(Q)$ by the group of isomorphisms
$\mathrm{GL}(\vec{d})=\prod_{i\in Q_{0}}\mathrm{GL}(d_{i},\mathbb{C})$
[Kin94], where $\mathrm{GL}(\vec{d})$ acts on $R_{\vec{d}}(Q)$ via
$g\cdot(V(a):a\in Q_{1})=(g_{h(a)}\cdot V(a)\cdot g_{t(a)}^{-1}:a\in Q_{1}).$
(2.1)
In the applications we consider in this paper, since the vector space over the
input and output vertices are equipped with fixed basis with physical
meanings, we need to use framed quiver representations [Nak94, Nak01, CB03,
Rei08].
Let $\vec{d},\vec{n}\in\mathbb{Z}_{\geq 0}^{Q_{0}}$. $\vec{n}$ will be the
dimension vector for the framing, which is a linear map
$e^{(i)}:\mathbb{C}^{n_{i}}\to V_{i}$ at each $i\in Q_{0}$ (where
$V_{i}=\mathbb{C}^{d_{i}}$). Since we will take a quotient by
$\mathrm{GL}(\vec{d})$, we shall think of $V_{i}$ as a vector space without a
preferred basis, while $\mathbb{C}^{n_{i}}$ is equipped with the standard
basis.
###### Definition 2.1.
The vector space of framed representations is given by
$R_{\vec{n},\vec{d}}=R_{\vec{d}}\times\bigoplus_{i\in
Q_{0}}\mathrm{Hom}(\mathbb{C}^{n_{i}},\mathbb{C}^{d_{i}}).$
It carries a natural action of $\mathrm{GL}(\vec{d})$ given by
$g\cdot(V,e)=(g\cdot V,(ge^{(i)}:i\in Q_{0}))$, where $g\cdot V$ is given by
Equation (2.1).
We need to remove unstable framed representations from $R_{\vec{n},\vec{d}}$
in order to get a nice quotient by $\mathrm{GL}(\vec{d})$.
###### Theorem 2.2 ([Nak96]).
$(V,e)\in R_{\vec{n},\vec{d}}$ is called stable if there is no proper
subrepresentation $U$ of $V$ which contains $\mathrm{Im}\,e$. The set of all
stable points of $R_{\vec{n},\vec{d}}$ is denoted by
$R_{\vec{n},\vec{d}}^{s}$. Then the quotient
$\mathcal{M}_{\vec{n},\vec{d}}:=R_{\vec{n},\vec{d}}^{s}/\mathrm{GL}(\vec{d})$
is a smooth variety, which is called to be a framed quiver moduli.
Actually $\mathcal{M}_{\vec{n},\vec{d}}$ can be formulated as a GIT quotient
[CB03, Rei08]. Namely, by adding an extra vertex labeled as $\infty$ to the
quiver and $n_{i}$ arrows from the vertex $\infty$ to the vertex $i$, $(V,e)$
can be identified as a usual representation of this bigger quiver with the
dimension vector $(\vec{d},1)$. The above stability condition can be rewritten
as slope stability, and hence it is a GIT quotient [Kin94].
Since $(d,1)$ is a primitive vector, $\mathcal{M}_{\vec{n},\vec{d}}$ is a
smooth fine moduli. There are universal vector bundles $\mathcal{V}_{i}$ over
$\mathcal{M}_{\vec{n},\vec{d}}$ corresponding to each vertex $i$, with fibers
$\mathcal{V}_{i}|_{[V,e]}=V_{i}$.
###### Example 2.3.
For the quiver with a single vertex and no arrow, and $n>d$,
$\mathcal{M}_{n,d}=\operatorname{Gr}(n,d)=\\{e\in\mathrm{Hom}(\mathbb{C}^{n},\mathbb{C}^{d}):e\textrm{
is surjective}\\}/\mathrm{GL}_{\vec{d}}$
is the (dual) Grassmannian. We have the tautological bundle $\mathcal{V}$ over
$\operatorname{Gr}(n,d)$. (Note that this tautological bundle is dual to the
one on $\operatorname{Gr}(d,n)\cong\operatorname{Gr}(n,d)$.)
The topology of $\mathcal{M}_{\vec{n},\vec{d}}$ is well-understood. Let’s make
an ordering of the vertices. Namely the vertices are labeled by
$\\{1,\ldots,N\\}$, such that $i<j$ implies there is no arrow going from $j$
to $i$. Such a labeling exists if $Q$ has no oriented cycle.
###### Theorem 2.4 (Reineke [Rei08]).
Assume $Q$ has no oriented cycle. Consider the chain of iterated Grassmann
bundles $M^{(N)}\stackrel{{\scriptstyle
p_{N}}}{{\to}}M^{(N-1)}\stackrel{{\scriptstyle
p_{N-1}}}{{\to}}\ldots\stackrel{{\scriptstyle
p_{2}}}{{\to}}M^{(1)}\stackrel{{\scriptstyle p_{1}}}{{\to}}\mathrm{pt}$ (where
$\mathrm{pt}$ denotes a singleton) defined by induction:
$M^{(i)}=\operatorname{Gr}_{M^{(i-1)}}\left(\underline{\mathbb{C}^{n_{i}}}\oplus\bigoplus_{j\to
i}p_{i-1}^{*}\dots p_{j+1}^{*}(S_{j}),d_{i}\right)\to M^{(i-1)},$
where $S_{i}$ denotes the tautological bundle on $M_{i}$ (as a Grassmann
bundle over $M_{i-1}$). (The direct sum is over each arrow $j\to i$.) Then
$\mathcal{M}_{\vec{n},\vec{d}}\cong M^{(N)}$, with universal bundles
$\mathcal{V}_{i}\cong p_{N}^{*}\dots p_{i+1}^{*}S_{i}$ for all $i\in Q_{0}$.
###### Corollary 2.5 (Reineke [Rei08]).
The Poincare polynomial of nonempty $\mathcal{M}_{\vec{n},\vec{d}}$ is given
by
$\prod_{i\in Q_{0}}\binom{n_{i}+\sum_{j\to i}d_{j}}{d_{i}}_{q^{2}}$
where
$\binom{n}{d}_{q}=\prod_{k=1}^{d}\frac{q^{n-d+k}-1}{q^{k}-1}.$
###### Remark 2.6.
In [Rei08], the framing $e$ goes in the other direction (from
$\mathbb{C}^{d_{i}}$ to $\mathbb{C}^{n_{i}}$). The above theorem is stated in
the dual way, which is the convention we take for the rest of this paper.
###### Example 2.7.
Consider the $A_{3}$-quiver which has three vertices $i=1,2,3$ and two arrows
$a_{1}:1\to 2,a_{2}:2\to 3$. Suppose $n_{1}=d_{1}$, $n_{2}=d_{2}+1$ and
$n_{3}=d_{3}$. Then the iterated Grassmann bundle is $M^{(3)}\to M^{(2)}\to
M^{(1)}$, where $M^{(1)}=\operatorname{Gr}(d_{1},d_{1})=\mathrm{pt}$ (and its
tautological bundle is the vector space $\mathbb{C}^{d_{1}}$);
$M^{(2)}=\operatorname{Gr}(d_{2}+1+d_{1},d_{2})$ is equipped with the
tautological bundle $S_{2}$ of rank $d_{2}$;
$M^{(3)}=\operatorname{Gr}_{\operatorname{Gr}(d_{2}+1+d_{1},d_{2})}(\underline{\mathbb{C}}^{d_{3}}\oplus
S_{2},d_{3})$ is a Grassmannian bundle over
$\operatorname{Gr}(d_{2}+1+d_{1},d_{2})$ with fibers
$\operatorname{Gr}(d_{2}+d_{3},d_{3})$. The corresponding Poincare Polynomial
will be
$\left(\prod_{k=1}^{d_{3}}\frac{q^{2(d_{2}+k)}-1}{q^{2k}-1}\right)\left(\prod_{k=1}^{d_{2}}\frac{q^{2(d_{1}+1+k)}-1}{q^{2k}-1}\right).$
See Figure 2.
${1}$${Gr(d_{1},d_{1})=\mathrm{pt}}$${2}$${Gr(d_{2}+1+d_{1},d_{2})}$${3}$${\operatorname{Gr}_{\operatorname{Gr}(d_{2}+1+d_{1},d_{2})}(\underline{\mathbb{C}}^{d_{3}}\oplus
S_{2},d_{3})}$$\scriptstyle{a_{1}}$$\scriptstyle{a_{2}}$ Figure 2. The
iterated Grassmann bundles associated to the $A_{3}$ quiver.
## 3\. Hermitian Metric over framed quiver moduli
In constructing fiber-bundle endomorphisms, it will be crucial to consider
Kähler metrics on universal bundles. In this section, we find a beautiful
formula for the canonical metric on the universal bundle $\mathcal{V}_{i}$
over $\mathcal{M}_{n,d}$ written in homogeneous coordinates. Using this
formula, we then show that the sum of Ricci curvatures over the vertices $i$
give a Kähler metric on $\mathcal{M}_{\vec{n},\vec{d}}$.
First, let us begin by recalling the typical example $\operatorname{Gr}(n,k)$.
### 3.1. The Grassmannian
Consider
$\operatorname{Gr}\left(n,k\right)=\operatorname{Mat}^{\mathbb{C}}_{n,k}\sslash_{\chi=1}U(k)=\left\\{e\in\operatorname{Mat}^{\mathbb{C}}_{n,k}:ee^{*}=I_{k}\right\\}\big{/}U\left(k\right)$
for $n\geq k$. Here we have used the dual description which better matches the
frame convention used in this paper. Namely,
$\operatorname{Gr}\left(n,k\right)$ parametrizes $k$-dimensional quotient
vector spaces of a fixed $n$-dimensional vector space as opposed to
$k$-dimensional subspaces. The moment map for the standard $U(k)$-action on
$\operatorname{Mat}^{\mathbb{C}}_{n,k}$ is
$ee^{*}:\operatorname{Mat}^{\mathbb{C}}_{n,k}\to\mathbf{i}\mathfrak{u}_{k}$.
We have taken the moment map level $\chi=1$ in the above symplectic reduction.
Note that $U(k)$ is acting on the left, although in the above expression
$U(k)$ appears on the right.
Writing $e=\left(b,p\right)$ where $b\in\operatorname{Mat}^{\mathbb{C}}_{k,k}$
and $p\in\operatorname{Mat}^{\mathbb{C}}_{n-k,k}$, the moment-map equation
$ee^{*}=I_{k}$ becomes
$bb^{*}+pp^{*}=I_{k}.$
We shall consider the chart defined by
$U=\\{[b,p]\in\operatorname{Gr}(n,k):\det
b\not=0\\}\cong\operatorname{Mat}^{\mathbb{C}}_{n-k,k}$
where the identification is given by the holomorphic coordinates
$\zeta^{h}=b^{-1}p\in\operatorname{Mat}^{\mathbb{C}}_{n-k,k}.$
We also have the symplectic coordinates
$\zeta^{u}=\left(b^{*}b\right)^{\frac{1}{2}}b^{-1}p\in\operatorname{Mat}^{\mathbb{C}}_{n-k,k}.$
The entries of $\zeta^{u}$ are not meromorphic functions. On the other hand,
$\zeta^{u}$ has the advantage that it satisfies the moment-map equation
$b^{*}b+\zeta^{u}(\zeta^{u})^{*}=I_{k}.$ (3.1)
(Note that the first term is $b^{*}b$ instead of $bb^{*}$.)
The construction of $\zeta^{u}$ uses the polar decomposition
$b=\left(b\left(b^{*}b\right)^{-\frac{1}{2}}\right)\left(b^{*}b\right)^{\frac{1}{2}}$
where $\left(b\left(b^{*}b\right)^{-\frac{1}{2}}\right)\in U(k)$ and
$\left(b^{*}b\right)^{\frac{1}{2}}\in\mathbf{i}\mathfrak{u}(k)$ is positive
definite. We obtain the coordinates $\zeta^{u}$ by observing
$[b,p]=\left[\left(b^{*}b\right)^{\frac{1}{2}},\zeta^{u}\right]$
using the left-$U(k)$-action, such that the first component
$\left(b^{*}b\right)^{\frac{1}{2}}$ is Hermitian, and is determined
$\zeta^{u}$ due to the moment-map equation (3.1).
The two coordinate systems are related by
$\zeta^{u}=\left(b^{*}b\right)^{\frac{1}{2}}\cdot\zeta^{h}.$ (3.2)
Let $S$ be the tautological vector bundle whose fibers are the quotient vector
spaces. (This is dual to the tautological bundle of
$\operatorname{Gr}(k,n)\cong\operatorname{Gr}(n,k)$.) It can be written as the
quotient of the trivial bundle:
$S=\left(\left\\{ee^{*}=I_{k}\right\\}\times\mathbb{C}^{k}\right)\big{/}U\left(k\right)$
where the left action of $U(k)$ on $\mathbb{C}^{k}$ is the standard one.
We now take the standard metric on $\mathbb{C}^{k}$, which is preserved by
$U(k)$ and hence descends to a metric $H$ of $S$.
Denote the standard basis of $\mathbb{C}^{k}$ by $\epsilon_{j}$ for
$j=1,\ldots,k$. Under this metric, we have the lifting of a local Hermitian
frame over the chart $U=\\{\det b\not=0\\}$ being
$u_{i}=b\left(b^{*}b\right)^{-\frac{1}{2}}\cdot\epsilon_{i}$
since
$\left[b,p,b\left(b^{*}b\right)^{-\frac{1}{2}}\epsilon_{i}\right]\sim\left[\left(b^{*}b\right)^{\frac{1}{2}},\zeta^{u},\epsilon_{i}\right]$.
We also have the lifting of a local holomorphic frame
$h_{i}=b\epsilon_{i}=b\left(b^{*}b\right)^{\frac{1}{2}}b^{-1}u_{i}$
since
$\left[b,p,b\epsilon_{i}\right]\sim\left(I_{k},\zeta^{h},\epsilon_{i}\right)$.
The two frames are related as follows.
###### Lemma 3.1.
$h_{i}=u_{j}a_{i}^{j}$ where
$\left(a_{i}^{j}\right)=\left(b^{*}b\right)^{\frac{1}{2}}$, $i$ is indexing
the colomns and $j$ is indexing the rows.
###### Proof.
Consider
$\left(h_{1}\ldots
h_{k}\right)=b\left(b^{*}b\right)^{\frac{1}{2}}b^{-1}\left(u_{1}\ldots
u_{k}\right)=\left(u_{1}\ldots u_{k}\right)\left(a_{i}^{j}\right).$
Thus $\left(a_{i}^{j}\right)=\left(u_{1}\ldots
u_{k}\right)^{-1}b\left(b^{*}b\right)^{\frac{1}{2}}b^{-1}\left(u_{1}\ldots
u_{k}\right)$.
$\left(u_{1}\ldots u_{k}\right)=b\left(b^{*}b\right)^{-\frac{1}{2}}$
since $u_{i}=b\left(b^{*}b\right)^{-\frac{1}{2}}\epsilon_{i}$. Result follows.
∎
###### Proposition 3.2.
The metric $H$ defined above on the tautological bundle $S$ is represented by
the matrix $\left(I_{k}+\zeta^{h}\left(\zeta^{h}\right)^{*}\right)^{-1}$ in
the local holomorphic frame $h_{i}$ and the local coordinates $\zeta^{h}$.
###### Proof.
Using Lemma 3.1,
$\left(H\left(h_{i},h_{p}\right)\right)=\sum_{j,l}\left(H\left(a_{i}^{j}u_{j},a_{p}^{l}u_{l}\right)\right)=\sum_{j,l}\left(\overline{a_{i}^{j}}a_{p}^{l}H\left(u_{j},u_{l}\right)\right)=\sum_{j}\left(\overline{a_{i}^{j}}a_{p}^{j}\right)=b^{*}b$
where $i$ is indexing the rows and $p$ is indexing the columns. By the moment
map equation (3.1) and the relation (3.2),
$b^{*}b+\left(b^{*}b\right)^{\frac{1}{2}}\cdot\zeta^{h}\left(\zeta^{h}\right)^{*}\left(b^{*}b\right)^{\frac{1}{2}}=I_{k}.$
Then
$I_{k}+\zeta^{h}\left(\zeta^{h}\right)^{*}=\left(b^{*}b\right)^{-1}.$
Hence
$H=\left(I_{k}+\zeta^{h}\left(\zeta^{h}\right)^{*}\right)^{-1}.$
∎
###### Example 3.3.
Let’s consider the simplest example:
$\mathbb{P}^{1}=\operatorname{Gr}(2,1)=(\mathbb{C}^{2}-\\{0\\})/\mathbb{C}^{\times}=\mathbb{S}^{3}/\mathrm{U}\left(1\right).$
The tautological bundle for $\operatorname{Gr}(2,1)$ is
$S=\left(\mathbb{S}^{3}\times\mathbb{C}\right)/U\left(1\right)$
where $U(1)$ acts on $\mathbb{C}$ in the standard way, and it acts on both
factors on the left. (Note that this is dual to the usual notion of the
tautological bundle of $\operatorname{Gr}(1,2)=\mathbb{P}^{1}$, since we are
now considering the family of quotient lines of $\mathbb{C}^{2}$, which are
dual to subspaces of $\mathbb{C}^{2}$.)
Let’s take the standard metric on $\mathbb{C}$. We have the local Hermitian
frame (over $z_{1}\neq 0$) $u$ given by
$\left(z_{1},z_{2},z_{1}/\left|z_{1}\right|\right)\stackrel{{\scriptstyle
U(1)}}{{\sim}}\left(\left|z_{1}\right|,\zeta^{u},1\right)$
where $\zeta^{u}$ is the coordinate of $\mathbb{P}^{1}$ which belongs to the
open unit disc, and $|z_{1}|$ is determined by the moment-map equation
$\left|z_{1}\right|^{2}+\left|z_{2}\right|^{2}=\left|z_{1}\right|^{2}+\left|\zeta^{u}\right|^{2}=1.$
We also have the local holomorphic frame $h$ defined by
$\left(z_{1},z_{2},z_{1}\right)\stackrel{{\scriptstyle\mathbb{C}^{\times}}}{{\sim}}\left(1,\zeta^{h},1\right).$
The unitary and holomorphic coordinates are related by
$\zeta^{u}=\left|z_{1}\right|\cdot\zeta^{h}=(1-|\zeta^{u}|^{2})\cdot\zeta^{h}$
for $\zeta^{h}\in\mathbb{C}$. The frames are related by
$h=\left|z_{1}\right|\cdot u.$
The Hermitian frame $u$ always have length one. Writing the metric in the
holomorphic frame $h$:
$\displaystyle\left|h\right|^{2}$
$\displaystyle=\left|z_{1}\right|^{2}=1-\left|\zeta^{u}\right|^{2}=1-\left|z_{1}\right|^{2}\left|\zeta^{h}\right|^{2}$
$\displaystyle=1-\left(1-\left|z_{1}\right|^{2}\left|\zeta^{h}\right|^{2}\right)\left|\zeta^{h}\right|^{2}=\ldots$
$\displaystyle=\frac{1}{1+\left|\zeta^{h}\right|^{2}}.$
This is the standard metric on $\mathcal{O}_{\mathbb{P}^{1}}(1)$, whose
curvature gives the Fubini-Study metric on $\mathbb{P}^{1}$.
### 3.2. Metric on framed quiver moduli
We have seen that the standard metric on the trivial bundle over
$\operatorname{Mat}^{\mathbb{C}}_{n,k}$ descends to give the standard metric
on $\operatorname{Gr}(n,k)$. However, it turns out that for the framed quiver
moduli, the standard metric on the trivial bundle over $R_{\vec{n},\vec{d}}$
is not good from the GIT quotient point of view, namely it is not equivariant
under $\mathrm{GL}_{\vec{d}}$. In this section, we find a nice metric over
$R_{\vec{n},\vec{d}}$ which is equivariant under $\mathrm{GL}_{\vec{d}}$.
Recall from the last section that
$\mathcal{M}_{\vec{n},\vec{d}}=R_{\vec{n},\vec{d}}^{s}/\mathrm{GL}_{\vec{d}}$.
The universal bundle over the vertex $i$ is given by
$\mathcal{V}_{i}=\left(R_{\vec{n},\vec{d}}^{s}\times\mathbb{C}^{d_{i}}\right)\big{/}\mathrm{GL}_{\vec{d}}$
where $\mathrm{GL}_{\vec{d}}$ acts diagonally on the left, the factor
$\mathrm{GL}(d_{i},\mathbb{C})$ of $\mathrm{GL}_{\vec{d}}$ acts on
$\mathbb{C}^{d_{i}}$ in the standard way, and other factors of
$\mathrm{GL}_{\vec{d}}$ act trivially on $\mathbb{C}^{d_{i}}$.
There is an equivalent description of
$\mathcal{M}=\mathcal{M}_{\vec{n},\vec{d}}$ and the universal bundle
$\mathcal{V}_{i}$ in terms of symplectic quotient. Namely, let
$\mu:R_{\vec{n},\vec{d}}\to\mathbf{i}\mathfrak{u}_{\vec{d}}$ be the moment
map. Explicitly, $\mu=(\mu_{i})_{i\in Q_{0}}$ where
$\mu_{i}=e^{(i)}(e^{(i)})^{*}-\sum_{t(a)=i}V_{a}^{*}V_{a}+\sum_{h(a^{\prime})=i}V_{a^{\prime}}V_{a^{\prime}}^{*}.$
Then define
$\mathcal{M}_{\vec{n},\vec{d}}=\mu^{-1}\left\\{-c\right\\}\big{/}U_{\vec{d}}$
for the following level $c$.
###### Lemma 3.4.
The slope stability condition
$\left(1,\vec{0}\right)\in\mathbb{C}^{\hat{Q}_{0}}$ corresponds to the moment-
map level $c=\left(-I_{d_{i}}\right)_{i\in
Q_{0}}\in\mathbf{i}\mathfrak{u}_{\vec{d}}$, where $I_{k}$ denotes the identity
matrix of rank $k$.
###### Proof.
The character taken in King’s stability [Kin94] corresponding to
$\left(1,\vec{0}\right)$ is
$\left(1+\Sigma\vec{d}\right)\left(\left(1,\vec{0}\right)-\frac{1}{1+\Sigma\vec{d}}\left(1,\vec{1}\right)\right)=\left(\Sigma\vec{d},-1,\ldots,-1\right)$
where $\Sigma\vec{d}=\sum_{i\in Q_{0}}d_{i}$, and the first entry is over the
root vertex. (Note that there is no group action over the root vertex.) Thus
we should take $c$ to be $-I_{d_{i}}$ over each vertex $i$. ∎
The universal bundle over the vertex $i$ is then given by
$\mathcal{V}_{i}=\left(\mu^{-1}\left\\{I_{\vec{d}}\right\\}\times\mathbb{C}^{d_{i}}\right)\big{/}U_{\vec{d}}.$
Let’s review some very basic definitions about group actions.
###### Definition 3.5.
Suppose a Lie group $G$ acts on a vector bundle
$V\stackrel{{\scriptstyle\pi}}{{\to}}M$ equivariantly, namely,
$g\circ\pi=\pi\circ g$ for all $g\in G$, and the action is fiberwise linear. A
metric $H$ on $V$ is said to be $G$-equivariant if
$H_{x}\left(v,w\right)=H_{g\cdot x}\left(g\cdot v,g\cdot w\right).$
Writing in matrix form when $G=\mathrm{GL}(n,\mathbb{C})$, the above equation
is $v^{*}\cdot H_{x}\cdot w=v^{*}\cdot\left(g^{*}\cdot H_{g\cdot x}\cdot
g\right)\cdot w,$ that is,
$(g^{*})^{-1}\cdot H_{x}\cdot g^{-1}=H_{g\cdot x}.$ (3.3)
The following easily follows from the definition.
###### Lemma 3.6.
Suppose a Lie group $G$ acts on a vector bundle
$V\stackrel{{\scriptstyle\pi}}{{\to}}M$ equivariantly and fiberwise linearly,
and the action of $G$ on $M$ is free and proper. A Hermitian form $H$ on $V$
descends to the corresponding bundle over the quotient $M/G$ if and only if
$H$ is $G$-equivariant.
For framed quiver varieties, we have the framing map $e^{(j)}\colon
R_{\vec{n},\vec{d}}^{s}\rightarrow\mathrm{Hom}\left(\mathbb{C}^{n_{j}},\mathbb{C}^{d_{j}}\right)$
for each vertex $j$. Using this, we cook up a
$\mathrm{GL}_{\vec{d}}$-invariant Hermitian form on the trivial bundle
$\underline{\mathbb{C}^{d_{i}}}\to R_{\vec{n},\vec{d}}$, which descends to a
metric on $\mathcal{V}_{i}\to\mathcal{M}_{\vec{n},\vec{d}}$.
###### Theorem 3.7.
Suppose $Q$ has no oriented cycle. Fix $i\in Q_{0}$. Let $\rho$ be the row
vector whose entries are $V_{\gamma}e^{\left(t(\gamma)\right)}$, where
$\gamma$ is any path whose head $h(\gamma)$ is $i$ (including the trivial
path), $t(\gamma)$ denotes its tail, and
$V_{\gamma}\in\mathrm{Hom}(\mathbb{C}^{d_{t(\gamma)}},\mathbb{C}^{d_{h(\gamma)}})$
is the representing matrix of $\gamma$. This defines a map
$V_{\gamma}e^{\left(t(\gamma)\right)}:R_{\vec{n},\vec{d}}\to\mathrm{Hom}(\mathbb{C}^{n_{t(\gamma)}},\mathbb{C}^{d_{i}}).$
Take
$\rho\rho^{*}=\sum_{h(\gamma)=i}\left(V_{\gamma}e^{\left(t(\gamma)\right)}\right)\left(V_{\gamma}e^{\left(t(\gamma)\right)}\right)^{*}:R_{\vec{n},\vec{d}}\to\mathrm{End}(\mathbb{C}^{d_{i}}).$
Then $H=(\rho\rho^{*})^{-1}$ is $\mathrm{GL}_{\vec{d}}$-equivariant, and it
descends to a metric on $\mathcal{V}_{i}$ over
$\mathcal{M}_{\vec{n},\vec{d}}$.
###### Proof.
$(\rho\rho^{*})^{-1}$ is $\mathrm{GL}_{\vec{d}}$-equivariant:
$\displaystyle H_{\left(g_{h(a)}V_{a}g^{-1}_{t(a)},g_{j}e^{(j)}\right)_{a\in
Q_{1},j\in Q_{0}}}$
$\displaystyle=\left(\sum_{\gamma}\left(g_{h(\gamma)}V_{\gamma}e^{\left(t(\gamma)\right)}\right)\left(g_{h(\gamma)}V_{\gamma}e^{\left(t(\gamma)\right)}\right)^{*}\right)^{-1}$
$\displaystyle=(g^{*}_{i})^{-1}\left(\sum_{\gamma}\left(V_{\gamma}e^{\left(t(\gamma)\right)}\right)\left(V_{\gamma}e^{\left(t(\gamma)\right)}\right)^{*}\right)^{-1}g_{i}^{-1}$
for all $g\in\mathrm{GL}_{\vec{d}}$. By Lemma 3.6, it descends to the bundle
$\mathcal{V}_{i}$ of the quotient.
Then we prove that the matrix-valued function $H_{(V_{a},e^{(j)})_{a\in
Q_{1},j\in Q_{0}}}=(\rho\rho^{*})^{-1}$ defines a positive-definite metric on
the trivial bundle $\underline{\mathbb{C}^{d_{i}}}$ _over the moment map level
$\mu^{-1}(I_{\vec{d}})$_ (rather than the whole $R_{\vec{n},\vec{d}}$). We
prove by induction on the vertices that $\rho\rho^{*}=I_{d_{i}}+B$ where $B$
is a semi-positive-definite Hermitian matrix, and hence $\rho\rho^{*}$ is
positive definite (and so does $(\rho\rho^{*})^{-1}$).
Since the quiver does not have oriented cycle, $Q_{0}$ can be ordered such
that $i<j$ whenever there is an arrow $i\to j$. Let $i_{0}$ be the minimal
vertex.
At $i_{0}$, there is no incoming arrow (other than the framing), and the
moment-map equation reads
$e^{(i_{0})}(e^{(i_{0})})^{*}=I_{d_{i_{0}}}+\sum_{t(a)=i_{0}}V_{a}^{*}V_{a}.$
$\sum_{t(a)=i_{0}}V_{a}^{*}V_{a}$ is semi-positive definite: $v^{*}\cdot
V_{a}^{*}V_{a}\cdot v=\|V_{a}\cdot v\|^{2}\geq 0$ for any column vector $v$.
Thus the statement is true for $\rho\rho^{*}=e_{i_{0}}e_{i_{0}}^{*}$.
Suppose the statement is true for all vertices less than $i\in Q_{0}$. At $i$,
the moment-map equation is
$e^{(i)}(e^{(i)})^{*}=I_{d_{i}}+\sum_{t(a)=i}V_{a}^{*}V_{a}-\sum_{h(a^{\prime})=i}V_{a^{\prime}}V_{a^{\prime}}^{*}.$
Then
$\displaystyle\rho\rho^{*}$
$\displaystyle=e^{(i)}(e^{(i)})^{*}+\sum_{h(a)=i}V_{a}\rho_{(t(a))}\rho_{(t(a))}^{*}V_{a}^{*}$
$\displaystyle=I_{d_{i}}+\sum_{t(a)=i}V_{a}^{*}V_{a}-\sum_{h(a^{\prime})=i}V_{a^{\prime}}V_{a^{\prime}}^{*}+\sum_{h(a^{\prime})=i}V_{a^{\prime}}(I_{d_{t(a^{\prime})}}+B_{t(a^{\prime})})V_{a^{\prime}}^{*}$
$\displaystyle=I_{d_{i}}+\sum_{t(a)=i}V_{a}^{*}V_{a}+\sum_{h(a^{\prime})=i}V_{a^{\prime}}B_{t(a^{\prime})}V_{a^{\prime}}^{*}$
where
$\rho_{(t(a))}=\left(V_{\gamma}e^{\left(t(\gamma)\right)}\right)_{h(\gamma)=t(a)}$,
which by inductive assumption can be written as
$I_{d_{t(a^{\prime})}}+B_{t(a^{\prime})}$ where $B_{t(a^{\prime})}$ is semi-
positive definite. The matrices $V_{a}^{*}V_{a}$ and
$V_{a^{\prime}}B_{t(a^{\prime})}V_{a^{\prime}}^{*}$ are semi-positive
definite:
$v^{*}V_{a^{\prime}}B_{t(a^{\prime})}V_{a^{\prime}}^{*}v=(V_{a^{\prime}}^{*}v)^{*}B_{t(a^{\prime})}(V_{a^{\prime}}^{*}v)\geq
0$
for all $v$. This proves the statement for the vertex $i$. ∎
The expression $(\rho\rho^{*})^{-1}$ can be understood as follows. $\rho^{*}$
embeds the dual $V_{i}^{*}$ into the dual frame which is a trivial bundle
equipped with the standard metric. This gives an induced metric on
$V_{i}^{*}$, which is $\rho\rho^{*}$ written in matrix form. Taking the dual,
we get the metric $H_{i}=(\rho\rho^{*})^{-1}$ on $V_{i}$.
By construction, the metrics on the dual $\mathcal{V}_{i}^{*}$ (still denoted
as $H_{i}$) have the following nice property. Inductively, it gives nice
expressions of $H_{i}$ in terms of _holomorphic coordinates_.
###### Proposition 3.8.
Suppose $Q$ has no oriented cycle. For $v,w\in(\mathcal{V}_{i})^{*}$,
$H_{i}(v,w)=H_{0}((e^{(i)})^{*}(v),(e^{(i)})^{*}(w))+\sum_{h(a)=i}H_{t(a)}(a^{*}(v),a^{*}(w))$
where $H_{0}$ denotes the trivial metric on the trivial bundle, and
$e^{(i)},a$ are denoting the holomorphic bundle maps corresponding to the
framing and arrow maps respectively.
###### Proof.
The metric on $\mathcal{V}_{i}^{*}$ is given by the matrix $\rho\rho^{*}$.
Then the above equation follows from
$\rho\rho^{*}=e^{(i)}(e^{(i)})^{*}+\sum_{h(a)=i}V_{a}\rho_{(t(a))}\rho_{(t(a))}^{*}V_{a}^{*}.$
∎
###### Remark 3.9.
As we have seen, the Grassmannian $\operatorname{Gr}(n,k)$ can be understood
as the framed moduli for the quiver which has one vertex and no arrow. The
matrix $e\in\mathrm{Hom}(\mathbb{C}^{n},\mathbb{C}^{k})$ is the framing map.
Then the moment map equation implies
$\rho\rho^{*}=ee^{*}=I_{k}$
in the above proposition. This is the standard metric on the trivial bundle
$\underline{\mathbb{C}^{k}}$ that we have used in the last subsection. In
particular $\rho\rho^{*}=(\rho\rho^{*})^{-1}$ in this case. But this is not
true for other quivers.
###### Remark 3.10.
Note that the above becomes an infinite sum if the quiver has oriented cycles.
The $\mathrm{GL}_{\vec{d}}$-equivariance still holds. We should restrict to
the open subset of $R^{s}_{n,d}$ that $(\rho\rho^{*})^{-1}$ is convergent. In
the next subsection, we will prove that the same expression defines a metric
for any given quiver.
The GIT description will be important to the proof of Theorem 3.14.
There is a residual symmetry $U_{\vec{n}}=\prod_{i\in Q_{0}}U({n_{i}})$ acting
on $\mathcal{M}_{\vec{n},\vec{d}}$. Actually, there is a bigger symmetry by
the non-compact group $\mathrm{GL}(\hat{W})$ [Rei08]. $U_{\vec{n}}$ is
considered here since this is the symmetry of the metric $H$ on
$\mathcal{V}_{i}$ as we shall see.
###### Definition 3.11.
The right residual action of $U_{\vec{n}}$ on $\mathcal{M}$ is defined as
$\left[(V_{a},e^{(j)})_{a\in Q_{1},j\in Q_{0}}\right]\cdot
g=\left[(V_{a},e^{(j)}\circ g_{j})_{a\in Q_{1},j\in Q_{0}}\right]$
for $g=(g_{j}\in U(d_{i}))_{j\in Q_{0}}\in U_{\vec{n}}$.
Since the above commutes with the left action of $\mathrm{GL}_{\vec{d}}$, the
action is well-defined on $\mathcal{M}$.
###### Lemma 3.12.
There is a canonical lift of the action of $U_{\vec{n}}$ on
$\mathcal{M}_{\vec{n},\vec{d}}$ to the universal bundle $\mathcal{V}_{i}$, so
that the bundle map $\mathcal{V}_{i}\rightarrow\mathcal{M}_{\vec{n},\vec{d}}$
is equivariant.
###### Proof.
$\mathcal{V}_{i}$ is the $\mathrm{GL}_{\vec{d}}$-quotient of the trivial
bundle $R^{s}_{n,d}\times V_{i}$. $U_{\vec{n}}$ acts on this by acting on the
component $V_{i}$ trivially. This action commutes with the left action of
$\mathrm{GL}_{\vec{d}}$ on $R^{s}_{n,d}\times V_{i}$, and hence descends to
act on $\mathcal{V}_{i}$. ∎
###### Lemma 3.13.
The metric defined in Theorem 3.7 are $U_{\vec{n}}$-invariant.
###### Proof.
For any $g\in U_{\vec{n}}$, since $g_{j}g_{j}^{*}=I_{n_{j}}$ for any $j\in
Q_{0}$,
$\displaystyle H_{\left(V_{a},e^{(j)}\right)_{a\in Q_{1},j\in Q_{0}}\cdot g}$
$\displaystyle=\left(\sum_{\gamma}\left(V_{\gamma}e^{\left(t(\gamma)\right)}\cdot
g_{t(\gamma)}\right)\left(V_{\gamma}e^{\left(t(\gamma)\right)}\cdot
g_{t(\gamma)}\right)^{*}\right)^{-1}$
$\displaystyle=\left(\sum_{\gamma}\left(V_{\gamma}e^{\left(t(\gamma)\right)}\right)\left(V_{\gamma}e^{\left(t(\gamma)\right)}\right)^{*}\right)^{-1}=H_{\left(V_{a},e^{(j)}\right)_{a\in
Q_{1},j\in Q_{0}}}.$
∎
Recall from Theorem 2.4 that $\mathcal{M}_{\vec{n},\vec{d}}$ is the total
space of an iterated Grassmann bundle $M^{(N)}\stackrel{{\scriptstyle
p_{N}}}{{\to}}M^{(N-1)}\stackrel{{\scriptstyle
p_{N-1}}}{{\to}}\ldots\stackrel{{\scriptstyle
p_{2}}}{{\to}}M^{(1)}\stackrel{{\scriptstyle p_{1}}}{{\to}}\mathrm{pt}$.
Moreover, $\mathcal{V}_{i}$ is the pull-back of the tautological bundle
$S_{i}$ of the Grassmann bundle
$M^{(i)}=\operatorname{Gr}_{M^{(i)}}(\underline{\mathbb{C}^{n_{i}}}\oplus\bigoplus_{j\to
i}p_{i}^{*}\ldots p_{j+1}^{*}(S_{j}),d_{i})\to M^{(i-1)}$. The tautological
bundle of the Grassmannian is equipped with a standard metric as illustrated
in Section 3.1. Inductively, $\mathcal{V}_{i}$ is also equipped with a pull-
back metric. We show that this equals to the metric $H$ we defined by an
explicit formula.
###### Theorem 3.14.
For all $i\in Q_{0}$, the metric
$H=\left(\sum_{h(\gamma)=i}\left(V_{\gamma}e^{\left(t(\gamma)\right)}\right)\left(V_{\gamma}e^{\left(t(\gamma)\right)}\right)^{*}\right)^{-1}$
equals to the metric on $\mathcal{V}_{i}$ constructed from the iterated
Grassmann bundle.
###### Proof.
As the proof of Theorem 3.7, we do induction on the vertices, which are
totally ordered such that $i<j$ whenever there is an arrow $i\to j$.
First, we have the GIT fiber-bundle map $\mathcal{M}\to M^{(j)}$, where
$M^{(j)}$ is the framed moduli of the quiver $Q^{(j)}$ (which is obtained by
removing all vertices $k>j$ and the corresponding arrows from $Q$). This map
$(V,e)\mapsto(V^{\prime},e^{\prime})$ is simply forgetting all the irrelevant
arrow maps and frame maps that are not supported on the subquiver $Q^{(j)}$.
Note that the stability condition is preserved: any subrepresentation
$R^{\prime}\subset V^{\prime}$ can be extended to a subrepresentation
$R\subset V$ by assigning the whole $V_{k}$ to the additional vertices $k>j$
(and the arrow maps just come from restriction). If
$\mathrm{Im}(e^{\prime})\subset R^{\prime}$, then $\mathrm{Im}(e)\subset R$.
Note that here we use the GIT description instead of symplectic reduction
since $(V^{\prime},e^{\prime})$ no longer satisfies the moment-map equation in
defining $M^{(j)}$ (even when $(V,e)$ satisfies the moment-map equation for
$\mathcal{M}$).
We start with the minimal vertex $i_{0}$. The quiver $Q^{(i_{0})}$ is simply a
single vertex, and the corresponding framed moduli is
$M^{(i_{0})}=\operatorname{Gr}(n_{i_{0}},d_{i_{0}}).$ The universal bundle
$\mathcal{V}_{i_{0}}^{Q^{(i_{0})}}$ is the tautological bundle of
$M^{(i_{0})}=\operatorname{Gr}(n_{i_{0}},d_{i_{0}})$. From the last
subsection, the standard metric of $\mathcal{V}_{i_{0}}^{Q^{(i_{0})}}$ is
descended from $I_{d_{i_{0}}}$, which is exactly
$(e^{Q^{(i_{0})}}(e^{Q^{(i_{0})}})^{*})^{-1}$ by the moment map equation for
$Q^{(i_{0})}$. The statement is trivial in this case.
Now consider the vertex $i$. Denote the vertex right before $i$ by $i-1$. For
the quiver $Q^{(i-1)}$, assume that the two metrics on the universal bundle
$\mathcal{V}^{Q^{(i-1)}}_{j}$ agree for every $j\in Q^{(i-1)}_{0}$. We have
the bundle map $\pi:M^{(i)}\to M^{(i-1)}$, and
$\mathcal{V}^{Q^{(i)}}_{j}=\pi^{*}\mathcal{V}^{Q^{(i-1)}}_{j}$ for all $j<i$.
Moreover, the metric on $\mathcal{V}^{Q^{(i)}}_{j}$ is pull-back from
$\mathcal{V}^{Q^{(i-1)}}_{j}$, which equals to
$\left(\sum_{h(\gamma)=j}\left(V_{\gamma}e^{\left(t(\gamma)\right)}\right)\left(V_{\gamma}e^{\left(t(\gamma)\right)}\right)^{*}\right)^{-1}$
by inductive assumption. The pull-back map does not change the arrow and
framing maps of $Q^{(i-1)}$. This proves the statement for $S^{Q^{(i)}}_{j}$
for $j<i$.
Consider $\mathcal{V}^{Q^{(i)}}_{i}$, which is the tautological bundle
associated to the Grassmannian bundle over $M^{(i-1)}$ parametrizing quotients
of
$(e^{(i)},(V_{a})_{h(a)=i}):\left.\left(\underline{\mathbb{C}^{d_{i}}}\oplus\left(\bigoplus_{h(a)=i}\mathcal{V}^{Q^{(i-1)}}_{t(a)}\right)\right)\right|_{x\in
M^{(i-1)}}\to\mathbb{C}^{d_{i}}.$
The metric on $(\mathcal{V}^{Q^{(i)}}_{i})^{*}$ is induced by the embedding
$(e^{(i)},(V_{a})_{h(a)=i})^{*}$ to
$\left(\underline{\mathbb{C}^{d_{i}}}\oplus\left(\bigoplus_{h(a)=i}\mathcal{V}_{t(a)}\right)\right)^{*}$,
whose metric is given by
$I_{d_{i}}\oplus\bigoplus_{h(a)=i}\left(\sum_{h(\gamma)=t(a)}\left(V_{\gamma}e^{\left(t(\gamma)\right)}\right)\left(V_{\gamma}e^{\left(t(\gamma)\right)}\right)^{*}\right)$
by inductive assumption. Thus the induced metric on
$(\mathcal{V}^{Q^{(i)}}_{i})^{*}$ is
$\displaystyle(e^{(i)},(V_{a})_{h(a)=i})\cdot\left(I_{d_{i}}\oplus\bigoplus_{h(a)=i}\left(\sum_{h(\gamma)=t(a)}\left(V_{\gamma}e^{\left(t(\gamma)\right)}\right)\left(V_{\gamma}e^{\left(t(\gamma)\right)}\right)^{*}\right)\right)\cdot(e^{(i)},(V_{a})_{h(a)=i})^{*}$
$\displaystyle=e^{(i)}(e^{(i)})^{*}+\sum_{\begin{subarray}{c}h(a)=i\\\
h(\gamma)=t(a)\end{subarray}}V_{a}\cdot\left(V_{\gamma}e^{\left(t(\gamma)\right)}\right)\left(V_{\gamma}e^{\left(t(\gamma)\right)}\right)^{*}\cdot
V_{a}^{*}$
$\displaystyle=\sum_{h(\gamma)=i}\left(V_{\gamma}e^{\left(t(\gamma)\right)}\right)\left(V_{\gamma}e^{\left(t(\gamma)\right)}\right)^{*}.$
Taking reciprocal gives the metric on $\mathcal{V}^{Q^{(i)}}_{i}$. This proves
the metric has the given expression. ∎
###### Theorem 3.15.
The Ricci curvature of the metric on $\bigotimes_{i\in Q_{0}}\mathcal{V}_{i}$
given in Theorem 3.7 defines a Kähler metric on $\mathcal{M}$.
###### Proof.
As in Theorem 3.7, denote
$\rho=\rho^{\left(i\right)}=\left(V_{\gamma}e^{\left(t(\gamma)\right)}\right)_{\gamma:h(\gamma)=i}$
which is a matrix-valued function on the vector space $R_{\vec{n},\vec{d}}$.
At each point of $R_{\vec{n},\vec{d}}$, $\rho$ is a linear map from
$\widehat{W}_{i}:=\bigoplus_{\gamma:h(\gamma)=i}\mathbb{C}^{n_{t(\gamma)}}$
(3.4)
to $V_{i}$. The Ricci curvature of the metric $(\rho\rho^{*})^{-1}$ is given
by $i\partial\overline{\partial}\log\det\rho\rho^{\mathrm{*}}$. We have
$\displaystyle\partial\overline{\partial}\log\det\rho\rho^{\mathrm{*}}$
$\displaystyle=\partial\left(\mathrm{tr~{}}\left(\left(\rho\rho^{\mathrm{*}}\right)^{-1}\overline{\partial}\left(\rho\rho^{\mathrm{*}}\right)\right)\right)$
$\displaystyle=\mathrm{tr}\left(\partial\left(\left(\rho\rho^{\mathrm{*}}\right)^{-1}\rho\left(\partial\rho\right)^{\mathrm{*}}\right)\right)$
$\displaystyle=\mathrm{tr~{}}\left(\left(\rho\rho^{\mathrm{*}}\right)^{-1}\partial\rho\left(\partial\rho\right)^{\mathrm{*}}+\left(\partial\left(\rho\rho^{\mathrm{*}}\right)^{-1}\right)\rho\left(\partial\rho\right)^{\mathrm{*}}\right)$
$\displaystyle=\mathrm{tr~{}}\left(\left(\partial\rho\right)^{\mathrm{*}}\left(\rho\rho^{\mathrm{*}}\right)^{-1}\partial\rho\right)-\mathrm{tr~{}}\left(\left(\rho\rho^{\mathrm{*}}\right)^{-1}\left(\partial\left(\rho\rho^{\mathrm{*}}\right)\right)\left(\rho\rho^{\mathrm{*}}\right)^{-1}\rho\left(\partial\rho\right)^{\mathrm{*}}\right)$
$\displaystyle=\mathrm{tr~{}}\left(\left(\partial\rho\right)^{\mathrm{*}}\left(\rho\rho^{\mathrm{*}}\right)^{-1}\partial\rho\right)-\mathrm{tr~{}}\left(\left(\rho\rho^{\mathrm{*}}\right)^{-1}\rho\left(\partial\rho\right)^{\mathrm{*}}\left(\rho\rho^{\mathrm{*}}\right)^{-1}\left(\left(\partial\rho\right)\rho^{\mathrm{*}}\right)\right)$
$\displaystyle=\mathrm{tr~{}}\left(\left(\partial\rho\right)^{\mathrm{*}}\left(\rho\rho^{\mathrm{*}}\right)^{-1}\partial\rho\right)-\mathrm{tr~{}}\left(\left(\partial\rho\cdot\left(\rho^{\mathrm{*}}\left(\rho\rho^{\mathrm{*}}\right)^{-\frac{1}{2}}\right)\right)^{\mathrm{*}}\left(\rho\rho^{\mathrm{*}}\right)^{-1}\left(\partial\rho\cdot\left(\rho^{\mathrm{*}}\left(\rho\rho^{\mathrm{*}}\right)^{-\frac{1}{2}}\right)\right)\right)$
where $\overline{\partial}\rho=0$ since the matrix $\rho$ has polynomial
entries in holomorphic coordinates.
We can take the singular value decomposition
$\rho=U\cdot\left(\text{diag}\left(\lambda_{1},\ldots,\lambda_{d_{i}}\right)\,\,\,0\right)\cdot
V^{\mathrm{*}}$
where $U\in U\left(d_{i}\right),~{}V\in U\left(\dim\widehat{W}_{i}\right)$,
and $\lambda_{i}>0$. ($\lambda_{i}\neq 0$ since $\rho$ is surjective.) Then
$\displaystyle\rho\rho^{\mathrm{*}}$
$\displaystyle=U\left(\text{diag}\left(\lambda_{1}^{2},\ldots,\lambda_{d_{i}}^{2}\right)\right)U^{\mathrm{*}}.$
$\displaystyle\rho^{\mathrm{*}}\left(\rho\rho^{\mathrm{*}}\right)^{-\frac{1}{2}}$
$\displaystyle=V\left(\begin{array}[]{c}\text{diag}\left(\lambda_{1},\ldots,\lambda_{\alpha\left(i\right)}\right)\\\
0\end{array}\right)\left(\text{diag}\left(\lambda_{1}^{-1},\ldots,\lambda_{\alpha\left(i\right)}^{-1}\right)\right)U^{\mathrm{*}}=V\left(\begin{array}[]{c}I_{\alpha\left(i\right)}\\\
0\end{array}\right)U^{\mathrm{*}}.$
In other words,
$\rho^{\mathrm{*}}=\left(\rho^{\mathrm{*}}\left(\rho\rho^{\mathrm{*}}\right)^{-\frac{1}{2}}\right)\left(\rho\rho^{\mathrm{*}}\right)^{\frac{1}{2}}$
is decomposed into the rescaling $\rho\rho^{\mathrm{*}}$ and the orthogonal
embedding
$\left(\rho^{\mathrm{*}}\left(\rho\rho^{\mathrm{*}}\right)^{-\frac{1}{2}}\right)$
to $\mathrm{Im}\rho^{\mathrm{*}}\subset\widehat{W}_{i}$.
Now take a vector $v\in T^{1,0}R_{\alpha,d}\cong TR_{\alpha,d}$, and evaluate
the above two-form by $(v,\bar{v})$. The first term
$\mathrm{tr~{}}\left(\left(\partial_{v}\rho\right)^{\mathrm{*}}\left(\rho\rho^{\mathrm{*}}\right)^{-1}\partial_{v}\rho\right)$
is the square norm of the linear map
$\partial_{v}\rho\colon\left(\widehat{W}_{i},h_{\mathrm{std}}\right)\rightarrow\left(V_{i},~{}h_{{\left(\rho\rho^{\mathrm{*}}\right)^{-1}}}\right).$
Namely we take the standard basis in $\widehat{W}_{i}$ (which is orthonormal
under the standard metric $h_{\mathrm{std}}$), map it to $V_{i}$ by
$\partial_{v}\rho$, and take the sum of their square norms with respect to the
metric $h_{{\left(\rho\rho^{\mathrm{*}}\right)^{-1}}}$.
The second term
$\mathrm{tr~{}}\left(\left(\partial_{v}\rho\cdot\left(\rho^{\mathrm{*}}\left(\rho\rho^{\mathrm{*}}\right)^{-\frac{1}{2}}\right)\right)^{\mathrm{*}}\left(\rho\rho^{\mathrm{*}}\right)^{-1}\left(\partial_{v}\rho\cdot\left(\rho^{\mathrm{*}}\left(\rho\rho^{\mathrm{*}}\right)^{-\frac{1}{2}}\right)\right)\right)$
is the square norm of the following component
$\left(\partial_{v}\rho\right)_{1}$ of $\partial_{v}\rho$. Namely, we
decompose
$\widehat{W}_{i}=\left(\mathrm{Im}~{}\rho^{\mathrm{*}}\right)\oplus\left(\mathrm{Im}~{}\rho^{\mathrm{*}}\right)^{\bot}$
and write
$\partial_{v}\rho=\left(\left(\partial_{v}\rho\right)_{1},\left(\partial_{v}\rho\right)_{2}\right)$
where
$\left(\partial_{v}\rho\right)_{1}\colon\mathrm{Im}~{}\rho^{\mathrm{*}}\rightarrow
V_{i}$ and
$\left(\partial_{v}\rho\right)_{2}\colon\left(\mathrm{Im}~{}\rho^{\mathrm{*}}\right)^{\bot}\rightarrow
V_{i}$.
We have
$\partial_{v}\overline{\partial_{v}}\log\det\rho\rho^{\mathrm{*}}=\left\|\partial_{v}\rho\right\|_{H}^{2}-\left\|\left(\partial_{v}\rho\right)_{1}\right\|_{H}^{2}=\left\|\left(\partial_{v}\rho\right)_{2}\right\|_{H}^{2}\geq
0$ (3.5)
for all $v\in T^{1,0}$. ($H$ stands for the metric
$(\rho\rho^{\mathrm{*}})^{-1}$.) This proves that the Ricci curvature of the
metric for each $i$ is semi-positive definite.
Now consider
$H_{T}=\sum_{i\in
Q_{0}}\partial\overline{\partial}\log\det\rho^{(i)}\left(\rho^{(i)}\right)^{\mathrm{*}}.$
Suppose it is zero when evaluated at $(v,\bar{v})$. Then each individual term
equals to zero. This forces $\left(\partial_{v}\rho^{(i)}\right)_{2}=0$ for
all $i$, that is, image of
$\left(\partial_{v}\rho^{(i)}\right)^{\mathrm{*}}=\partial_{v}(\rho^{(i)})^{*}$
sits in the image of $(\rho^{(i)})^{\mathrm{*}}$. This exactly means $v$
descends to the zero tangent vector in the quotient $\mathcal{M}$: $v$ does
not alter the subspaces given by $(\rho^{(i)})^{*}:V_{i}\to\hat{W}_{i}$ for
all $i$. By the identification of $\mathcal{M}$ as a quiver Grassmannian
[Rei08], it means $v$ does not change the position of the point
$((\rho^{(i)})^{*}:i\in Q_{0})$ in the quiver Grassmannian, and hence must be
the zero tangent vector. This proves the above expression is positive
definite. ∎
By Equation (3.5), the metric on $T\mathcal{M}$ produced from the Ricci
curvature of $\mathcal{V}_{i}$ is
$H_{T}(v,v)=\sum_{i\in Q_{0}}\|(\partial_{v}\rho^{(i)})_{2}\|_{H_{i}}^{2}.$
We have the tautological exact sequence of vector bundles over $\mathcal{M}$:
$0\to\bigoplus_{i\in Q_{0}}\mathrm{End}(\mathcal{V}_{i})\to\bigoplus_{a\in
Q_{1}}\mathrm{Hom}(\mathcal{V}_{t_{a}},\mathcal{V}_{h_{a}})\oplus\bigoplus_{i\in
Q_{0}}\mathcal{V}_{i}^{n_{i}}\to T\mathcal{M}\to 0$
where the second arrow is given by sending
$X|_{[\phi,e]\in\mathcal{M}}\in\bigoplus_{i\in
Q_{0}}\mathrm{End}(\mathcal{V}_{i})$ to
$\left((X_{h_{a}}\phi_{a}-\phi_{a}X_{t_{a}})_{a\in Q_{1}},(X_{i}e^{(i)})_{i\in
Q_{0}}\right)$
(which is the derivative of the action of $\mathrm{GL}_{\vec{d}}$), and
$T\mathcal{M}$ is obtained as the quotient bundle (of the middle one by the
first one). $H(v,v)$ can be defined for $v\in\bigoplus_{a\in
Q_{1}}\mathrm{Hom}(\mathcal{V}_{t_{a}},\mathcal{V}_{h_{a}})\oplus\bigoplus_{i\in
Q_{0}}\mathcal{V}_{i}^{n_{i}}$. $H_{TM}$ is zero on $\bigoplus_{i\in
Q_{0}}\mathrm{End}(\mathcal{V}_{i})$: the action of
$\mathrm{GL}(d_{j},\mathbb{C})$ does not change $\rho^{(i)}$ for $j\not=i$;
for $X\in\mathrm{End}(\mathcal{V}_{i})$, $(X\cdot\rho^{(i)}\cdot
w)^{*}v=w^{*}(\rho^{(i)})^{*}(X^{*}\cdot v)=0$ for all
$v\in\mathcal{V}_{i},w\in(\operatorname{Im}(\rho^{(i)})^{*})^{\perp}$, and so
$(\partial_{X}\rho^{(i)})_{2}=0$.
### 3.3. Quiver with oriented cycles
When the quiver $Q$ has an oriented cycle, the framed moduli
$\mathcal{M}_{\vec{n},\vec{d}}$ is no longer projective. Examples of such
quivers were studied algebraically by [Fed13, ER09] along the line of Reineke.
On the other hand, the metric given in Theorem 3.7 still makes sense for
quiver with oriented cycles, as long as we stay in the domain of convergence
and prove that it is positive-definite. Below we will prove this for any given
quiver.
Denote the moment-map level by $R^{\mu=I}_{n,d}=\\{\mu=1\\}\subset
R_{\vec{n},\vec{d}}$. Let $\|A\|=\sup_{\|v\|=1}\|A\cdot v\|$ be the operator
norm of a matrix $A$. We take the following open subset of $R^{\mu=I}_{n,d}$.
###### Definition 3.16.
Define
$R_{\vec{n},\vec{d}}^{\mu=I,\circ}:=\\{(V,e)\in
R^{\mu=I}_{n,d}:\|V(\gamma)\|<1\textrm{ for every oriented cycle }\gamma\\}$
where $V(\gamma)=V(a_{k})\ldots V(a_{1})$ for $\gamma=a_{k}\ldots a_{1}$.
$R_{\vec{n},\vec{d}}^{\circ}:=\mathrm{GL}_{\vec{d}}\cdot
R_{\vec{n},\vec{d}}^{\mu=I,\circ}.$
and
$\mathcal{M}_{\vec{n},\vec{d}}^{\circ}:=R_{\vec{n},\vec{d}}^{\mu=I,\circ}/U_{\vec{d}}=R_{\vec{n},\vec{d}}^{\circ}/\mathrm{GL}_{\vec{d}}.$
The above definition of $\mathcal{M}_{\vec{n},\vec{d}}^{\circ}$ makes sense
because of the following.
###### Lemma 3.17.
$R_{\vec{n},\vec{d}}^{\mu=I,\circ}$ is invariant under $U_{\vec{d}}$.
###### Proof.
For every oriented cycle $\gamma$ at $i\in Q_{0}$, $(g\cdot
V)(\gamma)=g_{i}V(\gamma)g_{i}^{-1}$, and hence the condition
$\|V(\gamma)\|<1$ is respected for $g\in U_{\vec{d}}$. ∎
The main theorem in this section is the following.
###### Theorem 3.18.
Let $Q$ be an arbitrary quiver. As in Theorem 3.7, for each $i\in Q_{0}$, set
$H_{i}=(\rho_{i}\rho_{i}^{*})^{-1}=\left(\sum_{h(\gamma)=i}\left(V_{\gamma}e^{\left(t(\gamma)\right)}\right)\left(V_{\gamma}e^{\left(t(\gamma)\right)}\right)^{*}\right)^{-1}$
which is an infinite sum, whose terms are ordered by the length of the path
$\gamma$. (There are just finitely many paths under each fixed length.) This
gives a convergent function
$H_{i}:R_{\vec{n},\vec{d}}^{\circ}\to\mathrm{End}(\mathbb{C}^{d_{i}})$.
$H_{i}$ is $\mathrm{GL}_{\vec{d}}$-equivariant, and it descends to a metric on
$\mathcal{V}_{i}$ over $\mathcal{M}_{\vec{n},\vec{d}}^{\circ}$.
We break into several steps to prove the above theorem. First, consider the
convergence.
###### Lemma 3.19.
$\rho\rho^{*}$ is absolutely convergent over
$R_{\vec{n},\vec{d}}^{\mu=I,\circ}$. Hence $(\rho\rho^{*})^{-1}$ is well-
defined and $\mathrm{GL}_{\vec{d}}$-equivariant on
$R_{\vec{n},\vec{d}}^{\circ}$.
###### Proof.
For $(V,e)\in R_{\vec{n},\vec{d}}^{\mu=I,\circ}$, we consider the expression
$\sum_{h(\gamma)=i}\left\|V_{\gamma}e^{\left(t(\gamma)\right)}\right\|\left\|\left(V_{\gamma}e^{\left(t(\gamma)\right)}\right)^{*}\right\|\leq\sum_{h(\gamma)=i}\left\|V_{\gamma}e^{\left(t(\gamma)\right)}\right\|^{2}.$
There are only finitely many paths $\gamma_{1},\ldots,\gamma_{k}$ with
$h(\gamma_{l})=i$ which do not contain any oriented cycle. Any other path
(with $h(\gamma)=i$) can be written as concatenation of one of these
$\gamma_{l}$ and some oriented cycles at some vertices. Thus
$\sum_{h(\gamma)=i}\left\|V_{\gamma}e^{\left(t(\gamma)\right)}\right\|^{2}\leq\sum_{l=1}^{k}\|\gamma_{l}e^{(t(\gamma_{l}))}\|^{2}\sum_{p=0}^{\infty}(1-\epsilon)^{p}=\sum_{l=1}^{k}\|\gamma_{l}e^{(t(\gamma_{l}))}\|^{2}\sum_{p=0}^{\infty}(1-\epsilon)^{p}=\sum_{l=1}^{k}\frac{\|\gamma_{l}e^{(t(\gamma_{l}))}\|^{2}}{\epsilon}<\infty$
where given $V$, there is a fixed $\epsilon\in(0,1)$ such that
$\|V_{\gamma}\|^{2}<1-\epsilon$ for all oriented cycles $\gamma$. Hence
$\rho\rho^{*}$ is absolutely convergent for every $(V,e)\in
R_{\vec{n},\vec{d}}^{\mu=I,\circ}$.
Every element in $R_{\vec{n},\vec{d}}^{\circ}$ can be written as $g\cdot(V,e)$
for $g\in\mathrm{GL}_{\vec{d}}$ and $(V,e)\in
R_{\vec{n},\vec{d}}^{\mu=I,\circ}$.
$(\rho\rho^{*})^{-1}|_{g\cdot(V,e)}=(g_{i}^{*})^{-1}(\rho\rho^{*})^{-1}|_{(V,e)}g_{i}^{-1}$
where $(\rho\rho^{*})^{-1}|_{(V,e)}$ is convergent. ∎
It remains to prove positive definiteness of $H_{i}$. First we consider the
following specific quiver which is simply a single oriented cycle.
###### Lemma 3.20.
If $Q$ is a single oriented cycle with $N$ vertices, then for each vertex $i$,
$\rho_{i}\rho_{i}^{*}$ is positive definite. (In particular, when $N=1$, $Q$
consists of one vertex and a self loop.)
###### Proof.
By symmetry, we just need to prove for $i=N$. Let $l=a_{N}\ldots a_{1}$, where
$a_{k}$ is the arrow $(k-1)\to k$.
$\rho_{N}\rho_{N}^{*}=\sum_{k=1}^{N}\left(a_{N}\ldots
a_{k+1}e_{k}e_{k}^{*}a_{k+1}^{*}\ldots a_{N}^{*}+\sum_{p>0}l^{p}a_{N}\ldots
a_{k+1}e_{k}e_{k}^{*}a_{k+1}^{*}\ldots a_{N}^{*}(l^{p})^{*}\right).$
The moment map equation at the vertex $k$ is
$e_{k}e_{k}^{*}=I-a_{k}a_{k}^{*}+a_{k+1}^{*}a_{k+1}$. Then the first term
gives
$\displaystyle\sum_{k=1}^{N}a_{N}\ldots
a_{k+1}(I-a_{k}a_{k}^{*}+a_{k+1}^{*}a_{k+1})a_{k+1}^{*}\ldots a_{N}^{*}$
$\displaystyle=$ $\displaystyle\sum_{k=1}^{N}a_{N}\ldots
a_{k+1}a_{k+1}^{*}\ldots a_{N}^{*}-\sum_{k=1}^{N}a_{N}\ldots
a_{k+1}a_{k}a_{k}^{*}a_{k+1}^{*}\ldots a_{N}^{*}+\sum_{k=1}^{N}a_{N}\ldots
a_{k+1}a_{k+1}^{*}a_{k+1}a_{k+1}^{*}\ldots a_{N}^{*}$ $\displaystyle=$
$\displaystyle I-ll^{*}+\sum_{k=1}^{N}a_{N}\ldots
a_{k+1}a_{k+1}^{*}a_{k+1}a_{k+1}^{*}\ldots a_{N}^{*}.$
Similarly, the second term gives
$\displaystyle\sum_{k=1}^{N}\sum_{p>0}l^{p}a_{N}\ldots
a_{k+1}(I-a_{k}a_{k}^{*}+a_{k+1}^{*}a_{k+1})a_{k+1}^{*}\ldots
a_{N}^{*}(l^{p})^{*}$ $\displaystyle=$
$\displaystyle\sum_{p>0}l^{p}(l^{p})^{*}-\sum_{p>0}l^{p}ll^{*}(l^{p})^{*}+\sum_{p>0}\sum_{k=1}^{N}l^{p}a_{N}\ldots
a_{k+1}a_{k+1}^{*}a_{k+1}a_{k+1}^{*}\ldots a_{N}^{*}(l^{p})^{*}$
$\displaystyle=$ $\displaystyle
ll^{*}+\sum_{p>0}\sum_{k=1}^{N}l^{p}a_{N}\ldots
a_{k+1}a_{k+1}^{*}a_{k+1}a_{k+1}^{*}\ldots a_{N}^{*}(l^{p})^{*}.$
Combining the two terms,
$\rho_{N}\rho_{N}^{*}=I+\sum_{p\geq 0}\sum_{k=1}^{N}l^{p}a_{N}\ldots
a_{k+1}a_{k+1}^{*}a_{k+1}a_{k+1}^{*}\ldots a_{N}^{*}(l^{p})^{*}$
and the second term is semi-positive-definite. Hence $\rho_{N}\rho_{N}^{*}$ is
positive definite. ∎
The following is the key lemma to prove positive-definiteness for a general
quiver.
###### Lemma 3.21.
Suppose $Q$ has the property that for every $i\in Q_{0}$, restricted to the
intersection of the moment map locus and $R^{Q,\circ}_{n,d}$,
$\rho^{Q}_{i}(\rho^{Q}_{i})^{*}=I+B_{i}$ for some semi-positive definite
matrix $B_{i}$. Let $Q^{\prime}$ be obtained by concatenating to $Q$ a chain
$x_{1}\stackrel{{\scriptstyle
a_{1}^{\prime}}}{{\to}}\ldots\stackrel{{\scriptstyle
a_{k-1}^{\prime}}}{{\to}}x_{k}$ where $x_{1}$ and $x_{k}$ are certain vertices
in $Q$. ($x_{1}$ can be equal to $x_{k}$, meaning what we have added is an
oriented cycle. When $k=1$, we have added a loop;
$a^{\prime}_{0}=a^{\prime}_{1}$. When $k=2$, there is no intermediate vertex
in the chain.) Then $Q^{\prime}$ has the same property. Namely, for every
$i\in Q^{\prime}_{0}$, restricted to the intersection of the moment map locus
and $R^{Q^{\prime},\circ}_{n,d}$,
$\rho^{Q^{\prime}}_{i}(\rho^{Q^{\prime}}_{i})^{*}=I+B_{i}^{\prime}$ for some
semi-positive-definite matrix $B_{i}^{\prime}$.
###### Proof.
First, consider the case that the given vertex $i\in Q_{0}^{\prime}$ belongs
to $Q_{0}$. For the original quiver $Q$,
$\rho^{Q}_{i}(\rho^{Q}_{i})^{*}=I+B_{i}$ where $B_{i}$ is semi-positive
definite. After concatenating the chain, the terms in $\rho^{Q}_{i}$,
$\gamma^{x_{1}\to i}_{Q}e^{(1)}(e^{(1)})^{*}(\gamma^{x_{1}\to
i}_{Q})^{*}\textrm{ and }\gamma^{x_{k}\to
i}_{Q}e^{(k)}(e^{(k)})^{*}(\gamma^{x_{k}\to i}_{Q})^{*}$
where $\gamma^{x_{1}\to i}_{Q}$ ($\gamma^{x_{k}\to i}_{Q}$ resp.) is a path
from $x_{1}$ (from $x_{k}$ resp.) to $i$, get affected. Namely, the moment map
equation for $e^{(1)}(e^{(1)})^{*}$ (or $e^{(k)}(e^{(k)})^{*}$) gets an extra
term $(a_{1}^{\prime})^{*}a_{1}^{\prime}$
($-a_{k-1}^{\prime}(a_{k-1}^{\prime})^{*}$ resp.). (If $k=1$, $x_{1}=x_{k}$
and $a^{\prime}_{1}=a^{\prime}_{0}$, and the moment map equation for
$e^{(1)}(e^{(1)})^{*}$ gets both the extra terms
$(a_{0}^{\prime})^{*}a_{0}^{\prime}$ and
$-a_{0}^{\prime}(a_{0}^{\prime})^{*}$.) As a result, we have an extra negative
term
$-\gamma^{x_{k}\to
i}_{Q}a_{k-1}^{\prime}(a_{k-1}^{\prime})^{*}(\gamma^{x_{k}\to i}_{Q})^{*}$
(3.6)
for each path $\gamma^{x_{k}\to i}_{Q}$. We shall show that these negative
terms can be canceled.
We also have additional paths in $Q^{\prime}$ heading to $i$, which can be
divided into the following types:
1. (1)
$\gamma_{Q}^{x_{k}\to i}a_{k-1}^{\prime}\ldots a_{l}^{\prime}e^{(x_{l})}$ for
$l=2,\ldots,k-1$. (This is an empty case when $k=1,2$.)
2. (2)
$\gamma_{Q}^{x_{k}\to i}\gamma_{\mathrm{new}}\gamma_{Q}^{j\to x_{1}}e_{j}$
where $\gamma_{\mathrm{new}}:=a_{k-1}^{\prime}\ldots a_{1}^{\prime}$ and
$\gamma_{Q}^{j\to x_{1}}$ is any path in $Q$ from $j\in Q_{0}$ to $x_{1}$.
($\gamma_{Q}^{x_{1}\to x_{1}}$ can be the trivial path.
$\gamma_{\mathrm{new}}=a_{0}^{\prime}$ when $k=1$.)
3. (3)
$\gamma_{Q}^{x_{k}\to
i}\left(\prod_{r=1}^{p}\gamma_{\mathrm{new}}\gamma_{Q,r}^{x_{k}\to
x_{1}}\right)a_{k-1}^{\prime}\ldots a_{l}^{\prime}e^{(x_{l})}$ for some $p>0$
and $l=2,\ldots,k-1$.
4. (4)
$\gamma_{Q}^{x_{k}\to
i}\left(\prod_{r=1}^{p}\gamma_{\mathrm{new}}\gamma_{Q,r}^{x_{k}\to
x_{1}}\right)\gamma_{\mathrm{new}}\gamma_{Q}^{j\to x_{1}}e_{j}$ for some $p>0$
and $j\in Q_{0}$.
For (1), by the moment-map equation
$e^{(x_{l})}(e^{(x_{l})})^{*}=I+(a^{\prime}_{l})^{*}a^{\prime}_{l}-a^{\prime}_{l-1}(a^{\prime}_{l-1})^{*}$,
we have
$\displaystyle\gamma_{Q}^{x_{k}\to i}a_{k-1}^{\prime}\ldots
a_{l}^{\prime}\cdot e^{(x_{l})}(e^{(x_{l})})^{*}\cdot(\gamma_{Q}^{x_{k}\to
i}a_{k-1}^{\prime}\ldots a_{l}^{\prime})^{*}$ $\displaystyle=$
$\displaystyle\gamma_{Q}^{x_{k}\to i}a_{k-1}^{\prime}\ldots
a_{l}^{\prime}(\gamma_{Q}^{x_{k}\to i}a_{k-1}^{\prime}\ldots
a_{l}^{\prime})^{*}+\gamma_{Q}^{x_{k}\to i}a_{k-1}^{\prime}\ldots
a_{l}^{\prime}(a^{\prime}_{l})^{*}a^{\prime}_{l}(\gamma_{Q}^{x_{k}\to
i}a_{k-1}^{\prime}\ldots a_{l}^{\prime})^{*}$
$\displaystyle-\gamma_{Q}^{x_{k}\to i}a_{k-1}^{\prime}\ldots
a_{l}^{\prime}a^{\prime}_{l-1}(a^{\prime}_{l-1})^{*}(\gamma_{Q}^{x_{k}\to
i}a_{k-1}^{\prime}\ldots a_{l}^{\prime})^{*}.$
The first term above for $l=k-1$ cancel with the extra negative term (3.6) for
$\rho_{i}^{Q}$. The third term (which is negative) for
$l\in\\{3,\ldots,k-1\\}$ cancel with the first term of $l-1$. As a result,
after combining (1) with the modified $\rho_{i}^{Q}$, the remaining negative
terms are
$-\gamma_{Q}^{x_{k}\to i}a_{k-1}^{\prime}\ldots
a^{\prime}_{1}(\gamma_{Q}^{x_{k}\to i}a_{k-1}^{\prime}\ldots
a^{\prime}_{1})^{*}=-\gamma_{Q}^{x_{k}\to
i}\gamma_{\mathrm{new}}(\gamma_{Q}^{x_{k}\to i}\gamma_{\mathrm{new}})^{*}.$
(3.7)
(For the case $k=1$, this trivially holds since
$\gamma_{\mathrm{new}}=a^{\prime}_{0}$, and the above equals to (3.6).)
Now consider (2): $\gamma_{Q}^{x_{k}\to
i}\gamma_{\mathrm{new}}\gamma_{Q}^{j\to
x_{1}}e_{j}e_{j}^{*}(\gamma_{Q}^{x_{k}\to
i}\gamma_{\mathrm{new}}\gamma_{Q}^{j\to x_{1}})^{*}$. The moment map equation
for $e_{j}e_{j}^{*}$ when $j\not=x_{1},x_{k}$ are the same for $Q$ and
$Q^{\prime}$. Summing $\gamma_{Q}^{x_{k}\to
i}\gamma_{\mathrm{new}}\gamma_{Q}^{j\to
x_{1}}e_{j}e_{j}^{*}(\gamma_{Q}^{x_{k}\to
i}\gamma_{\mathrm{new}}\gamma_{Q}^{j\to x_{1}})^{*}$ over arbitrary
$\gamma_{Q}^{j\to x_{1}}$ and $j\in Q_{0}$, we obtain
$(\gamma_{Q}^{x_{k}\to
i}\gamma_{\mathrm{new}})\cdot\rho^{Q}_{x_{1}}\cdot(\gamma_{Q}^{x_{k}\to
i}\gamma_{\mathrm{new}})^{*}=(\gamma_{Q}^{x_{k}\to
i}\gamma_{\mathrm{new}})\cdot(I+B_{1})\cdot(\gamma_{Q}^{x_{k}\to
i}\gamma_{\mathrm{new}})^{*}$
plus
$\sum_{\gamma_{Q}^{x_{1}\to x_{1}}}\gamma_{Q}^{x_{k}\to
i}\gamma_{\mathrm{new}}\gamma_{Q}^{x_{1}\to
x_{1}}(a_{1}^{\prime})^{*}a_{1}^{\prime}(\gamma_{Q}^{x_{k}\to
i}\gamma_{\mathrm{new}}\gamma_{Q}^{x_{1}\to
x_{1}})^{*}-\sum_{\gamma_{Q}^{x_{k}\to x_{1}}}\gamma_{Q}^{x_{k}\to
i}\gamma_{\mathrm{new}}\gamma_{Q}^{x_{k}\to
x_{1}}a_{k-1}^{\prime}(a_{k-1}^{\prime})^{*}(\gamma_{Q}^{x_{k}\to
i}\gamma_{\mathrm{new}}\gamma_{Q}^{x_{k}\to x_{1}})^{*}$
which is due to the additional terms in the moment map equations for
$e^{(x_{1})}(e^{(x_{1})})^{*}$ and $e^{(x_{k})}(e^{(x_{k})})^{*}$. (When
$k=1$, $\gamma_{Q}^{x_{0}\to x_{0}}$ being the trivial path at $x_{0}$ is one
of the possibilities.) In above, $B_{1}$ is semi-positive-definite by the
assumption on $Q$. Then the negative terms (3.7) cancel with the first term.
After combining the paths in $Q$ and (1) and (2), the remaining negative terms
are
$-\sum_{\gamma_{Q}^{x_{k}\to x_{1}}}\gamma_{Q}^{x_{k}\to
i}\gamma_{\mathrm{new}}\gamma_{Q}^{x_{k}\to
x_{1}}a_{k-1}^{\prime}(a_{k-1}^{\prime})^{*}(\gamma_{Q}^{x_{k}\to
i}\gamma_{\mathrm{new}}\gamma_{Q}^{x_{k}\to x_{1}})^{*}.$ (3.8)
(If there is no path $\gamma_{Q}^{x_{k}\to x_{1}}$ in $Q$ from $x_{k}$ to
$x_{1}$, then this is zero, and we do not have (3) nor (4). We stop here and
get that $\rho_{i}^{Q^{\prime}}(\rho_{i}^{Q^{\prime}})^{*}=I+B_{i}^{\prime}$
for a semi-positive-definite matrix $B_{i}^{\prime}$.)
The terms in (3) for $p=1$ and $l=k-1$ cancel with the above (3.8):
$\displaystyle\gamma_{Q}^{x_{k}\to
i}\left(\prod_{r=1}^{p}\gamma_{\mathrm{new}}\gamma_{Q,r}^{x_{k}\to
x_{1}}\right)a_{k-1}^{\prime}\ldots a_{l}^{\prime}\cdot
e^{(x_{l})}(e^{(x_{l})})^{*}\cdot\left(\gamma_{Q}^{x_{k}\to
i}\left(\prod_{r=1}^{p}\gamma_{\mathrm{new}}\gamma_{Q,r}^{x_{k}\to
x_{1}}\right)a_{k-1}^{\prime}\ldots a_{l}^{\prime}\right)^{*}$
$\displaystyle=$ $\displaystyle\gamma_{Q}^{x_{k}\to
i}\left(\prod_{r=1}^{p}\gamma_{\mathrm{new}}\gamma_{Q,r}^{x_{k}\to
x_{1}}\right)a_{k-1}^{\prime}\ldots a_{l}^{\prime}\left(\gamma_{Q}^{x_{k}\to
i}\left(\prod_{r=1}^{p}\gamma_{\mathrm{new}}\gamma_{Q,r}^{x_{k}\to
x_{1}}\right)a_{k-1}^{\prime}\ldots a_{l}^{\prime}\right)^{*}$
$\displaystyle+\gamma_{Q}^{x_{k}\to
i}\left(\prod_{r=1}^{p}\gamma_{\mathrm{new}}\gamma_{Q,r}^{x_{k}\to
x_{1}}\right)a_{k-1}^{\prime}\ldots
a_{l}^{\prime}(a^{\prime}_{l})^{*}a^{\prime}_{l}\left(\gamma_{Q}^{x_{k}\to
i}\left(\prod_{r=1}^{p}\gamma_{\mathrm{new}}\gamma_{Q,r}^{x_{k}\to
x_{1}}\right)a_{k-1}^{\prime}\ldots a_{l}^{\prime}\right)^{*}$
$\displaystyle-\gamma_{Q}^{x_{k}\to
i}\left(\prod_{r=1}^{p}\gamma_{\mathrm{new}}\gamma_{Q,r}^{x_{k}\to
x_{1}}\right)a_{k-1}^{\prime}\ldots
a_{l}^{\prime}a^{\prime}_{l-1}(a^{\prime}_{l-1})^{*}\left(\gamma_{Q}^{x_{k}\to
i}\left(\prod_{r=1}^{p}\gamma_{\mathrm{new}}\gamma_{Q,r}^{x_{k}\to
x_{1}}\right)a_{k-1}^{\prime}\ldots a_{l}^{\prime}\right)^{*}.$
Like in (1), for each $p$, the third term (which is negative) for
$l\in\\{3,\ldots,k-1\\}$ cancel with the first term for $l-1$. (When $p=0$ and
$l=k-1$, the third term is exactly (3.8).) Then the remaining negative terms
are (3.7) modified by inserting the loops
$\left(\prod_{r=1}^{p}\gamma_{\mathrm{new}}\gamma_{Q,r}^{x_{k}\to
x_{1}}\right)$ for $p>0$, that is, $-\gamma_{Q}^{x_{k}\to
i}\left(\prod_{r=1}^{p}\gamma_{\mathrm{new}}\gamma_{Q,r}^{x_{k}\to
x_{1}}\right)\gamma_{\mathrm{new}}(\gamma_{Q}^{x_{k}\to
i}\left(\prod_{r=1}^{p}\gamma_{\mathrm{new}}\gamma_{Q,r}^{x_{k}\to
x_{1}}\right)\gamma_{\mathrm{new}})^{*}$. Then like in (2), these negative
terms cancel with terms in (4). Summing up to finite $p$, the only negative
terms left are (3.8) modified by inserting the loops
$\left(\prod_{r=1}^{p}\gamma_{\mathrm{new}}\gamma_{Q,r}^{x_{k}\to
x_{1}}\right)$:
$-\sum_{\gamma_{Q}^{x_{k}\to x_{1}}}\gamma_{Q}^{x_{k}\to
i}\left(\prod_{r=1}^{p}\gamma_{\mathrm{new}}\gamma_{Q,r}^{x_{k}\to
x_{1}}\right)\gamma_{\mathrm{new}}\gamma_{Q}^{x_{k}\to
x_{1}}a_{k-1}^{\prime}(a_{k-1}^{\prime})^{*}\left(\gamma_{Q}^{x_{k}\to
i}\left(\prod_{r=1}^{p}\gamma_{\mathrm{new}}\gamma_{Q,r}^{x_{k}\to
x_{1}}\right)\gamma_{\mathrm{new}}\gamma_{Q}^{x_{k}\to x_{1}}\right)^{*}$
which cancel with terms in (3) for $(p+1)$. As $p\to\infty$,
$\|\left(\prod_{r=1}^{p}\gamma_{\mathrm{new}}\gamma_{Q,r}^{x_{k}\to
x_{1}}\right)\|\to 0$. This finishes the proof that
$\rho^{Q^{\prime}}_{i}(\rho^{Q^{\prime}}_{i})^{*}=I+B_{i}$, for $i\in Q_{0}$.
For the case that $i=x_{j}$ for $j=2,\ldots,k-1$, the proof is similar. (We do
not need to consider this case when $k=1,2$.) The paths in $Q^{\prime}$
heading to $x_{j}$ are divided into the following types:
1. (1)
$a_{j-1}^{\prime}\ldots a_{l}^{\prime}e^{(x_{l})}$ for $l=2,\ldots,j$.
2. (2)
$a_{j-1}^{\prime}\ldots
a_{1}^{\prime}\left(\prod_{r=1}^{p}\gamma_{Q,r}^{x_{k}\to
x_{1}}\gamma_{\mathrm{new}}\right)\gamma_{Q}^{j\to x_{1}}e_{j}$ for some
$p\geq 0$, $j\in Q_{0}$.
3. (3)
$a_{j-1}^{\prime}\ldots
a_{1}^{\prime}\left(\prod_{r=1}^{p}\gamma_{Q,r}^{x_{k}\to
x_{1}}\gamma_{\mathrm{new}}\right)\gamma_{Q}^{x_{k}\to
x_{1}}a_{k-1}^{\prime}\ldots a_{l}^{\prime}e^{(x_{l})}$ for some $p\geq 0$,
$l=2,\ldots,k-1$.
The cancellation is similar and we do not repeat here. ∎
Similarly, adding a chain at a single vertex of $Q$ preserves the positive-
definiteness property.
###### Lemma 3.22.
Suppose $Q$ as in Lemma 3.21. Let $Q^{\prime}$ be obtained by concatenating to
$Q$ a chain $x_{1}\stackrel{{\scriptstyle
a_{1}^{\prime}}}{{\to}}\ldots\stackrel{{\scriptstyle
a_{k-1}^{\prime}}}{{\to}}x_{k}$ at either $x_{1}$ or $x_{k}$ in $Q_{0}$. Then
for every $i\in Q^{\prime}_{0}$,
$\rho^{Q^{\prime}}_{i}(\rho^{Q^{\prime}}_{i})^{*}=I+B_{i}^{\prime}$ for some
semi-positive-definite matrix $B_{i}^{\prime}$.
###### Proof.
The proof in this case is simpler than that of Lemma 3.21, since there is no
new oriented cycle.
Consider the case that $x_{k}\in Q_{0}$. If $i$ belongs to the chain, then the
only paths that head to $i$ are contained in the chain. Since no oriented
cycle is involved in all such paths, Theorem 3.7 already gives the result.
If $i\in Q_{0}$, then
$\displaystyle\rho^{Q^{\prime}}_{i}(\rho^{Q^{\prime}}_{i})^{*}=$
$\displaystyle\sum_{j\in Q_{0}-\\{x_{k}\\}}\sum_{\gamma_{Q}^{j\to
i}}(\gamma_{Q}^{j\to i}e^{(j)})(\gamma_{Q}^{j\to
i}e^{(j)})^{*}+\sum_{\gamma_{Q}^{x_{k}\to i}}\gamma_{Q}^{x_{k}\to i}\cdot
e^{(x_{k})}(e^{(x_{k})})^{*}\cdot(\gamma_{Q}^{x_{k}\to i})^{*}$
$\displaystyle+\sum_{\gamma_{Q}^{x_{k}\to
i}}\sum_{r=1}^{k-1}(\gamma_{Q}^{x_{k}\to i}a_{k-1}\ldots
a_{r})e_{x_{r}}e_{x_{r}}^{*}\cdot(\gamma_{Q}^{x_{k}\to
i}a^{\prime}_{k-1}\ldots a^{\prime}_{r})^{*}$ $\displaystyle=$
$\displaystyle\rho_{i}^{Q}(\rho_{i}^{Q})^{*}-\sum_{\gamma_{Q}^{x_{k}\to
i}}\gamma_{Q}^{x_{k}\to
i}(a^{\prime}_{k-1}(a^{\prime}_{k-1})^{*})(\gamma_{Q}^{x_{k}\to i})^{*}$
$\displaystyle+\sum_{\gamma_{Q}^{x_{k}\to
i}}\sum_{r=1}^{k-1}(\gamma_{Q}^{x_{k}\to i}a^{\prime}_{k-1}\ldots
a^{\prime}_{r})(I-a^{\prime}_{r-1}(a^{\prime}_{r-1})^{*}+(a^{\prime}_{r})^{*}a^{\prime}_{r})\cdot(\gamma_{Q}^{x_{k}\to
i}a^{\prime}_{k-1}\ldots a^{\prime}_{r})^{*}$ $\displaystyle=$
$\displaystyle\rho_{i}^{Q}(\rho_{i}^{Q})^{*}+\sum_{\gamma_{Q}^{x_{k}\to
i}}\sum_{r=1}^{k-1}(\gamma_{Q}^{x_{k}\to i}a^{\prime}_{k-1}\ldots
a^{\prime}_{r})(a^{\prime}_{r})^{*}a^{\prime}_{r}\cdot(\gamma_{Q}^{x_{k}\to
i}a^{\prime}_{k-1}\ldots a^{\prime}_{r})^{*}.$
Since $\rho_{i}^{Q}(\rho_{i}^{Q})^{*}=I+B_{i}$ for some semi-positive-definite
matrix $B_{i}$, and the second term is semi-positive-definite,
$\rho^{Q^{\prime}}_{i}(\rho^{Q^{\prime}}_{i})^{*}$ satisfies the requirement.
The case that $x_{1}\in Q_{0}$ is similar and the proof is omitted. ∎
###### Proof of Theorem 3.18.
Without loss of generality, suppose $Q$ is connected. (Otherwise
$\mathcal{M}_{\vec{n},\vec{d}}$ and $\mathcal{V}_{i}$ decompose into products
coming from the connected components, and we just need to study each
component.) The case without oriented cycle is given in Theorem 3.7. Suppose
$Q$ has at least one oriented cycle. By Lemma 3.20, the statement is true for
this oriented cycle as a quiver. There must be additional arrows if this
single oriented cycle is not yet the whole $Q$. Then we can either add a chain
as in Lemma 3.21 or 3.22, and the statement still holds. (Both the cases of
loop at a vertex or multiple edge are covered by Lemma 3.21.) Inductively the
statement holds for $Q$. ∎
###### Example 3.23.
For the $A_{2}$-quiver, $\rho_{i}\rho_{i}^{*}$ gives a metric for all $i$. By
Lemma 3.21, this is still true if we add an oriented cycle with arrows
$l_{1},\ldots,l_{p}$. See Figure 3.
${1}$${4}$${2}$${5}$${3}$$\scriptstyle{e^{(1)}}$$\scriptstyle{a_{1}}$$\scriptstyle{e_{2,1}}$$\scriptstyle{e_{2}}$$\scriptstyle{a_{2}}$$\scriptstyle{\ell_{1}}$$\scriptstyle{e_{3}}$$\scriptstyle{\ell_{p}}$$\scriptstyle{e_{2,p-1}}$
Figure 3. $A_{2}$ modified by adding an oriented cycle.
Note that the following equality still holds over $R^{\circ}_{n,d}$:
$\rho\rho^{*}=e^{(i)}(e^{(i)})^{*}+\sum_{h(a)=i}V_{a}\rho_{(t(a))}\rho_{(t(a))}^{*}V_{a}^{*}.$
Thus Proposition 3.8 still holds for quivers with oriented cycles.
###### Proposition 3.24.
For any quiver $Q$ and every $v,w\in(\mathcal{V}_{i})^{*}$,
$H_{i}(v,w)=H_{0}((e^{(i)})^{*}(v),(e^{(i)})^{*}(w))+\sum_{h(a)=i}H_{t(a)}(a^{*}(v),a^{*}(w)).$
Now we consider a version of Theorem 3.15 in this case. We define
$\hat{W}_{i}$ by Equation (3.4). But this time, it is an infinite direct sum
of Hilbert spaces (meaning that it consists of infinite sequence
$w=(w_{\gamma}:h(\gamma)=i)$ with
$\|w\|^{2}=\sum_{\gamma}\|w_{\gamma}\|^{2}<\infty$).
###### Lemma 3.25.
For each $(V,e)\in R^{\circ}_{n,d}$, $\rho_{i}(V,e)$ defines a bounded linear
map $\hat{W}_{i}\to V_{i}$. Its adjoint
$\rho_{i}(V,e)^{*}:V_{i}\to\hat{W}_{i}$ has a singular-value decomposition.
###### Proof.
For $w=(w_{\gamma}:h(\gamma)=i)\in\hat{W}_{i}$, $\rho_{i}(V,e)$ maps it to
$\sum_{\gamma:h(\gamma)=i}V_{\gamma}e^{(t(\gamma))}w_{\gamma}.$
Like in the proof of Lemma 3.19, consider
$\sum_{\gamma:h(\gamma)=i}\|V_{\gamma}e^{(t(\gamma))}\|\|w_{\gamma}\|\leq\|w\|\sum_{\gamma:h(\gamma)=i}\|V_{\gamma}e^{(t(\gamma))}\|<+\infty.$
This also shows that if $\|w\|=1$, then the image of $w$ is also bounded.
$\rho_{i}(V,e)^{*}$ has image being finite-dimensional (since $V_{i}$ is
finite-dimensional), and hence is a compact operator. Thus it has a singular-
value decomposition. ∎
###### Proposition 3.26.
The Ricci curvature of the metric given by $(\rho_{i}\rho_{i}^{*})^{-1}$ is
semi-positive definite on $\mathcal{M}^{\circ}$.
###### Proof.
By the previous lemma, the proof of Theorem 3.15 on semi-positive definiteness
still works. Namely,
$\partial\overline{\partial}\log\det\rho\rho^{\mathrm{*}}=\mathrm{tr~{}}\left(\left(\partial\rho\right)^{\mathrm{*}}\left(\rho\rho^{\mathrm{*}}\right)^{-1}\partial\rho\right)-\mathrm{tr~{}}\left(\left(\partial\rho\cdot\left(\rho^{\mathrm{*}}\left(\rho\rho^{\mathrm{*}}\right)^{-\frac{1}{2}}\right)\right)^{\mathrm{*}}\left(\rho\rho^{\mathrm{*}}\right)^{-1}\left(\partial\rho\cdot\left(\rho^{\mathrm{*}}\left(\rho\rho^{\mathrm{*}}\right)^{-\frac{1}{2}}\right)\right)\right).$
Note that the two terms on the RHS are finite:
$\mathrm{tr~{}}\left(\left(\partial\rho\right)^{\mathrm{*}}\left(\rho\rho^{\mathrm{*}}\right)^{-1}\partial\rho\right)=\mathrm{tr~{}}\left(\rho\rho^{\mathrm{*}}\right)^{-1}\partial\rho\left(\left(\partial\rho\right)^{\mathrm{*}}\right)=\sum_{j=1}^{d_{i}}\langle\partial\rho\left(\partial\rho\right)^{\mathrm{*}}\epsilon_{j},\left(\rho\rho^{\mathrm{*}}\right)^{-1}\epsilon_{j}\rangle_{V_{i}}$
which is a finite sum, and similar for the second term.
$\rho^{\mathrm{*}}=\left(\rho^{\mathrm{*}}\left(\rho\rho^{\mathrm{*}}\right)^{-\frac{1}{2}}\right)\left(\rho\rho^{\mathrm{*}}\right)^{\frac{1}{2}}$
is decomposed into the rescaling $(\rho\rho^{\mathrm{*}})^{\frac{1}{2}}$ and
the orthogonal embedding
$\left(\rho^{\mathrm{*}}\left(\rho\rho^{\mathrm{*}}\right)^{-\frac{1}{2}}\right)$
to $\mathrm{Im}\,\rho^{\mathrm{*}}\subset\widehat{W}_{i}$. Then the above
equals to $\|(\partial\rho)_{2}\|_{H}^{2}\geq 0$ as in Theorem 3.15. ∎
## 4\. Fiberwise Nonlinearity
In the mathematical study of quivers, we mostly focused on linear
representations. In particular, the morphisms between universal vector bundles
are linear along fibers. On the other hand, nonlinear ‘activation functions’
play a key role in machine learning. In this section, we construct some
natural non-linear fiber-bundle endomorphisms of the universal bundles
$\mathcal{V}_{i}$ over $\mathcal{M}_{\vec{n},\vec{d}}$ by using fiberwise
symplectomorphisms.
For simplicity, we shall take $n_{i}=d_{i}+1$ for all $i\in Q_{0}$ in this
section.
### 4.1. Activation functions arising from toric moment maps and
symplectomorphisms
In this section, we make the observation that several activation functions
commonly used in machine learning actually belong to a much bigger class,
namely the $T$-equivariant symplectomorphisms on open subsets of a symplectic
toric variety.
First, let’s recall the basic setup for toric varieties. Let’s equip
$\mathbb{C}^{m}$ with the standard Kähler structure. We obtain a symplectic
toric variety $(X,\omega_{X})$ as a symplectic quotient by the real torus
$T^{m-d}$. We assume $X$ is smooth. The $T^{m-d}$-action can be specified by
an injective homomorphism $\mathbb{Z}^{m-d}\to\mathbb{Z}^{m}$, which induces a
map $T^{m-d}\to T^{m}$, and $T^{m}$ acts on $\mathbb{C}^{m}$ by coordinate-
wise multiplication. We assume that the quotient of $\mathbb{Z}^{m}$ by the
image of $\mathbb{Z}^{m-d}$ is again a lattice, which we identify as
$\mathbb{Z}^{d}$. We denote by $v_{i}\in\mathbb{Z}^{d}$ the images of the
standard basic vectors of $\mathbb{Z}^{m}$ under the quotient map
$\mathbb{Z}^{m}\to\mathbb{Z}^{d}$.
The residual action of $T^{d}$ on $X$ gives a moment-map fibration over a
polytope $P$, which is given by the intersection of $m$ half-spaces in
$\mathbb{R}^{d}$:
$\\{x\in\mathbb{R}^{d}:\ell_{j}(x):=v_{j}\cdot x-c_{j}\geq 0\\}$
where the constants $c_{j}\in\mathbb{R}^{d}$ are determined by the level taken
in the symplectic quotient. We assume that the level is chosen such that for
all $j$, $\\{\ell_{j}(x)=0\\}\cap P$ is a (non-empty) codimension-one boundary
of the polytope $P$.
Consider the open toric orbit of $X$, which can be identified as
$(\mathbb{C}^{\times})^{d}$ by fixing a basis of $\mathbb{Z}^{d}$. Denote by
$\omega_{X}$ the Kähler form induced on the symplectic quotient. Let
$\omega_{\mathrm{std}}=\sum_{i=1}^{d}dx_{i}\wedge d\theta_{i}$ be the standard
symplectic form on $\mathbb{R}^{d}\times T^{d}$ (where $T^{d}$ denotes the
real $d$-torus). The symplectic form $\omega_{X}|_{(\mathbb{C}^{\times})^{d}}$
has an explicit description by the following beautiful formula.
###### Theorem 4.1 ([Gui94, Abr98]).
$\left(\frac{1}{2}\left(\sum_{i=1}^{m}v_{i}\log\ell_{i}(x)\right),\mathrm{Id}\right):(P^{\circ}\times
T^{d},\omega_{\mathrm{std}}|_{P^{\circ}\times T^{d}})\to\mathbb{R}^{d}\times
T^{d}\stackrel{{\scriptstyle\exp}}{{\cong}}((\mathbb{C}^{\times})^{d},\omega_{X}|_{(\mathbb{C}^{\times})^{d}})$
is a symplectomorphism.
Taking the universal cover $\mathbb{R}^{d}\to T^{d}$ and lifting the above,
one obtains the following.
###### Corollary 4.2.
The inverse of
$\left(\frac{1}{2}\left(\sum_{i=1}^{m}v_{i}\log\ell_{i}(x)\right),\,\mathrm{Id}\right)$
gives a symplectomorphism
$\sigma_{\mathbb{C}}=(\sigma(\mathrm{Re}(\vec{z})),\mathrm{Im}(\vec{z})):(\mathbb{C}^{d},\exp^{*}\omega_{X})\to(P^{\circ}\times\mathbb{R}^{d},\omega_{\mathrm{std}})$
where $\exp:\mathbb{C}^{d}\to(\mathbb{C}^{\times})^{d}\subset X$, and
$\sigma:\mathbb{R}^{d}\to P^{\circ}$ is the inverse of
$\left(\frac{1}{2}\left(\sum_{i=1}^{m}v_{i}\log\ell_{i}(x)\right),\,\mathrm{Id}\right)$.
###### Example 4.3.
For $\mathbb{C}^{d}$, the moment polytope $P$ is $\mathbb{R}_{\geq
0}^{d}=\\{x_{i}\geq 0:i=1,\ldots,d\\}$. $v=(1,\ldots,1)$. The above map is
simply $\left(\frac{\log
x_{i}}{2}\right)_{i=1}^{d}:\mathbb{R}_{>0}^{d}\to\mathbb{R}^{d}$. The
symplectomorphism
$((\mathbb{C}^{\times})^{d},\omega_{\mathbb{C}^{n}}|_{\mathbb{C}^{\times})^{d}})\stackrel{{\scriptstyle\cong}}{{\to}}(\mathbb{R}_{>0}^{d}\times
T^{d},\omega_{\mathrm{std}}|_{P^{\circ}\times T^{d}})$ is
$\left(|z_{i}|^{2},\frac{\log z_{i}-\log\bar{z_{i}}}{2i}\right)_{i=1}^{d}$.
###### Example 4.4.
For the complex projective space $\mathbb{P}^{d}$, the corresponding moment
polytope $P$ (for a chosen level) is the $d$-simplex given by $\ell_{i}\geq 0$
where $\ell_{i}(x)=x_{i}$ for $i=1,\ldots,d$, and
$\ell_{d+1}(x)=1-x_{1}-x_{2}-\dots-x_{d}$. The generators are
$v_{i}=\epsilon_{i}$ for $i=1,\ldots,d$ (the standard basis) and
$v_{d+1}=-\sum_{i=1}^{d}v_{i}$. We have
$d\left(\frac{1}{2}\sum_{i=1}^{d+1}\ell_{i}(x)\log\ell_{i}(x)\right)=\frac{1}{2}\sum_{i=1}^{d}\log\left(\frac{x_{i}}{1-\sum_{j=1}^{d}x_{j}}\right)dx_{i}$
as a map $P^{\circ}\to\mathbb{R}^{d}$. By direct computation, the inverse of
this map equals to
$\sigma(\vec{r})=\left(\frac{e^{2r_{i}}}{1+\sum_{j=1}^{d}e^{2r_{j}}}\right)_{i=1}^{d}:\mathbb{R}^{d}\to
P^{\circ}.$ (4.1)
Written in terms of the complex coordinates
$\vec{z}\in(\mathbb{C}^{\times})^{d}$, the symplectomorphism
$((\mathbb{C}^{\times})^{d},\omega_{\mathbb{P}^{n}}|_{(\mathbb{C}^{\times})^{d}})\to(P^{\circ}\times
T^{d},\omega_{\mathrm{std}}|_{P^{\circ}\times T^{d}})$ is given by
$\left(\frac{|z_{i}|^{2}}{1+\sum_{j=1}^{d}|z_{j}|^{2}},\frac{\log
z_{i}-\log\bar{z_{i}}}{2i}\right)_{i=1}^{d}.$
Pulling back by $\mathbb{C}^{d}\to(\mathbb{C}^{\times})^{d}$, we have the
symplectomorphism
$\sigma_{\mathbb{C}}=\left(\left(\frac{e^{2r_{i}}}{1+\sum_{j=1}^{d}e^{2r_{j}}}\right)_{i=1}^{d},\mathrm{Id}\right):(\mathbb{C}^{d},\exp^{*}\omega_{X})\to(P^{\circ}\times\mathbb{R}^{d},\omega_{\mathrm{std}})$.
When $d=1$, $\frac{e^{2r}}{1+e^{2r}}$ is a commonly-used activation function.
See Figure 4. By taking a direct product,
$\left(\frac{e^{2r_{i}}}{1+e^{2r_{i}}}\right)_{i=1}^{d}:\mathbb{R}^{d}\to[0,1]^{d}$
corresponds to $(\mathbb{P}^{1})^{d}$.
$/2\pi i\mathbb{Z}$$\cong$$\left(\frac{e^{2x}}{1+e^{2x}},y\right)$ Figure 4.
$\mathbb{C}$ as a Covering Space Mapped to an Open Strip
###### Remark 4.5.
In above, we have taken the quotient Kähler structure from $\mathbb{C}^{m}$.
For a general toric Kähler structure, the symplectomorphism in Theorem 4.1 is
given by
$\left(d\left(\frac{1}{2}\left(\sum_{i=1}^{m}\ell_{i}\log\ell_{i}(x)\right)+h\right),\mathrm{Id}\right)$
where $h$ is a smooth function on the closed polytope $P$ such that the
Hessian of
$\left(\frac{1}{2}\left(\sum_{i=1}^{m}\ell_{i}\log\ell_{i}(x)\right)+h\right)$
is positive definite in $P^{\circ}$ [Abr98].
In particular, for a general projective toric variety $X$, we can take an
embedding of $X$ to $\mathbb{P}^{N}$ by toric holomorphic sections of a very
ample line bundle $L$, and use the induced toric Kähler structure from
$\mathbb{P}^{N}$. Then the symplectomorphism $\sigma_{\mathbb{C}}$ is given by
$(\sigma,\mathrm{Id})$ where
$\sigma(\vec{r})=\frac{\sum_{i=1}^{d}e^{2(\vec{u_{i}},\vec{r})}\vec{u_{i}}}{\sum_{j=1}^{d}e^{2(\vec{u_{j}},\vec{r})}}:\mathbb{R}^{d}\to
P^{\circ},$
$\vec{u_{i}}$ are points such that their convex hull equal to $P$, and
$(\vec{u},\vec{r})$ is the standard dot product on $\mathbb{R}^{d}$. See
[Ful93, Section 4.2].
Now we have the symplectomorphisms $\phi_{X}:(P^{\circ}\times
T^{d},\omega_{\mathrm{std}}|_{P^{\circ}\times
T^{d}})\stackrel{{\scriptstyle\cong}}{{\to}}((\mathbb{C}^{\times})^{d},\omega_{X}|_{(\mathbb{C}^{\times})^{d}})$
and $\phi_{\mathbb{C}^{d}}:\mathbb{R}_{>0}^{d}\times
T^{d}\stackrel{{\scriptstyle\cong}}{{\to}}((\mathbb{C}^{\times})^{d},\omega_{\mathbb{C}^{d}}|_{(\mathbb{C}^{\times})^{d}})$
(Example 4.3). For the toric structure of $X$, let’s arrange the order of the
indices such that the first $d$ vectors $v_{i}$ for $i=1,\ldots,d$ form a
basis of $\mathbb{Z}^{d}$. (We assume $m\geq d$.) Moreover, we take the first
$d$ constants $c_{j}=0$ for $j=1,\ldots,d$. Then
$P^{\circ}\subset\mathbb{R}_{>0}^{d}$.
Consider the composition
$\phi_{\mathbb{C}^{d}}\circ\phi_{X}^{-1}:((\mathbb{C}^{\times})^{d},\omega_{X}|_{(\mathbb{C}^{\times})^{d}})\to((\mathbb{C}^{\times})^{d},\omega_{\mathbb{C}^{d}}|_{(\mathbb{C}^{\times})^{d}})$.
It is a symplectomorphism onto the image
$\phi_{\mathbb{C}^{d}}(P^{\circ}\times T^{d})$.
###### Proposition 4.6.
$\phi_{\mathbb{C}^{d}}\circ\phi_{X}^{-1}$ extends to a $T$-equivariant
symplectomorphism
$\psi:(\mathbb{C}^{d},\omega_{X}|_{\mathbb{C}^{d}})\stackrel{{\scriptstyle\cong}}{{\to}}(\pi_{\mathbb{C}^{d}}^{-1}(P-B),\omega_{\mathbb{C}^{d}}|_{\pi_{\mathbb{C}^{d}}^{-1}(P-B)})$
where $B=\bigcup_{i=d+1}^{m}\\{\ell_{i}(x)=0\\}$, and
$\pi_{\mathbb{C}^{d}}=(|z_{i}|^{2})_{i=1}^{d}:\mathbb{C}^{d}\to\mathbb{R}_{\geq
0}^{d}$ is the moment map for $\mathbb{C}^{d}$.
###### Proof.
$\phi_{X}$ is given by $2r^{X}_{i}=\log
x_{i}+\sum_{j=d+1}^{m}(v_{j}^{(i)}\log\ell_{j}(x))$, where
$v_{j}=(v_{j}^{(1)},\ldots,v_{j}^{(d)})$. $\phi_{\mathbb{C}^{d}}$ is given by
$2r^{\mathbb{C}^{d}}_{i}=\log x_{i}$. Hence
$e^{2r^{X}_{i}}=e^{2r^{\mathbb{C}^{d}}_{i}}\prod_{j=d+1}^{m}\ell_{j}^{v_{j}^{(i)}}(e^{2r^{\mathbb{C}^{d}}_{1}},\ldots,e^{2r^{\mathbb{C}^{d}}_{d}})$.
In terms of the complex coordinates, this gives
$z^{X}_{i}=z^{\mathbb{C}^{d}}_{i}\left(\prod_{j=d+1}^{m}\ell_{j}^{v_{j}^{(i)}}\left(|z^{\mathbb{C}^{d}}_{1}|^{2},\ldots,|z^{\mathbb{C}^{d}}_{d}|^{2}\right)\right)^{1/2}.$
It is obviously well-defined over $\pi_{\mathbb{C}^{d}}^{-1}(P-B)$. We need to
show that it has inverse, which gives the required extension of
$\phi_{\mathbb{C}^{n}}\circ\phi_{X}^{-1}$. Since
$(\phi_{\mathbb{C}^{n}}\circ\phi_{X}^{-1})^{*}(\omega_{\mathbb{C}^{n}})=\omega_{X}$
on $(\mathbb{C}^{\times})^{d}$, this still holds over $\mathbb{C}^{d}$ as the
equality is a closed condition.
Consider the Jacobian of
$\vec{z}^{X}(\vec{z}^{\mathbb{C}^{d}},\overline{\vec{z}^{\mathbb{C}^{d}}})$.
We shall show it is positive definite, and hence invertible. To simplify, we
write $z=z^{\mathbb{C}^{d}}$. Denote
$G=\frac{1}{2}\left(\sum_{i=1}^{m}\ell_{i}(x)\log\ell_{i}(x)-v\cdot x\right)$.
For any non-zero vector $(a_{1},\ldots,a_{d})$,
$\displaystyle\sum_{i,j}\overline{a_{j}}a_{i}\partial_{z_{i}}z_{j}^{X}$
$\displaystyle=$
$\displaystyle\sum_{i,j}\overline{a_{j}}a_{i}\partial_{z_{i}}\left(\exp(\partial_{x_{j}}|_{x_{p}=z_{p}\bar{z_{p}}}G)\cdot\frac{z_{j}}{|z_{j}|}\right)$
$\displaystyle=$
$\displaystyle\sum_{i,j}\overline{a_{j}}a_{i}\exp(\partial_{x_{j}}|_{x_{p}=z_{p}\bar{z_{p}}}G)\cdot\partial_{z_{i}}\left(\frac{z_{j}}{|z_{j}|}\right)+\sum_{i,j}\overline{a_{j}}a_{i}\frac{z_{j}}{|z_{j}|}\cdot\exp(\partial_{x_{j}}|_{x_{p}=z_{p}\bar{z_{p}}}G)\cdot\frac{\partial(z_{i}\bar{z_{i}})}{\partial
z_{i}}\cdot\frac{\partial^{2}G}{\partial x_{i}\partial x_{j}}$
$\displaystyle=$
$\displaystyle\sum_{i}|a_{i}|^{2}|z_{i}^{X}|\cdot\left(\frac{1}{2|z_{i}|}\right)+\sum_{i,j}|z_{j}|^{-1}\cdot|z_{j}^{X}|\cdot\overline{a_{j}}z_{j}\cdot
a_{i}\bar{z_{i}}\cdot\frac{\partial^{2}G}{\partial x_{i}\partial x_{j}}.$
Note that
$|z_{j}|^{-1}|z_{j}^{X}|=\left(\prod_{k=d+1}^{m}\ell_{k}^{v_{k}^{(j)}}\left(|z^{\mathbb{C}^{d}}_{1}|^{2},\ldots,|z^{\mathbb{C}^{d}}_{d}|^{2}\right)\right)^{1/2}$
which is positive. Let $c=\min\\{|z_{j}|^{-1}|z_{j}^{X}|:j=1,\ldots,d\\}$.
Then the second term is no less than
$ca_{i}\bar{z_{i}}\cdot\frac{\partial^{2}G}{\partial x_{i}\partial x_{j}}$.
Since $\frac{\partial^{2}G}{\partial x_{i}\partial x_{j}}$ is positive
definite on $P^{\circ}$, it is semi-positive definite on $P-B$. Thus this term
is non-negative. The first term is positive. Thus
$\sum_{i,j}\overline{a_{j}}a_{i}\partial_{z_{i}}z_{j}^{X}>0$. Similarly
$\sum_{i,j}a_{j}\overline{a_{i}}\overline{\partial_{z_{i}}}z_{j}^{X}$. Hence
the Jacobian is positive-definite and hence invertible. ∎
###### Example 4.7.
We continue to consider $\mathbb{P}^{d}$. From Example 4.4,
$x_{i}=\frac{|z^{\mathbb{P}^{d}}_{i}|^{2}}{1+\sum_{j=1}^{d}|z^{\mathbb{P}^{d}}_{j}|^{2}}$.
$\pi^{-1}_{\mathbb{C}^{d}}(P-B)=\\{\|\vec{z}^{\mathbb{C}^{d}}\|<1\\}$. From
Example 4.3, $x_{i}=|z^{\mathbb{C}^{d}}_{i}|^{2}$. Hence the symplectomorphism
$(\mathbb{C}^{d},\omega_{\mathbb{P}^{d}}|_{\mathbb{C}^{d}})\stackrel{{\scriptstyle\cong}}{{\to}}(\\{\|\vec{z}^{\mathbb{C}^{d}}\|<1\\},\omega_{\mathbb{C}^{n}}|_{\pi_{\mathbb{C}^{d}}^{-1}(P-B)})$
is
$z^{\mathbb{C}^{d}}_{i}=\frac{z^{\mathbb{P}^{d}}_{i}}{\sqrt{1+\|\vec{z}^{\mathbb{P}^{d}}\|^{2}}}.$
(4.2)
When $d=1$, this gives $\frac{z}{\sqrt{1+|z|^{2}}}$ which is another
activation function used in machine learning. ($z$ is restricted in
$\mathbb{R}$ in most algorithms.) The symplectomorphism can be easily checked
in this case: ($z=z^{\mathbb{P}^{1}}$ for simplicity)
$\displaystyle dz^{\mathbb{C}}\wedge\overline{dz^{\mathbb{C}}}=$
$\displaystyle d\frac{z}{\sqrt{1+|z|^{2}}}\wedge
d\frac{\overline{z}}{\sqrt{1+|z|^{2}}}$ $\displaystyle=$
$\displaystyle\frac{(1+z\bar{z})dz-(|z|^{2}dz+z^{2}\overline{dz})/2}{(1+z\bar{z})^{3/2}}\wedge\frac{(1+z\bar{z})\overline{dz}-(|z|^{2}\overline{dz}+\bar{z}^{2}dz)/2}{(1+z\bar{z})^{3/2}}=\frac{dz\wedge\overline{dz}}{(1+z\bar{z})^{2}}$
giving the Fubini-Study metric. See Figure 5.
By taking the direct product, the symplectomorphism for the case
$X=(\mathbb{P}^{1})^{d}$ is
$\left(\frac{z_{j}}{\sqrt{1+|z_{j}|^{2}}}\right)_{i=1}^{d}$.
$\frac{z}{\sqrt{1+|z|^{2}}}$ Figure 5. $\mathbb{C}$ as a Chart in
$\mathbb{P}^{1}$ Mapped to an Open Disk
Due to the nice fact that the Fubini-Study metric on $\mathbb{P}^{d}$ is
$U(d)$-invariant (and so does the standard metric on $\mathbb{C}^{d}$), we
have the following (which is not true for $(\mathbb{P}^{1})^{d}$ nor general
toric manifolds).
###### Lemma 4.8.
For $X=\mathbb{P}^{d}$, the symplectomorphism
$\psi:(\mathbb{C}^{d},\omega_{\mathbb{P}^{d}}|_{\mathbb{C}^{d}})\stackrel{{\scriptstyle\cong}}{{\to}}\left(\pi_{\mathbb{C}^{d}}^{-1}(P-B),\omega_{\mathbb{C}^{n}}|_{\pi_{\mathbb{C}^{d}}^{-1}(P-B)}\right)$
in Proposition 4.6 is $U(d)$-equivariant.
###### Proof.
Using the explicit expression (4.2),
$\psi(U\cdot\vec{z}^{\mathbb{P}^{d}})=\frac{U\cdot\vec{z}^{\mathbb{P}^{d}}}{\sqrt{1+\|U\cdot\vec{z}^{\mathbb{P}^{d}}\|^{2}}}=\frac{U\cdot\vec{z}^{\mathbb{P}^{d}}}{\sqrt{1+\|\vec{z}^{\mathbb{P}^{d}}\|^{2}}}=U\cdot\psi(\vec{z}^{\mathbb{P}^{d}})$
for all $U\in U(d)$. ∎
As explained in Remark 4.5, we can also equip $X$ with another $T$-invariant
Kähler form (that do not come from the standard Kähler structure on
$\mathbb{C}^{m}$). The symplectomorphism is given by
$\left(\frac{1}{2}d\left(\sum_{i=1}^{m}\ell_{i}(x)\log\ell_{i}(x)+h(x)\right),\mathrm{Id}\right):(P^{\circ}\times
T^{d},\omega_{\mathrm{std}}|_{P^{\circ}\times T^{d}})\to\mathbb{R}^{d}\times
T^{d}\stackrel{{\scriptstyle\exp}}{{\cong}}((\mathbb{C}^{\times})^{d},\omega^{h}|_{(\mathbb{C}^{\times})^{d}}).$
Thus the toric construction is rather flexible.
###### Example 4.9.
The ‘softplus’ function $x=\log(1+e^{2y}):\mathbb{R}\to\mathbb{R}_{>0}$ gives
an example of such a Kähler structure on $\mathbb{C}$ (by identifying it with
$\mathbb{R}_{>0}\times\mathbb{R}$ with the standard symplectic structure). The
inverse is $y=\frac{1}{2}\log(e^{x}-1)$, whose difference with
$\frac{1}{2}\log x$ is
$h^{\prime}=\frac{1}{2}\log\frac{e^{x}-1}{x}=\frac{1}{2}\log\left(1+\sum_{k=1}^{\infty}\frac{x^{k}}{(k+1)!}\right)$
which is indeed a smooth function on $\mathbb{R}_{\geq 0}$. Moreover,
$y^{\prime}=\frac{e^{x}}{2(e^{x}-1)}>0$ on $\mathbb{R}_{>0}$.
### 4.2. Symplectomorphisms of fiber bundles over the moduli
In the last subsection, we have exhibited various symplectic embeddings for
$\mathbb{C}^{n}$. Now we want to make a family version of these maps over the
framed quiver moduli $\mathcal{M}_{\vec{n},\vec{d}}$. The last subsection can
be understood as constructing self-maps on a fiber of a vector bundle over
$\mathcal{M}_{\vec{n},\vec{d}}$.
To globalize (4.2), we consider the universal bundle $\mathcal{V}_{i}$
equipped with a Hermitian metric $H_{i}$. (We have constructed a nice
Hermitian metric on $\mathcal{V}_{i}$ in Section 3.1.) We have a fiberwise
symplectic structure $\omega_{\mathcal{V}_{i}}$ induced from the Hermitian
metric. Moreover, we have the projective bundle
$\mathbb{P}(\mathcal{V}_{i}\oplus\mathcal{O}_{\mathcal{M}})$ which is a
fiberwise compactification of $\mathcal{V}_{i}$. Then the fiber bundle
$\mathbb{P}(\mathcal{V}_{i}\oplus\mathcal{O}_{\mathcal{M}})$ is equipped with
a fiberwise Kähler metric
$\omega_{\mathbb{P}(\mathcal{V}_{i}\oplus\mathcal{O}_{\mathcal{M}})}$ induced
from $H_{i}$ (namely, $\frac{i}{2}\partial\bar{\partial}\log(H_{i}\oplus
H_{0})$ where $H_{0}$ is the trivial metric on $\mathcal{O}_{M}$).
###### Proposition 4.10.
There is a fiberwise symplectomorphism
$\psi_{\mathcal{V}_{i}}:(\mathcal{V}_{i},\omega_{\mathbb{P}(\mathcal{V}_{i}\oplus\mathcal{O}_{\mathcal{M}})}|_{\mathcal{V}_{i}})\stackrel{{\scriptstyle\cong}}{{\to}}\left(\\{v\in\mathcal{V}_{i}:H_{i}(v,v)<1\\},\omega_{\mathcal{V}_{i}}|_{\\{H_{i}(v,v)<1\\}}\right).$
###### Proof.
For each $p\in\mathcal{M}$, we have computed the symplectomorphism
$(\mathcal{V}_{i}|_{p},\omega_{\mathbb{P}(\mathcal{V}_{i}\oplus\mathcal{O}_{\mathcal{M}})}|_{\mathcal{V}_{i}|_{p}})\stackrel{{\scriptstyle\cong}}{{\to}}\left(\\{v\in\mathcal{V}_{i}|_{p}:H_{i}(v,v)<1\\},\omega_{\mathcal{V}_{i}}|_{\\{H_{i}(v,v)<1\\}}\right)$
in (4.2), with the metric given by $H_{i}|_{p}$ here. Thus
$\psi_{\mathcal{V}_{i}}(v)=\frac{v}{\sqrt{1+H_{i}(v,v)}}$ (4.3)
gives a fiberwise symplectomorphism whose image is ${\\{H_{i}(v,v)<1\\}}$. ∎
Recall that the universal bundle $\mathcal{V}_{i}\to\mathcal{M}$ admits an
action of $U_{\vec{n}}$ coming from framing (Definition 3.11). One advantage
of $\psi_{\mathcal{V}_{i}}$ is that it is equivariant under this action.
###### Lemma 4.11.
For $g\in U_{\vec{n}}$,
$\psi_{\mathcal{V}_{i}}\circ g=g\circ\psi_{\mathcal{V}_{i}}$
if we used the metric $H_{i}$ given in Theorem 3.7.
###### Proof.
This follows from (4.3) and $H_{i}(g\cdot v,g\cdot v)=H_{i}(v,v)$ by Lemma
3.13. ∎
###### Remark 4.12.
Equation (4.3) has an alternative derivation using the framing. Namely, we
have the surjective morphism $\rho:\underline{\hat{W}_{i}}\to\mathcal{V}_{i}$
(see Equation (3.4)), whose dual give a fiberwise-linear embedding
$\rho^{*}:\mathcal{V}_{i}\stackrel{{\scriptstyle
H_{i}}}{{\cong}}\mathcal{V}_{i}^{*}\to\underline{\hat{W}_{i}}^{*}$. (The
underline means the trivial bundle over $\mathcal{M}$ associated with the
vector space.) Then $\mathbb{P}(\underline{\hat{W}_{i}}^{*}\oplus\mathcal{O})$
(with the standard metric) induces a fiberwise Kähler form on
$\mathcal{V}_{i}^{*}$, and we have a fiberwise symplectic embedding
$(\mathcal{V}_{i},\omega_{\mathbb{P}(\underline{\hat{W}_{i}}^{*}\oplus\mathcal{O})})\hookrightarrow(\mathcal{V}_{i},\omega_{\mathcal{V}_{i}})$.
This gives
$\frac{\rho\rho^{*}\cdot H_{i}\cdot v}{\sqrt{1+H_{0}(\rho^{*}\cdot H_{i}\cdot
v,\rho^{*}\cdot H_{i}\cdot v)}}=\frac{\rho\rho^{*}\cdot H_{i}\cdot
v}{\sqrt{1+v^{*}\cdot H_{i}^{*}\cdot\rho\rho^{*}\cdot H_{i}\cdot v}}.$
Now if we use the metric $H_{i}=(\rho\rho^{*})^{-1}$ given by Theorem 3.7,
then the above equals to the expression in (4.3).
We also have a more flexible construction using the framing, which globalize
_any given non-linear continuous map_
$\sigma_{\mathbb{C}}:\mathbb{C}^{n_{i}}\to\mathbb{C}^{n_{i}}$. Namely,
$\sigma_{\mathbb{C}}$ can be regarded as a fiberwise non-linear self-map on
the trivial bundle
$\underline{\mathbb{C}^{n_{i}}}\to\underline{\mathbb{C}^{n_{i}}}$ (still
denoted by $\sigma_{\mathbb{C}}$). Then we take the composition
$\mathcal{V}_{i}\stackrel{{\scriptstyle
H_{i}}}{{\cong}}\mathcal{V}_{i}^{*}\stackrel{{\scriptstyle(e^{(i)})^{*}}}{{\to}}\underline{\mathbb{C}^{n_{i}}}\stackrel{{\scriptstyle\sigma_{\mathbb{C}}}}{{\to}}\underline{\mathbb{C}^{n_{i}}}\stackrel{{\scriptstyle
e^{(i)}}}{{\to}}\mathcal{V}_{i}$
and denote it by $\sigma_{\mathcal{V}_{i}}$. See Figure 6.
${\underline{\mathbb{C}^{n}}}$${\underline{\mathbb{C}^{n}}^{*}}$${\mathcal{V}_{i}}$${\mathcal{V}_{i}^{*}}$$\scriptstyle{e^{(i)}}$$\scriptstyle{\sigma_{\mathbb{C}}}$$\scriptstyle{\cong}$$\scriptstyle{\cong}$$\scriptstyle{(e^{(i)})^{*}}$
Figure 6. Using the framing to globalize an activation function
$\sigma_{\mathbb{C}}:\mathbb{C}^{n}\to\mathbb{C}^{n}$ over $\mathcal{M}$.
It is easy to get the following explicit expression in terms of
$\sigma_{\mathbb{C}}$.
###### Lemma 4.13.
The above fiber-bundle map
$\sigma_{\mathcal{V}_{i}}:\mathcal{V}_{i}\to\mathcal{V}_{i}$ equals to
$\sigma_{\mathcal{V}_{i}}(v)=\sum_{k=1}^{n_{i}}(\sigma_{\mathbb{C}})_{k}\left(H_{i}(e^{(i)}_{1},v),\ldots,H_{i}(e^{(i)}_{n_{i}},v)\right)\cdot
e^{(i)}_{k}$ (4.4)
where we write
$\sigma_{\mathbb{C}}=((\sigma_{\mathbb{C}})_{1},\ldots,(\sigma_{\mathbb{C}})_{n_{i}})$
and $e^{(i)}=(e^{(i)}_{1}\ldots e^{(i)}_{n_{i}})$.
For instance, we can take $\sigma$ to be the one coming from the
symplectomorphism in Corollary 4.2. We shall prove the universal approximation
theorem for such $\sigma$ in Section 5.
### 4.3. A machine learning program using the framed quiver moduli
Let $Q$ be a digraph and denote by $\mathbb{C}\cdot Q$ its path algebra over
$\mathbb{C}$. Let $\vec{d}\in\mathbb{Z}_{\geq 0}^{Q_{0}}$ be a dimension
vector. We take the framing dimension vector to be $\vec{n}=\vec{d}+\vec{1}$,
where $\vec{1}_{i}:=1$ for all $i\in Q_{0}$. The one additional framing vector
is used for translation (called a ‘bias’ vector).
We fix a collection of input vertices and a collection of output vertices
$I_{\mathrm{in}},I_{\mathrm{out}}\subset Q_{0}$, and
$\gamma\in I_{\mathrm{out}}\cdot(\mathbb{C}\cdot Q)\cdot
I_{\mathrm{in}}=\bigoplus_{i\in I_{\mathrm{in}},j\in
I_{\mathrm{out}}}j\cdot(\mathbb{C}\cdot Q)\cdot i.$
(The trivial path at a vertex $i$ is again denoted by $i$.)
#### 4.3.1. Machine learning using a flat space.
Let’s first formulate a typical machine learning program in the quiver setup.
The following flat space
$U\cong\prod_{a\in Q_{1}}\mathrm{Hom}(V_{t(a)},V_{h(a)})\times\prod_{i\in
Q_{0}}V_{i}$
is used frequently in the subject.
###### Lemma 4.14.
The open subset
$U=\\{[V,e]\in\mathcal{M}_{\vec{d}+\vec{1},\vec{d}}:(e^{(i)}_{1},\ldots,e^{(i)}_{d_{i}})=I_{d_{i}}\textrm{
for all }i\in Q_{0}\\}\subset\mathcal{M}$
gives a coordinate chart of $\mathcal{M}$. ($I_{d_{i}}$ denotes the identity
matrix of rank $d_{i}$.)
###### Proof.
First, such $(V,e)$ are stable: $\mathrm{Im}(e)$ is the whole $V$. Second, for
distinct $(V,e),(V^{\prime},e^{\prime})$ satisfying the above condition,
$[V,e]\not=[V^{\prime},e^{\prime}]$: since they are stable, their orbits are
closed. Suppose $g\cdot(V,e)=(V^{\prime},e^{\prime})$ for some
$g\in\mathrm{GL}_{\vec{d}}$. Then $g_{i}\cdot e^{(i)}=(e^{\prime})^{(i)}$. But
since
$(e^{(i)}_{1},\ldots,e^{(i)}_{d_{i}})=I_{d_{i}}=((e^{\prime})^{(i)}_{1},\ldots,(e^{\prime})^{(i)}_{d_{i}})$,
this forces $g=\mathrm{Id}$ and so $(V,e)=(V^{\prime},e^{\prime})$,
contradicting that they are distinct.
Then we have the chart map $U\stackrel{{\scriptstyle\cong}}{{\to}}\prod_{a\in
Q_{1}}\mathrm{Hom}(V_{t(a)},V_{h(a)})\times\prod_{i\in Q_{0}}V_{i}$ defined by
$W_{a}:=(e^{(i)}_{1},\ldots,e^{(i)}_{d_{i}})^{-1}V(a),\,\,b_{i}:=(e^{(i)}_{1},\ldots,e^{(i)}_{d_{i}})^{-1}e^{(i)}_{d_{i}+1}.$
∎
Now we fix a path $\gamma$ from the input vertices $I_{\mathrm{in}}$ to the
output vertices $I_{\mathrm{out}}$. For each element $[V,e]\in U$, by
composing the affine linear maps $V_{a}(\cdot)+e^{(t(a))}_{d_{t(a)}+1}$
attached to arrows $a$ in the path $\gamma$, together with some non-linear
functions $\sigma_{i}:V_{i}\to V_{i}$ that are called ‘activation functions’,
one obtains a non-linear function
$f^{U}_{[V,e]}:\bigoplus_{i\in
I_{\mathrm{in}}}\mathbb{C}^{d_{i}}\stackrel{{\scriptstyle(e^{(i)}_{1},\ldots,e^{(i)}_{d_{i}})}}{{\cong}}\bigoplus_{i\in
I_{\mathrm{in}}}V_{i}\to\bigoplus_{j\in
I_{\mathrm{out}}}V_{j}\stackrel{{\scriptstyle(e^{(j)}_{1},\ldots,e^{(j)}_{d_{i}})}}{{\cong}}\bigoplus_{j\in
I_{\mathrm{out}}}\mathbb{C}^{d_{j}}$
which is used to approximate a non-explicitly given function $f$. A stochastic
gradient flow of the error function on $U$ is employed to find the optimal
point in $U$.
Let’s write down the symmetry in Lemma 4.11 in the chart $U$.
###### Lemma 4.15.
The chart $U$ is invariant under the right action by
$\prod_{i\in Q_{0}}U(d_{i})\subset\prod_{i\in Q_{0}}U(d_{i}+1)=U_{\vec{n}}$
where $U(d_{i})\subset U(d_{i}+1)$ is embedded as
$\left(\begin{array}[]{cc}U(d_{i})&0\\\ 0&1\end{array}\right)$.
Consider the trivialization $\mathcal{V}_{i}|_{U}\cong\left(\prod_{a\in
Q_{1}}\mathrm{Hom}(V_{t(a)},V_{h(a)})\times\prod_{j\in
Q_{0}}V_{j}\right)\times V_{i}$. The right action of $g\in U(d_{i})$ on
$\mathcal{V}_{i}|_{U}$ is given by
$(W_{a},b_{j},v)_{\begin{subarray}{c}a\in Q_{1}\\\ j\in Q_{0},b_{j}\in
V_{j}\\\ v\in V_{i}\end{subarray}}\cdot g=(g^{-1}\cdot W_{a},g^{-1}\cdot
b_{j},g^{-1}v)_{\begin{subarray}{c}a\in Q_{1}\\\ j\in Q_{0},b_{j}\in V_{j}\\\
v\in V_{i}\end{subarray}}$
where $g^{-1}\cdot W_{a}$ equals to $g^{-1}W_{a}$ if $h(a)=i$, $W_{a}g$ if
$t(a)=i$, and $W_{a}$ otherwise; $g^{-1}\cdot b_{j}$ equals to $g^{-1}b_{j}$
if $j=i$, and $b_{j}$ if $j\not=i$.
###### Proof.
$U$ consists of points $[V,e]$ where $(e^{(j)}_{1},\ldots,e^{(j)}_{d_{i}})$
are invertible for all $j\in Q_{0}$. This property is invariant under the
action of $\prod_{i\in Q_{0}}U(d_{i})$. Hence $U$ is an invariant subset.
$(W_{a},b_{j},v)_{\begin{subarray}{c}a\in Q_{1}\\\ j\in Q_{0},b_{j}\in
V_{j}\\\ v\in V_{i}\end{subarray}}$ corresponds to the point
$[W,e,v]\in\mathcal{V}_{i}=(R^{s}_{n,d}\times V_{i})/\mathrm{GL}_{\vec{d}}$.
where $e^{(j)}=(I_{d_{j}}\,\,b_{j})$.
$\displaystyle[W,e,v]\cdot g=[W,e\cdot g,v]=[W,(g\,\,b_{j})_{j\in Q_{1}},v]$
$\displaystyle=$ $\displaystyle[g^{-1}\cdot W,(I_{d_{j}}\,\,g^{-1}\cdot
b_{j})_{j\in Q_{1}},g^{-1}\cdot v]$
where the left action by $g^{-1}$ (as an element in $\mathrm{GL}_{\vec{d}}$)
is as specified by definition. ∎
Now we prove an important symmetric property that the activation function
(4.2) enjoys, which can be used to reduce the dimensions.
###### Proposition 4.16.
Suppose the activation functions $\sigma_{i}:V_{i}\to V_{i}$ are taken to be
the one given in Equation (4.2). Then $f^{U}$ is invariant under
$\prod_{j\not\in I_{\mathrm{in}}\cup I_{\mathrm{out}}}U(d_{j})\subset
U_{\vec{n}}$. (See the embedding in Lemma 4.15.)
###### Proof.
The terms of $f^{U}$ are of the form
$\sigma_{h(a_{k})}\left(W_{a_{k}}\ldots\left(\sigma_{h(a_{1})}\left(W_{a_{1}}(v)+b_{h(a_{1})}\right)\ldots\right)+b_{h(a_{k})}\right)$.
By the above lemma, for $g\in U(d_{h(a_{i})})$ where $h(a_{i})\not\in
I_{\mathrm{in}}\cup I_{\mathrm{out}}$, the action of $g$ results in
$\sigma_{h(a_{i})}\mapsto g\cdot\sigma_{h(a_{i})}g^{-1}$ in the above
expression and does not affect any other part. By Lemma 4.8, $\sigma_{i}$ is
$U(d_{i})$-equivariant, and hence $f^{U}$ remains invariant. ∎
By the above proposition, we can descend $f^{U}$ to the orbit space
$U/\prod_{j\not\in I_{\mathrm{in}}\cup I_{\mathrm{out}}}U(d_{j})$ to reduce
the dimensions. However, the quotient space will be highly singular. Instead,
we can realize the dimension reduction by restricting $f^{U}$ to a submanifold
$U^{\prime}$ of $U$ whose orbit occupies the whole $U$, and do stochastic
gradient flow on $U^{\prime}$ instead of $U$. The following gives one simple
possibility.
###### Proposition 4.17.
Consider a subset $\\{a_{1},\ldots,a_{p}\\}$ of arrows whose heads and tails
do not belong to $I_{\mathrm{in}}\cup I_{\mathrm{out}}$, and for any two
distinct arrows $a_{1},a_{2}$ in the subset, $t(a_{1})\not=h(a_{2})$. Then the
vector subspace
$U^{\prime}:=\\{(W_{a},b_{j})_{\begin{subarray}{c}a\in Q_{1}\\\ j\in
Q_{0},b_{j}\in V_{j}\end{subarray}}\in U:W_{a_{i}}\textrm{ is of the form
}\left(D\,\,W_{a_{i}}^{\prime}\right)\,\,\forall i=1,\ldots,p\\}$
has its orbit being the whole $U$, that is, $\left(\prod_{j\not\in
I_{\mathrm{in}}\cup I_{\mathrm{out}}}U(d_{j})\right)\cdot U^{\prime}=U$. In
above, $D$ is a diagonal matrix (of the maximum possible size) and
$W_{a_{i}}^{\prime}$ is any matrix occupying the rest.
###### Proof.
This follows from the singular-value decomposition of a matrix $W$ as
$A\cdot\left(D\,\,W^{\prime}\right)\cdot B$ where $A$ and $B$ are unitary
matrices of appropriate sizes. ∎
#### 4.3.2. Machine learning using the quiver moduli.
The quiver moduli $\mathcal{M}=\mathcal{M}_{\vec{n},\vec{d}}$ gives a
compactification of $U$. Compactness is important for the formulation of Morse
theory and convergence of a gradient flow. We would like to use the whole
$\mathcal{M}$ in application of machine learning. Non-trivial metrics over the
moduli will play a crucial role.
First, consider the situation before adding in activation functions. Each
arrow $a\in Q_{1}$ is associated with a vector-bundle morphism
$a_{\mathcal{M}}:\mathcal{V}_{t(a)}\to\mathcal{V}_{h(a)}$. For each path
$a_{k}\ldots a_{1}$, we take the map
$(a_{k})_{\mathcal{M}}\left(\ldots\left((a_{2})_{\mathcal{M}}\left((a_{1})_{\mathcal{M}}(v)+e^{(h(a_{1}))}_{d_{h(a_{1})}+1}\right)+e^{(h(a_{2}))}_{d_{h(a_{2})}+1}\right)\ldots\right)+e^{(h(a_{k}))}_{d_{h(a_{k})}+1}$
(4.5)
which is fiberwise affine linear. Thus a path $\gamma$ gives an affine bundle
morphism
$\gamma_{\mathcal{M}}:\mathcal{V}_{I_{\mathrm{in}}}:=\bigoplus_{i\in
I_{\mathrm{in}}}\mathcal{V}_{i}\to\mathcal{V}_{I_{\mathrm{out}}}:=\bigoplus_{j\in
I_{\mathrm{out}}}\mathcal{V}_{j}.$
Then we have
$L_{\gamma}\left(\left(s^{(i)}_{k}\right)_{\begin{subarray}{c}i\in
I_{\mathrm{in}}\\\
k\in\\{1,\ldots,d_{i}\\}\end{subarray}}\right)=\left(H_{j}\left(e^{(j)}_{p},\sum_{i\in
I_{\mathrm{in}}}\gamma_{\mathcal{M}}\cdot\sum_{k=1}^{d_{i}}s^{(i)}_{k}e^{(i)}_{k}\right)\right)_{\begin{subarray}{c}j\in
I_{\mathrm{out}},\\\ p\in\\{1,\ldots,d_{j}\\}\end{subarray}}:\bigoplus_{i\in
I_{\mathrm{in}}}\underline{\mathbb{C}^{d_{i}}}\to\bigoplus_{j\in
I_{\mathrm{out}}}\underline{\mathbb{C}^{d_{j}}}.$ (4.6)
Suppose a continuous function $f=K\to\bigoplus_{j\in
I_{\mathrm{out}}}\mathbb{C}^{d_{j}}$ is given, where $K$ is a compact subset
of $\bigoplus_{i\in I_{\mathrm{in}}}\mathbb{C}^{d_{i}}$. Then the fiberwise
integral
$\mathcal{E}:=\int_{K}\left\|f-L_{\gamma}\right\|^{2}_{\mathcal{V}_{I_{\mathrm{out}}}}d\mu_{K}$
(4.7)
gives a smooth function on $\mathcal{M}$.
###### Remark 4.18.
Alternatively, we can define the fiber-bundle morphism
$f_{\mathcal{M}}:\sum_{j\in
I_{\mathrm{out}}}\sum_{l=1}^{d_{j}}f^{(j)}_{l}\cdot
e^{(j)}_{l}:\underline{K}\to\mathcal{V}_{I_{\mathrm{out}}}$
and take
$\int_{K}\left\|f_{\mathcal{M}}\left(\left(s^{(i)}_{k}\right)_{\begin{subarray}{c}i\in
I_{\mathrm{in}}\\\ k\in\\{1,\ldots,d_{i}\\}\end{subarray}}\right)-\sum_{i\in
I_{\mathrm{in}}}\gamma_{\mathcal{M}}\cdot\sum_{k=1}^{d_{i}}s^{(i)}_{k}e^{(i)}_{k}\right\|^{2}_{\mathcal{V}_{I_{\mathrm{out}}}}d\mu_{K}.$
However, with such a definition, we need to worry that
$(e^{(j)}_{1},\ldots,e^{(j)}_{d_{j}})$ for some output $j$ degenerates (as a
frame), in which case approximating $f$ and approximating $f_{\mathcal{M}}$
are different.
We have a gradient flow $r:\mathbb{R}\to\mathcal{M}$ which can be used to
minimize $\mathcal{E}$:
$\frac{dr}{dt}=-(\nabla\mathcal{E})(r(t))$
where $\nabla\mathcal{E}=(d\mathcal{E})^{\\#_{g}}$ where
$(\cdot)^{\\#}_{g}:T^{*}\mathcal{M}\stackrel{{\scriptstyle
g}}{{\cong}}T\mathcal{M}$ is the identification by a metric $g$ on
$\mathcal{M}$. ($(\nabla\mathcal{E})^{p}=g^{pq}\partial_{q}\mathcal{E}$ in
local coordinates.) $g$ can be taken to be the induced metric from the trivial
metric on the vector space $R_{\vec{n},\vec{d}}$ via symplectic reduction.
Alternatively, $g$ can be taken to be the metric given by the Ricci curvature
in Theorem 3.15 (when $Q$ has no oriented cycle), which has a better
expression in homogeneous coordinates.
Note that $(L_{\gamma})|_{[V,e]}$ is affine linear on $\bigoplus_{i\in
I_{\mathrm{in}}}\underline{\mathbb{C}^{d_{i}}}$ for every
$[V,e]\in\mathcal{M}$, which is not good enough for the purpose of
approximating $f$. We introduce fiberwise non-linearity below.
###### Definition 4.19.
Let $A$ be a finite set whose every element is associated with two vertices
(head $h$ and tail $t$) in $Q_{0}$. Elements in $A$ are called activation
arrows. (These are not arrows in $Q_{1}$.) The semiring generated by $Q_{1}$
and $A$, denoted by $\Gamma(Q,A)$, has the underlying vector space spanned by
the independent set $\coprod_{p=0}^{\infty}S_{p}$ where $S_{p}$ is defined
inductively as follows.
1. (1)
$S_{0}$ consists of all paths of $Q$.
2. (2)
Suppose $S_{p}$ has been defined, and each element in $S_{p}$ has a head $h$
and a tail $t$. $S_{p+1}$ consists of $\gamma\cdot\alpha\cdot\tilde{\gamma}$,
where $\alpha\in A$, $\gamma$ is any path of $Q$ with $h(\alpha)=t(\gamma)$,
and $\tilde{\gamma}\in t(\alpha)\cdot(\mathbb{C}\cdot S_{p})$.
($\mathbb{C}\cdot S_{p}$ denotes the vector space generated by $S_{p}$;
$t(\alpha)\cdot(\mathbb{C}\cdot S_{p})$ is the subspace generated by elements
of $S_{p}$ with head being $t(\alpha)$.) The above element has the head
$h(\gamma)$ and the tail $t(\tilde{\gamma})$.
The above vector space $\Gamma(Q,A)$ has an obvious product by concatenation.
($s_{1}\cdot s_{2}=0$ if $h(s_{2})\not=t(s_{1})$.) Note that for $\alpha\in
A$, $(s_{1}+cs_{2})\cdot\alpha=s_{1}\cdot\alpha+cs_{2}\cdot\alpha$, but
$\alpha\cdot(s_{1}+cs_{2})\not=\alpha\cdot s_{1}+c\alpha\cdot s_{2}$.
Now, suppose each element $\alpha\in A$ is associated with a fiber-bundle
morphism
$\alpha_{\mathcal{M}}:\mathcal{V}_{t(\alpha)}\to\mathcal{V}_{h(\alpha)}.$
Then by composing the corresponding affine linear morphisms (as in Equation
(4.5)) and $\alpha_{\mathcal{M}}$, an element $\tilde{\gamma}\in\Gamma(Q,A)$
induces a fiber-bundle morphism
$\tilde{\gamma}_{\mathcal{M}}:\mathcal{V}_{t(\tilde{\gamma})}\to\mathcal{V}_{h(\tilde{\gamma})}.$
In particular, if we fix $\tilde{\gamma}\in
I_{\mathrm{out}}\cdot\Gamma(Q,A)\cdot I_{\mathrm{in}}$, then we define
$f_{\tilde{\gamma}}:\bigoplus_{i\in
I_{\mathrm{in}}}\underline{\mathbb{C}^{d_{i}}}\to\bigoplus_{j\in
I_{\mathrm{out}}}\underline{\mathbb{C}^{d_{j}}}$
like in the definition of $L_{\gamma}$ in (4.6) by replacing
$\gamma_{\mathcal{M}}$ by $\tilde{\gamma}_{\mathcal{M}}$. Then a stochastic
gradient flow for the corresponding error function (by replacing $L_{\gamma}$
by $f_{\tilde{\gamma}}$ in (4.7)) can be carried out.
For now, we set $A$ to be $Q_{0}$ as a set, and each element
$\mathfrak{o}_{i}$ has the head and tail being $i\in Q_{0}$. We can associate
$\mathfrak{o}_{i}$ with the fiber-bundle morphism $\psi_{\mathcal{V}_{i}}$
given in (4.3), or $\sigma_{\mathcal{V}_{i}}$ in (4.4). Then we obtain
$f_{\tilde{\gamma}}$ above which is non-linear along fibers for the purpose of
machine learning.
If we associate $\mathfrak{o}_{i}$ with $\psi_{\mathcal{V}_{i}}$ given in
(4.3), then the symmetry of $\prod_{j\not\in I_{\mathrm{in}}\cup
I_{\mathrm{out}}}U(d_{j})\subset U_{\vec{n}}$ is respected, by Lemma 4.11. The
proof is similar to that for Proposition 4.16 and is omitted here.
###### Proposition 4.20.
Suppose $\mathfrak{o}_{i}$ is assigned as $\psi_{\mathcal{V}_{i}}$ given in
(4.3), and $H_{i}$ is taken to be the metric in Theorem 3.7. Then
$f_{\tilde{\gamma}}$ is invariant under $\prod_{j\not\in I_{\mathrm{in}}\cup
I_{\mathrm{out}}}U(d_{j})\subset U_{\vec{n}}$. (See the embedding in Lemma
4.15.)
###### Remark 4.21.
Since we are taking affine linear morphisms
$(a_{i})_{\mathcal{M}}(v)+e^{(h(a_{i}))}_{d_{h(a_{i})}+1}$ for the arrows
which involves the term $e^{(h(a_{i}))}_{d_{h(a_{i})}+1}$, only the symmetry
$U(d_{j})$ rather than $U(d_{j}+1)=U(n_{j})$ is respected. If the bias vector
$e^{(h(a_{i}))}_{d_{h(a_{i})}+1}$ is not used in the program, then
$f_{\tilde{\gamma}}$ will be invariant under the bigger group $\prod_{j\not\in
I_{\mathrm{in}}\cup I_{\mathrm{out}}}U(n_{j})$.
#### 4.3.3. A simple example
Recall the quiver in Example 2.7. Let’s take
$n_{1}=d_{1},n_{2}=d_{2}+1,n_{3}=d_{3}$ instead of $n_{i}=d_{i}+1\,\forall i$,
since we do not need to use bias vectors at the input and output vertices. The
path $\gamma$ is simply $a_{2}a_{1}$, and
$\tilde{\gamma}=a_{2}\cdot\mathfrak{o}_{1}\cdot a_{1}$.
We have the universal bundles
$\mathcal{V}_{1}\cong\underline{\mathbb{C}}^{d_{1}}$, $\mathcal{V}_{2}$ and
$\mathcal{V}_{3}$. We have $\rho^{(1)}=e^{(1)}$,
$\rho^{(2)}=(e^{(2)}\,\,a_{1}e^{(1)})$,
$\rho^{(3)}=(e^{(3)}\,\,a_{2}e^{(2)}\,\,a_{2}a_{1}e^{(1)})$. In terms of
homogeneous coordinates (namely the coordinates
$(e^{(1)},e^{(2)},e^{(3)},a_{1},a_{2})$ on the vector space
$R_{\vec{n},\vec{d}}$, where each entry is a matrix of suitable size), the
metrics given in Theorem 3.7 are
$\displaystyle H_{1}$
$\displaystyle=\left(e^{(1)}(e^{(1)})^{*}\right)^{-1},\,H_{2}=\left(e^{(2)}(e^{(2)})^{*}+a_{1}e^{(1)}(e^{(1)})^{*}a_{1}^{*}\right)^{-1},\,$
$\displaystyle H_{3}$
$\displaystyle=\left(e^{(3)}(e^{(3)})^{*}+a_{2}e^{(2)}(e^{(2)})^{*}a_{2}^{*}+a_{2}a_{1}e^{(1)}(e^{(1)})^{*}a_{1}^{*}a_{2}^{*}\right)^{-1}.$
on $\mathcal{V}_{i},i=1,2,3$ respectively. The activation functions we
constructed in the previous subsection are
$\psi_{i}(v)=\frac{v}{\sqrt{1+v^{*}H_{i}v}}\textrm{ and
}\sigma_{i}(v)=\sum_{k=1}^{d_{i}+1}(\sigma_{\mathbb{C}})_{k}\left((e^{(i)}_{1})^{*}H_{i}v,\ldots,(e^{(i)}_{d_{i}+1})^{*}H_{i}v\right)\cdot
e^{(i)}_{k},$
where we can take
$(\sigma_{\mathbb{C}})_{k}(\vec{z})=\frac{e^{2\mathrm{Re}(z_{k})}}{1+\sum_{j=1}^{d_{i}+1}e^{2\mathrm{Re}(z_{j})}}+\mathbf{i}\,\mathrm{Im}(z_{k})$
for instance. Both have the $\mathrm{GL}$-equivariance property
$\psi_{i}(g\cdot v)=g\cdot\psi_{i}(v)$ and $\sigma_{i}(g\cdot
v)=g\cdot\sigma_{i}(v)$.
The function (over $\mathcal{M}$) cooked up from this quiver is
$f_{\tilde{\gamma}}:\mathbb{C}^{d_{1}}\to\mathbb{C}^{d_{3}}$,
$f_{\tilde{\gamma}}(s_{1},\ldots,s_{d_{1}})=\left(H_{3}\left(e^{(3)}_{p},a_{2}\cdot\sigma_{2}\left(a_{1}\cdot\sum_{k=1}^{d_{1}}s_{k}e^{(1)}_{k}+e^{(2)}_{d_{2}+1}\right)\right)\right)_{p=1}^{d_{3}}$
if we use $\sigma_{2}$ as the activation function, or the same expression with
$\sigma_{2}$ replaced by $\psi_{2}$. Then we run a stochastic gradient flow on
$\mathcal{M}$ (or on the vector space $R_{\vec{n},\vec{d}}$ upstairs) to
minimize the distance of $f_{\tilde{\gamma}}$ and a function $f$ coming from
reality.
To run the gradient flow, we need to take a metric on $\mathcal{M}$. Recall
that we have the metric on the tangent bundle of $\mathcal{M}$ coming from the
Ricci curvatures of $H_{i}$:
$H_{T}=\sum_{i=1}^{3}\mathrm{tr~{}}\left(\left(\partial_{v}\rho^{(i)}\right)^{\mathrm{*}}\cdot
H_{i}\cdot\partial_{w}\rho^{(i)}\right)-\sum_{i=1}^{3}\mathrm{tr~{}}\left(\left(\partial_{v}\rho^{(i)}\cdot(\rho^{(i)})^{\mathrm{*}}\cdot
H_{i}^{\frac{1}{2}}\right)^{\mathrm{*}}\cdot
H_{i}\cdot\left(\partial_{w}\rho^{(i)}\cdot(\rho^{(i)})^{\mathrm{*}}\cdot
H_{i}^{\frac{1}{2}}\right)\right).$ (4.8)
Consider the open subset of the vector space $R_{\vec{n},\vec{d}}$ in which
$(e_{1}^{(i)},\ldots,e_{d_{i}}^{(i)})$ is invertible. (This is the preimage of
the chart $U\subset\mathcal{M}$.) A tangent vector $v$ of $\mathcal{M}$ is
lifted as $(\delta a_{1},\delta a_{2},\delta e_{d_{2}+1}^{(2)})$ (and all
other components are set to be zero). Since $\rho^{(1)}=e^{(1)}$,
$\rho^{(2)}=(e^{(2)},a_{1}e^{(1)})$,
$\rho^{(3)}=(e^{(3)},a_{2}e^{(2)},a_{2}a_{1}e^{(1)})$, we have
$\partial_{v}\rho^{(1)}=0$, $\partial_{v}\rho^{(2)}=\left((0\,\,\delta
e_{d_{2}+1}^{(2)}),\,\,(\delta a_{1})e^{(1)}\right)$, and
$\partial_{v}\rho^{(3)}=\left(0,\,\,(\delta a_{2})e^{(2)}+(0\,\,a_{2}\delta
e_{d_{2}+1}^{(2)}),\,\,\,(\delta a_{2})a_{1}e^{(1)}+a_{2}(\delta
a_{1})e^{(1)}\right).$
Then the above metric $H_{T}$ can be computed explicitly in terms of the
homogeneous coordinates.
###### Remark 4.22.
In above, if we use trivial metrics over the vector space
$R_{\vec{n},\vec{d}}$ instead of $H_{i}$ and $H_{T}$, the expressions will get
simpler; however they will only be $U_{\vec{d}}$-equivariant rather than
$\mathrm{GL}_{\vec{d}}$-equivariant. Then we need to restrict to the moment-
map level $\mu^{-1}(I)\subset R_{\vec{n},\vec{d}}$ and its tangent bundle, in
order to stay in the same moduli $\mathcal{M}$ downstairs. This would increase
the computational complexity.
We can also write in inhomogeneous coordinates $(W_{1},W_{2},b)$ in the chart
$U\subset\mathcal{M}$, where $(e_{1}^{(i)},\ldots,e_{d_{i}}^{(i)})=I_{d_{i}}$
for all $i=1,2,3$. $W_{1},W_{2}$ are the matrices of the arrows $a_{1},a_{2}$
and $e_{d_{2}+1}^{(2)}=b$ is the bias vector. Then $H_{1}=I$,
$H_{2}=(I+bb^{*}+W_{1}W_{1}^{*})^{-1}$, and
$H_{3}=(I+W_{2}W_{2}^{*}+W_{2}bb^{*}W_{2}^{*}+W_{2}W_{1}W_{1}^{*}W_{2}^{*})^{-1}$.
Then we can run the gradient flow in $U\subset\mathcal{M}$ (which has lower
dimensions than $R_{\vec{n},\vec{d}}$).
Note that when $W_{1},W_{2},b$ are close to zero, $H_{i}$ are close to the
identity matrix, and $\sigma_{i}$ is close to $\sigma_{\mathbb{C}}$. Moreover,
the second term of (4.8) is close to zero, and the first term is close to the
standard metric. Thus when $W_{1},W_{2},b$ are small, the function
$f_{\tilde{\gamma}}$ is close to the commonly used one
$f^{U}_{W_{1},W_{2},b}(v)=W_{2}\cdot(\sigma_{\mathbb{C}}(W_{1}\cdot v+b)),$
(4.9)
and the gradient flow is close to the usual one on the flat space $U$ (see
Section 4.3.1). The additional terms can be understood as modifications to
ensure the flow converges in $\mathcal{M}$.
### 4.4. A discussion on Morse inequalities
By the work of Reineke [Rei08], the framed quiver moduli $\mathcal{M}$ is a
tower of Grassmannians (Theorem 2.4), and hence its Poincaré polynomial is a
product of that of Grassmannians (Corollary 2.5). Such topological invariants
give important information about a gradient flow on $\mathcal{M}$.
In particular the Morse inequalities for a Morse function $\mathcal{E}$ on a
compact manifold $\mathcal{M}$ state as follows. Let $c^{j}(\mathcal{E})$ be
the number of critical points of index $j$ for $\mathcal{E}$. Then for every
$j$,
$c^{j}(\mathcal{E})\geq h^{j}(\mathcal{M})$
where $h^{j}(\mathcal{M})$ denotes the cohomological numbers (which are
coefficients of the Poincaré polynomial).
Given a gradient flow, which is a path $\gamma:\mathbb{R}\to\mathcal{M}$
satisfying the gradient flow equation, $\lim_{t\to\pm\infty}\gamma(t)$ are
critical points. Moreover, critical points carry important effect to the rate
of the gradient flow. Namely, when the flow $\gamma$ gets close to a critical
point with index being $1,\ldots,\dim\mathcal{M}-1$,
$\|\gamma^{\prime}(t)\|=\|\mathrm{grad}\,\mathcal{E}(\gamma(t))\|$ becomes
small. In other words the flow slows down when it passes through a
neighborhood of a critical point. Such a slowing-down effect of saddle points
was studied in machine learning in [PDGB14, DPG+14].
The cohomological numbers $h^{j}(\mathcal{M})$ give the minimum number of
critical points and hence are important invariants of a neural network (which
simply means a directed graph $Q$ together with a dimension vector $\vec{d}$
here). Over $\mathbb{C}$, $h^{j}(\mathcal{M})=0$ when $j$ is odd. Thus the
Euler characteristic $\chi_{Q,\vec{d}}$ equals to
$\sum_{j}h^{j}(\mathcal{M})$, which is the minimal total number of critical
points. It is computed by simply setting $q=1$ in Corollary 2.5. Another
important invariant is $\dim\mathcal{M}$ (that is, the number of training
parameters of the network), which is simply
$\mathcal{D}_{Q,\vec{d}}=\sum_{i\in
Q_{0}}d_{i}\left(n_{i}+\sum_{\begin{subarray}{c}j\to i\\\
j\not=i\end{subarray}}d_{j}\right)$
using the notation of Corollary 2.5. (We take $n_{i}=d_{i}+1$ in this
section.) These are illustrated in the two practical examples below.
###### Example 4.23.
Consider $Q$ being the $A_{k+2}$ quiver, which has $k+2$ vertices labeled by
$0,\ldots,k+1$, and there is exactly one arrow from $i$ to $i+1$ for
$i=0,\ldots,k$, and no arrow otherwise. Set $d_{-1}:=0$. Then the minimal
total number of critical points is
$\chi_{Q,\vec{d}}=\prod_{i=-1}^{k}\binom{d_{i}+d_{i+1}+1}{d_{i+1}}$
and
$\mathcal{D}_{Q,\vec{d}}=\sum_{i=-1}^{k}d_{i+1}(d_{i}+1).$
###### Example 4.24.
Now consider the following quiver $A_{k+2}^{\prime}$, which has vertices
labeled by $0,\ldots,k+1$, and there is one arrow from vertex $i$ to vertex
$j$ for every $i<j$. Set $d_{-1}:=0$. Then
$\chi_{Q,\vec{d}}=\prod_{i=-1}^{k}\binom{\sum_{j=0}^{i+1}d_{j}+1}{d_{i+1}}$
and
$\mathcal{D}_{Q,\vec{d}}=\sum_{i=-1}^{k}d_{i+1}\left(\sum_{j=0}^{i}d_{j}+1\right).$
Figure 7 shows the graph of $\log\chi_{Q,\vec{d}}$ versus
$\mathcal{D}_{Q,\vec{d}}$ for the two examples, where we set $k=3$,
$d_{1}=600,d_{5}=10$, and $d_{2}=d_{3}=d_{4}$.
Note that the quiver denoted by $A_{k+2}^{\prime}$ in Example 4.24 is a simple
analog of the network known as ResNet, which adds arrows to the
$A_{k+2}$-quiver that skip the middle vertices to get around with the
‘gradient-vanishing problem’. Namely, in the $A_{k+2}$ case, the derivatives
of $\mathcal{E}$ with respect to matrix entries for arrows in the early stage
are typically very small by chain rule, which is not good for the flow rate.
Arrows that skip the middle vertices are added, so that there are short paths
which involve the early arrows.
From Figure 7, we see that in the same dimensions, the minimal number of
critical points in $\mathcal{M}$ is smaller for $A_{5}^{\prime}$ than that for
$A_{5}$. (We have numerically verified this for general $k$.) This gives a
supporting evidence that $\chi_{Q,\vec{d}}$ is an important invariant in
applications to machine learning.
Figure 7. A plot of $\log\chi_{Q,\vec{d}}$ (y-axis) versus
$\mathcal{D}_{Q,\vec{d}}$ (x-axis) for $A_{5}$ and $A_{5}^{\prime}$.
### 4.5. A remark on Abelianization
In many basic neural networks, each vertex of $Q$ is associated with a vector
space of only dimension one. When $\vec{d}=\vec{1}$, that is, all entries of
the dimension vector equal to one, $\mathcal{M}_{\vec{n},\vec{1}}$ is a
quotient by the Abelian group $(\mathbb{C}^{\times})^{\Sigma\vec{d}}$, and
hence a toric variety. Indeed, by Theorem 2.4, $\mathcal{M}_{\vec{n},\vec{1}}$
is a tower of projective spaces $\mathbb{P}^{k}$ for a sequence of $k$.
Given $Q$ and $\vec{d}$, we can always construct a bigger quiver
$Q^{\textrm{Ab},\vec{d}}$as follows. For each vertex $i\in Q_{0}$, we make
$d_{i}$ copies indexed by $(i,p)$ for $p=1,\ldots,d_{i}$. For each arrow of
$Q$ from $i$ to $j$, we make a corresponding arrow for
$Q^{\textrm{Ab},\vec{d}}$ from $(i,p)$ to $(j,q)$ for every $p=1,\ldots,d_{i}$
and $q=1,\ldots,d_{j}$. See Figure 8 for an example.
Figure 8. An example of Abelianization. The LHS shows a quiver $Q$ together
with the dimension vector $\vec{d}$. The RHS shows $Q^{\mathrm{Ab},\vec{d}}$.
Given a dimension vector $\vec{n}\in Q_{0}$, define $\vec{n}^{\textrm{Ab}}\in
Q^{\textrm{Ab},\vec{d}}_{0}$ by $\vec{n}^{\textrm{Ab}}_{(i,p)}=\vec{n}_{i}$
for all $p$. The relation between $\mathcal{M}^{Q}_{\vec{n},\vec{d}}$ and the
toric variety
$\mathcal{M}^{Q^{\textrm{Ab},\vec{d}}}_{\vec{n}^{\textrm{Ab}},\vec{1}}$ is
known as Abelianization and is well-studied in [Mar00]. The basic example is
$\operatorname{Gr}(n,d)$ (which is the framed moduli for the quiver with a
single vertex), whose Abelianization is $(\mathbb{P}^{n})^{d}$ (the
disconnected quiver with $d$ vertices and no arrow).
Namely, the moduli spaces $\mathcal{M}^{Q}_{\vec{n},\vec{d}}$ and
$\mathcal{M}^{Q^{\textrm{Ab},\vec{d}}}_{\vec{n}^{\textrm{Ab}},\vec{1}}$ are
GIT quotients of the same vector space
$R^{Q}_{\vec{n},\vec{d}}=R^{Q^{\textrm{Ab},\vec{d}}}_{\vec{n}^{\textrm{Ab}},\vec{1}}$
by $\mathrm{GL}_{\vec{d}}$ and $(\mathbb{C}^{\times})^{\Sigma\vec{d}}$
respectively. More precisely, we have the fiber bundle
$\mu_{U_{\vec{d}}}^{-1}(\\{I\\})/T\to\mathcal{M}^{Q}_{\vec{n},\vec{d}}$ with
fibers being a product of complete flags $U_{\vec{d}}/T^{\Sigma\vec{d}}$, and
the inclusion
$\mu_{U_{\vec{d}}}^{-1}(\\{I\\})/T\subset\mathcal{M}^{Q^{\textrm{Ab},\vec{d}}}_{\vec{n}^{\textrm{Ab}},\vec{1}}$.
The universal bundles $\mathcal{V}_{i}$ over
$\mathcal{M}^{Q}_{\vec{n},\vec{d}}$ is descended from the direct sum of
universal line bundles $\bigoplus_{p=1}^{d_{i}}\mathcal{V}_{(i,p)}$ over
$\mathcal{M}^{Q^{\textrm{Ab},\vec{d}}}_{\vec{n}^{\textrm{Ab}},\vec{1}}$
(restricted to the above subset). The cohomology of
$\mathcal{M}^{Q}_{\vec{n},\vec{d}}$ is generated by the Chern classes
$c_{k}(\mathcal{V}_{i})$, which can be written as the $k$-th elementary
symmetric polynomials in $c_{1}(\mathcal{V}_{(i,p)})$ for $p=1,\ldots,d_{i}$.
On the other side, the cohomology of
$\mathcal{M}^{Q^{\textrm{Ab},\vec{d}}}_{\vec{n}^{\textrm{Ab}},\vec{1}}$ is
generated by $c_{1}(\mathcal{V}_{(i,p)})$.
Note that the functions $f^{Q}_{\tilde{\gamma}}$ and $\mathcal{E}^{Q}$ over
$\mathcal{M}^{Q}_{\vec{n},\vec{d}}$ constructed in Section 4.3.2 _cannot_ be
lifted to
$\mathcal{M}^{Q^{\textrm{Ab},\vec{d}}}_{\vec{n}^{\textrm{Ab}},\vec{1}}$. The
reason is that, $f^{Q}_{\tilde{\gamma}}$ and $\mathcal{E}^{Q}$ are
$\mathrm{GL}_{\vec{d}}$-equivariant functions on $R^{Q,s}_{n,d}$, the subset
of stable representations, rather than the whole $R^{Q}_{n,d}$. The definition
of $f^{Q}_{\tilde{\gamma}}$ involves the metrics on the universal bundles
$\mathcal{V}_{i}$, which take the expression $(\rho_{i}\rho_{i}^{*})^{-1}$,
and it is only defined over $R^{Q,s}_{n,d}$ where $\rho_{i}$ is surjective.
Rather, we have the functions $f^{Q^{\textrm{Ab},\vec{d}}}_{\tilde{\gamma}}$
and $\mathcal{E}^{Q^{\textrm{Ab},\vec{d}}}$ on
$\mathcal{M}^{Q^{\textrm{Ab},\vec{d}}}_{\vec{n}^{\textrm{Ab}},\vec{1}}$, which
uses the metrics $(\rho_{i,p}\rho_{i,p}^{*})^{-1}$ on the line bundles
$\mathcal{V}_{i,p}$.
Previously we have taken $\vec{n}=\vec{d}+\vec{1}$ for
$\mathcal{M}^{Q}_{n,d}$. After Abelianization, the dimension vectors
$\vec{n}^{\mathrm{Ab}}$ and $\vec{1}$ for
$\mathcal{M}^{Q^{\textrm{Ab},\vec{d}}}_{\vec{n}^{\textrm{Ab}},\vec{1}}$ no
longer satisfy such equality. This is actually not a problem, since the
function $f^{Q^{\textrm{Ab},\vec{d}}}_{\tilde{\gamma}}$ defined on
$\mathcal{M}^{Q^{\textrm{Ab},\vec{d}}}_{\vec{2},\vec{1}}$ can be lifted to
$\mathcal{M}^{Q^{\textrm{Ab},\vec{d}}}_{\vec{n}^{\textrm{Ab}},\vec{1}}$.
Alternatively, we can set the $k$-th framing vectors to be zero for all
$k=2,\ldots,d_{i}+1$ and for all vertices $(i,p)\in
Q^{\textrm{Ab},\vec{d}}_{0}$. This gives a subvariety of
$\mathcal{M}^{Q^{\textrm{Ab},\vec{d}}}_{\vec{n}^{\textrm{Ab}},\vec{1}}$ which
is isomorphic to $\mathcal{M}^{Q^{\textrm{Ab},\vec{d}}}_{\vec{2},\vec{1}}$.
## 5\. Universal Approximation Theorem
In Section 4.1, we have introduced the multi-variable functions $\sigma$
coming from moment maps of toric varieties. For instance,
$\sigma_{k}(x)=\frac{e^{2x_{k}}}{1+\sum_{j=1}^{d}e^{2x_{j}}}$ for
$X=\mathbb{P}^{d}$. In this section, we will give a theoretical basis for
using this as an activation function, by proving the universal approximation
theorem for this function.
The universal approximation theorem ensures that in theory, any given function
on a compact set can be approximated (as close as you want) by the functions
produced from directed graphs (denoted by $f^{U}_{[V,e]}$ in Section 4.3.1).
There are several different versions of this theorem [Cyb89, LLPS93, LJ18,
LPW+17]. To the authors’ knowledge, the past works have focused on proving the
theorem for single-variable activation functions.
In the work of Cybenko in proving the theorem below, rescaling on the domain
of the activation function plays a key role. The rescaling technique will also
be very useful in our situation.
###### Theorem 5.1 ([Cyb89]).
Let $\phi:\mathbb{R}\to\mathbb{R}$ be any continuous function with
$\lim_{x\to\infty}\phi(x)=1$ and $\lim_{x\to-\infty}\phi(x)=0$. Let $K$ be a
compact set in $\mathbb{R}^{d}$. Then the collection of functions
$G:K\to\mathbb{R}$ of the form
$G(x)=\sum_{j=1}^{N}\alpha_{j}\phi(y_{j}^{T}x+\theta_{j})$
where $N\in\mathbb{Z}_{>0}$,$y_{j}\in\mathbb{R}^{d}$, and
$\theta_{j},\alpha_{j}\in\mathbb{R}$, are dense in the space of continuous
functions $C(K)$.
The above function $G$ can be understood as $f^{U}_{W_{1},W_{2},b}$ (4.9)
produced from the graph $A_{3}$, when the dimension at the output vertex is
$d_{3}=1$, and $\sigma:\mathbb{R}^{d_{2}}\to\mathbb{R}^{d_{2}}$ is taken to be
$\sigma(\vec{x})=(\phi(x_{1}),\ldots,\phi(x_{d_{2}}))$. (Take $d_{2}=N$,
$W_{1}=(y_{j}^{T})_{j=1,\ldots,N}$, $b=(\theta_{j})_{j=1,\ldots,N}$, and
$W_{2}=(\alpha_{j})_{j=1,\ldots,N}$.) For general dimension $d_{3}$, we simply
have $(G_{1},\ldots,G_{d_{3}})$, where $G_{i}$ are of the same form as above
(with different $\alpha_{j,i}$) which can be used to approximate any given
continuous function $K\to\mathbb{R}^{d_{3}}$.
We will prove the following theorem. Consider the quiver with three vertices
as in Section 4.3.3, and the function
$f^{U}_{W_{1},W_{2},b}(v)=W_{2}(\sigma(W_{1}\cdot v+b))$ in (4.9), where
$\sigma$ is the multi-variable activation function on $\mathbb{R}^{d_{2}}$
made from $\mathbb{P}^{d_{2}}$.
###### Theorem 5.2.
Let $K$ be a compact set of $\mathbb{R}^{d_{1}}$, and
$f:K\to\mathbb{R}^{d_{3}}$ a continuous function. For any $\epsilon>0$, there
exists $d_{2}>0$ and $W_{1}\in\operatorname{Mat}(d_{2},d_{1})$,
$W_{2}\in\operatorname{Mat}(d_{3},d_{2})$, $b\in\mathbb{R}^{d_{2}}$ such that
$\|f^{U}_{W_{1},W_{2},b}-f\|_{L^{2}(K)}<\epsilon$.
The compact set is given as a subset in $\mathbb{R}^{n}$. Thus from now on we
restrict to the real field, which will suffice for the theorem. This means we
take real-valued matrices and the real part $\sigma$ of $\sigma_{\mathbb{C}}$.
### 5.1. Tropical limit
A crucial idea in the work of [Cyb89] is to compose $\phi$ with a rescaling,
so that it tends to a step function in the limit. We can apply such a
rescaling to the multi-variable function $\sigma:\mathbb{R}^{d}\to P^{\circ}$.
This is well-known in toric geometry and is called the tropical limit.
Let $\Sigma$ be the dual fan of the moment polytope $P$. $\Sigma$ is the
collection of cones that are dual to the boundary strata of the polytope $P$.
In particular, maximal cones of $\Sigma$ are one-to-one corresponding to
corners of $P$.
We assume that $|\Sigma|=\mathbb{R}^{d}$. $\mathbb{R}^{d}$ is stratified into
the relative interiors of cones in $\Sigma$. We recall the following
interesting fact from toric geometry. It plays an important role in the study
of holomorphic discs and Lagrangian Floer theory for toric varieties.
###### Lemma 5.3.
Let $X_{\Sigma}$ be a toric variety (equipped with any toric Kähler form). Let
$C$ be a cone in $\Sigma$. For any $x\not=0$ which lies in the relative
interior of $C$, $p_{C}=\lim_{t\to-\infty}\sigma(tx)$ exists and equal to a
point (which is independent of $x$) in the boundary stratum of $P$ that is
dual to $C$.
In other words, the family of functions $\sigma_{t}(x)=\sigma(tx)$ converges
(as $t\to+\infty$) to the discontinuous function $\sigma_{\infty}$, where
$\sigma_{\infty}(x)=p_{C}$ if $x$ belongs to the relative interior of $C$.
###### Proof.
The cone $C$ corresponds to a complex torus orbit of the toric variety $X$. To
be more explicit, consider a maximal cone $C^{\mathrm{max}}$ that contains
$C$. Without loss of generality, let $C^{\mathrm{max}}=\mathbb{R}_{\geq
0}\cdot\\{v_{1},\ldots,v_{d}\\}$, and $C=\mathbb{R}_{\geq
0}\cdot\\{v_{1},\ldots,v_{k}\\}$. $C^{\mathrm{max}}$ gives a local chart
$\mathbb{C}^{d}$ of the toric variety, and the complex torus orbit
corresponding to $C$ is given by $z_{1}=\ldots=z_{k}=0$.
We have a special point given by
$z_{1}=\ldots=z_{k}=0,\,z_{k+1}=\ldots=z_{d}=1$ in the orbit. We assert that
$p_{C}=\lim_{t\to-\infty}\sigma(tx)\in P$ (for any $x$ in the relative
interior of $C$) is the moment-map image of this point.
To see this, we write $x=\sum_{i=1}^{k}x_{i}v_{i}$ where $x_{i}\not=0$ for all
$i=1,\ldots,k$. Consider the lifting of
$tx=(tx_{1},\ldots,tx_{k},0,\ldots,0)$:
$(e^{tx_{1}},\ldots,e^{tx_{k}},1,\ldots,1)$ in the chart
$\mathbb{C}^{d}\subset X$. Then $\sigma(tx)$ is the moment-map image of
$(e^{tx_{1}},\ldots,e^{tx_{k}},1,\ldots,1)$. Taking $t\to-\infty$,
$(e^{tx_{1}},\ldots,e^{tx_{k}},1,\ldots,1)\to(z_{1}=\ldots=z_{k}=0,z_{k+1}=\ldots=z_{d}=1)$.
Thus $\sigma(tx)$ converges to the above special point $p_{C}$. ∎
In terms of solving equations, $p_{C}$ is the solution of the simultaneous
equations $x_{1}=\ldots=x_{k}=0$ and
$x_{i}\prod_{j=d+1}^{m}\ell_{j}^{v_{j,i}}\left(0,\ldots,0,x_{k+1},\ldots,x_{d}\right)=1$
for $i=k+1,\ldots,m$. (We have used the dual basis of
$\\{v_{1},\ldots,v_{d}\\}$ to write the coordinates of $P$, and
$v_{j}=\sum_{i=1}^{d}v_{j,i}v_{i}$.) By above, the solution exists and is
unique.
###### Example 5.4.
Consider $X=\mathbb{P}^{d}$. Denote the coordinates of $\mathbb{R}^{d}$ by
$(x_{1},\ldots,x_{d})$, and set $x_{0}\equiv 0$. The $(d+1-l)$-cones $C$ of
$\Sigma$ are given by $\left\\{x_{i_{1}}=\ldots=x_{i_{l}}>x_{k}\textrm{ for
all }k\in\\{0,\ldots,d\\}-\\{i_{1},\ldots,i_{l}\\}\right\\}$, where
$i_{1},\ldots,i_{l}\in\\{0,\ldots,d\\}$ are fixed, $l=1,\ldots,d+1$. For
$\sigma=\left(\frac{e^{2x_{p}}}{1+\sum_{j=1}^{d}e^{2x_{j}}}\right)_{p=1}^{d}$,
the point $p_{C}=\lim_{t\to-\infty}\sigma(tx)$ has coordinates
$(p_{C})_{i_{r}}=1/l$ for $r=1,\ldots,l$ and $i_{r}\not=0$, and
$(p_{C})_{j}=0$ for all other $j\not=i_{1},\ldots,i_{l},0$. In particular, for
the maximal cones $S_{i}=\\{x_{i}>x_{k}\textrm{ for all
}k\in\\{0,\ldots,d\\}-\\{i\\}\\}$, $p_{S_{i}}=\epsilon_{i}$ for $i=0,\ldots,d$
where $\epsilon_{0}:=0$ and $\\{\epsilon_{1},\ldots,\epsilon_{d}\\}$ denotes
the standard basis.
Figure 9. The left shows the fan picture of $\mathbb{P}^{2}$, and the right
shows the moment-map polytope. The dots show the limit points $p_{C}$ for each
cone $C$ of the fan.
###### Corollary 5.5.
Let $K$ be any compact set in $\mathbb{R}^{d}$. For any $\epsilon>0$ and an
open neighborhood of the union $U$ of codimension-one strata of $\Sigma$,
there exists $t\gg 0$ such that $|\sigma_{t}(x)-\sigma_{\infty}(x)|<\epsilon$
for all $x\in K-U$. ($\sigma_{\infty}$ is defined in Lemma 5.3.)
###### Proof.
Any $x\in K-U$ belongs to one of the maximal cones $C$. By Lemma 5.3,
$\sigma_{t}(x)$ converges to $p_{C}$. Moreover, both $\sigma_{t}$ and
$\sigma_{\infty}$ are continuous on $K-U$. Then the result follows from the
compactness of $K-U$. ∎
In order to prove Theorem 5.2, we consider a particular type of polyhedral
decompositions of $\mathbb{R}^{n}$, which we call to be a centered simplicial
web.
### 5.2. Centered polyhedral web
###### Definition 5.6.
A centered simplicial web with $N$ ordered compact chambers in
$\mathbb{R}^{n}$ is a polyhedral decomposition of $\mathbb{R}^{n}$ whose
vertices are all trivalent, defined inductively on the number of compact
chambers as follows.
A centered simplicial web with zero compact chamber is the polyhedral
decomposition given by the fan of $\mathbb{P}^{n}$, up to an affine linear
isomorphism in $\mathrm{GL}(n,\mathbb{R})\ltimes\mathbb{R}^{n}$.
Now suppose the notion of a centered simplicial web with $N$ ordered compact
chambers has been defined, which has exactly $(n+1)$ non-compact rays (which
we call the outer rays), whose corresponding infinite lines intersect at
exactly one point called the $N$-th center that lies in the union of the $N$
compact chambers. Moreover, the web is required to have $(n+1)$ non-compact
chambers; each non-compact chamber is adjacent to $n$ outer rays and opposite
to the remaining one outer ray. (‘Opposite’ here means that the non-compact
chamber is disjoint from the corresponding outer ray, whose infinite line
intersects the chamber at a half-line.) The outer rays are one-to-one
corresponding to their opposite non-compact chambers.
A centered simplicial web with $(N+1)$ ordered compact chambers is defined as
follows. First, take a centered simplicial web with $N$ ordered compact
chambers. Second, we choose a non-compact chamber, and denotes the direction
of its opposite ray by a non-zero vector $v$. Third, we take an affine
hyperplane which intersects all the relative interior of the $n$ adjacent rays
of the non-compact chamber. This bounds a new compact chamber and the
intersection points $V_{i}$ are the new vertices. Finally, we choose the
$(N+1)$-th center to be $c_{N+1}=c_{N}-tv$, where $c_{N}$ is the $N$-th
center, and $t\in\mathbb{R}_{>0}$ is taken such that $c_{N+1}$ lies in the
union of the compact chambers (including the new one). Then we have $n$ new
rays emanated from the vertices $V_{i}$ whose infinite lines pass through
$c_{N+1}$. This gives a new web with $(N+1)$ ordered compact chambers, and it
still has $(n+1)$ non-compact chambers, each of which is adjacent to $n$ outer
rays and opposite to one outer ray.
See Figure 10 for some examples of centered simplicial webs in
$\mathbb{R}^{2}$.
Figure 10. Examples of centered simplicial web in $\mathbb{R}^{2}$. They have
zero,one, and two compact chambers respectively.
From the above definition, there is a one-to-one correspondence between the
centers and compact chambers. Moreover, the $(k+1)$-th compact chamber
$C_{k+1}$ (for $k=1,\ldots,N-1$) is associated with a one-strata of the web,
which is a subset of the opposite ray of the non-compact chamber containing
$C_{k+1}$ in the $(k+1)$-th inductive step. Furthermore, both the $k$-th and
$(k+1)$-th centers lie in the infinite line of the associated 1-strata of
$C_{k+1}$.
###### Remark 5.7.
The above notion is closely related to tropical subvarieties. In the tropical
context, there is an integral structure on the ambient space and the balancing
condition (whose definition requires the integral structure) is imposed on
each vertex of a tropical variety. However, we do not have an integral
structure here, since $\sigma_{\mathbb{C}}$ is defined on the universal cover
$\mathbb{C}^{d}$ rather than $(\mathbb{C}^{\times})^{d}$ (see Corollary 4.2
and Example 4.4). It means affine linear maps are taken over $\mathbb{R}$
rather than over $\mathbb{Z}$. Instead of the balancing condition, we impose
the notion of centers in the above definition.
###### Theorem 5.8.
Given a centered simplicial web $A$ with $N$ ordered compact chambers in
$\mathbb{R}^{n}$, there exists an affine-linear embedding
$L:\mathbb{R}^{n}\to\mathbb{R}^{d}$, where $d=n+N$, such that the $L$-preimage
of the fan of $\mathbb{P}^{d}$ in $\mathbb{R}^{d}$ equals to the web $A$.
###### Proof.
We shall prove the following statement: given a centered simplicial web $A$
with $N\geq 1$ compact chambers in $\mathbb{R}^{n}$, there exists a centered
simplicial web $B$ with $N-1$ compact chambers in $\mathbb{R}^{n+1}$ such that
the intersection of $\mathbb{R}^{n}\cong\mathbb{R}^{n}\times\\{0\\}$ with $B$
equals to $A$. Then by applying this statement $N$ times, we obtain a web
$B^{(N)}$ with zero compact chamber in $\mathbb{R}^{n+N}$ whose intersection
with $\mathbb{R}^{n}\times\\{0\\}$ gives $A$. By an affine linear isomorphism
on $\mathbb{R}^{n+N}$, $B^{(N)}$ is identified with the fan of
$\mathbb{P}^{n+N}$. The required map $L$ is given by the composition of the
inclusion $\mathbb{R}^{n}\times\\{0\\}\subset\mathbb{R}^{n+N}$ with this
linear isomorphism.
The above statement is proved by induction on $N$. First consider the case
$N=1$. We take a point $V$ away from
$\mathbb{R}^{n}\times\\{0\\}\subset\mathbb{R}^{n+1}$. Then we take a cone at
$V$ over the compact simplicial chamber of $A$. Moreover, the line joining $V$
with the given center of $A$ intersects with the complement of the cone and
produces a ray emanated from $V$. This gives a simplicial web in
$\mathbb{R}^{n+1}$ with no compact chamber, whose intersection with
$\mathbb{R}^{n}\times\\{0\\}$ is exactly $A$. (See Figure 11.)
Now suppose it is true for $N$. Consider a centered simplicial web $A$ in
$\mathbb{R}^{n}\cong\mathbb{R}^{n}\times\\{0\\}$ with $N+1$ ordered compact
chambers. We can take away the $(N+1)$-th compact chamber $C=C_{N+1}$ (and
forget the corresponding center $c_{N+1}$) and obtain a centered simplicial
web $A^{\prime}$ with $N$ compact chambers. By inductive hypothesis, there
exists a centered simplicial web $B^{\prime}$ in $\mathbb{R}^{n+1}$ with $N-1$
compact chambers whose intersection with $\mathbb{R}^{n}\times\\{0\\}$ gives
$A^{\prime}$.
The compact chamber $C$ of $A$ is contained in a non-compact chamber
$C^{\prime}$ of $A^{\prime}$, which is the intersection of
$\mathbb{R}^{n}\times\\{0\\}$ with a non-compact chamber $D^{\prime}$ of
$B^{\prime}$. $C^{\prime}$ is opposite to an outer ray $R$ of $A^{\prime}$,
which is the intersection of $\mathbb{R}^{n}\times\\{0\\}$ with a non-compact
2-plane $P$ of $B^{\prime}$. Note that the last two centers $c^{A}_{N}$ and
$c^{A}_{N+1}$ of $A$ are contained in the line of $R$, and hence contained in
the infinite 2-plane of $P$. Consider the two rays of $B$ that are adjacent to
$P$. Denote the one which is opposite to the chamber $D^{\prime}$ by $L$. The
other one is denoted by $L^{\prime}$, which must be adjacent to $D^{\prime}$.
Now we construct a web $B$ in $\mathbb{R}^{n+1}$ whose intersection with
$\mathbb{R}^{n}\times\\{0\\}$ gives $A$. A point $V$ in the relative interior
of the ray $L^{\prime}$ is taken to be a new vertex. Consider the line passing
through $V$ and the last center $c^{A}_{N+1}$. This line lies in the infinite
2-plane of $P$. Thus for a generic choice of $V$, it must intersect with the
infinite line of $L$ at a point, which we shall define as the new center
$c^{B}_{N}$ for $B$. $V$ is taken far away enough in the ray $L^{\prime}$ so
that the intersection point $c^{B}_{N}$ equals to $c^{B}_{N-1}-v$ for some
vector $v$ in the direction of $L$.
Consider the $n$ outer rays of $A^{\prime}$ that are adjacent to the chamber
$C^{\prime}$. They are the intersections of $\mathbb{R}^{n}\times\\{0\\}$ with
the corresponding $n$ outer-2-planes of $B^{\prime}$ that are adjacent to
$L^{\prime}$. The last chamber $C$ of $A$ is formed by the hyperplane through
the vertices taken in the relative interior of the $n$ outer rays of
$A^{\prime}$. The lines joining $V$ to these vertices in $A$ lie in the
outer-2-planes of $B^{\prime}$, and hence intersect with the corresponding
outer rays of $B^{\prime}$ at certain points, which we take to be new vertices
of $B$. The hyperplane through $V$ and these new vertices bounds a new
chamber. The lines joining $c^{B}_{N}$ with the new vertices produce the outer
rays of the new web $B$. This gives $B$ whose intersection with
$\mathbb{R}^{n}\times\\{0\\}$ equals to $A$. ∎
Figure 11. Construction of a tropical web with zero compact chamber whose
intersection with a hyperplane equals to a given tropical web with one compact
chamber.
The inductive step in the above proof is illustrated by Figure 12.
Figure 12. Construction of a tropical web with $3$ compact chambers whose
intersection with a hyperplane equals to a given tropical web with $2$ compact
chambers.
Next, we consider polytopes rather than simplices and the corresponding webs
formed from polytopes. Motivated from the well-known fact below, we define a
centered polyhedral web to be the intersection of a centered simplicial web
with an affine subspace.
###### Proposition 5.9.
For a polytope $P$ with $m$ facets in $\mathbb{R}^{n}$ where $m>n+1$, there
exists a simplex $S$ in $\mathbb{R}^{m-1}$ such that
$S\cap(\mathbb{R}^{n}\times\\{0\\})=P$ (where $\mathbb{R}^{n}$ is identified
with $\mathbb{R}^{n}\times\\{0\\}$).
The simplex $S$ in Proposition 5.9 can be constructed as follows. Without loss
of generality suppose $0\in P$. Consider the dual polytope
$P^{\vee}=\\{\nu\in(\mathbb{R}^{n})^{*}:(\nu,v)\leq 1\textrm{ for all }v\in
P\\}$, which is the convex hull of its vertices $\nu_{i}$ for $i=1,\ldots,m$.
Then we have a surjective map from the standard simplex
$S^{\vee}=\\{\sum_{i}a_{i}\epsilon_{i}^{*}\in(\mathbb{R}_{\geq
0}^{m})^{*}:\sum_{i}a_{i}=1\\}$ (where $\\{\epsilon_{i}^{*}:i=1,\ldots,m\\}$
is the standard basis) to $P^{\vee}$ by sending $\epsilon_{i}^{*}$ to
$\nu_{i}$. $S^{\vee}$ in the affine subspace $\\{\sum_{i}a_{i}=1\\}$ can be
identified as a simplex in $(\mathbb{R}_{\geq 0}^{m-1})^{*}$ by the projection
along the direction $-\sum_{i}\epsilon_{i}$. Then the dual linear map gives
the desired linear injection $\mathbb{R}^{n}\to\mathbb{R}^{m-1}$ which sends
$P$ into $S$. By composing with a linear isomorphism, the image of
$\mathbb{R}^{n}$ can be made to be $\mathbb{R}^{n}\times\\{0\\}$.
As a result, a centered polygonal web with one compact chamber (which is
constructed by taking a polygon with a chosen center $c$ and outer rays at
vertices whose lines pass through $c$) can be obtained as an intersection with
$\mathbb{R}^{n}\times\\{0\\}$ of a centered simplicial web with one compact
chamber in $\mathbb{R}^{m-1}$ (where $m$ is the number of non-compact
chambers). See the left of Figure 13 for an example.
Figure 13. The left shows an example of a polyhedral web with one compact
chamber, which is given as an intersection of a simplicial web with a
subspace. The right two figures show a concentric simplicial and a polygonal
web.
The following degenerate configuration will be helpful. In Definition 5.6,
suppose we take all the centers to be the same. Moreover, suppose the new
hyperplane introduced to bound a new chamber is allowed to intersect the outer
rays at the original vertices (rather than their relative interior). Then we
can construct the following configuration.
###### Definition 5.10.
Let’s take $(n+1)$ rays emanated from $0\in\mathbb{R}^{n}$, such that any $n$
of them are linearly independent. For each ray, we take a sequence of distinct
points $V_{i,k}$ (where $i=1,\ldots,n+1$ is indexing the ray) such that
$V_{i,k+1}-V_{i,k}$ is pointing in the ray direction. Then for each $k>0$, we
take a simplex with vertices at $V_{i,k}$ for $i=1,\ldots,n+1$. This gives a
polyhedral decomposition of $\mathbb{R}^{n}$. This is called a concentric
simplicial web.
By taking an intersection of the above degenerate simplicial configuration
with a subspace (that passes through the center $0$), we get a configuration
made from a sequence of polytopes whose vertices lie in a fixed collection of
$p$ rays, where $p$ is the number of vertices of each polytope. We call this a
concentric polyhedral web. See the right of Figure 13.
### 5.3. Proof of the approximation theorem
We are now ready to prove Theorem 5.2.
###### Proof of Theorem 5.2.
For any $\delta>0$, we can take a concentric polyhedral web $B$ in
$\mathbb{R}^{d_{1}}$, such that for every chamber $C$ of $B$, $C\cap K$ is
contained in a $\delta$-ball. $B$ is constructed as follows. Without loss of
generality, let $K\ni 0$. First, we take a polytope $P$ that lies in a
$\delta$-ball centered at $0\in\mathbb{R}^{d_{1}}$. $P$ induces a subdivision
on the unit sphere $\mathbb{S}^{d_{1}-1}$ by projecting its boundary strata
onto $\mathbb{S}^{d_{1}-1}$. $P$ is taken with sufficiently many vertices such
that the induced subdivision on the unit sphere $\mathbb{S}^{d_{1}-1}$ lies in
a $\delta^{\prime}$-ball for a chosen $\delta^{\prime}$. By Proposition 5.9,
$P=S\cap(\mathbb{R}^{d_{1}}\times\\{0\\})$ for some simplex $S$ in
$\mathbb{R}^{m-1}$ where $m$ is the number of facets of $P$. We take $0\in S$
to be the center. Then we take rays from $0$ through the vertices of $S$ and
construct a concentric simplicial web. Since $K$ is compact, by taking
$\delta^{\prime}$ sufficiently small, and the sequences of vertices in the
rays sufficiently close to each other, the resulting concentric polyhedral web
$B$ can be made such that every chamber intersects $K$ in a $\delta$-ball.
Next, we take a centered simplicial web $A$ in $\mathbb{R}^{m-1}$ whose
centers are chosen sufficiently close to each other, and the vertices in the
inductive steps are taken such that $A\cap K$ is sufficiently close $B\cap K$.
Namely, for every chamber $C$ of $A$, $C\cap K$ lies in the
$\delta$-neighborhood of $C^{\prime}\cap K$ for the corresponding chamber
$C^{\prime}$ of $B$. In particular, $C\cap K$ is contained in a
$2\delta$-ball.
By Theorem 5.8, there exists
$L^{\prime}:\mathbb{R}^{m-1}\to\mathbb{R}^{d_{2}}$ such that $A$ is the
$L^{\prime}$-preimage of the fan $\Sigma_{\mathbb{P}^{d_{2}}}$. By composing
$L^{\prime}$ with $\mathbb{R}^{d_{1}}\times\\{0\\}\subset\mathbb{R}^{m-1}$, we
obtain $L:\mathbb{R}^{d_{1}}\to\mathbb{R}^{d_{2}}$ such that for every maximal
cone $S_{i}$ of $\Sigma_{\mathbb{P}^{d_{2}}}$, $L^{-1}(S_{i})\cap K$ lies in a
$(2\delta)$-ball.
Since $f$ is uniformly continuous in $K$, for every $\epsilon>0$, $\delta$ can
be taken such that $|f(x)-f(y)|<\epsilon$ for every $x,y\in K$ lying in a
$2\delta$-ball. In particular, we have a step function
$s=\sum_{C}r_{C}\delta_{C}$ supported over $A$ (where $\delta_{C}(x)=1$ for
$x\in C$ and $0$ otherwise, and $C$ are chambers of $A$) such that
$\|f-s\|_{L^{2}(K)}<\epsilon\sqrt{\mathrm{Vol}(K)}$.
We have the step function $\sigma_{\infty}$ which sends the interior of the
maximal cones $S_{i}$ of $\Sigma_{\mathbb{P}^{d_{2}}}$ to $e_{i}$ for
$i=0,\ldots,d_{2}$, where $e_{0}=0$. (See Example 5.4.) The cone $S_{i}$
corresponds to chambers $L^{-1}(S_{i})$ of $A$ under $L$. Since
$\\{e_{i}-e_{0}:i=1,\ldots,d_{2}\\}$ forms a basis, there exists a unique
affine linear map $W_{2}:\mathbb{R}^{d_{2}}\to\mathbb{R}^{d_{3}}$ which sends
$e_{i}$ to $r_{L^{-1}(S_{i})}$ for all $i=0,\ldots,d_{2}$. Thus
$s=W_{2}\circ\sigma_{\infty}\circ L$.
Finally, by Corollary 5.5, there exists $t\gg 0$ such that
$|W_{2}\circ\sigma_{t}-W_{2}\circ\sigma_{\infty}|<\epsilon/2$ on $L(K)-U$,
where $U$ is an arbitrary open neighborhood of the codimension-one strata of
$\Sigma_{\mathbb{P}^{d}_{2}}$. Moreover,
$|W_{2}\circ\sigma_{t}-W_{2}\circ\sigma_{\infty}|$ is bounded. Hence by taking
$\mathrm{Vol}(L^{-1}(U)\cap K)$ sufficiently small, we have
$\|W_{2}\circ\sigma_{t}\circ L-W_{2}\circ\sigma_{\infty}\circ
L\|_{L^{2}(K)}<\epsilon$. In conclusion,
$\|W_{2}\circ\sigma\circ(tL)-f\|_{L^{2}(K)}\leq\|W_{2}\circ\sigma\circ(tL)-W_{2}\circ\sigma_{\infty}\circ
L\|_{L^{2}(K)}+\|W_{2}\circ\sigma_{\infty}\circ L-f\|_{L^{2}(K)}$ can be made
arbitrarily small. ∎
## References
* [Abo06] M. Abouzaid, _Homogeneous coordinate rings and mirror symmetry for toric varieties_ , Geom. Topol. 10 (2006), 1097–1157.
* [Abr98] M. Abreu, _Kähler geometry of toric varieties and extremal metrics_ , Internat. J. Math. 9 (1998), no. 6, 641–651.
* [AJ20] M. A. Armenta and P.-M. Jodoin, _The representation theory of neural networks_ , preprint (2020), arXiv:2007.12213.
* [CAW+19] M.C.N. Cheng, V. Anagiannis, M. Weiler, P. de Haan, T.S. Cohen, and M. Welling, _Covariance in physics and convolutional neural networks_ , preprint (2019), arXiv:1906.02481.
* [CB03] W. Crawley-Boevey, _Normality of Marsden-Weinstein reductions for representations of quivers_ , Math. Ann. 325 (2003), no. 1, 55–79. MR 1957264
* [CDS15a] X. Chen, S. Donaldson, and S. Sun, _Kähler-Einstein metrics on Fano manifolds. I: Approximation of metrics with cone singularities_ , J. Amer. Math. Soc. 28 (2015), no. 1, 183–197.
* [CDS15b] by same author, _Kähler-Einstein metrics on Fano manifolds. II: Limits with cone angle less than $2\pi$_, J. Amer. Math. Soc. 28 (2015), no. 1, 199–234.
* [CDS15c] by same author, _Kähler-Einstein metrics on Fano manifolds. III: Limits as cone angle approaches $2\pi$ and completion of the main proof_, J. Amer. Math. Soc. 28 (2015), no. 1, 235–278.
* [CGKW18] T.S. Cohen, M. Geiger, J. Koehler, and M. Welling, _Spherical cnns_ , ICLR (2018).
* [CGW19] T.S. Cohen, M. Geiger, and M. Weiler, _A general theory of equivariant cnns on homogeneous spaces_ , NeurlPS (2019), arXiv:1811.02017.
* [CLL12] K. Chan, S.-C. Lau, and N.C. Leung, _SYZ mirror symmetry for toric Calabi-Yau manifolds_ , J. Differential Geom. 90 (2012), no. 2, 177–250.
* [CSW18] X. Chen, S. Sun, and B. Wang, _Kähler-Ricci flow, Kähler-Einstein metric, and K-stability_ , Geom. Topol. 22 (2018), no. 6, 3145–3173.
* [CW16] T.S. Cohen and M. Welling, _Group equivariant convolutional networks_ , Proceedings of The 33rd International Conference on Machine Learning, vol. 48, 2016, pp. 2990–2999.
* [CWKW19] T.S. Cohen, M. Weiler, B. Kicanaoglu, and M. Welling, _Gauge equivariant convolutional networks and the icosahedral cnn_ , Proceedings of the International Conference on Machine Learning (ICML), 2019.
* [Cyb89] G. Cybenko, _Approximation by superpositions of a sigmoidal function_ , Math. Control Signals Systems 2 (1989), no. 4, 303–314.
* [CYRL19] I. Chami, Z. Ying, C. Ré, and J. Leskovec, _Hyperbolic graph convolutional neural networks_ , Advances in neural information processing systems (2019).
* [dHCW20] P. de Haan, T. Cohen, and M. Welling, _Natural graph networks_ , preprint (2020), arXiv:2007.08349.
* [Don85] S.K. Donaldson, _Anti self-dual Yang-Mills connections over complex algebraic surfaces and stable vector bundles_ , Proc. London Math. Soc. (3) 50 (1985), no. 1, 1–26.
* [Don99] by same author, _Symmetric spaces, Kähler geometry and Hamiltonian dynamics_ , Northern California Symplectic Geometry Seminar, Amer. Math. Soc. Transl. Ser. 2, vol. 196, Amer. Math. Soc., Providence, RI, 1999, pp. 13–33.
* [Don12] by same author, _Stability, birational transformations and the Kahler-Einstein problem_ , Surveys in differential geometry. Vol. XVII, Surv. Differ. Geom., vol. 17, Int. Press, Boston, MA, 2012, pp. 203–228.
* [DPG+14] Y. Dauphin, R. Pascanu, C. Gulcehre, K. Cho, S. Ganguli, and Y. Bengio, _Identifying and attacking the saddle point problem in high-dimensional non-convex optimization_ , preprint (2014), arXiv:1406.2572.
* [ER09] J. Engel and M. Reineke, _Smooth models of quiver moduli_ , Math. Z. 262 (2009), no. 4, 817–848.
* [Fed13] S. Fedotov, _Framed moduli and Grassmannians of submodules_ , Trans. Amer. Math. Soc. 365 (2013), no. 8, 4153–4179.
* [FHS95] A. Floer, H. Hofer, and D. Salamon, _Transversality in elliptic Morse theory for the symplectic action_ , Duke Math. J. 80 (1995), no. 1, 251–292.
* [Flo89] A. Floer, _Witten’s complex and infinite-dimensional Morse theory_ , J. Differential Geom. 30 (1989), no. 1, 207–221.
* [FLTZ12] B. Fang, C.-C. M. Liu, D. Treumann, and E. Zaslow, _T-duality and homological mirror symmetry for toric varieties_ , Adv. Math. 229 (2012), no. 3, 1875–1911.
* [FO99] K. Fukaya and K. Ono, _Arnold conjecture and Gromov-Witten invariant_ , Topology 38 (1999), no. 5, 933–1048.
* [Ful93] W. Fulton, _Introduction to toric varieties_ , Annals of Mathematics Studies, vol. 131, Princeton University Press, Princeton, NJ, 1993, The William H. Roever Lectures in Geometry. MR 1234037 (94g:14028)
* [GBH18] O. Ganea, G. Becigneul, and T. Hofmann, _Hyperbolic neural networks_ , Advances in Neural Information Processing Systems (S. Bengio, H. Wallach, H. Larochelle, K. Grauman, N. Cesa-Bianchi, and R. Garnett, eds.), vol. 31, Curran Associates, Inc., 2018, pp. 5345–5355.
* [GGL19] M. Goyal, R. Goyal, and B. Lall, _Learning activation functions: A new paradigm for understanding neural networks_ , preprint (2019), arXiv:1906.09529.
* [Gui94] V. Guillemin, _Kaehler structures on toric varieties_ , J. Differential Geom. 40 (1994), no. 2, 285–309.
* [Ham82] R.S. Hamilton, _Three-manifolds with positive Ricci curvature_ , J. Differential Geometry 17 (1982), no. 2, 255–306.
* [HKL18] H. Hansol, Y. Kim, and S.-C. Lau, _Immersed two-spheres and SYZ with application to Grassmannians_ , arXiv:1805.11738 (2018).
* [HSTT18a] K. Hashimoto, S. Sugishita, A. Tanaka, and A. Tomiya, _Deep learning and holographic QCD_ , Phys. Rev. D 98 (2018), no. 10, 106014, 15.
* [HSTT18b] by same author, _Deep learning and the ads/cft correspondence_ , Physical Review D 98 (2018), no. 4.
* [HY20] Y.-H. He and S.-T. Yau, _Graph laplacians, riemannian manifolds and their machine-learning_ , preprint (2020), arXiv:2006.16619.
* [Kin94] A.D. King, _Moduli of representations of finite-dimensional algebras_ , Quart. J. Math. Oxford Ser. (2) 45 (1994), no. 180, 515–530.
* [LJ18] H. Lin and S. Jegelka, _Resnet with one-neuron hidden layers is a universal approximator_ , Advances in Neural Information Processing Systems (S. Bengio, H. Wallach, H. Larochelle, K. Grauman, N. Cesa-Bianchi, and R. Garnett, eds.), vol. 31, Curran Associates, Inc., 2018, pp. 6169–6178.
* [LLPS93] M. Leshno, V.Ya. Lin, A. Pinkus, and S. Schocken, _Multilayer feedforward networks with a nonpolynomial activation function can approximate any function_ , Neural Networks 6 (1993), no. 6, 861 – 867.
* [LPW+17] Z. Lu, H. Pu, F. Wang, Z. Hu, and L. Wang, _The expressive power of neural networks: A view from the width_ , preprint (2017), arXiv:1709.02540.
* [Mar00] S. Martin, _Symplectic quotients by a nonabelian group and by its maximal torus_ , preprint (2000), arXiv:math/0001002.
* [MM92] H.N. Mhaskar and C.A. Micchelli, _Approximation by superposition of sigmoidal and radial basis functions_ , Adv. in Appl. Math. 13 (1992), no. 3, 350–373.
* [Nak94] H. Nakajima, _Instantons on ALE spaces, quiver varieties, and Kac-Moody algebras_ , Duke Math. J. 76 (1994), no. 2, 365–416.
* [Nak96] by same author, _Varieties associated with quivers_ , Representation theory of algebras and related topics (Mexico City, 1994), CMS Conf. Proc., vol. 19, Amer. Math. Soc., Providence, RI, 1996, pp. 139–157.
* [Nak01] by same author, _Quiver varieties and finite-dimensional representations of quantum affine algebras_ , J. Amer. Math. Soc. 14 (2001), no. 1, 145–238.
* [NNU10] T. Nishinou, Y. Nohara, and K. Ueda, _Toric degenerations of Gelfand-Cetlin systems and potential functions_ , Adv. Math. 224 (2010), no. 2, 648–706.
* [Ono95] K. Ono, _On the Arnold conjecture for weakly monotone symplectic manifolds_ , Invent. Math. 119 (1995), no. 3, 519–537.
* [PDGB14] R. Pascanu, Y.N. Dauphin, S. Ganguli, and Y. Bengio, _On the saddle point problem for non-convex optimization_ , preprint (2014), arXiv:1405.4604.
* [Per02] G. Perelman, _The entropy formula for the ricci flow and its geometric applications_ , preprint (2002), arXiv:math/0211159.
* [Per03] by same author, _Ricci flow with surgery on three-manifolds_ , preprint (2003), arXiv:math/0303109.
* [Pet99] P.P. Petrushev, _Approximation by ridge functions and neural networks_ , SIAM J. Math. Anal. 30 (1999), no. 1, 155–189.
* [Pin99] A. Pinkus, _Approximation theory of the MLP model in neural networks_ , Acta numerica, 1999, Acta Numer., vol. 8, Cambridge Univ. Press, Cambridge, 1999, pp. 143–195.
* [Rei08] M. Reineke, _Framed quiver moduli, cohomology, and quantum groups_ , J. Algebra 320 (2008), no. 1, 94–115.
* [Sem92] S. Semmes, _Complex Monge-Ampère and symplectic manifolds_ , Amer. J. Math. 114 (1992), no. 3, 495–550.
* [Tia97] G. Tian, _Kähler-Einstein metrics with positive scalar curvature_ , Invent. Math. 130 (1997), no. 1, 1–37.
* [UY86] K. Uhlenbeck and S.-T. Yau, _On the existence of Hermitian-Yang-Mills connections in stable vector bundles_ , Comm. Pure Appl. Math. 39 (1986), no. S, suppl., S257–S293, Frontiers of the mathematical sciences: 1985 (New York, 1985).
* [Yau96] S.-T. Yau, _Review of kähler-einstein metrics in algebraic geometry_ , Israel Math. Conference Proceedings, Bar Ilan Univ., 1996, pp. 433–443.
|
# Forward Beam Monitor for the KATRIN experiment
A. Beglarian E. Ellinger111Corresponding author. N. Haußmann K. Helbing S.
Hickford U. Naumann H.-W. Ortjohann M. Steidl J. Wolf and S. Wüstling
###### Abstract
The KArlsruhe TRItium Neutrino (KATRIN) experiment aims to measure the
neutrino mass with a sensitivity of $0.2\text{\,}\mathrm{eV}$
($90\text{\,}\mathrm{\char 37\relax}$ CL). This will be achieved by a
precision measurement of the endpoint region of the $\upbeta$-electron
spectrum of tritium decay. The $\upbeta$-electrons are produced in the
Windowless Gaseous Tritium Source (WGTS) and guided magnetically through the
beamline. In order to accurately extract the neutrino mass the source activity
is required to be stable and known to a high precision. The WGTS therefore
undergoes constant extensive monitoring from several measurement systems. The
Forward Beam Monitor (FBM) is one such monitoring system.
The FBM system comprises a complex mechanical setup capable of inserting a
detector board into the KATRIN beamline with a positioning precision of better
than $0.3\text{\,}\mathrm{mm}$. The electron flux density at that position is
on the order of
${10}^{6}\text{\,}{\mathrm{s}}^{-1}\text{\,}{\mathrm{mm}}^{-2}$. The detector
board contains two silicon detector chips of p-i-n diode type which can
measure the $\upbeta$-electron flux from the source with a precision of
$0.1\text{\,}\mathrm{\char 37\relax}$ within $60\text{\,}\mathrm{s}$ with an
energy resolution of FWHM $=$ $2\text{\,}\mathrm{keV}$.
The unique challenge in developing the FBM arise from its designated operating
environment inside the Cryogenic Pumping Section which is a potentially
tritium contaminated ultra-high vacuum chamber at cryogenic temperatures in
the presence of a $1\text{\,}\mathrm{T}$ strong magnetic field. Each of theses
parameters do strongly limit the choice of possible materials which e.g.
caused difficulties in detector noise reduction, heat dissipation and
lubrication. In order to completely remove the FBM from the beam tube a
$2\text{\,}\mathrm{m}$ long traveling distance into the beamline is needed
demanding a robust as well as highly precise moving mechanism.
## 1 Introduction
The KATRIN experiment will improve the sensitivity of neutrino mass
measurements to $m_{\nu}=$ $0.2\text{\,}\mathrm{eV}$
($90\text{\,}\mathrm{\char 37\relax}$ C.L.) corresponding to a
$5\text{\,}\sigma$ discovery potential for a mass signal of $m_{\nu}=$
$0.35\text{\,}\mathrm{eV}$ [1, 2] in the most sensitive direct neutrino mass
experiment to date. The neutrino mass will be derived from a precise
measurement of the shape of the tritium $\upbeta$-decay spectrum near its
endpoint at $E_{0}=$ $18\,573.7\pm 0.1\text{\,}\mathrm{eV}$ [3]. The source of
$\upbeta$-electrons is a Windowless Gaseous Tritium Source (WGTS) which has an
activity of ${10}^{11}\text{\,}\mathrm{Bq}$.
The layout of the KATRIN beamline [4] is shown in figure 1. The Source and
Transport Section (STS) consists of the WGTS, the Differential Pumping Section
(DPS), the Cryogenic Pumping Section (CPS), and several source monitoring and
calibration systems [5]. Along the beamline superconducting solenoids generate
a magnetic field of several Tesla strength which adiabatically guides the
$\upbeta$-electrons towards the spectrometers while excess tritium is pumped
out of the system. The Spectrometer and Detector Section (SDS) consists of the
pre-spectrometer, the main-spectrometer, the monitor-spectrometer, and the
Focal Plane Detector (FPD). All spectrometers are of MAC-E-Filter type which
transmit electrons with energies above a chosen retarding energy [6], and
reject those with lower energies. The main-spectrometer can perform an energy
analysis of the $\upbeta$-electrons with an energy resolution of
$0.93\text{\,}\mathrm{eV}$ at $18.6\text{\,}\mathrm{keV}$.
Figure 1: The KATRIN beamline. The FBM is located at the end of the CPS and
represents the final source monitoring system before the $\upbeta$-electrons
enter the spectrometer and detector section.
The source-related parameters associated with the main systematic
uncertainties in the determination of the neutrino mass are activity
fluctuations of the WGTS, energy loss corrections (of $\upbeta$-electron
scattering in the WGTS), the final state distribution, the source magnetic
field, and the source plasma condition.
In order to analyse the tritium $\upbeta$-spectrum and determine the neutrino
mass the WGTS needs to be extremely stable, particularly in its activity and
isotopic composition. Therefore, the WGTS properties need to be known with
high precision, and are continuously monitored for short and long term
fluctuations. There are several monitoring and calibration subsystems
associated with the WGTS [5].
Results from the various subsystems are combined over long time periods during
extended measurement time. This paper focuses on one such activity monitoring
system, the Forward Beam Monitor (FBM). The FBM is the final monitoring
subsystem for $\upbeta$-electrons from the source before they enter into the
spectrometer and detector section. It has been commissioned prior to the
KATRIN krypton measurement campaign in June 2017 [7]. Initial data was then
obtained during the krypton measurement campaign and during the KATRIN first
tritium measurement campaign in May 2018 [8]. The FBM is capable of
continuously monitoring variations of the electron flux and changes in the
measured shape of the $\upbeta$-decay spectrum during the KATRIN neutrino mass
measurement phases.
This paper is organised as follows. In section 2 the WGTS and its operating
parameters are introduced and in section 3 the FBM measurement principle for
the monitoring of the relevant WGTS parameters is explained. Section 4
contains a technical description of the FBM. In section 5 the FBM
commissioning and results from the krypton and first tritium measurement
phases are presented, and section 6 contains the summary.
## 2 Tritium source
The WGTS is the origin of $\upbeta$-electrons whose observed spectrum will
ultimately lead to the measurement of the neutrino mass [9]. The general setup
of the WGTS is shown in figure 2. It is a column of tritium gas inside a
cylinder with a diameter of $90\text{\,}\mathrm{mm}$ and a length of
$10\text{\,}\mathrm{m}$. The latter is situated in a homogeneous magnetic
field of $3.6\text{\,}\mathrm{T}$ generated by superconducting solenoid
magnets. The tritium gas is injected in the middle of the beam tube with an
adjustable pressure $p_{\text{in}}$ = ${10}^{-3}\text{\,}\mathrm{mbar}$, and
is pumped out at both ends with a constant outlet pressure of
$p_{\text{out}}=$ $0.05$ $p_{\text{in}}$.
Figure 2: Setup of the WGTS. Tritium is injected into the centre of the
cylinder and pumped out at both ends. The flux tube is surrounded by
superconducting magnets to guide the $\upbeta$-electrons. The longitudinal
density profile of the tritium molecules along the column is shown above.
### 2.1 Column density
The column density is defined as tritium molecule density integrated along the
central axis of the source, i.e., the number of tritium molecules per source
cross section area. The neutrino mass measurement depends on the accurate
description of inelastic scattering of electrons by the gas molecules inside
the source. There are several key parameters of the WGTS that need to be kept
stable with high precision in order to achieve a high sensitivity in the
neutrino mass measurement. These include
* •
Beam tube temperature
The molecular tritium gas must be at cryogenic temperatures of
$<80\text{\,}\mathrm{K}$ to minimise corrections to the electrons energy due
to thermal movement of the decaying mother atoms. The cooling concept is based
on a two-phase liquid neon thermosiphon [10, 11].
* •
Pressure
The amount of tritium inside the source scales with the inlet pressure.
Stabilisation is achieved using a pressurised control vessel from which
tritium flows via a capillary to the beam tube.
* •
Tritium purity
A high isotopic purity of molecular tritium gas ($>95\text{\,}\mathrm{\char
37\relax}$) is required. The tritium purity $\mathbf{\epsilon_{T}}$ is given
by the ratio of the number of tritium atoms to the total sum of atoms in the
WGTS. In addition to T2 other isotopolouges include DT, HT, D2, HD, and H2.
The tritiated hydrogen isotopolouges differ in their mass, recoil energies,
and the rotational and vibrational final state distributions of their daughter
molecules following tritium decay. The gas composition is measured via LAser
RAman spectroscopy (LARA) [12, 13].
These key parameters have an effect on the rate and/or energy of the electrons
emitted from the source. There are several control and monitoring systems in
the KATRIN experiment with the purpose of meeting the precision and stability
requirements of the key source parameters.
The column density, $\mathcal{N}$, can be obtained by combining an in-situ
measurement of the tritium purity with an activity (decay rate) measurement.
The count rate $S$ of $\upbeta$-electrons from the source as measured by
dedicated particle detectors (activity monitors, see section 2.3) scales as
$S=C\cdot\epsilon_{T}\cdot\mathcal{N}$ (2.1)
where $C$ is a proportionality constant encompassing experimental properties
such as detector efficiency and acceptance, and the half-life of tritium.
Small fluctuations of the source parameters lead to changes in the observed
shape of the differential $\upbeta$-electron spectrum. Fluctuations in the
column density are expected to be in the ${10}^{-3}$ regime. Given the
targeted sensitivity for the neutrino mass measurement, column density and
tritium purity must not give rise to an uncertainty beyond $\delta
m^{2}_{\nu}=$ $7.5\text{\times}{10}^{-3}\text{\,}{\mathrm{eV}}^{2}$ to the
neutrino mass analysis.
### 2.2 Electron transport
The $\upbeta$-electrons resulting from the decay of the tritium are
adiabatically guided towards the spectrometer and detector section. The
transport section is also used to eliminate the tritium flow towards the
spectrometers which must be free of tritium in order to meet the necessary
background requirements for neutrino mass measurements. The transport section
consists of a Differential Pumping Section (DPS) and a Cryogenic Pumping
Section (CPS).
The DPS consists of five beam tube segments within superconducting solenoids
with turbomolecular pumps between each [14]. The CPS traps all remaining
traces of tritium by cryo-sorption on argon frost at $4\text{\,}\mathrm{K}$
condensed on the gold plated surfaces of the beam tube [15, 16]. Both the DPS
and CPS have $20\text{\,}\mathrm{\SIUnitSymbolDegree}$ chicanes to block the
line of sight for the diffusing tritium gas and to increase the probability
that the tritium molecules get pumped away or hit the walls of the beam tube.
At the end of the transport section the tritium flow is suppressed by $14$
orders of magnitude compared to the center of the WGTS. The electron flow is
unaffected and all electrons are guided adiabatically towards the spectrometer
and detector section.
### 2.3 Activity monitors
The activity of the tritium source is monitored by two systems. These activity
monitors
1. 1.
provide information about fluctuations of the WGTS activity on a timescale of
minutes and
2. 2.
are used (together with the measured tritium purity) to monitor the column
density with $0.1\text{\,}\mathrm{\char 37\relax}$ precision, via equation
(2.1).
One of these activity monitors is located at the rear wall upstream of the
WGTS. This detector measures the X-rays created when the $\upbeta$-electrons
impact on the rear wall [5]. The second activity monitor is called the Forward
Beam Monitor (FBM). It is located downstream towards the main-spectrometer in
the transport section, mounted between the last two superconducting solenoids
of the CPS. Here the tritium flow has been suppressed by a factor of
$10^{$14$}$, to approximately
${10}^{-14}\text{\,}\mathrm{mbar}\text{\,}\mathrm{l}\text{\,}{\mathrm{s}}^{-1}$,
which minimises background effects and contamination from tritium. The
magnetic field in this position is axially symmetric with a magnitude of
$0.84\text{\,}\mathrm{T}$ so the spatial homogeneity of the source profile can
be studied. The FBM is the final measurement component before the spectrometer
and detector section.
## 3 Measurement principle
The FBM measures $\upbeta$-electrons from the tritium source as they are
guided to the spectrometer and detector section. Hence, the
$\upbeta$-electrons are following the beamline when they are detected by the
FBM. Such a detector must not shadow any part of the electron flux tube that
will be used for the measurement of the neutrino mass. Therefore, the FBM
configuration is such that the detector is located in the outer rim of the
electron flux during neutrino mass measurements. The active radius of the flux
tube used for measurement is approximately $71\text{\,}\mathrm{mm}$ and the
outer rim in which the detector is situated is up to $7\text{\,}\mathrm{mm}$
wide.
The p-i-n diode detectors have an energy threshold of approximately
$5\text{\,}\mathrm{keV}$, dependent on the background noise and the type of
diode used. This lower energy value is determined during calibration of each
diode. For an accurate rate measurement the lower energy threshold needs to be
stable.
Figure 3: Cross section of the FBM setup with the electron flux tube in the
KATRIN beamline. During nominal monitoring operation the FBM is situated in
the outer rim of the flux tube, up to approximately $7\text{\,}\mathrm{mm}$ in
thickness.
It is assumed that the activity measurement in the outer rim of the flux tube
is representative of the activity across the entire beamline cross section.
Variations of the column density in the radial direction are expected to be on
the ${10}^{-4}$ level [17]. The assumption that the outer rim is
representative of the entire flux tube is verified during repeated calibration
runs when the FBM is moved across the beamline. These two operation modes of
the FBM are standard “monitoring mode” and calibration “scanning mode” and are
described in the following sections.
### 3.1 Monitoring mode
Monitoring mode is the standard mode of operation for the FBM. It is intended
for permanent and continuous monitoring of the source activity and the main
observable is the electron count rate. Together with the measurement of the
tritium purity, the FBM monitoring mode provides continuously information on
the column density of the source.
### 3.2 Scanning mode
Flux tube scans are performed during calibration of the KATRIN experiment. The
purpose of scanning is to
1. 1.
confirm that the activity in the beamline outer rim is representative of the
entire flux tube,
2. 2.
map any irregularities in the cross section of the flux tube, and
3. 3.
investigate the area of the flux tube entering the spectrometer and detector
section (i.e. measure possible shadow effects by STS instrumentation).
During the KATRIN experiment calibration runs are performed between neutrino
mass measurement runs once every $\sim
60\text{\,}\mathrm{d}\mathrm{a}\mathrm{y}\mathrm{s}$. During commissioning and
initial measurement campaigns the scanning mode was used more frequently.
## 4 Technical description
In the following sections a technical description of the FBM is given. A more
detailed description can be found in ref. [18]. Further information on the
basic concept and the early development of the FBM can be found in ref. [19]
and [20].
### 4.1 Vacuum manipulator
The measurement of the electron flux is performed under ultra high vacuum
(UHV) conditions in a potentially tritium contaminated environment. The main
mechanical requirements for the vacuum manipulator are:
1. 1.
to situate the FBM detector board in the outer rim of the flux tube without
shadowing the main detector and additionally to move it throughout the cross
section of the flux tube,
2. 2.
to be capable of removing all FBM components out of the CPS allowing full
metal sealed vacuum gate valves to separate the FBM volume from the CPS
volume, and
3. 3.
to provide a safe enclosure for tritium, complying with all radiation safety
regulations of the tritium laboratory.
An overview of the complete FBM setup is shown in figure 4 and figure 5.
Figure 4: The FBM hardware setup. The FBM is attached to the CPS perpendicular
to the beamline. The CPS and FBM gate valves separate the FBM from the CPS if
the detector is in parking position within the FBM six-way cross. With the
help of the $2\text{\,}\mathrm{m}$ long bellow the detector arm can be driven
into the flux tube within the CPS.
The vacuum components of the FBM setup are separated from the CPS by a gate
valve. Behind this valve the FBM detector board is completely removed from the
KATRIN beamline. Attached to this volume are a turbomolecular pump and
pressure gauges. Behind the main FBM vacuum volume are bellows, support
structures, stepper motors, rotary encoders, and electrical feedthroughs.
These components provide the movement of the FBM detector board and the
readout of the measured data.
The movement of the detector board is realised by combining two linear drive
mechanisms. A long stainless steel support tube with an outer diameter of
$54\text{\,}\mathrm{mm}$ can be moved over a distance of
$1.8\text{\,}\mathrm{m}$ along its symmetry axis. At its forward end the
detector holder (hereafter known as the “front end”, see figure 6) is
attached. The support tube provides space for electrical feeding and a driving
rod which can be moved coaxially along the tube by approximately
$10\text{\,}\mathrm{cm}$. The latter linear movement is converted by the front
end into a rotary movement with a rack and pinion drive such that the
combination of these two movements enable the positioning of the detector
board in a two-dimensional plane.
Two edge-welded bellows are used to realise the linear movements in the
vacuum. The large bellow has an unusually long extended length of
$2223\text{\,}\mathrm{mm}$ with a working-stroke of
$1800\text{\,}\mathrm{mm}$. The back end provides electrical feedthroughs as
well as the mechanics for the rotary movement. The system is moved with a
$2\text{\,}\mathrm{m}$ long spindle drive featuring low play and two carriages
for more stability. To prevent the long bellows, the driving rod, and the
support tube from sagging and hanging down, several supports are added to the
setup. These include 3D printed trolleys outside the vacuum which can move
freely over the slider and are automatically pulled along from the motion of
the bellows, and structures with full ceramic ball bearings for supporting the
long tube and driving rod inside the vacuum chamber.
The front end which contains the FBM detector board is the mechanical and
electrical connection between the detector board and the manipulator. It
converts the linear movement of the driving rod into a rotary movement with a
low play rack-and-pinion drive to allow the movement in the vertical
direction. It is shown in figure 6.
To reduce magnetic force acting on the system, as well as to reduce influences
onto the electron guiding magnetic field, the front end, similar to all other
vacuum parts of the FBM, is made of low permeability $\mu_{\text{p}}$
materials (such as stainless steel $1.4429$ with $\mu_{\text{p}}$ $<1.005$).
To prevent cold welding of moveable parts the materials of the pinion
(stainless steel), rack (titanium), and the front end’s cylindrically shaped
main body (stainless steel), are alternated. A precise groove in the main
cylinder allows leading the rack with low play. To reduce friction, an ultra
low friction and UHV compatible dry lubrication is added, which mostly
consists of a coating with tungsten disulfide.
Figure 5: CAD drawing of the FBM as it is inserted into the CPS. In contrast
to figure 4 parts of the vacuum equipment, of the CPS and half of the FBM’s
long bellows are invisible for a better illustration of the mechanics.
To facilitate an easy slipping onto the second support flange the cylinder has
a chamber at its forward end. Two cut-outs extend the movement limits in
$y$-direction and provide space for the electrical feeding.
The axis of the detector holder is made of steel $1.4429$ like the pinion and
is mounted via dry full-ceramic ball bearings. The lever arm is also made of
steel $1.4429$, but the detector board holder (back plate) of aluminum, to
reduce weight. To shield the detector board from radio frequency and, even
more importantly, from the electron beam, a steel $1.4429$ cover was designed
featuring two small holes which allow electrons to reach the p-i-n diode
chips. The full lever arm length from the axis to the tip of the FBM
(including the cover, compare figure 6 ) is $130\text{\,}\mathrm{mm}$ and the
maximum height of the detector equals the height of the cover which is
$50\text{\,}\mathrm{mm}$. The electrical connector is shielded from the
electron beam by a thin steel plate.
The turbomolecular pump is located vertically above the main FBM vacuum volume
and is capable of pumping speeds up to
$260\text{\,}\mathrm{l}\text{\,}{\mathrm{s}}^{-1}$ (nitrogen). Two pressure
gauges are mounted below the FBM vacuum volume which cover the range from
$1.3\text{\times}{10}^{-10}\text{\,}\mathrm{mbar}$ to
$1.3\text{\times}{10}^{-2}\text{\,}\mathrm{mbar}$. In order to reach the
required vacuum level the setup is baked out periodically after being exposed
to atmosphere.
Figure 6: The FBM manipulator front end. The detector board is fixed on the
end of a lever arm which is rotated by a rack and pinion drive.
### 4.2 Motion control
The two stepper motors mentioned in subsection 4.1
($12.1\text{\,}\mathrm{N}\text{\,}\mathrm{m}$ and
$2.7\text{\,}\mathrm{N}\text{\,}\mathrm{m}$ holding torque,
$1.8\text{\,}\mathrm{\SIUnitSymbolDegree}$ resolution) are not directly acting
on the spindle axes but with one stage transmissions using toothed wheels.
Since the FBM is not equipped with motor breaks the $x$-transmission is chosen
such that the torque at the motor is sufficiently small to withstand the
vacuum forces even if it is not powered anymore.
Since it is possible that the stepper motors miss steps without being noticed,
absolute rotary encoders are used to determine the position of the FBM because
they retain the full information of the position even during a power cut.
These optical encoders work with up to $16$-bit single turn and $14$-bit multi
turn resolution, i.e. $2^{16}$ steps per revolution can be counted. This sums
up to an overall resolution of $2^{30}$ steps. To minimise mechanical play
both encoders are connected directly to their corresponding spindle axes. The
main spindle has a slope of $2.5\text{\,}\mathrm{mm}$, hence a theoretical
precision of ${10}^{-5}\text{\,}\mathrm{\SIUnitSymbolMicro m}$ can be reached.
However, due to mechanical tolerances the actual precision is significantly
lower as will be described in subsection 5.1.
To fulfill stringent safety requirements the motion control of the FBM is
implemented on a Field-Programmable Gate Array (FPGA) which continues to run
during power cuts with the help of three backup accumulators. It directly
monitors and controls the motor, encoders, and sensors and also includes a
fast full safety retraction of the FBM which allows closure of the safety gate
valves to separate the FBM volume from the CPS.
The FPGA communicates with two KATRIN internal database systems: the ZEntrale
datenerfassung Und Steuerung (ZEUS) server and the Advanced Data Extraction
Infrastructure (ADEI) server [21]. All data obtained by the FPGA is
automatically transferred and available on both servers. Safety-critical
systems, such as vacuum pumps, valves, pressure gauges, and end switches, are
integrated within the KATRIN PCS7 safety system.
### 4.3 Detector
The main tasks of the FBM are to monitor the electron flux within the electron
beam and to obtain the beta spectrum of tritium. Detector chips with a thin
entrance window (dead layer) are used to allow the detection of electrons with
energies below $10\text{\,}\mathrm{keV}$. In addition this also allows
detection of low energy ($<60\text{\,}\mathrm{keV}$) photons which is
important for calibrating the detector. The FBM features a UHV compatible two
channel detector board, including detector chips of silicon type and
additional sensors, as described below.
Figure 7: Left: The FBM detector board is made of polyimide and equipped with
SMD parts. The two p-i-n diodes are glued to the tip of the board and their
signals are amplified by two separate transimpedance amplifiers. Close to the
p-i-n diodes a PT-1000 temperature sensor and a Hall sensor are located.
Center: Collection of p-i-n diodes in TO-18 casing as they are delivered by
the producer [28]. Right: Picture of the Hamamatsu S9055 p-i-n diode with the
lid removed. The silicon diode itself is mounted on a ceramic carrier which
can be taken out of the casing.
#### 4.3.1 Detector board and back plate
The detector board (PCB) is made of polyimide which meets the vacuum and
material requirements because of its relatively low outgassing rate compared
to most other polymers and excellent thermal resistance. To enhance the
thermal through-plane conductivity of the board for dissipating the heat
produced by the electrical components, the PCB is a flexible, thin
($0.2\text{\,}\mathrm{mm}$) multilayer board which consists of alternating
polyimide and copper layers. Here the back plate acts as a passive heat sink
which is permanently cooled in the cold environment of the CPS. The board
contains two detector channels, each consisting of a detector chip and a
preamplifier.
A Hall sensor on the detector board determines the local magnetic field. In
this region of the CPS the magnetic field is approximately
$0.84\text{\,}\mathrm{T}$ in the centre of the flux tube and is axially
symmetric. The magnetic field is measured in only one axis and the electron
flux should follow this magnetic field exactly with the exception of upstream
blockages. The measurement of the magnetic field is therefore also useful for
additional positioning and alignment measurements.
Temperature stabilisation is important as the p-i-n diode leakage current
rises exponentially with detector temperature. Therefore, the energy
resolution and stability of the energy threshold are dependent on the detector
temperature and effect the spectra obtained. To record the temperature a
PT-1000 sensor is placed on the detector board near the p-i-n diodes and the
Hall sensor.
The PCB is mounted on a $5\text{\,}\mathrm{mm}$-thick aluminum back plate
attached to the moving components. It is glued to the back plate with a UHV
compatible two-component adhesive to ease the mounting of the electrical parts
and to improve the thermal contact. The electronics are covered by a stainless
steel metal shield to protect them from electrons and ions in the beamline as
well as from radio frequency interference. To allow electrons to reach the
p-i-n diodes the cover features two holes. The detector board has “cut out”
corners in order to reduce the area of the flux tube that is covered. The
electronics and detectors on the FBM detector board are connected via a
custom-made PEEK connector with cabling running through the FBM manipulator to
the vacuum feedthroughs.
#### 4.3.2 Preamplifiers and p-i-n diodes
The preamplifiers of the two p-i-n diode detector channels are DC coupled
charge sensitive amplifiers which operate in a continuous reset mode. Each
preamplifier consists of a low-noise JFET front end in common-source
configuration and an operational amplifier (op-amp) connected in a non-
inverting scheme. The feedback loop stretching across both stages consists of
a $R=$ $1\text{\,}\mathrm{G\SIUnitSymbolOhm}$ resistor in parallel with a $C=$
$0.5\text{\,}\mathrm{pF}$ capacitor, forming a time constant of $\tau=C\cdot
R=$ $0.5\text{\,}\mathrm{ms}$. Thanks to the DC coupled circuitry, not only
individual charge-generating events can be read out with a $F_{AC}=U/Q=1/C=$
$2\text{\,}\mathrm{V}\text{\,}{\mathrm{pC}}^{-1}$ translation factor, but also
a current readout can be performed by looking at the DC voltage offset at the
output of the preamplifier with $F_{DC}=$
$1\text{\,}\mathrm{V}\text{\,}{\mathrm{nA}}^{-1}$.
The fundamental components of the FBM are the p-i-n diode detector chips.
There are two silicon p-i-n diodes mounted on the detector board which detect
the $\upbeta$-electrons from the tritium source. These two p-i-n diodes can
have different active sensitive areas. The silicon p-i-n diodes are
manufactured by Hamamatsu Photonics and can be type S5971, S5972, S5973, or
S9055-01 which have sensitive areas of different sizes (see table 1). One
advantage of these detectors is that their casing and properties are all
identical, the only difference is their respective sensitive area. This means
the electronic design of the detector board can remain the same and the board
with the p-i-n diodes that most suits the measurement purposes can be mounted
and inserted into the flux tube. Furthermore the dead layer does not exceed
$1\text{\,}\mathrm{\SIUnitSymbolMicro m}$.
Diode | $\boldsymbol{A_{s}}$ [${\mathrm{mm}}^{2}$] | $\boldsymbol{C}$ [$\mathrm{pF}$] | $\boldsymbol{I_{\text{dark}}}$ [$\mathrm{pA}$] | $\boldsymbol{d_{\text{dead}}}$ [$\mathrm{nm}$] | $\boldsymbol{T_{s}}$ [$\mathrm{s}$] | $\boldsymbol{T_{m}}$ [$\mathrm{s}$]
---|---|---|---|---|---|---
| data sheet | (at $10\text{\,}\mathrm{V}$) | (at $10\text{\,}\mathrm{V}$) | data sheet | measured | |
S9055-01 | 0.008 | 0.5 | 2.0 | <1000 | 300–500 | 498 | 252
S9055 | 0.031 | 0.8 | 2.0 | 129 | 64.6
S5973 | 0.126 | 1.6 | 1.9 | 31.7 | 15.9
S5972 | 0.503 | 3.0 | 10 | 8.0 | 4.0
S5971 | 1.131 | 3.0 | 70 | 3.5 | 1.8
Table 1: Parameters of the FBM p-i-n diodes. Capacitance $C$ and dark current
$I_{\text{dark}}$ are taken from the data sheets. The thickness of the dead
layer $d_{\text{dead}}$ was determined with an electron gun and Monte Carlo
simulations. The two right columns show the time to build a sufficiently
detailed spectrum in monitoring mode ($T_{m}$) and scanning mode ($T_{s}$).
The casing of these diodes is metal and includes a large glass window. Since
the windows of these TO-18 casings would prevent the detection of any
electrons the diodes are removed from the housing and directly mounted (using
two-component adhesive) onto the FBM detector board. The Hamamatsu S5971 p-i-n
diode detector chip is shown in figure 7.
The choice of the p-i-n diode size is based on the expected rate from the
tritium source within each measurement phase (larger diodes are used for
commissioning measurements where the amount of tritium is lower).
The statistical error of the measurement is dominated by the number of
electrons that are counted by the detector and is given by
$\frac{\Delta N}{N}=\frac{1}{\sqrt{N}}=\frac{1}{\sqrt{A\phi\epsilon t}}$ (4.1)
where $A$ is the sensitive area of the p-i-n diode, $\phi$ is the electron
flux density, $\epsilon$ is the detector efficiency, and $t$ is the
measurement time. The detector efficiency includes losses due to back
reflected electrons and pile-up effects. To reach the required precision of
$\Delta N/N=$ $0.1\text{\,}\mathrm{\char 37\relax}$ the measurement time is
$t=\frac{1}{0.001^{2}A\phi\epsilon}$ (4.2)
Assuming an energy threshold of $7\text{\,}\mathrm{keV}$ approximately
$\frac{1}{3}$ of the tritium spectrum is measured. Using this reduction
factor, an electron flux density of
${10}^{6}\text{\,}{\mathrm{s}}^{-1}\text{\,}{\mathrm{mm}}^{-2}$ and a detector
efficiency of $\epsilon=$ $65\text{\,}\mathrm{\char 37\relax}$ the measurement
time needed to reach the required $0.1\text{\,}\mathrm{\char 37\relax}$
precision for each of these p-i-n diodes is calculated and listed in table 1.
The one unknown property of these p-i-n diodes is their individual dead layer.
During manufacturing the thickness of the dead layer is not measured and
therefore not available a priori, but limited to $1000\text{\,}\mathrm{nm}$.
The thickness of the dead layer is indicated by the minimum energy that can be
detected. The measurement of the dead layer is done by analysing the shape of
the peak from monoenergetic electrons originating from an electron gun (see
section 5.2). Figure 10 illustrates such an analysis. Measurements of the dead
layer are performed for each p-i-n diode before they are mounted on the FBM
detector board. It is assumed that the dead layer remains constant over time,
even after bakeout cycles of the vacuum setup. This is because the dead layer
is silicon oxide which is not affected by heat and requires approximately
${10}^{13}$ electrons (on the order of several years in the FBM location) to
suffer from radiation damage.
### 4.4 Data acquisition
For the two p-i-n diode detector channels an Amptek [27] PX5 and an Amptek DP5
are used for the data readout. These are digital pulse processors with build-
in amplifiers used to amplify the signal by up to a factor of $100$. These
Amptek devices are connected to an Apple (Mac mini) computer running the
Object-orientated Real-time Control and Acquisition (ORCA) software [22]. An
ORCA readout module was specifically designed for the FBM Amptek devices. The
raw ORCA data is converted into ROOT [29] files for analysis. The preamplifier
outputs of the two p-i-n diode detector channels can also be connected to
separate low-pass filters to measure the DC offset occurring from the event
rate on the respective p-i-n diode chip.
The pulse processing parameters of each detector channel can be optimised to
obtain either the count rate or a spectrum of the $\upbeta$-electrons from the
source. The peaking time is set to
* •
Fast channel: $1.0\text{\,}\mathrm{\SIUnitSymbolMicro s}$ to measure the count
rate (larger p-i-n diode with higher count rate)
* •
Slow channel: $3.2\text{\,}\mathrm{\SIUnitSymbolMicro s}$ to measure the
spectrum (smaller p-i-n diode with lower count rate)
During scanning the required measurement time at each point is reduced due to
the increased electron flux towards the centre of the beam tube. The analysis
of the FBM data is based on the established analysis systems of the KATRIN
experiment. Therefore, all data, slow control, and run files are available on
the ADEI server and KATRIN databases.
## 5 Measurements
This section presents selected results [18] of the measurements performed with
the FBM during its commissioning phases as well as during the first KATRIN
measurement campaigns. These results serve as an evaluation tool for the
positioning accuracy of the vacuum manipulator and the performance of the
detector. In some cases the data is compared to the results of numerical
simulations of the detector response.
### 5.1 Alignment and positioning precision
Figure 8: Left: Data and fit result of the calibrated and temperature
corrected $z$-component of the magnetic field in the CPS. Right: The residuals
of the simulated magnetic field which shows a good agreement with the data.
Positioning reproducibility is the ability of the FBM to find a position
relative to a former position. This is different to the absolute positioning
accuracy which includes external reference points with respect to the KATRIN
coordinate system. The reproducibility is validated by using a laser setup as
well as a portable Coordinate Measuring Machine (CMM). It was determined to be
better than $0.1\text{\,}\mathrm{mm}$. However, the overall alignment
uncertainties (also CMM) dominate the absolute positioning accuracy as shown
in table 2.
| $\boldsymbol{\sigma_{\text{alignment}}}$ [$\mathrm{mm}$] | $\boldsymbol{\tilde{\sigma}_{\text{max}}}$ [$\mathrm{mm}$] | $\boldsymbol{\sigma_{\text{full}}}$ [$\mathrm{mm}$] | Offset [$\mathrm{mm}$]
---|---|---|---|---
$x$ | $0.28$ | $0.042$ | $0.28$ | $-1.2\pm 0.2$
$y$ | $0.1$ | $0.07$ | $0.13$ | $4.2\pm 0.2$
Table 2: The overall positioning accuracy $\sigma_{\text{full}}$ results from
the combination of the uncertainties of the alignment
$\sigma_{\text{alignment}}$ and the positioning reproducibility
$\tilde{\sigma}_{\text{max}}$. There is a misalignment between the FBM and the
flux tube expressed by a constant offset which was determined from flux tube
scans.
To calibrate the movement system, as well as to find the center of the flux
tube, the magnetic field in the CPS can be used (see left panel in figure 8).
The shape of the magnetic flux can be described by a two-dimensional Gaussian.
The required calibration values, namely the encoder value for the horizontal
lever arm and the offset of the magnetic flux center to the FBM system (listed
in the last row in table 2), are given by the free parameters in a fit of data
taken during a flux tube scan.
To demonstrate the excellent positioning accuracy of the manipulator a thin
($0.14\text{\,}\mathrm{mm}$ diameter) electron beam was scanned with the FBM
by moving the detector (type S5971 with $1.2\text{\,}\mathrm{mm}$ diameter)
through the fixed beam in a grid pattern with $0.1\text{\,}\mathrm{mm}$
spacing [18]. Since the beam is far smaller than the p-i-n diode, it is rather
the diode being scanned by the beam than vice versa. The plot in figure 9
shows the measured intensities as a function of detector position. The
background is removed by a noise cut at
$100\text{\,}\mathrm{c}\mathrm{p}\mathrm{s}$ and as a result only positions
where electrons were detected are shown. The large circular contours represent
the entrance window of the diode (small, $1.2\text{\,}\mathrm{mm}$ diameter)
as it is stated in the data sheets and the visual surface (large,
$1.3\text{\,}\mathrm{mm}$) of the diode as it was measured. The position of
the contours is adjusted such that the number of events within the contours is
maximised. The center represents the actual position of the beam at
$x_{\text{FBM}}=$ $-1.2\text{\,}\mathrm{mm}$ and $y_{\text{FBM}}=$
$7.6\text{\,}\mathrm{mm}$. One spot close to the center has low statistics,
which is associated with a small dirt particle which was found to be located
at the equivalent position of the p-i-n diode chips’s active surface.
### 5.2 Detector response and dead layer
For calibration KATRIN is equipped with an electron gun which is situated in
the rear section and can provide a mono-energetic electron beam with energies
up to $20\text{\,}\mathrm{keV}$. In the left panel of figure 10 the measured
detector response to $18.6\text{\,}\mathrm{keV}$ electrons is shown. Due to
partial charge collection the peak is shifted to lower energies, widens and
develops a long low energy tail descending into an almost flat plateau. To
understand the related effects and to reach the required precision for the
FBM, numerical simulations [18] were performed using Geant4 [23]. The model of
the detector comprises a detection volume and a fully insensitive dead layer
(also known as slab-model). Within the detector the electrons undergo elastic
and inelastic scattering losing their energy subsequently and are finally
either completely stopped in the detector or reflected from it. For the
interaction model the Penelope physics package was used. The low energy region
A (see figure 10) is predominantly produced by reflected electrons and region
B mainly by the dead layer losses which also cause the shift of the main peak
in region C by approximately $0.6\text{\,}\mathrm{keV}$ (mean energy loss in
the dead layer).
The measured electron spectra were compared to a library of simulated electron
spectra for different dead layers, with a $\chi^{2}$ test used to find the
most accurate match. The best result was obtained with a dead layer thickness
of $340\text{\,}\mathrm{nm}$. The simulations always overestimated the data in
the low energy tail at approximately $5\text{\,}\mathrm{keV}$ (region A). This
is caused by an incomplete model which does not include the magnetic field
configuration in the CPS for the purpose of simplicity. Due to magnetic
mirroring in the CPS electrons which were initially reflected from the
detector are guided back within the peaking time of the DAQ contributing to
regions B and C instead of A.
It was possible to determine the dead layers of the FBM p-i-n diodes which
range from $300\text{\,}\mathrm{nm}500\text{\,}\mathrm{nm}$ causing an energy
dependent shift of the measured peak of
$0.5\text{\,}\mathrm{keV}2\text{\,}\mathrm{keV}$ for electron energies up to
$20\text{\,}\mathrm{keV}$. With these simulations the detection efficiency for
electrons as a function of kinetic energy could also be determined as shown in
the right plot in figure 10.
Figure 9: Scan of stationary electron gun beam. The FBM detector board is
moved through the beam in a grid with $0.1\text{\,}\mathrm{mm}$ step length.
Each colored dot represents the measured rate at this detector position. The
background is removed by a noise cut at
$100\text{\,}\mathrm{c}\mathrm{p}\mathrm{s}$. Note that the size of the data
points is arbitrary and does not represent the size of the beam spot. The
inner circle represents the active surface (here
$1.13\text{\,}{\mathrm{mm}}^{2}$) as stated by the manufacturer and the outer
circle the visual surface of the p-i-n diode. The detector chip profiles are
positioned such that they comprise the highest rate.
Figure 10: Left: Measured and simulated electron gun peaks obtained during the
first tritium campaign. The simulation includes a detector energy resolution
with FWHM $=$ $2.35\text{\,}\mathrm{keV}$. The best match was obtained with a
dead layer thickness of $340\text{\,}\mathrm{nm}$. Right: The simulated
efficiencies for electrons not to get reflected from the detector
($\epsilon_{r}$), not to get stopped in the dead layer ($\epsilon_{d}$), and
for exceeding the energy threshold of $5\text{\,}\mathrm{keV}$
($\epsilon_{\text{th}}$ ($5\text{\,}\mathrm{keV}$)). The intrinsic efficiency
of the detector is then given by $\varepsilon_{i}$($5\text{\,}\mathrm{keV}$)
$=\varepsilon_{r}\cdot\varepsilon_{d}\cdot\varepsilon_{\text{th}}$($5\text{\,}\mathrm{keV}$).
### 5.3 First tritium measurement campaign
Figure 11: Data and fit of the ${}^{241}\text{Am}$ spectrum. The spectrum is
recorded with the FBM with an energy resolution of
$\sigma_{\text{FWHM}}\approx$ $2\text{\,}\mathrm{keV}$. The positions of all
${}^{241}\text{Am}$ lines are identified and labeled. The full fit function
consists of a combination of Gaussians and error functions for each (strong)
${}^{241}\text{Am}$ line in the peak. The global fit comprises $33$ free fit
parameters. Only the Gaussian parts representing the strongest lines are used
for the calibration of the FBM.
Before the actual tritium measurement an alternative front end, equipped with
a Faraday cup, was installed to the FBM in order to check ion blocking,
measure the radial ion distribution in the beamline, and check the simulated
source gas models by measuring secondary electrons [24]. The measurements with
the p-i-n diode detector started with the “first tritium measurement campaign“
[25] which took place from the 5 to the 20 of May 2018 with a gas mixture of
$0.5\text{\,}\mathrm{\char 37\relax}$ tritium in deuterium. In the following
sections the results of this first data-taking period with tritium are
presented.
#### 5.3.1 Configuration
With a fraction of only $0.5\text{\,}\mathrm{\char 37\relax}$ of tritium in
the source gas an electron flux of approximately
$5000\text{\,}{\mathrm{s}}^{-1}\text{\,}{\mathrm{mm}}^{-2}$ was expected at
the FBM measuring plane. Therefore, the largest p-i-n diodes have been chosen
($1.1\text{\,}{\mathrm{mm}}^{2}$) to optimise counting statistics. The peaking
time of the DAQ for both channels was $6.4\text{\,}\mathrm{\SIUnitSymbolMicro
s}$, resulting in a pile-up rate of about $3\text{\,}\mathrm{\char 37\relax}$
which can be neglected for stability analyses (see section 5.3.4).
Acceptance tests were performed prior to the campaign to extract calibration
parameters, energy resolutions, and noise thresholds of the detectors. These
measurements were performed with an 241Am source in the vented system with the
FBM in parking position. The source was placed at a close distance between the
two p-i-n diodes. The desired diode could then be irradiated using the
movement mechanics and be adjusted to find the maximum count rate. In figure
11 one of the 241Am spectra extracted from these measurements is shown. The
calibration parameters are obtained by a global fit to the whole spectrum.
#### 5.3.2 Spectrum
The spectrum shown in figure 12 is the first tritium spectrum recorded with
the FBM. Between $6\text{\,}\mathrm{keV}20\text{\,}\mathrm{keV}$ the spectrum
agrees with the expectation, however below $6\text{\,}\mathrm{keV}$ the slope
is unexpectedly increasing. This is probably due to background counts from
noise and edge effects from the diodes. This may also explain why the spectra
of the two channels do not match perfectly for lower energies. Other likely
sources for this mismatch, which is also the reason for about
$2\text{\,}\mathrm{\char 37\relax}$ lower rate in channel $1$ than in channel
$2$ during the whole campaign, are
* •
uncertainties in the energy calibrations which cause the deviations among the
channels for lower energies,
* •
small differences in the active area, or
* •
small differences in the dead layer thickness of the two p-i-n diodes.
Figure 12: Tritium $\upbeta$-spectrum measured with both detector channels of
the FBM during the first tritium campaign. The rate deviation of approximately
$2\text{\,}\mathrm{\char 37\relax}$ between the two channels is probably
caused by the uncertainties in the calibrations or differences in the active
surface or dead layer thickness of the p-i-n diodes.
#### 5.3.3 Flux tube scans
Several scans of the $\upbeta$-electron flux cross section were performed
recording the tritium count rate, the magnetic field, and the temperature.
During a scan, the temperature usually drops by about
$1\text{\,}\mathrm{\SIUnitSymbolCelsius}$. This occurs when the detector is
moved further into the cold CPS where the detector directly faces the
$4\text{\,}\mathrm{K}$ cold beam tube of the CPS in which the argon frost
layer is prepared. To check the calibration of the magnetic field sensor as
well as to verify our understanding of the magnetic field configuration in the
FBM measuring plane the magnetic field data is compared with simulations. The
residual analysis displayed in figure 8 (right) shows a good agreement between
simulation and data.
Figure 13: Radial dependence of the count rate derived from a cross scan
during the first tritium campaign with channel $1$ of the FBM. Top: 1D
Gaussians are fit to the data for each horizontal ($\text{X}_{1,2}$) and
vertical ($\text{Y}_{1,2}$) scan. The Gaussian means are compatible with the
results from magnetic field measurements. One can clearly see that for
identical positions slightly different rates are measured, for example the
rate increased during the $x$-scans such that the mean of the $\text{X}_{2}$
fit is lower than for $\text{X}_{1}$. The Gaussian widths are approximately
$\sigma=$ $165\text{\,}\mathrm{mm}$. Bottom: 2D scatter plot of the same data.
The scans for $y$ are not perfectly on a vertical line due to the chosen scan
pattern which explains the larger uncertainties in the fits.
In figure 13 (top and bottom) the results of a cross scans over the cross
section of the flux tube for both detector channels are shown. The electron
flux shows the expected Gaussian shape where the rate drops from the center to
the outer rim by approximately $10\text{\,}\mathrm{\char 37\relax}$ as
predicted by simulations [26]. It can be seen that the event rate for
identical positions changes during the scans which affects the extracted mean
of the fits. Nevertheless, the means are compatible to the results from the
alignment measurements in section 5.1 which use the magnetic field data. This
is expected as the electron flux scales with the magnetic flux.
#### 5.3.4 Rate stability
Figure 14: $\upbeta$-electron rate trend summary of the first tritium
campaign. The full available data from the stability measurements at the
monitoring position for both channels is plotted. The count rates for channel
1 are approximately $0.7\text{\,}\mathrm{\char 37\relax}$ smaller compared to
channel 2 using the same energy threshold (here $5.3\text{\,}\mathrm{keV}$).
Apart from that, the channels follow the same trend. The full linear fit
reveals a mean relative increase of about $0.02\text{\,}\mathrm{\char
37\relax}$ per hour while for the single regions this value is smaller than
$0.01\text{\,}\mathrm{\char 37\relax}$ per hour.
During the two weeks of the first tritium campaign the FBM was mainly
monitoring the flux in the CPS at position $x_{\text{FBM}}=$
$65\text{\,}\mathrm{mm}$ (outer rim of the flux tube, see figure 3 and 13).
From time to time background measurements were taken slightly out of the beam
at $x_{\text{FBM}}=$ $80\text{\,}\mathrm{mm}$. The full rate trend graphs are
shown in figure 14 for both detector channels including linear fits to the
data. The entire monitoring time is separated into six time regions. There is
a long term drift of approximately $0.02\text{\,}\mathrm{\char
37\relax}\mathrm{/}\mathrm{h}$ determined from all regions, while for single
regions the drift is generally smaller, especially for the longer regions $2$,
$3$, $5$, and $6$, hence the reason for the larger long term drift must mainly
originate from incidences which occur between the regions. Several
investigations have been performed to find the source of this long-term drift,
and there are hints that the detector response changes over time due to an
increase in the noise level and degrading effects of the detector chip. Hence
this drift is probably caused by the FBM and not by a change of the incoming
electron flux. The latter assumption is supported by the results of the other
monitoring systems which do not observe such a drift. However, this long-term
drift is sufficiently small as the FBM is designed to monitor relative source
fluctuations over short time intervals, such as seconds, minutes and at
maximum a few hours. Within these time ranges the drift is within the required
sensitivity of $0.1\text{\,}\mathrm{\char 37\relax}$. Therefore, despite the
observed long term drift, the FBM shows a stability fulfilling its design
goal.
In the first tritium campaign of KATRIN the FBM was utilised to reduce
systematic uncertainties in the tritium concentration $\epsilon_{\text{T}}$
measurement performed by the LARA system. In this campaign the tritium amount
was limited to about $1\text{\,}\mathrm{\char 37\relax}$ in deuterium, and
consequentially statistical fluctuations in the determination of the
concentration of the tritiated hydrogen isotopolouges were much stronger than
it is the case for standard operation ($\epsilon_{\text{T}}>$ $0.95$). An
average tritium concentration was determined over a long duration ($\approx
3\text{\,}\mathrm{h}$) with LARA, and this average tritium concentration was
fluctuated according to the higher statistics FBM data for short durations.
Note that these are time-scales over which the drift of the FBM is negligible.
This way the uncertainty of the short-term fluctuations measurement of the
tritium concentration were reduced from about $2\text{\,}\mathrm{\char
37\relax}$ down to about $0.5\text{\,}\mathrm{\char 37\relax}$. The
cooperation of these two monitoring systems was crucial to reduce the tritium
concentration systematic input for an upcoming $\mathrm{keV}$-scale sterile
neutrino analysis of the first tritium data.
## 6 Summary
The KATRIN experiment aims for a precise measurement of the electron
antineutrino mass with a sensitivity of $0.2\text{\,}\mathrm{eV}$
($90\text{\,}\mathrm{\char 37\relax}$ CL). One of the systematic uncertainties
in this measurement arises from fluctuations of the column density of high
luminosity tritium source. In order to reach the design goal of KATRIN, the
latter must be measured on the per-mille level over time scales of a few
minutes. Therefore the source is continuously monitored by several monitoring
systems, one of which is the Forward Beam Monitor (FBM). The FBM has the
advantage of being capable of continuously monitoring variations of the
electron flux and changes in the observed shape of the $\upbeta$-decay
spectrum with high accuracy on short time scales.
A UHV compatible vacuum manipulator was commissioned. It is able to place a
detector board directly into the beta-electron flux originating from the
tritium source. Although the mounting position of the apparatus demands a
movement mechanism with a working stroke of $1.8\text{\,}\mathrm{m}$ the FBM
is able to reach any position within the electron flux cross-section with a
precision of better than $0.3\text{\,}\mathrm{mm}$ which can be determined
with magnetic field measurements.
The detector board at the tip of the FBM manipulator measures the electron
flux with two silicon p-i-n diodes. The FBM reaches an energy resolution of
about $\sigma_{\text{FWHM}}=$ $2\text{\,}\mathrm{keV}$ at an energy threshold
of $5\text{\,}\mathrm{keV}$. The readout electronics are optimised to register
electron events at a rate of
$\mathcal{O}$(${10}^{4}\text{\,}\mathrm{c}\mathrm{p}\mathrm{s}$) and thus to
measure relative changes in the electron flux with $0.1\text{\,}\mathrm{\char
37\relax}$ precision in about $100\text{\,}\mathrm{s}$.
The entrance window (dead layer) of the p-i-n diodes has a large impact on the
detector response when measuring electrons. It was found that the dead layer
thickness of the p-i-n diodes used for the FBM range from
$300\text{\,}\mathrm{nm}500\text{\,}\mathrm{nm}$.
After commissioning, the FBM was employed for several KATRIN measurement
campaigns. The capabilities of the FBM detector were confirmed as well as the
positioning accuracy of the manipulator. A small long term (days to weeks)
drift of the rate was observed which correlates to a drift of the noise level
of the electronics. On short time scales (hours) the FBM is stable to the per-
mille level. With this the FBM is a monitoring device which reaches all its
design goals.
With its good performance the FBM data already played a key role in reducing
the systematic uncertainties of the tritium concentration
$\epsilon_{\text{T}}$ fluctuations during the first tritium campaign. This was
achieved by combining it with the LARA data which featured a relative
statistical uncertainty of only a few percent on time scales of minutes due to
the low amount of source gas molecules. This will be important for an upcoming
$\mathrm{keV}$-scale sterile neutrino analysis of the first tritium data.
## Acknowledgments
We acknowledge the support of the Ministry of Education and Research BMBF
(05A14PX3, 05A17PX3) and the Helmholtz Association.
## References
* [1] A. Osipowicz et al. KATRIN: A Next generation tritium beta decay experiment with sub-eV sensitivity for the electron neutrino mass. Letter of intent. 9 2001.
* [2] J. Angrik et al. KATRIN design report 2004. 2005\.
* [3] M. Aker, K. Altenmüller, M. Arenz, M. Babutzka, J. Barrett, S. Bauer et al. Improved upper limit on the neutrino mass from a direct kinematic method by KATRIN. Phys. Rev. Lett., 123:221802, Nov 2019.
* [4] M. Arenz, M. Babutzka, M. Bahr, J. Barrett, S. Bauer, M. Beck et al. Commissioning of the vacuum system of the KATRIN main spectrometer. Journal of Instrumentation, 11(04):P04011–P04011, apr 2016.
* [5] M. Babutzka et al. Monitoring of the operating parameters of the KATRIN Windowless Gaseous Tritium Source. New J. Phys., 14:103046, 2012.
* [6] A. Picard et al. A solenoid retarding spectrometer with high resolution and transmission for kev electrons. Nuclear Instruments and Methods in Physics Research Section B: Beam Interactions with Materials and Atoms, 63(3):345 – 358, 1992.
* [7] M. Arenz, W.-J. Baek, M. Beck, A. Beglarian, J. Behrens, T. Bergmann et al. First transmission of electrons and ions through the KATRIN beamline. Journal of Instrumentation, 13(04):P04020–P04020, apr 2018.
* [8] Max Aker, K. Altenmüller, M. Arenz, W.-J. Baek, J. Barrett, A. Beglarian et al. First operation of the KATRIN experiment with tritium. The European Physical Journal C, 80(3):264, 2020.
* [9] Lutz Bornschein, Beate Bornschein, Michael Sturm, Marco Röllig, and Florian Priester for the KATRIN collaboration. Status of the neutrino mass experiment KATRIN. Fusion Science and Technology, 67(2):274–277, 2015.
* [10] S. Grohmann. Stability analyses of the beam tube cooling system in the KATRIN source cryostat. Cryogenics, 49(8):413 – 420, 2009.
* [11] S. Grohmann, T. Bode, M. Hötzel, H. Schön, M. Süßer, and T. Wahl. The thermal behaviour of the tritium source in KATRIN. Cryogenics, 55-56:5 – 11, 2013.
* [12] M. Sturm, M. Schlösser, R. J. Lewis, B. Bornschein, G. Drexlin, and H. H. Telle. Monitoring of all hydrogen isotopologues at tritium laboratory Karlsruhe using Raman spectroscopy. Laser Physics, 20(2):493–507, 2010.
* [13] Magnus Schlösser et al. Accurate calibration of the laser Raman system for the Karlsruhe Tritium Neutrino Experiment. Journal of Molecular Structure, 1044:61 – 66, 2013. MOLECULAR SPECTROSCOPY AND MOLECULAR STRUCTURE 2012.
* [14] A. Marsteller et al. Neutral tritium gas reduction in the KATRIN differential pumping sections. Vacuum, 184:109979, 2021.
* [15] W. Gil, J. Bonn, B. Bornschein, R. Gehring, O. Kazachenko, J. Kleinfeller, and S. Putselyk for the KATRIN collaboration. The Cryogenic Pumping Section of the KATRIN experiment. IEEE Transactions on Applied Superconductivity, 20(3):316–319, 2010\.
* [16] Carsten Röttele. Results of the first Cool-down of the KATRIN Cryogenic Pumping Section. Journal of Physics: Conference Series, 888:012228, sep 2017.
* [17] Markus Hötzel. Simulation and analysis of source-related effects for KATRIN. Ph.D. thesis, KIT, Karlsruhe, 2012.
* [18] Enrico Ellinger. Development and investigation of the Forward Beam Monitor for the KATRIN experiment. Ph.D. thesis, Bergische Universität Wuppertal (BUW), 2019. urn:nbn:de:hbz:468-20191127-123110-5.
* [19] Udo Schmitt. Entwicklung eines Monitordetektors für das KATRIN-Experiment. PhD thesis, KIT, Karlsruhe, EKP, 2008.
* [20] Martin Babutzka. Untersuchung eines verfahrbaren Monitordetektors zur Überwachung der Aktivität des $\beta$-Zerfalls in der kryogenen Pumpstrecke des KATRIN-Experiments. Diploma thesis, Karlsruher Institut für Technologie (KIT), 2010.
* [21] Marco Kleesiek. A Data-Analysis and Sensitivity-Optimization Framework for the KATRIN Experiment. Ph.D. thesis, Karlsruhe U., EKP, 2014.
* [22] M.A. Howe, G.A. Cox, P.J. Harvey, F. McGirt, K. Rielage, J.F. Wilkerson, and J.M. Wouters. Sudbury Neutrino Observatory neutral current detector acquisition software overview. IEEE Transactions on Nuclear Science, 51(3):878–883, 2004.
* [23] Agostinelli, S. et al. Geant4 - a simulation toolkit. Nuclear Instruments and Methods in Physics Research Section A: Accelerators, Spectrometers, Detectors and Associated Equipment, 506(3):250–303, 2003.
* [24] M. Klein. Tritium ions in KATRIN: blocking, removal and detection. PhD thesis, Karlsruher Institut für Technologie (KIT), 2018. 10.5445/IR/1000093526.
* [25] Max Aker et al. First operation of the KATRIN experiment with tritium. The European Physical Journal C, 80(3):264, 2020.
* [26] Norman Haußmann. Simulation and measurement of the Forward Beam Monitor detector signal for the KATRIN experiment. Ph.D. thesis, Bergische Universität Wuppertal (BUW), 2019. urn:nbn:de:hbz:468-20191009-085446-9.
* [27] Amptek, Inc. PX5 DPP and DP5 DPP www.amptek.com.
* [28] Hamamatsu Photonics K.K. Collection of PIN photodiodes www.hamamatsu.com.
* [29] Rene Brun and Fons Rademakers. ROOT - An Object Oriented Data Analysis Framework. Proceedings AIHENP’96 Workshop, Lausanne, Sep. 1996, Nuclear Instruments and Methods in Physics Research A 389 (1997) 81-86. See also https://root.cern/
|
# Some cluster tilting modules for weighted surface algebras
Karin Erdmann Mathematical Institute, Oxford University, ROQ, Oxford OX2 6GG,
United Kingdom<EMAIL_ADDRESS>
###### Abstract.
Non-singular weighted surface algebras satisfy the necessary condition found
in [6] for existence of cluster tilting modules. We show that any such algebra
whose Gabriel quiver is bipartite, has a module satisfying the necessary ext
vanishing condition. We show that it is 3-cluster tilting precisely for non-
singular triangular or spherical algebras, but not for any other weighted
surface algebra with bipartite Gabriel quiver.
Keywords: Symmetric algebra, Surface algebra, cluster tilting modules.
2020 MSC: 16D50, 16E30, 16G20, 16G60
###### 2020 Mathematics Subject Classification:
16D50, 16E30, 16G20, 16G60
## 1\. Introduction
A module $M$ of a finite-dimensional algebra $A$ is an $n$-cluster tilting
module (or maximal $(n-1)$-orthogonal) provided
$\displaystyle{\rm add}(M)=$ $\displaystyle\\{N\mid{\rm Ext}^{i}(M,N)=0\mbox{
for }1\leq i\leq n-1\\}$ $\displaystyle=$ $\displaystyle\\{N\mid{\rm
Ext}^{i}(N,M)=0\mbox{ for }1\leq i\leq n-1\\},$
(see [12], [13]). We would like to know whether non-singular weighted surface
algebras have cluster tilting modules. Weighted surface algebras are a class
of tame symmetric algebras, periodic as bimodules, of period $4$ (see [7] and
[8], [9]). This means that they satisfy the necessary condition found in [6],
requiring that all non-projective modules should have bounded periodic
resolutions. As observed in [6] if such an algebra has an $n$-cluster tilting
module then the only option is $n=3$.
Here we study weighted surface algebras which have a bipartite Gabriel quiver,
which means that in the presentation as in [8] (see also [9]) it has many
virtual arrows. We introduce a module $M$, defined in the same way for each of
the algebras, which satisfies ${\rm Ext}^{1}(M,M)=0$ and ${\rm
Ext}^{2}(M,M)=0$. We show that it is 3-cluster tilting when $\Lambda$ is
either a triangle algebra $T(\lambda)$, or a spherical algebra $S(\lambda)$
(see §3 for the definition) with $\lambda\in K$ and $\lambda\neq 0,1$. We also
show that for any other weighted surface algebra whose Gabriel quiver is
bipartite, $M$ cannot be a direct summand of a 3-cluster tilting module.
The algebra $T(\lambda)$ occurs in various places in the literature. It is an
algebra with $k=1$ in the family $Q(3\mathcal{A})_{1}$ of algebras of
quaternion type, in [5]. Furthermore, it occurs with the name
$B_{1,1}(\lambda)$ in [3]. As well, it occurs in [1] with the name
$A_{1}(\lambda)$. In [8] it is called the triangle algebra $T(\lambda)$ (in
Example 3.4). Similarly the spherical algebra $S(\lambda)$ was introduced in
[8] (Example 3.6).
Spherical algebras are a special case of the family of algebras which come
from the triangulation $T(n)$ of the sphere as defined in Example 7.5 of [10].
We call these algebras $n$-spherical; when $n=2$ they are the same as the
spherical algebras. One may also observe that the spherical algebra with $n=1$
with the multiplicities $2,2,1$ is the same as the triangle algebra (see [8],
Example 3.3).
Algebras whose Gabriel quiver is the same as that of $T(\lambda)$, allowing a
multiplicity $k>1$, coincide (up to a scalar parameter) with the algebras
$Q(3\mathcal{A})_{2}^{k}$ in the labelling of [11]. When the characteristic of
$K$ is $2$ these occur as the basic algebras of blocks of finite groups. Very
recently B. Böhmler and R. Marcinczik proved using computer calculations that
for $k=2$, it has a 3-cluster tilting module (see [2]).
Much of this note was written five years ago, when talking to Idun Reiten
about [3], and it was extended first when spherical algebras had been
discovered, and then again, inspired by an email from R. Marcinczik (for which
I am grateful).
## 2\. Preliminaries
Throughout $K$ is an algebraically closed field, of arbitrary characteristic.
Assume $\Lambda$ is a finite-dimensional symmetric $K$-algebra. We recall some
identities for the stable category $\underline{\rm mod}\Lambda$.
(1) $D{\rm Ext}^{1}(M,N)\cong\underline{\rm Hom}(\tau^{-1}N,M)$, and in this
case $\tau\cong\Omega^{2}$.
(2) ${\rm Ext}^{i}(U,V)\cong\underline{\rm
Hom}(\Omega^{i}U,V)\cong\underline{\rm Hom}(U,\Omega^{-i}V)$.
This implies that $\dim{\rm Ext}^{i}(M,N)=\dim{\rm Ext}^{1}(N,\Omega^{i+2}M)$.
The algebras we consider have the property that all non-projective
indecomposable right $\Lambda$-modules are $\Omega$ periodic of periods
dividing $4$. This gives us the following, we refer to this as ext symmetry.
###### Corollary 2.1.
Assume $\Lambda$ is symmetric and all modules have $\Omega$-period dividing
$4$. Then for all $M,N$ we have $\dim{\rm Ext}^{2}(M,N)=\dim{\rm
Ext}^{1}(N,M)$ as vector spaces.
This simplifies the search for 3-cluster tilting modules. If we know that
${\rm Ext}^{1}(N,X)=0$ and ${\rm Ext}^{1}(X,N)=0$ then automatically ${\rm
Ext}^{2}(X,N)=0$ and ${\rm Ext}^{2}(N,X)=0$.
## 3\. The algebras
### 3.1. Weighted surface algebras
We review the definition from [9], for details see [7], [8], [9].
Assume $Q$ is a finite quiver. Denote by $KQ$ the path algebra of $Q$ over
$K$. We will consider algebras of the form $A=KQ/I$ where $I$ is an ideal of
$KQ$ which contains all paths of length $\geq m$ for some $m>>0$, so that the
algebra is finite-dimensional and basic. The Gabriel quiver $Q_{A}$ of $A$ is
then the full subquiver of $Q$ obtained from $Q$ by removing all arrows
$\alpha$ with $\alpha+I\in R_{Q}^{2}+I$.
A quiver $Q$ is _$2$ -regular_ if for each vertex $i\in Q_{0}$ there are
precisely two arrows starting at $i$ and two arrows ending at $i$. Such a
quiver has an involution on the arrows, $\alpha\mapsto\bar{\alpha}$, such that
for each arrow $\alpha$, the arrow $\bar{\alpha}$ is the arrow $\neq\alpha$
such that $s(\alpha)=s(\bar{\alpha})$.
A _triangulation quiver_ is a pair $(Q,f)$ where $Q$ is a (finite) connected
2-regular quiver, with at least two vertices, and where $f$ is a fixed
permutation of the arrows such that $t(\alpha)=s(f(\alpha))$ for each arrow
$\alpha$, and such that $f^{3}$ is the identity. The permutation $f$ uniquely
determines a permutation $g$ of the arrows, defined by
$g(\alpha):=\overline{f(\alpha)}$ for any arrow $\alpha$. We assume throughout
that $(Q,f)$ is a triangulation quiver. To give the presentations of the
algebras in question, we use the following notation. For each arrow $\alpha$,
we fix
$\displaystyle m_{\alpha}\in\mathbb{N}^{*}$ a weight, constant on $g$-cycles,
and $\displaystyle c_{\alpha}\in K^{*}$ a parameter, constant on $g$-cycles,
and define $\displaystyle n_{\alpha}:=$ the length of the $g$-cycle of
$\alpha$, $\displaystyle B_{\alpha}:=\alpha g(\alpha)\ldots
g^{m_{\alpha}n_{\alpha}-1}(\alpha)$ $\displaystyle\mbox{ the path along the
$g$-cycle of $\alpha$ of length $m_{\alpha}n_{\alpha}$},$ $\displaystyle
A_{\alpha}:=\alpha g(\alpha)\ldots g^{m_{\alpha}n_{\alpha}-2}(\alpha)$ the
path along the $g$-cycle of $\alpha$ of length $m_{\alpha}n_{\alpha}-1$.
###### Definition 3.1.
We say that an arrow $\alpha$ of $Q$ is virtual if $m_{\alpha}n_{\alpha}=2$,
that is $A_{\alpha}$ has length $1$. Note that this condition is preserved
under the permutation $g$, and that virtual arrows form $g$-orbits of sizes 1
or 2.
We assume that the following conditions hold.
(1) $m_{\alpha}n_{\alpha}\geq 2$ for all arrows $\alpha$, and
(2) $m_{\alpha}n_{\alpha}\geq 3$ for all arrows $\alpha$ such that
$\bar{\alpha}$ is virtual and $\bar{\alpha}$ is not a loop, and
$m_{\alpha}n_{\alpha}\geq 4$ for all arrows $\alpha$ such that $\bar{\alpha}$
is virtual and $\bar{\alpha}$ is a loop.
Condition (1) is a general assumption, and (2) is needed to eliminate two
small algebras (see [8]). We also assume that $Q$ has at least three vertices.
With this, the definition of a weighted surface algebra (as revised in [9]) is
as follows.
###### Definition 3.2.
The algebra $\Lambda=\Lambda(Q,f,m_{\bullet},c_{\bullet})=KQ/I$ is a weighted
surface algebra if $(Q,f)$ is a triangulation quiver, with $|Q_{0}|\geq 2$,
and $I=I(Q,f,m_{\bullet},c_{\bullet})$ is the ideal of $KQ$ generated by:
1. (1)
$\alpha f(\alpha)-c_{\bar{\alpha}}A_{\bar{\alpha}}$ for all arrows $\alpha$ of
$Q$,
2. (2)
$\alpha f(\alpha)g(f(\alpha))$ for all arrows $\alpha$ of $Q$ unless
$f^{2}(\alpha)$ is virtual, or unless $f(\bar{\alpha})$ is virtual and
$m_{\bar{\alpha}}=1,\ n_{\bar{\alpha}}=3$.
3. (3)
$\alpha g(\alpha)f(g(\alpha))$ for all arrows $\alpha$ of $Q$ unless
$f(\alpha)$ is virtual, or unless $f^{2}(\alpha)$ is virtual and
$m_{f(\alpha)}=1,\ n_{f(\alpha)}=3$.
The Gabriel quiver $Q_{\Lambda}$ is the subquiver of $Q$ obtained by removing
all virtual arrows.
We recall a few properties.
(1) Any such algebra is symmetric and tame.
(2) The dimension of $e_{i}\Lambda$ is equal to
$m_{\alpha}n_{\alpha}+m_{\bar{\alpha}}n_{\bar{\alpha}}$ where
$\alpha,\bar{\alpha}$ are the arrows starting at $i$.
The relations also imply that
$c_{\alpha}B_{\alpha}=c_{\bar{\alpha}}B_{\bar{\alpha}}$ in $\Lambda$. One can
show that this spans the socle of $e_{i}\Lambda$.
We wish to define a module $M$ such that ${\rm Ext}^{1}(M,M)=0$ and ${\rm
Ext}^{2}(M,M)=0$, as a candidate to be 3-cluster tilting. This can be done for
a weighted surface algebra whose quiver is bipartite; this requires that each
triangle of $f$ must contain a virtual arrow. Such a quiver can be thought of
made up of three building blocks, first a quiver of the form
$a_{1}$$b_{1}$$d_{1}$$a_{2}$$b_{2}$$d_{2}$$a_{3}$$\cdots$$a_{n}$$b_{n}$$d_{n}$$a_{n+1}\\!\\!\\!\\!\\!\\!\\!\\!\\!\\!\\!\\!\\!\\!\\!\\!\\!\\!\\!\\!$$\xi_{1}$$\delta_{1}$$\gamma_{1}$$\eta_{1}$$\sigma_{1}$$\varrho_{1}$$\xi_{2}$$\delta_{2}$$\gamma_{2}$$\eta_{2}$$\sigma_{2}$$\varrho_{2}$$\xi_{n}$$\delta_{n}$$\gamma_{n}$$\eta_{n}$$\sigma_{n}$$\varrho_{n}$
,
where the shaded triangles define the $f$-orbits.
Next, quivers of the form
$\textstyle{1\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\scriptstyle{\varepsilon}$$\scriptstyle{\alpha}$$\textstyle{2\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\scriptstyle{\beta}$
or
$\textstyle{2^{\prime}\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\scriptstyle{\gamma}$$\textstyle{3\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\scriptstyle{\delta}$$\scriptstyle{\varepsilon^{\prime}}$
We describe the quivers and algebras we consider. We always take the
multiplicities at $2$-cycles of $g$ equal to $1$, and at loops we take
multiplicity $2$. That is, all arrows in 2-cycles and loops are virtual and
not part of the Gabriel quiver.
### 3.2. Algebras with Gabriel quiver $3\mathcal{A}$
We take the quiver $Q$ obtained by glueing the second and the third type
above, identifying vertex $2$ with vertex $2^{\prime}$.
$\textstyle{1\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\scriptstyle{\varepsilon}$$\scriptstyle{\alpha}$$\textstyle{2\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\scriptstyle{\beta}$$\scriptstyle{\gamma}$$\textstyle{3\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\scriptstyle{\delta}$$\scriptstyle{\varepsilon^{\prime}}$
The permutation $g$ is of the form $(\alpha\ \gamma\ \delta\
\beta)(\varepsilon)(\varepsilon^{\prime})$. Let $m_{\alpha}=k\geq 2$.
The case $k=1$ is special, this gives the triangular algebra, called
$T(\lambda)$ in [8], here $\lambda\neq 1$. With suitable choice of
$c_{\bullet}$, the presentation of the weighted surface algebra induces the
(Gabriel) presentation of $T(\lambda)$
$\displaystyle\alpha\beta\alpha$ $\displaystyle=\alpha\gamma\delta,$
$\displaystyle\delta\beta\alpha$ $\displaystyle=\lambda(\delta\gamma\delta),$
$\displaystyle\beta\alpha\beta$ $\displaystyle=\gamma\delta\beta,$
$\displaystyle\beta\alpha\gamma$ $\displaystyle=\lambda(\gamma\delta\gamma),$
$\displaystyle\alpha\beta\alpha\gamma$ $\displaystyle=0,$
$\displaystyle\beta\alpha\beta\alpha\beta$ $\displaystyle=0,$
$\displaystyle\delta\gamma\delta\beta$ $\displaystyle=0,$
$\displaystyle\gamma\delta\gamma\delta\gamma$ $\displaystyle=0,$
$\displaystyle\alpha\beta\alpha\beta\alpha$ $\displaystyle=0,$
$\displaystyle\delta\gamma\delta\gamma\delta$ $\displaystyle=0,$
$\displaystyle\delta\beta\alpha\beta$ $\displaystyle=0.$
One can show that $T(\lambda)$ and $T(\mu)$ are not isomorphic for
$\lambda\neq\mu$. The weighted surface algebras with the same quiver and
$\varepsilon,\varepsilon^{\prime}$ virtual loops have Gabriel quiver denoted
by $Q(3\mathcal{A})_{2}^{k}$ in [11] (which is the algebra with parameter $k$
in the the family $Q(3\mathcal{A})_{2}$ of [5]).
### 3.3. Spherical algebras
We have the algebras whose quiver is given by the first building block where
we identify $a_{1}=a_{n+1}$, for $n\geq 2$. The case $n=2$ gives the algebra
$S(\lambda)$, called spherical algebra, introduced in [8], Example 3.6, as
follows.
123456$\alpha$$\xi$$\delta$$\eta$$\beta$$\nu$$\varrho$$\varepsilon$$\sigma$$\mu$$\omega$$\gamma$
where the four shaded triangles denote the $f$-orbits. We take all
multiplicities equal to $1$, the presentation induced by the weighted surface
algebra presentation is, with suitable choice of $c_{\bullet}$,
$\displaystyle\alpha\beta\nu$ $\displaystyle=\varrho\omega\nu,$
$\displaystyle\beta\nu\delta$ $\displaystyle=\lambda\beta\gamma\sigma,$
$\displaystyle\nu\delta\alpha$ $\displaystyle=\lambda\gamma\sigma\alpha,$
$\displaystyle\delta\alpha\beta$ $\displaystyle=\delta\varrho\omega,$
$\displaystyle\gamma\sigma\varrho$ $\displaystyle=\nu\delta\varrho,$
$\displaystyle\sigma\varrho\omega$ $\displaystyle=\lambda\sigma\alpha\beta,$
$\displaystyle\varrho\omega\gamma$ $\displaystyle=\lambda\alpha\beta\gamma,$
$\displaystyle\omega\gamma\sigma$ $\displaystyle=\omega\nu\delta,$
$\displaystyle\alpha\beta\nu\delta\alpha$ $\displaystyle=0,$
$\displaystyle\beta\nu\delta\varrho$ $\displaystyle=0,$
$\displaystyle\nu\delta\alpha\beta\nu$ $\displaystyle=0,$
$\displaystyle\delta\alpha\beta\gamma$ $\displaystyle=0,$
$\displaystyle\gamma\sigma\varrho\omega\gamma$ $\displaystyle=0,$
$\displaystyle\sigma\varrho\omega\nu$ $\displaystyle=0,$
$\displaystyle\varrho\omega\gamma\sigma\varrho$ $\displaystyle=0,$
$\displaystyle\omega\gamma\sigma\alpha$ $\displaystyle=0,$
$\displaystyle\beta\gamma\sigma\varrho$ $\displaystyle=0,$
$\displaystyle\sigma\alpha\beta\nu$ $\displaystyle=0,$
$\displaystyle\delta\varrho\omega\gamma$ $\displaystyle=0,$
$\displaystyle\omega\nu\delta\alpha$ $\displaystyle=0,$
$\displaystyle\beta\nu\delta\alpha\beta$ $\displaystyle=0,$
$\displaystyle\delta\alpha\beta\nu\delta$ $\displaystyle=0,$
$\displaystyle\sigma\varrho\omega\gamma\sigma$ $\displaystyle=0,$
$\displaystyle\omega\gamma\sigma\varrho\omega$ $\displaystyle=0.$
### 3.4. The $n$-spherical algebra
When $n\geq 2$, the permutation $g$ is of the form
$\prod_{i=1}^{n}(\xi_{i}\,\eta_{i})\cdot(\gamma_{1}\,\sigma_{1}\,\gamma_{2}\,\sigma_{2}\,\dots\,\gamma_{n}\,\sigma_{n})\cdot(\varrho_{n}\,\delta_{n}\,\varrho_{n-1}\,\delta_{n-1}\,\dots\,\varrho_{1}\,\delta_{1}).$
We take the multiplicities for the $2n$-cycles to be $m,m^{\prime}\geq 1$, and
write $c,c^{\prime}$ for the parameters at these cycles.
### 3.5. A mixed algebra
We can glue together the three building blocks by identifying $2=a_{1}$, and
$2^{\prime}=a_{n+1}$. In this case, the permutation $g$ is the product of one
large cycle with $n$ cycles of length $2$, and two loops:
$\prod_{i=1}^{n}(\xi_{i}\
\eta_{i})\cdot(\gamma_{1}\,\sigma_{1}\,\gamma_{2}\,\sigma_{2}\,\dots\,\gamma_{n}\,\sigma_{n}\
\gamma\ \delta\
\varrho_{n}\,\delta_{n}\,\varrho_{n-1}\,\delta_{n-1}\,\dots\,\varrho_{1}\,\delta_{1}\
\beta\ \alpha)(\varepsilon)(\varepsilon^{\prime}).$
We take again the multiplicities equal to $1$ on 2-cycles of $g$, or
$m_{\gamma_{1}}=m$, and the parameter function with value $1$ on each virtual
arrow. These algebras were not studied in previous papers but they fit into
the same scheme.
## 4\. Construction of the module $M$
Let $\Lambda$ be one of the algebras as described above. Let $\Gamma\subset
Q_{0}$ be the set of vertices which are not adjacent to a virtual arrow.
###### Definition 4.1.
Let $M$ be the (right) $\Lambda$-module
$M:=\Lambda\oplus[\bigoplus_{i\in\Gamma}S_{i}]\oplus[\bigoplus_{\nu\not\in\Gamma}\Omega^{2}(S_{\nu})]$
In the following we write down the details for the case of the $n$-spherical
algebra, for the other algebras they are essentially the same. In this case
$\Gamma=\\{a_{i}\mid 1\leq i\leq n\\}$.
### 4.1. The $\Omega$-translates of the simple modules
For the algebra in question, the dimensions of the indecomposable projectives
are:
$\dim P_{a_{i}}=2n(m+m^{\prime}),\ \ \dim P_{b_{i}}=2nm+2,\ \ \dim
P_{d_{i}}=2nm^{\prime}+2.$
Let $a_{i}\in\Gamma$. The structure of $\Omega^{\pm 1}(S_{a_{i}})$ can be seen
from the presentation of the algebra. The module $\Omega^{2}(S_{a_{i}})$ has
dimension 5, the Loewy structure is
$\begin{matrix}b_{i-1}&&d_{i}\cr&a_{i}&\cr b_{i}&&d_{i-1}\end{matrix}$
That is, the module has a ’simple waist’. Now let $\nu\in Q\setminus\Gamma$ we
set $U_{\nu}:=\Omega^{2}(S_{\nu})$. Then
$\Omega(U_{\nu})=\Omega^{-1}(S_{\nu})$ and
$\Omega^{-1}(U_{\nu})=\Omega(S_{\nu})$, their structure can also be seen from
the presentation. We describe $U_{\nu}$.
###### Lemma 4.2.
The module $U_{b_{i}}$ is uniserial of length $2nm^{\prime}-1$, with
composition series
$U_{b_{i}}=\mathcal{U}(a_{i},d_{i-1},a_{i-1},d_{i-2},a_{i-2},\ldots,a_{i+1})$
The module $U_{d_{i}}$ is uniserial of length $2nm-1$, with composition series
$U_{d_{i}}=\mathcal{U}(a_{i+1},b_{i+1},a_{i+2},b_{i+2},\ldots,a_{i})$
(taking indices modulo $n$ and writing $a_{i}$, $d_{i}$, $b_{i}$ meaning the
corresponding simple module).
Proof We compute $U_{b_{1}}$, that is
$\Omega^{2}(S_{b_{1}})=\Omega(S_{b_{1}})=\\{x\in
P_{a_{2}}\mid\sigma_{1}x=0\\}\subset P_{a_{2}}$. From the relations for the
algebra, we have
$\sigma_{1}\varrho_{1}\delta_{1}=cA_{\sigma_{1}}=c\sigma_{1}A_{\sigma_{1}}^{\prime}$
Hence $\sigma_{1}\psi=0$ if we set
$\psi=\psi_{\varrho_{1}\delta_{1}}:=\varrho_{1}\delta_{1}-cA_{\sigma_{1}}^{\prime}.$
One exhibits a basis for $\psi\Lambda$, showing that it has the same dimension
as $\Omega^{2}(S_{b_{1}})$, hence we have equality. The submodule structure
follows directly. The case $U_{d_{i}}$ is similar. $\Box$
###### Proposition 4.3.
We have ${\rm Ext}^{1}(M,M)=0$ and ${\rm Ext}^{2}(M,M)=0$.
Proof By ext symmetry, it suffices to show that for any non-projective
indecomposable summand $X$ of $M$ we have ${\rm Ext}^{1}(M,X)=0$ and ${\rm
Ext}^{1}(X,M)=0$. For this, we use the following short exact sequences:
Let $a_{i}\in\Gamma$,
$None$ $0\to\Omega(S_{a_{i}})\to P_{a_{i}}\to S_{a_{i}}\to 0,\ \
0\to\Omega^{2}(S_{a_{i}})\ \to P_{c_{i}}\oplus
P_{d_{i}}\to\Omega(S_{a_{i}})\to 0$
Consider a vertex $\nu$ not in $\Gamma$, let $\nu=c_{i}$
$None$ $0\to\Omega^{-1}S_{b_{i}}\to P_{a_{i}}\to U_{b_{i}}\to 0,\ \ 0\to
S_{b_{i}}\to P_{b_{i}}\to\Omega^{-1}(S_{b_{i}})\to 0$
Let $\nu=d_{i}$, then
$None$ $0\to\Omega^{-1}(S_{d_{i}})\to P_{a_{i}+1}\to U_{d_{i}}\to 0,\ \ 0\to
S_{d_{i}}\to P_{d_{i}}\to\Omega^{-1}(S_{d_{i}})\to 0$
We apply the functor ${\rm Hom}_{A}(-,X)$ to the above exact sequences.
(I) Assume $X=S_{a_{j}}$ for some $j$. We know from the quiver that ${\rm
Ext}^{1}(S_{a_{i}},S_{a_{j}})=0$ already. To show that ${\rm
Ext}^{2}(S_{a_{i}},S_{a_{j}})=0$ we apply the functor to the second sequence
in (1). From the structure of $\Omega^{2}(S_{a_{i}})$ we see directly that
${\rm Hom}(\Omega^{2}(S_{a_{i}}),S_{a_{j}})=0$ and hence ${\rm
Ext}^{2}(S_{a_{i}},S_{a_{j}})=0$.
We have ${\rm Ext}^{1}(U_{b_{i}},X)=0$ since ${\rm
Hom}(\Omega^{-1}(S_{b_{i}}),S_{a_{j}})=0$ Furthermore ${\rm
Ext}^{2}(U_{b_{i}},X)=0$ since ${\rm Hom}(S_{b_{i}},S_{a_{j}})=0$. Similarly
one shows ${\rm Ext}^{1}(U_{d_{i}},X)=0$ and ${\rm Ext}^{2}(U_{d_{i}},X)=0$.
(II) Now assume $X=U_{b_{j}}$ for some $j$. First, by dimension shift ${\rm
Ext}^{1}(U_{\nu},U_{b_{j}})\cong{\rm Ext}^{1}(S_{\nu},S_{b_{j}})=0$ for any
$\nu$ of valency $1$, from the quiver. Next, consider ${\rm
Ext}^{2}(U_{\nu},X)$, by applying the functor ${\rm Hom}(-,X)$ to the second
exact sequence in (2). We have ${\rm Hom}(S_{\mu},U_{b_{j}})=0$ (the socle of
$U_{b_{i}}$ is always some $S_{a}$), and hence ${\rm Ext}^{2}(U_{\nu},X)=0$.
Now consider ${\rm Ext}^{t}(S_{a_{i}},X)$ for $t=1,2$. By the ext symmetry, it
is isomorphic to ${\rm Ext}^{t}(X,S_{a_{i}})$ for $t=2,1$. By part (I) we know
that it is zero.
The proof for $X=U_{d_{i}}$ is analogous. $\Box$
###### Remark 4.4.
For possible later use, we write down sequences which may be used to show
${\rm Ext}^{1}(X,M)=0$ and ${\rm Ext}^{2}(X,M)=0$: Let $a_{i}\in\Gamma$,
$None$ $0\to S_{a_{i}}\to P_{a_{i}}\to\Omega^{-1}(S_{a_{i}})\to 0,\ \
0\to\Omega^{-1}(S_{a_{i}})\to P_{b_{i}}\oplus
P_{d_{i}}\to\Omega^{-2}(S_{a_{i}})\to 0$
Consider a vertex $\nu$ not in $\Gamma$, let $\nu=c_{i}$
$None$ $0\to U_{b_{i}}\to P_{a_{i+1}}\to\Omega(S_{b_{i}})\to 0,\ \
0\to\Omega(S_{b_{i}})\to P_{b_{i}}\to S_{b_{i}}\to 0$
Let $\nu=d_{i}$, then
$None$ $0\to U_{d_{i}}\to P_{a_{i}}\to\Omega(S_{d_{i}})\to 0,\ \
0\to\Omega(S_{d_{i}})\to P_{d_{i}}\to S_{d_{i}}\to 0.$
## 5\. Ext vanishing and 3-cluster tilting
We would like to determine when $M$ is 3-cluster tilting. Hence take $X$
indecomposable and not projective, and assume
${\rm Ext}^{1}(M,X)=0={\rm Ext}^{2}(M,X).$
By ext symmetry, we get for free that ${\rm Ext}^{1}(X,M)=0={\rm
Ext}^{2}(X,M).$ The aim is to show that $X$ is in add$(M)$, or if not, to
identify $X$.
###### Lemma 5.1.
The socle and the top of $X$ belong to ${\rm add}(\oplus_{i}S_{a_{i}})$.
Proof Let $\nu$ be a vertex $\neq a_{i}$ for any $i$. Apply the functor ${\rm
Hom}(-,X)$ to the second sequence of (2), this gives the exact sequence
$0\to{\rm Hom}(\Omega^{-1}(S_{b_{i}}),X)\to{\rm Hom}(P_{b_{i}},X)\to{\rm
Hom}(S_{b_{i}},X)\to 0$
Any homomorphsim $P_{b_{i}}\to X$ must map the socle to zero, otherwise it
would be split. Hence it lies in ${\rm Hom}(\Omega^{-1}(S_{b_{i}}),X)$ and
therefore the first two terms are isomorphic. Hence the last term is zero, as
required. To show that also ${\rm Hom}(X,S_{\nu})=0$ we use a sequence from
$(2^{*})$. $\Box$
###### Lemma 5.2.
We have ${\rm Hom}(\Omega(X),S_{a_{i}})=0$ and ${\rm
Hom}(S_{a_{i}},\Omega^{-1}(X))=0$.
Proof Since ${\rm Ext}^{1}(X,S_{a_{i}})=0$, from a minimal projective cover
of $X$ we obtain the exact sequence
$0\to{\rm Hom}(X,S_{a_{i}})\to{\rm Hom}(P_{X},S_{a_{i}})\to{\rm
Hom}(\Omega(X),S_{a_{i}})\to 0$
The first two terms are isomorphic since we start with a projective cover.
Hence the last term is zero. Similarly by using an injective hull we get ${\rm
Hom}(\Omega^{-1}(X),S_{a_{i}})=0$. $\Box$
Let $\mathcal{X}$ be the category of $A$-modules which have socle and top in
${\rm add}(S_{a_{i}})$. This category is equivalent to ${\rm mod}-e\Lambda e$.
where $e$ is the idempotent $e:=\sum_{i}e_{a_{i}}$. An equivalence is given by
the functor $V\mapsto Ve$, with inverse the composite of $(-)\otimes_{e\Lambda
e}(e\Lambda)$ follows by factoring out the largest $A$-submodule $V^{\prime}$
with $V^{\prime}e=0$ (see for example [4]).
We may write down quiver and presentation of the algebra $e\Lambda e$. The
arrows are $x_{i}:=\gamma_{i}\sigma_{i}$ and $y_{i}:=\varrho_{i}\delta_{i}$,
for $1\leq i\leq n$ where $x_{i}:a_{i}\mapsto a_{i+1}$ and $y_{i}:a_{i+1}\to
a_{i}$. From the relations for $\Lambda$ we see We claim that $x_{i}y_{i}=0$
and $y_{i}x_{i-1}=0$. That is, $e\Lambda e$ is special biserial. Moreover, for
any $i$, the longest non-zero monomial $x_{i}x_{i+1}\ldots$ is up to a scalar
equal to the longest non-zero monomial $y_{i-1}y_{i-2}\ldots$, and this gives
the socle relations.
###### Lemma 5.3.
The module $X$ has simple socle and top.
Proof The module $X$, and as well, all projectives (injectives) $P_{a_{i}}$
belong to the category $\mathcal{X}$, and hence we may fix an injective hull,
or projective cover, of $X$ by identifying with the image of a suitable
injective hull, or projective cover, of $Xe$, in ${\rm mod}e\Lambda e$. The
indecomposable $e\Lambda e$-modules are ’strings’ or ’bands’, and their
injective hulls or projective covers may be written down explicitly.
Assume the socle of $X$ is not simple, then consider the injective hull
$I_{X}$, it has at least two indecomposable summands, say it is $\oplus_{i\in
R}P_{a_{i}}$. We may assume, with the above convention, and taking $X\to
I_{X}$ as inclusion, that $X$ has a generator
$\omega=(\omega_{1},\omega_{2},0,\ldots)$ such that $\omega A$ has socle of
length two, and moreover, that $\omega x_{j}=(\omega_{1}x_{j},0,\ldots)$ and
$\omega y_{j-1}=(0,\omega_{2}y_{j-1},0,\ldots)$ and $\omega x_{r}=0$, $\omega
y_{s}=0$ for all other generators $x_{r},y_{s}$ of $e\Lambda e$. This implies
then that $\omega J\subseteq X$ where $J$ is the radical of $\Lambda$. Now
consider $\pi:I_{X}\to\Omega(X)$. The element $\pi(\omega_{1},0,\ldots)$ is
non-zero (since $\omega$ is a generator for $X$). Furthermore,
$[\pi(\omega_{1},0,\ldots)]J=\pi[(\omega_{1},0)J]=0$ since
$(\omega_{1},0,\ldots)J$ is contained in $X$. Now
$\pi(\omega_{1},0)=\pi(\omega_{1},0)e$, (since ${\rm top}X$ is in add$(\oplus
S_{a_{i}})$. Hence for some $i$ we have ${\rm
Hom}(S_{a_{i}},\Omega^{-1}(X))\neq 0$. This contradicts the previous Lemma.
Similarly by exploiting a projective cover, one shows that the top of $X$ must
be simple. $\Box$
###### Proposition 5.4.
The module $X$ is uniserial.
Proof If $X$ is not uniserial then $Xe$ is not uniserial (using the structure
of the projectives in this case). Then $Xe$ is a ’band module’. This means
that $X$ contains a submodule isomorphic to the second socle of some
$P_{a_{j}}$. That is ${\rm Hom}(\Omega^{2}(S_{a_{j}}),X)\neq 0$.
Applying ${\rm Hom}(-,X)$ to the exact sequence
$0\to\Omega^{2}(S_{a_{j}})\stackrel{{\scriptstyle\iota}}{{\to}}P:=P_{b_{j}}\oplus
P_{d_{j-1}}\longrightarrow\Omega(S_{a_{j}})\to 0$
gives an excact sequence, that is a non-zero homomorphism
$\theta:\Omega^{2}(S_{a_{j}})\to X$ factors through $\iota$, say
$\theta=\psi\circ\iota$. The kernel of $\theta$ is the socle of
$\Omega^{2}(S_{a_{j}})$ which also is the socle of $P$. We factor out these
socles, then for the induced maps we have
$\bar{\theta}=\bar{\psi}\circ\bar{\iota}.$
Now, the map $\bar{\psi}$ on the socle of $\bar{P}$ is non-zero on each
component. It follows that the image of $\bar{\psi}$ has Loewy length equal to
the Loewy length of $P/{\rm soc}P$.
Note that all modules $P_{a_{i}}$ have the same Loewy length $\ell$ say. As
well $P_{b_{j}}\oplus P_{d_{j-1}}$ has Loewy length $\ell$. Hence the Loewy
length of $P/{\rm soc}P$ is $\ell-1$. The image of $\bar{\psi}$ is contained
in the radical of $X$, which is the unique maximal submodule. It follows that
the Loewy length of $X$ is $\ell$. But this means that $X$ must be projective,
a contradiction. This shows that $Xe$ is uniserial, and then from the
structure of the projectives, also $X$ is uniserial. $\Box$
We summarize. We have shown that if $X$ is indecomposable and not projective
such that ${\rm Ext}^{1}(M,X)=0={\rm Ext}^{2}(M,X)$ then
$(*)$ $X$ is uniserial, and ${\rm soc}X$ and ${\rm top}X$ are in
add$(\oplus_{i}S_{a_{i}})$. That is, $X$ is a subquotient of some $U_{b_{i}}$
or $U_{d_{j}}$.
We show now that if $X$ is any module satifying $(*)$ then ${\rm
Ext}^{1}(M,X)=0$ and ${\rm Ext}^{2}(M,X)=0$.
###### Lemma 5.5.
Let $X=\mathcal{U}(a_{j},b_{j},a_{j+1},b_{j+1},\ldots a_{l})$, a subquotient
of some $U_{\nu}$. Then
${\rm Ext}^{1}(M,X)=0$ and ${\rm Ext}^{2}(M,X)=0$.
Proof We use the sequences in the proof of Proposition 4.3. We apply the
functor $(-,X):={\rm Hom}(-,X)$ to the exact sequences in (1). We start with
the second, this gives
$0\to(\Omega(S_{a_{i}}),X)\longrightarrow(P_{b_{i}}\oplus
P_{d_{i-1}},X)\longrightarrow(\Omega^{2}(S_{a_{i}}),X)\to{\rm
Ext}^{1}(\Omega(S_{a_{i}}),X)\to 0$
We see that ${\rm Hom}(\Omega^{2}(S_{a_{i}}),X)=0$ ($X$ is uniserial). Hence
the ext space is zero. Moreover, it follows that the first two terms are
isomorphic, which we can use for the first sequence:
$0\to(S_{a_{i}},X)\longrightarrow Xe_{a_{i}}\longrightarrow Xe_{b_{i}}\oplus
Xe_{d_{i}}\to{\rm Ext}^{1}(S_{a_{i}},X)\to 0$
In our case, $Xe_{d_{i}}=0$. Note that in the composition series we have
length two subquotients $a_{r},b_{r}$, except that for $l=r$ we have an extra
copy of $a_{l}$. Hence if $i=l$ then the first term is $K$, and
$|Xe_{a_{l}}|=1+|Xe_{b_{l}}|$ and ext is zero. Suppose $i\neq\ell$, then the
first term is zero and the second and third are isomorphic. Again ext is zero.
Next, we apply $(-,X)$ to the sequences in (3). Since $S_{d_{i}}$ does not
occur in $X$, the functor takes the second sequence to zero. From the first
sequence we get
$0\to{\rm Hom}(U_{d_{i}},X)\to Xe_{a_{i}+1}\to 0\to{\rm
Ext}^{1}(U_{d_{i}},X)\to 0$
and the ext space is zero.
Now consider $(-,X)$ applied to sequences in (2). The second sequence gives
${\rm Hom}(\Omega^{-1}(S_{b_{i}}),X)\cong Xe_{b_{i}}$ and ${\rm
Ext}^{1}(\Omega^{-1}(S_{b_{i}}),X)=0$. Consider the first sequence, this gives
$0\to{\rm Hom}(U_{b_{i}},X)\to Xe_{a_{i}}\to Xe_{b_{i}}\to{\rm
Ext}^{1}(U_{b_{i}},X)\to 0$
If the top (ie $S_{a_{i}}$) of $U_{b_{i}}$ is not the same as the socle of $X$
then the hom space is zero and the second and third term are isomorphic, and
ext is zero. Supoose $i=l$, then the first term is $K$, and
$|Xe_{a_{i}}|=1+|Xe_{b_{i}}|$ and again the ext space is zero. $\Box$
###### Corollary 5.6.
Assume $\Lambda$ is the triangle algebra, or the spherical algebra. Then $M$
is 3-cluster tilting.
Proof For these algebras, all indecomposables satisfying $(*)$ are in
add$(M)$. $\Box$
Consider an $n$-spherical algebra for $n\geq 3$ and $m=m^{\prime}=1$. Then the
(finite) set of modules $X$ satifying $(*)$ contains all modules of the form
$\mathcal{U}(a_{i},b_{i},a_{i+1}),\ \ \mathcal{U}(a_{i},d_{i-1},a_{i-1}).$
To have a 3-cluster tilting module with $M$ as a summand, we would need to
take $\widetilde{M}=M\oplus\mathcal{V}$ where $\mathcal{V}$ is the direct sum
of all modules satisfying $(*)$. However, $\widetilde{M}$ has self-extensions.
For example there is a non-split exact sequence
$0\to\mathcal{U}(a_{2},b_{2},a_{3})\to
S_{a_{2}}\oplus\mathcal{U}(a_{1},b_{1},a_{2},b_{2},a_{3})\to\mathcal{U}(a_{1},b_{1},a_{2})\to
0$
Hence $M$ cannot be extended to a 3-cluster tilting module for the
$n$-spherical algebra when $n\geq 3$.
We also consider the algebra with triangular quiver and $k\geq 2$. In this
case the list of uniserial modules $X$ which are subquotients of $U_{1}$ and
$U_{3}$ contains the modules
$\mathcal{U}(2,3,2),\ \mathcal{U}(2,1,2)$
Let $\widetilde{M}=M\oplus\mathcal{V}$ where $\mathcal{V}$ is the direct sum
of all indecomposable modules satisfying $(*)$. This is not a 3-cluster
tilting module since it has self-extensions: we have the non-split exact
sequence
$0\to\mathcal{U}(2,3,2)\to
S_{2}\oplus\mathcal{U}(2,1,2,3,2)\to\mathcal{U}(2,1,2)\to 0$
## References
* [1] J. Bialkowski, A. Skowroński, On tame weakly symmetric algebras having only periodic modules. Arch. Math. 81(2003), 142-154.
* [2] B. Böhmler, R. Marczinzik, A cluster tilting module for a representation-infinite block of a group algebra. arxiv:2101.10217
* [3] I. Burban, O. Iyama, B. Keller, I. Reiten, Cluster tilting for one-dimensional hypersurface singularities, Adv. Math. 217 (2008) 2443–2484.
* [4] J. Brundan, R. Dipper, A. Kleshchev, Quantum linear groups and representations of $GL_{n}(\mathbb{F}_{q})$, AMS Memoirs vol. 149(2001) no. 706.
* [5] K. Erdmann, Blocks of Tame Representation Type and Related Algebras, in: Lecture Notes in Math., vol. 1428, Springer-Verlag, Berlin-Heidelberg, 1990.
* [6] K. Erdmann, T. Holm Maximal $n$-orthogonal modules for selfinjective algebras, Proc. Amer. Math. Soc. 136(2008), no. 9, 3069-3078.
* [7] K. Erdmann, A. Skowroński, Weighted surface algebras, J. Algebra 505 (2018) 490–558.
* [8] K. Erdmann, A. Skowroński, Weighted surface algebras: general version, J. Algebra 544 (2020), 170-227.
* [9] K. Erdmann, A. Skowroński, Weighted surface algebras: general version, Corrigendum. J. Algebra 569(2021), 875-889.
* [10] K. Erdmann, A. Skowroński, Algebras of generalized dihedral type. Nagoya Math. J. 240(2020), 181-236.
* [11] T. Holm, Derived equivalence classification of algebras of dihedral, semidihedral, and quaternion type. J. Algebra 211(1999), 159-205.
* [12] O. Iyama, Higher-dimensional Auslander-Reiten theory on maximal orthogonal subcategories. Adv. Math. 210 (2007), 22-50.
* [13] O. Iyama, Auslander correspondence, Adv. Math. 210(2007), 51-82.
|
# A Deterministic Algorithm for
the Discrete Logarithm Problem in a Semigroup
Simran Tinani Institute of Mathematics, University of Zurich, Switzerland
Joachim Rosenthal Institute of Mathematics, University of Zurich, Switzerland
###### Abstract
The discrete logarithm problem in a finite group is the basis for many
protocols in cryptography. The best general algorithms which solve this
problem have time complexity of $\mathcal{O}(\sqrt{N}\log N)$, and a space
complexity of $\mathcal{O}(\sqrt{N})$ where $N$ is the order of the group. (If
$N$ is unknown, a simple modification would achieve a time complexity of
$\mathcal{O}(\sqrt{N}(\log N)^{2})$.) These algorithms require the inversion
of some group elements or rely on finding collisions and the existence of
inverses, and thus do not adapt to work in the general semigroup setting. For
semigroups, probabilistic algorithms with similar time complexity have been
proposed. The main result of this paper is a deterministic algorithm for
solving the discrete logarithm problem in a semigroup. Specifically, let $x$
be an element in a semigroup having finite order $N_{x}$. The paper provides
an algorithm, which, given any element $y\in\langle x\rangle$, provides all
natural numbers $m$ with $x^{m}=y$, and has time complexity
$O(\sqrt{N_{x}}(\log N_{x})^{2})$ steps. The paper also gives an analysis of
the success rates of the existing probabilistic algorithms, which were so far
only conjectured or stated loosely.
## 1 Introduction
Let $G$ be a group and assume $x,y\in G$ are two elements of the group. We
refer to $x$ as the base element. The discrete logarithm problem (referred to
henceforth as DLP) asks for the computation of an integer $m\in\mathbb{Z}$
(assuming such integers exist) such that $x^{m}=y$. The DLP plays an important
role in a multitude of algebraic and number theoretic cryptographic systems.
Its use was introduced in the Diffie-Hellman protocol for public key exchange
[7] and has since seen a tremendous amount of development, generalisations and
extensions [12]. Many modern-day systems for public key exchange use the
discrete logarithm problem in a suitable group. The most commonly used groups
have been the multiplicative group of finite fields and the group of points on
an elliptic curve. The DLP in Jacobians of hyperelliptic curves and more
general abelian varieties has also been studied extensively [5].
In this paper, we compute complexities using group multiplications as one
fundamental step. Thus, an exponentiation $x^{e}$ is performed in
$\mathcal{O}(\log e)$ steps. We will use the fact that for two lists of length
$n$ in which a match exists, a match can be found in $\mathcal{O}(n\log n)$
steps using standard sorting and searching algorithms (for details, the
interested reader may refer to [6]). For a general finite group of order $N$,
there exist algorithms that solve the DLP in $\mathcal{O}(\sqrt{N}\log N)$
steps. Such algorithms are said to produce a square root attack. The most
well-known examples are Shank’s Baby Step-Giant Step algorithm [18] and the
Pollard-Rho algorithm [16]. Note that Shank’s algorithm is a deterministic
algorithm having time complexity $\mathcal{O}(\sqrt{N}\log N)$ space
complexity $\mathcal{O}(\sqrt{N})$. In contrast, Pollard’s algorithm is a
probabilistic algorithm having time complexity $\mathcal{O}(\sqrt{N}\log N)$
group multiplications and space complexity $\mathcal{O}(1)$. If $N$ is
unknown, a simple modification of these algorithms would achieve a time
complexity of $\mathcal{O}(\sqrt{N}(\log N)^{2})$.
Elliptic curve groups have been widely implemented in practice since for a
carefully selected elliptic curve group the best known classical algorithm for
solving DLP has running time $\mathcal{O}(\sqrt{N}\log N)$, where $N$ is the
group order. This is in contrast to many other finite groups such as the
multiplicative group of a finite field and the group of invertible matrices
over a finite field where algorithms with subexponential running time are
known [1].
In cryptography the Diffie-Hellman protocol using a finite group has been
generalized to situations where the underlying problem is a discrete logarithm
problem in a semigroup or even to situations where a semigroup acts on a set
[10, 11]. The interested reader will find more material in a recent survey by
Goel et al. [8].
It is naturally interesting to ask whether the DLP also has a square root
attack in more generalized structures such as semigroups. Here, we define a
semigroup as any set of elements with an associative binary operation. Since
the best algorithms for the DLP all make use of the existence of inverses, it
is unclear whether they can be generalized to a semigroup. However, when a
special type of semigroup element, called a torsion element, is used as the
base, it turns out that the DLP is reducible in polynomial time to the DLP in
a finite group. A torsion element is one whose powers eventually repeat to
form a cycle, and will be defined more precisely in Section 2. This section
also elaborates more on why the standard collision-based algorithms are not
directly adaptable to the semigroup case. A semigroup in which every element
is torsion is called a torsion semigroup.
The DLP in semigroups with a torsion base element, in a classical setting, was
first discussed by Chris Monico [14] in 2002, and later in a paper by Banin
and Tsaban [3] in 2016. While the discussion in the present paper is entirely
on classical algorithms, it is also worth mentioning the paper [4], where the
authors independently provide a quantum algorithm that solves the DLP in a
torsion semigroup.
Both the algorithm of Monico and the one of Banin and Tsaban are probabilistic
and might fail with low probability. Further, some of their methods are
heuristic, dependent on an oracle or some additional assumption, and their
success rates and expected number of steps are either conjectured or stated
loosely. It is therefore of interest to come up with an algorithm which
deterministically computes the discrete logarithm in a semigroup. In this
regard we like to make some analogy to the problem of determining if an
integer is a prime number, a problem of great importance in cryptography.
Nowadays in practice the algorithm of Miller and Rabin [13, 17] has been
implemented for many years. Still it was a great result when Agrawal, Kayal
and Saxena [2] came up with a deterministic polynomial time algorithm to
achieve this goal.
A key step in finding the discrete logarithm in a semigroup is computing the
cycle length of an element. Both the algorithms of [3] and [14] rely on
computing some multiple of the cycle length, and then removing “extra” factors
by taking gcd’s until the cycle length is obtained. Once the cycle length
value is obtained, the discrete logarithm may easily be computed with a few
more simple steps. While Monico does not provide further elaboration on how
this is done, the paper by Banin and Tsaban bridges this knowledge gap by
showing how the problem is reduced to a DLP in a group once the cycle length
and start values are known. Denote by $N_{x}$ the order of $x$ (formally
defined in Definition 4). The complexity of the algorithm in [3] is
$\mathcal{O}(\sqrt{N_{x}}(\log N_{x})^{2}\log\log N_{x})$, and the of the one
in [14] is $\mathcal{O}(\sqrt{N_{x}}(\log N_{x})^{2})$. While both of the
existing methods seem to succeed with high probability for practical values,
we show that the process of taking successive gcd’s/factors is unnecessary,
and that one can deterministically find the cycle length. The main
contribution of this paper will be a deterministic algorithm for computing the
discrete logarithm of an element $y$ in some semigroup $S$ with respect to
some torsion base element $x\in S$. The time complexity of our algorithm is
$\mathcal{O}(\sqrt{N_{x}}(\log N_{x})^{2})$.
The paper is structured as follows: After providing preliminaries and basic
definitions in Section 2, we will analyse in Section 3 the success rates and
expected number of steps involved in the probabilistic algorithms for cycle
length by Banin and Tsaban (Algorithm 1) and Monico (Algorithm 3).
In Section 4, which is the main section of this paper, we provide a
deterministic algorithm to calculate the cycle length $L_{x}$ of a torsion
element $x$ of a semigroup and thus to also solve the DLP, without the use of
an oracle. This algorithm has complexity
$\mathcal{O}\left(\sqrt{N_{x}}\cdot(\log N_{x})^{2}\ \right).$ For
completeness, we will also demonstrate the use of Pohlig–Hellman algorithm
[15] for a semigroup.
## 2 Preliminaries
A semigroup $S$ is a set together with an associative binary operation. Like
in group theory where a torsion group consists of elements of finite order
only we define:
###### Definition 1 (Torsion Element).
Let $S$ be a semigroup. An element $x\in S$ is called a torsion element if the
sub-semigroup $\langle x\rangle:=\\{x^{k}\mid k\in\mathbb{N}\\}$ generated by
$x$, is finite. $S$ is called a torsion semigroup if every $x\in S$ is a
torsion element.
Throughout the paper the following definitions will be assumed:
###### Definition 2 (Cycle Start).
Let $x\in S$. The cycle start $s_{x}$ of $x$ is defined as the smallest
positive integer such that $x^{s_{x}}=x^{b}$ for some $b\in\mathbb{N}$,
$b>s_{x}$.
###### Definition 3 (Cycle Length).
Let $x\in S$. The cycle length $L_{x}$ of $x$ is defined as the smallest
positive integer such that $x^{s_{x}+L_{x}}=x^{s_{x}}$.
###### Definition 4 (Element order).
Let $x\in S$. With notation as above, we define the order $N_{x}$ of $x$ as
the cardinality of the sub-semigroup $\langle x\rangle$. Note that
$N_{x}=s_{x}+L_{x}-1.$
###### Definition 5 (Semigroup DLP).
Let $S$ be a semigroup and $x\in S$. The semigroup DLP is defined as follows.
Given $y\in\langle x\rangle:=\\{x^{k}\mid k\in\mathbb{N}\\}$, find all $m\in
N$ such that $x^{m}=y$.
We state below a key result first proved in [3].
###### Lemma 1 ([3]).
Let $S$ be a semigroup and $x\in S$ be an element with cycle start $s_{x}$.
The set of powers $G_{x}=\\{x^{s_{x}+k}\mid k\geq 0\\}$ of $x$ forms a finite
cyclic group. The identity element of $G_{x}$ is given by $x^{tL_{x}}$, where
$t$ is the minimum positive integer such that $x^{tL_{x}}\in G_{x}$.
The following result is stated in [14] in a slightly different formulation. We
provide an equivalent proof based on the group structure of $G_{x}$.
###### Lemma 2 ([14]).
Let $x\in S$ have cycle start $s_{x}$ and cycle length $L_{x}$. For all
integers $n,m\geq s_{x}$, we have $x^{m}=x^{n}\iff n\equiv m\mod L_{x}$.
###### Proof.
We can assume without loss of generality that $n\geq m$, and so we can write
$n=m+kL_{x}+u$, with $k\geq 0$ and $0\leq u<L_{x}$. First suppose that
$n\equiv m\mod L_{x}$, i.e. $u=0$. Since $m,n\geq s_{x}$, we have
$x^{n}=x^{m+kL_{x}}=x^{m}$.
Conversely, if $x^{n}=x^{m}$, write $n_{1}=n-s_{x}\geq 0$, and
$m_{1}=m-s_{x}\geq 0$. We have
$x^{s_{x}+m_{1}}=x^{s_{x}+n_{1}}=x^{s_{x}+m_{1}+kL_{x}+u}=x^{s_{x}+m_{1}+u}.$
Now, without loss of generality, $m_{1}\geq s_{x}$, because if not, one can
always increment $m_{1}$ and $n_{1}$ by multiples of $L_{x}$ until this
happens. So, we can assume that $x^{m_{1}}$ lies in $G_{x}$ and is thus
invertible. We multiply by the inverse on both sides to finally get
$x^{s_{x}}=x^{s_{x}+u}.$
Thus, we must have $u=0$ or $n\equiv m\mod L_{x}$, as required.
∎
###### Remark 1.
It becomes clear from the above discussion that the standard collision-based
algorithms for order and discrete log computations in a group do not adapt
directly to a general semigroup. Collision-based algorithms for the
computation of the order $N$ of a group element $x$ (for instance, see [19])
are based on the principle that whenever $N$ can be expressed in the form
$N=A-B$ for non-negative integers $A$ and $B$, the collision $x^{A}=x^{B}$
always occurs. However, this principle does not work in a semigroup, where
there are two independent components of the order. More specifically, for a
semigroup element $x$ with cycle length $L_{x}$ and cycle start $s_{x}$,
whenever $L_{x}$ may be expressed in the form $A-B$ for non-negative integers
$A$ and $B$, the equality $x^{A}=x^{B}$ holds if and only if $A,B\geq s_{x}$.
As an example, consider a semigroup element $x$ with cycle length $L_{x}=12$
and cycle start $s_{x}=5$. Then, $L_{x}=15-3$, but $x^{15}\neq x^{3}$. Thus
without prior knowledge of the cycle start, the semigroup order $N_{x}$ or
cycle length $L_{x}$ cannot directly be found using the same collision-based
algorithms for groups.
Similarly, collision-based algorithms fail for discrete log computations in a
semigroup. As an example, consider a semigroup element $x$ with cycle length
$L_{x}=15$ and cycle start $s_{x}=10$, and suppose that the discrete log of
$y=x^{5}$ is to be found. Then $y\cdot x^{6}=x^{11}=x^{26}$ is obtained as a
collision. However, unlike in the group case, the conclusion
$y=x^{26-6}=x^{20}$ is wrong since $x^{5}\neq x^{20}$. This happens because
even though $x$ is torsion and forms a cycle of powers, it is not invertible.
This concludes the prerequisite knowledge on torsion elements in semigroups.
In the next section, we study the existing probabalistic algorithms for cycle
lengths, and analyse their assumptions, working and complexities.
## 3 Existing Probabalistic Agorithms
### 3.1 Banin and Tsaban’s Algorithm
In this section, we study the probabalistic algorithm described in [3] for
computing the cycle length of a torsion element in a semigroup. While the
authors of the original paper describe their theory only for torsion
semigroups, it will become clear that the same discussion holds true for any
semigroup when the base element chosen is torsion.
Let $S$ be a semigroup and $x$ be a torsion element of $S$. Let $s_{x}$ denote
the cycle start of $x$ and $L_{x}$ its cycle length. Then, recall from Lemma 1
that $G_{x}:=\\{x^{s_{x}},x^{s_{x}+1},\ldots,x^{s_{x}+L_{x}-1}\\}$ is a cyclic
group, and that it has order $L_{x}$. The authors of [3] assume the
availability of a ‘Discrete Logarithm Oracle’ for the group $G_{x}$, which
returns values $\log_{x}h$ for $h\in G_{x}$. They state that these values need
not be smaller than the group order but are polynomial in the size of $G_{x}$
and the element $x$. The representation of the identity in $G_{x}$ is unknown,
and a method to compute inverses is not available.
The authors claim that the well-known algorithms for discrete logarithm
computations in groups do not explicitly require inverses, or can easily be
modified to work without the use of inverses. While it is true that these
algorithms make use of mainly the existence of inverses rather than their
explicit computation, we believe that the fact that easy modification is
possible is not immediate without some justification. In fact, it will become
clear in the later sections that the modified Baby-Step-Giant-Step algorithm
devised by Monico [14] (and also the deterministic algorithm presented in
Section 4) is a crucial and non-trivial part of any such modification.
We make the following observation from the proof of Lemma 1 found in [3]. For
any $k\geq 0$, denote by $v_{k}$ the smallest positive integer such that
$v_{k}L_{x}\geq 2s_{x}+k.$
We then have $x^{v_{k}L_{x}-s_{x}-k}\in G_{x}$ and
$x^{s_{x}+k}x^{v_{k}L_{x}-s_{x}-k}=x^{v_{k}L_{x}}=x^{tL_{x}},$ (1)
so the inverse of the element $x^{s_{x}+k}$ of $G_{x}$ is given by
$x^{v_{k}(L_{x})-s_{x}-k}$. In particular, the computation of inverses
requires prior knowledge of the cycle start. As will be explained below, the
cycle start may be computed only once the value of the cycle length is known,
using a binary search. This explains why the authors insist that their
Discrete Logarithm Oracle does not need to use the computation of inverses.
Below, we describe Algorithm 1, which is the algorithm suggested in [3] to
compute the order of the group $G_{x}$, i.e. the cycle length $L_{x}$ of $x$.
Input A semigroup $S$ and a torsion element $x\in S$; a DLP oracle for groups
Output The cycle length $L_{x}$ of $x$
1: Initialize $i,j,g,L_{x}\leftarrow 1$, $N>>s_{x}+L_{x}$. Fix bounds
$r>1,s>1$.
2: while $j<s$
1. 1.
Fix a random $z\in\\{\lfloor M/2\rfloor,\ldots,M\\}$ and set $h=x^{z}$.
2. 2.
while $i<r$
1. (a)
Choose a random number $k_{i}>0$.
2. (b)
Use the DLP oracle to compute $k_{i}^{\prime}=\log_{h}(h^{k_{i}})$.
3. (c)
Set $g\leftarrow\gcd\limits_{j\leq
i}(k_{j}-k_{j}^{\prime})=\gcd\left(\gcd\limits_{j<i}(k_{j}-k_{j}^{\prime}),\
k_{i}-{k_{i}}^{\prime}\right)$.
4. (d)
Set $i\leftarrow i+1$.
3. 3.
end while
4. 4.
Set $L_{x}\leftarrow lcm(L_{x},g)$, $j\leftarrow j+1$.
6: Return $L_{x}$.
Algorithm 1 Banin-Tsaban Algorithm for Cycle Length
We first note that the authors state complexities in terms of $L_{x}$, which
are valid when the bound $N$ for $N_{x}$ is known. If the algorithm fails for
a value of $N$, the authors suggest to double $N$ and try again. In this case,
which we will assume from now on, we assert that the complexities need to be
taken in terms of $N_{x}$ instead of $L_{x}$. The oracle may be assumed to
have the standard complexity of $O(\sqrt{N_{x}}\log N_{x})$ steps for discrete
logarithm calculations. Step 2.2(c) takes $\mathcal{O}(\log(\max_{j\leq
i}(k_{j}-k_{i}))=\mathcal{O}(\log N_{x})$ integer operations by the assumption
on the oracle, which does not contribute to the total complexity. Thus, the
total complexity of step (2.2) comes from the oracle alone, and is
$\mathcal{O}(\sqrt{N_{x}}\log N_{x})$. Now, the authors of [3] remark that $r$
and $s$ can be taken to satisfy $r=\mathcal{O}(1)$ and $s=\mathcal{O}(\log\log
N_{x})$. Thus, the total complexity is $\mathcal{O}(\log N_{x})$ times the
complexity of Algorithm 1, and thus $\mathcal{O}(\log\log(N_{x})\log N_{x})$
times the complexity of step (2.2). Therefore, we get the total complexity of
$\mathcal{O}(\log\log N_{x}(\log N_{x})^{2}\sqrt{N_{x}})$.
Finally, in Algorithm 2, we present the application of the binary search
method to find the cycle start once $L_{x}$ is known. This algorithm is
formulated as below for this purpose in [3], though the idea to use a binary
search is also originally mentioned in [14].
Input A semigroup element $x$ with cycle length $L_{x}$
Output Cycle start $s_{x}$ of $x$
1: Initialize $s_{x}\leftarrow 1$
2: while $x^{s_{x}+L_{x}}\neq x^{s_{x}}$ do
$s_{x}\leftarrow 2s_{x}$
3: end while
4: Set $a\leftarrow s_{x}/2$
5: while $|a-s_{x}|\geq 2$
$c\leftarrow(a+s_{x})/2$
if $x^{c+L_{x}}\neq x^{c}$ then
$a\leftarrow c$
else
$s_{x}\leftarrow c$
6: end while
Algorithm 2 Calculating Cycle Start (Binary Search)
###### Lemma 3.
Let $N_{x}$ be the order of the element $x$. Then Algorithm 2 requires
$\mathcal{O}\left((\log N_{x})^{2}\right).$
steps.
###### Proof.
Each of Steps (2) and (5) involves $\mathcal{O}(\log N_{x})$ rounds, each of
which computes requires $\mathcal{O}(\log N_{x})$ semigroup multiplications
and one comparison. The total complexity is thus $\mathcal{O}\left((\log
N_{x})^{2}\right)$. ∎
### 3.2 Monico’s Algorithm
In his PhD thesis [14], Chris Monico provides a probabilistic algorithm
(described below as Algorithm 3) that calculates the cycle length of an
element in a finite ring of order $N$. This algorithm makes use of the
multiplicative semigroup structure of the finite ring, and of the availability
of the explicit bound $N$ for every cycle length, and is in fact applicable to
any semigroup where such a bound $N$ is available. In this subsection, we
analyse this algorithm, provide a more concrete bound on its success rate, and
compute its complexity in terms of $N$. We will discuss this algorithm in
terms of torsion semigroups, as opposed to finite rings.
Input A finite semigroup $S$ with $\lvert S\rvert=N$ and an element $x\in S$
Output The cycle length $L_{x}$ of $x$
1: Set $m=\lceil\sqrt{N}\rceil$. Choose a prime $q>N$.
2: For $0\leq i\leq m$, compute and store in a table the pairs $(i;x^{q+im})$.
Sort the table by the second component.
3: Find the least positive integer $b_{1}$ such that $x^{q+b_{1}}$ is in the
table: $x^{q+b_{1}}=x^{q+a_{1}m}$. (Note: $0<b_{1}<m$).
4: Find the least positive integer $b_{2}$ such that $x^{2q+b_{2}}$ is in the
table: $x^{2q+b_{2}}=x^{q+a_{2}m}$. (Again, $0<b_{2}<m$).
5: Compute $g=\gcd(a_{1}m-b_{1},a_{2}m-b_{2}-q)$.
6: For each divisor $d$ of $g$ below some bound $B$, do the following.
If $x^{N+g/d}=x^{N}$:
set $g\leftarrow g/d$;
7: Output $L_{x}=g$ and stop.
Algorithm 3 Monico’s Baby-Step Giant-Step for Cycle Length
We first note that if $L_{x}>m$ and the table in Step (2) has repeated entries
$x^{q+i_{1}m}=x^{q+i_{2}m}$, then numbers $b_{1}$ and $b_{2}$ may not exist
below $m$. In this case the algorithm needs to be modified to take
$g\leftarrow(i_{1}-i_{2})m$. However, whenever this case does not arise, it
can be shown that steps 3 and 4 are always successful in finding a collision.
We further remark that in step 6, the list of divisors of $g$ is kept fixed,
while $g$ is updated to $g/d$ whenever the condition is satisfied. In the
subsequent steps, non-divisors of $g/d$ can be immediately discarded. However,
the end result depends on the order in which divisors are tested, which the
algorithm does not mention explicitly. However, we note that it is, in fact,
possible to restrict the testing to only the prime power divisors of $g$ below
$B$, and with this setting, the optimal performance is obtained by taking
divisors in decreasing order. We will assume this set-up for the rest of the
analysis.
Step (2) involves $\mathcal{O}(\log N)$ steps compute $x^{q}$ and $x^{m}$ and
another $\mathcal{O}(\sqrt{N})$ multiplications to compute $x^{q},x^{q}\cdot
x^{m},x^{q}\cdot x^{2}m,\ldots,x^{q}\cdot x^{m^{2}}$. Step 3 involves at most
$m$ multiplications $x^{q+1}=x^{q}\cdot x,x^{q+1}\cdot x,\ldots,x^{q+m-1}$,
with complexity $\mathcal{O}(\sqrt{N})$, and match-finding with the first
list, with complexity $\mathcal{O}(\sqrt{N}\log N)$ with standard sorting and
search algorithms. The same is true for step (4). Step (5) has complexity
$\mathcal{O}(\log\max(a_{1}m-b_{1},a_{2}m-b_{2}-q))=\mathcal{O}(\log N)$ and
so does not contribute to the overall complexity. Step 6 involves $B$
iterations of a multiplication and an exponentiation $x^{g/d}$, and thus has a
time complexity of $\mathcal{O}(B(\log g+1))=\mathcal{O}(B\log N)$
multiplications.
In the original work, Monico states that the bound $B$ of Algorithm 3 can
always be chosen so that $B<\sqrt{a_{1}m-b_{1}}$. We remark that this claim
does not hold in the current setting of the algorithm. For example, with a
cycle length value of 4, and $a_{1}m-b_{1}=104$, $a_{2}m-b_{2}-q=52$, we get
$g=52$. If $B<\sqrt{a_{1}m-b_{1}}=\sqrt{104}<11$, then we would only test
divisors $d$ below 11, and would never factor out 13 to obtain the true cycle
length. For such a bound to work, one needs to modify the algorithm to test
both divisors $g$ and $g/d$ in step 6. However, we will show in Lemma 4 that
it is almost always sufficient to take $B$ to be a reasonably large fixed
constant, thus the complexity of step 6 can be counted as $\mathcal{O}(\log
N)$, and does not contribute to the overall complexity. Thus, the overall time
complexity is $\mathcal{O}(\sqrt{N}\log N)$. If $N$ is unavailable, the
algorithm can also be modified to update the value of $N$ step-by-step until a
large enough value is found. In this case, Algorithm 3 has a total complexity
of $\mathcal{O}(\sqrt{N_{x}}(\log N_{x})^{2})$.
Further, Monico suggests a modification to the above algorithm, viz. to find
several such $a_{i}$ and $b_{i}$ and compute all the gcd’s. It is clear that
this suggestion is exactly the method used in Banin and Tsaban’s algorithm as
discussed in Section 3.1.
We now analyze the probability of success. The algorithm first looks for a
collisions of the form $x^{q+{a_{1}}m}=x^{q+b_{1}}$. The working principle is
that in this case, the cycle length $L_{x}$ divides $a_{1}m-b_{1}$. Similarly,
if also $x^{q+{a_{2}}m}=x^{2q+b_{2}}$ then
$g=\gcd(a_{1}m-b_{1},a_{2}m-b_{2}-q)$ is a multiple of $L_{x}$.
So far, the process is essentially the same in both Algorithms 1 and 3: while
the former uses a discrete logarithm oracle to obtain multiples of the cycle
length, the latter directly finds these multiples by finding collisions.
However, in Algorithm 3, we do not proceed with computing multiple factors of
$L_{x}$, but work with the fixed multiple $g$ of $L_{x}$, whereas in Algorithm
1 this multiple shrinks several times.
Algorithm 3 then proceeds by fixing a bound $B$ and iterating over every
number $d$ below $B$ to check if $d\mid g$. If yes, it executes the next part,
i.e. checks if $x^{N+D/d}=x^{N}$, and if this holds, it sets $D\leftarrow
D/d$. Note that if the factorization of the number $g$ is known (or if $g$ can
be factored in time negligible compared to $O(\sqrt{N})$, then we do not need
this fixed bound $B$, and can instead iterate over every prime factor $d$ of
$g$. It is well-known that the number of prime factors of $g$ counted with
multiplicity is $\mathcal{O}(\log g)$, so Step (5) of the algorithm can find
$L_{x}$ in $\mathcal{O}(\log N)$ steps. However, in general, factoring $g$ may
be difficult, so we assume from here on that the algorithm proceeds by fixing
a bound $B$ for the divisors of $g$. Below we analyse the probability of the
algorithm succeeding in terms of $B$ and $g$.
###### Lemma 4.
The probability that Algorithm 3 succeeds is bounded below by
$\left(1-\frac{1}{B}\right)^{\log g}$.
###### Proof.
We write $g=L_{x}\cdot F$ for some number $F$ and suppose that the algorithm
fails. This means that there is a divisor, and hence also a prime power
divisor of $F$, which the algorithm fails to factor out. Let $p$ be a prime
dividing $F$, $\alpha_{p}$ denote its largest power dividing $F$, and
$\beta_{p}$ be its largest power below the fixed bound $B$. So, we have
$p^{\alpha_{p}}\mid F$, $p^{\alpha_{p}+1}\nmid F$, $p^{\beta_{p}}<B$,
$p^{\beta_{p}+1}>B$.
Since the number of times the algorithm divides $g$ by $p$ is
$\sum\limits_{i=1}^{\beta_{p}}i=\beta_{p}\cdot(\beta_{p}+1)/2,$
we must have $\beta_{p}\cdot(\beta_{p}+1)/2<\alpha_{p}$ if the algorithm
fails. So, the algorithm succeeds as long as
$\beta_{p}\cdot(\beta_{p}+1)/2\geq\alpha_{p}$ for every prime divisor $p$ of
$F$. Thus, the probability of success for the algorithm can be bounded below
by
$\displaystyle\prod_{p\mid
g}\mathrm{Prob}\left(\frac{\beta_{p}\cdot(\beta_{p}+1)}{2}\geq\alpha_{p}\right).$
Write $v_{p}=\frac{\beta_{p}(\beta_{p}+1)}{2}$ for simplicity. We may assume
that $g$ is a random multiple of $L_{x}$ below the bound $B$, so $F$ is a
random number in $\\{1,\ldots,\frac{B}{L_{x}}\\}$. We have,
$\displaystyle\mathrm{Prob}(\alpha_{p}\leq v_{p})=$ $\displaystyle
1-\mathrm{Prob}(p^{v_{p}+1}\mid F)$ $\displaystyle=$ $\displaystyle
1-\left(\frac{B/L_{x}}{p^{v_{p}+1}(B/L_{x})}\right)$ $\displaystyle=$
$\displaystyle
1-1/p^{v_{p}+1}=1-\dfrac{1}{p^{\frac{\beta_{p}(\beta_{p}+1)}{2}+1}}.$
Hence, a lower bound for the probability of the algorithm’s success is
$\prod\limits_{p\mid
F}\left(1-\dfrac{1}{p^{\frac{\beta_{p}\cdot(\beta_{p}+1)}{2}+1}}\right).$
Now, we have,
$\displaystyle p^{\beta_{p}+1}>B\;\iff\dfrac{1}{p^{\beta_{p}+1}}<\dfrac{1}{B}$
$\displaystyle\implies$ $\displaystyle
1-\dfrac{1}{p^{\frac{\beta_{p}(\beta_{p}+1)}{2}+1}}>1-\dfrac{1}{B^{\frac{\beta_{p}}{2}+1}}>1-\dfrac{1}{B}.$
We further make the following observation. Let $\omega(n)$ denote the number
of distinct prime divisors of integer $n$ (note, however, that the same
statement also holds if counted with multiplicity). Then clearly,
$2^{\omega(n)}\leq n,$ and so, taking logarithms, $\omega(n)\leq\log_{2}n.$
Collecting all the above results, we conclude that the probability of success
Prob (success) of Algorithm 3 is bounded below as follows.
Prob (success) $\displaystyle\geq\prod_{p\mid F}\left(1-\frac{1}{B}\right)$
$\displaystyle=\left(1-\frac{1}{B}\right)^{\omega(F)}\geq\left(1-\frac{1}{B}\right)^{\log
F}$ $\displaystyle\geq\left(1-\frac{1}{B}\right)^{\log g}.$
∎
Note that this bound shows that Algorithm 3 is indeed successful with
overwhelming probability, as conjectured by the author. For example, with
$B=10^{6}$, even when $g$ is several orders of magnitude larger than $B$, say
$g=2^{4000}$, the probability of success is greater than 99.6 percent, by the
bound derived in Lemma 4.
## 4 Deterministic Solution of the DLP
The solution of the DLP in a semigroup involves two parts: the calculation of
the cycle length and start of the base element $x$, and the use of this value
to find the discrete log.
### 4.1 Deterministic Algorithm for Cycle Length Computation
We now present our deterministic algorithm for the computation of the cycle
length. It works by finding a suitable collision, and also guarantees finding
the actual cycle length rather than just a multiple of it, in a fixed number
of steps.
Input A semigroup $S$ and a torsion element $x\in S$. Assume $N_{x}$ is the
order of $x$.
Output Cycle length $L_{x}$ of $x$
1: Initialize $N\leftarrow 1$.
2: Set $q\leftarrow\lceil\sqrt{N}\rceil$.
3: Compute, one by one, $x^{N},x^{N+1},\ldots,x^{N+q}$ and check for the
equality $x^{N}=x^{N+j}$ at each step $j\geq 1$. Store these values in a table
as pairs $(N+j,x^{N+j})$, $0\leq j<q$. If $x^{N}=x^{N+j}$ for any $j<q$, then
set $L_{x}\leftarrow j$ and end the process. If not, proceed to the next step.
4: For $0\leq i\leq q$, compute, one by one, the values
$x^{N+q},x^{N+2q},\ldots,x^{N+iq}$ and at each step $i$, look for a match in
the table of values calculated in Step (3).
5: Suppose that a match $x^{N+iq}=x^{N+j}$ is found, and $i$ is the smallest
integer such that this happens. Set $L_{x}\leftarrow iq-j$ and end the
process.
6: If no match is found in steps 3 or 5, set $N\leftarrow 2\cdot N$ and go
back to Step (2).
Algorithm 4 Deterministic Algorithm for Cycle Length
###### Theorem 1.
Let $S$ be a semigroup and $x\in S$ a torsion element with order $N_{x}$. If
an upper bound on $N_{x}$ is known, Algorithm 4 returns the correct value of
the cycle length $L_{x}$ with
$\mathcal{O}\left(\sqrt{N_{x}}\cdot(\log N_{x})^{2}\ \right)$
steps. The total space complexity is $\mathcal{O}\left(\sqrt{N_{x}}\right)$
semigroup elements.
###### Proof.
We first assume $N\geq\max(L_{x},s_{x})$ and show that steps 1 to 5 succeed in
finding $L_{x}$. We have $q=\lceil\sqrt{N}\rceil$. If $L_{x}<q$, then the
equality $x^{N}=x^{N+L_{x}}$ is found in the first step and the statement of
the theorem follows. Else if $L_{x}\geq q$, we can write uniquely
$L_{x}=iq-j,$
for some positive integers $i>0$, $0\leq j<q$. Now, we must have $i\leq q$,
because otherwise if $i\geq q+1$, we would have
$L_{x}\geq(q+1)q-j>q^{2}+q-q=q^{2}\geq N,$
a contradiction.
We have
$\displaystyle L_{x}=iq-j,\;0<i\leq q,0\leq j<q$ $\displaystyle\implies$
$\displaystyle N+j+L_{x}=N+iq$ $\displaystyle\implies$ $\displaystyle
x^{N+j}=x^{N+j+L_{x}}=x^{N+iq},$
where the last step follows because $N>s_{x}$ by assumption. So, such a
collision always occurs between elements of the two lists in the algorithm.
We now claim that for the smallest such integer $i$ computed in Step (5) of
Algorithm 4, $L_{x}=iq-j$.
To see this, let $i$ be the smallest positive integer such that
$x^{N+j}=x^{N+iq}.$
Also let $L_{x}=i^{\prime}q-j^{\prime}$, $0<i^{\prime}\leq q$, $0\leq
j^{\prime}<q$. We have already shown above that such integers $i^{\prime}$ and
$j^{\prime}$ exist for our choice of $N$. By the definition of $L_{x}$, we
must have $L_{x}\mid iq-j$. Now suppose that $i^{\prime}>i$. Then,
$\displaystyle i^{\prime}q-j^{\prime}\geq$ $\displaystyle(i+1)q-j^{\prime}$
$\displaystyle=$ $\displaystyle iq+(q-j^{\prime})>iq$ $\displaystyle\geq$
$\displaystyle iq-j.$
But, $L_{x}=i^{\prime}q-j^{\prime}\mid iq-j$, so we must have
$iq-j=i^{\prime}q-j^{\prime}$. Since $i^{\prime}>i$, this means that
$q\leq(i^{\prime}-i)q=(j^{\prime}-j)<j^{\prime},$
which is a contradiction because $0\leq j^{\prime}<q$. So, we must have
$i^{\prime}=i$, $j^{\prime}=j$. This proves the claim.
We have shown above that the algorithm finds the correct cycle length when
$N>\max(s_{x},L_{x})$. Since the algorithm doubles the value of $N$ until a
match is found, it always terminates and outputs the correct cycle length. We
now look at the time complexity.
For a given $N$, step (2) involves one exponentiation, or $\mathcal{O}(\log
N)$ multiplications to find $x^{N}$ and then at most another
$q=\mathcal{O}(\sqrt{N})$ multiplications and equality checks for $x^{N}\cdot
x,x^{N}\cdot x^{2},\ldots,x^{N}\cdot x^{q}$. This step also needs a storage
space of at most $q=\mathcal{O}(\sqrt{N})$ elements. Step 5 needs one
exponentiation or $\mathcal{O}(\log N)$ multiplications to find $x^{q}$, and
then another $q=\mathcal{O}(\sqrt{N})$ multiplications to find $x^{N+q}\cdot
x^{q},x^{N+q}\cdot x^{2q}\ldots,x^{N+q^{2}}$. Finding matches in steps 3 and 5
can be done in $\mathcal{O}(q\log q)=\mathcal{O}(\sqrt{N}\log\sqrt{N})$
comparisons with the use of sorting and efficient look-up methods. Thus,
clearly, steps 1 to 5 in algorithm 4 have a total complexity of
$\mathcal{O}(\sqrt{N}\log N)$.
Moreover, the algorithm starts at $N=1$ and doubles $N$ until the cycle length
is found, i.e. until $N>\max(s_{x},L_{x})$. Thus, the number of times steps 2
to 5 are performed is
$\displaystyle\left\lceil\log\left(\max\left(L_{x},s_{x}\right)\right)\right\rceil=\mathcal{O}\left(\max\left(\log\left(L_{x}\right),\log(s_{x})\right)\right)=\mathcal{O}(\log
N_{x})$
Thus, the total number of steps involved is
$\mathcal{O}\left(\sqrt{N_{x}}\cdot(\log N_{x})^{2}\right).$
Clearly, Step (3) involves the storage of
$q=\lceil\sqrt{N}\rceil=\mathcal{O}\left(\sqrt{\max(s_{x},L_{x})}\right)=\mathcal{O}\left(\sqrt{N_{x}}\right)$
elements, so this value gives the total space complexity. This completes the
proof. ∎
###### Remark 2.
If a bound $N$ on the order $N_{x}$ is known a priori, then Algorithm 4 can
clearly be completed in a single round, with time complexity
$\mathcal{O}\left(\sqrt{N}\cdot(\log N)\right)$.
###### Remark 3.
For the case of a group, there exist better algorithms for the computation of
the order of an element even when the total group order is unbounded. For
instance, Algorithm 3.3 in [19] uses a growth function $d(t)$, which
generalizes the square root function used above, to compute the order $N$ of a
group element $x$, and achieves time and space complexities of
$\mathcal{O}\left(\sqrt{N}\right)$, thus eliminating the additional $\log N$
multiplier introduced by the method in Algorithm 4.
However, this method fails when used for a general semigroup due to the
presence of two independent unknown components of the order. To see this, note
that the algorithm would need to be modified for a semigroup as follows. At
stage $t$, one has $g(t-1)\leq N_{x}<g(t)$. On the completion of the baby
steps, one has a table with the powers
$x^{g(t)},x^{g(t)+1},\ldots,x^{g(t)+b(t)}$ (the addition of $g(t)$ is
necessary in the semigroup case to ensure that the loop is entered). The giant
steps compute $x^{g(t)+g(t-1)+\cdot b(t)},x^{g(t)+g(t-1)+2\cdot
b(t)},\ldots,x^{g(t)+g(t-1)+d(t)\cdot b(t)}=x^{2g(t)}$. Now, while $N_{x}$ is
guaranteed to have a unique expression as $g(t-1)+ib(t)-j$ with $0<i\leq d(t)$
and $0\leq j\leq b(t)$, this does not necessarily lead to a collision. In
fact, if $b(t)<L_{x}<g(t-1)$ and $2L_{x}>g(t)=g(t-1)+d(t)\cdot b(t)$, then
neither the baby steps nor the giant steps leads to a collision, and the cycle
length is never found (note that this can happen only if $L_{x}>s_{x}$).
Moreover, if a collision $x^{g(t)+g(t-1)+ib(t)}=x^{g(t)+j}$ is obtained in the
giant step phase, the only conclusion that can be drawn is that $L_{x}\mid
g(t-1)+ib(t)-j$. If instead we forced the condition $g(t-1)\leq N_{x}<g(t)$, a
collision again may never occur because there is no control on the cycle start
(For instance, in matrix semigroups over finite simple semirings, the cycle
start is often found to be much larger than the cycle length. In such cases,
adapting group-based algorithms would fail). See Remark 1 for further details.
#### 4.1.1 Experimental Results for Cycle Length Computations
We used Algorithm 4 to compute cycle length values in several common
semigroups, such as matrix semigroups over finite fields, matrix semigroups
over the finite simple semiring $S_{20}$ (see [20] for a construction and [11]
for the addition and multiplication tables), and the symmetric and alternating
groups (where the cycle length is precisely the order of the element). We
further used the obtained cycle lengths to compute the cycle start values
using Algorithm 2. The working code may be found at
https://github.com/simran-tinani/semigroup-cycle-length.
### 4.2 Solving the DLP once the Cycle Length is known
In this section, we demonstrate the solution of the DLP for a torsion element
$x$ in the semigroup $S$ once the cycle length is known. As before let $N_{x}$
be the order of the sub-semigroup $\langle x\rangle$, let $L_{x}$ be the cycle
length of the torsion element $x$ (which we assume is already computed) and
let $y\in\langle x\rangle$ be an element.
In [3], the authors demonstrate the next steps in solving for $\log_{x}(y)$,
via a reduction to a DLP in the group $G_{x}$, once $L_{x}$ and $s_{x}$ are
known. The procedure is described in Algorithm 5 below, which has been adapted
from the original formulation in [3].
Input A semigroup $S$, a torsion element $x\in S$, with cycle length $L_{x}$
and cycle start $s_{x}$, and $y\in S$ with $y=x^{m}$
Output The discrete logarithm $m$ of $y$ with base $x$
1: Compute $t=\left\lceil\frac{s_{x}}{L_{x}}\right\rceil$ and define
$x^{\prime}=x^{tL_{x}+1}\in G_{x}$.
2: Find the minimum number $0\leq b\leq t$ such that $y^{\prime}=y\cdot
x^{bL_{x}}\in G_{x}$ using binary search.
3: Use Shank’s Baby-Step Giant-Step algorithm for the group $\langle
x^{\prime}\rangle\subseteq G_{x}$ to compute
$m^{\prime}\in\\{0,1,\ldots,L_{x}-1\\}$ such that
${(x^{\prime})}^{m^{\prime}}={y}^{\prime}$.
4: Find the maximum number $c\geq 0$ such that
$x^{(tL_{x}+1)m^{\prime}-cL_{x}}\in G_{x}$ using binary search.
5: Return $m=m^{\prime}(tL_{x}+1)-(b+c)L_{x}$.
Algorithm 5 Algorithm for Discrete Logarithm
Since authors of [3] do not give an explicit proof of correctness Step 5 in
Algorithm 5, we provide it in Theorem 2. Before this, we will need the
following technical result.
###### Lemma 5.
Let $L_{x}$ be the cycle length of $x\in S$, and $n$, $a$, and $a^{\prime}$ be
fixed positive integers. Suppose that $x^{bL_{x}+n}=x^{a}\in G_{x}$, where $b$
is the minimum number such that $x^{bL_{x}+n}\in G_{x}$, and
$x^{n-cL_{x}}=x^{a^{\prime}}\in G_{x}$, where $c$ the maximum number such that
$x^{n-cL_{x}}\in G_{x}$. Then
$bL_{x}+n\leq a,\ \text{and}\ n-cL_{x}\leq a^{\prime}.$
###### Proof.
First let $x^{bL_{x}+n}=x^{a}$ with $b$ minimal such that $x^{bL_{x}+n}\in
G_{x}$. Suppose, to the contrary, that $bL_{x}+n>a$. We must have, by the
minimality of $b$, $x^{(b-1)L_{x}+n}\not\in G_{x}$, so $(b-1)L_{x}+n<a$.
$\displaystyle\text{But},\;\;\;$ $\displaystyle x^{bL_{x}+n}=x^{a}\in G_{x}$
$\displaystyle\implies$ $\displaystyle bL_{x}+n-a=kL_{x},\ k\geq 1$
$\displaystyle\implies$ $\displaystyle(b-k)L_{x}+n=a$ $\displaystyle\implies$
$\displaystyle x^{(b-k)L_{x}+n}=x^{a}\in G_{x},\ k\geq 1.$
This is a contradiction to the minimality of $b$. So, $bL_{x}+n\leq a$. Now
suppose that $x^{x-cL_{x}}=x^{a}\in G_{x}$, with $c$ maximal, and suppose that
$n-cL_{x}>a^{\prime}$. We argue as above:
$\displaystyle L_{x}\mid n-cL_{x}-a^{\prime}$ $\displaystyle\implies$
$\displaystyle n-(k+c)L_{x}=a^{\prime},\ \text{for some}\ k\geq 1$
$\displaystyle\implies$ $\displaystyle x^{n-(k+c)L_{x}}=x^{a^{\prime}}\in
G_{x},$
which is a contradiction to the maximality of $c$. Thus $n-cL_{x}\leq
a^{\prime}$. ∎
###### Theorem 2.
Let $S$ be a semigroup, $x\in S$ a torsion element and $y\in\langle x\rangle$
any element. Assume the cycle length $L_{x}$ and cycle start $s_{x}$ of $x$
are known. Then Algorithm 5 returns the correct values of the discrete
logarithm $m=\log_{x}(y)$ in $\mathcal{O}\left(\sqrt{L_{x}}+(\log
N_{x})^{2}\right)$ semigroup multiplications, with a required storage of
$\mathcal{O}\left(\sqrt{L_{x}}\right)$ semigroup elements.
###### Proof.
We use the notations of Algorithm 5, and also write $n=\log_{x}y$. We will
show that the output $m$ is equal to the correct discrete logarithm value $n$.
Recall that we have a group $G_{x}$, generated by
${x}^{\prime}:=x^{tL_{x}+1}$, and with identity $x^{tL_{x}}$. The parameter
$t$ is given by the formula $t=\left\lceil\frac{s_{x}}{L_{x}}\right\rceil$.
Inverses in $G_{x}$ can be computed in polynomial time using the formula (1).
Note that membership in $G_{x}$ can be tested with one equality check: $y\in
G_{x}\iff y\cdot x^{L_{x}}=y$. There are now two cases:
1. 1.
When $y\in G_{x}$, we have $b=0$. Here, it is possible to use Shank’s Baby
Step-Giant Step algorithm [18] which is a deterministic algorithm and which
requires $\mathcal{O}\left(\sqrt{L_{x}}\right)$ semigroup multiplications and
storage space $\mathcal{O}\left(\sqrt{L_{x}}\right)$, in order to compute
$\log_{{x}^{\prime}}(y)$. This is done in Step (3). From this value,
$n=\log_{x}(y)$ is readily computed, as shown below. Note that in this case,
$\log_{x}(y)$ is determined modulo $L_{x}$.
2. 2.
When $y\not\in G_{x}$, Algorithm 5 first computes, using binary search, the
smallest power $b$ of $x^{L_{x}}$ such that the product $y\cdot x^{bL_{x}}$
lies in the group $G_{x}$, and then proceeds as in case 1 via the Baby Step-
Giant Step algorithm to find the discrete logarithm $m^{\prime}$ of $y\cdot
x^{bL_{x}}$ with base $x^{\prime}$ (i.e. ${(x^{\prime})}^{m^{\prime}}=y\cdot
x^{bL_{x}}$). Note that in this case, the value of $\log_{x}(y)$ is less than
$s_{x}$, and is thus determined uniquely in $\mathbb{N}$. Again, the time and
space complexity are both $\mathcal{O}\left(\sqrt{L_{x}}\right)$.
In both cases above, we have the maximal value $c$ such that
$x^{m^{\prime}(tL_{x}+1)-cL_{x}}\in G_{x}$, and so $c\leq
L_{x}+s_{x}+1=N_{x}+1$, since $m^{\prime}\leq L_{x}$ and $tL_{x}\leq
L_{x}+s_{x}$. We also clearly have $b\leq t\leq N_{x}$. Since the computations
of both $b$ and $c$ are done via binary searches, they contribute
$\mathcal{O}((\log N_{x})^{2})$ steps to the overall time complexity. Now,
$x^{m^{\prime}(tL_{x}+1)-cL_{x}}=x^{m^{\prime}(tL_{x}+1)}=(x^{\prime})^{m^{\prime}}=x^{bL_{x}+n}.$
Applying Lemma 5 to the above equation, we must have
$m^{\prime}(tL_{x}+1)-cL_{x}\leq bL_{x}+n,\ \text{and}\ bL_{x}+n\leq
m^{\prime}(tL_{x}+1)-cL_{x}.$
Therefore, $bL_{x}+n=m^{\prime}(tL_{x}+1)-cL_{x}$, or
$n=m^{\prime}(tL_{x}+1)-(b+c)L_{x},$ which is precisely equal to $m$, the
value returned by the Algorithm 5. Thus, $m=n$. This completes the proof. ∎
Combining Theorem 1, Lemma 3 and Theorem 2 we arrive at the main proposition
of the paper:
###### Proposition 1.
Let $S$ be a semigroup, $x\in S$ a torsion element and $y\in\langle x\rangle$
any element. The discrete logarithm $m=\log_{x}(y)$ can be computed
deterministically in
$\mathcal{O}\left(\sqrt{N_{x}}\cdot(\log N_{x})^{2}\ \right)$
steps, with a required storage of $\mathcal{O}\left(\sqrt{N_{x}}\right)$
semigroup elements.
###### Proof.
For the solution, one begins by finding $L_{x}$. This can be done using
Algorithm 4 and according to Theorem 1 this requires
$\mathcal{O}\left(\sqrt{N_{x}}\cdot(\log N_{x})^{2}\right)$ steps and the
storage of $\mathcal{O}\left(\sqrt{N_{x}}\right)$ elements.
By Lemma 3 the computation of the cycle start $s_{x}$ is achieved in
$\mathcal{O}((\log N_{x})^{2})$ semigroup multiplications, which does not
contribute to the overall cost of the algorithm.
By Theorem 2, the discrete logarithm $m$ can then be retrieved using Algorithm
5, in $\mathcal{O}\left((\log N_{x})^{2}+\sqrt{L_{x}}\right)$ steps, with a
required storage of $\mathcal{O}\left(\sqrt{L_{x}}\right)$ semigroup elements.
As $L_{x}\leq N_{x}$, the overall complexity is dominated by the computation
of the cycle length, and the proof of the result is now clear. ∎
### 4.3 Solving the DLP once the Factorization of the Cycle Length is known
We mentioned in the introduction that for a general group of order $N$ the
best general known algorithms for solving the discrete logarithm problem have
complexity $\mathcal{O}(\sqrt{N})$ operations.
In case the order $N$ has a prime factorization into small primes there is the
famous Pohlig–Hellman algorithm [15] for solving the DLP whose complexity is
dominated by the largest prime factor in the integer factorization of $N$.
In case that we have available the integer factorization of the cycle length
$L_{x}$ we can adapt the Pohlig–Hellman algorithm for groups to a
Pohlig–Hellman algorithm for solving the DLP in a semigroup. Algorithm 6
represents this adapted Pohlig–Hellman algorithm.
Input A semigroup $S$, a torsion element $x\in S$, with cycle length
$L_{x}=\prod_{i=1}^{r}p_{i}^{e_{i}}$ and cycle start $s_{x}$, and $y\in S$
with $y=x^{m}$
Output The discrete logarithm $m$ of $y$ with base $x$
1: Compute $t=\left\lceil\frac{s_{x}}{L_{x}}\right\rceil$ and define
$x^{\prime}=x^{tL_{x}+1}\in G_{x}$.
2: Find the minimum number $0\leq b\leq t$ such that $y^{\prime}=y\cdot
x^{bL_{x}}\in G_{x}$ using binary search.
3: for $i\in\\{1,\ldots,r\\}$
1. 1.
Compute the values $x_{i}^{\prime}=(x^{\prime})^{L_{x}/p_{i}^{e_{i}}}$,
$y_{i}^{\prime}=(y^{\prime})^{L_{x}/p_{i}^{e_{i}}}$, and
$\gamma_{i}:=(x_{i}^{\prime})^{p^{e_{i}-1}}$.
2. 2.
Calculate the inverse $z_{i}$ of ${x_{i}}^{\prime}$ in $G_{x}$ using (1).
3. 3.
Set $k\leftarrow 0$ and $n_{0}\leftarrow 0$.
4. 4.
while $k<e_{i}$ do
1. (a)
Compute
${y_{k}^{\prime}}=(y_{i}^{\prime}z_{i}^{n_{k}})^{p^{e_{i}-1-k}}\in\langle\gamma_{i}\rangle$.
2. (b)
Use Shank’s Baby-Step Giant-Step algorithm for the group
$\langle\gamma_{i}\rangle\subseteq G_{x}$ to compute
$d_{k}\in\\{0,1,\ldots,p_{i}-1\\}$
such that ${\gamma_{i}}^{d_{k}}={y_{k}}^{\prime}$.
3. (c)
Set $n_{k+1}\leftarrow n_{k}+p_{i}^{k}d_{k}$, and $k\leftarrow k+1$.
5. 5.
end while
6. 6.
Set $m_{i}:=n_{e_{i}}$.
4: end for
5: Use the Chinese Remainder Theorem to solve the congruence equations
${\displaystyle m^{\prime}\equiv m_{i}{\pmod{p_{i}^{e_{i}}},}\quad\forall\
i\in\\{1,\dots,r\\}}$ uniquely for $m^{\prime}\mod L_{x}$. This gives the
discrete logarithm of $y^{\prime}$ with respect to the base $x^{\prime}$ in
the group $G_{x}$.
6: Find the maximum number $c\geq 0$ such that
$x^{(tL_{x}+1)m^{\prime}-cL_{x}}\in G_{x}$ using binary search.
7: Return $m=m^{\prime}(tL_{x}+1)-(b+c)L_{x}$.
Algorithm 6 Pohlig–Hellman Algorithm for solving the Discrete Logarithm
Problem in a Semigroup
###### Theorem 3.
Let $S$ be a semigroup, $x\in S$ a torsion element and $y\in\langle x\rangle$
any element. Assume the cycle start $s_{x}$ of $x$ is known and assume the
integer factorization of the cycle length $L_{x}$ is known to be
$L_{x}=\prod_{i=1}^{r}p_{i}^{e_{i}}$. Then Algorithm 6 computes the discrete
logarithm $\log_{x}y$ requiring
$\mathcal{O}\left(\sum\limits_{i=1}^{r}e_{i}\left(\log
L_{x}+\sqrt{p_{i}}\right)+\left(\log N_{x}\right)^{2}\right)$ steps. The space
complexity of the algorithm consists in
$\mathcal{O}\left(\sum\limits_{i=1}^{r}e_{i}\sqrt{p_{i}}\right)$ semigroup
elements.
###### Proof.
Steps 1 and 2 are in analogy to the corresponding steps of Algorithm 5. Steps
3 to 5 represent the Pohlig–Hellman algorithm for groups with the implied
complexity dominated by the largest prime factor $p_{i}$ of the integer
factorization of $L_{x}$ (for a reference on Pohlig–Hellman in groups, see in
[9, Theorem 2.32]). It follows that the running time of the algorithm is
$\mathcal{O}\left(\sum\limits_{i=1}^{r}e_{i}\left(\log
L_{x}+\sqrt{p_{i}}\right)\right)$ steps. The computation of $b$ and $c$
require in addition $\left(\log N_{x}\right)^{2}$ steps. The total space
complexity is $\mathcal{O}\left(\sum\limits_{i=1}^{r}e_{i}\sqrt{p_{i}}\right)$
semigroup elements and that completes the proof. ∎
## 5 Conclusion
The DLP in a finite group has noteworthy significance for cryptography, and so
an extension of existing solutions to other algebraic structures like
semigroups, where inverses may not be available, is of natural interest. In
particular, the DLP in a semigroup has been discussed before in two places,
namely [3] and [14]. Both these authors provide probabilistic generalizations
of existing collision-based methods for the case of a semigroup. The time
complexity of the algorithm in [3] is $\mathcal{O}(\sqrt{N_{x}}(\log
N_{x})^{2}\log\log N_{x})$, and the one in [14] is
$\mathcal{O}(\sqrt{N_{x}}(\log N_{x})^{2})$. Both these methods rely on
computing a multiple of the cycle length and then taking gcd’s or factors, and
could fail with a small probability that depends on the parameters chosen. In
this paper, we provided a deterministic solution of the semigroup DLP, which
computes the cycle length directly and does not rely on finding a factor of
it. The time complexity of our algorithm is $\mathcal{O}(\sqrt{N_{x}}(\log
N_{x})^{2})$. We further demonstrated the application of the Pohlig-Hellman
algorithm to semigroups. A direct consequence of our findings is that for
cryptographic purposes, generalizing the type of algebraic structure for the
DLP offers no additional advantage, at least in the torsion case, both in a
classical and a quantum setting.
## 6 Acknowledgement
This research is supported by armasuisse Science and Technology. The second
author is also supported by Swiss National Science Foundation grant n. 188430.
## References
* [1] L. M. Adleman and J. DeMarrais. A subexponential algorithm for discrete logarithms over all finite fields. Math. Comp., 61(203):1–15, 1993.
* [2] M. Agrawal, N. Kayal, and N. Saxena. PRIMES is in P. Ann. of Math. (2), 160(2):781–793, 2004.
* [3] M. Banin and B. Tsaban. A reduction of semigroup DLP to classic DLP. Des. Codes Cryptography, 81(1):75–82, October 2016.
* [4] A.M. Childs and G. Ivanyos. Quantum computation of discrete logarithms in semigroups. Journal of Mathematical Cryptology, 8(4):405 – 416, 01 Dec. 2014\.
* [5] H. Cohen, G. Frey, R. Avanzi, C. Doche, T. Lange, K. Nguyen, and F. Vercauteren, editors. Handbook of Elliptic and Hyperelliptic Curve Cryptography. Discrete Mathematics and its Applications (Boca Raton). Chapman & Hall/CRC, Boca Raton, FL, 2006.
* [6] Thomas H Cormen, Charles E Leiserson, Ronald L Rivest, and Clifford Stein. Introduction to algorithms. MIT press, 2009.
* [7] W. Diffie and M. E. Hellman. New directions in cryptography. IEEE Trans. Inform. Theory, IT-22(6):644–654, 1976.
* [8] N. Goel, I. Gupta, and B. K. Dass. Survey on SAP and its application in public-key cryptography. J. Math. Cryptol., 14(1):144–152, 2020.
* [9] Jeffrey Hoffstein, Jill Pipher, Joseph H Silverman, and Joseph H Silverman. An introduction to mathematical cryptography, volume 1. Springer, 2008.
* [10] D. Kahrobaei, C. Koupparis, and V. Shpilrain. Public key exchange using matrices over group rings. Groups Complex. Cryptol., 5(1):97–115, 2013.
* [11] G. Maze, C. Monico, and J. Rosenthal. Public key cryptography based on semigroup actions. Adv. in Math. of Communications, 1(4):489–507, 2007.
* [12] A. J. Menezes, P. C. van Oorschot, and S. A. Vanstone. Handbook of applied cryptography. CRC Press Series on Discrete Mathematics and its Applications. CRC Press, Boca Raton, FL, 1997.
* [13] G. L. Miller. Riemann’s hypothesis and tests for primality. J. Comput. System Sci., 13(3):300–317, 1976.
* [14] C. Monico. Semirings and semigroup actions in public-key cryptography. PhD thesis, University of Notre Dame Notre Dame, 2002.
* [15] S. Pohlig and M. Hellman. An improved algorithm for computing logarithms over GF(p) and its cryptographic significance (corresp.). IEEE Transactions on information Theory, 24(1):106–110, 1978.
* [16] J. M. Pollard. Monte Carlo methods for index computation. Mathematics of computation, 32(143):918–924, 1978.
* [17] M. O. Rabin. Probabilistic algorithm for testing primality. J. Number Theory, 12(1):128–138, 1980.
* [18] D. Shanks. Class number, a theory of factorization, and genera. In Proc. of Symp. Math. Soc., 1971, volume 20, pages 41–440, 1971\.
* [19] Andrew V. Sutherland. Order computations in generic groups. 2007\.
* [20] J. Zumbrägel. Classification of finite congruence-simple semirings with zero. Journal of Algebra and Its Applications, 07(03):363–377, 2008.
|
# Compact groups with a set of positive Haar measure satisfying a nilpotent
law
Alireza Abdollahi Department of Pure Mathematics, Faculty of Mathematics and
Statistics, University of Isfahan, Isfahan 81746-73441, Iran.
<EMAIL_ADDRESS>and Meisam Soleimani Malekan Department of
Mathematics, University of Isfahan, Isfahan 81746-73441, Iran; Institute for
Research in Fundamental Sciences, School of Mathematics, Tehran, Iran.
<EMAIL_ADDRESS>
###### Abstract.
The following question is proposed in [4, Question 1.20].
Let $G$ be a compact group, and suppose that
${\mathcal{N}}_{k}(G)=\\{(x_{1},\dots,x_{k+1})\in
G^{k+1}\;|\;[x_{1},\dots,x_{k+1}]=1\\}$ has positive Haar measure in
$G^{k+1}$. Does $G$ have an open $k$-step nilpotent subgroup?
The case $k=1$ is already known. We positively answer it for $k=2$.
###### Key words and phrases:
Compact groups, subsets with positive Haar measure, $2$-step nilpotent groups
###### 2010 Mathematics Subject Classification:
20E18; 20P05
## 1\. Introduction and Results
Let $G$ be a (Hausdorff) compact group. Then $G$ has a unique normalized Haar
measure denoted by ${\mathbf{m}}_{G}$. The following question is proposed in
[4].
###### Question 1.1.
[4, Question 1.20] Let $G$ be a compact group, and suppose that
${\mathcal{N}}_{k}(G)=\\{(x_{1},\dots,x_{k+1})\in
G^{k+1}\;|\;[x_{1},\dots,x_{k+1}]=1\\}$ has positive Haar measure in
$G^{k+1}$. Does $G$ have an open $k$-step nilpotent subgroup?
Positive answer to Question 1.1 is known for $k=1$ (see [3, Theorem 1.2]). It
follows from [4, Theorem 1.19] that Question 1.1 has positive answer for
arbitrary $k$ whenever we further assume that $G$ is totally disconnected i.e.
$G$ is a profinite group. Here we answer positively Question 1.1 for $k=2$
(see Theorem 3.2 below).
## 2\. A preliminary lemma
We need the following lemma in the proof of our main result.
###### Lemma 2.1.
Suppose that $G$ is a group and $x_{1},x_{2},x_{3},g_{1},g_{2},g_{3}\in G$ are
such that
$\displaystyle 1$
$\displaystyle\overset{1}{=}[x_{1},x_{2},x_{3}]\overset{2}{=}[x_{1}g_{1},x_{2}g_{2},x_{3}g_{3}]\overset{3}{=}[x_{1}g_{1},x_{2}g_{2},x_{3}]\overset{4}{=}[x_{1}g_{1},x_{2},x_{3}g_{2}]$
$\displaystyle\overset{5}{=}[x_{1}g_{1},x_{2},x_{3}]\overset{6}{=}[x_{1}g_{1},x_{2},x_{3}g_{3}]\overset{7}{=}[x_{1},x_{2},x_{3}g_{1}]\overset{8}{=}[x_{1},x_{2}g_{2},x_{3}g_{1}]$
$\displaystyle\overset{9}{=}[x_{1},x_{2}g_{2},x_{3}]\overset{10}{=}[x_{1},x_{2},x_{3}g_{2}]\overset{11}{=}[x_{1},x_{2}g_{2},x_{3}g_{3}]\overset{12}{=}[x_{1},x_{2},x_{3}g_{3}].$
Then $[g_{1},g_{2},g_{3}]=1$.
###### Proof.
We will throughout using famous commutator calculus identities.
$\displaystyle 1$
$\displaystyle=[x_{1}g_{1},x_{2}g_{2},g_{3}]=[[x_{1}g_{1},g_{2}][x_{1}g_{1},x_{2}]^{g_{2}},g_{3}]\;\;\text{\rm
by (2) and (3)}$
$\displaystyle=[[x_{1}g_{1},g_{2}][x_{1}g_{1},x_{2}],g_{3}]\;\;\textbf{\rm by
(4) and (5)}$
$\displaystyle=[x_{1}g_{1},g_{2},g_{3}]=[[x_{1},g_{2}]^{g_{1}}[g_{1},g_{2}],g_{3}]\;\;\textbf{\rm
by (5) and (6). \;\; (I)}$
On the other hand,
$\displaystyle 1$ $\displaystyle=[x_{1},x_{2}g_{2},g_{1}]\;\;\text{\rm by (8)
and (9)}$
$\displaystyle=[[x_{1},g_{2}][x_{1},x_{2}]^{g_{2}},g_{1}]=[[x_{1},g_{2}][x_{1},x_{2}],g_{1}]\;\;\text{by
(1) and (10)}$ $\displaystyle=[x_{1},g_{2},g_{1}]\;\;\text{by (1) and (7).
\;\; (II)}$
Also,
$\displaystyle 1$ $\displaystyle=[x_{1},x_{2}g_{2},g_{3}]\;\;\text{\rm by (9)
and (11)}$
$\displaystyle=[[x_{1},g_{2}][x_{1},x_{2}]^{g_{2}},g_{3}]=[[x_{1},g_{2}][x_{1},x_{2}],g_{3}]\;\;\text{by
(1) and (10)}$ $\displaystyle=[x_{1},g_{2},g_{3}]\;\;\text{by (1) and (12).
\;\; (III)}$
Now it follows from (I), (II) and (III) that $[g_{1},g_{2},g_{3}]=1$. ∎
###### Remark 2.2.
The “left version” ($g_{i}x_{j}$ instead of $x_{j}g_{i}$) of Lemma 3.2 is not
clear to hold. The validity of a similar result to Lemma 3.2 for commutators
with length more than $3$ is also under question.
## 3\. Compact groups with many elements satisfying the $2$-step nilpotent
law
We need the “right version” of [5, Theorem 2.3] as follows.
###### Theorem 3.1.
If $A$ is a measurable subset with positive Haar measure in a compact group
$G$, then for any positive integer $k$ there exists an open subset $U$ of $G$
containing $1$ such that $\mathbf{m}_{G}(A\cap Au_{1}\cap\cdots\cap Au_{k})>0$
for all $u_{1},\dots,u_{k}\in U$.
###### Proof.
Since $\mathbf{m}_{G}(A)=\mathbf{m}_{G}(A^{-1})$, it follows from Theorem 2.3
of [5] that there exists an open subset $V$ of $G$ containing $1$ such that
$\mathbf{m}_{G}(A^{-1}\cap v_{1}A^{-1}\cap\cdots\cap v_{k}A^{-1})>0$
for all $v_{1},\dots,v_{k}\in V$. By [2, Theorem 4.5], there exists an open
subset $U\subseteq V$ such that $1\in U$ and $U=U^{-1}$. Thus for all
$u_{1},\dots,u_{k}\in U$
$\displaystyle 0<$ $\displaystyle\mathbf{m}_{G}(A^{-1}\cap
u_{1}^{-1}A^{-1}\cap\cdots\cap u_{k}^{-1}A^{-1})$ $\displaystyle=$
$\displaystyle\mathbf{m}_{G}((A\cap Au_{1}\cap\cdots\cap Au_{k})^{-1})$
$\displaystyle=$ $\displaystyle\mathbf{m}_{G}(A\cap Au_{1}\cap\cdots\cap
Au_{k})$
This completes the proof. ∎
Now we can prove our main result.
###### Theorem 3.2.
Let $G$ be a compact group, and suppose that
${\mathcal{N}}_{2}(G)=\\{(x_{1},x_{2},x_{3})\in G\times G\times
G\;|\;[x_{1},x_{2},x_{3}]=1\\}$ has positive Haar measure in $G\times G\times
G$. Then $G$ has an open $2$-step nilpotent subgroup.
###### Proof.
Let $X:={\mathcal{N}}_{2}(G)$. It follows from Theorem 3.1 and [2, Theorem
4.5] that there exists an open subset $U=U^{-1}$ of $G$ containing $1$ such
that
$X\cap X\bar{u}_{1}\cap\cdots\cap X\bar{u}_{11}\neq\varnothing\;\;\;(*)$
for all $\bar{u}_{1},\dots,\bar{u}_{11}\in U\times U\times U$. Now take
arbitrary elements $g_{1},g_{2},g_{3}\in U$ and consider
$\displaystyle\bar{u}_{1}$
$\displaystyle=(g_{1}^{-1},g_{2}^{-1},g_{3}^{-1}),\bar{u}_{2}=(g_{1}^{-1},g_{2}^{-1},1),\bar{u}_{3}=(g_{1}^{-1},1,g_{2}^{-1})$
$\displaystyle\bar{u}_{4}$
$\displaystyle=(g_{1}^{-1},1,1),\bar{u}_{5}=(g_{1}^{-1},1,g_{3}^{-1}),\bar{u}_{6}=(1,1,g_{1}^{-1}),\bar{u}_{7}=(1,g_{2}^{-1},g_{1}^{-1})$
$\displaystyle\bar{u}_{8}$
$\displaystyle=(1,g_{2}^{-1},1),\bar{u}_{9}=(1,1,g_{2}^{-1}),\bar{u}_{10}=(1,g_{2}^{-1},g_{3}^{-1}),\bar{u}_{11}=(1,1,g_{3}^{-1}).$
By $(*)$, there exists $(x_{1},x_{2},x_{3})\in X$ such that all the following
3-tuples are in $X$.
$\displaystyle(x_{1}g_{1},x_{2}g_{2},x_{3}g_{3}),(x_{1}g_{1},x_{2}g_{2},x_{3}),(x_{1}g_{1},x_{2},x_{3}g_{2})$
$\displaystyle(x_{1}g_{1},x_{2},x_{3}),(x_{1}g_{1},x_{2},x_{3}g_{3}),(x_{1},x_{2},x_{3}g_{1}),(x_{1},x_{2}g_{2},x_{3}g_{1})$
$\displaystyle(x_{1},x_{2}g_{2},x_{3}),(x_{1},x_{2},x_{3}g_{2}),(x_{1},x_{2}g_{2},x_{3}g_{3}),(x_{1},x_{2},x_{3}g_{3}).$
Now Lemma 2.1 implies that $\langle g_{1},g_{2},g_{3}\rangle$ is nilpotent of
class at most $2$. Therefore the subgroup $H:=\langle U\rangle$ generated by
$U$ is $2$-step nilpotent. Since $H=\bigcup_{n\in{\mathbb{N}}}U^{n}$, $H$ is
open in $G$. This completes the proof. ∎
## References
* [1] G. B. Folland, A Course in Abstract Harmonic Analysis, Stud. Adv. Math., Taylor & Francis, London, 1994.
* [2] E. Hewitt and K. A. Ross, Abstract Harmonic Analysis: Structure and Analysis for Compact Groups Analysis on Locally Compact Abelian Groups, Grundlehren Math. Wiss., Springer, Berlin, 2013.
* [3] K. H. Hofmann and F. G. Russo, The probability that $x$ and $y$ commute in a compact group, Math. Proc. Cambridge Philos. Soc. 153 (2012), no. 3, 557-571.
* [4] A. Martino, M. C. H. Tointon, M. Valiunas and E. Ventura, Probabilistic Nilpotence in infinite groups, to appear in Israel J. Math.
* [5] M. Soleimani Malekan, A. Abdollahi and M. Ebrahimi, Compact groups with many elements of bounded order, J. Group Theory, 23 (2020) no. 6 991–998.
|
††thanks: This work has been partially funded by the Ministry of Science,
Innovation and Universities of Spain through the grant PGC2018-095998-B-I00
and by the Agency for Management of University and Research Grants of
Catalonia through the grant 2017SGR1725.
# Geometry of certain foliations on the complex projective plane
Samir Bedrouni Faculté de Mathématiques, USTHB, BP $32$, El-Alia, $16111$ Bab-
Ezzouar, Alger, Algérie<EMAIL_ADDRESS><EMAIL_ADDRESS>David Marín
Departament de Matemàtiques, Edifici Cc, Universitat Autònoma de Barcelona,
08193 Cerdanyola del Vallès (Barcelona), Spain. Centre de Recerca Matemàtica,
Edifici Cc, Campus de Bellaterra, 08193 Cerdanyola del Vallès (Barcelona),
Spain<EMAIL_ADDRESS>
(August 27, 2024)
###### Abstract
Let $d\geq 2$ be an integer. The set $\mathbf{F}(d)$ of foliations of degree
$d$ on the complex projective plane can be identified with a Zariski’s open
set of a projective space of dimension $d^{2}+4d+2$ on which
$\mathrm{Aut}(\mathbb{P}^{2}_{\mathbb{C}})$ acts. We show that there are
exactly two orbits $\mathcal{O}(\mathcal{F}_{1}^{d})$ and
$\mathcal{O}(\mathcal{F}_{2}^{d})$ of minimal dimension $6$, necessarily
closed in $\mathbf{F}(d)$. This generalizes known results in degrees $2$ and
$3.$ We deduce that an orbit $\mathcal{O}(\mathcal{F})$ of an element
$\mathcal{F}\in\mathbf{F}(d)$ of dimension $7$ is closed in $\mathbf{F}(d)$ if
and only if $\mathcal{F}_{i}^{d}\not\in\overline{\mathcal{O}(\mathcal{F})}$
for $i=1,2.$ This allows us to show that in any degree $d\geq 3$ there are
closed orbits in $\mathbf{F}(d)$ other than the orbits
$\mathcal{O}(\mathcal{F}_{1}^{d})$ and $\mathcal{O}(\mathcal{F}_{2}^{d}),$
unlike the situation in degree $2.$ On the other hand, we introduce the notion
of the basin of attraction $\mathbf{B}(\mathcal{F})$ of a foliation
$\mathcal{F}\in\mathbf{F}(d)$ as the set of $\mathcal{G}\in\mathbf{F}(d)$ such
that $\mathcal{F}\in\overline{\mathcal{O}(\mathcal{G})}.$ We show that the
basin of attraction $\mathbf{B}(\mathcal{F}_{1}^{d})$, resp.
$\mathbf{B}(\mathcal{F}_{2}^{d})$, contains a quasi-projective subvariety of
$\mathbf{F}(d)$ of dimension greater than or equal to
$\dim\mathbf{F}(d)-(d-1)$, resp. $\dim\mathbf{F}(d)-(d-3)$. In particular, we
obtain that the basin $\mathbf{B}(\mathcal{F}_{2}^{3})$ contains a non-empty
Zariski open subset of $\mathbf{F}(3)$. This is an analog in degree $3$ of a
result on foliations of degree $2$ due to Cerveau, Déserti, Garba Belko and
Meziani.
2010 Mathematics Subject Classification. — 37F75, 32S65, 32M25, 32M05.
###### keywords:
foliation, singularity, inflection point, orbit, isotropy group
## Introduction
The set $\mathbf{F}(d)$ of holomorphic foliations of degree $d$ on
$\mathbb{P}^{2}_{\mathbb{C}}$ is identified with a Zariski open subset of the
projective space $\mathbb{P}_{\mathbb{C}}^{\hskip 0.56905ptd^{2}+4d+2}$. We
are interested here in the action of the group
$\mathrm{Aut}(\mathbb{P}^{2}_{\mathbb{C}})=\mathrm{PGL}_{3}(\mathbb{C})$ on
$\mathbf{F}(d).$ We generalize to arbitrary degree some results known in small
degrees [9, 1, 5] on this action.
For $\mathcal{F}\in\mathbf{F}(d)$, we will respectively denote by
$\mathcal{O}(\mathcal{F})$ and $\mathrm{Iso}(\mathcal{F})$ the orbit and the
isotropy group of $\mathcal{F}$ under the action of
$\mathrm{Aut}(\mathbb{P}^{2}_{\mathbb{C}}),$ i.e.
$\displaystyle\mathcal{O}(\mathcal{F}):=\\{\varphi^{*}\mathcal{F}\in\mathbf{F}(d)\hskip
2.84526pt|\hskip
2.84526pt\varphi\in\mathrm{Aut}(\mathbb{P}^{2}_{\mathbb{C}})\\}$ and
$\displaystyle\mathrm{Iso}(\mathcal{F}):=\\{\varphi\in\mathrm{Aut}(\mathbb{P}^{2}_{\mathbb{C}})\hskip
2.84526pt|\hskip 2.84526pt\varphi^{*}\mathcal{F}=\mathcal{F}\\}.$
$\mathcal{O}(\mathcal{F})$ is a Zariski irreducible subset of $\mathbf{F}(d)$
and $\mathrm{Iso}(\mathcal{F})$ is an algebraic subgroup of
$\mathrm{Aut}(\mathbb{P}^{2}_{\mathbb{C}}).$
Following [15] we will say that a foliation of $\mathbf{F}(d)$ is convex if
its leaves other than straight lines have no inflection points. We will denote
by $\mathbf{FC}(d)$ the subset of $\mathbf{F}(d)$ consisting of convex
foliations, which is a Zariski closed subset of $\mathbf{F}(d).$
According to [7, Proposition 2.2] every foliation of degree $0$ or $1$ is
convex. For $d\geq 2$, $\mathbf{FC}(d)$ is a proper closed subset of
$\mathbf{F}(d)$ and it contains the foliation $\mathcal{F}_{1}^{d}$ defined in
the affine chart $(x,y)$ by the $1$-form (see [3, page 75])
$\omega_{1}^{d}=y^{d}\mathrm{d}x+x^{d}(x\mathrm{d}y-y\mathrm{d}x).$
We know from [9, Proposition 2.3] that if $\mathcal{F}$ is an element of
$\mathbf{F}(d)$ with $d\geq 2,$ then the dimension of
$\mathcal{O}(\mathcal{F})$ is at least $6,$ or equivalently, the dimension of
$\mathrm{Iso}(\mathcal{F})$ is at most $2$. In addition these bounds are
attained by the convex foliation $\mathcal{F}_{1}^{d}$ and the non convex
foliation $\mathcal{F}_{2}^{d}$ defined by the $1$-form (see [3])
$\omega_{2}^{d}=x^{d}\mathrm{d}x+y^{d}(x\mathrm{d}y-y\mathrm{d}x).$
The main result of this paper is the following.
###### Theorem A.
Let $d$ be an integer greater than or equal to $2$ and let $\mathcal{F}$ be an
element of $\mathbf{F}(d).$ Assume that the isotropy group
$\mathrm{Iso}(\mathcal{F})$ of $\mathcal{F}$ has dimension $2$. Then
$\mathcal{F}$ is linearly conjugated to one of the two foliations
$\mathcal{F}_{1}^{d}$ and $\mathcal{F}_{2}^{d}$ defined respectively by the
$1$-forms
* 1\.
$\omega_{1}^{d}=y^{d}\mathrm{d}x+x^{d}(x\mathrm{d}y-y\mathrm{d}x);$
* 2\.
$\omega_{2}^{d}=x^{d}\mathrm{d}x+y^{d}(x\mathrm{d}y-y\mathrm{d}x).$
In other words, $\mathcal{O}(\mathcal{F}_{1}^{d})$ and
$\mathcal{O}(\mathcal{F}_{2}^{d})$ are the only orbits of dimension $6$. They
are closed in $\mathbf{F}(d).$ Moreover we have
$\displaystyle\mathrm{Iso}(\mathcal{F}_{1}^{d})=\left\\{\left(\frac{\alpha^{d-1}x}{1+\beta
x},\frac{\alpha^{d}y}{1+\beta x}\right)\hskip 2.84526pt\Big{|}\hskip
2.84526pt\alpha\in\mathbb{C}^{*},\hskip 2.84526pt\beta\in\mathbb{C}\right\\},$
$\displaystyle\mathrm{Iso}(\mathcal{F}_{2}^{d})=\left\\{\left(\frac{\alpha^{d+1}x}{1+\beta
x},\frac{\alpha^{d}y}{1+\beta x}\right)\hskip 2.84526pt\Big{|}\hskip
2.84526pt\alpha\in\mathbb{C}^{*},\hskip 2.84526pt\beta\in\mathbb{C}\right\\};$
these two groups are not conjugated.
This theorem is a generalization in arbitrary degree of previous results on
foliations of degrees $d=2$ ([9, Proposition 2.7]) and $d=3$ ([1, Theorem 10],
[5, Corollary B]).
We also obtain the following corollary, which generalizes [5, Corollary 3.9]:
###### Corollary B.
Let $d$ be an integer greater than or equal to $2$ and let $\mathcal{F}$ be an
element of $\mathbf{F}(d).$ If
$\dim\mathcal{O}(\mathcal{F})\leq\leavevmode\nobreak\ 7$, then
$\overline{\mathcal{O}(\mathcal{F})}\subset\mathcal{O}(\mathcal{F})\cup\mathcal{O}(\mathcal{F}_{1}^{d})\cup\mathcal{O}(\mathcal{F}_{2}^{d}).$
In particular, when $\dim\mathcal{O}(\mathcal{F})=7,$ the orbit
$\mathcal{O}(\mathcal{F})$ of $\mathcal{F}$ is closed in $\mathbf{F}(d)$ if
and only if $\mathcal{F}_{i}^{d}\not\in\overline{\mathcal{O}(\mathcal{F})}$
for $i=1,2.$
In the spirit of Corollary B we can ask under what condition the closure in
$\mathbf{F}(d)$ of the orbit $\mathcal{O}(\mathcal{F})$ of an element
$\mathcal{F}$ of $\mathbf{F}(d)$ contains the foliations $\mathcal{F}_{1}^{d}$
and $\mathcal{F}_{2}^{d}$, a question that we have already asked and studied
in degree $3$ in [5, Section 3]. In Section §3, we extend (Propositions 3.4
and 3.11) in arbitrary degree $d$ our previous results in [5, Propositions
3.10, 3.12, 3.15, 3.17] concerning this question.
For $\mathcal{F}\in\mathbf{F}(d),$ we call basin of attraction of
$\mathcal{F}$ the subset $\mathbf{B}(\mathcal{F})$ of $\mathbf{F}(d)$ defined
by
$\displaystyle\mathbf{B}(\mathcal{F}):=\\{\mathcal{G}\in\mathbf{F}(d)\hskip
2.84526pt|\hskip
2.84526pt\mathcal{F}\in\overline{\mathcal{O}(\mathcal{G})}\\}.$
It follows from [9, Theorem 2.15] that in degree $2$ the basin
$\mathbf{B}(\mathcal{F}_{1}^{2})$ contains a quasi-projective subvariety of
$\mathbf{F}(2)$ of dimension greater than or equal to $\dim\mathbf{F}(2)-1.$
In Section §3, we establish an analogous result in any degree greater than
$2.$
###### Theorem C (Theorem 3.10).
For any integer $d\geq 2$, the basin of attraction
$\mathbf{B}(\mathcal{F}_{1}^{d})$ of $\mathcal{F}_{1}^{d}$ contains a quasi-
projective subvariety of $\mathbf{F}(d)$ of dimension greater than or equal to
$\dim\mathbf{F}(d)-(d-1).$
Notice that the non-convexity of $\mathcal{F}_{2}^{d}$ and the fact that
$\mathbf{FC}(d)$ is closed in $\mathbf{F}(d)$ imply that
(0.1)
$\displaystyle\mathbf{B}(\mathcal{F}_{2}^{d})\subset\mathbf{F}(d)\setminus\mathbf{FC}(d).$
In degree $2$, according to [9, Theorem 3], inclusion (0.1) is an equality:
(0.2)
$\displaystyle\mathbf{B}(\mathcal{F}_{2}^{2})=\mathbf{F}(2)\setminus\mathbf{FC}(2).$
It follows in particular from equality (0.2) that the basin
$\mathbf{B}(\mathcal{F}_{2}^{2})$ is a Zariski open subset of $\mathbf{F}(2).$
For $d\geq 3$ we show the following result.
###### Theorem D (Theorem 3.18).
In any degree $d\geq 3$, the basin of attraction
$\mathbf{B}(\mathcal{F}_{2}^{d})$ of $\mathcal{F}_{2}^{d}$ contains a quasi-
projective subvariety of $\mathbf{F}(d)$ of dimension greater than or equal to
$\dim\mathbf{F}(d)-(d-3)$. In particular, the basin
$\mathbf{B}(\mathcal{F}_{2}^{3})$ contains a non-empty Zariski open subset of
$\mathbf{F}(3).$
Along the same order of ideas, we prove the following result.
###### Theorem E (Theorem 3.21).
For any integer $d\geq 2$, the intersection
$\mathbf{B}(\mathcal{F}_{1}^{d})\cap\mathbf{B}(\mathcal{F}_{2}^{d})$ is non-
empty and it contains a quasi-projective subvariety of $\mathbf{F}(d)$ of
dimension equal to $\dim\mathbf{F}(d)-3d.$
By combining equality (0.2) with the classification of C. Favre and J. V.
Pereira of convex foliations of degree two (_cf._ [10, Proposition 7.4] or [6,
Theorem A]), we see that the only closed orbits in $\mathbf{F}(2)$ under the
action of $\mathrm{Aut}(\mathbb{P}^{2}_{\mathbb{C}})$ are those of
$\mathcal{F}_{1}^{2}$ and $\mathcal{F}_{2}^{2}.$ We show in Section §4 that in
any degree $d\geq 3$ there are closed orbits in $\mathbf{F}(d)$ other than the
orbits $\mathcal{O}(\mathcal{F}_{1}^{d})$ and
$\mathcal{O}(\mathcal{F}_{2}^{d}),$ unlike the situation in degree $2.$ More
precisely, we will consider a family of elements of $\mathbf{F}(d)$ which has
been already studied in degree $d=2$ in [9, page 189], namely the family
$(\mathcal{F}_{0}^{d}(\lambda))_{\lambda\in\mathbb{C}^{*}}$ of foliations of
degree $d$ on $\mathbb{P}^{2}_{\mathbb{C}}$ defined by the $1$-form
$\omega_{0}^{d}(\lambda)=x\mathrm{d}y-\lambda y\mathrm{d}x+y^{d}\mathrm{d}y.$
We will see that, for $\lambda=1,$ $\mathcal{F}_{0}^{d}(1)$ is linearly
conjugated to the foliation $\mathcal{F}_{1}^{d}$ and that, for any
$\lambda\neq 1,$
$\dim\mathcal{O}\big{(}\mathcal{F}_{0}^{d}(\lambda)\big{)}=7.$ Moreover, we
will show (Proposition 4.2) that the orbit
$\mathcal{O}\big{(}\mathcal{F}_{0}^{d}(\lambda)\big{)}$ is closed for any
$d\geq 3$ and $\lambda=-\frac{1}{d-1}$, resp. for any $d\in\\{3,4,5\\}$ and
any $\lambda\in\mathbb{C}^{*},$ and we conjecture that it is so for any $d\geq
6$ and any $\lambda\in\mathbb{C}^{*}$ (see Conjectures 1 and 2).
## 1 Some definitions and notations
### 1.1 Singularities and local invariants
A degree $d$ holomorphic foliation $\mathcal{F}$ on
$\mathbb{P}^{2}_{\mathbb{C}}$ is defined in homogeneous coordinates $[x:y:z]$
by a $1$-form
$\omega=a(x,y,z)\mathrm{d}x+b(x,y,z)\mathrm{d}y+c(x,y,z)\mathrm{d}z,$
where $a,$ $b$ and $c$ are homogeneous polynomials of degree $d+1$ without
common factor and satisfying the Euler condition $i_{\mathrm{R}}\omega=0$,
where
$\mathrm{R}=x\frac{\partial{}}{\partial{x}}+y\frac{\partial{}}{\partial{y}}+z\frac{\partial{}}{\partial{z}}$
denotes the radial vector field and $i_{\mathrm{R}}$ is the interior product
by $\mathrm{R}$.
Dually the foliation $\mathcal{F}$ can also be defined by a homogeneous vector
field
$\hskip
14.22636pt\mathrm{Z}=U(x,y,z)\frac{\partial{}}{\partial{x}}+V(x,y,z)\frac{\partial{}}{\partial{y}}+W(x,y,z)\frac{\partial{}}{\partial{z}},$
the coefficients $U,V$ and $W$ are homogeneous polynomials of degree $d$
without common factor. The relation between $\mathrm{Z}$ and $\omega$ is given
by
$\omega=i_{\mathrm{R}}i_{\mathrm{Z}}(\mathrm{d}x\wedge\mathrm{d}y\wedge\mathrm{d}z).$
The singular locus $\mathrm{Sing}\mathcal{F}$ of $\mathcal{F}$ is the
projectivization of the singular locus of $\omega$
$\mathrm{Sing}\,\omega=\\{(x,y,z)\in\mathbb{C}^{3}\,|\,a(x,y,z)=b(x,y,z)=c(x,y,z)=0\\}.$
Let $\mathcal{C}\subset\mathbb{P}^{2}_{\mathbb{C}}$ be an algebraic curve with
homogeneous equation $F(x,y,z)=0.$ We say that $\mathcal{C}$ is an invariant
curve by $\mathcal{F}$ if $\mathcal{C}\smallsetminus\mathrm{Sing}\mathcal{F}$
is a union of (ordinary) leaves of the regular foliation
$\mathcal{F}|_{\mathbb{P}^{2}_{\mathbb{C}}\smallsetminus\mathrm{Sing}\mathcal{F}}$.
In algebraic terms, this is equivalent to require that the $2$-form
$\omega\wedge\mathrm{d}F$ is divisible by $F$, i.e. it vanishes along each
irreducible component of $\mathcal{C}.$
Let $p$ be an arbitrary point of $\mathcal{C}.$ When each irreducible
component of $\mathcal{C}$ passing through $p$ is not $\mathcal{F}$-invariant,
we define the tangency order $\mathrm{Tang}(\mathcal{F},\mathcal{C},p)$ of
$\mathcal{F}$ with $\mathcal{C}$ at $p$ as follows. We fix a local chart
$(\mathrm{u},\mathrm{v})$ such that $p=(0,0)$; let
$f(\mathrm{u},\mathrm{v})=0$ be a reduced local equation of $\mathcal{C}$ in a
neighborhood of $p$ and let $\mathrm{X}$ be a vector field defining the germ
of $\mathcal{F}$ at $p.$ We denote by $\mathrm{X}(f)$ the Lie derivative of
$f$ along $\mathrm{X}$ and by $\langle f,\mathrm{X}(f)\rangle$ the ideal of
$\mathbb{C}\\{\mathrm{u},\mathrm{v}\\}$ generated by $f$ and $\mathrm{X}(f)$.
Then
$\mathrm{Tang}(\mathcal{F},\mathcal{C},p)=\dim_{\mathbb{C}}\frac{\mathbb{C}\\{\mathrm{u},\mathrm{v}\\}}{\langle
f,\mathrm{X}(f)\rangle}.$
Notice that $\mathrm{Tang}(\mathcal{F},\mathcal{C},p)$ coincides with the
intersection multiplicity $(\mathcal{C}.\mathcal{C}^{\prime})_{p}$ at $p$ of
the two algebraic curves $\mathcal{C}=\\{F=0\\}$ and
$\mathcal{C}^{\prime}=\\{\mathrm{Z}(F)=0\\}.$ Moreover,
$\mathrm{Tang}(\mathcal{F},\mathcal{C},p)<+\infty$ by the non-invariance of
the irreducible components of $\mathcal{C}$ passing through $p.$ By
convention, we put $\mathrm{Tang}(\mathcal{F},\mathcal{C},p)=+\infty$ if there
is at least one irreducible component of $\mathcal{C}$ invariant by
$\mathcal{F}$ and passing through $p.$
Let us recall some local notions attached to the pair $(\mathcal{F},s)$, where
$s\in\mathrm{Sing}\mathcal{F}$. The germ of $\mathcal{F}$ at $s$ is defined,
up to multiplication by a unity in the local ring $\mathcal{O}_{s}$ at $s$, by
a vector field
$\mathrm{X}=A(\mathrm{u},\mathrm{v})\frac{\partial{}}{\partial{\mathrm{u}}}+B(\mathrm{u},\mathrm{v})\frac{\partial{}}{\partial{\mathrm{v}}}$.
The algebraic multiplicity $\nu(\mathcal{F},s)$ of $\mathcal{F}$ at $s$ is
given by
$\nu(\mathcal{F},s)=\min\\{\nu(A,s),\nu(B,s)\\},$
where $\nu(g,s)$ denotes the algebraic multiplicity of the function $g$ at
$s.$ Let us denote by $\mathfrak{L}_{s}(\mathcal{F})$ the family of straight
lines through $s$ which are not invariant by $\mathcal{F}.$ For any line
$\ell$ of $\mathfrak{L}_{s}(\mathcal{F}),$ we have the inequalities
$1\leq\mathrm{Tang}(\mathcal{F},\ell_{s},s)\leq d$. This allows us to
associate to the pair $(\mathcal{F},s)$ the following (invariant) integers
$\displaystyle\hskip
22.76228pt\tau(\mathcal{F},s)=\min\\{\mathrm{Tang}(\mathcal{F},\ell,s)\hskip
2.84526pt|\hskip 2.84526pt\ell\in\mathfrak{L}_{s}(\mathcal{F})\\},$
$\displaystyle\hskip
14.22636pt\kappa(\mathcal{F},s)=\max\\{\mathrm{Tang}(\mathcal{F},\ell,s)\hskip
2.84526pt|\hskip 2.84526pt\ell\in\mathfrak{L}_{s}(\mathcal{F})\\}.$
The invariant $\tau(\mathcal{F},s)$ represents the tangency order of
$\mathcal{F}$ with a generic line passing through $s.$ It is easy to see that
$\tau(\mathcal{F},s)=\min\\{k\geq 1\hskip 2.84526pt|\hskip
2.84526pt\det(J^{k}_{s}\,\mathrm{X},\mathrm{R}_{s})\not\equiv
0\\}\geq\nu(\mathcal{F},s),$
where $J^{k}_{s}\,\mathrm{X}$ denotes the $k$-jet of $\mathrm{X}$ at $s$ and
$\mathrm{R}_{s}$ is the radial vector field centered at $s.$ The Milnor number
of $\mathcal{F}$ at $s$ is the integer
$\mu(\mathcal{F},s)=\dim_{\mathbb{C}}\mathcal{O}_{s}/\langle A,B\rangle,$
where $\langle A,B\rangle$ denotes the ideal of $\mathcal{O}_{s}$ generated by
$A$ and $B$.
The singularity $s$ is called radial of order $n-1$ if $\nu(\mathcal{F},s)=1$
and $\tau(\mathcal{F},s)=n.$
The singularity $s$ is called non-degenerate if $\mu(\mathcal{F},s)=1,$ or
equivalently if the Jacobian matrix of $\mathrm{X}$ at $s$, denoted by
$\mathrm{Jac}\hskip 0.28453pt\mathrm{X}(s),$ possesses two nonzero eigenvalues
$\lambda,\mu.$ In this case, the quantity
$\displaystyle\mathrm{BB}(\mathcal{F},s)=\frac{\mathrm{tr}^{2}(\mathrm{Jac}\hskip
0.28453pt\mathrm{X}(s))}{\det(\mathrm{Jac}\hskip
0.28453pt\mathrm{X}(s))}=\frac{\lambda}{\mu}+\frac{\mu}{\lambda}+2$
is called the Baum-Bott index of $\mathcal{F}$ at $s,$ see [2].
We will say that the singularity $s$ is quasi-radial of order $n-1$ if
$\mu(\mathcal{F},s)=1,$ $\mathrm{BB}(\mathcal{F},s)=4$ and
$\kappa(\mathcal{F},s)=n.$ In the sequel we will denote by
$\mathrm{QRad}(\mathcal{F},n-1)$ the set of quasi-radial singularities of
$\mathcal{F}$ of order $n-1$ and by $\widehat{\mathrm{QRad}}(\mathcal{F},n-1)$
the subset of $\mathrm{Sing}(\mathcal{F})\times\mathfrak{L}_{s}(\mathcal{F})$
defined by
$\displaystyle\widehat{\mathrm{QRad}}(\mathcal{F},n-1):=\Big{\\{}(s,\ell)\in\mathrm{Sing}(\mathcal{F})\times\mathfrak{L}_{s}(\mathcal{F})\hskip
2.84526pt|\hskip 2.84526pt\mu(\mathcal{F},s)=1,\hskip
2.84526pt\mathrm{BB}(\mathcal{F},s)=4,\hskip
2.84526pt\mathrm{Tang}(\mathcal{F},\ell,s)=n\Big{\\}}.$
###### Remark 1.1.
Every radial singularity of order $n-1$ of a foliation $\mathcal{F}$ of degree
$d\geq 2$ on $\mathbb{P}^{2}_{\mathbb{C}}$ is quasi-radial of order $n-1.$ The
converse is false: for instance, for the foliation defined in the affine chart
$z=1$ by the $1$-form
$(x+y)\mathrm{d}y-y\mathrm{d}x+(x^{n}+y^{d})\mathrm{d}x,$ with
$n\in\\{2,3,\ldots,d\\},$ the point $[0:0:1]$ is a quasi-radial singularity of
order $n-1,$ but it is not radial.
### 1.2 Inflection points
Let us consider a foliation $\mathcal{F}$ of degree $d$ on
$\mathbb{P}^{2}_{\mathbb{C}}$ and let $p$ be a regular point of $\mathcal{F}$.
Let us denote by $\mathrm{T}^{\mathbb{P}}_{p}\mathcal{F}$ the tangent line to
the leaf of $\mathcal{F}$ passing through $p$; it is the straight line of
$\mathbb{P}^{2}_{\mathbb{C}}$ passing through $p$ with direction
$\mathrm{T}_{p}\mathcal{F}.$ If $k\in\\{2,\ldots,d\\},$ we will say that $p$
is a (transverse) inflection point of order $k-1$ of $\mathcal{F}$ if
$\mathrm{Tang}(\mathcal{F},\mathrm{T}^{\mathbb{P}}_{p}\mathcal{F},p)=k,$ in
which case the line $\mathrm{T}^{\mathbb{P}}_{p}\mathcal{F}$ is not invariant
by $\mathcal{F}.$ When $\mathrm{T}^{\mathbb{P}}_{p}\mathcal{F}$ is
$\mathcal{F}$-invariant, the point $p$ will be called a trivial inflection
point of $\mathcal{F}$. If we denote by $\mathrm{Inv}(\mathcal{F})$ the set of
invariant lines of $\mathcal{F},$ then the set of trivial inflection points of
$\mathcal{F}$ is precisely
$\mathrm{Inv}(\mathcal{F})\setminus\mathrm{Sing}(\mathcal{F}).$ In the sequel,
we will denote by $\mathrm{Flex}(\mathcal{F})$ the set of inflection points of
$\mathcal{F}$ and by $\mathrm{Flex}(\mathcal{F},k-1)$ the subset of
$\mathrm{Flex}(\mathcal{F})$ consisting of transverse inflection points of
$\mathcal{F}$ of order $k-1,$ i.e.
$\displaystyle\mathrm{Flex}(\mathcal{F},k-1):=\Big{\\{}p\in\mathbb{P}^{2}_{\mathbb{C}}\hskip
2.84526pt|\hskip 2.84526ptp\not\in\mathrm{Sing}(\mathcal{F}),\hskip
2.84526pt\mathrm{Tang}(\mathcal{F},\mathrm{T}^{\mathbb{P}}_{p}\mathcal{F},p)=k\Big{\\}}.$
Let us recall the notion of inflection divisor of $\mathcal{F}$, introduced by
Pereira [16], which allows to determine the set $\mathrm{Flex}(\mathcal{F}).$
Let $\mathrm{Z}$ be a homogeneous vector field of degree $d$ on
$\mathbb{C}^{3}$ defining $\mathcal{F}.$ The inflection divisor of
$\mathcal{F}$, denoted by $\mathrm{I}_{\mathcal{F}}$, is the divisor of
$\mathbb{P}^{2}_{\mathbb{C}}$ defined by the homogeneous equation
(1.1) $\left|\begin{array}[]{ccc}x&\mathrm{Z}(x)&\mathrm{Z}^{2}(x)\\\
y&\mathrm{Z}(y)&\mathrm{Z}^{2}(y)\\\
z&\mathrm{Z}(z)&\mathrm{Z}^{2}(z)\end{array}\right|=0.$
According to [16], $\mathrm{I}_{\mathcal{F}}$ satisfies the following
properties:
1. 1.
The support of $\mathrm{I}_{\mathcal{F}}$ is exactly the closure of the set
$\mathrm{Flex}(\mathcal{F})$ of inflection points of $\mathcal{F}.$ More
precisely, $\mathrm{I}_{\mathcal{F}}$ can be decomposed as
$\mathrm{I}_{\mathcal{F}}=\mathrm{I}_{\mathcal{F}}^{\mathrm{inv}}+\mathrm{I}_{\mathcal{F}}^{\hskip
0.56905pt\mathrm{tr}},$ where the support of
$\mathrm{I}_{\mathcal{F}}^{\mathrm{inv}}$ is the set
$\mathrm{Inv}(\mathcal{F})$ of $\mathcal{F}$-invariant lines and the support
of $\mathrm{I}_{\mathcal{F}}^{\hskip 0.56905pt\mathrm{tr}}$ is the closure of
the set of transverse inflection points of $\mathcal{F}.$
2. 2.
If $\mathcal{C}$ is an algebraic curve invariant by $\mathcal{F},$ then
$\mathcal{C}\subset\mathrm{I}_{\mathcal{F}}$ if and only if
$\mathcal{C}\subset\mathrm{Inv}(\mathcal{F}).$
3. 3.
The degree of the divisor $\mathrm{I}_{\mathcal{F}}$ is $3d.$
The foliation $\mathcal{F}$ will be called convex if its inflection divisor
$\mathrm{I}_{\mathcal{F}}$ is totally invariant by $\mathcal{F}$, i.e. if
$\mathrm{I}_{\mathcal{F}}$ is a product of invariant lines.
## 2 Description of the foliations $\mathcal{F}$ of degree greater than or
equal to $2$ such that $\dim\mathcal{O}(\mathcal{F})=6$
Recall that the foliations $\mathcal{F}_{1}^{d}$ and $\mathcal{F}_{2}^{d}$ are
respectively defined in the affine chart $z=1$ by the $1$-forms
$\displaystyle\omega_{1}^{d}=y^{d}\mathrm{d}x+x^{d}(x\mathrm{d}y-y\mathrm{d}x)$
and
$\displaystyle\omega_{2}^{d}=x^{d}\mathrm{d}x+y^{d}(x\mathrm{d}y-y\mathrm{d}x).$
The foliation $\mathcal{F}_{1}^{d}$ is convex with inflection divisor
$\mathrm{I}_{\mathcal{F}_{1}^{d}}=\mathrm{I}_{\mathcal{F}_{1}^{d}}^{\hskip
0.56905pt\mathrm{inv}}=x^{d+1}y^{2d-1}$ and it has two singular points
$s_{1}=[0:0:1]$ and $s_{2}=[0:1:0]$; the singularity $s_{1}$ has maximal
algebraic multiplicity $d$ and $s_{2}$ is radial of maximal order $d-1.$ The
foliation $\mathcal{F}_{2}^{d}$ is not convex with invariant inflection
divisor $\mathrm{I}_{\mathcal{F}_{2}^{d}}^{\hskip
0.56905pt\mathrm{inv}}=x^{2d+1}$ and transverse inflection divisor
$\mathrm{I}_{\mathcal{F}_{2}^{d}}^{\hskip 0.56905pt\mathrm{tr}}=y^{d-1}.$ The
singular locus $\mathrm{Sing}(\mathcal{F}_{2}^{d})$ is reduced to the point
$s_{1}=[0:0:1]$; moreover $\nu(\mathcal{F}_{2}^{d},s_{1})=d.$ We note that the
$1$-forms $\dfrac{\omega_{1}^{d}}{x^{2}y^{d}}$ and
$\dfrac{\omega_{2}^{d}}{x^{d+2}}$ are closed and they respectively admit as
first integrals
$\displaystyle\frac{1}{d-1}\left(\frac{x}{y}\right)^{d-1}+\frac{1}{x}$ and
$\displaystyle\frac{1}{d+1}\left(\frac{y}{x}\right)^{d+1}-\frac{1}{x};$
this allows to check that
$\displaystyle\mathrm{Iso}(\mathcal{F}_{1}^{d})=\left\\{\left(\frac{\alpha^{d-1}x}{1+\beta
x},\frac{\alpha^{d}y}{1+\beta x}\right)\hskip 2.84526pt\Big{|}\hskip
2.84526pt\alpha\in\mathbb{C}^{*},\hskip 2.84526pt\beta\in\mathbb{C}\right\\}$
and
$\displaystyle\mathrm{Iso}(\mathcal{F}_{2}^{d})=\left\\{\left(\frac{\alpha^{d+1}x}{1+\beta
x},\frac{\alpha^{d}y}{1+\beta x}\right)\hskip 2.84526pt\Big{|}\hskip
2.84526pt\alpha\in\mathbb{C}^{*},\hskip 2.84526pt\beta\in\mathbb{C}\right\\}.$
In particular, $\dim\mathrm{Iso}(\mathcal{F}_{i}^{d})=2$ for $i=1,2.$ Thus the
orbits $\mathcal{O}(\mathcal{F}_{1}^{d})$ and
$\mathcal{O}(\mathcal{F}_{2}^{d})$ are both of dimension $6,$ which is the
minimal dimension possible in any degree $d\geq 2$ ([9, Proposition 2.3]).
Theorem A announced in the Introduction shows that the orbits
$\mathcal{O}(\mathcal{F}_{1}^{d})$ and $\mathcal{O}(\mathcal{F}_{2}^{d})$ are
the only orbits having minimal dimension $6.$ The goal of this section is to
prove this theorem.
Let us denote by $\chi(\mathbb{P}^{2}_{\mathbb{C}})$ the Lie algebra of
holomorphic vector fields on $\mathbb{P}^{2}_{\mathbb{C}}$:
$\chi(\mathbb{P}^{2}_{\mathbb{C}})$ is of course the Lie algebra of the
automorphism group of $\mathbb{P}^{2}_{\mathbb{C}}.$ Let $\mathcal{F}$ be a
foliation on $\mathbb{P}^{2}_{\mathbb{C}}$ and let $\mathrm{X}$ be an element
of $\chi(\mathbb{P}^{2}_{\mathbb{C}}).$ Following [9] we will say that
$\mathrm{X}$ is a symmetry of the foliation $\mathcal{F}$ if the flow
$\exp(t\mathrm{X})$ is, for each $t,$ in the isotropy group
$\mathrm{Iso}(\mathcal{F})$ of $\mathcal{F}.$ If $\omega$ defines
$\mathcal{F}$ in an affine chart, $\mathrm{X}$ is a symmetry of $\mathcal{F}$
if and only if $\mathrm{L}_{\mathrm{X}}\omega\wedge\omega=0,$ where
$\mathrm{L}_{\mathrm{X}}\omega$ denotes the Lie derivative of $\omega$ along
$\mathrm{X}.$
###### Lemma 2.1.
Let $\mathcal{F}$ be a foliation of degree $d$ on
$\mathbb{P}^{2}_{\mathbb{C}}$ and let $\mathrm{X}$ be a symmetry of
$\mathcal{F}.$ Assume that there is an affine chart
$\mathbb{C}^{2}\subset\mathbb{P}^{2}_{\mathbb{C}}$ such that the vector field
$\mathrm{X}$ is affine (i.e. $\deg X\leq 1$) and let $\omega$ be a $1$-form
defining $\mathcal{F}$ in this chart. Then there is a constant
$\lambda\in\mathbb{C}$ such that
$\mathrm{L}_{\mathrm{X}}\omega=\lambda\omega.$
###### Proof.
We will use an argument similar to one in [9, Proposition 2.5]. Since
$\mathrm{L}_{\mathrm{X}}\omega\wedge\omega=0$ and $\omega$ has isolated
singularities, the De Rham-Saito division theorem (_cf._ [17] or [8,
Proposition 1.14]) ensures the existence of a holomorphic function $g$ on
$\mathbb{C}^{2}$ such that $\mathrm{L}_{\mathrm{X}}\omega=g\omega.$ The
$1$-form $\omega$ and the vector field $\mathrm{X}$ being polynomials,
$\mathrm{L}_{\mathrm{X}}\omega$ is also polynomial; therefore $g$ is rational
and holomorphic on $\mathbb{C}^{2}$ hence polynomial. The vector field
$\mathrm{X}$ being affine we have
$\deg\mathrm{L}_{\mathrm{X}}\omega\leq\deg\omega$; the equality
$\mathrm{L}_{\mathrm{X}}\omega=g\omega$ implies that $g$ is constant. ∎
If $\mathcal{F}$ is a foliation on $\mathbb{P}^{2}_{\mathbb{C}}$, we will
denote by $\mathfrak{iso}(\mathcal{F})$ the Lie algebra of the algebraic group
$\mathrm{Iso}(\mathcal{F})$; $\mathfrak{iso}(\mathcal{F})$ is a Lie subalgebra
of $\chi(\mathbb{P}^{2}_{\mathbb{C}})$ and it consists of symmetries of
$\mathcal{F}.$ We know from [9, Proposition 2.5] that if
$\dim\mathfrak{iso}(\mathcal{F})=2$ then $\mathfrak{iso}(\mathcal{F})$ is
affine, i.e. generated by two vector fields $\mathrm{X}$ and $\mathrm{Y}$ such
that $[\mathrm{X},\mathrm{Y}]=\mathrm{Y}.$ The following lemma classifies the
affine Lie subalgebras of $\chi(\mathbb{P}^{2}_{\mathbb{C}})$ and it will be
used to prove Theorem A.
###### Lemma 2.2.
Every affine Lie subalgebra of $\chi(\mathbb{P}^{2}_{\mathbb{C}})$ is linearly
conjugated to one of the following models
* ($\mathfrak{a}$)
$\big{\langle}\gamma\,x\frac{\partial}{\partial x}+y\frac{\partial}{\partial
y},\frac{\partial}{\partial y}\big{\rangle}$ with $\gamma\in\mathbb{C}^{*};$
* ($\mathfrak{b}$)
$\big{\langle}y\frac{\partial}{\partial y},\frac{\partial}{\partial
y}\big{\rangle};$
* ($\mathfrak{c}$)
$\big{\langle}\frac{\partial}{\partial x}+y\frac{\partial}{\partial
y},\frac{\partial}{\partial y}\big{\rangle};$
* ($\mathfrak{d}$)
$\big{\langle}x\frac{\partial}{\partial x}+(x+y)\frac{\partial}{\partial
y},\frac{\partial}{\partial y}\big{\rangle};$
* ($\mathfrak{e}$)
$\big{\langle}x\frac{\partial}{\partial x}+2y\frac{\partial}{\partial
y},\frac{\partial}{\partial x}+x\frac{\partial}{\partial y}\big{\rangle}.$
###### Proof.
Let $\mathfrak{g}$ be an affine Lie subalgebra of
$\chi(\mathbb{P}^{2}_{\mathbb{C}}).$ Then there exist $\mathrm{X}$ and
$\mathrm{Y}$ in $\chi(\mathbb{P}^{2}_{\mathbb{C}})$ such that
$\mathfrak{g}=\langle\mathrm{X},\mathrm{Y}\rangle$ and
$[\mathrm{X},\mathrm{Y}]=\mathrm{Y}.$ Fixing homogeneous coordinates $[x:y:z]$
in $\mathbb{P}^{2}_{\mathbb{C}}$ we have an isomorphism of Lie algebras
$\tau\hskip
2.84544pt\colon\mathfrak{sl}_{3}(\mathbb{C})\rightarrow\chi(\mathbb{P}^{2}_{\mathbb{C}})$
defined, for $A\in\mathfrak{sl}_{3}(\mathbb{C}),$ by
$\tau(A)=(x\hskip 5.69054pty\hskip
5.69054ptz)A\left(\begin{array}[]{c}\vspace{0.5mm}\frac{\partial}{\partial
x}\\\ \vspace{0.5mm}\frac{\partial}{\partial y}\\\
\vspace{0.5mm}\frac{\partial}{\partial z}\end{array}\right).$
Notice that if $A=\left(\begin{array}[]{ccc}a_{11}&a_{12}&a_{13}\\\
a_{21}&a_{22}&a_{23}\\\
a_{31}&a_{32}&a_{33}\end{array}\right)\in\mathfrak{sl}_{3}(\mathbb{C}),$ then
in the affine chart $z=1$ the vector field
$\tau(A)\in\chi(\mathbb{P}^{2}_{\mathbb{C}})$ writes as
$\displaystyle\left(a_{31}+(a_{11}-a_{33})x+a_{21}y-a_{13}x^{2}-a_{23}xy\right)\frac{\partial}{\partial
x}+\left(a_{32}+a_{12}x+(a_{22}-a_{33})y-a_{13}xy-
a_{23}y^{2}\right)\frac{\partial}{\partial y}.$
Let $M$ and $N$ be the matrices of $\mathfrak{sl}_{3}(\mathbb{C})$ associated
to the vector fields $\mathrm{X}$ and $\mathrm{Y}$ respectively, i.e.
$M=\tau^{-1}(\mathrm{X})$ and $N=\tau^{-1}(\mathrm{Y})$. Then the Lie bracket
$[\mathrm{X},\mathrm{Y}]$ corresponds to $[M,N]:=MN-NM$ and therefore
$[M,N]=N.$
Let us write $M=\left(\begin{array}[]{ccc}-m_{22}-m_{33}&m_{12}&m_{13}\\\
m_{21}&m_{22}&m_{23}\\\ m_{31}&m_{32}&m_{33}\end{array}\right).$ Taking into
account the possible Jordan forms of a matrix of
$\mathfrak{sl}_{3}(\mathbb{C}),$ it suffices us to treat the following
possibilities
$\displaystyle N=\left(\begin{array}[]{ccc}-a-b&0&0\\\ 0&a&0\\\
0&0&b\end{array}\right),$ $\displaystyle
N=\left(\begin{array}[]{ccc}-2c&0&0\\\ 0&c&0\\\ 0&1&c\end{array}\right),$
$\displaystyle N=\left(\begin{array}[]{ccc}0&0&0\\\ 0&0&0\\\
0&1&0\end{array}\right),$ $\displaystyle N=\left(\begin{array}[]{ccc}0&0&0\\\
1&0&0\\\ 0&1&0\end{array}\right).$
where $a,b\in\mathbb{C},c\in\mathbb{C}^{*},$ with $(a,b)\neq(0,0).$
1. If $N=\left(\begin{array}[]{ccc}-a-b&0&0\\\ 0&a&0\\\ 0&0&b\end{array}\right)$ then the equality $[M,N]=N$ implies that $a=b=0$: contradiction.
2. If $N=\left(\begin{array}[]{ccc}-2c&0&0\\\ 0&c&0\\\ 0&1&c\end{array}\right)$ then the $(1,1)$ coefficient of the matrix $[M,N]-N$ is equal to $2c$ and is therefore nonzero: contradiction.
3. Assume that $N=\left(\begin{array}[]{ccc}0&0&0\\\ 0&0&0\\\ 0&1&0\end{array}\right)$; the equality $[M,N]=N$ then leads to $M=\left(\begin{array}[]{ccc}1-2m_{33}&m_{12}&0\\\ 0&m_{33}-1&0\\\ m_{31}&m_{32}&m_{33}\end{array}\right)$. Up to replacing $M$ by $M-m_{32}N$ we can assume that $m_{32}=0.$ Now we will distinguish several eventualities:
3.1. When $(3m_{33}-1)(3m_{33}-2)\neq 0$ the matrix
$P=\left(\begin{array}[]{ccc}3m_{33}-1&m_{12}&0\\\ 0&3m_{33}-2&0\\\
-m_{31}&-m_{31}m_{12}&3m_{33}-2\end{array}\right)$ commutes with $N$ and
$P^{-1}MP=\left(\begin{array}[]{ccc}1-2m_{33}&0&0\\\ 0&m_{33}-1&0\\\
0&0&m_{33}\end{array}\right)$. Thus $\mathfrak{g}$ is linearly conjugated to
$\displaystyle\big{\langle}\tau(P^{-1}MP),\tau(N)\big{\rangle}=\big{\langle}(1-3m_{33})x\tfrac{\partial}{\partial
x}-y\tfrac{\partial}{\partial y},\tfrac{\partial}{\partial
y}\big{\rangle}=\big{\langle}\gamma\,x\tfrac{\partial}{\partial
x}+y\tfrac{\partial}{\partial y},\tfrac{\partial}{\partial
y}\big{\rangle},\qquad\text{where }\gamma=3m_{33}-1\in\mathbb{C}^{*}.$
3.2. Assume that $m_{33}=\frac{1}{3}.$ If $\delta\in\mathbb{C}^{*}$ then the
matrix $P=\left(\begin{array}[]{ccc}\frac{1}{\delta}&-m_{12}&0\\\ 0&1&0\\\
0&m_{12}m_{31}&1\end{array}\right)$ commutes with $N$ and
$P^{-1}MP=\left(\begin{array}[]{ccc}\frac{1}{3}&0&0\\\ 0&-\frac{2}{3}&0\\\
\frac{m_{31}}{\delta}&0&\frac{1}{3}\end{array}\right)$. As a result
$\mathfrak{g}$ is linearly conjugated to
$\displaystyle\big{\langle}\tau(P^{-1}MP),\tau(N)\big{\rangle}=\big{\langle}\tfrac{m_{31}}{\delta}\tfrac{\partial}{\partial
x}-y\tfrac{\partial}{\partial y},\tfrac{\partial}{\partial
y}\big{\rangle}=\big{\langle}-\tfrac{m_{31}}{\delta}\tfrac{\partial}{\partial
x}+y\tfrac{\partial}{\partial y},\tfrac{\partial}{\partial y}\big{\rangle}.$
The case where $m_{31}=0$ leads to the model $(\mathfrak{b}).$ If $m_{31}\neq
0$ then by taking $\delta=-m_{31}$ we get the model $(\mathfrak{c}).$
3.3. Assume that $m_{33}=\frac{2}{3}.$ If $\delta\in\mathbb{C}^{*}$ then the
matrix $P=\left(\begin{array}[]{ccc}\delta&0&0\\\ 0&1&0\\\ -\delta
m_{31}&-m_{12}m_{31}&1\end{array}\right)$ commutes with $N$ and
$P^{-1}MP=\left(\begin{array}[]{ccc}-\frac{1}{3}&\frac{m_{12}}{\delta}&0\\\
0&-\frac{1}{3}&0\\\ 0&0&\frac{2}{3}\end{array}\right)$. As a consequence
$\mathfrak{g}$ is linearly conjugated to
$\displaystyle\big{\langle}\tau(P^{-1}MP),\tau(N)\big{\rangle}=\big{\langle}-x\tfrac{\partial}{\partial
x}+(\tfrac{m_{12}}{\delta}x-y)\tfrac{\partial}{\partial
y},\tfrac{\partial}{\partial
y}\big{\rangle}=\big{\langle}x\tfrac{\partial}{\partial
x}+(y-\tfrac{m_{12}}{\delta}x)\tfrac{\partial}{\partial
y},\tfrac{\partial}{\partial y}\big{\rangle}.$
The case $m_{12}=0$ leads to the model $(\mathfrak{a})$ with $\gamma=1.$ If
$m_{12}\neq 0$ then by taking $\delta=-m_{12}$ we obtain the model
$(\mathfrak{d}).$
4. Assume that $N=\left(\begin{array}[]{ccc}0&0&0\\\ 1&0&0\\\ 0&1&0\end{array}\right)$; then the equality $[M,N]=N$ implies that $M=\left(\begin{array}[]{ccc}-1&0&0\\\ m_{32}&0&0\\\ m_{31}&m_{32}&1\end{array}\right)$. Up to replacing $M$ by $M-m_{32}N$ we can assume that $m_{32}=0.$ The matrix $P=\left(\begin{array}[]{ccc}1&0&0\\\ 0&1&0\\\ -\frac{m_{31}}{2}&0&1\end{array}\right)$ commutes with $N$ and $P^{-1}MP=\left(\begin{array}[]{ccc}-1&0&0\\\ 0&0&0\\\ 0&0&1\end{array}\right)$. Therefore $\mathfrak{g}$ is linearly conjugated to
$\displaystyle\big{\langle}\tau(P^{-1}MP),\tau(N)\big{\rangle}=\big{\langle}-2x\tfrac{\partial}{\partial
x}-y\tfrac{\partial}{\partial y},y\tfrac{\partial}{\partial
x}+\tfrac{\partial}{\partial
y}\big{\rangle}=\big{\langle}2x\tfrac{\partial}{\partial
x}+y\tfrac{\partial}{\partial y},y\tfrac{\partial}{\partial
x}+\tfrac{\partial}{\partial y}\big{\rangle}.$
By permuting the coordinates $x$ and $y$ we obtain the model $(\mathfrak{e}).$
∎
###### Proof of Theorem A.
Since $\dim\mathfrak{iso}(\mathcal{F})=\dim\mathrm{Iso}(\mathcal{F})=2,$ [9,
Proposition 2.5] implies that $\mathfrak{iso}(\mathcal{F})$ is affine.
Therefore, up to linear conjugation, $\mathfrak{iso}(\mathcal{F})$ is one of
the models ($\mathfrak{a}$)–($\mathfrak{e}$) of Lemma 2.2.
Let $\omega$ be a $1$-form defining $\mathcal{F}$ in the affine chart $z=1$
$\hskip 28.45274pt\omega=A(x,y)\mathrm{d}x+B(x,y)\mathrm{d}y,\quad
A,B\in\mathbb{C}[x,y],\hskip 5.69054pt\gcd(A,B)=1.$
We will study the five possible models ($\mathfrak{a}$)–($\mathfrak{e}$) of
the Lie algebra $\mathfrak{iso}(\mathcal{F})$ and show that $\omega$ is
linearly conjugated to one of the two $1$-forms $\omega_{1}^{d}$ or
$\omega_{2}^{d}.$
1. Assume that $\mathfrak{iso}(\mathcal{F})$ is of one of the types ($\mathfrak{a}$)–($\mathfrak{d}$), i.e. that $\mathfrak{iso}(\mathcal{F})=\big{\langle}\mathrm{X},\mathrm{Y}\big{\rangle}$ where $\mathrm{X}\in\\{\gamma\,x\frac{\partial}{\partial x}+y\frac{\partial}{\partial y},\varepsilon\frac{\partial}{\partial x}+y\frac{\partial}{\partial y},x\frac{\partial}{\partial x}+(x+y)\frac{\partial}{\partial y}\\},\mathrm{Y}=\frac{\partial}{\partial y}$ with $\varepsilon\in\\{0,1\\}$ and $\gamma\in\mathbb{C}^{*}.$ By Lemma 2.1 there exist $\lambda,\mu\in\mathbb{C}$ such that $\mathrm{L}_{\mathrm{X}}\omega=\lambda\omega$ and $\mathrm{L}_{\mathrm{Y}}\omega=\mu\omega.$ Since $\mathrm{L}_{\mathrm{Y}}\mathrm{d}x=\mathrm{d}\mathrm{L}_{\mathrm{Y}}x=0$ and $\mathrm{L}_{\mathrm{Y}}\mathrm{d}y=\mathrm{d}\mathrm{L}_{\mathrm{Y}}y=0,$ we have $\mathrm{L}_{\mathrm{Y}}\omega=\mathrm{Y}(A)\mathrm{d}x+\mathrm{Y}(B)\mathrm{d}y=\frac{\partial A}{\partial y}\mathrm{d}x+\frac{\partial B}{\partial y}\mathrm{d}y.$ Therefore $\mathrm{L}_{\mathrm{Y}}\omega=\mu\omega$ if and only if $\frac{\partial A}{\partial y}=\mu A$ and $\frac{\partial B}{\partial y}=\mu B.$ Since $A,B\in\mathbb{C}[x,y]$ and $\mu\in\mathbb{C},$ it follows that $\mu=0$, $A(x,y)=A(x)$ and $B(x,y)=B(x).$ Thus
$\hskip 28.45274pt\omega=A(x)\mathrm{d}x+B(x)\mathrm{d}y,\quad
A,B\in\mathbb{C}[x],\hskip 5.69054pt\gcd(A,B)=1.$
1.1. Let us consider the case where
$\mathrm{X}=\gamma\,x\frac{\partial}{\partial x}+y\frac{\partial}{\partial y}$
with $\gamma\in\mathbb{C}^{*}.$ We have
$\displaystyle\mathrm{L}_{\mathrm{X}}\omega=\mathrm{X}(A)\mathrm{d}x+A\mathrm{d}\mathrm{X}(x)+\mathrm{X}(B)\mathrm{d}y+B\mathrm{d}\mathrm{X}(y)=\big{(}\gamma\,xA^{\prime}+\gamma\,A\big{)}\mathrm{d}x+\big{(}\gamma\,xB^{\prime}+B\big{)}\mathrm{d}y,$
so that $\mathrm{L}_{\mathrm{X}}\omega=\lambda\omega$ if and only if
$\gamma\,xA^{\prime}=(\lambda-\gamma)A$ and
$\gamma\,xB^{\prime}=(\lambda-1)B.$ By putting
$\kappa=\frac{\lambda-\gamma}{\gamma}$ and $\nu=\frac{\lambda-1}{\gamma},$ the
last two equations can be rewritten as $xA^{\prime}=\kappa\,A$ and
$xB^{\prime}=\nu B$ and can be immediately integrated to give
$A(x)=\alpha\,x^{\kappa}$ and $B(x)=\beta\,x^{\nu},$ where
$\alpha,\beta\in\mathbb{C}.$ Since $A,B\in\mathbb{C}[x]$ and $\gcd(A,B)=1,$ we
deduce that $\alpha,\beta\in\mathbb{C}^{*},$ $\kappa,\nu\in\mathbb{N}$ and
$\kappa\,\nu=0.$ The equality $\deg\mathcal{F}=d$ then implies that
* •
either $\kappa=0$ and $\nu=d,$ in which case
$\omega=\alpha\mathrm{d}x+\beta\,x^{d}\mathrm{d}y;$
* •
or $\nu=0$ and $\kappa=d,$ in which case
$\omega=\alpha\,x^{d}\mathrm{d}x+\beta\mathrm{d}y.$
If $\omega=\alpha\mathrm{d}x+\beta\,x^{d}\mathrm{d}y,$ resp.
$\omega=\alpha\,x^{d}\mathrm{d}x+\beta\mathrm{d}y,$ by making the change of
coordinates
$\left(x,y\right)\mapsto\left(\frac{y}{x},-\frac{\alpha}{\beta\,x}\right)$, we
reduce ourselves to
$\omega=\omega_{1}^{d}=y^{d}\mathrm{d}x+x^{d}(x\mathrm{d}y-y\mathrm{d}x),$
resp.
$\omega=\omega_{2}^{d}=x^{d}\mathrm{d}x+y^{d}(x\mathrm{d}y-y\mathrm{d}x).$
1.2. Let us examine the case where $X=\varepsilon\frac{\partial}{\partial
x}+y\frac{\partial}{\partial y}$ with $\varepsilon\in\\{0,1\\}.$ Since
$\mathrm{L}_{\mathrm{X}}\mathrm{d}x=\mathrm{d}\mathrm{L}_{\mathrm{X}}x=0$ and
$\mathrm{L}_{\mathrm{X}}\mathrm{d}y=\mathrm{d}\mathrm{L}_{\mathrm{X}}y=\mathrm{d}y,$
we have
$\mathrm{L}_{\mathrm{X}}\omega=\mathrm{X}(A)\mathrm{d}x+\mathrm{X}(B)\mathrm{d}y+B\mathrm{d}y=\varepsilon\,A^{\prime}\mathrm{d}x+(\varepsilon\hskip
0.56905ptB^{\prime}+B)\mathrm{d}y.$ Therefore
$\mathrm{L}_{\mathrm{X}}\omega=\lambda\omega$ if and only if
$\varepsilon\,A^{\prime}=\lambda\,A$ and $\varepsilon\hskip
0.56905ptB^{\prime}=(\lambda-1)B.$ Since $A,B\in\mathbb{C}[x]$ and
$\lambda\in\mathbb{C},$ it follows that $AB=0$: contradiction with
$\gcd(A,B)=1.$
1.3. Let us study the eventuality: $\mathrm{X}=x\frac{\partial}{\partial
x}+(x+y)\frac{\partial}{\partial y}.$ We have
$\mathrm{d}\mathrm{X}(x)=\mathrm{d}x$ and
$\mathrm{d}\mathrm{X}(y)=\mathrm{d}x+\mathrm{d}y,$ so that
$\mathrm{L}_{\mathrm{X}}\omega=\mathrm{X}(A)\mathrm{d}x+A\mathrm{d}x+\mathrm{X}(B)\mathrm{d}y+B(\mathrm{d}x+\mathrm{d}y)=(xA^{\prime}+A+B)\mathrm{d}x+(xB^{\prime}+B)\mathrm{d}y.$
The condition $\mathrm{L}_{\mathrm{X}}\omega=\lambda\omega$ is then equivalent
to the system of differential equations $xA^{\prime}+B=(\lambda-1)A$ and
$xB^{\prime}=(\lambda-1)B,$ which can be easily integrated to yield
$A(x)=(a-b\ln x)x^{\lambda-1}$ and $B(x)=bx^{\lambda-1},$ where
$a,b\in\mathbb{C}.$ Since $A\in\mathbb{C}[x]$, we deduce that $b=0$ and
therefore $B\equiv 0$: contradiction with $\gcd(A,B)=1.$
2. Assume that $\mathfrak{iso}(\mathcal{F})$ is of type ($\mathfrak{e}$), i.e. $\mathfrak{iso}(\mathcal{F})=\big{\langle}\mathrm{X},\mathrm{Y}\big{\rangle}$ where $\mathrm{X}=x\frac{\partial}{\partial x}+2y\frac{\partial}{\partial y},\mathrm{Y}=\frac{\partial}{\partial x}+x\frac{\partial}{\partial y}.$ As before by writing explicitly that $\mathrm{L}_{\mathrm{X}}\omega=\lambda\omega$ and $\mathrm{L}_{\mathrm{Y}}\omega=\mu\omega,$ with $\lambda,\mu\in\mathbb{C}$ (Lemma 2.1), we obtain the system of partial differential equations
$\displaystyle x\tfrac{\partial A}{\partial x}+2y\tfrac{\partial A}{\partial
y}=(\lambda-1)A,$ $\displaystyle x\tfrac{\partial B}{\partial
x}+2y\tfrac{\partial B}{\partial y}=(\lambda-2)B,$
$\displaystyle\tfrac{\partial A}{\partial x}+x\tfrac{\partial A}{\partial
y}=\mu\hskip 1.13809ptA-B,$ $\displaystyle\tfrac{\partial B}{\partial
x}+x\tfrac{\partial B}{\partial y}=\mu\hskip 0.56905ptB.$
It follows in particular that
$\displaystyle(x^{2}-2y)\tfrac{\partial B}{\partial y}=(\mu\,x+2-\lambda)B$
and $\displaystyle(x^{2}-2y)\tfrac{\partial A}{\partial
y}=(\mu\,x+1-\lambda)A-xB.$
Elementary integrations then lead to
$\displaystyle B(x,y)=b(x)(x^{2}-2y)^{\frac{\lambda-2-\mu\,x}{2}}$ and
$\displaystyle
A(x,y)=\left(a(x)\sqrt{x^{2}-2y}-xb(x)\right)(x^{2}-2y)^{\frac{\lambda-2-\mu\,x}{2}},$
for some functions $a$ and $b$ of the coordinate $x.$ Since
$A,B\in\mathbb{C}[x,y]$ and $\gcd(A,B)=1,$ we deduce that $\lambda-2-\mu\,x=0$
and $a(x)=0$ for any $x\in\mathbb{C}$, hence $\lambda=2,\mu=0$ and $a\equiv
0.$ Therefore $B(x,y)=b(x)$ and $A(x,y)=-xb(x)=-xB(x,y)$: contradiction with
$\gcd(A,B)=1.$ ∎
## 3 Foliations of $\mathbf{F}(d)$ degenerating onto $\mathcal{F}_{1}^{d}$
and $\mathcal{F}_{2}^{d}$
In this section we will study the problem of knowing whether the closure of
the orbit of a foliation of $\mathbf{F}(d)$ contains the foliations
$\mathcal{F}_{1}^{d}$ and $\mathcal{F}_{2}^{d}.$ The following definition will
be useful.
###### Definition 3.1 ([9]).
Let $\mathcal{F}$ and $\mathcal{F}^{\prime}$ be two foliations of
$\mathbf{F}(d).$ We say that $\mathcal{F}$ degenerates onto
$\mathcal{F}^{\prime}$ if the closure $\overline{\mathcal{O}(\mathcal{F})}$
(in $\mathbf{F}(d)$) of the orbit $\mathcal{O}(\mathcal{F})$ contains
$\mathcal{O}(\mathcal{F}^{\prime})$ and
$\mathcal{O}(\mathcal{F})\not=\mathcal{O}(\mathcal{F}^{\prime}).$
###### Remarks 3.2.
Let $\mathcal{F}$ and $\mathcal{F}^{\prime}$ be two foliations such that
$\mathcal{F}$ degenerates onto $\mathcal{F}^{\prime}$. Then:
* (i)
$\dim\mathcal{O}(\mathcal{F}^{\prime})<\dim\mathcal{O}(\mathcal{F})$;
* (ii)
$\deg\mathrm{I}_{\mathcal{F}}^{\mathrm{inv}}\leq\deg\mathrm{I}_{\mathcal{F}^{\prime}}^{\mathrm{inv}},$
which is equivalent to $\deg\mathrm{I}_{\mathcal{F}}^{\hskip
0.56905pt\mathrm{tr}}\geq\deg\mathrm{I}_{\mathcal{F}^{\prime}}^{\hskip
0.56905pt\mathrm{tr}}$. In particular, if $\mathcal{F}$ is convex,
$\mathcal{F}^{\prime}$ is also convex.
Corollary B is an immediate consequence of Theorem A and Remark 3.2 (i).
###### Remark 3.3.
Corollary B applies particularly to any foliation of $\mathbf{F}(d)$ which is
homogeneous, i.e. which is given, for a suitable choice of affine coordinates
$(x,y),$ by a homogeneous $1$-form
$\omega=A(x,y)\mathrm{d}x+B(x,y)\mathrm{d}y,$ where
$A,B\in\mathbb{C}[x,y]_{d}$ and $\gcd(A,B)=1.$ Indeed, for such a foliation
$\mathcal{H}\in\mathbf{F}(d)$, we have (_cf._ [4])
$\mathrm{Iso}(\mathcal{H})=\\{(\alpha\,x,\alpha\,y)\hskip 2.84526pt|\hskip
2.84526pt\alpha\in\mathbb{C}^{*}\\};$
in particular, $\dim\mathcal{O}(\mathcal{H})=7$ and consequently
$\hskip
15.649pt\overline{\mathcal{O}(\mathcal{H})}\subset\mathcal{O}(\mathcal{H})\cup\mathcal{O}(\mathcal{F}_{1}^{d})\cup\mathcal{O}(\mathcal{F}_{2}^{d}).$
Assertion 1. (resp. 2.) of the following proposition gives a necessary (resp.
sufficient) condition for a foliation of $\mathbf{F}(d)$ to degenerate onto
the foliation $\mathcal{F}_{1}^{d}.$
###### Proposition 3.4.
Let $\mathcal{F}$ be an element of $\mathbf{F}(d)$ such that
$\mathcal{F}_{1}^{d}\not\in\mathcal{O}(\mathcal{F}).$ The following assertions
hold:
1. If $\mathcal{F}$ degenerates onto $\mathcal{F}_{1}^{d}$, then $\mathcal{F}$ possesses a non-degenerate singularity $m$ satisfying $\mathrm{BB}(\mathcal{F},m)=4.$
2. If $\mathcal{F}$ possesses a quasi-radial singularity of maximal order $d-1$, i.e. if $\mathrm{QRad}(\mathcal{F},d-1)\neq\emptyset$, then $\mathcal{F}$ degenerates onto $\mathcal{F}_{1}^{d}.$
###### Proof.
1. Assume that $\mathcal{F}$ degenerates onto $\mathcal{F}_{1}^{d}$. Then there is an analytic family of foliations $(\mathcal{F}_{\varepsilon})$ defined by a family of $1$-forms $(\omega_{\varepsilon})$ such that $\mathcal{F}_{\varepsilon}$ belongs to $\mathcal{O}(\mathcal{F})$ for $\varepsilon\neq 0$ and $\mathcal{F}_{\varepsilon=0}=\mathcal{F}_{1}^{d}.$ The non-degenerate singular point of $\mathcal{F}_{1}^{d}$, denoted by $m_{0},$ is $\scriptscriptstyle\langle\\!\langle$ stable $\\!\scriptscriptstyle\,\rangle\\!\rangle$ in the sense that there is an analytic family $(m_{\varepsilon})$ of non-degenerate singular points of $\mathcal{F}_{\varepsilon}$ such that $m_{\varepsilon=0}=m_{0}.$ The $\mathcal{F}_{\varepsilon}$’s being conjugated to $\mathcal{F}$ for $\varepsilon\neq 0$, the foliation $\mathcal{F}$ admits a non-degenerate singular point $m$ such that
$\forall\hskip 2.84526pt\varepsilon\in\mathbb{C}^{*},\hskip
2.84526pt\mathrm{BB}(\mathcal{F}_{\varepsilon},m_{\varepsilon})=\mathrm{BB}(\mathcal{F},m).$
Since $\mu(\mathcal{F}_{\varepsilon},m_{\varepsilon})=1$ for any
$\varepsilon\in\mathbb{C}$, the function
$\varepsilon\mapsto\mathrm{BB}(\mathcal{F}_{\varepsilon},m_{\varepsilon})$ is
continuous, hence constant on $\mathbb{C}$. As a consequence
$\displaystyle\mathrm{BB}(\mathcal{F},m)=\mathrm{BB}(\mathcal{F}_{\varepsilon=0},m_{\varepsilon=0})=\mathrm{BB}(\mathcal{F}_{1}^{d},m_{0})=4.$
2. Assume that $\mathcal{F}$ has a quasi-radial singularity $m$ of order $d-1$. Then $\mu(\mathcal{F},m)=1,\mathrm{BB}(\mathcal{F},m)=4$ and $\kappa(\mathcal{F},m)=d.$ This last equality ensures the existence of a line $\ell_{m}$ passing through $m$, not invariant by $\mathcal{F}$ and such that $\mathrm{Tang}(\mathcal{F},\ell_{m},m)=d.$ Let us choose an affine coordinate system $(x,y)$ such that $m=(0,0)$ and $\ell_{m}=\\{x=0\\}.$ The foliation $\mathcal{F}$ is defined in these coordinates by a $1$-form $\omega$ of type
$\displaystyle\omega=C_{d}(x,y)(x\mathrm{d}y-y\mathrm{d}x)+\sum_{i=1}^{d}\big{(}A_{i}(x,y)\mathrm{d}x+B_{i}(x,y)\mathrm{d}y\big{)},$
$\displaystyle\text{where }A_{i},\,B_{i}\in\mathbb{C}[x,y]_{i},\hskip
2.84526ptC_{d}\in\mathbb{C}[x,y]_{d}.$
We have
$\displaystyle\omega\wedge\mathrm{d}x\Big{|}_{x=0}=\sum_{i=1}^{d}B_{i}(0,y)\mathrm{d}y\wedge\mathrm{d}x=\sum_{i=1}^{d}B_{i}(0,1)y^{i}\mathrm{d}y\wedge\mathrm{d}x.$
Then the equality $\mathrm{Tang}(\mathcal{F},\ell_{m},m)=d$ translates into
$B_{i}(0,1)=0$ for $i\in\\{1,2,\ldots,d-1\\}$ and $B_{d}(0,1)\neq 0.$ This
allows to write
$\displaystyle B_{1}(x,y)=\alpha\hskip 0.56905ptx,$ $\displaystyle
B_{d}(x,y)=x\widehat{B}_{d-1}(x,y)+\gamma\,y^{d},$ $\displaystyle
B_{i}(x,y)=x\widetilde{B}_{i-1}(x,y)\hskip 2.84526pt\text{for
}i\in\\{2,3,\ldots,d-1\\},$
where $\widetilde{B}_{i-1}\in\mathbb{C}[x,y]_{i-1},\hskip
2.84526pt\widehat{B}_{d-1}\in\mathbb{C}[x,y]_{d-1},\hskip
2.84526pt\gamma\in\mathbb{C}^{*},\hskip 2.84526pt\alpha\in\mathbb{C}.$ The
equalities $\mu(\mathcal{F},m)=1$ and $\mathrm{BB}(\mathcal{F},m)=4$ imply
that $\alpha\neq 0$ and $A_{1}(x,y)=\delta\hskip 0.56905ptx-\alpha y$ for some
$\delta\in\mathbb{C}.$ Thus $\omega$ is of type
$\displaystyle\omega=\delta\hskip
0.56905ptx\mathrm{d}x+\big{(}x\widehat{B}_{d-1}(x,y)+\gamma\,y^{d}\big{)}\mathrm{d}y+\big{(}C_{d}(x,y)+\alpha\big{)}\big{(}x\mathrm{d}y-y\mathrm{d}x\big{)}+\sum_{i=2}^{d}A_{i}(x,y)\mathrm{d}x+x\sum_{i=2}^{d-1}\widetilde{B}_{i-1}(x,y)\mathrm{d}y,$
where $A_{i}\in\mathbb{C}[x,y]_{i},\hskip
2.84526pt\widetilde{B}_{i-1}\in\mathbb{C}[x,y]_{i-1},\hskip
2.84526pt\widehat{B}_{d-1}\in\mathbb{C}[x,y]_{d-1},\hskip
2.84526pt\delta\in\mathbb{C},\hskip 2.84526pt\alpha,\gamma\in\mathbb{C}^{*}.$
By putting $\varphi=\left(\varepsilon^{d}\hskip 0.28453ptx,\varepsilon\hskip
0.28453pty\right)$ and
$\theta=\alpha(x\mathrm{d}y-y\mathrm{d}x)+\gamma\,y^{d}\mathrm{d}y$, we obtain
$\displaystyle\frac{1}{\varepsilon^{d+1}}\varphi^{*}\omega=\theta+\varepsilon^{d-1}\left(\delta\hskip
0.56905ptx\mathrm{d}x+x\widehat{B}_{d-1}(\varepsilon^{d-1}x,y)\mathrm{d}y\right)+\varepsilon^{d}C_{d}(\varepsilon^{d-1}x,y)\left(x\mathrm{d}y-y\mathrm{d}x\right)+\sum_{i=2}^{d}\varepsilon^{i-1}A_{i}(\varepsilon^{d-1}x,y)\mathrm{d}x+x\sum_{i=2}^{d-1}\varepsilon^{i-1}\widetilde{B}_{i-1}(\varepsilon^{d-1}x,y)\mathrm{d}y$
which tends to $\theta$ as $\varepsilon$ tends to $0.$ By making the change of
coordinates $\left(x,y\right)\mapsto\left(\frac{x}{y}-\frac{\gamma}{\alpha
y},\frac{x}{y}\right)$, we reduce ourselves to
$\theta=\omega_{1}^{d}=y^{d}\mathrm{d}x+x^{d}(x\mathrm{d}y-y\mathrm{d}x)$. As
a result $\mathcal{F}$ degenerates onto $\mathcal{F}_{1}^{d}.$ ∎
###### Example 3.5.
Let us consider the homogeneous foliation $\mathcal{H}_{1}^{d}$ defined in the
affine chart $z=1$ by the $1$-form
${\mspace{2.0mu}\overline{\mspace{-1.4mu}\omega\mspace{-1.4mu}}\mspace{2.0mu}}_{1}^{d}=y^{d}\mathrm{d}x-x^{d}\mathrm{d}y.$
We know from [4, Proposition 4.1] that $\mathcal{H}_{1}^{d}$ is convex and
admits the points $[1:0:0]$ and $[0:1:0]$ as radial singularities of maximal
order $d-1.$ Therefore $\mathcal{H}_{1}^{d}$ degenerates onto
$\mathcal{F}_{1}^{d}$ (Proposition 3.4) and it does not degenerate onto
$\mathcal{F}_{2}^{d}$, because $\mathcal{F}_{2}^{d}$ is not convex. Thus,
according to Remark 3.3, we have
$\hskip
42.67912pt\overline{\mathcal{O}(\mathcal{H}_{1}^{d})}=\mathcal{O}(\mathcal{H}_{1}^{d})\cup\mathcal{O}(\mathcal{F}_{1}^{d}).$
###### Example 3.6.
Let us consider the family $(\mathcal{G}^{d}(\gamma))_{\gamma\in\mathbb{C}}$
of foliations of degree $d$ on $\mathbb{P}^{2}_{\mathbb{C}}$ defined in the
affine chart $z=1$ by
$\eta^{d}(\gamma)=(x-\gamma\,y)\mathrm{d}y-y\mathrm{d}x+x^{d}\mathrm{d}x-y^{d}\mathrm{d}y.$
We remark that the point $m=[0:0:1]$ is a non-degenerate singularity of
$\mathcal{G}^{d}(\gamma)$ with Baum-Bott index $4.$ Moreover, along the line
$\ell=\\{y=0\\}$ we have
$\eta^{d}(\gamma)\wedge\mathrm{d}y\Big{|}_{y=0}=x^{d}\mathrm{d}x\wedge\mathrm{d}y,$
so that $\mathrm{Tang}(\mathcal{G}^{d}(\gamma),\ell,m)=d$. It follows that the
singularity $m$ of $\mathcal{G}^{d}(\gamma)$ is quasi-radial of maximal order
$d-1.$ As a consequence $\mathcal{G}^{d}(\gamma)$ degenerates onto
$\mathcal{F}_{1}^{d}$ (Proposition 3.4).
The converse of assertion 2. of Proposition 3.4 is false as the following
example shows.
###### Example 3.7.
Let $\mathcal{F}$ be the foliation of degree $d\geq 2$ on
$\mathbb{P}^{2}_{\mathbb{C}}$ defined in the affine chart $z=1$ by
$\displaystyle\omega=x\mathrm{d}y-y\mathrm{d}x+P(y)\mathrm{d}y,$
where $P$ is a polynomial of $\mathbb{C}[y]$ of degree $d$ admitting $0$ as a
root of multiplicity $\leq d-1,$ i.e. $P$ is of the form
$\displaystyle P(y)=y^{\nu}(a_{0}+a_{1}y+\cdots+a_{d-\nu}y^{d-\nu}),$
$\displaystyle\text{where}\hskip 2.84526pt\nu\in\\{1,2,\ldots,d-1\\},\hskip
2.84526pta_{i}\in\mathbb{C},\hskip 2.84526pta_{0}a_{d-\nu}\neq 0.$
The singular locus of $\mathcal{F}$ consists of the two points $m=[0:0:1]$ and
$m^{\prime}=[1:0:0]$; moreover
$\displaystyle\mu(\mathcal{F},m)=1,$
$\displaystyle\mathrm{BB}(\mathcal{F},m)=4,$
$\displaystyle\kappa(\mathcal{F},m)=\nu<d,$
$\displaystyle\mu(\mathcal{F},m^{\prime})>1.$
It follows that $\mathcal{F}$ has no quasi-radial singularity of maximal order
$d-1,$ i.e. $\mathrm{QRad}(\mathcal{F},d-1)=\emptyset.$ However, $\mathcal{F}$
degenerates onto $\mathcal{F}_{1}^{d}$. Indeed, by putting
$\varphi=$$\left(\dfrac{a_{d-\nu}}{\varepsilon^{d}}x,\dfrac{1}{\varepsilon}y\right)$,
we see that
$\displaystyle\lim_{\varepsilon\to
0}\frac{\varepsilon^{d+1}}{a_{d-\nu}}\varphi^{*}\omega=x\mathrm{d}y-y\mathrm{d}x+y^{d}\mathrm{d}y.$
###### Question 1.
Let $\mathcal{F}$ be a foliation of degree $d\geq 2$ on
$\mathbb{P}^{2}_{\mathbb{C}}.$ Is it true that if $\mathcal{F}$ degenerates
onto $\mathcal{F}_{1}^{d}$ then
* –
either $\mathcal{F}$ admits a quasi-radial singularity of maximal order $d-1$,
* –
or $\mathcal{F}$ is conjugated to Example 3.7, i.e. up to linear conjugation
$\mathcal{F}$ is given by a $1$-form of type
$x\mathrm{d}y-y\mathrm{d}x+P(y)\mathrm{d}y$ with $P\in\mathbb{C}[y],$ $\deg
P=d$ and $P(0)=0$?
###### Proposition 3.8.
Let $d$ be an integer greater than or equal to $2.$ Let us denote by
$U_{1}(d)$ the subset of $\mathbf{F}(d)$ defined by
$\displaystyle U_{1}(d):=\Big{\\{}\mathcal{F}\in\mathbf{F}(d)\hskip
2.84526pt|\hskip 2.84526pt\forall\hskip
0.56905pts\in\mathrm{Sing}(\mathcal{F}),\hskip
2.84526pt\mu(\mathcal{F},s)=1,\tau(\mathcal{F},s)=1\Big{\\}}.$
Then:
* (i)
$U_{1}(d)$ is a non-empty Zariski open subset of $\mathbf{F}(d);$ in
particular, for any $\gamma\in\mathbb{C},$ $\mathcal{G}^{d}(\gamma)\in
U_{1}(d)$ if and only if
$\gamma\left(\gamma^{d+1}+\dfrac{(d+1)^{d+1}}{d^{d}}\right)\neq 0.$
* (ii)
Let $\mathcal{F}$ be an element of $U_{1}(d).$ For any singular point
$s\in\mathrm{Sing}(\mathcal{F}),$ the set
$\Lambda(\mathcal{F},s):=\Big{\\{}\ell_{s}\in\mathfrak{L}_{s}(\mathcal{F})\hskip
2.84526pt|\hskip 2.84526pt\mathrm{Tang}(\mathcal{F},\ell_{s},s)>1\Big{\\}}$
has at most $2$ elements. In particular, the set
$\bigcup\limits_{n=2}^{d}\widehat{\mathrm{QRad}}(\mathcal{F},n-1)$ is finite.
To prove this proposition, we need the following lemma.
###### Lemma 3.9.
Let $\mathcal{F}$ be a foliation of degree $d\geq 2$ on
$\mathbb{P}^{2}_{\mathbb{C}},$ $s$ a singular point of $\mathcal{F},$
$\ell_{s}$ a line passing through $s$ and not invariant by $\mathcal{F}$ and
$\mathrm{X}=A(x,y)\frac{\partial{}}{\partial{x}}+B(x,y)\frac{\partial{}}{\partial{y}}$
a polynomial vector field defining $\mathcal{F}$ in an affine chart $(x,y)$
containing $s.$ Let us denote by $(x_{0},y_{0})$ the coordinates of $s$ and
let $a(x-x_{0})+b(y-y_{0})=0$ be an equation of the line $\ell_{s}.$ Then, for
any $n\in\\{2,3,\ldots,d\\},$ $\mathrm{Tang}(\mathcal{F},\ell_{s},s)\geq n$ if
and only if
$\displaystyle\frac{\mathrm{d}^{j}}{\mathrm{d}t^{j}}\Big{(}a\hskip
0.56905ptA(x_{0}+b\hskip 0.56905ptt,y_{0}-a\hskip 0.56905ptt)+bB(x_{0}+b\hskip
0.56905ptt,y_{0}-a\hskip 0.56905ptt)\Big{)}\Big{|}_{t=0}=0,\quad\forall\hskip
0.56905ptj\in\\{1,2,\ldots,n-1\\}.$
In particular, the set
$\Lambda(\mathcal{F},s):=\Big{\\{}\ell_{s}\in\mathfrak{L}_{s}(\mathcal{F})\hskip
2.84526pt|\hskip
2.84526pt\mathrm{Tang}(\mathcal{F},\ell_{s},s)>\tau(\mathcal{F},s)\Big{\\}}$
is finite and its cardinality is at most $\tau(\mathcal{F},s)+1.$
###### Proof.
The $1$-form $\omega=A(x,y)\mathrm{d}y-B(x,y)\mathrm{\mathrm{d}}x$ also
defines the foliation $\mathcal{F}$ because $i_{\mathrm{X}}\omega=0.$ We have
$\displaystyle\omega\wedge\mathrm{d}\big{(}a(x-x_{0})+b(y-y_{0})\big{)}\Big{|}_{(x,y)=(x_{0}+b\hskip
0.56905ptt,y_{0}-a\hskip 0.56905ptt)}=P(t)\mathrm{d}y\wedge\mathrm{d}x,$
where $P(t)=a\hskip 0.56905ptA(x_{0}+b\hskip 0.56905ptt,y_{0}-a\hskip
0.56905ptt)+bB(x_{0}+b\hskip 0.56905ptt,y_{0}-a\hskip 0.56905ptt)$. Thus
$\mathrm{Tang}(\mathcal{F},\ell_{s},s)=\nu(P(t),0).$ Notice that
$P(0)=\leavevmode\nobreak\ 0$ because the point $s$ being singular for
$\mathcal{F},$ we have $A(x_{0},y_{0})=B(x_{0},y_{0})=0.$ Then
$\mathrm{Tang}(\mathcal{F},\ell_{s},s)\geq n$ if and only if the root $t=0$ of
the polynomial $P$ has multiplicity at least $n$, that is if and only if
$P^{\prime}(0)=P^{\prime\prime}(0)=\cdots=P^{(n-1)}(0)=0$, hence the announced
equivalence holds.
By conjugating $\omega$ by the translation $(x+x_{0},y+y_{0})$, we can assume
that $s=(0,0)$. Let us denote $\tau(\mathcal{F},s)$ simply by $\tau$. Then the
vector field $\mathrm{X}$ decomposes in the form
$\displaystyle\mathrm{X}=C_{\tau-2}(x,y)\mathrm{R}+\sum_{i=\tau}^{d+1}\mathrm{X}_{i},$
where
$\mathrm{R}=x\frac{\partial{}}{\partial{x}}+y\frac{\partial{}}{\partial{y}}$,
$C_{\tau-2}$ is a polynomial of degree $\leq\tau-2,$
$\mathrm{X}_{i}=A_{i}(x,y)\frac{\partial{}}{\partial{x}}+B_{i}(x,y)\frac{\partial{}}{\partial{y}}$
is a homogeneous vector field of degree $i$, with
$\det(\mathrm{X}_{\tau},\mathrm{R})\not\equiv 0.$ Thus, we have
$\displaystyle a\hskip 0.56905ptA(b\hskip 0.56905ptt,-a\hskip
0.56905ptt)+bB(b\hskip 0.56905ptt,-a\hskip 0.56905ptt)$
$\displaystyle=\bigg{(}a\Big{(}xC_{\tau-2}(x,y)+\sum_{i=\tau}^{d+1}A_{i}(x,y)\Big{)}+b\Big{(}yC_{\tau-2}(x,y)+\sum_{i=\tau}^{d+1}B_{i}(x,y)\Big{)}\bigg{)}\bigg{|}_{(x,y)=(b\hskip
0.56905ptt,-a\hskip 0.56905ptt)}$
$\displaystyle=\sum_{i=\tau}^{d+1}\Big{(}a\hskip 0.56905ptA_{i}(b\hskip
0.56905ptt,-a\hskip 0.56905ptt)+bB_{i}(b\hskip 0.56905ptt,-a\hskip
0.56905ptt)\Big{)}$ $\displaystyle=\sum_{i=\tau}^{d+1}t^{i}Q_{i+1}(a,b),$
where $Q_{i+1}(a,b):=a\hskip 0.56905ptA_{i}(b,-a)+bB_{i}(b,-a)$ is a
homogeneous polynomial of degree $i+1$ in $(a,b).$ From this, we deduce that
$\mathrm{Tang}(\mathcal{F},\ell_{s},s)>\tau$ if and only if
$Q_{\tau+1}(a,b)=0.$ As a result
$\displaystyle\Lambda(\mathcal{F},s)=\Big{\\{}\ell_{s}=\\{ax+by=0\\}\in\mathfrak{L}_{s}(\mathcal{F})\hskip
2.84526pt|\hskip 2.84526ptQ_{\tau+1}(a,b)=0\Big{\\}}.$
Now, the polynomial $Q_{\tau+1}$ is not identically zero because
$Q_{\tau+1}(a,b)=-\det(\mathrm{X}_{\tau},\mathrm{R})\big{|}_{(x,y)=(b,-a)}\not\equiv
0.$ It follows that $\Lambda(\mathcal{F},s)$ has cardinality at most $\tau+1.$
∎
###### Proof of Proposition 3.8.
We have
$\displaystyle U_{1}(d)=\Big{\\{}\mathcal{F}\in\mathbf{F}(d)\hskip
2.84526pt|\hskip 2.84526pt\forall\hskip
0.56905pts\in\mathrm{Sing}(\mathcal{F}),\hskip
2.84526pt\det(\mathrm{Jac}\hskip 0.28453pt\mathrm{X}(s))\neq 0,\hskip
2.84526pt\det(J^{1}_{s}\mathrm{X},\mathrm{R}_{s})\not\equiv 0\Big{\\}},$
where $\mathrm{X}$ denotes a polynomial vector field defining $\mathcal{F}$ in
an affine chart containing $s$ and $\mathrm{R}_{s}$ is the radial vector field
centered at $s.$ It follows that $U_{1}(d)$ is a Zariski open subset of
$\mathbf{F}(d).$ To establish assertion (i), it remains to show that for any
$\gamma\in\mathbb{C},$ $\mathcal{G}^{d}(\gamma)\in U_{1}(d)$ if and only if
$\gamma\left(\gamma^{d+1}+\frac{(d+1)^{d+1}}{d^{d}}\right)\neq 0.$ In
homogeneous coordinates, the foliation $\mathcal{G}^{d}(\gamma)$ is defined by
the $1$-form
$\displaystyle\hskip
28.45274pt\Omega^{d}(\gamma)=z\big{(}x^{d}-yz^{d-1}\big{)}\mathrm{d}x-z\big{(}y^{d}+\gamma\,yz^{d-1}-xz^{d-1}\big{)}\mathrm{d}y+\big{(}y^{d+1}-x^{d+1}+\gamma\,y^{2}z^{d-1}\big{)}\mathrm{d}z\hskip
1.42262pt.$
The singular locus $\mathrm{Sing}\big{(}\mathcal{G}^{d}(\gamma)\big{)}$
consists of the points
$\displaystyle s_{0}=[0:0:1],$ $\displaystyle s_{k}=[x_{k}:x_{k}^{d}:1],$
$\displaystyle s^{\prime}_{l}=[1:\xi^{l}:0],$ $\displaystyle
k\in\\{1,2,\ldots,d^{2}-1\\},\,l\in\\{0,1,\ldots,d\\},$
where $\xi=\exp\left(\frac{2\mathrm{i}\pi}{d+1}\right)$ and the $x_{k}$’s are
the roots of the polynomial $P(x)=x^{d^{2}-1}+\gamma\,x^{d-1}-1.$
In the affine chart $z=1,$ resp. $x=1,$ $\mathcal{G}^{d}(\gamma)$ is given by
the vector field
$\displaystyle\mathrm{Y}=\big{(}y^{d}+\gamma\,y-x\big{)}\frac{\partial{}}{\partial{x}}+\big{(}x^{d}-y\big{)}\frac{\partial{}}{\partial{y}},$
$\displaystyle\text{resp.
}\mathrm{Z}=\big{(}y^{d+1}+\gamma\,y^{2}z^{d-1}-1\big{)}\frac{\partial{}}{\partial{y}}+z\big{(}y^{d}+\gamma\,yz^{d-1}-z^{d-1}\big{)}\frac{\partial{}}{\partial{z}}\hskip
1.42262pt.$
A direct computation show that $\det(\mathrm{Jac}\hskip
0.28453pt\mathrm{Y}(s_{0}))=1\neq 0$,
$\det(J^{1}_{s_{0}}\mathrm{Y},\mathrm{R}_{s_{0}})=\gamma\,y^{2}$ and
$\displaystyle\det(\mathrm{Jac}\hskip
0.28453pt\mathrm{Z}(s^{\prime}_{l}))=(d+1)\xi^{-2l}\neq 0,$
$\displaystyle\det(\mathrm{Jac}\hskip
0.28453pt\mathrm{Y}(s_{k}))=1-d\gamma\,x_{k}^{d-1}-d^{2}x_{k}^{d^{2}-1}=(d-1)\big{(}d\gamma\,x_{k}^{d-1}-d-1\big{)},\hskip
2.84526pt\text{\normalsize because}\hskip 2.84526ptP(x_{k})=0,$
$\displaystyle\det(J^{1}_{s^{\prime}_{l}}\mathrm{Z},\mathrm{R}_{s^{\prime}_{l}})=d\xi^{-l}\big{(}y-\xi^{l}\big{)}z\not\equiv
0,$
$\displaystyle\det(J^{1}_{s_{k}}\mathrm{Y},\mathrm{R}_{s_{k}})=\big{(}dx_{k}^{d^{2}-d}+\gamma\big{)}\big{(}y-x_{k}^{d}\big{)}^{2}-dx_{k}^{d-1}\big{(}x-x_{k}\big{)}^{2}\not\equiv
0,\hskip 2.84526pt\text{\normalsize because}\hskip 2.84526ptx_{k}\neq 0.$
From these we deduce that $\mathcal{G}^{d}(\gamma)\in U_{1}(d)$ if and only if
$\gamma\neq 0$ and $d\gamma\,x_{k}^{d-1}-d-1\neq 0$, i.e. if and only if
$\gamma\neq 0$ and $x_{k}^{d-1}\neq\frac{d+1}{d\gamma}.$ Now, by putting
$Q(t)=t^{d+1}+\gamma\,t-1,$ we have $P(x)=Q(x^{d-1})$ so that
$t_{0}\in\mathbb{C}$ is a root of the polynomial $Q(t)$ if and only if there
exists $k\in\\{1,2,\ldots,d^{2}-1\\}$ such that $t_{0}=x_{k}^{d-1}.$ It
follows that
$\displaystyle\mathcal{G}^{d}(\gamma)\in
U_{1}(d)\Longleftrightarrow\gamma\,Q\left(\frac{d+1}{d\gamma}\right)\neq
0\Longleftrightarrow\gamma\left(\gamma^{d+1}+\frac{(d+1)^{d+1}}{d^{d}}\right)\neq
0.$
Assertion (ii) is an immediate consequence of Lemma 3.9. ∎
###### Theorem 3.10.
Let $d$ be an integer greater than or equal to $2.$ Let us denote by
$\Sigma_{1}(d)$ the subset of $\mathbf{F}(d)$ defined by
$\displaystyle\Sigma_{1}(d):=\Big{\\{}\mathcal{F}\in\mathbf{F}(d)\hskip
2.84526pt|\hskip
2.84526pt\mathrm{QRad}(\mathcal{F},d-1)\neq\emptyset\Big{\\}}.$
Then
* (a)
$\emptyset\neq\Sigma_{1}(d)\varsubsetneq\mathbf{B}(\mathcal{F}_{1}^{d});$
* (b)
$\Sigma_{1}(d)$ is a constructible subset of $\mathbf{F}(d)$ of dimension
greater than or equal to $\dim\mathbf{F}(d)-(d-1).$
###### Proof.
(a) $\Sigma_{1}(d)$ contains the foliations $\mathcal{H}_{1}^{d}$ and
$\mathcal{G}^{d}(\gamma),\gamma\in\mathbb{C}$ (Examples 3.5 and 3.6) and is
therefore non-empty. Assertion 2. of Proposition 3.4 ensures that
$\Sigma_{1}(d)\subset\mathbf{B}(\mathcal{F}_{1}^{d})$; this inclusion is
strict as Example 3.7 shows.
(b) Let us denote by $\mathbb{\check{P}}^{2}_{\mathbb{C}}$ the dual projective
plane of $\mathbb{P}^{2}_{\mathbb{C}}.$ Let $\pi\hskip
2.84526pt\colon\mathbf{F}(d)\times\mathbb{P}^{2}_{\mathbb{C}}\times\mathbb{\check{P}}^{2}_{\mathbb{C}}\to\mathbf{F}(d)$
be the projection onto the first factor; we have
$\Sigma_{1}(d)=\pi(W_{1}(d)),$ where
$\displaystyle W_{1}(d):$
$\displaystyle=\bigcup_{\mathcal{F}\in\Sigma_{1}(d)}\\{\mathcal{F}\\}\times\widehat{\mathrm{QRad}}(\mathcal{F},d-1)$
$\displaystyle=\Big{\\{}(\mathcal{F},s,\ell)\in\mathbf{F}(d)\times\mathbb{P}^{2}_{\mathbb{C}}\times\mathbb{\check{P}}^{2}_{\mathbb{C}}\hskip
2.84526pt|\hskip 2.84526pts\in\mathrm{Sing}(\mathcal{F}),\hskip
2.84526pt\ell\in\mathfrak{L}_{s}(\mathcal{F}),\hskip
2.84526pt\mu(\mathcal{F},s)=1,\hskip
2.84526pt\mathrm{BB}(\mathcal{F},s)=4,\hskip
2.84526pt\mathrm{Tang}(\mathcal{F},\ell,s)=d\Big{\\}}.$
According to Lemma 3.9, $W_{1}(d)$ can be rewritten as
(3.5) $\displaystyle
W_{1}(d)=\left\\{(\mathcal{F},s,\ell)\in\mathbf{F}(d)\times\mathbb{P}^{2}_{\mathbb{C}}\times\mathbb{\check{P}}^{2}_{\mathbb{C}}\hskip
1.42262pt\left|\begin{array}[c]{l}\vspace{2mm}s=(x_{0},y_{0})\in\ell=\\{a(x-x_{0})+b(y-y_{0})=0\\}\\\
\vspace{2mm}A(x_{0},y_{0})=0,\hskip 2.84526ptB(x_{0},y_{0})=0,\hskip
2.84526pt\det(\mathrm{Jac}\hskip 0.28453pt\mathrm{X}(s))\neq 0,\hskip
2.84526pt\dfrac{\mathrm{tr}^{2}(\mathrm{Jac}\hskip
0.28453pt\mathrm{X}(s))}{\det(\mathrm{Jac}\hskip 0.28453pt\mathrm{X}(s))}=4\\\
\vspace{2mm}a\hskip 0.56905ptA(x_{0}+b\hskip 0.56905ptt,y_{0}-a\hskip
0.56905ptt)+bB(x_{0}+b\hskip 0.56905ptt,y_{0}-a\hskip 0.56905ptt)\not\equiv
0\\\ \dfrac{\mathrm{d}^{j}}{\mathrm{d}t^{j}}\Big{(}a\hskip
0.56905ptA(x_{0}+b\hskip 0.56905ptt,y_{0}-a\hskip 0.56905ptt)+bB(x_{0}+b\hskip
0.56905ptt,y_{0}-a\hskip
0.56905ptt)\Big{)}\Big{|}_{t=0}=0,j=1,\ldots,d-1\end{array}\right.\right\\},$
where
$\mathrm{X}=A(x,y)\frac{\partial{}}{\partial{x}}+B(x,y)\frac{\partial{}}{\partial{y}}$
is a polynomial vector field defining $\mathcal{F}$ in an affine chart $(x,y)$
containing $s.$ It follows that $W_{1}(d)$ is a quasi-projective subvariety of
$\mathbf{F}(d)\times\mathbb{P}^{2}_{\mathbb{C}}\times\mathbb{\check{P}}^{2}_{\mathbb{C}}.$
Thus, by Chevalley’s Theorem [11, Exercise II.3.19], the set
$\Sigma_{1}(d)=\pi(W_{1}(d))$ is constructible.
According to the above discussion and Proposition 3.8 (i), the intersection
$U_{1}(d)\cap\Sigma_{1}(d)$ contains the foliations $\mathcal{G}^{d}(\gamma)$,
with $\gamma\left(\gamma^{d+1}+\frac{(d+1)^{d+1}}{d^{d}}\right)\neq 0,$ and is
therefore non-empty ($U_{1}(d)$ being the set of $\mathcal{F}\in\mathbf{F}(d)$
such that for any $s\in\mathrm{Sing}\mathcal{F},$ $\mu(\mathcal{F},s)=1$ and
$\tau(\mathcal{F},s)=1$). Then there exists an irreducible component
$\Sigma_{1}^{0}(d)$ of $\Sigma_{1}(d)$ such that
$U_{1}(d)\cap\Sigma_{1}^{0}(d)\neq\emptyset.$ Let
$W_{1}(d)=\bigcup\limits_{i=1}^{k}W_{1}^{i}(d)$ be the decomposition of
$W_{1}(d)$ into its irreducible components. Let us denote by $\pi_{0}\hskip
2.84526pt\colon W_{1}(d)\to\mathbf{F}(d)$ the restriction of $\pi$ to
$W_{1}(d).$ Then, there is $n\in\\{1,\ldots,k\\}$ such that
$\overline{\pi_{0}(W_{1}^{n}(d))}=\overline{\Sigma_{1}^{0}(d)}.$ Indeed, since
$\Sigma_{1}(d)=\pi_{0}(W_{1}(d)),$ we have
$\overline{\Sigma_{1}^{0}(d)}\subset\overline{\Sigma_{1}(d)}=\bigcup\limits_{i=1}^{k}\overline{\pi_{0}(W_{1}^{i}(d))}.$
The irreducibility of $\Sigma_{1}^{0}(d)$ therefore ensures the existence of
$n\in\\{1,\ldots,k\\}$ such that
$\overline{\Sigma_{1}^{0}(d)}\subset\overline{\pi_{0}(W_{1}^{n}(d))}\subset\overline{\Sigma_{1}(d)}.$
Since $\overline{\Sigma_{1}^{0}(d)}$ is an irreducible component of
$\overline{\Sigma_{1}(d)}$ and since $\overline{\pi_{0}(W_{1}^{n}(d))}$ is
irreducible by continuity of $\pi_{0},$ we deduce that
$\overline{\pi_{0}(W_{1}^{n}(d))}=\overline{\Sigma_{1}^{0}(d)}.$
Thus, since $U_{1}(d)$ is a Zariski open subset of $\mathbf{F}(d)$
(Proposition 3.8 (i)), the morphism $\pi_{0}$ induces by restriction a
dominant morphism of quasi-projective varieties $\pi_{0}^{n}\hskip
2.84526pt\colon
W_{1}^{n}(d)\cap\pi_{0}^{-1}(U_{1}(d))\to\overline{\Sigma_{1}^{0}(d)}\cap
U_{1}(d).$
Notice that all the fibers of $\pi_{0}$ over the elements of
$U_{1}(d)\cap\Sigma_{1}(d)$ are finite and non-empty. Indeed, if
$\mathcal{F}\in U_{1}(d)\cap\Sigma_{1}(d)$ then, by Proposition 3.8 (ii), the
set $\widehat{\mathrm{QRad}}(\mathcal{F},d-1)$ is finite and non-empty;
therefore so is
$\pi_{0}^{-1}(\mathcal{F})=\\{\mathcal{F}\\}\times\widehat{\mathrm{QRad}}(\mathcal{F},d-1).$
Since $\pi_{0}(W_{1}^{n}(d)\cap\pi_{0}^{-1}(U_{1}(d)))\subset
U_{1}(d)\cap\Sigma_{1}(d),$ it follows that all the non-empty fibers of
$\pi_{0}^{n}$ are finite and therefore zero-dimensional. The fiber dimension
theorem (_cf._ [14, Theorem 3, page 49]) then implies that
$\dim(W_{1}^{n}(d)\cap\pi_{0}^{-1}(U_{1}(d)))=\dim(\overline{\Sigma_{1}^{0}(d)}\cap
U_{1}(d));$ since $W_{1}^{n}(d)\cap\pi_{0}^{-1}(U_{1}(d))$ and
$\overline{\Sigma_{1}^{0}(d)}\cap U_{1}(d)$ are non-empty open subsets of the
irreducible varieties $W_{1}^{n}(d)$ and $\overline{\Sigma_{1}^{0}(d)}$
respectively, we have
$\displaystyle\hskip
65.44142pt\dim\overline{\Sigma_{1}^{0}(d)}=\dim(\overline{\Sigma_{1}^{0}(d)}\cap
U_{1}(d))=\dim(W_{1}^{n}(d)\cap\pi_{0}^{-1}(U_{1}(d)))=\dim W_{1}^{n}(d).$
Now, from (3.5) we deduce that each irreducible component $W_{1}^{i}(d)$ of
$W_{1}(d)$ has dimension
$\hskip 41.25641pt\dim
W_{1}^{i}(d)\geq\dim(\mathbf{F}(d)\times\mathbb{P}^{2}_{\mathbb{C}}\times\mathbb{\check{P}}^{2}_{\mathbb{C}})-4-(d-1)=\dim\mathbf{F}(d)-(d-1),$
hence
$\hskip
44.67102pt\dim\Sigma_{1}(d)=\dim\overline{\Sigma_{1}(d)}\geq\dim\overline{\Sigma_{1}^{0}(d)}=\dim
W_{1}^{n}(d)\geq\dim\mathbf{F}(d)-(d-1).$
∎
Assertion 1. (resp. 2.) of the following proposition gives a necessary (resp.
sufficient) condition for a foliation of $\mathbf{F}(d)$ to degenerate onto
the foliation $\mathcal{F}_{2}^{d}.$
###### Proposition 3.11.
Let $\mathcal{F}$ be an element of $\mathbf{F}(d)$ such that
$\mathcal{F}_{2}^{d}\not\in\mathcal{O}(\mathcal{F}).$ The following assertions
hold:
1. If $\mathcal{F}$ degenerates onto $\mathcal{F}_{2}^{d}$, then $\deg\mathrm{I}_{\mathcal{F}}^{\hskip 0.56905pt\mathrm{tr}}\geq d-1.$
2. If $\mathcal{F}$ admits an inflection point of maximal order $d-1$, i.e. if $\mathrm{Flex}(\mathcal{F},d-1)\neq\emptyset,$ then $\mathcal{F}$ degenerates onto $\mathcal{F}_{2}^{d}.$
###### Proof.
1. If $\mathcal{F}$ degenerates onto $\mathcal{F}_{2}^{d},$ then $\deg\mathrm{I}_{\mathcal{F}}^{\hskip 0.56905pt\mathrm{tr}}\geq\deg\mathrm{I}_{\mathcal{F}_{2}^{d}}^{\hskip 0.56905pt\mathrm{tr}}.$ An immediate computation shows that $\mathrm{I}_{\mathcal{F}_{2}^{d}}^{\hskip 0.56905pt\mathrm{tr}}=\leavevmode\nobreak\ y^{d-1}$ so that $\deg\mathrm{I}_{\mathcal{F}_{2}^{d}}^{\hskip 0.56905pt\mathrm{tr}}=d-1$, hence the announced inequality holds.
2. Assume that $\mathcal{F}$ possesses such a point. We choose an affine coordinate system $(x,y)$ such that $p=(0,0)$ is an inflection point of order $d-1$ of $\mathcal{F}$ and $x=0$ is the tangent line to the leaf of $\mathcal{F}$ passing through $p.$ Let $\omega$ be a $1$-form defining $\mathcal{F}$ in these coordinates. Since $\mathrm{T}^{\mathbb{P}}_{p}\mathcal{F}=\\{x=0\\}$, $\omega$ is of type
$\displaystyle\omega=C_{d}(x,y)(x\mathrm{d}y-y\mathrm{d}x)+\alpha\mathrm{d}x+\sum_{i=1}^{d}\big{(}A_{i}(x,y)\mathrm{d}x+B_{i}(x,y)\mathrm{d}y\big{)},$
$\displaystyle\text{where }A_{i},\,B_{i}\in\mathbb{C}[x,y]_{i},\hskip
2.84526ptC_{d}\in\mathbb{C}[x,y]_{d},\hskip 2.84526pt\alpha\in\mathbb{C}^{*}.$
We have
$\displaystyle\omega\wedge\mathrm{d}x\Big{|}_{x=0}=\sum_{i=1}^{d}B_{i}(0,y)\mathrm{d}y\wedge\mathrm{d}x=\sum_{i=1}^{d}B_{i}(0,1)y^{i}\mathrm{d}y\wedge\mathrm{d}x.$
Therefore the hypothesis that $(0,0)$ is an inflection point of order $d-1$ of
$\mathcal{F}$ translates into $B_{i}(0,1)=0$ for $i\in\\{1,2,\ldots,d-1\\}$
and $B_{d}(0,1)\neq 0.$ Then we can write
$\displaystyle B_{d}(x,y)=x\widehat{B}_{d-1}(x,y)+\beta y^{d},$ $\displaystyle
B_{i}(x,y)=x\widetilde{B}_{i-1}(x,y)\hskip 2.84526pt\text{for
}i\in\\{1,2,\ldots,d-1\\},$
where $\widetilde{B}_{i-1}\in\mathbb{C}[x,y]_{i-1},\hskip
2.84526pt\widehat{B}_{d-1}\in\mathbb{C}[x,y]_{d-1},\hskip
2.84526pt\beta\in\mathbb{C}^{*}.$ Thus $\omega$ is of type
$\displaystyle\omega=\alpha\mathrm{d}x+\big{(}x\widehat{B}_{d-1}(x,y)+\beta
y^{d}\big{)}\mathrm{d}y+C_{d}(x,y)\big{(}x\mathrm{d}y-y\mathrm{d}x\big{)}+\sum_{i=1}^{d}A_{i}(x,y)\mathrm{d}x+x\sum_{i=1}^{d-1}\widetilde{B}_{i-1}(x,y)\mathrm{d}y,$
where $A_{i}\in\mathbb{C}[x,y]_{i},\hskip
2.84526pt\widetilde{B}_{i-1}\in\mathbb{C}[x,y]_{i-1},\hskip
2.84526pt\widehat{B}_{d-1}\in\mathbb{C}[x,y]_{d-1},\hskip
2.84526pt\alpha,\beta\in\mathbb{C}^{*}.$
Let us consider the family of automorphisms
$\varphi=\varphi_{\varepsilon}=(\varepsilon^{d+1}x,\varepsilon y).$ We have
$\displaystyle\frac{1}{\varepsilon^{d+1}}\varphi^{*}\omega=\alpha\mathrm{d}x+\Big{(}\varepsilon^{d}x\widehat{B}_{d-1}(\varepsilon^{d}x,y)+\beta
y^{d}\Big{)}\mathrm{d}y+\varepsilon^{d+1}C_{d}(\varepsilon^{d}x,y)\left(x\mathrm{d}y-y\mathrm{d}x\right)+\sum_{i=1}^{d}\varepsilon^{i}A_{i}(\varepsilon^{d}x,y)\mathrm{d}x+x\sum_{i=1}^{d-1}\varepsilon^{i}\widetilde{B}_{i-1}(\varepsilon^{d}x,y)\mathrm{d}y$
which tends to $\alpha\mathrm{d}x+\beta y^{d}\mathrm{d}y$ as $\varepsilon$
tends to $0.$ Clearly $\alpha\mathrm{d}x+\beta y^{d}\mathrm{d}y$ defines a
foliation conjugated to $\mathcal{F}_{2}^{d}$; as a result $\mathcal{F}$
degenerates onto $\mathcal{F}_{2}^{d}.$ ∎
###### Example 3.12.
Let us consider the homogeneous foliation $\mathcal{H}_{2}^{d}$ defined in the
affine chart $z=1$ by the $1$-form
${\mspace{2.0mu}\overline{\mspace{-1.4mu}\omega\mspace{-1.4mu}}\mspace{2.0mu}}_{2}^{d}=x^{d}\mathrm{d}x-y^{d}\mathrm{d}y.$
We know from [4, Proposition 4.1] that $\mathcal{H}_{2}^{d}$ has no non-
degenerate singularity with Baum-Bott index $4$ and that
$\hskip
123.48485pt\mathrm{Flex}(\mathcal{H}_{2}^{d},d-1)=\\{xy=0\\}\setminus\\{[0:0:1]\\}\neq\emptyset.$
Thus $\mathcal{H}_{2}^{d}$ degenerates onto $\mathcal{F}_{2}^{d}$ (Proposition
3.11) and it does not degenerate onto $\mathcal{F}_{1}^{d}$ (Proposition 3.4).
Consequently, according to Remark 3.3, we have
$\hskip
42.67912pt\overline{\mathcal{O}(\mathcal{H}_{2}^{d})}=\mathcal{O}(\mathcal{H}_{2}^{d})\cup\mathcal{O}(\mathcal{F}_{2}^{d}).$
###### Example 3.13 (Jouanolou’s foliation).
Let us consider the foliation $\mathcal{F}_{J}^{d}$ of degree $d\geq 2$ on
$\mathbb{P}^{2}_{\mathbb{C}}$ defined, in the affine chart $z=1,$ by
$\omega_{J}^{d}=(x^{d}y-1)\mathrm{d}x+(y^{d}-x^{d+1})\mathrm{d}y.$
This example is due to Jouanolou and is historically the first explicit
example of foliation without invariant algebraic curve ([12]). The point
$p=(0,0)$ is an inflection point of maximal order $d-1$ of
$\mathcal{F}_{J}^{d}$ because
$\mathrm{T}^{\mathbb{P}}_{p}\mathcal{F}_{J}^{d}=\\{x=0\\}\hskip
5.69054pt\text{and}\hskip
5.69054pt\omega_{J}^{d}\wedge\mathrm{d}x\Big{|}_{x=0}=y^{d}\mathrm{d}y\wedge\mathrm{d}x.$
As a result $\mathcal{F}_{J}^{d}$ degenerates onto $\mathcal{F}_{2}^{d}$
(Proposition 3.11). However, we know from [13, Section 3] that every
singularity $s$ of $\mathcal{F}_{J}^{d}$ is non-degenerate with Baum-Bott
index
$\mathrm{BB}(\mathcal{F}_{J}^{d},s)=\frac{(d+2)^{2}}{d^{2}+d+1}\neq 4,$
so that $\mathcal{F}_{J}^{d}$ does not degenerate onto $\mathcal{F}_{1}^{d}$
(Proposition 3.4).
The converse of assertion 2. of Proposition 3.11 is false as the following
example shows.
###### Example 3.14.
Let $\mathcal{F}$ be the foliation of degree $d\geq 2$ on
$\mathbb{P}^{2}_{\mathbb{C}}$ defined in the affine chart $z=1$ by
$\displaystyle\omega=\mathrm{d}x+P(y)\mathrm{d}y,$ $\displaystyle\text{where
}P\in\mathbb{C}[y],\hskip 2.84526pt\deg P=d.$
It is easy to check that
$\mathrm{Sing}(\mathcal{F})=\big{\\{}[1:0:0]\big{\\}}$ and
$\mathrm{I}_{\mathcal{F}}^{\hskip 0.56905pt\mathrm{tr}}=P^{\prime}(y).$ If the
derivative $P^{\prime}$ has a single root, i.e if $P$ is of the form
$P(y)=a(y-\alpha)^{d}+b,$ where $\alpha,a,b\in\mathbb{C},a\neq 0,$ then
$\mathcal{F}$ is conjugated to $\mathcal{F}_{2}^{d}$; indeed, we have
$\displaystyle\frac{1}{a}\varphi^{*}\omega=\mathrm{d}x+y^{d}\mathrm{d}y,\quad\text{where}\hskip
2.84526pt\varphi=(ax-by,y+\alpha).$
We assume that the derivative $P^{\prime}$ has at least two distinct roots;
this implies that $d\geq 3.$ A straightforward computation shows that
$\mathcal{F}$ has no inflection point of maximal order $d-1$, i.e.
$\mathrm{Flex}(\mathcal{F},d-1)=\emptyset.$ However, $\mathcal{F}$ degenerates
onto $\mathcal{F}_{2}^{d}$. Indeed, by writing
$P(y)=a_{0}+a_{1}y+\cdots+a_{d}y^{d}$, $a_{i}\in\mathbb{C},\,a_{d}\neq 0,$ and
by putting
$\psi=$$\left(\dfrac{a_{d}}{\varepsilon^{d+1}}x,\dfrac{1}{\varepsilon}y\right)$,
we obtain that
$\displaystyle\lim_{\varepsilon\to
0}\frac{\varepsilon^{d+1}}{a_{d}}\psi^{*}\omega=\mathrm{d}x+y^{d}\mathrm{d}y.$
###### Question 2.
Let $\mathcal{F}$ be a foliation of degree $d\geq 3$ on
$\mathbb{P}^{2}_{\mathbb{C}}.$ Is it true that if $\mathcal{F}$ degenerates
onto $\mathcal{F}_{2}^{d}$ then
* –
either $\mathcal{F}$ possesses an inflection point of maximal order $d-1$,
* –
or $\mathcal{F}$ is conjugated to Example 3.14, i.e. up to linear conjugation
$\mathcal{F}$ is given by a $1$-form of type $\mathrm{d}x+P(y)\mathrm{d}y$
with $P\in\mathbb{C}[y],$ $\deg P=d$?
###### Proposition 3.15.
Let $d$ be an integer greater than or equal to $2.$ Let us denote by
$U_{2}(d)$ the set of foliations $\mathcal{F}\in\mathbf{F}(d)$ whose
inflection divisor $\mathrm{I}_{\mathcal{F}}$ is transverse (i.e.
$\mathrm{I}_{\mathcal{F}}=\mathrm{I}_{\mathcal{F}}^{\hskip
0.56905pt\mathrm{tr}}$) and reduced. Then
* (i)
$U_{2}(d)$ contains the Jouanolou’s foliation $\mathcal{F}_{J}^{d}$ and it is
a (non-empty) Zariski open subset of $\mathbf{F}(d);$
* (ii)
for any $d\geq 3,$ every foliation $\mathcal{F}\in U_{2}(d)$ has a finite
number (possibly zero) of transverse inflection points of order greater than
or equal to $2$; in other words, the set
$\bigcup\limits_{k=3}^{d}\mathrm{Flex}(\mathcal{F},k-1)$ is finite.
To establish this proposition, let us first prove the following lemma.
###### Lemma 3.16.
Let $\mathcal{F}$ be a foliation of degree $d\geq 2$ on
$\mathbb{P}^{2}_{\mathbb{C}},$ $p$ a regular point of $\mathcal{F}$ and
$\mathrm{X}$ a polynomial vector field defining $\mathcal{F}$ in an affine
chart $(x,y)$ containing $p.$ Then, for any $k\in\\{2,3,\ldots,d\\},$
$\mathrm{Tang}(\mathcal{F},\mathrm{T}^{\mathbb{P}}_{p}\mathcal{F},p)\geq k$ if
and only if the matrix
$\left(\begin{array}[]{cccc}\mathrm{X}(x)&\mathrm{X}^{2}(x)&\cdots&\mathrm{X}^{k}(x)\\\
\mathrm{X}(y)&\mathrm{X}^{2}(y)&\cdots&\mathrm{X}^{k}(y)\end{array}\right)\Big{|}_{p}$
has rank $1.$
###### Remark 3.17.
If
$\mathrm{X}=\sum\limits_{i=1}^{n}\mathrm{X}_{i}(z_{1},\ldots,z_{n})\dfrac{\raisebox{-1.42262pt}{$\partial$}}{\partial
z_{i}}$ is a holomorphic vector field on $\mathbb{C}^{n}$ and if
$t\mapsto\alpha(t)$ is an integral curve of $\mathrm{X},$ then we have the
following formula which can be easily proved by induction on $j:$
(3.6)
$\displaystyle\frac{\raisebox{-2.27621pt}{$\mathrm{d}^{j}$}}{\mathrm{d}\hskip
0.56905ptt^{j}}\alpha(t)=(\mathrm{X}^{j}(z_{1}),\ldots,\mathrm{X}^{j}(z_{n}))\circ\alpha(t).$
###### Proof.
Let $t\mapsto\alpha(t)$ be the integral curve of $\mathrm{X}$ passing through
$p$ at $t=0.$ The point $p$ being regular for $\mathcal{F},$ we have
$\mathrm{T}_{p}\mathcal{F}\ni\alpha^{\prime}(0)=\mathrm{X}(p)\neq 0.$ Up to
linear conjugation, we can assume that $p=(0,0)$ and
$\mathrm{T}^{\mathbb{P}}_{p}\mathcal{F}=\\{y=0\\}.$ We can then write
$\alpha(t)=\left(\sum\limits_{i\geq 1}x_{i}\frac{t^{i}}{i!},\sum\limits_{i\geq
1}y_{i}\frac{t^{i}}{i!}\right)$ with $y_{1}=0$ and $x_{1}\neq 0.$ Thus,
$\mathrm{Tang}(\mathcal{F},\mathrm{T}^{\mathbb{P}}_{p}\mathcal{F},p)=\nu(g(t),0),$
where $g(t)=\sum\limits_{i\geq 2}y_{i}\frac{t^{i}}{i!}.$ As a result,
$\mathrm{Tang}(\mathcal{F},\mathrm{T}^{\mathbb{P}}_{p}\mathcal{F},p)\geq k$ if
and only if $y_{2}=y_{3}=\cdots=y_{k}=0$, or equivalently if and only if the
matrix $\left(\begin{array}[]{cccc}x_{1}&x_{2}&\cdots&x_{k}\\\
0&y_{2}&\cdots&y_{k}\end{array}\right)$ has rank $1.$ Now, by using formula
(3.6), we see that
$\left(\begin{array}[]{cccc}x_{1}&x_{2}&\cdots&x_{k}\\\
0&y_{2}&\cdots&y_{k}\end{array}\right)=\left(\begin{array}[]{cccc}\mathrm{X}(x)&\mathrm{X}^{2}(x)&\cdots&\mathrm{X}^{k}(x)\\\
\mathrm{X}(y)&\mathrm{X}^{2}(y)&\cdots&\mathrm{X}^{k}(y)\end{array}\right)\Big{|}_{(x,y)=(0,0)},$
hence the lemma follows. ∎
###### Proof of Proposition 3.15.
(i) For $\mathcal{F}\in\mathbf{F}(d),$ to say that $\mathrm{I}_{\mathcal{F}}$
is transverse and reduced means that $\mathcal{F}$ has no invariant line and
that $\mathrm{I}_{\mathcal{F}}$ has no multiple component, which shows that
$U_{2}(d)$ is a Zariski open subset of $\mathbf{F}(d).$
As we have already mentioned in Example 3.13, the Jouanolou’s foliation
$\mathcal{F}_{J}^{d}$ has no invariant algebraic curve [12]; in particular, it
has no invariant line and consequently
$\mathrm{I}_{\mathcal{F}_{J}^{d}}=\mathrm{I}_{\mathcal{F}_{J}^{d}}^{\hskip
0.56905pt\mathrm{tr}}.$ To establish the first announced assertion, it remains
to prove that $\mathrm{I}_{\mathcal{F}_{J}^{d}}$ is reduced. In homogeneous
coordinates, the foliation $\mathcal{F}_{J}^{d}$ is defined by the vector
field $y^{d}\frac{\partial}{\partial x}+z^{d}\frac{\partial}{\partial
y}+x^{d}\frac{\partial}{\partial z}$; an immediate computation, using formula
(1.1), shows that $\mathrm{I}_{\mathcal{F}_{J}^{d}}$ has equation
$F(x,y,z)=0,$ where
$F(x,y,z)=x^{2d+1}z^{d-1}+y^{2d+1}x^{d-1}+z^{2d+1}y^{d-1}-3x^{d}y^{d}z^{d}.$
We must show that $F$ has no multiple factor in $\mathbb{C}[x,y,z]$. Since
$F\in\mathbb{Z}[x,y,z],$ it suffices to show that $F$ has no multiple factor
in $\mathbb{F}_{2}[x,y,z].$ Indeed, if $F$ had a multiple factor in
$\mathbb{C}[x,y,z],$ then one of the resultants
$\mathrm{Res}_{x}(F,\frac{\partial F}{\partial x})\in\mathbb{Z}[y,z]\hskip
1.13809pt$ or $\hskip 1.13809pt\mathrm{Res}_{y}(F,\frac{\partial F}{\partial
y})\in\mathbb{Z}[x,z]\hskip 1.13809pt$ or $\hskip
1.13809pt\mathrm{Res}_{z}(F,\frac{\partial F}{\partial z})\in\mathbb{Z}[x,y]$
would be identically zero and therefore so would be its reduction modulo $2$;
so that $F$ would also have a multiple factor in
$\mathbb{F}_{\mathrm{2}}[x,y,z].$ We have to show that
$\mathrm{gcd}(F,\frac{\partial F}{\partial x},\frac{\partial F}{\partial
y},\frac{\partial F}{\partial z})=1$ in $\mathbb{F}_{2}[x,y,z],$ or
equivalently that
$\displaystyle\mathrm{gcd}(F,\tfrac{\partial F}{\partial x})=1\hskip
2.84526pt\text{in}\hskip 2.84526pt\mathbb{F}_{2}(y,z)[x],$
$\displaystyle\mathrm{gcd}(F,\tfrac{\partial F}{\partial y})=1\hskip
2.84526pt\text{in}\hskip 2.84526pt\mathbb{F}_{2}(x,z)[y],$
$\displaystyle\mathrm{gcd}(F,\tfrac{\partial F}{\partial z})=1\hskip
2.84526pt\text{in}\hskip 2.84526pt\mathbb{F}_{2}(x,y)[z].$
The coordinates $x,y,z$ playing a symmetric role, it suffices again to show
that $\mathrm{gcd}(F,\tfrac{\partial F}{\partial x})=1$ in
$\mathbb{F}_{2}(y,z)[x].$ In $\mathbb{F}_{2}[x,y,z]$ we have
$\displaystyle
F=x^{2d+1}z^{d-1}+y^{2d+1}x^{d-1}+z^{2d+1}y^{d-1}+x^{d}y^{d}z^{d}$ and
$\displaystyle\tfrac{\partial F}{\partial
x}=x^{d-2}\left(x^{d+2}z^{d-1}+dxy^{d}z^{d}+(d+1)y^{2d+1}\right).$
Then $x=0$ is not a root of $F\in\mathbb{F}_{2}(y,z)[x]$ and consequently
$\displaystyle\mathbb{F}_{2}(y,z)[x]\ni\gcd(F,\tfrac{\partial F}{\partial
x})=\gcd(F,\varphi),\quad\text{where}\quad\varphi=x^{d+2}+dxzy^{d}+(d+1)\frac{y^{2d+1}}{z^{d-1}}.$
Moreover, a straightforward computation shows that
$\displaystyle
x^{3}F=\left(x^{d+2}z^{d-1}-(d-1)xy^{d}z^{d}-dy^{2d+1}\right)\varphi+y^{d-1}z^{2d+1}\left(x+\frac{y^{d+1}}{z^{d}}\right)\left(x^{2}+(d^{2}-d-1)\frac{y^{d+1}}{z^{d}}x+d(d+1)\frac{y^{2d+2}}{z^{2d}}\right),$
so that
$\displaystyle\mathbb{F}_{2}(y,z)[x]\ni\gcd(F,\varphi)$
$\displaystyle=\gcd\left(\Big{(}x+\frac{y^{d+1}}{z^{d}}\Big{)}\Big{(}x-\frac{y^{d+1}}{z^{d}}\Big{)},\,\varphi\right),\hskip
2.84526pt\text{because}\hskip 2.84526ptd^{2}-d\equiv d(d+1)\equiv 0\mod 2$
$\displaystyle=\gcd\left(x-\frac{y^{d+1}}{z^{d}},\,x^{d+2}+dxzy^{d}+(d+1)\frac{y^{2d+1}}{z^{d-1}}\right)$
$\displaystyle=\gcd\left(x-\frac{y^{d+1}}{z^{d}},\,x^{d+2}-\frac{y^{2d+1}}{z^{d-1}}\right)$
$\displaystyle=1,$
because
$\left(\dfrac{y^{d+1}}{z^{d}}\right)^{d+2}\neq\dfrac{y^{2d+1}}{z^{d-1}}$ in
the field $\mathbb{F}_{2}(y,z).$ As a result
$\mathbb{F}_{2}(y,z)[x]\ni\gcd(F,\tfrac{\partial F}{\partial x})=1.$
(ii) Let $\mathcal{F}$ be a foliation of degree $d\geq 3$ on
$\mathbb{P}^{2}_{\mathbb{C}}$ with reduced and transverse inflection divisor
$\mathrm{I}_{\mathcal{F}}$, i.e. $\mathcal{F}\in U_{2}(d).$ We want to show
that the set
$\Gamma(\mathcal{F}):=\bigcup\limits_{k=3}^{d}\mathrm{Flex}(\mathcal{F},k-1)$
is finite. By definition of $\Gamma(\mathcal{F})$ we have
(3.7)
$\displaystyle\Gamma(\mathcal{F})\subset\Big{\\{}p\in\mathbb{P}^{2}_{\mathbb{C}}\hskip
2.84526pt|\hskip 2.84526ptp\not\in\mathrm{Sing}(\mathcal{F}),\hskip
2.84526pt\mathrm{Tang}(\mathcal{F},\mathrm{T}^{\mathbb{P}}_{p}\mathcal{F},p)\geq
3\Big{\\}}.$
Let $\mathrm{X}$ be a vector field defining $\mathcal{F}$ in an affine chart
$\mathbb{C}^{2}=\\{(x,y)\\}\subset\mathbb{P}^{2}_{\mathbb{C}}.$ Lemma 3.16 and
inclusion (3.7) imply that $\Gamma(\mathcal{F})\cap\mathbb{C}^{2}$ is
contained in the set of points $p\in\mathbb{C}^{2}$ such that
$\displaystyle\left(\begin{array}[]{c}\mathrm{X}(x)\\\
\mathrm{X}(y)\end{array}\right)(p)\neq\left(\begin{array}[]{c}0\\\
0\end{array}\right),$
$\displaystyle\mathrm{I_{X}}(p):=\left|\begin{array}[]{cc}\mathrm{X}(x)&\mathrm{X}^{2}(x)\\\
\mathrm{X}(y)&\mathrm{X}^{2}(y)\end{array}\right|(p)=0,$
$\displaystyle\mathrm{X}(\mathrm{I_{X}})(p)=\left|\begin{array}[]{cc}\mathrm{X}(x)&\mathrm{X}^{3}(x)\\\
\mathrm{X}(y)&\mathrm{X}^{3}(y)\end{array}\right|(p)=0.$
Now, the affine chart
$\mathbb{C}^{2}=\\{(x,y)\\}\subset\mathbb{P}^{2}_{\mathbb{C}}$ being
arbitrary, $\Gamma(\mathcal{F})$ is finite if and only if
$\Gamma(\mathcal{F})\cap\mathbb{C}^{2}$ is finite. It suffices therefore to
show that the algebraic curves
$\mathrm{I}_{\mathcal{F}}\cap\mathbb{C}^{2}=\\{\mathrm{I_{X}}(x,y)=0\\}$ and
$\mathcal{C}:=\\{\mathrm{X}(\mathrm{I}_{\mathrm{X}})(x,y)=0\\}$ intersect at a
finite number of points, i.e. that they have no common component. Let us argue
by contradiction and assume that there exist
$K,L,L^{\prime}\in\mathbb{C}[x,y],$ with $\deg K>0,$ such that
$\mathrm{I_{X}}=KL$ and $\mathrm{X}(\mathrm{I}_{\mathrm{X}})=KL^{\prime}.$
Then $KL^{\prime}=\mathrm{X}(KL)=\mathrm{X}(K)L+K\mathrm{X}(L)$ and therefore
$\mathrm{X}(K)L=K(L^{\prime}-\mathrm{X}(L)).$ Moreover, the hypothesis that
$\mathrm{I}_{\mathcal{F}}$ is reduced implies that $\mathrm{gcd}(K,L)=1.$ It
follows that there is $L^{\prime\prime}\in\mathbb{C}[x,y]$ such that
$\mathrm{X}(K)=KL^{\prime\prime}$, which means that the algebraic curve
$\mathcal{C}^{\prime}:=\\{K(x,y)=0\\},$ contained in
$\mathrm{I}_{\mathcal{F}}$, is invariant by $\mathcal{F},$ contradicting the
hypothesis that $\mathrm{I}_{\mathcal{F}}$ is transverse. ∎
###### Theorem 3.18.
Let $d$ be an integer greater than or equal to $2.$ Let us denote by
$\Sigma_{2}(d)$ the subset of $\mathbf{F}(d)$ defined by
$\displaystyle\Sigma_{2}(d):=\Big{\\{}\mathcal{F}\in\mathbf{F}(d)\hskip
2.84526pt|\hskip
2.84526pt\mathrm{Flex}(\mathcal{F},d-1)\neq\emptyset\Big{\\}}.$
Then
* (a)
$\mathbf{B}(\mathcal{F}_{2}^{2})=\mathbf{F}(2)\setminus\mathbf{FC}(2)=\Sigma_{2}(2)$
and, for any $d\geq 3$, we have
$\emptyset\neq\Sigma_{2}(d)\varsubsetneq\mathbf{B}(\mathcal{F}_{2}^{d});$
* (b)
$\Sigma_{2}(d)$ is a constructible subset of $\mathbf{F}(d)$;
* (c)
for any $d\geq 3$, we have $\dim\Sigma_{2}(d)\geq\dim\mathbf{F}(d)-(d-3).$
In particular, the set $\Sigma_{2}(3)$, and therefore
$\mathbf{B}(\mathcal{F}_{2}^{3}),$ contains a non-empty Zariski open subset of
$\mathbf{F}(3).$
###### Proof.
(a) As we have already said in Introduction, the first equality
$\mathbf{B}(\mathcal{F}_{2}^{2})=\mathbf{F}(2)\setminus\mathbf{FC}(2)$ follows
from [9, Theorem 3]. The second equality
$\mathbf{F}(2)\setminus\mathbf{FC}(2)=\Sigma_{2}(2)$ is a consequence of the
following obvious remark: if
$\mathcal{F}\in\mathbf{F}(2)\setminus\mathbf{FC}(2)$ then every transverse
inflection point of $\mathcal{F}$ is of order $1$.
The set $\Sigma_{2}(d)$ contains the foliations $\mathcal{H}_{2}^{d}$ and
$\mathcal{F}_{J}^{d}$ (Examples 3.12 and 3.13) and is therefore non-empty.
According to assertion 2. of Proposition 3.11, we have
$\Sigma_{2}(d)\subset\mathbf{B}(\mathcal{F}_{2}^{d})$; this inclusion is
strict for any $d\geq 3$ as Example 3.14 shows.
(b) Let $\pi\hskip
2.84526pt\colon\mathbf{F}(d)\times\mathbb{P}^{2}_{\mathbb{C}}\to\mathbf{F}(d)$
be the projection onto the first factor; notice that
$\Sigma_{2}(d)=\pi(W_{2}(d)),$ where
$\displaystyle W_{2}(d):$
$\displaystyle=\bigcup_{\mathcal{F}\in\Sigma_{2}(d)}\\{\mathcal{F}\\}\times\mathrm{Flex}(\mathcal{F},d-1)$
$\displaystyle=\Big{\\{}(\mathcal{F},p)\in\mathbf{F}(d)\times\mathbb{P}^{2}_{\mathbb{C}}\hskip
2.84526pt|\hskip 2.84526ptp\not\in\mathrm{Sing}(\mathcal{F}),\hskip
2.84526pt\mathrm{Tang}(\mathcal{F},\mathrm{T}^{\mathbb{P}}_{p}\mathcal{F},p)=d\Big{\\}}.$
By Lemma 3.16, $W_{2}(d)$ can be rewritten as
(3.14) $\displaystyle
W_{2}(d)=\Big{\\{}(\mathcal{F},p)\in\mathbf{F}(d)\times\mathbb{P}^{2}_{\mathbb{C}}\hskip
2.84526pt|\hskip 2.84526pt\left(\begin{array}[]{c}\mathrm{X}(x)\\\
\mathrm{X}(y)\end{array}\right)(p)\neq\left(\begin{array}[]{c}0\\\
0\end{array}\right),\hskip
2.84526pt\left|\begin{array}[]{cc}\mathrm{X}(x)&\mathrm{X}^{j}(x)\\\
\mathrm{X}(y)&\mathrm{X}^{j}(y)\end{array}\right|(p)=0,j=2,\ldots,d\Big{\\}},$
where $\mathrm{X}$ denotes a polynomial vector field defining $\mathcal{F}$ in
an affine chart $(x,y)$ containing $p.$ It follows that $W_{2}(d)$ is a quasi-
projective subvariety of $\mathbf{F}(d)\times\mathbb{P}^{2}_{\mathbb{C}}.$
Therefore, by Chevalley’s theorem [11, Exercise II.3.19], the set
$\Sigma_{2}(d)=\pi(W_{2}(d))$ is constructible.
(c) From the above discussion and Proposition 3.15 (i), we have
$\mathcal{F}_{J}^{d}\in U_{2}(d)\cap\Sigma_{2}(d)\neq\emptyset$ ($U_{2}(d)$
being the set of foliations of $\mathbf{F}(d)$ with reduced and transverse
inflection divisor). Therefore there exists an irreducible component
$\Sigma_{2}^{0}(d)$ of $\Sigma_{2}(d)$ such that
$U_{2}(d)\cap\Sigma_{2}^{0}(d)\neq\emptyset.$ We denote by $\pi_{0}\hskip
2.84526pt\colon W_{2}(d)\to\mathbf{F}(d)$ the restriction of $\pi$ to
$W_{2}(d).$ Let $W_{2}(d)=\bigcup\limits_{i=1}^{n}W_{2}^{i}(d)$ be the
decomposition of $W_{2}(d)$ into its irreducible components. Then, by arguing
as in the proof of Theorem 3.10, we see that there is $k\in\\{1,\ldots,n\\}$
such that $\overline{\pi_{0}(W_{2}^{k}(d))}=\overline{\Sigma_{2}^{0}(d)}.$
Since $U_{2}(d)$ is a Zariski open subset of $\mathbf{F}(d)$ (Proposition 3.15
(i)), the morphism $\pi_{0}$ therefore induces by restriction a dominant
morphism of quasi-projective varieties $\pi_{0}^{k}\hskip 2.84526pt\colon
W_{2}^{k}(d)\cap\pi_{0}^{-1}(U_{2}(d))\to\overline{\Sigma_{2}^{0}(d)}\cap
U_{2}(d).$
Notice that, for any $\mathcal{F}\in U_{2}(d)\cap\Sigma_{2}(d),$ the fiber
$\pi_{0}^{-1}(\mathcal{F})$ is finite and non-empty, because
$\pi_{0}^{-1}(\mathcal{F})=\\{\mathcal{F}\\}\times\mathrm{Flex}(\mathcal{F},d-1)$
and $\mathrm{Flex}(\mathcal{F},d-1)$ is finite and non-empty by assertion (ii)
of Proposition 3.15. Since
$\pi_{0}(W_{2}^{k}(d)\cap\pi_{0}^{-1}(U_{2}(d)))\subset
U_{2}(d)\cap\Sigma_{2}(d),$ we deduce that all the non-empty fibers of
$\pi_{0}^{k}$ are finite and therefore zero-dimensional. The fiber dimension
theorem (_cf._ [14, Theorem 3, page 49]) then ensures that
$\dim(W_{2}^{k}(d)\cap\pi_{0}^{-1}(U_{2}(d)))=\dim(\overline{\Sigma_{2}^{0}(d)}\cap
U_{2}(d))$; since $W_{2}^{k}(d)\cap\pi_{0}^{-1}(U_{2}(d))$ and
$\overline{\Sigma_{2}^{0}(d)}\cap U_{2}(d)$ are non-empty open subsets of the
irreducible varieties $W_{2}^{k}(d)$ and $\overline{\Sigma_{2}^{0}(d)}$
respectively, we have
$\displaystyle\hskip
65.44142pt\dim\overline{\Sigma_{2}^{0}(d)}=\dim(\overline{\Sigma_{2}^{0}(d)}\cap
U_{2}(d))=\dim(W_{2}^{k}(d)\cap\pi_{0}^{-1}(U_{2}(d)))=\dim W_{2}^{k}(d).$
Now, it follows from (3.14) that each irreducible component $W_{2}^{i}(d)$ of
$W_{2}(d)$ has dimension
$\dim
W_{2}^{i}(d)\geq\dim(\mathbf{F}(d)\times\mathbb{P}^{2}_{\mathbb{C}})-(d-1)=\dim\mathbf{F}(d)-(d-3),$
hence
$\hskip
44.10185pt\dim\Sigma_{2}(d)=\dim\overline{\Sigma_{2}(d)}\geq\dim\overline{\Sigma_{2}^{0}(d)}=\dim
W_{2}^{k}(d)\geq\dim\mathbf{F}(d)-(d-3).$
The subset $\Sigma_{2}(d)\subset\mathbf{F}(d)$ being constructible, it
contains a dense open subset of its closure $\overline{\Sigma_{2}(d)}$. In
degree $d=3$ we have $\dim\overline{\Sigma_{2}(3)}\geq\dim\mathbf{F}(3)$ and
therefore $\dim\overline{\Sigma_{2}(3)}=\dim\mathbf{F}(3),$ so that
$\overline{\Sigma_{2}(3)}=\mathbf{F}(3)$ because $\mathbf{F}(3)$ is
irreducible. It follows that $\Sigma_{2}(3)$ contains a dense open subset of
$\mathbf{F}(3)$. This ends the proof of the theorem. ∎
###### Remark 3.19.
The set $\mathbf{F}(d)$ contains elements which degenerate onto both
$\mathcal{F}_{1}^{d}$ and $\mathcal{F}_{2}^{d},$ e.g. the family of foliations
$\mathcal{G}^{d}(\gamma)$, $\gamma\in\mathbb{C}.$ Indeed, on the one hand, we
have seen (Example 3.6) that $\mathcal{G}^{d}(\gamma)$ degenerates onto
$\mathcal{F}_{1}^{d}.$ On the other hand, by putting
$\varphi=(\frac{x}{\varepsilon},\frac{y}{\varepsilon})$ we obtain that
$\lim\limits_{\varepsilon\to
0}\varepsilon^{d+1}\varphi^{*}\eta^{d}(\gamma)={\mspace{2.0mu}\overline{\mspace{-1.4mu}\omega\mspace{-1.4mu}}\mspace{2.0mu}}_{2}^{d},$
which shows that $\mathcal{G}^{d}(\gamma)$ degenerates onto the homogeneous
foliation $\mathcal{H}_{2}^{d}$ (Example 3.12) and therefore, by transitivity,
onto $\mathcal{F}_{2}^{d}.$
###### Example 3.20.
Let us consider the homogeneous foliation $\mathcal{H}_{1,2}^{d}$ defined in
the affine chart $z=1$ by the $1$-form
${\mspace{2.0mu}\overline{\mspace{-1.4mu}\omega\mspace{-1.4mu}}\mspace{2.0mu}}_{1,2}^{d}=(x^{d}+y^{d})\mathrm{d}x+x^{d}\mathrm{d}y.$
This foliation degenerates onto both $\mathcal{F}_{1}^{d}$ and
$\mathcal{F}_{2}^{d}.$ Indeed, on the one hand, $\mathcal{H}_{1,2}^{d}$ is
given in the affine chart $y=1$ by
$\hskip
29.87547pt{\mspace{2.0mu}\overline{\mspace{-1.4mu}\theta\mspace{-1.4mu}}\mspace{2.0mu}}_{1,2}^{d}=x\mathrm{d}z-z\mathrm{d}x+x^{d}\mathrm{d}z+x^{d}(x\mathrm{d}z-z\mathrm{d}x);$
we see that the point $[0:1:0]$ is a radial singularity of maximal order $d-1$
of $\mathcal{H}_{1,2}^{d}.$ Thus, by Proposition 3.4, $\mathcal{H}_{1,2}^{d}$
degenerates onto $\mathcal{F}_{1}^{d}.$ On the other hand, a straightforward
computation shows that
$\hskip
56.9055pt\mathrm{Flex}(\mathcal{H}_{1,2}^{d},d-1)=\\{y=0\\}\setminus\\{[0:0:1]\\}\neq\emptyset;$
consequently, $\mathcal{H}_{1,2}^{d}$ also degenerates onto
$\mathcal{F}_{2}^{d}$ (Proposition 3.11).
Since
$\overline{\mathcal{O}(\mathcal{H}_{1,2}^{d})}\subset\mathcal{O}(\mathcal{H}_{1,2}^{d})\cup\mathcal{O}(\mathcal{F}_{1}^{d})\cup\mathcal{O}(\mathcal{F}_{2}^{d})$
(Remark 3.3), we deduce that in fact
$\hskip
19.91684pt\overline{\mathcal{O}(\mathcal{H}_{1,2}^{d})}=\mathcal{O}(\mathcal{H}_{1,2}^{d})\cup\mathcal{O}(\mathcal{F}_{1}^{d})\cup\mathcal{O}(\mathcal{F}_{2}^{d}).$
###### Theorem 3.21.
Let $d$ be an integer greater than or equal to $2.$ Then
* (a)
$\emptyset\neq\Sigma_{1}(d)\cap\Sigma_{2}(d)\subset\mathbf{B}(\mathcal{F}_{1}^{d})\cap\mathbf{B}(\mathcal{F}_{2}^{d})\supset\mathbf{B}(\mathcal{H}_{1,2}^{d});$
* (b)
$\mathbf{B}(\mathcal{H}_{1,2}^{d})$ contains a quasi-projective subvariety of
$\mathbf{F}(d)$ of dimension equal to $\dim\mathbf{F}(d)-3d.$
###### Proof.
(a) The intersection $\Sigma_{1}(d)\cap\Sigma_{2}(d)$ contains the homogeneous
foliation $\mathcal{H}_{1,2}^{d}$ (Example 3.20) and is therefore non-empty.
The inclusion
$\Sigma_{1}(d)\cap\Sigma_{2}(d)\subset\mathbf{B}(\mathcal{F}_{1}^{d})\cap\mathbf{B}(\mathcal{F}_{2}^{d})$
follows from Theorems 3.10 and 3.18.
Let us show the inclusion
$\mathbf{B}(\mathcal{H}_{1,2}^{d})\subset\mathbf{B}(\mathcal{F}_{1}^{d})\cap\mathbf{B}(\mathcal{F}_{2}^{d}).$
Let $\mathcal{F}\in\mathbf{B}(\mathcal{H}_{1,2}^{d})$, i.e.
$\mathcal{F}\in\mathbf{F}(d)$ such that
$\mathcal{H}_{1,2}^{d}\in\overline{\mathcal{O}(\mathcal{F})}.$ Since
$\mathcal{H}_{1,2}^{d}$ degenerates onto $\mathcal{F}_{i}^{d},i=1,2,$ it
follows that
$\mathcal{F}_{i}^{d}\in\overline{\mathcal{O}(\mathcal{H}_{1,2}^{d})}\subset\overline{\mathcal{O}(\mathcal{F})}$,
hence
$\mathcal{F}\in\mathbf{B}(\mathcal{F}_{1}^{d})\cap\mathbf{B}(\mathcal{F}_{2}^{d}).$
(b) Let us denote by $\Sigma(\mathcal{H}_{1,2}^{d})$ the subset of
$\mathbf{F}(d)$ defined as follows: an element $\mathcal{F}$ of
$\mathbf{F}(d)$ belongs to $\Sigma(\mathcal{H}_{1,2}^{d})$ if and only if
* (1)
$\mathcal{F}$ admits an invariant line $\ell$;
* (2)
there is a system of homogeneous coordinates
$[x:y:z]\in\mathbb{P}^{2}_{\mathbb{C}}$ in which $\ell=\\{z=0\\}$ and
$\mathcal{F}$ is defined in the affine chart $z=1$ by a $1$-form $\omega$ of
type
$\displaystyle\omega=\sum\limits_{i=0}^{d-1}\omega_{i}+\lambda{\mspace{2.0mu}\overline{\mspace{-1.4mu}\omega\mspace{-1.4mu}}\mspace{2.0mu}}_{1,2}^{d}=\sum\limits_{i=0}^{d-1}\omega_{i}+\lambda\left((x^{d}+y^{d})\mathrm{d}x+x^{d}\mathrm{d}y\right),$
where $\lambda\in\mathbb{C}^{*}$ and the $\omega_{i}$’s are homogeneous
$1$-forms of degree $i.$
Notice that
$\Sigma(\mathcal{H}_{1,2}^{d})\subset\mathbf{B}(\mathcal{H}_{1,2}^{d}).$
Indeed, by putting $\varphi=(\frac{x}{\varepsilon},\frac{y}{\varepsilon})$ and
by writing $\omega_{i}=P_{i}(x,y)\mathrm{d}x+Q_{i}(x,y)\mathrm{d}y,$ where
$P_{i},Q_{i}\in\mathbb{C}[x,y]_{i},$ we obtain
$\varepsilon^{d+1}\varphi^{*}\omega=\sum_{i=0}^{d-1}(\varepsilon^{d-i}P_{i}(x,y)\mathrm{d}x+\varepsilon^{d-i}Q_{i}(x,y)\mathrm{d}y)+\lambda{\mspace{2.0mu}\overline{\mspace{-1.4mu}\omega\mspace{-1.4mu}}\mspace{2.0mu}}_{1,2}^{d}$
which tends to
$\lambda{\mspace{2.0mu}\overline{\mspace{-1.4mu}\omega\mspace{-1.4mu}}\mspace{2.0mu}}_{1,2}^{d}$
as $\varepsilon$ tends to $0$. It follows that
$\mathcal{H}_{1,2}^{d}\in\overline{\mathcal{O}(\mathcal{F})}$ for any
$\mathcal{F}\in\Sigma(\mathcal{H}_{1,2}^{d}),$ hence the inclusion
$\Sigma(\mathcal{H}_{1,2}^{d})\subset\mathbf{B}(\mathcal{H}_{1,2}^{d})$ holds.
Moreover, every foliation $\mathcal{F}\in\mathbf{F}(d)$ is given in the affine
chart $z=1$ by a $1$-form of type
$\displaystyle\sum\limits_{i=0}^{d}\big{(}A_{i}(x,y)\mathrm{d}x+B_{i}(x,y)\mathrm{d}y\big{)}+C_{d}(x,y)(x\mathrm{d}y-y\mathrm{d}x),$
where $A_{i},B_{i}\in\mathbb{C}[x,y]_{i},C_{d}\in\mathbb{C}[x,y]_{d}$ with
$\mathrm{gcd}\big{(}yC_{d}-\sum\limits_{i=0}^{d}A_{i},xC_{d}+\sum\limits_{i=0}^{d}B_{i}\big{)}=1$.
Condition (2) is then equivalent to taking $C_{d}\equiv
0,\,A_{d}(x,y)=\lambda(x^{d}+y^{d}),\,B_{d}(x,y)=\lambda\,x^{d}.$ Since the
set of foliations of $\mathbf{F}(d)$ admitting an invariant line is a Zariski
closed subset of $\mathbf{F}(d),$ we deduce that
$\Sigma(\mathcal{H}_{1,2}^{d})$ is a quasi-projective subvariety of
$\mathbf{F}(d).$
Since $\omega$ and $\mu\omega$ define the same foliation if $\mu\neq 0,$ and
the choice of a line $\ell\subset\mathbb{P}^{2}_{\mathbb{C}}$ is equivalent to
the choice of a point in $\mathbb{\check{P}}^{2}_{\mathbb{C}},$ conditions (1)
and (2) imply that
$\dim\Sigma(\mathcal{H}_{1,2}^{d})=2+2\sum_{i=0}^{d-1}(i+1)=d^{2}+d+2=\dim\mathbf{F}(d)-3d.$
∎
## 4 A family of foliations of $\mathbf{F}(d)$ with orbits of dimension less
than or equal to $7$
In this section we will establish some properties of the family
$(\mathcal{F}_{0}^{d}(\lambda))_{\lambda\in\mathbb{C}^{*}}$ of foliations of
degree $d$ on $\mathbb{P}^{2}_{\mathbb{C}}$ defined in the affine chart $z=1$
by
$\omega_{0}^{d}(\lambda)=x\mathrm{d}y-\lambda y\mathrm{d}x+y^{d}\mathrm{d}y.$
In homogeneous coordinates, $\mathcal{F}_{0}^{d}(\lambda)$ is given by
$\Omega_{0}^{d}(\lambda)=-\lambda
yz^{d}\mathrm{d}x+z\left(xz^{d-1}+y^{d}\right)\mathrm{d}y+y\left((\lambda-1)xz^{d-1}-y^{d}\right)\mathrm{d}z.$
Thus, the singular locus of $\mathcal{F}_{0}^{d}(\lambda)$ consists of the two
points $s_{1}=[0:0:1]$ and $s_{2}=[1:0:0]$. The singularity $s_{1}$ is non-
degenerate with Baum-Bott index
$\mathrm{BB}(\mathcal{F}_{0}^{d}(\lambda),s_{1})=2+\lambda+\frac{1}{\lambda}$
and the singular point $s_{2}$ has maximal algebraic multiplicity $d.$ We see
that for $\lambda=1$ the $1$-form $\Omega_{0}^{d}(1)$ writes in the affine
chart $x=1$ as
$z^{d}\mathrm{d}y+y^{d}(z\mathrm{d}y-y\mathrm{\mathrm{d}}z);$
we deduce that $\mathcal{F}_{0}^{d}(1)$ is conjugated to the foliation
$\mathcal{F}_{1}^{d}$ and is therefore convex.
In the sequel we assume that $\lambda\in\mathbb{C}\setminus\\{0,1\\}$. A
direct computation, using formula (1.1), leads to
(4.1) $\displaystyle\mathrm{I}_{\mathcal{F}_{0}^{d}(\lambda)}^{\hskip
0.56905pt\mathrm{inv}}=yz^{2d-1}$ and
$\displaystyle\mathrm{I}_{\mathcal{F}_{0}^{d}(\lambda)}^{\hskip
0.56905pt\mathrm{tr}}=(\lambda-1)x-\big{(}(d-1)\lambda+1\big{)}y^{d};$
it follows that, for any $\lambda\in\mathbb{C}\setminus\\{0,1\\}$,
$\mathcal{F}_{0}^{d}(\lambda)$ is not convex.
A straightforward computation shows that the algebraic curve $(1-\lambda\hskip
0.56905ptd)x+y^{d}=0$ is invariant by $\mathcal{F}_{0}^{d}(\lambda).$ What is
more, the rational $1$-form
$\eta_{0}^{d}(\lambda)=\dfrac{\omega_{0}^{d}(\lambda)}{y\left(\left(1-\lambda\hskip
0.56905ptd\right)x+y^{d}\right)}$ is closed. For $\lambda=\frac{1}{d}$ we note
that $\eta_{0}^{d}(\frac{1}{d})=\dfrac{\omega_{0}^{d}(\lambda)}{y^{d+1}}$ has
as first integral $\dfrac{x}{dy^{d}}-\ln y;$ this allows to see that
$\mathrm{Iso}(\mathcal{F}_{0}^{d}(\frac{1}{d}))$ is the group
$\\{(\alpha^{d}x,\alpha y)\hskip 2.84526pt|\hskip
2.84526pt\alpha\in\mathbb{C}^{*}\\}.$ When
$\lambda\in\mathbb{C}\setminus\\{0,1,\frac{1}{d}\\}$ a straightforward
computation shows that $\eta_{0}^{d}(\lambda)$ integrates into
$\lambda\ln\left(\left(1-\lambda\hskip 0.56905ptd\right)x+y^{d}\right)-\ln y,$
which allows to verify that the isotropy group is here again
$\mathrm{Iso}(\mathcal{F}_{0}^{d}(\lambda))=\\{(\alpha^{d}x,\alpha y)\hskip
2.84526pt|\hskip 2.84526pt\alpha\in\mathbb{C}^{*}\\}.$
It follows in particular that, for any
$\lambda\in\mathbb{C}\setminus\\{0,1\\}$,
$\mathcal{O}(\mathcal{F}_{0}^{d}(\lambda))$ has dimension $7.$
Notice that two foliations $\mathcal{F}_{0}^{d}(\lambda)$ and
$\mathcal{F}_{0}^{d}(\lambda^{\prime})$ are conjugated if and only if
$\lambda=\lambda^{\prime}.$
###### Proposition 4.1.
Let $\lambda$ be a nonzero complex number. Let $\mathcal{F}$ be an element of
$\mathbf{F}(d)$ such that
$\mathcal{F}_{0}^{d}(\lambda)\not\in\mathcal{O}(\mathcal{F}).$
1. If $\mathcal{F}$ degenerates onto $\mathcal{F}_{0}^{d}(\lambda)$, then $\mathcal{F}$ admits a non-degenerate singular point $m$ satisfying $\mathrm{BB}(\mathcal{F},m)=2+\lambda+\frac{1}{\lambda}$.
2. If $\mathcal{F}$ possesses a non-degenerate singular point $m$ such that
$\mathrm{BB}(\mathcal{F},m)=2+\lambda+\frac{1}{\lambda}\qquad\text{and}\qquad\kappa(\mathcal{F},m)=d,$
then $\mathcal{F}$ degenerates onto $\mathcal{F}_{0}^{d}(\lambda).$
###### Proof.
It suffices to argue as in the proof of Proposition 3.4, replacing the
foliation $\mathcal{F}_{1}^{d}$ by $\mathcal{F}_{0}^{d}(\lambda)$ and the
equality $\mathrm{BB}(\mathcal{F},m)=4$ by
$\mathrm{BB}(\mathcal{F},m)=2+\lambda+\frac{1}{\lambda}.$ ∎
###### Proposition 4.2.
The orbit $\mathcal{O}\big{(}\mathcal{F}_{0}^{d}(\lambda)\big{)}$ is closed in
$\mathbf{F}(d)$ in the following two cases:
* (i)
$d\geq 3$ and $\lambda=-\dfrac{1}{d-1};$
* (ii)
$d\in\\{3,4,5\\}$ and $\lambda\in\mathbb{C}^{*}.$
The proof of this proposition uses the following lemma.
###### Lemma 4.3.
Let $\lambda$ be a nonzero complex number. Then, the orbit
$\mathcal{O}\big{(}\mathcal{F}_{0}^{d}(\lambda)\big{)}$ is closed in
$\mathbf{F}(d)$ if and only if $\mathcal{F}_{0}^{d}(\lambda)$ does not
degenerate onto $\mathcal{F}_{2}^{d}.$
###### Proof.
The direct implication is obvious. Let us prove the converse. From the above
discussion, $\mathcal{F}_{0}^{d}(1)$ is conjugated to the convex foliation
$\mathcal{F}_{1}^{d}$; therefore its orbit
$\mathcal{O}\big{(}\mathcal{F}_{0}^{d}(1))$ is closed in $\mathbf{F}(d).$ For
any $\lambda\in\mathbb{C}\setminus\\{0,1\\}$, the unique non-degenerate
singular point $s_{1}=[0:0:1]$ of $\mathcal{F}_{0}^{d}(\lambda)$ has Baum-Bott
index
$\mathrm{BB}(\mathcal{F}_{0}^{d}(\lambda),s_{1})=2+\lambda+\frac{1}{\lambda}\neq
4$; this implies, according to assertion 1. of Proposition 3.4, that
$\mathcal{F}_{0}^{d}(\lambda)$ does not degenerate onto $\mathcal{F}_{1}^{d}.$
Moreover, for any $\lambda\in\mathbb{C}\setminus\\{0,1\\}$,
$\mathcal{O}(\mathcal{F}_{0}^{d}(\lambda))$ has dimension $7.$ The converse
implication then follows immediately from Corollary B. ∎
###### Proof of Proposition 4.2.
(i) Let us put $\lambda_{0}=-\frac{1}{d-1}$; according to (4.1) we have
$\mathrm{I}_{\mathcal{F}_{0}^{d}(\lambda_{0})}^{\hskip
0.56905pt\mathrm{tr}}=(\lambda_{0}-1)x,$ hence
$\deg\mathrm{I}_{\mathcal{F}_{0}^{d}(\lambda_{0})}^{\hskip
0.56905pt\mathrm{tr}}=1<d-1$ for any $d\geq 3.$ According to the first
assertion of Proposition 3.11, it follows that, for any $d\geq 3,$ the
foliation $\mathcal{F}_{0}^{d}(\lambda_{0})$ does not degenerate onto
$\mathcal{F}_{2}^{d},$ so that its orbit
$\mathcal{O}\big{(}\mathcal{F}_{0}^{d}(\lambda_{0})\big{)}$ is closed in
$\mathbf{F}(d)$ (Lemma 4.3).
(ii) Let $[x:y:z]$ be homogeneous coordinates in
$\mathbb{P}^{2}_{\mathbb{C}}.$ For $n\in\mathbb{N}$, let us denote by
$\Lambda^{1}_{n}$ the $\mathbb{C}$-vector space of $1$-forms in the variables
$x,y,z,$ whose coefficients are homogeneous polynomials of degree $n.$ Let us
put $\alpha=y\mathrm{d}z-z\mathrm{d}y,$ $\beta=z\mathrm{d}x-x\mathrm{d}z$ and
$\gamma=x\mathrm{d}y-y\mathrm{d}x.$ We have the identification
$\displaystyle\mathbf{F}(d)$ $\displaystyle=$
$\displaystyle\left\\{[\Omega]\in\mathbb{P}(\Lambda^{1}_{d+1})\hskip
2.84526pt|\hskip 2.84526pt\Omega=p\mathrm{d}x+q\mathrm{d}y+r\mathrm{d}z,\hskip
2.84526ptp,q,r\in\mathbb{C}[x,y,z]_{d+1},\hskip
2.84526ptxp+yq+zr=0,\gcd(p,q,r)=1\right\\}$ $\displaystyle=$
$\displaystyle\left\\{[\Omega]\in\mathbb{P}(\Lambda^{1}_{d+1})\hskip
2.84526pt|\hskip 2.84526pt\Omega=A\alpha+B\beta+C\gamma,\hskip
2.84526ptA,B\in\mathbb{C}[x,y,z]_{d},\hskip
2.84526ptC\in\mathbb{C}[x,y]_{d},\hskip 2.84526pt\gcd\big{(}yA-xB,zB-yC,xC-
zA\big{)}=1\right\\}.$
By writting
$\displaystyle A=\xi_{1}\hskip 0.28453ptx^{d}+\xi_{3}\hskip
0.28453ptx^{d-1}y+\cdots+\xi_{2d+1}\hskip
0.28453pty^{d}+\Big{(}\xi_{2d+3}\hskip 0.28453ptx^{d-1}+\xi_{2d+5}\hskip
0.28453ptx^{d-2}y+\cdots+\xi_{4d+1}\hskip
0.28453pty^{d-1}\Big{)}z+\Big{(}\xi_{4d+3}\hskip
0.28453ptx^{d-2}+\xi_{4d+5}\hskip 0.28453ptx^{d-3}y+\cdots+\xi_{6d-1}\hskip
0.28453pty^{d-2}\Big{)}z^{2}+\cdots+\xi_{d^{2}+3d+1}\hskip 0.28453ptz^{d},$
$\displaystyle B=\xi_{2}\hskip 0.28453ptx^{d}+\xi_{4}\hskip
0.28453ptx^{d-1}y+\cdots+\xi_{2d+2}\hskip
0.28453pty^{d}+\Big{(}\xi_{2d+4}\hskip 0.28453ptx^{d-1}+\xi_{2d+6}\hskip
0.28453ptx^{d-2}y+\cdots+\xi_{4d+2}\hskip
0.28453pty^{d-1}\Big{)}z+\Big{(}\xi_{4d+4}\hskip
0.28453ptx^{d-2}+\xi_{4d+6}\hskip 0.28453ptx^{d-3}y+\cdots+\xi_{6d}\hskip
0.28453pty^{d-2}\Big{)}z^{2}+\cdots+\xi_{d^{2}+3d+2}\hskip 0.28453ptz^{d},$
$\displaystyle C=\xi_{d^{2}+3d+3}\hskip 0.28453ptx^{d}+\xi_{d^{2}+3d+4}\hskip
0.28453ptx^{d-1}y+\xi_{d^{2}+3d+5}\hskip
0.28453ptx^{d-2}y^{2}+\cdots+\xi_{d^{2}+4d+2}\hskip
0.28453ptxy^{d-1}+\xi_{d^{2}+4d+3}\hskip 0.28453pty^{d},$
we can identify the class $[\Omega]$ of $\Omega=A\alpha+B\beta+C\gamma$ to the
element $[\xi_{1}:\xi_{2}:\cdots:\xi_{d^{2}+4d+3}]\in\mathbb{P}^{\hskip
0.56905ptd^{2}+4d+2}_{\mathbb{C}}.$ Thus, we can identify $\mathbf{F}(d)$ with
the Zariski open set:
$\displaystyle\left\\{[\xi_{1}:\xi_{2}:\cdots:\xi_{d^{2}+4d+3}]\in\mathbb{P}^{\hskip
0.56905ptd^{2}+4d+2}_{\mathbb{C}}\hskip
1.42262pt\left|\begin{array}[c]{l}\vspace{2mm}A=\xi_{1}\hskip
0.28453ptx^{d}+\xi_{3}\hskip 0.28453ptx^{d-1}y+\cdots+\xi_{2d+1}\hskip
0.28453pty^{d}+\big{(}\xi_{2d+3}\hskip 0.28453ptx^{d-1}+\xi_{2d+5}\hskip
0.28453ptx^{d-2}y+\cdots+\xi_{4d+1}\hskip
0.28453pty^{d-1}\big{)}z+\cdots+\xi_{d^{2}+3d+1}\hskip 0.28453ptz^{d}\\\
\vspace{2mm}B=\xi_{2}\hskip 0.28453ptx^{d}+\xi_{4}\hskip
0.28453ptx^{d-1}y+\cdots+\xi_{2d+2}\hskip
0.28453pty^{d}+\big{(}\xi_{2d+4}\hskip 0.28453ptx^{d-1}+\xi_{2d+6}\hskip
0.28453ptx^{d-2}y+\cdots+\xi_{4d+2}\hskip
0.28453pty^{d-1}\big{)}z+\cdots+\xi_{d^{2}+3d+2}\hskip 0.28453ptz^{d}\\\
\vspace{2mm}C=\xi_{d^{2}+3d+3}\hskip 0.28453ptx^{d}+\xi_{d^{2}+3d+4}\hskip
0.28453ptx^{d-1}y+\xi_{d^{2}+3d+5}\hskip
0.28453ptx^{d-2}y^{2}+\cdots+\xi_{d^{2}+4d+2}\hskip
0.28453ptxy^{d-1}+\xi_{d^{2}+4d+3}\hskip 0.28453pty^{d}\\\ \gcd\big{(}yA-
xB,zB-yC,xC-zA\big{)}=1\end{array}\right.\right\\}.$
Then, via this identification, we have
$\displaystyle\mathcal{F}_{2}^{d}=\big{[}\Omega_{2}^{d}\big{]}=\big{[}x^{d}\beta+y^{d}\gamma]=[0:1:0:0:\cdots:0:0:1\big{]}$
and
$\displaystyle\mathcal{F}_{0}^{d}(\lambda)=\big{[}\Omega_{0}^{d}(\lambda)\big{]}=\big{[}(y^{d}+xz^{d-1})\alpha+\lambda\,yz^{d-1}\beta\big{]}=\big{[}\underbrace{0:0:\cdots:0}_{2d}:1:\underbrace{0:0:\cdots:0}_{d^{2}+d-5}:1:0:0:\lambda:\underbrace{0:0:\cdots:0}_{d+3}\big{]}.$
In addition, the orbit of a foliation $\mathcal{F}=[\Omega]\in\mathbf{F}(d)$
is
$\displaystyle\mathcal{O}(\mathcal{F})=\left\\{[\varphi^{*}\Omega]\hskip
2.84526pt\Big{|}\hskip
2.84526pt\varphi=[a_{1}x+a_{2}y+a_{3}z:a_{4}x+a_{5}y+a_{6}z:a_{7}x+a_{8}y+a_{9}z]\in\mathrm{Aut}(\mathbb{P}^{2}_{\mathbb{C}})\right\\}.$
Let $[x_{1}:x_{2}:\cdots:x_{d^{2}+4d+3}]$ be a system of homogeneous
coordinates in $\mathbb{P}^{\hskip 0.56905ptd^{2}+4d+2}_{\mathbb{C}}.$ For
$d=3,$ let us consider the following homogeneous polynomial in
$x_{1},x_{2},\ldots,x_{24}$ of degree $5$:
$\displaystyle\hskip
8.5359ptP_{3}=-90x_{2}\Big{(}x_{1}\left(294x_{1}-269x_{4}\right)+10x_{2}\left(29x_{3}+4x_{6}\right)+86x_{4}^{2}\Big{)}x_{22}x_{24}-1125x_{2}^{2}\left(21x_{1}-23x_{4}\right)x_{23}x_{24}$
$\displaystyle\hskip
24.75375pt+45x_{2}\Big{(}2x_{3}\left(294x_{1}+13x_{4}\right)-x_{6}\left(552x_{1}-271x_{4}\right)+1125x_{2}x_{5}\Big{)}x_{21}x_{24}+28125x_{2}x_{10}x_{21}x_{23}x_{24}$
$\displaystyle\hskip
24.75375pt+25\Big{(}108\left(x_{9}-2x_{12}\right)\left(3x_{1}-4x_{4}\right)+9x_{10}\left(112x_{3}-93x_{6}\right)+675x_{2}x_{11}\Big{)}x_{21}^{2}x_{24}-6000x_{2}x_{10}x_{22}^{2}x_{24}$
$\displaystyle\hskip
24.75375pt-5625x_{5}x_{11}x_{21}^{3}+20\Big{(}\left(2x_{1}-x_{4}\right)\left(41x_{9}-7x_{12}\right)+30x_{10}\left(2x_{3}-3x_{6}\right)+50x_{2}x_{11}\Big{)}x_{22}^{3}-50625x_{2}^{3}x_{24}^{2}$
$\displaystyle\hskip
24.75375pt-5\Big{(}2x_{9}\left(207x_{1}-116x_{4}\right)-x_{12}\left(153x_{1}-314x_{4}\right)+5x_{10}\left(356x_{3}-359x_{6}\right)+1350x_{2}x_{11}\Big{)}x_{21}x_{22}x_{23}$
$\displaystyle\hskip
24.75375pt+1875\Big{(}x_{11}\left(2x_{3}-x_{6}\right)+x_{5}\left(2x_{9}-x_{12}\right)\Big{)}x_{21}^{2}x_{22}-375x_{2}\Big{(}2x_{1}\left(3x_{1}-7x_{4}\right)-x_{2}\left(3x_{3}-2x_{6}\right)+8x_{4}^{2}\Big{)}x_{23}^{2}$
$\displaystyle\hskip
24.75375pt+50\Big{(}5x_{10}\left(39x_{1}-38x_{4}\right)-3x_{2}\left(x_{9}-32x_{12}\right)\Big{)}x_{21}x_{23}^{2}-50\Big{(}x_{10}\left(14x_{1}-37x_{4}\right)-3x_{2}\left(7x_{9}+x_{12}\right)\Big{)}x_{22}^{2}x_{23}$
$\displaystyle\hskip
24.75375pt+15\Big{(}5x_{11}\left(21x_{1}+22x_{4}\right)-8x_{3}\left(14x_{9}-43x_{12}\right)+6x_{6}\left(13x_{9}-56x_{12}\right)-350x_{5}x_{10}\Big{)}x_{21}^{2}x_{23}+R\,x_{21}^{2}$
$\displaystyle\hskip
24.75375pt-5\Big{(}20x_{11}\left(24x_{1}-7x_{4}\right)+4x_{9}\left(97x_{3}-43x_{6}\right)+x_{12}\left(94x_{3}-211x_{6}\right)-600x_{5}x_{10}\Big{)}x_{21}x_{22}^{2}+S\,x_{21}x_{22}$
$\displaystyle\hskip
24.75375pt-75\Big{(}2x_{10}\left(78x_{1}-29x_{4}\right)-15x_{2}\left(2x_{9}-19x_{12}\right)\Big{)}x_{21}x_{22}x_{24}+125x_{2}x_{10}x_{22}x_{23}^{2}+Tx_{22}^{2}+Ux_{21}x_{23}$
$\displaystyle\hskip 24.75375pt+Vx_{22}x_{23},$
where
$\displaystyle\hskip
8.5359ptR=5568x_{6}x_{5}\left(3x_{1}-4x_{4}\right)-18x_{3}x_{5}\left(1612x_{1}-1941x_{4}\right)+6x_{3}^{2}\left(1952x_{3}-4389x_{6}\right)+3x_{6}^{2}\left(7057x_{3}-2136x_{6}\right)-11250x_{2}x_{5}^{2}$
$\displaystyle\hskip
23.33147pt+2700x_{7}\left(3x_{1}-4x_{4}\right)^{2}+54x_{8}\left(3x_{1}-4x_{4}\right)\left(106x_{3}-89x_{6}\right),$
$\displaystyle\hskip
8.5359ptS=27000x_{2}x_{7}\left(3x_{1}-4x_{4}\right)-24x_{3}^{2}\left(658x_{1}-249x_{4}\right)+1512x_{4}x_{8}\left(11x_{1}-4x_{4}\right)+252x_{1}^{2}\left(83x_{5}-36x_{8}\right)-90x_{2}x_{3}\left(329x_{5}-318x_{8}\right)$
$\displaystyle\hskip
23.33147pt-2x_{4}x_{5}\left(17073x_{1}-6047x_{4}\right)+3x_{1}x_{6}\left(8712x_{3}-3599x_{6}\right)-x_{4}x_{6}\left(11658x_{3}-6041x_{6}\right)+90x_{2}x_{6}\left(226x_{5}-267x_{8}\right),$
$\displaystyle\hskip
8.5359ptT=20x_{1}x_{3}\left(294x_{1}-253x_{4}\right)-40x_{1}x_{6}\left(159x_{1}-152x_{4}\right)+1900x_{2}x_{3}\left(x_{3}-x_{6}\right)+20x_{4}^{2}\left(68x_{3}-95x_{6}\right)-25x_{2}x_{6}\left(40x_{3}-33x_{6}\right)$
$\displaystyle\hskip
23.33147pt+60x_{1}x_{2}\left(361x_{5}-252x_{8}\right)-10x_{2}x_{4}\left(983x_{5}-756x_{8}\right)+67500x_{2}^{2}x_{7},$
$\displaystyle\hskip
8.5359ptU=90x_{1}x_{3}\left(98x_{1}-117x_{4}\right)-30x_{1}x_{6}\left(171x_{1}-284x_{4}\right)-150x_{2}x_{6}\left(68x_{3}-35x_{6}\right)-30x_{2}x_{4}\left(167x_{5}+396x_{8}\right)+7050x_{2}x_{3}^{2}$
$\displaystyle\hskip
23.33147pt+20x_{4}^{2}\left(73x_{3}-157x_{6}\right)+270x_{1}x_{2}\left(41x_{5}+33x_{8}\right),$
$\displaystyle\hskip
8.5359ptV=5x_{2}x_{4}\left(1604x_{3}-611x_{6}\right)-30x_{1}^{2}\left(294x_{1}-563x_{4}\right)-30x_{4}^{2}\left(355x_{1}-86x_{4}\right)-30x_{1}x_{2}\left(463x_{3}-242x_{6}\right)-75x_{2}^{2}\left(109x_{5}-198x_{8}\right).$
A computation carried out with Maple shows that evaluating $P_{3}$ at an
arbitrary element $[\xi_{1}:\xi_{2}:\cdots:\xi_{24}]$ of
$\mathcal{O}\big{(}\mathcal{F}_{0}^{3}(\lambda)\big{)},$ we find
$P_{3}\big{(}[\xi_{1}:\xi_{2}:\cdots:\xi_{24}]\big{)}=0,$ i.e.
$\mathcal{O}\big{(}\mathcal{F}_{0}^{3}(\lambda)\big{)}$ is contained in the
zero locus of $P_{3}$
$\displaystyle\mathrm{\text{Zeros}}(P_{3}):=\left\\{[x_{1}:x_{2}:\cdots:x_{24}]\in\mathbb{P}^{23}_{\mathbb{C}}\hskip
2.84526pt|\hskip
2.84526ptP_{3}\big{(}[x_{1}:x_{2}:\cdots:x_{24}]\big{)}=0\right\\},$
which is a Zariski closed subset of $\mathbb{P}^{23}_{\mathbb{C}}.$ Therefore
we have
$\overline{\mathcal{O}\big{(}\mathcal{F}_{0}^{3}(\lambda)\big{)}}\subset\mathrm{\text{Zeros}}(P_{3})$
for any $\lambda\in\mathbb{C}^{*}.$ Moreover, we have
$\displaystyle P_{3}\left(0,1,0,0,\cdots,0,0,1\right)=-50625\neq 0,$
hence $\mathcal{F}_{2}^{3}\not\in\mathrm{\text{Zeros}}(P_{3}).$ It follows
that, for any $\lambda\in\mathbb{C}^{*},$ we have
$\mathcal{F}_{2}^{3}\not\in\overline{\mathcal{O}\big{(}\mathcal{F}_{0}^{3}(\lambda)\big{)}}$,
so that $\mathcal{F}_{0}^{3}(\lambda)$ does not degenerate onto
$\mathcal{F}_{2}^{3}.$ Consequently, according to Lemma 4.3, the orbit
$\mathcal{O}\big{(}\mathcal{F}_{0}^{3}(\lambda)\big{)}$ is closed in
$\mathbf{F}(3)$.
To show that the orbit
$\mathcal{O}\big{(}\mathcal{F}_{0}^{4}(\lambda)\big{)},$ resp.
$\mathcal{O}\big{(}\mathcal{F}_{0}^{5}(\lambda)\big{)},$ is closed in
$\mathbf{F}(4),$ resp. $\mathbf{F}(5),$ it suffices to argue as in degree
$d=3$, replacing the polynomial $P_{3}$ by the following polynomial $P_{4},$
resp. $P_{5}$:
$\displaystyle
P_{4}=\Big{(}3x_{3}\left(129x_{3}-212x_{6}\right)+3x_{4}\left(178x_{5}+15x_{8}\right)+12x_{1}\left(22x_{5}-3x_{8}\right)+5184x_{2}x_{7}-20x_{6}^{2}\Big{)}x_{31}+1728x_{15}x_{31}^{2}$
$\displaystyle\hskip
16.21828pt-432\left(2x_{13}-x_{16}\right)x_{31}x_{32}+48\left(42x_{11}-31x_{14}\right)x_{31}x_{33}-18\left(24x_{11}-19x_{14}\right)x_{32}^{2}-162x_{2}\left(4x_{1}-15x_{4}\right)x_{34}$
$\displaystyle\hskip
16.21828pt-18\Big{(}2x_{1}\left(27x_{3}-20x_{6}\right)-x_{4}\left(15x_{3}-x_{6}\right)+x_{2}\left(170x_{5}-69x_{8}\right)\Big{)}x_{32}+4212x_{12}x_{31}x_{34}-486x_{12}x_{32}x_{33}$
$\displaystyle\hskip
16.21828pt+36\Big{(}3\left(x_{1}-x_{4}\right)\left(12x_{1}-x_{4}\right)+22x_{2}\left(3x_{3}-2x_{6}\right)\Big{)}x_{33}-10368x_{2}^{2}x_{35},$
resp. $\displaystyle
P_{5}=\Big{(}50x_{7}\left(4906x_{1}-4749x_{4}\right)-27040x_{10}\left(5x_{1}-6x_{4}\right)-5x_{5}\left(10596x_{3}-13469x_{6}\right)+20x_{8}\left(1019x_{3}-2028x_{6}\right)$
$\displaystyle\hskip
16.21828pt+569100x_{2}x_{9}\Big{)}x_{43}+142275x_{19}x_{43}^{2}-11690x_{17}x_{43}x_{44}+98140x_{14}x_{43}x_{47}-140x_{2}\left(2180x_{1}-1691x_{4}\right)x_{47}$
$\displaystyle\hskip
16.21828pt+35\left(1564x_{13}-1645x_{16}\right)x_{43}x_{46}+\Big{(}8620x_{8}\left(2x_{1}-x_{4}\right)-50x_{5}\left(141x_{1}-11x_{4}\right)+10x_{3}\left(513x_{3}-1580x_{6}\right)$
$\displaystyle\hskip
16.21828pt+70x_{2}\left(2779x_{7}-2704x_{10}\right)+9875x_{6}^{2}\Big{)}x_{44}-35\Big{(}\left(x_{1}-x_{4}\right)\left(295x_{1}+683x_{4}\right)-x_{2}\left(3776x_{3}-4427x_{6}\right)\Big{)}x_{46}$
$\displaystyle\hskip
16.21828pt+70\left(323x_{18}-253x_{15}\right)x_{43}x_{45}+7\left(686x_{13}-293x_{16}\right)x_{44}x_{45}-2975x_{15}x_{44}^{2}-15946x_{14}x_{45}^{2}-1422750x_{2}^{2}x_{48}$
$\displaystyle\hskip
16.21828pt+\Big{(}14x_{3}\left(15x_{1}+1124x_{4}\right)-14x_{6}\left(10x_{1}+1129x_{4}\right)-595x_{2}\left(221x_{5}-250x_{8}\right)\Big{)}x_{45}+49210x_{14}x_{44}x_{46}.$
∎
For $d\geq 6,$ we propose:
###### Conjecture 1.
Let $d$ be an integer greater than or equal to $6$ and $\lambda$ a nonzero
complex number. A homogeneous coordinate system
$[x_{1}:x_{2}:\cdots:x_{d^{2}+4d+3}]$ being fixed in $\mathbb{P}^{\hskip
0.56905ptd^{2}+4d+2}_{\mathbb{C}},$ there exists a homogeneous polynomial
$Q_{d}\in\mathbb{C}[x_{1},x_{2},\cdots,x_{d^{2}+4d+3}]$ of degree $3$, not
depending on $\lambda,$ which vanishes on the orbit
$\mathcal{O}\big{(}\mathcal{F}_{0}^{d}(\lambda)\big{)}$ and does not vanish at
the point $\mathcal{F}_{2}^{d}=[0:1:0:0:\cdots:0:0:1\big{]}.$
Computations made with Maple by the first author show the validity of this
conjecture for $d$ small ($d\leq 30$) by taking the polynomial $Q_{d}$ in the
following form:
$\displaystyle
Q_{d}=x_{d^{2}+3d+3}\left(\sum\limits_{i=1}^{d-1}\alpha_{i}\,x_{2d+2i+1}\,x_{d^{2}+4d+2-i}+\sum\limits_{i=0}^{4}\beta_{i}\,x_{2d+2i+4}\,x_{d^{2}+4d+2-i}\right)+(x_{1}\hskip
5.69054ptx_{2}\hskip 5.69054pt\cdots\hskip
5.69054ptx_{d+1})M\left(\begin{array}[]{c}x_{d^{2}+4d+3}\\\ x_{d^{2}+4d+2}\\\
\vdots\\\ x_{d^{2}+3d+3}\end{array}\right)$ $\displaystyle\hskip
20.48601pt+x_{d^{2}+3d+4}\left(\delta_{0}\,x_{2d+4}\,x_{d^{2}+4d+1}+\delta_{1}\,x_{2d+6}\,x_{d^{2}+4d}+\sum\limits_{i=1}^{d-3}\gamma_{i}\,x_{2d+2i+1}\,x_{d^{2}+4d+1-i}\right),$
where $M=\left(\begin{array}[]{c}L_{1}\\\ L_{2}\\\ \vdots\\\
L_{d+1}\end{array}\right)$ is a square matrix of order $d+1$ whose lines are
of the form:
$\displaystyle L_{1}=\left[0\hskip 5.69054pt0\hskip
5.69054pta_{1,3}x_{1}+b_{1,3}x_{4}\hskip
5.69054pta_{1,4}x_{3}+b_{1,4}x_{6}\hskip
5.69054pta_{1,5}x_{5}+b_{1,5}x_{8}\hskip 5.69054pt\cdots\hskip
5.69054pta_{1,\,d+1}x_{2d-3}+b_{1,\,d+1}x_{2d}\right]$ $\displaystyle
L_{2}=\left[b_{2,1}x_{2}\hskip 5.69054pta_{2,2}x_{1}+b_{2,2}x_{4}\hskip
5.69054pta_{2,3}x_{3}+b_{2,3}x_{6}\hskip
5.69054pta_{2,4}x_{5}+b_{2,4}x_{8}\hskip
5.69054pta_{2,5}x_{7}+b_{2,5}x_{10}\hskip 5.69054pt\cdots\hskip
5.69054pta_{2,\,d+1}x_{2d-1}+b_{2,\,d+1}x_{2d+2}\right]$ $\displaystyle\vdots$
$\displaystyle L_{2k-1}=\left[\underbrace{0\hskip 5.69054pt0\hskip
5.69054pt\cdots\hskip 5.69054pt0}_{\min(2k,\,d+1)}\hskip
5.69054pta_{2k-1,\,2k+1}x_{2k-1}+b_{2k-1,\,2k+1}x_{2k+2}\hskip
5.69054pta_{2k-1,\,2k+2}x_{2k+1}+b_{2k-1,\,2k+2}x_{2k+4}\hskip
5.69054pt\cdots\hskip
5.69054pta_{2k-1,\,d+1}x_{2d-2k-1}+b_{2k-1,\,d+1}x_{2d-2k+2}\right]$
$\displaystyle L_{2k}=\left[\underbrace{0\hskip 5.69054pt0\hskip
5.69054pt\cdots\hskip 5.69054pt0}_{2k-2}\hskip
5.69054ptb_{2k,\,2k-1}x_{2k}\hskip
5.69054pta_{2k,\,2k}x_{2k-1}+b_{2k,\,2k}x_{2k+2}\hskip
5.69054pta_{2k,\,2k+1}x_{2k+1}+b_{2k,\,2k+1}x_{2k+4}\hskip
5.69054pt\cdots\hskip
5.69054pta_{2k,\,d+1}x_{2d-2k+1}+b_{2k,\,d+1}x_{2d-2k+4}\right],$
where
$\alpha_{i},\beta_{i},\gamma_{i},\delta_{i},a_{i,j},b_{i,j}\in\mathbb{C}$ with
$b_{2,1}\neq 0.$
It is clear that Conjecture 1 and Lemma 4.3 imply the following conjecture.
###### Conjecture 2.
For any integer $d\geq 6$ and any $\lambda\in\mathbb{C}^{*},$ the orbit
$\mathcal{O}\big{(}\mathcal{F}_{0}^{d}(\lambda)\big{)}$ is closed in
$\mathbf{F}(d).$
## References
* [1] C. R. Alcántara and R. Ronzón-Lavie. Classification of foliations on $\mathbb{CP}^{2}$ of degree $3$ with degenerate singularities. J. Singul. 14:52–73, 2016.
* [2] P. Baum and R. Bott. Singularities of holomorphic foliations. J. Differential Geometry, 7:279–342, 1972.
* [3] S. Bedrouni. Feuilletages de degré trois du plan projectif complexe ayant une transformée de Legendre plate. PhD thesis, University of Sciences and Technology Houari Boumediene, 2017. Available on https://arxiv.org/abs/1712.03895.
* [4] S. Bedrouni and D. Marín. Tissus plats et feuilletages homogènes sur le plan projectif complexe. Bull. Soc. Math. France, 146(3):479–516, 2018.
* [5] S. Bedrouni and D. Marín. Classification of foliations of degree three on $\mathbb{P}^{2}_{\mathbb{C}}$ with a flat Legendre transform. To appear in Ann. Inst. Fourier (Grenoble), 2019.
* [6] S. Bedrouni and D. Marín. Une nouvelle démonstration de la classification des feuilletages convexes de degré deux sur $\mathbb{P}^{2}_{\mathbb{C}}$. Bull. Soc. Math. France, 148(4):613–622, 2020.
* [7] M. Brunella. Birational geometry of foliations, volume 1 of IMPA Monographs. Springer, Cham, 2015.
* [8] F. Cano, D. Cerveau, and J. Déserti. Théorie élémentaire des feuilletages holomorphes singuliers. Echelles. Belin, 2013.
* [9] D. Cerveau, J. Déserti, D. Garba Belko, and R. Meziani. Géométrie classique de certains feuilletages de degré deux. Bull. Braz. Math. Soc. (N.S.), 41(2):161–198, 2010.
* [10] C. Favre and J. V. Pereira. Webs invariant by rational maps on surfaces. Rend. Circ. Mat. Palermo (2), 64(3):403–431, 2015.
* [11] R. Hartshorne. Algebraic geometry. Springer-Verlag, New York-Heidelberg, 1977. Graduate Texts in Mathematics, No. 52.
* [12] J. P. Jouanolou. Équations de Pfaff algébriques, volume 708 of Lecture Notes in Mathematics. Springer, Berlin, 1979.
* [13] A. Lins Neto and J. V. Pereira. The generic rank of the Baum-Bott map for foliations of the projective plane. Compos. Math., 142(6):1549–1586, 2006.
* [14] D. Mumford. The red book of varieties and schemes, volume 1358 of Lecture Notes in Mathematics. Springer-Verlag, Berlin, 1988.
* [15] D. Marín and J. V. Pereira. Rigid flat webs on the projective plane. Asian J. Math. 17(1):163–191, 2013.
* [16] J. V. Pereira. Vector fields, invariant varieties and linear systems. Ann. Inst. Fourier (Grenoble), 51(5):1385–1405, 2001.
* [17] K. Saito. On a generalization of de-Rham lemma. Ann. Inst. Fourier (Grenoble), 26(2):vii, 165–170, 1976.
|
# State estimation with limited sensors – A deep learning based approach
Yash Kumar
Department of Mechanical Engineering
Delhi Technological University
Shahbad Daulatpur, Main Bawana Road, Delhi-110042, India
<EMAIL_ADDRESS>
&Pranav Bahl
Department of Mechanical Engineering
Delhi Technological University
Shahbad Daulatpur, Main Bawana Road, Delhi-110042, India
<EMAIL_ADDRESS>
&Souvik Chakraborty
Department of Applied Mechanics
Indian Institute of Technology Delhi
Hauz Khas - 110042, New Delhi, India
<EMAIL_ADDRESS>
###### Abstract
The importance of state estimation in fluid mechanics is well-established; it
is required for accomplishing several tasks, including design/optimization,
active control, and future state prediction. A common tactic in this regard is
to rely on reduced-order models. Such approaches, in general, use measurement
data of a one-time instance. However, often data available from sensors is
sequential, and ignoring it results in information loss. In this paper, we
propose a novel deep learning-based state estimation framework that learns
from sequential data. The proposed model structure consists of the recurrent
cell to pass information from different time steps, enabling this information
to recover the full state. We illustrate that utilizing sequential data allows
for state recovery from minimal and noisy sensor measurements. For efficient
recovery of the state, the proposed approach is coupled with an auto-encoder
based reduced-order model. We illustrate the performance of the proposed
approach using three examples, and it is found to outperform other
alternatives existing in the literature.
_K_ eywords state estimation $\cdot$ autoencoder $\cdot$ recurrent neural
network $\cdot$ limited sensors
## 1 Introduction
The integration of deep learning has benefited modern algorithms in modeling,
data processing, prediction, and control of various engineering systems. In
fluid mechanics, work on machine learning implementation started last decade
and has grown since then. [1] used a neural network to reconstruct turbulence
flow fields and the flow in the near-wall region of a channel flow using wall
information. [2] and [3], in two separate works, used deep learning algorithms
to improve a Reynolds-averaged Navier Stokes turbulence model. Recently, [4]
proposed a multi-fidelity deep learning framework for turbulent flow. [5] used
a shallow network for estimating 2D state from measurements over cylinder
surface. In this paper, we are particularly interested in applying deep
learning for state estimation in fluid mechanics, and hence, the discussion
hereafter is focused on the same.
State estimation is the ability to recover flow based on a few noisy
measurements. It is an inverse problem and arises in many engineering
applications such as remote sensing, medical imaging, ocean dynamics,
reservoir modeling, and blood flow modeling. Controlling the flow and
optimizing machine design in these applications depend upon the ability to
predict the state with given sensors. The challenge associated with state
estimation is two-fold. Firstly, for almost all practical cases, state-
estimation is an ill-posed problem, and hence, a unique solution to the
problem does not exist [6]. Secondly, for practical problems, the number of
sensors available is often limited. As a consequence, one has to deal with a
sparse data set [7].
Attempts for state estimation dates back to 1960 when the Kalman filter based
approaches were used for state estimation [8]. This method assumes the
system’s dynamics to produce full state and updates it based on new
measurements to reduce estimation error forming a closed feedback loop.
However, the classical Kalman filter based approaches are only applicable for
linear dynamical systems [9]. Improvements to Kalman filter algorithms, such
as Extended Kalman filter [10] and Unscented Kalman filter [11] algorithms can
also be found in the literature. [12] attempts to generalize Kalman filters by
using the Gaussian process. Approaches based on observer dynamical system uses
a reduced-order model to predict the future based on the past while
simultaneously corrected by receiving measurements. [13] applies dynamic mode
decomposition as a reduced-order model to Kalman smoother estimate to identify
coherent structures. [14] used a nonlinear observer-based on Galerkin
projection of Navier-Stokes equation to estimate POD coefficients. The use of
Bayes filters, such as the Kalman and particle filters, in conjugation with
POD based ROM on various flow problems, can also be found in the literature
[15, 16, 17].
Another major category of approaches includes library-based approaches and
stochastic approaches. Library-based approaches use offline data, and the
library consists of generic modes such as Fourier, wavelet, discrete cosine
transform basis or data specific POD or DMD modes, or training data. Library
based approaches using sparse representation assumes state can be expressed as
the combination of library elements. Sparse coefficients are obtained by
solving pursuit problem [18, 19]. [20] used sparse representation and training
data as the library with localized reconstruction for reconstructing complex
fluid flows. Gappy POD [21] estimates POD coefficients in a least-square sense
and uses a library of POD modes. However, it is prone to ill-conditioning and
is dealt with using the best sensor placements [22] to improve the condition
number[23].
The most explored approach for state estimation is perhaps the one based on
stochastic estimation. The idea was first proposed by [24] for a turbulence
study where the conditional mean was approximated using a power series. In a
linear setting, coefficients are computed using a two-point, second-order
correlation tensor. Other variants like quadratic stochastic estimation [25]
and spectral linear stochastic estimation [26] can also be found in the
literature. [27] proposed to include time delayed measurements to further
improve accuracy. [28] extended stochastic approach to estimate POD
coefficients. A linear mapping between sensors and coefficients was assumed.
Recently, [29] used a neural network to learn a non-linear mapping between
sensor measurements and POD coefficients. These approaches allow more
flexibility in sensor placements and have been applied for flow control over
airfoil [30] and analyzing isotropic turbulence [31, 32] and turbulent
boundary layers [27, 25].
One limitation associated with all the approaches mentioned above resides in
the fact that spatial information of a single sample is used to recover the
full state, but often, data is sequential. Ignoring the sequence of the data
during state information invariably results in information loss. To address
this apparent shortcoming, we propose a novel deep learning-based non-
intrusive framework for state estimation that learns from sequential data. The
proposed framework couples recurrent neural network (RNN) with auto-encoder
(AE). While AE is used to learn the nonlinear manifold, RNN is employed to
take advantage of the time series data. We illustrate that by utilizing
sequential data, the proposed framework is able to estimate the state in a
more accurate fashion. Perhaps, more importantly, the number of sensors
required is significantly less. For showcasing the performance of the proposed
framework, two benchmark problems involving periodic vortex shedding and
transient flow past a cylinder are considered. Results obtained are compared
with those obtained using other state-of-the-art techniques.
The remainder of the paper is organized as follows. In Section 2, details on
the problem statement is provided. A brief review of RNN and AE are furnished
in Section 3. Details on the proposed approach are provided in Section 4.
Section 5 presents two numerical examples to illustrate the performance of the
proposed approach. Finally, Section 6 provides the concluding remarks.
## 2 Problem statement
We consider a dynamical system obtained by partial discretization of the
governing differential equations:
$\dot{\bm{w}}=\bm{f}\left(\bm{w},t;\bm{\theta}\right),\;\;\;\bm{w}(t_{n},\bm{\theta})=\bm{w}^{n}\left(\bm{\theta}\right),$
(1)
where $\bm{w}\in\mathbb{R}^{N_{w}}$ represents the high-dimensional state
vector that depends on parameters $\bm{\theta}\in\mathbb{R}^{N_{d}}$ and time
$t\in\left[0,t_{max}\right]$. $\bm{f}$ in Eq. (1) is a nonlinear function that
governs the dynamical evolution of the state vector $\bm{w}$. Note that for
brevity of representation, we have not shown the dependence of the state
$\bm{w}$ on $t$ and $\bm{\theta}$, and the dependence of $\bm{f}$ on
$\bm{w},t$ and $\bm{\theta}$.
We note that the state vector $\bm{w}$ is high-dimensional in nature and it is
extremely difficult to directly work with $\bm{w}$. A commonly used strategy
in this regards is to approximate the high-dimensional state vector $\bm{w}$
on a low-dimensional manifold,
$\bm{w}\left(t;\bm{\theta}\right)\approx\bm{w}_{r}\left(t;\bm{\theta}\right)=\bm{\Phi}\left(\bm{a}\left(t;\bm{\theta}\right)\right),$
(2)
where $\bm{a}\left(t;\bm{\theta}\right)\in\mathbb{R}^{N_{a}}$ represents the
reduced space and
$\bm{\Phi}\left(\cdot\right):\bm{a}\left(t;\bm{\theta}\right)\mapsto\bm{w}\left(t;\bm{\theta}\right)$
is the manifold. $N_{a}$ in Eq. (2) represent the dimension of reduced space
such that $N_{a}\ll N_{w}$. Substituting Eq. (2) into Eq. (1), we obtain
$\dot{\bm{a}}=\bm{\Psi}\left(\bm{f}\left(\bm{\Phi}\left(a\right),t;\bm{\theta}\right)\right),\;\bm{a}(t_{n};\bm{\theta})=\bm{a}^{n}(\bm{\theta}),$
(3)
where
$\bm{\Psi}(\cdot)=\left(\bm{\Lambda}\bm{\Gamma}^{-1}\right)\circ\bm{\Lambda}\left(\cdot\right)$
and $\bm{\Lambda}:\mathbb{R}^{N_{w}}\mapsto\mathbb{R}^{N_{a}}$. In Eq. (3), we
have assumed that $\bm{\Phi}(\bm{a})$ is continuously differentiable such that
$\bm{\Gamma}(\dot{\bm{a}})=\dot{\bm{\Phi}}(\bm{a})$;
$\bm{\Gamma}:\mathbb{R}^{N_{a}}\mapsto\mathbb{R}^{N_{w}}$.
$a^{n}(\bm{\theta})$ in Eq. (3) represents the initial condition. With this
representation, the objective in state-estimation reduces to estimating the
reduced order state variable $\bm{a}^{n}\left(\bm{\theta}\right)$. Generally,
this is achieved by determining a mapping
$\mathcal{M}:\mathbb{R}^{N_{s}}\mapsto\mathbb{R}^{N_{a}}$ between the sensor
measurements and reduced state,
$\bm{a}^{n}=\mathcal{M}(\bm{s}^{n}(\bm{\theta})),$ (4)
where $\bm{s}^{n}\in\mathbb{R}^{N_{s}}$ represents the sensor measurements at
time-step $n$ and $N_{s}$ indicates the number of sensors present. A schematic
representation of the same is shown in Fig. 1.
Figure 1: Schematic representation of state estimation framework. It consists
of two components, (a) $\mathcal{M}(\bm{s})$ \- it maps the sensor data to a
reduced state and (b) a projection operator - it maps the predicted reduced
state to the original state. Performance of a state estimation framework is
dependent on performance of this two components and the quality and quantity
of data.
The state estimation framework discussed above has two major limitations.
* •
We note that the state estimation framework only relies on sensor responses at
the current step for predicting the state variables. In other words, the
sequential nature of the sensor measurements is ignored. This invariably
results in loss of information, and hence, the accuracy of the state
estimation is compromised.
* •
Secondly, as shown in Eq. (2), the use of reduced-order model results in
information loss. While completely avoiding this information loss is
unavoidable, it is necessary to ensure that this information loss is
minimized.
This paper aims to develop a deep learning-based framework for state
estimation that addresses the two limitations discussed above.
## 3 A brief review of AE and RNN
This section briefly reviews two poplar deep learning approaches, namely auto-
encoders (AE) and recurrent neural networks (RNN). It is to be noted that AE
and RNN form the backbone of the proposed approach.
### 3.1 Auto-encoders
AE is a class of unsupervised deep learning techniques trained to copy the
inputs to the output. It consists of a latent space/hidden layer
$\bm{h}\in\mathbb{R}^{h}$ that represents a compressed representation of the
input; this is often referred to as the bottleneck layer. The network
architecture for an AE can be viewed as having two parts, an encoder that maps
the input to the latent space $\bm{h}$ and a decoder that reconstructs the
inputs from the latent space. Mathematically, this is represented as
$\bm{h}=\bm{\omega}\left(\bm{\xi}\right),\;\hat{\bm{\xi}}=\bm{g}\left(\bm{h}\right),$
(5)
where $\bm{\omega}$ represents the encoder network and $\bm{g}$ represents the
decoder network. $\bm{\xi}\in\mathbb{R}^{N}$ in Eq. (5) represents the input
variables with $N$ being the number of variables. A schematic representation
of AE is shown in Fig. 2.
Figure 2: Schematic representation of an auto-encoder (AE). It consists of
encoder ($\omega$) and ($g$). The red nodes represents the bottleneck layer.
AE is integral to the neural network landscape and was initially developed for
model reduction and feature extraction. As far as training an AE is concerned,
an adaptive learning rate optimization algorithm(ADAM) is a popular choice.
The learning in AE is generally expressed as
$\bm{\alpha}=\arg\min_{\bm{\alpha}}\mathcal{L}(\bm{\xi}_{t},\bm{g}\left(\bm{\omega}(\bm{\xi})\right);\bm{\alpha}),$
(6)
where
$\mathcal{L}(\bm{\xi}_{t},\bm{g}\left(\bm{\omega}(\bm{\xi})\right);\bm{\alpha})$
represents the loss-function and
$\bm{\alpha}=\left[\bm{\alpha}_{e},\bm{\alpha}_{d}\right]$ are the
hyperparameters (weights and biases) of the neural network. $\bm{\alpha}_{e}$
corresponds to the hyperparameters of the encoder while $\bm{\alpha}_{d}$
corresponds to the hyperparameters of the decoder. Some important remarks on
AE are furnished below
Remark 1: A situation where
$\bm{g}\left(\bm{\omega}(\bm{\xi})\right)=\bm{\xi}$ everywhere needs to be
avoided [33]. In other words, the training algorithm is designed in such a way
to restrict direct copying of the input.
Remark 2: In AE, the dimensionality of the latent space
$\bm{h}\in\mathbb{R}^{h}$ is generally much smaller than the dimensionality of
the input variable $\bm{\xi}\in\mathbb{R}_{N}$, $h\ll N$. Therefore, $\bm{h}$
can be thought of as a reduced-order representation of $\bm{\xi}$.
Remark 3: When the decoder $\bm{g}$ is linear, and $\mathcal{L}$ is a mean-
squared error, an AE learns to span the same subspace as principal component
analysis.
Remark 4: AE with nonlinear encoder $\bm{\omega}$ and nonlinear decoder
$\bm{g}$ learns a nonlinear reduced-order manifold; however, an AE with too
much expressive capacity learns to copy the inputs (Remark 1).
It is to be noted that researchers are still working on developing AE for
various types of tasks. Some of the popular AE available in the literature
includes variational AE [34], sparse AE [35], stochastic AE [36] and capsule
AE [37] among other. For further details on different types of AE, interested
readers may refer [33].
### 3.2 Recurrent neural networks
Many of the learning tasks involved in artificial intelligence necessitate
handling data that are sequential in nature. Examples of sequential data
include image captioning, time-series forecasting, and speech synthesis, among
others. A recurrent neural network (RNN) is a type of neural network that is
particularly suitable for sequential data. RNN captures the time dynamics by
using cycles in the graph. Consider $\bm{\xi}$ to be inputs and $\bm{y}_{t}$
to be the output at time $t$. The output $\bm{y}_{t}$ in RNN is expressed as a
function of $\bm{\xi}$ and $\bm{h}_{t}$; however, owing to the cyclic graph in
RNN, the hidden state $\bm{h}_{t}$ is continuously updated as the sequence is
processed. A schematic representation of a simple RNN is shown in Fig. 3. Note
that different variants to the classical RNN can be found in the literature.
In this work, we have used a Long Short-Term Memory [38], and hence, the
discussion hereafter is focused on the same. Readers interested in other types
of RNN may refer to [39].
Figure 3: Schematic representation of a typical RNN. $\xi$ (blue node)
represents the input, $h_{i},\forall i$ (red node) represents the hidden layer
and $y_{i}$ (green node) is the output sequence.
LSTM, first proposed by [38], is a type of gated RNN cell that overcomes the
well-know issue of vanishing gradient with the help of the gates that control
the flow of information, i.e., differentiates between the information to be
updated and that to be deleted [40, 41, 42]. LSTM cell comprises a forget
gate, input gate, output gate, and a cell state. Each of these has its
significance. The cell state refers to the information that has to be
transferred in the sequence, and the respective gates determine the
information that has to updated or deleted from the cell state [43]. The
output of the current cell, also referred to as the hidden state, helps retain
the short-term memory, and the cell state, on the other hand, is used to
retain the Long-term Memory. Cell state in LSTM is multiplied with the forget
gate in each cell along with the addition from the input gate, and this
provides the opportunity for forgetting gate to eradicate the unimportant
information and input gate to enhance state with useful information [43]. A
schematic representation of LSTM cell is shown in Fig. 4.
Figure 4: Schematic of the LSTM cell
Mathematical, the operations being carried out inside a LSTM cell is
represented using the following equations.
$C_{t}=\mathbf{F}_{t}\odot\mathbf{C}_{t-1}+\mathbf{I}_{t}\odot\mathbf{\tilde{C}}_{t}$
(7a)
$\mathbf{\tilde{C}}_{t}=\tanh(\mathbf{\xi}_{t}\mathbf{W}_{xc}+\mathbf{h}_{t-1}\mathbf{W}_{hc}+\mathbf{b}_{c}$
(7b)
$\mathbf{I}_{t}=\sigma(\mathbf{\xi}_{t}\mathbf{\xi}_{xi}+\mathbf{h}_{t-1}\mathbf{W}_{hi}+\mathbf{b}_{i}$
(7c)
$\mathbf{F}_{t}=\sigma(\mathbf{\xi}_{t}\mathbf{\xi}_{xf}+\mathbf{h}_{t-1}\mathbf{W}_{hf}+\mathbf{b}),$
(7d)
where $\mathbf{I}_{t}$, $\mathbf{F}_{t}$ and $\tilde{\mathbf{C}}_{t}$ are
respectively the input gate, forget gate, and a candidate cell state. Note
that the update is carried out is additive in nature; this allows long-term
information to pass through and avoids the gradient from vanishing. The short
term state in LSTM is calculated as
$\mathbf{O}_{t}=\sigma\left(\mathbf{\xi}_{t}\mathbf{\xi}_{xo}+\mathbf{h}_{t-1}\mathbf{W}_{ho}+\mathbf{b}_{o}\right)$
(8a) $\mathbf{h}_{t}=\mathbf{O}_{t}\odot\tanh(\mathbf{C}_{t}),$ (8b)
where $\mathbf{O}_{t}$ represents the output gate. It is to be noted that
$\mathbf{h}_{t}$ is used as the output of the cell as well as the hidden state
for the next time-step; this is responsible for the short term memory of LSTM.
$\mathbf{C}_{t}$ on the other hand is responsible for long term memory.
The use of RNN for complex dynamical systems has attracted significant
interest from the research community; this is primarily because of its
capability in capturing temporal dependencies [40]. Multiple architectures
have been proposed for using RNN for accomplishing the task future state
prediction of a dynamical system. Recently [44] and [45] used LSTM and
transformers as time integrator to predict flow evolution state. [46] used
simple RNN for IMU modeling in deep Kalman filter. [47] used autoencoder with
linear recurrence to learns important dynamical features. In this work also,
we use LSTM for extracting useful information from sequential sensor data. The
next section provides more details on the same.
## 4 Proposed approach
In this section, we propose a novel deep learning based framework for state
estimation. The proposed approach integrates the two deep learning approaches
discussed in Section 3, namely AE and RNN. Within the proposed framework, AE
learns the reduced nonlinear subspace. It helps in reducing the information
loss due to the compressed representation. RNN, on the other hand, extends the
capability of the proposed approach and allows it to reuse sensor data
collected at previous time-steps. AE (by reducing the state variable) and RNN
(by incorporating information from the past) also helps address the ill-
posedness associated with solving a state estimation problem.
Consider
$\bm{s}=\left[\bm{s}_{1},\bm{s}_{2},\ldots,\bm{s}_{n}\right]\in\mathbb{R}^{N_{t}\times
N_{s}}$ represents the measurement data obtained from $N_{s}$ sensors over
$N_{t}$ time-steps. Also consider $\bm{w}\in\mathbb{R}_{N_{w}}$ to be state
variables. We can express the state variable $\bm{w}^{t}$ at time $t$ as
$\bm{w}^{t}=g\left(\mathcal{M}\left(\bm{s}^{t-k:t};\bm{\theta}_{M}\right);\bm{\theta}_{g}\right)$
(9)
where $\mathcal{M}\left(\cdot;\bm{\theta}_{M}\right)$ represents a mapping
between the sensor data and the reduced state variable and
$g\left(\cdot;\bm{\theta}_{g}\right)$ projects the reduced state variable back
to the original space. $\bm{\theta}_{M}$ and $\bm{\theta}_{g}$ represent
parameters associated with $\mathcal{M}\left(\cdot;\bm{\theta}_{M}\right)$ and
$g\left(\cdot;\bm{\theta}_{g}\right)$, respectively. Unlike existing methods,
sensor data corresponding to current and previous time-steps have been used
for predicting $\bm{w}^{t}$ in Eq. (9). A schematic representation of the same
is shown in Fig. 5.
Figure 5: Schematic representation of the proposed approach. The difference
with Fig. 1 resides in the fact that information from previous time-steps are
also utilized within the proposed framework. We propose to model
$\mathcal{M}\left(\cdot;\bm{\theta}_{M}\right)$ using RNN and
$g\left(\cdot;\bm{\theta}_{g}\right)$ by using AE.
We also note that the sensor data is sequential and hypothesize that modeling
this sequential nature of the sensor data will improve the predictive
capability of model $\mathcal{M}\left(\cdot;\bm{\theta}_{M}\right)$.
Therefore, we propose to model $\mathcal{M}\left(\cdot;\bm{\theta}_{M}\right)$
by using RNN. As stated in Section 3.2, RNN is suited explicitly for modeling
such sequential data. Another aspect in Eq. (9) is associated with the
projection operator $g\left(\cdot;\bm{\theta}_{g}\right)$. We reiterate that
accuracy and efficiency of Eq. (9) is significantly dependent on
$g\left(\cdot;\bm{\theta}_{g}\right)$. One popular choice among researchers is
to use proper orthogonal decomposition for computing the projection operator
$g$. However, proper orthogonal decomposition being a linear projection scheme
has limited expressive capability. In this work, we propose to use AE as
$g\left(\cdot;\bm{\theta}_{g}\right)$. Owing to the fact that AE is a
nonlinear reduced order model, we expect the accuracy to enhance. However, one
must note that training AE demands more computational effort as compared to
proper orthogonal decomposition. Hereafter, we refer to the proposed framework
as Auto-encoder and Recurrent neural network based state Estimation (ARE)
framework. Next, details on network architecture and training algorithm for
ARE are furnished.
### 4.1 Network architecture and training
The ARE architecture proposed in this paper involves an AE and an RNN. For
state estimation, the trained AE is split into the encoder and the decoder
parts. The encoder part is used during training the RNN within the ARE
framework. The decoder part is used while estimating the state variable. The
AE and RNN networks are trained separately, with the latter following the
former (see Fig. 6). The AE architecture considered in this paper consists of
5 hidden layers, with the 3rd layer being the bottleneck layer. The flow-field
is vectorized before providing it as an input to the AE. Rectified linear unit
(ReLU) activation function has been used for all but the last layer. For the
last layer, a linear activation function is used. For training the network,
ADAM [48] optimizer with a learning rate of $0.0007$ is used. We denote the
trained parameters of the AE as
$\bm{\theta}_{AE}=\left[\bm{\theta}_{w},\bm{\theta}_{g}\right]$, where
$\bm{\theta}_{w}$ and $\bm{\theta}_{g}$ corresponds to the network parameters
for encoder and decoder part respectively. For making the model robust to
noisy data, two batch-norm layers and one dropout layer is added to AE. Batch
normalization [49] is used to normalizes the activation distribution. It
reduces the model’s sensitivity to learning rate [50], reduces training time,
and increases stability. Additionally, this also acts as a regularizer.
Dropout [51] is also an effective regularization technique that works by
dropping connections between neurons during training with a specified
probability $p$. Using the dropout layer just before the bottleneck layer
proved to be most efficient in increasing the robustness of the network to
noise as it simulates the noise in latent vector from the RNN network. A
dropout probability of 0.35 is used in the network. Considering $Li$ to be the
$i$-th layer of AE, BN to be the batch normalization and DR to be the
dropbout, the architecture used in this paper is as follows:
$L1\rightarrow BN\rightarrow L2\rightarrow BN\rightarrow DR\rightarrow
L3\rightarrow L4\rightarrow L5\rightarrow L5\rightarrow Output$
Once the AE trains, we proceed to train the RNN part. The objective here is to
learn a mapping between the sequential sensor data and the reduced state
variable obtained by using the encoder part of the trained AE. First, the
sensor data passes through RNN. It helps in capturing information from the
sequential data. After that, the RNN outputs are mapped to the reduced states
by using a feed-forward neural network. Reduced states of the training outputs
are obtained by using the trained AE (encoder part). The parameters
$\bm{\theta}_{M}$ (see Eq. (9)) corresponds to the RNN and the feedforward
neural network and are obtained by solving the following optimization problem
$\bm{\theta}_{M}^{*}=\arg\min_{\bm{\theta}}\sum_{i=1}^{N_{\text{samp}}}\left\|h^{t}-\mathcal{N}\left(\bm{s}^{t-k:t};\bm{\theta}_{M}\right)\right\|+\lambda\left\|\bm{\theta}_{M}\right|_{2}^{2},$
(10)
Where $\mathcal{N}$ represents the combined RNN and feed-forward neural
network mapping. The second term in Eq. (10) represents $L^{2}$ regularization
and is adopted to avoid overfitting. $\lambda$ is a tuning parameter and needs
to be tuned manually. Similar to AE, the optimization problem is solved by
using ADAM optimizer [48]. We have used weight decay of $10^{-5}$ and a
learning rate of $0.0008$. Early stopping is used, which also acts as a
regularizer [52]. RNN training is schematically shown in Fig. 6(b). The steps
involved in training the proposed (ARE) are shown in Algorithm 1.
Figure 6: Schematic representation of ARE (training phase) - (a) Training AE
for reconstructing the flow-field. (b) Training the RNN for mapping the sensor
data to the reduced state.
1 Library data generation: Perform experiment or run CFD simulations to
generate a library of data, $\mathcal{D}=\left[\bm{s},\bm{w}\right]$ where,
$\bm{s}=[\bm{s}^{1}_{1:N_{s}},\ldots,\bm{s}^{N_{t}}_{1:N_{s}},\;\;\bm{w}=\left[\bm{w}^{1},\ldots,\bm{w}^{N_{t}}\right],$
where $\bm{w}^{i}\in\mathbb{R}^{N_{w}}$.
2 Initialize: Initialize $\bm{\theta}_{g}$ and $\bm{\theta}_{M}$. Provide
sequence length $k+1$ for RNN, learning rate parameters, network architectures
and number of training epochs.
3 Train AE for $\bm{w}=\left[\bm{w}^{1},\ldots,\bm{w}^{N_{t}}\right]$.
$\bm{a}^{t}\leftarrow\omega(\bm{w}^{t};\bm{\theta}_{w})$;
$\triangleright$ Reducing dimensionality using AE
Train RNN for sequence length $k+1$;
$\triangleright$ Eq. (10)
Algorithm 1 Training ARE
### 4.2 State estimation using the proposed approach
Once the proposed ARE is trained by following the procedure detailed in
Algorithm 1, one can use it for estimating the state. A trained ARE performs
state estimation in two simple steps. In the first step, the trained RNN is
used for estimating the reduced state based on the sensor measurement. Once
the reduced state has been estimated, the decoder part of the AE is used to
project the reduced state onto the original state. For clarity of readers, the
steps for predicting state using ARE are shown in Algorithm 2. A schematic
representation of the same is also shown in Fig. 7.
1 Pre-requisite: Trained ARE model,
$\bm{\theta}=\left[\bm{\theta}_{M},\bm{\theta}_{g}\right]$ and sensor data
$\bm{s}_{*}^{t-k:t}$.
$\bm{a}^{t}_{*}\leftarrow\mathcal{N}(\bm{s}_{*}^{t-k:t};\bm{\theta}_{M})$;
$\triangleright$ Reduced state prediction using RNN
$\bm{w}^{t}_{*}\leftarrow=g(\bm{a}^{t}_{*};\bm{\theta}_{g})$;
$\triangleright$ Full-state estimation using decoder part of AE
Algorithm 2 State estimation using ARE Figure 7: Schematic representation of
the proposed ARE for state estimation
## 5 Numerical experiments
In this section, two examples are presented to illustrate the performance of
the proposed approach. The examples selected are well-known benchmark problems
in the fluid mechanic’s community. For both examples, we have considered that
the sensor measures vorticity, and the objective is to reconstruct the
vorticity field. We present case studies by varying the number of sensors and
the number of sequences available. To illustrate the excellent performance of
our approach, a comparison with another state-of-the-art method has been
provided. Comparison among results is carried out based on a qualitative and
quantitative metric. To be specific, visual inspection is used as a
qualitative metric, and the relative error is used as a quantitative metric,
$\epsilon=\frac{\left\|\bm{\omega}^{n}(\bm{\theta}^{*})-\bm{\omega}^{n}_{r}(\bm{\theta}^{*})\right\|_{2}}{\left\|\bm{\omega}^{n}(\bm{\theta}^{*})\right\|_{2}}\times
100.$ (11)
where $\epsilon$ represents the error, $\bm{\omega}^{n}(\bm{\theta}^{*})$ is
the true state and $\bm{\omega}^{n}_{r}(\bm{\theta}^{*})$ is the state vector
predicted using the proposed approach. $\left\|\cdot\right\|_{2}$ represents
the $l_{2}$ norm. The dataset for solving the state estimation problems is
generated using OpenFoam [53]. The proposed approach has been implemented
using PyTorch [54]. The software associated with the proposed approach, along
with the implementation of both the examples, will be made available on
acceptance of the paper.
### 5.1 Periodic Vortex shedding past a cylinder
As the first example, we consider two-dimensional flow past a circular
cylinder at Reynolds’ number $Re=190$. It is a well known canonical problem
and is characterized by periodic laminar flow vortex shedding. A schematic
representation of the computational domain is shown in Fig. 8. The circular
cylinder is considered to have a diameter of $1$ unit. The center of the
cylinder is located at a distance of $8$ units from the inlet. The outlet is
located at a distance of $25$ units from the center of the cylinder. The
sidewalls are at $4$ units distance from the center of the cylinder. At the
inlet boundary, a uniform velocity of $1$ unit along the $X$-direction is
applied. Pressure boundary condition with $P=0$ is considered at the outlet. A
no-slip boundary at the cylinder surface is considered.
Figure 8: (a) Schematic representation of the computational domain with
boundary conditions at the inlet and the outlet. The cylinder is having a
diameter of 1 unit. A no-slip boundary is considered at the cylinder wall.
Zero pressure gradient at the inlet and zero velocity gradient at the outlet
are considered. (b) Schematic of the problem domain with snapshot cutout of
$a\times b$. For periodic vorticity problem, $a=6$ units and $b=4$ units. For
transient flow problem, $a=12$ units and $b=6$ units. The schematics are not
to scale.
The dataset necessary for training the proposed model is generated by using
Unsteady Reynolds-averaged Navier Stokes (URANS) simulation in OpenFoam [53].
The selection of URANS for data generation is based on the fact that results
obtained using URANS are highly accurate in $Re\in[30,300]$ [55]. Nonetheless,
the method proposed is not dependent on the fluid simulator used and can be
seamlessly used in conjunction with more accurate simulator like DNS. The
overall problem domain is discretized into 63420 elements with finer mesh near
the cylinder. Time step $\delta t=0.02$ units is considered.
For training the model, a library of 180 snapshots is generated by running
OpenFoam. Additional 120 snapshots, 60 for validation and 60 for testing, have
also been generated. Two consecutive snapshots are separated by 10$\delta t$.
Coordinate of the snapshot cutout stretches from $[0,-2]\times[6,2]$ which is
discretized into $168\times 251$ points in $x$ and $y$ directions (see Fig.
8(b)). The objective here is to recover the complete vorticity-field in the
cutout by using the sensor measurements. Details on the network architecture
are provided in Table 1.
Table 1: Network architecture of proposed ARE and SD for periodic vortex shedding problem. BN indicates batch normalization and DR represents dropout. HS is the number of features in the hidden state of lstm. Networks | Architecture
---|---
ARE(Auto-Encoder) | $251\times 168\rightarrow BN^{*}\rightarrow 1024\rightarrow BN^{*}\rightarrow DR(0.35)^{*}\rightarrow 256\rightarrow 25\rightarrow 256\rightarrow 1024\rightarrow 168\times 251$
ARE(RNN) | LSTM(HS=50) $\rightarrow 50\rightarrow 50\rightarrow 25$
SD | $N_{s}\rightarrow 35\rightarrow BN\rightarrow DR(0.1)^{*}\rightarrow 40\rightarrow BN\rightarrow 168\times 251$
Components with ∗ are not used in case of noise-free data.
To illustrate the superiority of the proposed approach, the results obtained
using the proposed approach are compared with those obtained using proper
orthogonal decomposition-based deep state estimation (PDS) proposed by [29]
and shallow decoder (SD) proposed by [5]. In PDS, the first 25 modes are used,
which is the same as the number of neurons in the bottleneck layer of the
proposed approach. The feed-forward neural network used in PDS for mapping the
sensor measurements to the latent state is the same as the feed-forward
network used in ARE. Note that comparison with other popular approaches such
as gappy-POD and linear stochastic estimation is not shown as it is already
established in [29] that PDS outperforms both the approaches. Brief details on
PDS, gappy-POD, and linear stochastic estimation are provided in A.
Fig. 9 shows the results obtained using different approaches. It is an
idealized case where we have considered the sensor data to be noise free. The
sequence length of four is used for training ARE. We have considered the
extreme case where data from only one sensor is available. We observe that the
proposed ARE, with only one senor, is able to recover the full state
accurately. PDS and SD, on the other hand, yields less accurate results.
(a) PDS (b) ARE (c) SD (d) Ground Truth
Figure 9: Figure depicts results of periodic vortex shedding for one sensor.
Orange dots in the images represents the location of sensor. SEQ-LEN used for
ARE model is 4.
(a) PDS (b) ARE (c) SD (d) Ground Truth
Figure 10: Figure depicts results of periodic vortex shedding for one sensor.
The data is corrupted with white Gaussian noise having SNR = 20. Orange dots
in the images represents the location of sensor. SEQ-LEN used for ARE model is
4.
Fig. 10 shows results corresponding to the case where the sensor is corrupted
by white Gaussian noise. This is a more realistic case. Again, a sequence
length of 4 is considered and it is assumed that data from only one sensor is
available. In this case also, we observe that ARE yields highly accurate
results and outperforms PDS and SD. The effect of noise on the proposed ARE is
shown in Fig. 11. We observe that both ARE (with one and two sensors) and SD
(with two sensors) yield the best results. The fact that the proposed approach
is able to correctly predict the state from only one sensor data (noisy) is
really impressive. SD with one sensor and PDS predicted results are
significantly less accurate as compared to the proposed approach.
Figure 11: (a) Performance of the proposed approach ARE, SD and PDS for
periodic vortex shedding problem at different noise level. For both ARE and
PDS, results corresponding to one and two sensors are presented. (b)
Performance of proposed ARE, SD and PDS with increase in number of sensors.
Data used is noise free.
Next, we investigate the effect of varying the number of sensors. Fig. 11
shows the performance of different methods with the increase in the number of
sensors. The data considered is idealistic with no noise. Cases corresponding
to one, two, five, and ten sensors are presented. For all the cases, the
proposed ARE is found to yield the best result. Results obtained using SD and
PDS follows a similar trend. We also carried out an additional case study
where we considered that sensor data from both past and future is available. A
bi-directional RNN (B-RNN) based ARE developed for the same. However, due to
the paucity of space, the same is not presented here. Those interested can
refer to B for details on the same.
### 5.2 Transient Flow past a cylinder
As the second example, we consider the problem involving transient flow past a
cylinder. Because of the transient nature of the flow, this is much more
challenging than the periodic vortex shedding problem in Section 5.1. The
problem domain, meshing, and solution strategy for this problem are considered
the same as the periodic vortex shedding problem. However, unlike the previous
problem, we have considered variation in the Reynolds’ number. This
exponentially increases the complexity of the problem.
The training library for this problem was created by running OpenFoam. Total
1200 snapshots consisting of $400$ snapshots at $Re=[180,190,200]$ were
generated. Given the fact that $Re$ considered for this problem, resides in
$[30,300]$, we have used URANS for generating the data [55]. Validation and
test set consists of $400$ sequential snapshots at $Re=[185,195]$. Similar to
the previous problem, the snapshots were separated by a time interval of
$10\delta t$. Co-ordinate of snapshot cutout stretches from
$\left[0,-3\right]\times\left[12.0,3.0\right]$ (see Fig. 8(b)). The cutout is
discretized into $252$ and $502$ points in the $x$ and $y$ direction,
respectively. Similar to the previous example, the objective here is to
recover the vorticity field based on sensor measurements. Similar to the
previous example, results obtained have been compared with PDS and SD. Details
on the network architecture used for this problem are provided in Table 2.
Table 2: Network architecture of proposed ARE for transient flow problem. HS is the number of features in the hidden state of lstm. Networks | Architecture
---|---
ARE(Auto-Encoder) | $252\times 502\rightarrow BN^{*}\rightarrow 2000\rightarrow BN^{*}\rightarrow DR(0.35)^{*}\rightarrow 300\rightarrow 25\rightarrow 300\rightarrow 2000\rightarrow 252\times 502$
ARE(RNN) | LSTM(HS=50) $\rightarrow 200\rightarrow 200\rightarrow 25$
SD | $N_{s}\rightarrow 350\rightarrow BN\rightarrow DR(0.1)^{*}\rightarrow 400\rightarrow BN\rightarrow 252\times 502$
Components with ∗ are not used in case of noise-free data.
(a) PDS (b) ARE (c) SD (d) Ground truth
Figure 12: Figure depicts results of Transient Flow for one sensor. Orange
dots in the images represents the location of sensor. SEQ-LEN used for ARE
model is 4
Fig. 12 shows the reconstructed vorticity field using ARE, SD and PDS. Ground
truth has also been reported. It is an idealized case where we have considered
the sensor data to be noise-free. A sequence length of four is used for ARE.
We have considered the extreme case where data from only one sensor is
available. For this problem also, the proposed is able to recover the full
vorticity field accurately. PDS and SD, on the other hand, fails to accurately
recover the vorticity field. Next, we consider a more realistic scenario where
the sensor data is corrupted by noise. Fig. 13 shows the results corresponding
to noisy sensor measurement. In this case, we have considered that
measurements from two sensors are available. For this case also, the results
obtained using the proposed ARE is found to be superior as compared to those
in the literature.
(a) PDS (b) ARE (c) SD (d) Ground truth
Figure 13: Figure depicts results of Transient Flow for two sensors. The data
is corrupted with white Gaussian noise with SNR = 28. Orange dots in the
images represents the location of sensor. SEQ-LEN used for ARE model is 4
The effect of noise on the performance is shown Fig. 14. Overall, results
obtained using the proposed ARE are found to be more accurate as compared to
those obtained using PDS and SD. The improved accuracy of ARE can be attribute
to the fact that it learns from a sequence of data. Surprisingly, we observe
that PDS with one sensor yields superior result as compared to PDS with two
sensors. This is because the POD modes for two sensor cases is susceptible to
the noise in the data. As noise in the data reduces, the POD modes are
identified correctly and performance of PDS with two sensors is found to be
superior as expected. The proposed ARE and SD are immune to such overfitting
phenomenon.
We also investigate the effect of a number of sensors and sequence length
considered in the proposed approach. Fig. 15 shows the performance of
different methods with an increase in the number of sensors. Similar to
previous example, cases corresponding to one, two, five, and ten sensors are
presented. As expected, increase in number of sensors results in a improved
accuracy. As for accuracy, for all the four cases, the proposed approach
yields the best result outperforming both PDS and SD. Fig. 15 illustrates the
performance of the proposed approach corresponding to the different sequence
length. As expected, initially, the results are found to improve with an
increase in the sequence length. ARE reaches a saturation point at sequence
length three, and no significant improvement in results is observed on further
increasing the sequence length.
Figure 14: Performance of the proposed approach ARE, SD and PDS for the
transient flow problem at different noise level. ARE1 corresponds to results
with one sensor and ARE2 corresponds to results with 2 sensors.
Figure 15: (a) Performance of PDS, SD, and proposed ARE with increase in
number of sensors. (b) Performance of proposed ARE with increase in sequence
length. Data used in both is without noise.
Lastly, similar to previous example, we have also considered the case where
both past and future data are available. B-RNN based ARE is used for solving
the same. For details on the same, interested readers may refer B.
### 5.3 NOAA Optimum Interpolation (OI) SST V2 Dataset
As the last example, we test our approach on the well-known Sear Surface
Temperature (SST) dataset. SST Data is publicly available and uploaded by the
Physical Sciences Division at National Oceanic and Atmospheric Administration
(NOAA). The data is made available by the respective organization after the
analysis of in-situ and satellite observation of the SST. The resolution of
the data available corresponding to the variable “time” is Weekly, Monthly and
monthly long-term mean. The resolution corresponding to the spatial domain is
of 1 deg latitude and 1 deg longitude, which translates to 180 x 360 grid
points. In this study we make use of weekly mean data from the year 1990 on-
wards. The data is said to be centred around Wednesday for this period. Data
for initial 400 weeks was chosen as the training data-set. Following 100 weeks
data was chosen as validation data-set and following 200 weeks data for the
testing. The problem description here is to reconstruct the SST data on the
spatial domain with the help of limited sensor data measurements. Details on
the network architecture used for this problem are provided in Table 3.
Table 3: Network architecture of proposed ARE and SD for SST dataset. BN indicates batch normalization and DR represents dropout. HS is the number of features in the hidden state of lstm. Networks | Architecture
---|---
ARE(Auto-Encoder) | $44219\rightarrow BN^{*}\rightarrow 512\rightarrow BN^{*}\rightarrow DR(0.35)^{*}\rightarrow 256\rightarrow 25\rightarrow 256\rightarrow 512\rightarrow 44219$
ARE(RNN) | LSTM(HS=50) $\rightarrow 100\rightarrow 100\rightarrow 25$
SD | $N_{s}\rightarrow 350\rightarrow BN\rightarrow DR(0.1)^{*}\rightarrow 400\rightarrow BN\rightarrow 44219$
Components with ∗ are not used in case of noise-free data.
Fig. 16 shows the comparison of results between three different architectures
we chose for our study i.e., PDS, SD, and ARE. To show the applicability of
the techniques in real world scenarios, we make use of corrupted or noisy data
at different signal to noise ratios and at three different sensor measurements
i.e. 2, 4 and 8. The proposed problem reconstructs the SST data over the globe
accurately for different sensor data input. We observe, with decreasing noise
in the data, either of the approaches perform well; however, when noise in the
data is significant, the proposed ARE is found to be significantly more
accurate as compared to PDS and SD. Even with higher number of sensors, the
proposed approach is found to be superior (at higher noise level) as compared
to PDS and SD. For better visualization, the contour plot of reconstructed sea
surface temperature is shown in Fig. 17. We have considered SNR to be 20 and
assumed data to be available from four sensors. Sequence length of four is
used for the proposed ARE. It is clearly visible that the reconstructed field
using the proposed ARE is more accurate compared to the PDS and SD
reconstructed fields.
(a) sensors 2 (b) sensors 4 (c) sensors 8
Figure 16: Change in performance of different approaches is presented with
increasing noise using three cases
(a) PDS (b) ARE (c) SD (d) Ground truth
Figure 17: Figure depicts results of SST with 4 sensors and at noise of SNR
20. Blue dots in the images represents the location of sensor. SEQ-LEN used
for ARE model is 4
(a) SNR 10 (b) SNR 20 (c) SNR 30
Figure 18: Performance of the three model is compared with increasing no. of
sensors at SNR 10, 20, 30
(a) sensor 2 (b) sensor 4 (c) sensor 8
Figure 19: Performance of ARE and PDS is compared at different number of
neurons in bottleneck layer with noise of SNR 20
In Fig. 18, we compare the results of the approaches corresponding to the
increasing number of sensors at three different SNR ratios. The investigation
further strengthens our claim for the robustness of our proposed approach,
ARE. As the number of sensors increases, it is evident that all the approaches
tend to present better results. For the sake of comparison, we considered
corrupted data at three different SNR, 10, 20 and 30, since the approaches
mentioned perform more or less the same without noise. At SNR 10, the proposed
approach tends to outperform its counterparts by a larger margin compared to
SNR 30. As the corruption in data increases, ARE tends to perform much better
than its counterparts. SD performs almost similar to ARE at lower noise level
and better than PDS in all cases.
Lastly, we investigate the sensitivity of the proposed approach to the number
of neurons in the bottleneck layer. For PDS, this corresponds to the number of
POD modes. The concept of bottleneck layer doesn’t exist for SD and hence, the
same is not reported. Results obtained are shown in Fig. 19. The results are
presented corresponding to cases with 2, 4 and 8 sensors. The SNR is kept
fixed at 20. The number of neurons in the bottleneck layer is varied in
$[5,50]$. We observe that the proposed approach is less sensitive to the
number of neurons in the bottleneck layer, although slight improvement can be
observed as the layer size is increased. This observation presents an
insightful understanding of our approach which could be translated to a
computationally less expensive approach, with decreased layer size of our
bottleneck layer. As for PDS, we observe that the accuracy reduces with
increase in the number of POD modes. This seems counter-intuitive as increase
in POD modes is expected to reduce the error. However, on a closer
examination, we observed that with more number of POD modes, the number of
parameters in neural network increases, resulting in overfitting. This can be
addressed by increasing the number of sensors. Overall, the results presented
in this section illustrate the accuracy and robustness of the proposed
approach in state estimation from limited sensor.
## 6 Conclusions
In this work, we introduced a novel deep learning based approach for state
estimation. The proposed approach uses an auto-encoder as a reduced-order
model and recurrent neural network to map the sensor measurements to the
reduced state. The proposed framework is superior to existing state estimation
frameworks in two aspects. First, auto-encoder, being a nonlinear manifold
learning framework, is superior to the usually used proper orthogonal
decomposition. Secondly, unlike existing state estimation frameworks, the
proposed approach utilizes present and past sensor measurements using the
recurrent neural network. It results in improved accuracy, specifically for
cases where limited sensors are deployed. Experiments performed on simulation
of flow past a cylinder and sea-surface data illustrated the capability of the
proposed approach in learning from sequential data. Comparison carried out
with respect to proper orthogonal decomposition-based deep state estimation
and shallow-decoder showed the superior accuracy of the proposed approach.
Moreover, the proposed approach was also found to be robust to the noise in
the sensor measurements.
Utilizing sequential information can prove beneficial in many other state
estimation tasks. Future work can be aimed at exploiting effects of other
models used on sequential information such as transformers, comparing between
other auto-encoder network variations, the effect of varying time step between
measurements, developing models capable of transfer learning, i.e., trained on
smaller time step but can use larger time step. Work on utilizing, recently
developed a physics-informed neural network for solving such problems can also
be pursued in the future.
Acknowledgements: The authors would like to thank Dr. Arghya Samanta and
Nirmal J Nair for the useful discussions during this paper’s preparation. SC
acknowledges the financial support of the I-Hub Foundation for Cobotics (IHFC)
and seed grant provided through Faculty initiation grant, IIT Delhi.
## References
* [1] Michele Milano and Petros Koumoutsakos. Neural network modeling for near wall turbulent flow. Journal of Computational Physics, 182(1):1 – 26, 2002.
* [2] Julia Ling, Andrew Kurzawski, and Jeremy Templeton. Reynolds averaged turbulence modelling using deep neural networks with embedded invariance. Journal of Fluid Mechanics, 807:155–166, 2016.
* [3] Nicholas Geneva and Nicholas Zabaras. Quantifying model form uncertainty in reynolds-averaged turbulence models with bayesian deep neural networks. Journal of Computational Physics, 383:125–147, 2019.
* [4] Nicholas Geneva and Nicholas Zabaras. Multi-fidelity generative deep learning turbulent flows. arXiv preprint arXiv:2006.04731, 2020.
* [5] N. Benjamin Erichson, Lionel Mathelin, Zhewei Yao, Steven L. Brunton, Michael W. Mahoney, and J. Nathan Kutz. Shallow learning for fluid flow reconstruction with limited sensors and limited data, 2019.
* [6] Lorenzo Rosasco, Andrea Caponnetto, Ernesto Vito, Francesca Odone, and Umberto Giovannini. Learning, regularization and ill-posed inverse problems. Advances in Neural Information Processing Systems, 17:1145–1152, 2004.
* [7] Rajdip Nayek, Suparno Mukhopadhyay, and Sriram Narasimhan. Mass normalized mode shape identification of bridge structures using a single actuator-sensor pair. Structural Control and Health Monitoring, 25(11):e2244, 2018.
* [8] Simo Särkkä. Bayesian filtering and smoothing. Number 3. Cambridge University Press, 2013.
* [9] Christopher M Bishop. Pattern recognition and machine learning. springer, 2006.
* [10] Konrad Reif, Stefan Gunther, Engin Yaz, and Rolf Unbehauen. Stochastic stability of the discrete-time extended kalman filter. IEEE Transactions on Automatic control, 44(4):714–728, 1999.
* [11] Eric A Wan, Rudolph Van Der Merwe, and Simon Haykin. The unscented kalman filter. Kalman filtering and neural networks, 5(2007):221–280, 2001.
* [12] Rajdip Nayek, Souvik Chakraborty, and Sriram Narasimhan. A gaussian process latent force model for joint input-state estimation in linear structural systems. Mechanical Systems and Signal Processing, 128:497–530, 2019.
* [13] Jonathan Tu, John Griffin, Adam Hart, Clarence Rowley, Louis Cattafesta, and Lawrence Ukeiley. Integration of non-time-resolved piv and time-resolved velocity point sensors for dynamic estimation of velocity fields. Experiments in Fluids, 54, 01 2012.
* [14] M. Buffoni, S. Camarri, A. Iollo, E. Lombardi, and M.V. Salvetti. A non-linear observer for unsteady three-dimensional flows. Journal of Computational Physics, 227(4):2626 – 2643, 2008.
* [15] Ryota Kikuchi, Takashi Misaka, and Shigeru Obayashi. Assessment of probability density function based on POD reduced-order model for ensemble-based data assimilation. Fluid Dynamics Research, 47(5):051403, sep 2015.
* [16] V. Mons, J.-C. Chassaing, T. Gomez, and P. Sagaut. Reconstruction of unsteady viscous flows using data assimilation schemes. Journal of Computational Physics, 316:255 – 280, 2016.
* [17] Andre Fernando De Castro da Silva and Tim Colonius. Ensemble-based state estimator for aerodynamic flows. AIAA Journal, 56:1–11, 06 2018.
* [18] J. A. Tropp and A. C. Gilbert. Signal recovery from random measurements via orthogonal matching pursuit. IEEE Transactions on Information Theory, 53(12):4655–4666, 2007\.
* [19] S. G. Mallat and Zhifeng Zhang. Matching pursuits with time-frequency dictionaries. IEEE Transactions on Signal Processing, 41(12):3397–3415, 1993\.
* [20] Jared L. Callaham, Kazuki Maeda, and Steven L. Brunton. Robust flow reconstruction from limited measurements via sparse representation. Physical Review Fluids, 4(10), Oct 2019.
* [21] Richard Everson and Lawrence Sirovich. Karhunen–loève procedure for gappy data. JOSA A, 12, 08 1995.
* [22] Kelly Cohen, Stefan Siegel, and Thomas McLaughlin. A heuristic approach to effective sensor placement for modeling of a cylinder wake. Computers & Fluids, 35:103–120, 01 2006.
* [23] K. Willcox. Unsteady flow sensing and estimation via the gappy proper orthogonal decomposition. Computers & Fluids, 35(2):208 – 226, 2006.
* [24] Ronald Adrian. On the role of conditional averages in turbulence theory. -1:323–332, 01 1977.
* [25] A. M. Naguib, C. E. Wark, and O. Juckenhöfel. Stochastic estimation and flow sources associated with surface pressure events in a turbulent boundary layer. Physics of Fluids, 13(9):2611–2626, 2001.
* [26] Dan Ewing and Joseph H. Citriniti. Examination of a lse/pod complementary technique using single and multi-time information in the axisymmetric shear layer. In J. N. Sørensen, E. J. Hopfinger, and N. Aubry, editors, IUTAM Symposium on Simulation and Identification of Organized Structures in Flows, pages 375–384, Dordrecht, 1999. Springer Netherlands.
* [27] Y. G. Guezennec. Stochastic estimation of coherent structures in turbulent boundary layers. Physics of Fluids A: Fluid Dynamics, 1(6):1054–1060, 1989.
* [28] J. P. Bonnet, D. R. Cole, J. Delville, M. N. Glauser, and L. S. Ukeiley. Stochastic estimation and proper orthogonal decomposition: Complementary techniques for identifying structure. Experiments in Fluids, 17(5):307–314, Sep 1994.
* [29] Nirmal J. Nair and Andres Goza. Leveraging reduced-order models for state estimation using deep learning. Journal of Fluid Mechanics, 897:R1, 2020.
* [30] Jeremy Pinier, Julie Ausseur, Mark Glauser, and Hiroshi Higuchi. Proportional closed-loop feedback control of flow separation. Aiaa Journal - AIAA J, 45:181–190, 01 2007.
* [31] Ronald Adrian. Conditional eddies in isotropic turbulence. Physics of Fluids, 22, 11 1979.
* [32] T. C. Tung and R. J. Adrian. Higher-order estimates of conditional eddies in isotropic turbulence. The Physics of Fluids, 23(7):1469–1470, 1980.
* [33] Kevin P Murphy. Machine learning: a probabilistic perspective. MIT press, 2012.
* [34] Ilyes Khemakhem, Diederik Kingma, Ricardo Monti, and Aapo Hyvarinen. Variational autoencoders and nonlinear ica: A unifying framework. In International Conference on Artificial Intelligence and Statistics, pages 2207–2217. PMLR, 2020.
* [35] Xavier Glorot, Antoine Bordes, and Yoshua Bengio. Deep sparse rectifier neural networks. In Proceedings of the fourteenth international conference on artificial intelligence and statistics, pages 315–323. JMLR Workshop and Conference Proceedings, 2011.
* [36] Hareesh Bahuleyan, Lili Mou, Hao Zhou, and Olga Vechtomova. Stochastic wasserstein autoencoder for probabilistic sentence generation. arXiv preprint arXiv:1806.08462, 2018.
* [37] Adam R Kosiorek, Sara Sabour, Yee Whye Teh, and Geoffrey E Hinton. Stacked capsule autoencoders. arXiv preprint arXiv:1906.06818, 2019.
* [38] Sepp Hochreiter and Jürgen Schmidhuber. Long short-term memory. Neural computation, 9:1735–80, 12 1997.
* [39] Zachary C Lipton, John Berkowitz, and Charles Elkan. A critical review of recurrent neural networks for sequence learning. arXiv preprint arXiv:1506.00019, 2015.
* [40] Suraj Pawar, Shady E. Ahmed, Omer San, Adil Rasheed, and Ionel M. Navon. Long short-term memory embedded nudging schemes for nonlinear data assimilation of geophysical flows. Physics of Fluids, 32(7):076606, 2020.
* [41] José del Águila Ferrandis, Michael S. Triantafyllou, Chryssostomos Chryssostomidis, and George Em Karniadakis. Learning functionals via LSTM neural networks for predicting vessel dynamics in extreme sea states. CoRR, abs/1912.13382, 2019.
* [42] Sepp Hochreiter. The vanishing gradient problem during learning recurrent neural nets and problem solutions. International Journal of Uncertainty, Fuzziness and Knowledge-Based Systems, 06(02):107–116, 1998.
* [43] Hamidreza Eivazi, Hadi Veisi, Mohammad Hossein Naderi, and Vahid Esfahanian. Deep neural networks for nonlinear model order reduction of unsteady flows. Physics of Fluids, 32(10):105104, 2020.
* [44] Nicholas Geneva and Nicholas Zabaras. Transformers for modeling physical systems, 2020.
* [45] Hamidreza Eivazi, Hadi Veisi, Mohammad Hossein Naderi, and Vahid Esfahanian. Deep neural networks for nonlinear model order reduction of unsteady flows. Physics of Fluids, 32(10):105104, Oct 2020.
* [46] Siavash Hosseinyalamdary. Deep kalman filter: Simultaneous multi-sensor integration and modelling; a gnss/imu case study. Sensors, 18(5), 2018.
* [47] Samuel E. Otto and Clarence W. Rowley. Linearly-recurrent autoencoder networks for learning dynamics, 2019.
* [48] Diederik Kingma and Jimmy Ba. Adam: A method for stochastic optimization. International Conference on Learning Representations, 12 2014.
* [49] Sergey Ioffe and Christian Szegedy. Batch normalization: Accelerating deep network training by reducing internal covariate shift, 2015.
* [50] Sanjeev Arora, Zhiyuan Li, and Kaifeng Lyu. Theoretical analysis of auto rate-tuning by batch normalization, 2018\.
* [51] Nitish Srivastava, Geoffrey Hinton, Alex Krizhevsky, Ilya Sutskever, and Ruslan Salakhutdinov. Dropout: A simple way to prevent neural networks from overfitting. Journal of Machine Learning Research, 15(56):1929–1958, 2014.
* [52] Jeffrey Heaton. Ian goodfellow, yoshua bengio, and aaron courville: Deep learning: The mit press, 2016, 800 pp, isbn: 0262035618. Genetic Programming and Evolvable Machines, 19, 10 2017.
* [53] Hrvoje Jasak. Openfoam: open source cfd in research and industry. International Journal of Naval Architecture and Ocean Engineering, 1(2):89–94, 2009.
* [54] Adam Paszke, Sam Gross, Soumith Chintala, Gregory Chanan, Edward Yang, Zachary DeVito, Zeming Lin, Alban Desmaison, Luca Antiga, and Adam Lerer. Automatic differentiation in pytorch. 2017\.
* [55] R.M. Stringer, J. Zang, and A.J. Hillis. Unsteady rans computations of flow around a circular cylinder for a wide range of reynolds numbers. Ocean Engineering, 87:1–9, 2014.
* [56] M. Schuster and K. K. Paliwal. Bidirectional recurrent neural networks. IEEE Transactions on Signal Processing, 45(11):2673–2681, 1997\.
* [57] Pierre Baldi, Søren Brunak, Paolo Frasconi, Giovanni Soda, and Gianluca Pollastri. Exploiting the past and the future in protein secondary structure prediction. Bioinformatics (Oxford, England), 15:937–46, 12 1999.
## Appendix A Proper orthogonal decomposition based deep state estimation
(PDS)
In this section, we briefly provide the theoretical background of PDS. We note
that development of PDS is motivated from gappy-POD and linear stochastic
estimation and hence, the discussion in this section also starts from gappy-
POD and then proceeds to PDS via linear stochastic estimation.
Let training set consist of $N_{t}$ flattened data images,
$\mathbf{W}=(\bm{w}_{1},\bm{w}_{2},\ldots,\bm{w}_{N_{t}})$, where
$\bm{w}_{t}\in\mathbb{R}^{N_{w}}$ and corresponding sensor measurements
$\bm{s}=(\bm{s}_{1},\bm{s}_{2},\dots,\bm{s}_{N_{t}})$. As already state in
Section 2, the aim in state estimation is to recover full state $\bm{w}$ via
sensor measurements $\bm{s}$.
### A.1 Gappy-POD
Consider, $\mathbf{H}\in\mathbb{R}^{N_{s}\times N_{w}}$ is measurement
operator that maps full state to measurements and have ones as sensor location
and zeros otherwise. Mathematically,
$\bm{s}=\mathbf{H}\bm{w},$ (12)
where $\bm{w}$ can be approximated by linear combination of $k$ modes. Modes
used here are $k$ most dominant modes, $\bm{\Phi}$, and are obtained by proper
orthogonal decomposition of training data $\mathbf{W}$ and selecting first $k$
singular vectors,
$\bm{w}\approx\sum_{i=1}^{k}\phi_{i}a_{i}=\bm{\Phi}\bm{A},$ (13)
where
$\mathbf{W}\approx\bm{\Phi}DB^{*},truncated\hskip 5.69054ptk-rank,$ (14)
and
$\bm{s}\approx\mathbf{H}\bm{\Phi}a.$ (15)
In Eq. (14), $\bm{\Phi}\in\mathbb{R}^{N_{w}\times k}$. During testing $\bm{a}$
is obtained by solving the following minimization problem.
$\bm{a}\in\underset{\tilde{\bm{a}}}{\arg\min}\left\|\bm{s}-\mathbf{H}\bm{\Phi}\tilde{\bm{a}}\right\||_{2}^{2}.$
(16)
Solution to this problem is obtained by taking Moore-Penrose pseudo inverse
$\bm{a}=(\mathbf{H}\bm{\Phi})^{+}\bm{s}$ (17)
This approach requires previous knowledge of operator $\mathbf{H}$. This
operator is only available for simple systems but is often unknown for systems
of practical interest.
### A.2 Linear stochastic estimation
Linear Stochastic Estimation overcomes this issue by defining another operator
which will map latent state to sensor. Operator
$P:\mathbb{R}^{k}\mapsto\mathbb{R}^{N_{s}}$ is learned from training data via
the following minimization problem.
$P\in\underset{\tilde{P}}{\arg\min}||\bm{s}-\tilde{P}\bm{A}||^{2}_{2},$ (18)
where $\mathbf{S}\in\mathbb{R}^{N_{s}\times N_{t}}$ and
$A\in\mathbb{R}^{k\times N_{t}}$. This approach completely skips over the
operator $\mathbf{H}$; instead, $P$ is considered to be the empirical estimate
of the linear operator $\mathbf{H}\Phi$. Subsequently, $\bm{a}$ is obtained as
$\bm{a}\in\underset{\tilde{\bm{a}}}{\arg\min}||\bm{s}-P\tilde{\bm{a}}||^{2}_{2}$
(19)
### A.3 Deep state estimation
Deep state estimation [29] approache replaces the linear mapping between
latent state and sensors with nonlinear mapping to further generalize the
approach. It uses a neural network
$G:\mathbb{R}^{N_{s}}\mapsto\mathbb{R}^{bn}$ parametrized by $\theta$ for
nonlinear mapping of sensor measurements to approximate embeddings.
$\bm{a}=G(\bm{s},\bm{\theta})$ (20)
The neural network is trained as
$\bm{\theta}\in\underset{\tilde{\bm{\theta}}}{\arg\min}\sum_{i=1}^{N_{t}}||\bm{a}^{(i)}-G(\bm{s}^{(i)},\tilde{\bm{\theta}})||_{2}^{2},$
(21)
where
$\bm{a}\in\underset{\tilde{\bm{a}}}{\arg\min}||\bm{w}-\Phi\tilde{\bm{a}}||^{2}_{2}.$
(22)
$\bm{\Phi}$ as before are the POD-modes. During testing trained neural network
$G$ approximate embeddings for sensor measurements which is used to recover
full state. We refer to this method as POD based deep state estimation (PDS).
## Appendix B Bi-directional RNN
The basic idea of bidirectional recurrent neural nets (B-RNN) [56, 57] is to
present each training sequence to two separate recurrent nets namely forward
and backward, both of which are connected to the same output layer. (In some
cases a third network is used in place of the output layer, but here we have
used the simpler model). This means that for every point in a given sequence,
the BRNN has complete, sequential information about all points before and
after it. In this work, we have used B-RNN for solving a special state
estimation problem where sensor measurements from both past and future states
are available.
A Schematic representation of B-RNN used is shown in the Fig. 20. It is
modified to stop propagating information after the middle time step in forward
and backward network. This lowers computational time and reduce number of
training weights. Thus architecture proposed is capable of using information
from both ahead and behind in time.
Figure 20: Schematic representation of bi-directional ARE. This is useful when
both past and future sensor measurements are available.
(a) Periodic vortex shedding (b) Transient flow
Figure 21: Comparison between B-RNN based ARE and RNN based ARE
To feed sensor measurements $s^{t-k,\ldots,t+k}$ into network, data is
splitted into $s^{t-k,\ldots,t}$ and $s^{t,\ldots,t+k}$ as shown in the Fig.
20. Note that although more accurate, this network can use only be used when
delay in time of reconstruction is acceptable also it uses odd sequence length
of sensor measurements. Output of forward and backward network is concatenated
and passed to feedforward network which yields the final estimated latent
representation of middle sensor measurement $s^{t}$. Thus training of
$\mathcal{N}$ can be formulated as follows
$\bm{\theta}_{M}^{*}=\arg\min_{\bm{\theta}}\sum_{i=1}^{N_{\text{samp}}}\left\|h^{t}-\mathcal{N}\left(\bm{s}^{t-k:t+k};\bm{\theta}_{M}\right)\right\|+\lambda\left\|\bm{\theta}_{M}\right|_{2}^{2},$
(23)
where $\theta_{M}$ are parameters of combined RNN and feed-forward network
$\mathcal{N}:\mathbb{R}^{N_{s}*(2k+1)}\mapsto\mathbb{R}^{h}$.
### B.1 Results
Fig. 21 shows a comparative assessment between B-RNN based ARE, and RNN based
ARE. Results corresponding to one, two, five, and ten sensors are presented.
For one sensor, results obtained using B-RNN based ARE significantly
outperforms those obtained using RNN based ARE. However, as the number of
sensors increases, the results obtained using the two approaches becomes
identical. One counter-intuitive result is obtained for the periodic vortex
shedding problem where the result obtained using B-RNN based ARE found to be
worse than that obtained using RNN based ARE. It is probably because the
neural network parameters for B-RNN has converged to a local minimum.
|
# Investigating Bi-Level Optimization for Learning and Vision from a Unified
Perspective:
A Survey and Beyond
Risheng Liu, Jiaxin Gao, Jin Zhang, Deyu Meng, and Zhouchen Lin, R. Liu and J.
Gao are with the DUT-RU International School of Information Science &
Engineering, Dalian University of Technology, and also with the Key Laboratory
for Ubiquitous Network and Service Software of Liaoning Province, Dalian
116024, China. E-mail<EMAIL_ADDRESS><EMAIL_ADDRESS>R. Liu is
the corresponding author. J. Zhang is with the Department of Mathematics,
Southern University of Science and Technology, and National Center for Applied
Mathematics Shenzhen, China, E-mail<EMAIL_ADDRESS>Meng is with
School of Mathematics and Statistics and Ministry of Education Key Lab of
Intelligent Networks and Network Security, Xi’an Jiaotong University, Xi’an,
Shaanxi, China. E-mail<EMAIL_ADDRESS>Lin is with the Key
Laboratory of Machine Perception (Ministry of Education), School of
Electronics Engineering and Computer Science, Peking University, Beijing
100871, China, and also with the Cooperative Medianet Innovation Center,
Shanghai Jiao Tong University, Shanghai 200240, China. E-mail:
<EMAIL_ADDRESS>received April 19, 2005; revised August 26, 2015.
###### Abstract
Bi-Level Optimization (BLO) is originated from the area of economic game
theory and then introduced into the optimization community. BLO is able to
handle problems with a hierarchical structure, involving two levels of
optimization tasks, where one task is nested inside the other. In machine
learning and computer vision fields, despite the different motivations and
mechanisms, a lot of complex problems, such as hyper-parameter optimization,
multi-task and meta learning, neural architecture search, adversarial learning
and deep reinforcement learning, actually all contain a series of closely
related subproblms. In this paper, we first uniformly express these complex
learning and vision problems from the perspective of BLO. Then we construct a
best-response-based single-level reformulation and establish a unified
algorithmic framework to understand and formulate mainstream gradient-based
BLO methodologies, covering aspects ranging from fundamental automatic
differentiation schemes to various accelerations, simplifications, extensions
and their convergence and complexity properties. Last but not least, we
discuss the potentials of our unified BLO framework for designing new
algorithms and point out some promising directions for future research.
###### Index Terms:
Bi-level optimization, Learning and vision applications, Value-function-based
reformulation, Best-response mapping, Explicit and implicit gradients.
## 1 Introduction
Bi-Level Optimization (BLO) is the hierarchical mathematical program where the
feasible region of one optimization task is restricted by the solution set
mapping of another optimization task (i.e., the second task is embedded within
the first one) [1]. The outer optimization task is commonly referred to as the
Upper-Level (UL) problem, and the inner optimization task is commonly referred
to as the Lower-Level (LL) problem. BLOs involve two kinds of variables,
referred to as the UL and LL variables, accordingly.
The origin of BLOs can be traced to the domain of game theory and is known as
Stackelberg competition [2]. Subsequently, it has been investigated in view of
many important applications in various fields of science and engineering,
particularly in economics, management, chemistry, optimal control, and
resource allocation problems [3, 4, 5, 6]. Especially, in recent years, a
great amount of modern applications in the fields of machine learning and
computer vision (e.g., hyper-parameter optimization [7, 8, 9, 10], multi-task
and meta learning [11, 12, 13], neural architecture search [14, 15, 16],
generative adversarial learning [17, 18, 19], deep reinforcement learning [20,
21, 22] and image processing and analysis [23, 24, 25], just name a few) have
arisen that fit the BLO framework.
In general, most of the earlier BLOs are highly complicated and
computationally challenging to solve due to their nonconvexity and non-
differentiability [26, 27]. Despite their apparent simplicity, BLOs are
nonconvex problems with an implicitly determined feasible region even if the
UL and LL subproblems are convex [28, 29]. Indeed, it has been proved that
even strictly checking the local optimality of the simplest BLO model (e.g.,
linear BLO) is still a NP-hard problem [30, 31]. In addition, the existence of
multiple optima for the LL subproblem can result in an inadequate formulation
of BLOs, which could aggravate the difficulty of theoretical analysis [32].
Despite the challenges, a lot of research topics consisting of methods and
applications of BLOs have followed in this field, see [27, 33]. Early studies
focused on numerical methods, including extreme-point methods [34], branch-
and-bound methods [3, 35], descent methods [36, 37], penalty function methods
[38, 39], trust-region methods [40, 41], and so on. The most often used
procedure is to replace the LL subproblem with its Karush–Kuhn–Tucker (KKT)
conditions, and if assumptions are made (such as smoothness, convexity, among
others) the BLOs can be transformed into single-level optimization problems
[42, 43, 44]. However, due to the high complexity of bi-level models, solving
BLOs for large-scale and high-dimensional practical applications in learning
and vision fields is still challenging [45].
The classical idea (e.g., the first-order approach in economics literature) to
reformulate BLO is to replace the LL subproblem Eq. (1) by its KKT conditions
and minimize over the original variables $\mathbf{x}$ and $\mathbf{y}$ as well
as the multipliers. The resulting problem is a so-called Mathematical Program
with Equilibrium Constraints (MPEC) [46, 47]. Unfortunately, MPECs are still a
challenging class of problems because of the presence of the complementarity
constraint [48]. Solution methods for MPECs can be categorized into two types
of approaches. The first one, namely, the nonlinear programming approach
rewrites the complementarity constraint into nonlinear inequalities, and then
allows to leverage powerful numerical nonlinear programming solvers. The other
one, namely, the combinatorial approach tackles the combinatorial nature of
the disjunctive constraint. Despite the difficulties, MPEC has been studied
intensively in the last three decades [49]. Recently, some progress on the
MPEC approach in dealing with BLOs have been witnessed by the community of
mathematical programming, in the context of selecting optimal hyper-parameters
in regression and classification problems. There are two issues caused by the
multipliers in the MPEC approach. First, in theory, if there exist more than
one multipliers for the LL subproblem, MPEC will not be equivalent to the
original BLO (in the local optimality scenario) [50]. Second, the introduced
auxiliary multiplier variables can limit the numerical efficiency when solving
the BLO problem.
In recent years, a variety of machine learning and computer vision tasks,
including but not limited to, hyper-parameter optimization [51, 52, 13, 53]),
multi-task and meta learning [54, 55, 56, 57], neural architecture search [14,
16, 58, 59], adversarial learning [18, 17, 60, 21], and deep reinforcement
learning [20, 61, 21, 62], have been investigated in application scenarios.
Despite the different motivations and mechanisms, all these problems contain a
series of closely related subproblems and have a natural hierarchical
optimization structure. However, although received increasing attentions in
both academic and industrial communities, there still lack a unified
perspective to understand and formulate these different categories of
hierarchical learning and vision problems.
We notice that most previous surveys on BLOs (e.g., [1, 63, 64, 65, 66, 67,
68]) are purely from the viewpoint of mathematical programming and mainly
focus on the formulations, properties, optimality conditions and these
classical solution algorithms, such as evolutionary methods [5]. In contrast,
the aim of this paper is to utilize BLO to express a variety of complex
learning and vision problems, which explicitly or implicitly contain closely
related subproblems. Furthermore, we present a unified perspective to
comprehensively survey different categories of gradient-based BLO
methodologies in specific learning and vision applications. In particular, we
first provide a literature review on various complex learning and vision
problems, including hyper-parameter optimization, multi-task and meta
learning, neural architecture search, adversarial learning, deep reinforcement
learning and so on. We demonstrate that all these tasks can be modeled as a
general BLO formulation. Following this perspective, we then establish a best-
response-based single-level reformulation to express these existing BLO
models. By further introducing a unified algorithmic framework on the single-
level reformulation, we can uniformly understand and formulate these existing
gradient-based BLOs and analyze their accelerations, simplifications,
extensions, and convergence and complexity proprieties. Finally, we
demonstrate the potentials of our framework for designing new algorithms and
point out some promising research directions for BLO in learning and vision
fields.
Compared with existing surveys on BLOs, our major contributions can be
summarized as follows:
1. 1.
To the best of our knowledge, this is the first survey paper to focus on
uniformly understanding and (re)formulating different categories of complex
machine learning and computer vision tasks and their solution methods
(especially in the context of deep learning) from the perspective of BLO.
2. 2.
By introducing a best-response-based single-level reformulation and
constructing a best-response-based algorithmic framework, we obtain a general
and flexible platform that can successfully unify different existing gradient-
based BLO methodologies and uniformly analyze these accelerations,
simplifications, and extensions in literature.
3. 3.
The convergence behaviors of gradient-based BLOs are comprehensively analyzed.
Especially, we establish a general convergence analysis template to
investigate the iteration behaviors of a series of gradient-based BLOs from a
unified perspective. The time and space complexity of various mainstream
schemes is also systematically analyzed.
4. 4.
Our gradient-based BLO platform not only comprehensively covers mainstream
gradient-based BLO methods, but also has potentials for designing new BLO
algorithms to deal with more challenging tasks. We also point out some
promising directions for future research.
We summarize our mathematical notations in Table I. The remainder of this
paper is organized as follows. We first introduce some necessary fundamentals
of BLOs in Section 2. Then, Section 3 provides a comprehensive survey of
various learning and vision applications that all can be modeled as BLOs. In
Section 4, we establish an algorithmic framework in a unified manner for
existing gradient-based BLO schemes. Within this framework, we further
understand and formulate two different categories of BLOs (i.e., explicit and
implicit gradients for best-response) in Section 5 and Section 6,
respectively. We also discuss the so-called lower-level singleton issue of
BLOs in Section 7. The convergence and complexity properties of these
gradient-based BLOs are discussed in Section 8. Section 9 puts forward
potentials of our framework for designing new algorithms to deal with more
challenging pessimistic BLOs. Finally, Section 10 points out some promising
directions for future research.
TABLE I: Summary of mathematical notations. Notation | Description | Notation | Description
---|---|---|---
$\mathcal{D}_{\mathtt{tr}}$/$\mathcal{D}_{\mathtt{val}}$ | Training/Validation data | $o^{ij}/\mathbf{x}_{o}^{ij}$ | Operations/Operation weights
$\pi/r$ | Policy/Reward | $s/a$ | State/Action
$Q^{\pi}$ | Q-function under $\pi$ | $Q^{\pi}(s,a)$ | State-action value-function
$G/D$ | Generator/Discriminator | $\mathbf{u}/\mathbf{v}$ | Real-world image/Random noise
$\rho_{t}$ | Aggregation parameters | $\mathbf{x}\in\mathbb{R}^{m}$/$\mathbf{y}\in\mathbb{R}^{n}$ | UL/LL variable
$F/f$ | UL/LL objective | $\mathcal{S}(\mathbf{x})$ | Solution set of the LL subproblem (given $\mathbf{x}$)
$\mathbf{y}^{*}(\mathbf{x})$ | BR mapping | $\widetilde{\mathcal{S}}(\mathbf{x})$ | Solution set of the ISB subproblem (given $\mathbf{x}$)
$\mathtt{dist}(\cdot)$ | Point-to-set distance | $\circ$ | Compound operation
$\Psi(\mathbf{x})$ | $\Psi=\Psi_{T}\circ\cdots\circ\Psi_{1}\circ\Psi_{0}$ | $\Psi_{\bm{\theta}}(\mathbf{x})$ | Hyper-network with parameters $\bm{\theta}$
$\Psi_{t}$ | Dynamical system at $t$-th stage | $\frac{\partial\varphi(\mathbf{x})}{\partial\mathbf{x}}$ | Gradient of $\mathbf{x}$
$\frac{\partial F(\mathbf{x},\mathbf{y}^{*}(\mathbf{x}))}{\partial\mathbf{x}}$ | Direct gradient of $\mathbf{x}$ | $\mathbf{G}(\mathbf{x})$ | Indirect gradient of $\mathbf{x}$
$\frac{\partial\mathbf{y}^{*}(\mathbf{x})}{\partial\mathbf{x}}$ | BR Jacobian | $\frac{\partial\varphi(\mathbf{x}^{k})}{\partial\mathbf{x}^{k}}$ | Numerical BR Jacobian w.r.t. $\mathbf{x}^{k}$
$\varphi(\mathbf{x})$ | UL value-function | $\psi(\mathbf{x})$ | LL value-function
$\mathbf{d}(\mathbf{y}_{t-1};\mathbf{x})$ | Optimistic aggregated gradient | $\widetilde{\mathbf{d}}(\mathbf{y}_{t-1};\mathbf{x})$ | Pessimistic aggregated gradient
$\left(\frac{\partial^{2}f}{\partial\mathbf{y}\partial\mathbf{y}^{\prime}}\right)^{-1}$ | Inverse Hessian matrix | $\left(\frac{\partial^{2}f}{\partial\mathbf{y}\partial\mathbf{y}^{\prime}}\right)^{-1}\frac{\partial F}{\partial\mathbf{y}}$ | Inverse Hessian-vector product
$t$/$k$ | Index of the LL/UL iteration | $T/K$ | Maximum LL/UL iteration number
$(\cdot)_{t}$ | $t$-th LL iteration | $(\cdot)^{k}$ | $k$-th UL iteration
$(\cdot)^{{}^{\prime}}$ | Transposition operation | $(\mathbf{x}^{*},\mathbf{y}^{*})$ | The optimal UL and LL solutions
$\mathbf{Z}_{T}$ | $\sum_{t=1}^{T}\left(\prod_{i=t+1}^{T}\mathbf{A}_{i}\right)\mathbf{B}_{t}$ | $\mathbf{Z}_{T-M}$ | $\sum_{t=T-M+1}^{T}\left(\prod_{i=t+1}^{T}\mathbf{A}_{i}\right)\mathbf{B}_{t}$
$\mathbf{P}(\mathbf{y}_{t-1},\bm{\omega})$ | Layer-wise transformation | $\sum\limits_{j=0}^{\infty}\left(\mathbf{I}-\frac{\partial^{2}f}{\partial\mathbf{y}\partial\mathbf{y}^{\prime}}\right)^{j}$ | Neumann series
$\mathbf{A}_{t}$ | $\frac{\partial\Psi_{t}(\mathbf{y}_{t-1};\mathbf{x})}{\partial\mathbf{y}_{t-1}}$ | $\mathbf{B}_{t}$ | $\frac{\partial\Psi_{t}(\mathbf{y}_{t-1};\mathbf{x})}{\partial\mathbf{x}}$
$\inf_{\mathbf{y}\in\mathcal{S}(\mathbf{x})}F(\mathbf{x},\mathbf{y})$ | Optimistic objective | $\psi_{\mu}(\mathbf{x})$ | parameterized LL value-function (with $\mu$)
$\sup_{\mathbf{y}\in\mathcal{S}(\mathbf{x})}F(\mathbf{x},\mathbf{y})$ | Pessimistic objective | $\varphi_{\mu,\theta,\tau}\left(\mathbf{x}\right)$ | parameterized UL value-function (with $\mu$, $\theta$ and $\tau$)
## 2 Fundamentals of Bi-Level Optimization
Bi-Level Optimization (BLO) contains two levels of optimization tasks, where
one is nested within the other as a constraint. The inner (or nested) and
outer optimization tasks are often respectively referred to as the Lower-Level
(LL) and Upper-Level (UL) subproblems [1]. Correspondingly, there are two
types of variables, namely, the LL ($\mathbf{y}\in\mathbb{R}^{n}$) and UL
($\mathbf{x}\in\mathbb{R}^{m}$) variables. Specifically, the LL subproblem can
be formulated as the following parametric optimization task
$\min_{\mathbf{y}\in\mathcal{Y}}f(\mathbf{x},\mathbf{y}),\
(\mbox{parameterized by $\mathbf{x}$}),$ (1)
where we consider a continuous function
$f:\mathbb{R}^{m}\times\mathbb{R}^{n}\to\mathbb{R}$ as the LL objective and
$\mathcal{Y}\subseteq\mathbb{R}^{n}$ is a nonempty set. By denoting the value-
function as
$\psi(\mathbf{x}):=\min_{\mathbf{y}\in\mathcal{Y}}f(\mathbf{x},\mathbf{y})$,
we can define the solution set of the LL subproblem with given $\mathbf{x}$ as
$\mathcal{S}(\mathbf{x}):=\left\\{\mathbf{y}\in\mathcal{Y}~{}|~{}f(\mathbf{x},\mathbf{y})\leq\psi(\mathbf{x})\right\\}.$
Then the standard BLO problem can be formally expressed as
$\min\limits_{\mathbf{x}\in\mathcal{X}}F(\mathbf{x},\mathbf{y}),\ s.t.\
\mathbf{y}\in\mathcal{S}(\mathbf{x}),$ (2)
where the UL objective $F:\mathbb{R}^{m}\times\mathbb{R}^{n}\to\mathbb{R}$ is
also a continuous function and the feasible set
$\mathcal{X}\subseteq\mathbb{R}^{m}$. In fact, a feasible solution to BLO in
Eq. (2) should be a vector of UL and LL variables, such that it satisfies all
the constraints in Eq. (2), and the LL variables are optimal to the LL
subproblem in Eq. (1) for the given UL variables as parameters. In Fig. 1, we
provide a simple visual illustration for BLOs stated in Eq. (2).
Figure 1: Illustrating the problem of BLO. (a) first shows a standard BLO
problem with the situation of multiple solutions of $f$. Green curves denote
LL objectives denoted by $f$, and their corresponding minimizers given by
$\mathcal{S}(\mathbf{x})$ were shown as green dots. The red curve represents
the UL objective $F$, whose minimizer is shown as the red dot. (b) further
illustrates that, in general, not all points (green dots) in
$\mathcal{S}(\mathbf{x})$ could minimize the UL objective denoted by $F$.
The above BLO problem has a natural interpretation as a non-cooperative game
between two players (i.e., Stackelberg game [1]). Correspondingly, we may also
call the UL and LL subproblems as the leader and follower, respectively. Then
the “leader” chooses the decision $\mathbf{x}$ first, and afterwards the
“follower” observes $\mathbf{x}$ so as to respond with a decision
$\mathbf{y}$. Therefore, the follower may depend on the leader’s decision.
Likewise, the leader has to satisfy a constraint that depends on the
follower’s decision.
It is worthwhile noting that the LL subproblem may have multiple solutions for
every (or some) fixed value of the UL decision making variable $\mathbf{x}$.
When the solution of the LL subproblem is not unique, it is difficult for the
leader to predict which point in $\mathcal{S}(\mathbf{x})$ the follower will
choose (see Fig. 1 (b) for example).
## 3 Understanding and Modeling Practical Problems by BLOs
In this section, we demonstrate that even with different motivations and
mechanisms, a variety of modern complex learning and vision tasks (e.g, hyper-
parameter optimization, multi-task and meta learning, neural architecture
search, adversarial learning, deep reinforcement learning and so on) actually
share close relationships from the BLO perspective. Moreover, we provide a
uniform BLO expression to (re)formulate all these problems. Table II provides
a summary of learning and vision applications, which can be understood and
modeled by BLO.
TABLE II: Summary of related learning and vision applications that can be (re)formulated as BLOs. The abbreviations are listed as follows: Hyper-parameter Optimization (HO), Meta-Feature Learning (MFL), Meta-Initialization Learning (MIL), Neural Architecture Search (NAS), Adversarial Learning (AL), and Deep Reinforcement Learning (DRL). Task | Important work | Other work
---|---|---
HO | [51] (ICML, 2017), [13] (AISTATS, 2019), [32] (ICML, 2020) | [69] (SIAM, 2014), [70] (EURO, 2020), [71] (ICML, 2015), [72] (AISTATS, 2020), [73] (ICML, 2018), [74] (ICML, 2016), [75] (arXiv, 2019), [9] (ICLR, 2019), [53] (ICML, 2021), [76] (ICML, 2017), [77] (NIPS, 2020), [12] (ICML, 2018)
MFL | [12] (ICML, 2018), [56] (ICML, 2017) | [11] (arXiv, 2017), [79] (CVPR, 2018), [80] (CVPR, 2018), [81] (ICLR, 2018), [82] (ICLR, 2017), [83] (ICLR, 2018), [84] (ICML, 2019)
MIL | [85] (NIPS, 2019), [57] (ICLR, 2019), [86] (ICLR, 2019), [87] (NIPS, 2019) | [88] (ICLR, 2017), [55] (ICML, 2017), [89] (ICML, 2017), [90] (arXiv, 2018), [91] (arXiv, 2019), [92] (CVPR, 2020),[93] (AAAI, 2020), [94] (arXiv, 2018), [95] (NIPS, 2019), [96] (ICLR, 2020), [97] (ICML, 2019),[98] (ICML, 2018), [99] (CVPR, 2020), [100] (AAAI, 2020), [101] (ICASSP, 2020)
NAS | [14] (ICLR, 2019), [102] (NIPS, 2018), [58] (TGRS, 2020), [103] (CVPR, 2020), [59] (ICLR, 2019) | [104] (ICLR, 2019), [105] (CVPR, 2018), [16] (ICLR, 2019), [106] (ICCV, 2019), [107] (AISTATS, 2020), [108] (CVPR, 2020), [109] (AAAI, 2020), [110] (CVPR, 2020), [111] (CVPR, 2019), [112] (NIPS, 2019), [113] (ICCV, 2019), [114] (NIPS, 2019), [115] (CVPR, 2020), [116] (arXiv, 2019), [117] (SIGKDD, 2020), [118] (CVPR, 2020), [119] (CVPR, 2020), [120] (arXiv, 2020)
AL | [121] (arXiv, 2016), [21] (arXiv, 2016) | [122] (AAAI, 2020), [123] (CVPR, 2020), [18] (arXiv, 2018), [17] (PR, 2019), [124] (ICML, 2020), [60] (CVPR, 2020), [19] (CVPR, 2020), [125] (ICML, 2018)
DRL | [20] (AAAI, 2020), [21] (arXiv, 2016), [61] (ICML, 2020) | [126] (CIRED, 2019), [62] (arXiv, 2020), [127] (NIPS, 2019), [128] (ICML, 2020), [129] (AAMAS, 2020), [130] (arXiv, 2019), [131] (TSG, 2019), [132] (NeurIPS, 2016), [133] (NeurIPS, 2017), [134] (arXiv, 2018), [61] (ICML, 2020), [135] (ICML, 2019)
Others | [136] (SIAM, 2013), [23] (SSVM, 2015), [137] (TIP, 2020), [138] (TNNLS, 2020), [139] (TIP, 2016) | [140] (ICML, 2016),[141] (ICLR, 2019), [25] (IJCAI, 2020), [142] (arXiv, 2019), [143] (UAI, 2020), [144] (arXiv, 2021), [145] (arXiv, 2020), [146] (TIP, 2020), [147] (arXiv, 2019), [148] (arXiv, 2020), [149] (T-RO, 2020), [150] (WACV, 2020), [151] (arXiv, 2020), [152] (NIPS, 2020), [153] (arXiv, 2020), [154] (arXiv, 2020), [155] (CVPR, 2020), [156] (ICLR, 2018)
### 3.1 Hyper-parameter Optimization
Hyper-parameter Optimization (HO) refers to the problem of identifying the
optimal set of hyper-parameters that can’t be learned using the training data
alone. Early in learning and vision areas, designing regularized models or
support vector machines are generally the recommended approaches of selecting
hyper-parameters [157]. Based on the representation of the hierarchical
structure, these approaches are first expressed as a BLO problem and then
transformed into the single-level optimization problem by replacing the LL
subproblem with its optimality condition [158]. Due to the high computational
cost, especially in high-dimensional hyper-parameter space, these original
methods even could not guarantee a local optimal solution [159].
In recent years, gradient-based HO methods with deep neural networks have
received extensive attention, which are generally divided into two categories:
iterative differentiation (i.e., [7, 160, 23, 71, 10, 11, 12, 75, 13, 73]) and
implicit differentiation (i.e., [161, 162, 136, 163, 9, 8, 74, 164, 72]),
depending on how the gradient (w.r.t. hyper-parameters) can be computed. The
former approximates the best-response function by performing several steps of
gradient descent on the loss function, while the latter derives the hyper-
gradients through the implicit function theory. One particular type of
gradient-based HO is the data hyper-cleaning problem [51, 13], which generally
trains a linear classifier with a cross-entropy function (w.r.t. parameters
$\mathbf{y}$) and learns to optimize the hyper-parameters $\mathbf{x}$ with a
$\ell_{2}$ regularization function.
---
Figure 2: Schematic diagram of HO. The UL subproblem involves optimization of
hyper-parameters $\mathbf{x}$ based on
$\left(\mathcal{D}_{\mathtt{tr}},\mathcal{D}_{\mathtt{val}}\right)$, while the
LL subproblem involves optimization of weight parameters $\mathbf{y}$, aiming
to find the learning algorithm $g_{\mathbf{y}}(\cdot)$ based on
$\mathcal{D}_{\mathtt{tr}}$.
Indeed, HO can be understood as the most straightforward application of BLO in
learning and vision fields [157]. Specifically, the UL objective
$F(\mathbf{x},\mathbf{y};\mathcal{D}_{\mathtt{val}})$ aims to minimize the
validation set loss with respect to the hyper-parameters (e.g. weight decay),
and the LL objective $f(\mathbf{x},\mathbf{y};\mathcal{D}_{\mathtt{tr}})$
needs to output a learning algorithm by minimizing the training loss with
respect to the model parameters (e.g. weights and biases). As illustrated in
Fig. 2, the full dataset $\mathcal{D}$ is divided into the training and
validation datasets (i.e.,
$\mathcal{D}_{\mathtt{tr}}\cup\mathcal{D}_{\mathtt{val}}$) and we instantiate
how to model the HO task from the perspective of BLO. Inspired by this nested
optimization, most HO applications can be characterized by the bi-level
structure and formulated as the BLO problems. The UL subproblem involves the
optimization of hyper-parameters $\mathbf{x}$ and the LL subproblem (w.r.t.
weight parameters $\mathbf{y}$) aims to find the learning algorithm
$g_{\mathbf{y}}(\cdot)$ by minimizing the training loss.
### 3.2 Multi-task and Meta Learning
The goal of meta learning (a.k.a., learning to learn) is to design models that
can learn new skills or adapt to new environments rapidly with a few training
examples (see Fig. 3 for a schematic diagram). As a variant of meta learning,
multi-task learning just intends to jointly perform all the given tasks [165,
166]. One of the most well-known instances of meta learning is few-shot
classification (i.e., $N$-way $M$-shot). Each task is a $N$-way classification
designed to learn the meta-parameter with $M$ training samples selected from
each of the class. Specially, the full meta training data set
$\mathcal{D}=\\{\mathcal{D}^{j}\\}$ ($j=1,\cdots,N$) can be segmented into
$\mathcal{D}^{j}=\mathcal{D}_{\mathtt{tr}}^{j}\cup\mathcal{D}_{\mathtt{val}}^{j}$,
where $\mathcal{D}^{j}$ is linked to the $j$-th task.
---
Figure 3: Illustrating the training process of meta learning. The whole
process is visualized to learn new tasks quickly by drawing upon related tasks
on corresponding data sets. It can be decomposed into two parts: the “base-
learner” trained for operating a given task and the “meta-learner” trained to
learn how to optimize the base-learner.
According to the dependency between the meta-parameters and the network
parameters, current meta learning based methods can be roughly categorized as
two groups, i.e., meta-feature learning and meta-initialization learning, as
can be seen in Fig. 4. Specifically, meta-initialization learning aims to
investigate the meta information of multiple tasks by the network
initialization, which can also be understood as the promotion of fine-tuning
[85, 87]. From the BLO perspective, we actually formulate the network
parameters and their initialization (based on multi-task information) by the
LL and UL subproblems, respectively. In contrast, meta-feature learning
methods first separate the network architecture as the meta feature extraction
part and the task-specific part. Then they formulate a hierarchical learning
process [11, 12, 56]. So in such tasks, we use the UL and LL subproblems to
model the meta-feature part and the task-specific part, respectively.
#### 3.2.1 Meta-feature Learning
Meta-Feature Learning (MFL) aims to learn a sharing meta feature
representation of all tasks. Recently, series of meta learning based
approaches show that multi-task with hard parameter sharing and meta-feature
representation are essentially similar [167, 168]. The optimization of meta-
learner with respect to meta-parameters based on the UL subproblem is similar
to HO [11, 12, 73]. The cross-entropy function
$\ell(\mathbf{x},\mathbf{y}^{j};\mathcal{D}_{\mathtt{tr}}^{j})$ is actually
considered as the task-specific loss for the $j$-th task on the meta training
data set to define the LL objective.
As illustrated in the subfigure (a) of Fig. 4, following the bi-level
framework, the network architecture in this category can be subdivided into
two groups. The first is the cross-task intermediate representation layer
parameterized by $\mathbf{x}$ (illustrated by the blue block), outputting the
meta features. The second is the logistic regression layer parameterized by
$\mathbf{y}^{j}$ (illustrated by the green block), as the ground classifier
for the $j$-th task. As can be seen, the feature layers are shared across all
episodes, while the softmax regression layer is episode (task) specific. We
can also observe that the process of network forward propagation corresponds
to the process of passing from the feature extraction part to the softmax
part.
#### 3.2.2 Meta-initialization Learning
Meta-Initialization Learning (MIL) aims to learn a meta initialization for all
tasks. MAML [89], known for its simplicity, estimates initialization
parameters with the cross-entropy and mean-squared error for supervised
classification and regression tasks purely by the gradient-based search.
Except for initial parameters, recent approaches have focused on learning
other meta variables, such as updating strategies (e.g., descent direction and
learning rate [169, 88, 91]) and an extra preconditioning matrix (i.e., [78,
83, 98]). Moreover, implicit gradient methods have a rapid development in the
context of few-shot meta learning. There exist a large variety of algorithms
replacing the gradient process of the optimization of base-learner through
calculation of implicit meta gradient [85, 95, 170, 86]. Due to the large
amount of computation required to calculate the Hessian vector product in the
training process, various Hessian-free algorithms have been proposed to
alleviate the costly computation of second-order derivatives, including but
not limited to [94, 54, 96, 57, 55, 56]. In particular, various first-order
approximation BLO algorithms have been proposed to avoid the time-consuming
calculation of second-order derivatives in [90]. For instance, a modularized
optimization library was proposed in [53] to unify several meta learning
algorithms into a common BLO framework111The code for this library is
available at https://github.com/dut-media-lab/BOML..
As can be shown in subfigure (b) of Fig. 4, $\mathbf{x}$ denoted by blue
blocks corresponds to network initialization parameters, and $\mathbf{y}$
denoted by green blocks corresponds to model parameters and is treated as the
updated variable satisfying the condition $\mathbf{y}_{0}^{j}=\mathbf{x}$.
Compared to MFL, there is no deeply intertwined and entangled relationship
between two variables $(\mathbf{x},\mathbf{y}^{j})$, and $\mathbf{x}$ is only
explicitly related to $\mathbf{y}$ in the initial state. As a bi-level coupled
nested loop strategy, the LL subproblem based on base-learner is trained for
operating a given task, and the UL subproblem based on meta-learner aims to
learn how to optimize the base-learner. Among the well-known approaches in
this direction, most recent approaches (i.e., [171, 90]) have claimed that the
LL objective is denoted by the task-specific loss on the training data set,
i.e.,
$f(\mathbf{x},\\{\mathbf{y}^{j}\\})=\ell(\mathbf{x},\mathbf{y}^{j};\mathcal{D}_{\mathtt{tr}}^{j})$.
By utilizing cross-entropy function, the UL objective is given by
$F(\mathbf{x},\\{\mathbf{y}^{j}\\})=\sum_{j}\ell(\mathbf{x},\mathbf{y}^{j};\mathcal{D}_{\mathtt{val}}^{j}).$
(a) MFL
(b) MIL
Figure 4: Illustration of two architectures that are generally applied to
multi-task and meta learning: MFL and MIL. Both of them can be separated into
two parts: meta-parameters denoted by $\mathbf{x}$ (blue blocks) and
parameters denoted by $\mathbf{y}^{j}$ (green blocks). (a) shows meta-
parameters for features shared across tasks and parameters of the logistic
regression layer. (b) shows meta (initial) parameters shared across tasks and
parameters of the task specific layer.
Both MFL and MIL are essential solution strategies of one optimizer based on
another optimizer, thus conforming to the construction of the BLO scheme. As a
task-specific loss associated with the $j$-th task, the LL objective can be
defined as
$\mathbf{y}^{j}\in\arg\min_{\mathbf{y}^{j}\in\mathcal{Y}}f\left(\mathbf{x},\mathbf{y}^{j};\mathcal{D}_{\mathtt{tr}}^{j}\right)$,
$j=1,\cdots,N$. Also, based on $\\{\mathcal{D}_{\mathtt{val}}^{j}\\}$, the UL
objective can be given by
$\min_{\mathbf{x}\in\mathcal{X}}F\left(\mathbf{x},\\{\mathbf{y}^{j}\\};\\{\mathcal{D}_{\mathtt{val}}^{j}\\}\right)$.
To summarize, the UL meta-learner performs gradient descent operations and
updates the meta-parameter with feedback from base-learners to extract
generalized meta knowledge. Subsequently, the better meta knowledge is fed
into the base-learner (i.e., the LL subproblem) as part of its model for
optimizing $\mathbf{y}$, thereby forming an optimization cycle.
### 3.3 Neural Architecture Search
Neural Architecture Search (NAS) seeks to automate the process of choosing the
optimal neural network architecture [172]. Recently, there has aroused a great
deal of interest in gradient-based differentiable NAS methods [14, 173, 174].
Specifically, these gradient-based differentiable NAS methods mainly contain
three main concepts: search space, search strategy and performance estimation
strategy. As shown in Fig. 5, by designing an architecture search space, they
generally use a certain search strategy to find the optimal network
architecture. Such a process can be regarded as the system of optimizing the
operation and connection of each node.
---
Figure 5: Schematic diagram of NAS. Derived from a predefined search space
$\mathcal{A}$, NAS first selects an architecture $A$ to transport into the
performance estimation strategy, then returns the estimated performance of $A$
to the search strategy.
DARTS [14], the most well-known instance, relaxed the search space to be
continuous and conducted searching for architectures in a differentiable way
to simultaneously optimize the architectures and weights. Actually, each
operation corresponds to a coefficient in DARTS. By denoting
$\mathbf{x}=\\{\mathbf{x}^{ij}\\}$ as the architecture parameters and
$\mathbf{x}^{ij}$ as the form of connection between two nodes, the expression
formula of mixed operations $\bar{o}^{ij}(\cdot)$ based on the softmax
function can be written as
$\bar{o}^{ij}(\cdot)=\sum\limits_{o\in\mathcal{O}}\frac{\exp(\mathbf{x}_{o}^{ij})}{\sum\limits_{o^{\prime}\in\mathcal{O}}\exp(\mathbf{x}_{o^{\prime}}^{ij})}o(\cdot),$
where $o$ and $o^{\prime}$ are operations and $\mathcal{O}$ is the set of all
candidate operations. Then,
$o^{ij}=\arg\max_{o\in\mathcal{O}}\mathbf{x}_{o}^{ij}$ is further evaluated
and performed in order to obtain the optimal architecture. However, due to the
sharp deterioration in performance caused by the large number of skip
connections, a great deal of improved approaches have emerged, such as ENAS
[105], PC-DARTS [59], P-DARTS [106], just to name a few.
Currently, a series of gradient-based differentiable NAS methods combined with
meta learning have been proposed, see [175, 16, 176, 108]. Based on the bi-
level coupling mechanism, these gradient-based differentiable NAS methods have
achieved promising results in the numerous visual and learning applications,
such as image classification [58], semantic segmentation [115, 177, 111],
object detection [112, 113, 118, 103, 117], medical image analysis [177, 115],
video classification [140], recommendation system [120], graph network [116,
130] and representation learning [130], etc.
Given the proper search space, it is helpful for these gradient-based
differentiable NAS methods to derive the optimal architecture for different
vision and learning tasks. From the BLO’s point of view, the UL objective
w.r.t. the architecture weights (e.g. block/cell) can be parameterized by
$\mathbf{x}$. And the LL objective w.r.t. the model weights can be
parameterized by $\mathbf{y}$. Therefore, the full searching process can
virtually be formulated as a BLO paradigm, where the UL objective is defined
by $F(\mathbf{x},\mathbf{y};\mathcal{D}_{\mathtt{val}})$ based on the
validation data set $\mathcal{D}_{\mathtt{val}}$, and the LL objective is
given by $f(\mathbf{x},\mathbf{y};\mathcal{D}_{\mathtt{tr}})$ based on the
training data set $\mathcal{D}_{\mathtt{tr}}$.
### 3.4 Adversarial Learning
Adversarial Learning (AL) is currently deemed as one of the most important
learning tasks. It has been applied in a large variety of application areas,
i.e., image generation [60, 18, 123], adversarial attacks [178] and face
verification [17]. For example, the work proposed in [60] introduced an
adaptive BLO model for image generation, which guided the generator to
reasonably modify the parameters in a complementary and promoting way.
Moreover, a new adversarial training strategy has been proposed by learning a
parametric optimizer with neural networks to study the adversarial attack
[18]. As the current influential model, Generative Adversarial Network (GAN)
can be deemed as deep generative models [179]. Recently, targeting at finding
pure Nash equilibrium of generator and discriminator, the author proposed to
exploit a fully differentiable search framework by formalizing as solving a
bi-level mini-max optimization problem [19].
---
Figure 6: Illustrating the architecture of GAN. The generator $G$ is
represented as a deterministic feed forward neural network (red blocks),
through which a fixed random noise $\mathbf{v}$ is passed to output
$G(\mathbf{v})$. The discriminator $D$ is another neural network (green
blocks) which maps the sampled real-world image $\mathbf{u}\sim p_{data}$ and
$G(\mathbf{v})$ to a binary classification probability.
Most of the AL approaches can formulate the unsupervised learning problem as a
bi-level game between two opponents: a generator which samples from a
distribution, and a discriminator which classifies the samples as real or
false, as shown in Fig. 6. The goal of GAN is to minimize the duality gap
denoted by $\mathcal{V}(D,G)$:
$\begin{split}\min\limits_{G}\max\limits_{D}\mathcal{V}(D,G)&=\mathbb{E}_{\mathbf{u}\sim
p_{data}(\mathbf{u})}\log D(\mathbf{u})\\\
&+\mathbb{E}_{\mathbf{v}\sim\mathcal{N}_{(0,1)}}\log(1-D(G(\mathbf{v}))),\end{split}$
where the fixed random noise source $\mathbf{v}$ obtained from
$\mathbf{v}\sim\mathcal{N}_{(0,1)}$ is input into the generator $G$, which,
together with the sampled real-world image $\mathbf{u}\sim p_{data}$, is then
authenticated by the discriminator $D$. Notice that $\mathbb{E}$ denotes the
expectation which implies that the average value of some functions under a
probability distribution.
Indeed, AL problems generally correspond to the mini-max BLO problems, where
the UL discriminator denoted by $F$ targets on learning a robust classifier,
and the LL generator denoted by $f$ tries to generate the adversarial samples.
Specifically, the UL and LL objectives can be respectively formulated as
$\begin{split}F(\mathbf{x},\mathbf{y})&=-\mathbb{E}_{\mathbf{u}\sim
p_{data}(\mathbf{u})}\log D(\mathbf{u})\\\
&-\mathbb{E}_{\mathbf{v}\sim\mathcal{N}_{(0,1)}}\log(1-D(G(\mathbf{v}))),\end{split}$
$f(\mathbf{x},\mathbf{y})=-\mathbb{E}_{\mathbf{v}\sim\mathcal{N}_{(0,1)}}\log(D(G(\mathbf{v}))),$
where $G$ and $D$ are parameterized with variables $\mathbf{y}$ and
$\mathbf{x}$, respectively. In other words, the UL subproblem aims to reduce
the duality gap $\mathcal{V}(D,G)$ and the LL subproblem interactively
optimizes the discriminator parameters denoted by $\mathbf{x}$ to obtain the
optimal solution.
### 3.5 Deep Reinforcement Learning
Deep Reinforcement Learning (DRL) aims to make optimal decisions by
interacting with the environment and learning from the experiences. Indeed, a
variety of DRL tasks, including Single-Agent Reinforcement Learning (SARL)
[21, 22, 62], Multi-Agent Reinforcement Learning (MARL) [126, 129, 20, 180],
Meta Reinforcement Learning (MRL) [61, 135, 181, 182], and Imitation Learning
(IL) [132, 134, 133], which all can be modeled and tackled by BLO techniques.
As for SARL problems, Actor-Critic (AC) type methods have been widely studied
and viewed as a bi-level or two-time-scale optimization problems [62, 22], as
illustrated in Fig. 7. Indeed, AC type DRL methods often aim to simultaneously
learn a state-action value-function $Q^{\pi}$ that predicts to expect the
discounted cumulative reward and a policy which is optimal for that value
function:
$\leavevmode\resizebox{446.2658pt}{}{$Q^{\pi}(s,a)=\mathbb{E}_{s_{i+j}\sim\mathcal{P},r_{i+j}\sim\mathcal{R},a_{i+j}\sim\pi}\left(\sum\limits_{k=0}^{\infty}\gamma^{j}r_{i+j}|s_{i}=s,a_{i}=a\right)$},$
where $\mathcal{P}$ and $\mathcal{R}$ denote dynamics of the environment and
reward function, $s$ and $a$ are the state and action, $i$ and $j$ represent
the i-th and j-th steps, and $\mathbb{E}$ is the expectation which implies
that the average value of some function under a probability distribution. The
policy maximizes the expected discounted cumulative reward for that state-
action value-function, i.e., $\pi^{*}=\arg\max_{\pi}\mathbb{E}_{s_{0}\sim
p_{0},a_{0}\sim\pi}\left(Q^{\pi}(s_{0},a_{0})\right),$ where $s_{0}$, $a_{0}$
and $p_{0}$ correspond to the initial state, initial action and the initial
state distribution, respectively. Under the BLO paradigm, the actor and critic
correspond to the UL and LL variables, respectively. Let $\mathbf{x}$ denote
the parameters of the state-action value-function and $\mathbf{y}$ denote the
parameters of the policy $\pi$. The UL and LL objectives respectively take the
form
$\displaystyle F(\mathbf{x},\mathbf{y})$
$\displaystyle=\mathbb{E}{s_{i},a_{i}\sim\pi}(\mathtt{div}(\mathbb{E}_{s_{i+1},a_{i+1},r_{i+1}}$
$\displaystyle\left(r_{i+1}+\gamma Q(s_{i+1},a_{i+1})\right)\parallel
Q(s_{i},a_{i}))),$ $f(\mathbf{x},\mathbf{y})=-\mathbb{E}_{s_{0}\sim
p_{0},a_{0}\sim\pi}Q^{\pi}(s_{0},a_{0}),$
where $\mathtt{div}(\cdot||\cdot)$ represents any divergence.
MARL studies how multiple agents can collectively learn, collaborate, and
interact with each other in an environment. In the classical MARL system,
agents are treated equally and the goal is to solve the Markov game to an
arbitrary Nash equilibrium when multiple equilibria exist, thus lacking a
solution for selection. To address this issue, the work in [20] formulates
MARL as the multi-state model-free Stackelberg equilibrium learning problem.
Thus, under Markov games, they construct a BLO formulation to find Stackelberg
equilibrium to address the MARL task. Similarly, a multi-agent bi-level
cooperative reinforcement learning algorithm was proposed in [126] to solve
the stochastic decision-making problem.
In recent years, MRL approaches (a.k.a., meta learning on reinforcement
learning tasks), which aim to learn a policy that adapts fast to new tasks
and/or environments, have achieved remarkable success [183, 181]. For example,
the work in [182] learns a policy that can quickly adapt to other related
models only with one policy gradient step. By adding control variables into
gradient estimation, the work in [135] can obtain low variance estimates for
policy gradients. While the work in [61] characterizes the optimality gap of
the stationary points attained by MAML for both reinforcement learning and
supervised learning. Since all these works are based on the meta-
initialization platform, it is also nature to formulate these meta
reinforcement learning methods from the perspective of BLOs.
Generally, IL techniques are very useful when it is easier for an expert to
demonstrate the desired behavior rather than to specify a reward function
which would generate the same behavior or to directly learn the policy in DRL
tasks [184]. In recent years, by connecting imitation learning with generative
adversarial learning, a series of Generative Adversarial Imitation Learning
(GAIL) techniques [132, 134, 133] have been investigated to imitate an expert
in a model-free DRL scenario. Since GAIL type methods have a natural
connection to the mechanism of GANs, we can definitely formulate these models
using BLOs.
---
Figure 7: Illustrating the schematic diagram of AC learning. First the actor
$\pi$ interacts with the environment to learn the state-action value-function
$Q^{\pi}(s,a)$, and then the actor $\pi$ is again obtained based on
$Q^{\pi}(s,a)$.
### 3.6 Other Related Applications
The rapid development of deep learning has claimed its domination in the area
of image processing and analysis. In addition to the above mentioned tasks,
there exist a significant amount of other related learning and vision tasks
that can be re(formulated) as BLO problems, such as image enhancement [136,
24, 139, 185, 142], image registration [25], image-to-image translation [186],
image recognition [187], image compression [188] and other related works [143,
152, 141]. For example, the earlier work presented in [136] considered the
problem of parameter learning for image denoising models and incorporated
$p$-norm–based analysis priors. Under a BLO formulation, the LL subproblem was
given by a variational model which consisted of the data fidelity and
regularization term, and the UL subproblem was expressed by the loss function.
Furthermore, the work proposed in [139] formulated the discriminant dictionary
learning method for image recognition tasks as a BLO. From this point of view,
the UL subproblem can directly minimize the classification error, while the LL
subproblem can use the sparsity term and the Laplacian term to characterize
the intrinsic data structure.
By addressing a unified BLO problem, the LL subproblem is usually expressed as
fundamental models that conform to the laws or principles of physics, while
the UL subproblem usually considers the further constraints on variables [23,
140].
## 4 Gradient-based BLOs
In past years, gradient-based techniques have became the most popular BLO
solution strategies in learning and vision fields. In fact, one of the first
gradient-based BLO methodology is [30]. Currently, a variety of explicit
gradient-based methods have been investigated to solve BLOs [71, 73, 189].
Specifically, the works in [12, 51] first calculate gradient flow of the LL
objective and then perform either reverse or forward gradient computations for
the UL subproblem. Similar ideas have also been considered in [190, 23, 75],
but with different specific implementations. On the other hand, there also
exist some implicit gradient based methods [72, 85, 191] to use the implicit
function theorem to obtain the gradient. In this section, we first review
three categories of mainstream BLO formulations, which have been considered in
various application scenarios. We then demonstrate how to uniformly
reformulate these different BLOs from a single-level optimization perspective
and investigate the intrinsic structures of existing gradient-based BLO
algorithms within a unified algorithmic platform.
### 4.1 Different Formulations of BLO
It is worthwhile to notice that the original BLO model given in Eq. (2) is not
clear in case of the multiple LL optimal solutions for some of the selections
of the UL decision maker [1]. Therefore, it is necessary to define, which
solution out of the multiple LL solutions in $\mathcal{S}(\mathbf{x})$ should
be considered. Here we actually consider three categories of viewpoints, i.e.,
singleton, optimistic and pessimistic BLOs.
The most straightforward idea in existing learning and vision literature is to
assume that $\mathcal{S}(\mathbf{x})$ is a singleton. Formally, we call the
BLO model is with the Lower-Level Singleton (LLS) condition if
$\forall\mathbf{x}\in\mathcal{X}$, the solution set of the LL subproblem
(i.e., $\mathcal{S}(\mathbf{x})$) is a singleton. In this case, we can
simplify the original model as
$\min\limits_{\mathbf{x}\in\mathcal{X}}F(\mathbf{x},\mathbf{y}),\ s.t.\
\mathbf{y}=\arg\min\limits_{\mathbf{y}\in\mathcal{Y}}f(\mathbf{x},\mathbf{y}).$
(3)
Such singleton version of BLOs is well-defined and could cover a variety of
learning and vision tasks (e.g., [7, 71, 74, 9], just name a few). Thus, in
recent years, dozens of methods have been developed to address this nested
optimization task in different application scenarios (see the following
sections for more details).
Furthermore, the situation becomes more intricate if the LL subproblem is not
uniquely solvable for each $\mathbf{x}\in\mathcal{X}$. Essentially, if the
follower can be motivated to select an optimal solution in
$\mathcal{S}(\mathbf{x})$ that is also best for the leader (i.e., with respect
to $F$), it yields the so-called optimistic (strong) formulation of BLO
$\min\limits_{\mathbf{x}\in\mathcal{X}}\left\\{\min\limits_{\mathbf{y}\in\mathcal{Y}}F(\mathbf{x},\mathbf{y}),\
s.t.\
\mathbf{y}\in\arg\min\limits_{\mathbf{y}\in\mathcal{Y}}f(\mathbf{x},\mathbf{y})\right\\}.$
(4)
The above stated optimistic viewpoint has drawn increasing attention in BLO
literature [192, 193, 194] and recently also been investigated in learning and
vision fields [32, 189, 195]. In Section 7, we will further explore how to
solve such optimistic BLOs in detail.
If the leader does not have the information whether the follower returns the
best response $\mathbf{y}$ from $\mathcal{S}(\mathbf{x})$ in terms of the UL
objective $F$, then we have to assume that the follower is not cooperate with
the leader. This is known as the pessimistic (weak) formulation of BLO [196,
197] and can be given as:
$\min\limits_{\mathbf{x}\in\mathcal{X}}\left\\{\max\limits_{\mathbf{y}\in\mathcal{Y}}F(\mathbf{x},\mathbf{y}),\
s.t.\
\mathbf{y}\in\arg\min\limits_{\mathbf{y}\in\mathcal{Y}}f(\mathbf{x},\mathbf{y})\right\\}.$
(5)
It should be pointed out that till now we still lack efficient gradient-based
algorithms to address the pessimistic BLO problems222In Section 9, we will
demonstrate that we can also obtain some practical gradient-based iteration
scheme within our general algorithmic platform for the pessimistic formulation
of BLO..
### 4.2 BR-based Single-Level Reformulation
In this work, we consider the optimal solution of the LL subproblem with a
given UL variable $\mathbf{x}$ as the Best-Response (BR) of the follower
(denoted as $\mathbf{y}^{*}(\mathbf{x})$). Then we can interpret BLO as a game
process, in which the leader $\mathbf{x}$ considers what BR of the follower
$\mathbf{y}$ is, i.e., how it will respond once it has observed the quantity
of the leader [198, 1]. Based on the above understanding, we can reformulate
the three different categories of BLOs as a unified single-level optimization
problem.
Specifically, given the UL variable $\mathbf{x}$, we denote the corresponding
BR mapping as $\mathbf{y}^{*}(\mathbf{x})$. In fact, if considering the
singleton BLO, $\mathbf{y}^{*}(\mathbf{x})$ can be directly obtained by the
unique LL solution. While for the optimistic and pessimistic BLOs, we actually
first define their Inner Simple Bi-level (ISB) subproblems (w.r.t.,
$\mathbf{y}$)333It is known that the simple bi-level optimization is just a
specific BLO problem with only one variable [32, 199]. as
$\small\mbox{Optimistic
ISB:}\min\limits_{\mathbf{y}\in\mathcal{S}(\mathbf{x})}F(\mathbf{x},\mathbf{y})\
\mbox{and}\ \mbox{Pessimistic
ISB:}\max\limits_{\mathbf{y}\in\mathcal{S}(\mathbf{x})}F(\mathbf{x},\mathbf{y}).$
(6)
Then by defining the solution set of ISB as
$\widetilde{\mathcal{S}}(\mathbf{x})$, we could consider any
$\mathbf{y}^{*}(\mathbf{x})\in\widetilde{\mathcal{S}}(\mathbf{x})$ as the BR
mapping, because points in $\widetilde{\mathcal{S}}(\mathbf{x})$ all obtain
the minimum/maximum of $F(\mathbf{x},\mathbf{y})$ in
$\mathcal{S}(\mathbf{x})$. Therefore, we can formulate the general BR mapping
for different categories of BLOs as follows:
$\left\\{\begin{array}[]{ll}\mathbf{y}^{*}(\mathbf{x}):=\arg\min\limits_{\mathbf{y}\in\mathcal{Y}}f(\mathbf{x},\mathbf{y}),\quad\mbox{Singleton},\\\
\mathbf{y}^{*}(\mathbf{x})\in\widetilde{\mathcal{S}}(\mathbf{x}):=\left\\{\begin{array}[]{l}\arg\min\limits_{\mathbf{y}\in\mathcal{S}(\mathbf{x})}F(\mathbf{x},\mathbf{y}),\quad\mbox{Optimistic},\\\
\arg\max\limits_{\mathbf{y}\in\mathcal{S}(\mathbf{x})}F(\mathbf{x},\mathbf{y}),\quad\mbox{Pessimistic}.\end{array}\right.\end{array}\right.$
(7)
Based on Eq. (7), we actually obtain the following value-function-based
reformulation (a single-level optimization model) for BLOs stated in Eq. (2),
i.e.,
$\min\limits_{\mathbf{x}\in\mathcal{X}}\varphi(\mathbf{x}):=F(\mathbf{x},\mathbf{y}^{*}(\mathbf{x})),$
(8)
in which $\varphi(\mathbf{x})$ actually can be used to uniformly represent the
UL value-function of $F$ from the singleton, optimistic (i.e.,
$\inf_{\mathbf{y}\in\mathcal{S}(\mathbf{x})}F(\mathbf{x},\mathbf{y})$) and
pessimistic (i.e.,
$\sup_{\mathbf{y}\in\mathcal{S}(\mathbf{x})}F(\mathbf{x},\mathbf{y})$)
viewpoints.
### 4.3 A Unified Platform for Gradient-based BLOs
Moving one step forward, the gradient of $\varphi$ w.r.t. the UL variable
$\mathbf{x}$ can be written as444Please notice that we actually do not
distinguish between the operation of the derivatives and partial derivatives
to simplify our presentation.
$\underbrace{\frac{\partial\varphi(\mathbf{x})}{\partial\mathbf{x}}}_{\text{grad.
of $\mathbf{x}$}}=\underbrace{\frac{\partial
F(\mathbf{x},\mathbf{y}^{*}(\mathbf{x}))}{\partial\mathbf{x}}}_{\text{direct
grad. of
$\mathbf{x}$}}\quad+\underbrace{\mathbf{G}(\mathbf{x}),}_{\text{indirect grad.
of $\mathbf{x}$}}$ (9)
where the indirect gradient $G(\mathbf{x})$ can be further specified as the
following two components:
$\mathbf{G}(\mathbf{x})=\underbrace{\overbrace{\left(\frac{\partial\mathbf{y}^{*}(\mathbf{x})}{\partial\mathbf{x}^{\prime}}\right)^{\prime}}^{\text{BR
Jacobian}}\overbrace{\frac{\partial
F(\mathbf{x},\mathbf{y}^{*}(\mathbf{x}))}{\partial\mathbf{y}}.}^{\text{direct
grad. of $\mathbf{y}$}}}_{\text{indirect grad. of $\mathbf{x}$}}$ (10)
Here we use “grad.” as the abbreviation of gradient and denote the transpose
operation as $(\cdot)^{\prime}$. Note that, $\mathbf{y}^{*}(\mathbf{x})$ as a
general mapping, can be given specific constraints and necessary assumptions
to fit their particular requirements for these specific gradient-based BLO
approaches in order to obtain different iteration formats and theoretical
properties. For details, please refer to the following contents. In fact, by
simple computation, the direct gradient is easy to obtain. However, the
indirect gradient is intractable to obtain because we must compute the
changing rate of the optimal LL solution with respect to the UL variable
(i.e., the BR Jacobian
$\frac{\partial\mathbf{y}^{*}(\mathbf{x})}{\partial\mathbf{x}}$). Please
notice that we will also call
$\frac{\partial\varphi(\mathbf{x}^{k})}{\partial\mathbf{x}^{k}}$ as the
practical BR Jacobian w.r.t. $\mathbf{x}^{k}$ in the following statements. The
computation of the indirect gradient $\mathbf{G}(\mathbf{x})$ naturally
motives formulating $\mathbf{y}^{*}(\mathbf{x})$ and hence
$\frac{\partial\mathbf{y}^{*}(\mathbf{x})}{\partial\mathbf{x}}$. For this
purpose, a series of techniques have recently been developed from either
explicit or implicit perspectives, which obtain their optimal solutions by
recurrent differentiation through dynamic system and based on implicit
differentiation theory, respectively.
Now we demonstrate how to formulate various existing gradient-based BLOs from
a unified algorithmic platform. We first summarize a general BLO updating
scheme in Alg. 1. It can be seen that the key component of this algorithm is
to calculate the BR Jacobian. Then with
$\frac{\partial\varphi(\mathbf{x}^{k})}{\partial\mathbf{x}^{k}}$, we can just
perform standard (stochastic) gradient descent schemes to update
$\mathbf{x}^{k}$. Based upon our general algorithmic platform, we can observe
that the main differences of these existing BLO approaches are just their
specific strategies for calculating Jacobian of the BR mapping under different
conditions (i.e., w/ LLS and w/o LLS).
Algorithm 1 A General Gradient-based BLO Scheme
0: The UL and LL initialization.
0: The optimal UL and LL solutions.
1: for $k=1,\cdots,K$ do
2: Calculate the BR Jacobian
$\frac{\partial\varphi(\mathbf{x}^{k})}{\partial\mathbf{x}^{k}}$. %
(Mainstream calculation strategies are summarized in Figs. 8-9 and thoroughly
surveyed in the following sections)
3: Perform (stochastic) gradient descent to update $\mathbf{x}^{k}$. % (based
on $\frac{\partial\varphi(\mathbf{x}^{k})}{\partial\mathbf{x}^{k}}$)
4: end for
In Fig. 8, we summarize mainstream gradient-based BLOs and illustrate their
intrinsic relationships within our general algorithmic platform. It can be
observed that in the LLS scenario, from the BR-based perspective, existing
gradient methods can be categorized as two groups: Explicit Gradient for Best-
Response (EGBR, stated in Section 5) and Implicit Gradient for Best-Response
(IGBR, stated in Section 6). As for EGBR, there are mainly three types of
methods, namely, recurrence-based EGBR (e.g., [51, 71, 12, 13, 14]),
initialization-based EGBR (e.g., [90, 94] ) and proxy-based EGBR methods
(e.g., [82, 171, 87, 78]), differing from each other in the way of formulating
the BR mapping. For IGBR, existing works consider two groups of techniques
(e.g., linear system [74, 85] and Neumann series [72]) to alleviate the
computational complexity issue for the BR Jacobian. We emphasize that the
validity of above BLO methodologies must depend on the singleton of their LL
solution set. When solving BLOs without the LLS assumption, recent works in
[32, 189] have demonstrated that we need to first construct BR mapping based
on both UL and LL subproblems, and then solve two optimization subproblems,
namely, the single-level optimization subproblem (w.r.t. $\mathbf{x}$) and the
ISB subproblem (w.r.t. $\mathbf{y}$). While the work in [195] has introduced a
series of barrier functions and utilized interior point methods to obtain the
BR mapping for each given $\mathbf{x}$.
To end up this section, we plot Fig. 9 to illustrate the optimization
processes of existing mainstream gradient-based BLO methods from the BR
mapping perspective and within our unified algorithmic platform. In the
following (i.e., Sections 5-7), we will thoroughly survey these different
categories of gradient-based BLO algorithms (including their acceleration,
simplification and extension techniques) and their theoretical properties
(convergence behaviors and computational complexity), accordingly.
---
Figure 8: Summary of the mainstream gradient-based BLOs. We categorize these
existing approaches into two main groups, i.e., w/ and w/o LLS assumptions.
When solving BLOs with LLS assumption, these methods can be further divided
into two categories: EGBR and IGBR. As for EGBRs, they can be solved by
different Automatic Differentiation (AD) techniques (as denoted by the dashed
rectangle). Very recently, two algorithms have also been proposed to address
BLOs without the LLS assumption. In particular, they actually introduce a bi-
level gradient aggregation or a value-function-based interior-point method to
calculate the indirect gradient.
---
Figure 9: Illustrating the roadmap of different categories of gradient-based
BLOs. In the left bottom region, the formulations in the solid rectangles
(i.e., singleton and optimistic) have been widely studied. In contrast, since
gradient-based methods for pessimistic BLOs have not been properly
investigated in existing literature, we denote this category of formulation by
a dashed rectangle. In Section 9, we demonstrate that we can also obtain a
practical pessimistic BLO scheme within our general algorithmic platform.
## 5 Explicit Gradient for Best-Response
With the LLS condition, we delve deep into the EGBR category of methods, which
aims to perform automatic differentiation through the LL dynamic system [200,
201] to solve the BLO problem. Specifically, given an initialization
$\mathbf{y}_{0}=\Psi_{0}(\mathbf{x})$ at $t=0$, the iteration process of EGBRs
can be generally written as
$\mathbf{y}_{t}=\Psi_{t}(\mathbf{y}_{t-1};\mathbf{x}),\ t=1,\cdots,T,$ (11)
where $\Psi_{t}$ denotes some given updating scheme (based on the LL
subproblem) at $t$-th stage and $T$ denotes the overall LL iterations number.
For example, we can formulate $\Psi_{t}$ based on the gradient descent rule,
i.e.,
$\Psi_{t}(\mathbf{y}_{t-1};\mathbf{x})=\mathbf{y}_{t-1}-\eta_{t}\mathbf{d}_{f}(\mathbf{y}_{t-1},\mathbf{x}),$
(12)
where $\mathbf{d}_{f}(\mathbf{y}_{t-1},\mathbf{x})$ is the descent mapping of
$f$ at $t$-th stage (e.g.,
$\mathbf{d}_{f}(\mathbf{y}_{t-1},\mathbf{x})=\frac{\partial
f(\mathbf{x},\mathbf{y}_{t-1})}{\partial\mathbf{y}_{t-1}}$) and $\eta_{t}$
denotes the corresponding step size . Then we can calculate
$\frac{\partial\varphi(\mathbf{x}^{k})}{\partial\mathbf{x}^{k}}$ by
substituting $\mathbf{y}_{T}:=\Psi(\mathbf{x})$ approximately for
$\mathbf{y}^{*}(\mathbf{x})$, and the full dynamical system can be defined as
$\Psi(\mathbf{x}):=\Psi_{T}\circ\cdots\circ\Psi_{1}\circ\Psi_{0}(\mathbf{x}).$
(13)
Here the notation $\circ$ represents the compound dynamical operation of the
entire iteration. That is, we actually consider the following optimization
model
$\min\limits_{\mathbf{x}\in\mathcal{X}}\varphi_{T}(\mathbf{x}):=F(\mathbf{x},\mathbf{y}_{T}(\mathbf{x})),$
(14)
and need to calculate
$\frac{\partial\varphi_{T}(\mathbf{x})}{\partial\mathbf{x}}$ (instead of Eq.
(9)) in the practical optimization scenario. Since it should be noted that
$\Psi$ actually obtains an explicit gradient for best-response of the
follower, we call this category of gradient-based BLOs as EGBR approaches
hereafter. Starting from the Eq. (11), it is obvious to notice that
$\mathbf{y}_{t}$ may be affected coupling with the variable $\mathbf{x}$
throughout the iteration. This coupling relationship will have a direct impact
on the optimization process of UL variable in Eq. (9). In fact, existing EGBR
algorithms can be summarized from three perspectives. The first is that, if
$\mathbf{x}$ closely acts on $\mathbf{y}_{t}$ during the whole iteration
process, the subsequent optimization of variable $\mathbf{x}$ will be carried
out recursively. The second is that when $\mathbf{x}$ only acts in the initial
step, the subsequent optimization of variable $\mathbf{x}$ will be simplified.
The third class is to replace the whole iterative process with a hyper-
network, so as to efficiently approximate the BR mapping. Ultimately, in such
cases, we divide them into three categories in terms of the coupling
dependence of the two variables and the solution procedures, namely
recurrence-based EGBR (stated in Section 5.1), initialization-based EGBR
(stated in Section 5.2) and proxy-based EGBR (stated in Section 5.3).
### 5.1 Recurrence-based EGBR
It can be seen from Eq. (11) that all the LL iterative variables
$\mathbf{y}_{0},\mathbf{y}_{1},\cdots,\mathbf{y}_{T}$ depend on $\mathbf{x}$,
and $\mathbf{x}$ acts as a recurrent variable of the dynamical system. One of
the most well-known approaches for calculating
$\frac{\partial\varphi_{T}(\mathbf{x})}{\partial\mathbf{x}}$ (with the above
recurrent structure) is Automatic Differentiation (AD) [160, 202], which is
also called algorithmic differentiation or simply “AutoDiff”. There exist two
diametrically opposite ways on computing gradients for recurrent neural
networks, of which one corresponds to back-propagation through time in a
reverse-mode way [203, 204], and the other corresponds to real-time recurrent
learning in a forward-mode way [205, 206]. Quite a number of methods, closely
related to this subject, have been proposed since then [71, 51, 12, 13]. Here
we would like to review recurrence-based BR methods, covering forward-mode,
reverse-mode AD, truncated and one-stage simplifications.
Forward-mode AD (FAD): To compute
$\frac{\partial\varphi_{T}(\mathbf{x})}{\partial\mathbf{x}}$, FAD appeals to
the chain rule for the derivative of the dynamical system [51]. Specifically,
recalling that $\mathbf{y}_{t}=\Psi_{t}(\mathbf{y}_{t-1},\mathbf{x})$, we have
that the operation $\Psi_{t}$ indeed depends on $\mathbf{x}$ both directly by
its expression and indirectly through $\mathbf{y}_{t-1}$. Hence, by drawing
upon the chain rule, the formulation is given as555Please notice that here we
actually require $\mathbf{y}_{t}(\mathbf{x})$ to be a continuously
differentiable function (w.r.t. $\mathbf{x}$) for all $t=1,\cdots,T$. In
existing EGBRs, they just introduce differentiable $\Psi_{t}$ to meet this
requirement.
$\frac{\partial\mathbf{y}_{t}}{\partial\mathbf{x}}=\frac{\partial\Psi_{t}(\mathbf{y}_{t-1};\mathbf{x})}{\partial\mathbf{y}_{t-1}}\frac{\partial\mathbf{y}_{t-1}}{\partial\mathbf{x}}+\frac{\partial\Psi_{t}(\mathbf{y}_{t-1};\mathbf{x})}{\partial\mathbf{x}}.$
(15)
To simplify the notation, we denote
$\mathbf{Z}_{t}=\frac{\partial\mathbf{y}_{t}}{\partial\mathbf{x}}$,
$\mathbf{A}_{t}=\frac{\partial\Psi_{t}(\mathbf{y}_{t-1};\mathbf{x})}{\partial\mathbf{y}_{t-1}}$,
$\mathbf{B}_{t}=\frac{\partial\Psi_{t}(\mathbf{y}_{t-1};\mathbf{x})}{\partial\mathbf{x}}$
for $t>0$ and
$\mathbf{Z}_{0}=\mathbf{B}_{0}=\frac{\partial\Psi_{0}(\mathbf{x})}{\partial\mathbf{x}}$.
Then we can rewrite Eq. (15) as
$\mathbf{Z}_{t}=\mathbf{A}_{t}\mathbf{Z}_{t-1}+\mathbf{B}_{t}$
$(t=1,\cdots,T)$. In this way, we have the following formulation to
approximate the BR Jacobian
$\frac{\partial\mathbf{y}_{T}(\mathbf{x})}{\partial\mathbf{x}}=\mathbf{Z}_{T}=\sum\limits_{t=0}^{T}\left(\prod\limits_{i=t+1}^{T}\mathbf{A}_{i}\right)\mathbf{B}_{t}.$
(16)
Based on the above derivation, it is apparent that
$\frac{\partial\varphi_{T}(\mathbf{x})}{\partial\mathbf{x}}$ can be computed
by an iterative algorithm summarized in Alg. 2. Actually, FAD allows the
program to update parameters after each step, which may significantly speed up
the dynamic iterator and take up less memory resources when the number of
hyper-parameters is much smaller than the number of parameters. It can be
time-prohibitive for many hyper-parameters with a more efficient and
convenient way.
Algorithm 2 Forward-mode AD (FAD)
0: The UL variable at the current stage $\mathbf{x}$ and the LL initialization
$\mathbf{y}_{0}$.
0: The gradient of $\varphi_{T}$ with respect to $\mathbf{x}$, i.e.,
$\frac{\partial\varphi_{T}}{\partial\mathbf{x}}$.
1: $\mathbf{Z}_{0}=\frac{\partial\Psi_{0}(\mathbf{x})}{\partial\mathbf{x}}$.
2: for $t=1,\cdots,T$ do
3: $\mathbf{y}_{t}=\Psi_{t}(\mathbf{y}_{t-1};\mathbf{x})$.
4: $\mathbf{Z}_{t}=\mathbf{A}_{t}\mathbf{Z}_{t-1}+\mathbf{B}_{t}$.
5: end for
6: return $\frac{\partial
F(\mathbf{x},\mathbf{y}_{T})}{\partial\mathbf{x}}+\mathbf{Z}_{T}^{\prime}\frac{\partial
F(\mathbf{x},\mathbf{y}_{T})}{\partial\mathbf{y}_{T}}$.
Reverse-mode AD (RAD): RAD is a generalization of the back-propagation
algorithm and based on a Lagrangian formulation associated with the parameter
optimization dynamics. By replacing $\mathbf{y}^{*}(\mathbf{x})$ by
$\mathbf{y}_{T}$ and incorporating Eq. (16) into Eq. (9), a series of RAD
works (e.g., [71, 12, 51]) derived
$\frac{\partial\varphi_{T}(\mathbf{x})}{\partial\mathbf{x}}=\frac{\partial
F(\mathbf{x},\mathbf{y}_{T})}{\partial\mathbf{x}}+\mathbf{Z}_{T}^{\prime}\frac{\partial
F(\mathbf{x},\mathbf{y}_{T})}{\partial\mathbf{y}_{T}}.$ (17)
Rather than calculating $\mathbf{Z}_{T}$ by forward propagation as that in FAD
(i.e., Alg. 2), the computation of Eq. (17) can also be implemented by back-
propagation. That is, we first define $\mathbf{g}_{T}=\frac{\partial
F(\mathbf{x},\mathbf{y}_{T})}{\partial\mathbf{x}}$ and
$\bm{\lambda}_{T}=\frac{\partial
F(\mathbf{x},\mathbf{y}_{T})}{\partial\mathbf{y}_{T}}$. Then we update
$\mathbf{g}_{t-1}=\mathbf{g}_{t}+\mathbf{B}_{t}^{\prime}\bm{\lambda}_{t}$, and
$\bm{\lambda}_{t-1}=\mathbf{A}_{t}^{\prime}\bm{\lambda}_{t}$, with
$t=T,\cdots,0$. Finally, we have that
$\frac{\partial\varphi_{T}(\mathbf{x})}{\partial\mathbf{x}}=\mathbf{g}_{-1}$.
Indeed, the above RAD calculation is structurally identical to back-
propagation through time [51]. Moreover, we can also derive it following the
classical Lagrangian approach. That is, we reformulate Eq. (14) as the
following constrained model
$\min\limits_{\mathbf{x}\in\mathcal{X}}\varphi_{T}(\mathbf{x})\ \ s.t.\ \
\left\\{\begin{array}[]{l}\mathbf{y}_{0}=\Psi_{0}(\mathbf{x}),\\\
\mathbf{y}_{t}=\Psi_{t}(\mathbf{y}_{t-1};\mathbf{x}),\
t=1,\cdots,T.\end{array}\right.$ (18)
The corresponding Lagrangian function can be written as
$\begin{array}[]{l}\mathcal{L}(\mathbf{x},\\{\mathbf{y}_{t}\\},\\{\bm{\lambda}_{t}\\})=\varphi_{T}(\mathbf{x})+\bm{\lambda}_{0}^{\prime}\left(\Psi_{0}(\mathbf{x})-\mathbf{y}_{0}\right)\\\
+\sum\limits_{t=1}^{T}\bm{\lambda}_{t}^{\prime}\left(\Psi_{t}(\mathbf{y}_{t-1};\mathbf{x})-\mathbf{y}_{t}\right),\end{array}$
(19)
where $\bm{\lambda}_{t}$ denotes the Lagrange multiplier associated with the
$t$-th stage of the dynamic system. The KKT optimality condition of Eq. (18)
is obtained by setting all derivatives of $\mathcal{L}$ to zero, satisfying
the condition that $\mathbf{y}_{t}(\mathbf{x})$ is a continuously
differentiable function w.r.t. $\mathbf{x}$ for the case that $t=1,\cdots,T$.
Then by some simple algebras, we have
$\frac{\partial\varphi_{T}(\mathbf{x})}{\partial\mathbf{x}}=\frac{\partial\mathcal{L}}{\partial\mathbf{x}}$.
Overall, we present the RAD algorithm in Alg. 3.
Algorithm 3 Reverse-mode AD (RAD)
0: The UL variable at the current stage $\mathbf{x}$ and the LL initialization
$\mathbf{y}_{0}$.
0: The gradient of $\varphi_{T}$ with respect to $\mathbf{x}$, i.e.,
$\frac{\partial\varphi_{T}}{\partial\mathbf{x}}$.
1: $\mathbf{y}_{0}=\Psi_{0}(\mathbf{x})$.
2: for $t=1,\cdots,T$ do
3: $\mathbf{y}_{t}=\Psi_{t}(\mathbf{y}_{t-1};\mathbf{x})$.
4: end for
5: $\mathbf{g}_{T}=\frac{\partial
F(\mathbf{x},\mathbf{y}_{T})}{\partial\mathbf{x}}$ and
$\bm{\lambda}_{T}=\frac{\partial
F(\mathbf{x},\mathbf{y}_{T})}{\partial\mathbf{y}_{T}}$.
6: for $t=T,\cdots,0$ do
7: $\mathbf{g}_{t-1}=\mathbf{g}_{t}+\mathbf{B}_{t}^{\prime}\bm{\lambda}_{t}$
and $\bm{\lambda}_{t-1}=\mathbf{A}_{t}^{\prime}\bm{\lambda}_{t}$.
8: end for
9: return $\mathbf{g}_{-1}$.
Truncated RAD (TRAD): The above two precise calculation methods in many
practical applications are tedious and time-consuming with full back-
propagation training. As aforementioned, due to the complicated long-term
dependencies of the UL subproblem on $\mathbf{y}_{T}(\mathbf{x})$, calculating
Eq. (17) in RAD is a challenging task. This difficulty is further aggravated
when both $\mathbf{x}$ and $\mathbf{y}$ are high-dimensional vectors. More
recently, the truncation idea has been revisited to address the above issue
and shows competitive performance with significantly less computation time and
memory [207, 208, 13]. Specifically, by ignoring the long-term dependencies
and approximating Eq. (17) with partial sums (i.e., storing only the last $M$
iterations), we have
$\frac{\partial\varphi_{T}(\mathbf{x})}{\partial\mathbf{x}}\approx\mathbf{g}_{T-M}:=\frac{\partial
F(\mathbf{x},\mathbf{y}_{T})}{\partial\mathbf{x}}+\mathbf{Z}_{T-M}^{\prime}\frac{\partial
F(\mathbf{x},\mathbf{y}_{T}(\mathbf{x}))}{\partial\mathbf{y}_{T}},$ (20)
where
$\mathbf{Z}_{T-M}=\sum_{t=T-M+1}^{T}\left(\prod_{i=t+1}^{T}\mathbf{A}_{i}\right)\mathbf{B}_{t}$.
It can be seen that ignoring the long-term dependencies can greatly reduce the
time and space complexity for computing the approximate gradients. Recently,
the work in [13] has investigated the theoretical properties of the above
truncated RAD scheme, and confirmed this fact that using few-step back-
propagation could perform comparably to optimization with the exact gradient,
while requiring far less memory and half computation time.
One-stage RAD: Limited and expensive memory is often a bottleneck in modern
massive-scale deep learning applications. For instance, multi-step iteration
of the inner program will cause a lot of memory consumption [89]. Inspired by
BLO, a variety of simplified and elegant techniques have been adopted to
circumvent this issue. The work in [14] proposes another simplification of
RAD, which considers a fixed initialization $\mathbf{y}_{0}$ and only performs
one-step iteration in Eq. (11) to remove the recurrent structure for the
gradient computation in Eq. (17), i.e.,
$\frac{\partial\varphi_{1}(\mathbf{x})}{\partial\mathbf{x}}=\frac{\partial
F(\mathbf{x},\mathbf{y}_{1}(\mathbf{x}))}{\partial\mathbf{x}}+\left(\frac{\partial\mathbf{y}_{1}(\mathbf{x})}{\partial\mathbf{x}^{\prime}}\right)^{\prime}\frac{\partial
F(\mathbf{x},\mathbf{y}_{1}(\mathbf{x}))}{\partial\mathbf{y}_{1}(\mathbf{x})}.$
(21)
By formulating the dynamical system as that in Eq. (12), we then write
$\frac{\partial\mathbf{y}_{1}(\mathbf{x})}{\partial\mathbf{x}}$ as
$\frac{\partial\mathbf{y}_{1}}{\partial\mathbf{x}^{\prime}}=\frac{\partial\left(\mathbf{y}_{0}-\frac{\partial
f(\mathbf{x},\mathbf{y}_{0})}{\partial\mathbf{y}_{0}}\right)}{\partial\mathbf{x}^{\prime}}=-\frac{\partial^{2}f(\mathbf{x},\mathbf{y}_{0})}{\partial\mathbf{y}_{0}\partial\mathbf{x}^{\prime}}.$
(22)
Since calculating Hessian in Eq. (22) is still time consuming, to further
simplify the calculation, we can adopt finite approximation [14] to cancel the
calculation of the Hessian matrix (e.g., central difference approximation).
The specific derivation can be formalized as follows:
$\frac{\partial
F(\mathbf{x},\mathbf{y}_{1})}{\partial\mathbf{y}_{1}}\frac{\partial^{2}f(\mathbf{x},\mathbf{y}_{0})}{\partial\mathbf{y}_{0}\partial\mathbf{x}^{\prime}}\approx\frac{\frac{\partial
f(\mathbf{x},\mathbf{y}_{0}^{+})}{\partial\mathbf{x}}-\frac{\partial
f(\mathbf{x},\mathbf{y}_{0}^{-})}{\partial\mathbf{x}}}{2\epsilon},$ (23)
in which $\mathbf{y}_{0}^{\pm}=\mathbf{y}_{0}\pm\epsilon\frac{\partial
F(\mathbf{x},\mathbf{y}_{1})}{\partial\mathbf{y}_{1}}$. Note that $\epsilon$
is set to be a small scalar equal to the learning rate [59].
### 5.2 Initialization-based EGBR
The research community has started moving towards the challenging goal of
building general purpose initialization-based optimization systems whose
ability to learn the initial parameters better. Regardless of the recurrent
structure, we need to consider the special setting to analyze a family of
algorithms for learning the initialization parameters, named initialization-
based EGBR methods. In this series, MAML [89] is considered as the most
representative and important work. By making more practical assumptions about
the coupling dependence of two variables, these methods no longer use the full
dynamical system to explicitly and accurately describe the dependency between
$\mathbf{x}$ and $\mathbf{y}$ as discussed above in Eq. (18), but adopt a
further simplified paradigm.
Specifically, by treating the iterative dynamical system with only the first
step that $\mathbf{y}$ is explicitly related to $\mathbf{x}$, this process can
be formulated as
$\min\limits_{\mathbf{x}\in\mathcal{X}}\varphi_{T}(\mathbf{x})\ \ s.t.\ \
\left\\{\begin{array}[]{l}\mathbf{y}_{0}=\Psi_{0}(\mathbf{x}),\\\
\mathbf{y}_{t}=\Psi_{t}(\mathbf{y}_{t-1}),\ t=1,\cdots,T,\end{array}\right.$
(24)
where $\mathbf{x}$ represents the network initialization parameters, and
$\mathbf{y}_{t}$ represents the network parameters after performing some sort
of update. Given initial condition $\Psi_{0}(\mathbf{x})$, then we obtain the
following simplified formula
$\mathbf{y}_{T}=\Psi_{0}(\mathbf{x})-\sum\limits_{t=1}^{T}\mathbf{d}_{f}(\mathbf{y}_{t-1}),$
(25)
where $\mathbf{d}_{f}(\mathbf{y}_{t-1})$ is the descent mapping of $f$ at the
$t$-th stage (e.g., $\mathbf{d}_{f}(\mathbf{y}_{t-1})=\frac{\partial
f(\mathbf{x},\mathbf{y}_{t-1})}{\partial\mathbf{y}_{t-1}}$). Finally, we have
the Jacobian matrix as follows
$\frac{\partial\mathbf{y}_{T}}{\partial\mathbf{x}}=\frac{\partial\left(\Psi_{0}(\mathbf{x})-\sum\limits_{t=1}^{T}\mathbf{d}_{f}(\mathbf{y}_{t-1})\right)}{\partial\mathbf{x}}.$
(26)
Then we have to calculate the Hessian matrix term
$\frac{\partial^{2}f}{\partial\mathbf{y}_{t-1}\partial\mathbf{x}^{\prime}}$,
which is time consuming in real computation scenario. To reduce the
computational load, we will introduce two remarkably simple algorithms via a
series of approximate transformation operations below. Among various schemes
to simplify the algorithm based on initialization-based EGBR approaches,
first-order approximation (e.g., [90, 94]) and layer-wise transformation
(e.g., [82, 171, 87, 78]) are among the more popular. Very recently, the works
in [209, 210] also consider the initialization as an auxiliary variable to
improve the performance of RAD.
First-order Approximation: For example, the most representative algorithms
(i.e., FOMAML [90] and Reptile [94]) adopted the operation by first-order
approximation, a way to alleviate the problem of Hessian term computation
while not sacrificing much performance. Specifically, this approximation
ignores the second derivative term by removing the Hessian matrix
$\frac{\partial^{2}f}{\partial\mathbf{y}_{t-1}\partial\mathbf{x}^{\prime}}$,
and then simplifies substitution of
$\frac{\partial\varphi_{T}(\mathbf{x})}{\partial\mathbf{x}}$ performed by
$\frac{\partial\varphi_{T}(\mathbf{x})}{\partial\mathbf{x}}=\frac{\partial
F(\mathbf{x},\mathbf{y}_{T}(\mathbf{x}))}{\partial\mathbf{x}}+\left(\frac{\partial\Psi_{0}(\mathbf{x})}{\partial\mathbf{x}^{\prime}}\right)^{\prime}\frac{\partial
F(\mathbf{x},\mathbf{y}_{T}(\mathbf{x}))}{\partial\mathbf{y}_{T}(\mathbf{x})}.$
(27)
In addition, there is another way of first-order extension to simplify Eq.
(26) through the operation of difference approximation [94]. It no longer
avoids the Hessian term but tries another soft way to approximate
$\frac{\partial\mathbf{y}_{T}}{\partial\mathbf{x}}$
(i.e.,$\mathbf{y}_{T}-\mathbf{x}$ and $(\mathbf{y}_{T}-\mathbf{x})/\alpha$),
in which $\alpha$ is the step size used in gradient decent operation. Unlike
[90], this method proposed to use different linear combinations of all steps
rather than using just the final step. But overall, the above algorithm could
significantly reduce the computing costs while keeping roughly equivalent
performance.
Layer-wise Transformation: Indeed, there are also a series of learning-based
BLOs related to layer-wise transformation, i.e., Meta-SGD [82], T-Net [171],
Meta-Curvature [87] and WarpGrad [78]. In addition to initial parameters, this
type of work focuses on learning some additional parameters (or
transformation) at each layer of the network. From the above Eq. (25), it can
be uniformly formulated as
$\mathbf{y}_{T}=\Psi_{0}(\mathbf{x})-\sum\limits_{t=1}^{T}\mathbf{P}(\mathbf{y}_{t-1},\bm{\omega})\mathbf{d}_{f}(\mathbf{y}_{t-1}),$
(28)
where $\mathbf{P}(\mathbf{y}_{t-1},\bm{\omega})$ defines the matrix
transformation learned at each layer and $\bm{\omega}$ is an auxiliary vector
(e.g., learning rate). For example, Meta-SGD [82] learns a vector
$\bm{\omega}$ of learning rates and $\mathbf{P}$ corresponded to
$\mathtt{diag}(\bm{\omega})$, and T-Net [171] aims to learn block-diagonal
preconditioning linear projections. Similarly, an additional the block-
diagonal preconditioning transformation is also performed by Meta-Curvature
[87]. WarpGrad [78] is closely related to the concurrent work of Meta-
Curvature [87], which defines the preconditions gradient from a geometrical
point of view and replaces the linear projection with a non-linear
preconditioning matrix as a warp layer.
### 5.3 Proxy-based EGBR
Generally speaking, calculating the BR mapping (or BR Jacobian) is key to
solve BLOs. Recently, several proxy-based EGBR methods (e.g., [9, 77, 164])
utilize the differentiable hyper-network (denoted as
$\Psi_{\bm{\theta}}(\mathbf{x})$ with parameters $\bm{\theta}$) to substitute
the dynamic system $\Psi(\mathbf{x})$ and then approximate the BR
mapping666Note that, these methods assume $\mathbf{y}^{*}(\mathbf{x})$ is a
continuously differentiable function and $\mathcal{X}$ and $\mathcal{Y}$
denote the whole space [9, 77, 164]., i.e.,
$\Psi_{\bm{\theta}}(\mathbf{x})\to\Psi(\mathbf{x})\approx\mathbf{y}^{*}(\mathbf{x}).$
(29)
Specifically, they train a hyper-network that takes hyper-parameters
$\mathbf{x}$ as input and outputs the approximate optimal set of weights as
the optimal solution of the LL subproblem.
In fact, both global and local proxy techniques have been considered to
approximate the BR mapping. From the perspective of global approximation,
first, if the distribution $p(\mathbf{x})\subseteq\mathcal{X}$ is fixed, they
learn $\bm{\theta}$ by minimizing $\mathbb{E}_{\mathbf{x}\sim
p(\mathbf{x})}f(\mathbf{x},\Psi_{\bm{\theta}}(\mathbf{x}))$, so that
$\Psi_{\bm{\theta}}(\mathbf{x})$ can approximate the BR mapping in a
neighborhood around the current $\mathbf{x}$, and second update $\mathbf{x}$
with $\Psi_{\bm{\theta}}$ as a proxy substituted into Eq. (14), i.e.,
$\mathbf{x}^{*}\approx\arg\min_{\mathbf{x}\in\mathcal{X}}F(\mathbf{x},\Psi_{\bm{\theta}}(\mathbf{x})).$
(30)
For local approximation, by introducing a small UL disturbing term, they first
minimize the objective $\mathbb{E}_{\epsilon\sim
p(\epsilon|\delta)}f(\mathbf{x}+\epsilon,\Psi_{\bm{\theta}}(\mathbf{x}+\epsilon))$,
where $\epsilon$ represents the perturbation noise added to $\mathbf{x}$, and
$p(\epsilon|\delta)$ is defined as a factorized Gaussian noise distribution
with a fixed scale parameter $\delta$. After that, the UL variable
$\mathbf{x}$ is updated by minimizing the proxy function, i.e., Eq. (30).
In comparison to other type EGBRs, proxy-based EGBRs can easily replace
existing modules in deep learning libraries with hyper-counterparts that
accept an additional vector of UL variable as input and adapt online, thereby
requiring less memory consumption to meet the performance requirements.
## 6 Implicit Gradient for Best-Response
In contrast to the EGBR methods surveyed above, IGBR methods in essence can be
interpreted as introducing Implicit Function Theory (IFT) to derive BR
Jacobian [211]. In particular, IGBR type BLOs only rely on the solution to the
LL optimization and can effectively decouple the UL gradient computation from
the choice of LL optimizer. Indeed, the gradient-based BLO methodologies with
implicit differentiation are radically different from EGBR methods, which have
been extensively applied in a string of applications (e.g., [72, 85, 191]). As
an example, a set of early IGBR approaches (e.g., [161, 162]) used implicit
differentiation to select hyper-parameters of kernel-based models. Recently,
IGBR type approaches have been applied in different application scenarios,
such as learning hyper-parameter for neural networks [72] and variational
models [136].
Now we demonstrate how to derive IGBRs to solve BLOs. Specifically, in the LLS
optimization scenario, we first require that $f(\mathbf{x},\mathbf{y})$
satisfies the smooth condition (or at least twice continuously differentiable)
w.t.r. both the UL and LL variables, and $\mathbf{y}^{*}(\mathbf{x})$ is a
continuously differentiable function w.r.t. $\mathbf{x}$. Then we can directly
obtain the implicit gradient of $\mathbf{x}$ (i.e., $\mathbf{G}(\mathbf{x})$)
based on the first-order optimality condition (i.e., $\frac{\partial
f(\mathbf{x},\mathbf{y}^{*}(\mathbf{x}))}{\partial\mathbf{y}^{*}(\mathbf{x})}=0$).
That is, by deriving the above equation w.r.t. $\mathbf{x}$, we have that
$\frac{\partial\mathbf{y}^{*}(\mathbf{x})}{\partial\mathbf{x}^{\prime}}+\left(\frac{\partial^{2}f(\mathbf{x},\mathbf{y}^{*}(\mathbf{x}))}{\partial\mathbf{y}^{*}(\mathbf{x})\partial\mathbf{y}^{*}(\mathbf{x})^{\prime}}\right)^{-1}\frac{\partial^{2}f(\mathbf{x},\mathbf{y}^{*}(\mathbf{x}))}{\partial\mathbf{y}^{*}(\mathbf{x})\partial\mathbf{x}^{\prime}}=0.$
By further assuming that
$\frac{\partial^{2}f(\mathbf{x},\mathbf{y}^{*}(\mathbf{x})}{\partial\mathbf{y}^{*}(\mathbf{x})\partial\mathbf{y}^{*}(\mathbf{x})^{\prime}}$
is invertible, and drawing upon the chain rule, the indirect gradient
$\mathbf{G}(\mathbf{x})$ can be obtained as follows:
$\displaystyle\mathbf{G}(\mathbf{x})=-\left(\frac{\partial^{2}f(\mathbf{x},\mathbf{y}^{*}(\mathbf{x}))}{\partial\mathbf{y}^{*}(\mathbf{x})\partial\mathbf{x}^{\prime}}\right)^{\prime}\left(\frac{\partial^{2}f(\mathbf{x},\mathbf{y}^{*}(\mathbf{x}))}{\partial\mathbf{y}^{*}(\mathbf{x})\partial\mathbf{y}^{*}(\mathbf{x})^{\prime}}\right)^{-1}$
(31) $\displaystyle\frac{\partial
F(\mathbf{x},\mathbf{y}^{*}(\mathbf{x}))}{\partial\mathbf{y}^{*}(\mathbf{x})}$
.
Intuitively, Eq. (31) has offered the exact indirect gradient formulation but
is generally calculated based on numerical approximations in practice. From a
computational point of view, due to involving a large number of repeated
product operations of Hessian-vector and Jacobian-vector, EGBRs based on high-
dimensional data are usually computationally expensive and time-consuming.
Thus a few implicit techniques, such as IGBR based on linear system [74, 85]
and Neumann series [72], have been proposed to address this computational
issue.
Based on Linear System: To calculate the Hessian matrix inverse more
efficiently, it is generally assumed that solving linear systems is a common
operation (e.g., HOAG [74], IMAML [85]). Specially,
$(\frac{\partial^{2}f}{\partial\mathbf{y}\partial\mathbf{y}^{\prime}})^{-1}\frac{\partial
F}{\partial\mathbf{y}}$ can be computed as the solution to the linear system
$(\frac{\partial^{2}f}{\partial\mathbf{y}\partial\mathbf{y}^{\prime}})\mathbf{q}=\frac{\partial
F}{\partial\mathbf{y}}$ for $\mathbf{q}$. Based on the above derivation, it is
apparent that $\frac{\partial F}{\partial\mathbf{x}}$ can be directly computed
by the algorithm summarized in Alg. 4.
Algorithm 4 Implicit Gradient by Solving Linear System
0: The UL variable at the current state, i.e., $\mathbf{x}$
0: The gradient of $F$ with respect to $\mathbf{x}$, i.e., $\frac{\partial
F}{\partial\mathbf{x}}$
1: Optimize the LL variable up to tolerance $\epsilon$. That is, find
${\mathbf{y}_{\varepsilon}}$ such that
$\|\mathbf{y}^{*}(\mathbf{x})-{\mathbf{y}_{\varepsilon}}\|\leq\epsilon.$
2: Solve the linear system
$\left(\frac{\partial^{2}f(\mathbf{x},\mathbf{y}_{\varepsilon})}{\partial\mathbf{y}_{\varepsilon}\partial\mathbf{y}_{\varepsilon}^{\prime}}\right)\mathbf{q}=\frac{\partial
F(\mathbf{x},\mathbf{y}_{\varepsilon})}{\partial\mathbf{y}_{\varepsilon}},$
for $\mathbf{q}$ up to the tolerance $\epsilon$, i.e.,
$\left\|\left(\frac{\partial^{2}f}{\partial\mathbf{y}_{\varepsilon}\partial\mathbf{y}_{\varepsilon}^{\prime}}\right)\mathbf{q}-\frac{\partial
F}{\partial\mathbf{y}_{\varepsilon}}\right\|\leq\epsilon.$
3: Compute approximate gradient by $\mathbf{p}=\frac{\partial
F(\mathbf{x},\mathbf{y}_{\varepsilon})}{\partial\mathbf{x}}-\left(\frac{\partial^{2}f(\mathbf{x},\mathbf{y}_{\varepsilon})}{\partial\mathbf{y}_{\varepsilon}\partial\mathbf{x}^{\prime}}\right)^{\prime}\mathbf{q}.$
4: return $\mathbf{p}$.
Based on Neumann Series: Instead of solving the linear system, another type of
IGBM (i.e., Neumann IFT [72]) method aims to calculate the Neumann series to
approximate the inverse of Hessian matrix. Specifically, rather than solving
the linear system in the second step of Alg. 4, the inverse Hessian is
expressed as the following Neumann series:
$\left(\frac{\partial^{2}f}{\partial\mathbf{y}\partial\mathbf{y}^{\prime}}\right)^{-1}=\lim_{i\to\infty}\sum\limits_{j=0}^{i}\left(\mathbf{I}-\frac{\partial^{2}f}{\partial\mathbf{y}\partial\mathbf{y}^{\prime}}\right)^{j},$
where $\mathbf{I}$ denotes an identity matrix with proper size. If the
operator
$\mathbf{I}-\frac{\partial^{2}f}{\partial\mathbf{y}\partial\mathbf{y}^{\prime}}$
is contractive, it leverages that unrolling differentiation for $i$ steps
around locally optimal weights $\mathbf{y}^{*}(\mathbf{x})$ is equivalent to
approximating the inverse with the first $i$ terms in Neumann series. In this
way, the entire computation can efficiently perform vector-Jacobian products,
thus providing a cheap approximation to the inverse-Hessian-vector product.
## 7 BLO beyond Lower-Level Singleton
As stated in the above Sections 4-6, different categories of gradient-based
algorithms have been proposed to address BLOs. However, most of these
approaches rely on the LLS assumption (i.e., the solution set of the LL
subproblem is a singleton) stated in Section 4 to simplify their optimization
process and theoretical analysis. That is to say, the sequence
$\\{\mathbf{y}_{t}\\}_{t=0}^{T}$ generated by these mainstream methods could
converge to the true optimal solution only if the LLS condition is satisfied.
Unfortunately, it has been demonstrated that such LLS assumption is too
restrictive to be satisfied in most real-world learning and vision
applications. For example, the works in [32, 189] have designed a series of
counter-examples to illustrate that these existing EGBRs cannot obtain the
correct solution if the LLS assumption is not satisfied.
In this section, we review some recent works [32, 189, 195], which can
efficiently address the LLS issue in the optimistic BLO scenario. The key
optimization process of these works is to obtain the solution set of the ISB
(i.e., Eq. (6)). That is, these works actually adopted different techniques,
such as the UL and LL gradient aggregation [32, 189] and value-function-based
interior-point method [195] to solve Eq. (6) for BLOs without the LLS
condition.
UL and LL Gradient Aggregation: Differing from previous EGBR type methods
which only rely on the gradient information of the LL subproblem to update
$\mathbf{y}$, a more generalized EGBR type method, Bi-level Descent
Aggregation (BDA) method [32], characterizes an aggregate computation of both
the LL and the UL descent information. With a given UL variable $\mathbf{x}$,
the aggregated descent direction w.r.t. the ISB subproblem (i.e., Eq. (6)) can
be defined as
$\mathbf{d}(\mathbf{y}_{t-1};\mathbf{x})=\rho_{t}\frac{\partial
F(\mathbf{x},\mathbf{y}_{t-1})}{\partial\mathbf{y}_{t-1}}+(1-\rho_{t})\frac{\partial
f(\mathbf{x},\mathbf{y}_{t-1})}{\partial\mathbf{y}_{t-1}},$ (32)
where ${\rho_{t}\in(0,1]}$ is the aggregation parameter (tending to zero [212,
213]), and $\frac{\partial
F(\mathbf{x},\mathbf{y}_{t-1})}{\partial\mathbf{y}_{t-1}}$ (or $\frac{\partial
f(\mathbf{x},\mathbf{y}_{t-1})}{\partial\mathbf{y}_{t-1}}$) stands for the
descent directions of the UL (or LL) objectives.
Value-Function-based Interior-point Method: Different from EGBRs and IGBRs, a
more recent Value-Function Best-Response (VFBR) type BLO methods reformulate
BLO into a ISB optimization problem by the value function of the UL objective.
After that, they further transform it into a single-level optimization problem
with an inequality constraint through the value function of the LL objective.
Recently, a typical VFBR work, named Bi-level Value-Function-based Interior-
point Method (BVFIM) [195], has designed a log-barrier penalty-based single-
level reformulation for Eq. (6) to address the LLS issue in the non-convex
scenario. Specifically, BVFIM first reformulates the ISB subproblem in Eq. (6)
as follows:
$\min\limits_{\mathbf{y}\in\mathcal{Y}}F(\mathbf{x},\mathbf{y}),\ \mathrm{\
s.t.\ }\ f(\mathbf{x},\mathbf{y})\leq\psi_{\mu}(\mathbf{x}),$ (33)
where $\psi_{\mu}(\mathbf{x})$ is a regularized value function of the LL
subproblem, i.e.,
$\psi_{\mu}(\mathbf{x})=\min\limits_{\mathbf{y}\in\mathcal{Y}}f(\mathbf{x},\mathbf{y})+\frac{\mu_{1}}{2}\|\mathbf{y}\|^{2}+\mu_{2}.$
(34)
Here $\mu_{1},\mu_{2}$ are two positive constants and we denote
$\mu=(\mu_{1},\mu_{2})$. Then the relaxed inequality constraint
$f(\mathbf{x},\mathbf{y})\leq\psi_{\mu}(\mathbf{x})$ is penalized to the
objective by a log-barrier penalty and thus Eq. (33) can be approximated by
$\small\varphi_{\mu,\theta,\tau}\left(\mathbf{x}\right)=\min_{\mathbf{y}\in\mathcal{Y}}F(\mathbf{x},\mathbf{y})+\frac{\theta}{2}\|\mathbf{y}\|^{2}-\tau\ln(\psi_{\mu}(\mathbf{x})-f(\mathbf{x},\mathbf{y})),$
(35)
where $(\mu,\theta,\tau)>0$. Finally, BVFIM proves that indirect gradient
$\mathbf{G}(\mathbf{x})$ in Eq. (10) can be obtained by solving a series of
Eq. (35) with decreasing parameters $(\mu,\theta,\tau)$ (tending to zero). It
should be noticed that BVFIM can successfully avoid these time-consuming
Hessian-vector and Jacobian-vector products, which are necessary in previous
gradient-based BLOs. So this method is more suitable for BLO tasks with
complex LL subproblems.
## 8 Theoretical Investigations
In addition to modeling various learning and vision applications from the
perspective of BLO and establishing a general algorithmic framework to unify
different categories of existing BLO algorithms, in this section, we further
investigate some important theoretical issues of BLOs, including the
convergence behaviors and computational complexity of gradient-based BLOs,
which actually can provide us insights and guidance in practical application
scenarios (e.g., adopt/design proper BLO methods).
### 8.1 Convergence Properties and Required Conditions
In existing literature, two categories of convergence properties have been
proved for gradient-based BLOs. The first type is “convergence towards
stationarity”, which guarantees that the UL value-function can converge to a
first-order stationary point satisfying
$\lim_{K\rightarrow\infty}\|\frac{\partial\varphi(\mathbf{x}_{T}^{K})}{\partial\mathbf{x}_{T}^{K}}\|=0$.
Here we actually consider the convergence property of the UL variable, i.e.,
the number of UL iteration $K$ tends to infinity (with fixed number of LL
iteration $T$). The other convergence results actually characterize the
following properties: $\mathbf{x}_{T}\xrightarrow[]{s}\mathbf{x}^{*}$777Here
we use “$\xrightarrow[]{s}$” to denote subsequential convergence. and
$\inf_{\mathbf{x}\in\mathcal{X}}\varphi_{T}(\mathbf{x})\rightarrow\inf_{\mathbf{x}\in\mathcal{X}}\varphi(\mathbf{x})$)
when $T\to\infty$. That is, they prove that for any limit point
$\bar{\mathbf{x}}$ of the sequence $\\{\mathbf{x}_{T}\\}$, if $\mathbf{x}_{T}$
is a global (resp. local) minimum of $\varphi_{T}(\mathbf{x})$, then
$\bar{\mathbf{x}}$ is a global (resp. local) minimum of $\varphi(\mathbf{x})$.
For convenience, we call this type of property as “convergence towards
global/local minimum”888We will provide more details on this convergence
property in the following subsection (i.e., Theorem 1).. In Table III, we
analyze the convergence properties and conditions required by the UL and LL
subproblems for different categories of gradient-based BLOs, including EGBRs
(e.g., RHG [12], TRAD [13], HF-MAML [214], STN [9] and BDA [32]), IGBRs (e.g.,
HOAG [74] and IMAML [85]) and VFBR (e.g., BVFIM [195]).
To guarantee the convergence to stationary solutions, some EGBRs (e.g., TRAD
[13], HF-MAML [214] and STN [9]) required the first-order Lipshitz assumption
for the UL and LL objectives (i.e., “${L}_{F}$” and “${L}_{f}$” for short) and
the twice continuously differentiable property for the LL objective. In
addition, there are also some EGBRs that require additional strong assumptions
to obtain the first-order stationary points. For instance, HF-MAML [214]
relies on second-order Lipshitz assumption (denoted as “Lipschitz-Hessian”)
for the LL objective, while STN [9] needs the nonsingular Hessian assumption
for the LL objective. As for IGBRs (e.g., HOAG [74] and IMAML [85]), they
generally require that the gradient (w.r.t. $\mathbf{y}$) of both the UL
objective and the LL objective are Lipschitz continuous. As another mainstream
EGBR, the work [12] requires that the LL dynamic system
$\\{\mathbf{y}_{T}(\mathbf{x})\\}$ is uniformly bounded on $\mathcal{X}$ and
$\mathbf{y}_{T}(\mathbf{x})$ uniformly converges to
$\mathbf{y}^{*}(\mathbf{x})$ when $T\rightarrow\infty$. Then we can obtain the
convergence towards the global/local minimum. As for IGBRs, both the Lipshitz
Hessian and nonsingular Hessian are key properties to guarantee their
stationarity convergence [74, 214, 85].
TABLE III: Summarizing the convergence results of mainstream gradient-based
methods for BLOs within our framework.
Category | Method | LLS | UL | LL | Main convergence results
---|---|---|---|---|---
${L}_{F}$ | SC | ${L}_{f}$ | $\mathcal{C}^{2}$ | SC | Lip-Hess | NS-Hess
EGBR | TRAD [13] | ✓ | ✗ | ✗ | ✓ | ✓ | ✓ | ✗ | ✓ | Stationarity: $\frac{\partial\varphi(\mathbf{x}_{T}^{K})}{\partial\mathbf{x}_{T}^{K}}\rightarrow 0$.
HF-MAML [214] | ✓ | ✓ | ✗ | ✓ | ✓ | ✗ | ✓ | ✗
STN [9] | ✓ | ✗ | ✗ | ✓ | ✓ | ✓ | ✗ | ✓
RHG [12] | ✓ | ✗ | ✗ | ✗ | ✗ | ✗ | ✗ | ✗ | Global/local minimum: $\mathbf{x}_{T}\xrightarrow[]{s}\mathbf{x}^{*},$ $\inf_{\mathbf{x}\in\mathcal{X}}\varphi_{T}(\mathbf{x})\rightarrow\inf_{\mathbf{x}\in\mathcal{X}}\varphi(\mathbf{x})$.
BDA [32] | ✓ | ✓ | ✗ | ✓ | ✗ | ✗ | ✗ | ✗
✗ | ✓ | ✓ | ✓ | ✗ | ✗ | ✗ | ✗
BDA [189] | ✗ | ✓ | ✗ | ✓ | ✗ | ✗ | ✗ | ✗
VFBR | BVFIM [195] | ✗ | ✗ | ✗ | ✗ | ✗ | ✗ | ✗ | ✗
IGBR | HOAG [74] | ✓ | ✓ | ✗ | ✓ | ✓ | ✓ | ✓ | ✓ | Stationarity: $\frac{\partial\varphi(\mathbf{x}_{T}^{K})}{\partial\mathbf{x}_{T}^{K}}\rightarrow 0$.
IMAML [85] | ✓ | ✓ | ✗ | ✓ | ✓ | ✓ | ✓ | ✓
* 1
Notice that $F(\mathbf{x},\mathbf{y})$ is continuously differentiable on
$\mathcal{X}\times\mathcal{Y}$ ($\mathcal{X}$ is a compact set) and
$f(\mathbf{x},\mathbf{y})$ is continuously differentiable on
$\mathbb{R}^{m}\times\mathcal{Y}$. The feasible solution set $\mathcal{Y}$
represents the whole space $\mathbb{R}^{n}$.
* 2
“${L}_{F}$ (resp. ${L}_{f}$)” means the gradient of $F(\mathbf{x},\cdot)$
(resp. $f(\mathbf{x},\cdot)$) is Lipschitz continuous with Lipschitz constant
${L}_{F}$ (resp. ${L}_{f}$). SC means strongly convex and $\mathcal{C}^{2}$
implies that $f(\mathbf{x},\cdot)$ is second-order continuously differentiable
w.r.t. $\mathbf{y}$. “NS-Hess” and “Lip-Hess” represent the nonsingularity and
Lipschitz properties of Hessian
$\frac{\partial^{2}f}{\partial\mathbf{y}\partial\mathbf{y}^{\prime}}$,
respectively. Please refer to [214, 85, 74] for more details on these
variational analysis concepts.
* 3
Here we respectively represent “required” and “not required” by “✓” and “✗”
for these properties.
* 4
We summarize two kinds of convergent properties, i.e., “stationarity” and
“global/local minimum”. The former implies that the gradient descent on the UL
value-function converges to first-order stationary points satisfying
$\lim_{K\rightarrow\infty}\|\frac{\partial\varphi(\mathbf{x}_{T}^{K})}{\partial\mathbf{x}_{T}^{K}}\|=0$
(with fixed number of LL iterations $T$), while the latter characterizes the
convergence towards global/local minimum satisfying
$\mathbf{x}_{T}\xrightarrow[]{s}\mathbf{x}^{*}$ and
$\inf_{\mathbf{x}\in\mathcal{X}}\varphi_{T}(\mathbf{x})\rightarrow\inf_{\mathbf{x}\in\mathcal{X}}\varphi(\mathbf{x})$
as $T\to\infty$.
### 8.2 A General Proof Template for EGBRs
In this subsection, we would like to further provide a general proof template
to analyze the convergence behaviors (i.e., convergence towards global/local
minimum) of EGBR methods in more detail. In particular, given the output of
the LL dynamic system (i.e., $\mathbf{y}_{T}(\mathbf{x})$), we first introduce
two elementary properties on it as follows:
1. (1)
Uniform approximation quality to the LL solution:
$\\{\mathbf{y}_{T}(\mathbf{x})\\}$ is uniformly bounded on $\mathcal{X}$, and
for any $\epsilon>0$, there exists $t(\epsilon)>0$ such that whenever
$T>t(\epsilon)$, we have
$\sup_{\mathbf{x}\in\mathcal{X}}\left\\{f(\mathbf{x},\mathbf{y}_{T}(\mathbf{x}))-\psi(\mathbf{x})\right\\}\leq\epsilon,$
or
$\sup_{\mathbf{x}\in\mathcal{X}}\|\frac{\partial
f(\mathbf{x},\mathbf{y}_{T}(\mathbf{x}))}{\partial\mathbf{y}_{T}(\mathbf{x})}\|\leq\epsilon,$
where $\psi(\mathbf{x})$ denotes the LL value-function, i.e.,
$\psi(\mathbf{x}):=\min_{\mathbf{y}\in\mathcal{Y}}f(\mathbf{x},\mathbf{y})$.
2. (2)
Point-wise approximation quality to the ISB solution: For each
$\mathbf{x}\in\mathcal{X}$, we have
$\lim_{T\rightarrow\infty}\mathtt{dist}(\mathbf{y}_{T}(\mathbf{x}),\widetilde{\mathcal{S}}(\mathbf{x}))=0,$
where $\widetilde{\mathcal{S}}(\mathbf{x})$ represents the solution set of the
ISB subproblem in Eq. (6) and $\mathtt{dist}(\cdot,\cdot)$ denotes the point-
to-set distance.
Equipped with the above two properties on $\\{\mathbf{y}_{T}(\mathbf{x})\\}$,
we can present general convergence results of Eqs. (11)-(14) in the following
theorem999Here we actually provide a brief proof roadmap, which is summarized
based on theoretical studies in existing works [51, 195, 32, 189]. .
###### Theorem 1.
(Convergence towards global/local minimum) Suppose that the generated sequence
$\left\\{\mathbf{y}_{t}(\mathbf{x})\right\\}$ satisfies the above two
properties. Let $\mathbf{x}_{T}$ be a global (resp. local) minimum of
$\varphi_{T}(\mathbf{x})$, i.e.,
$\mathbf{x}_{T}\in\arg\min_{\mathbf{x}\in\mathcal{X}}\varphi_{T}(\mathbf{x})$.
Then we have
* (1)
Any limit point $\bar{\mathbf{x}}$ of the sequence $\\{\mathbf{x}_{T}\\}$ is a
global (resp. local) minimum of $\varphi(\mathbf{x})$, i.e.,
$\bar{\mathbf{x}}\in\arg\min_{\mathbf{x}\in\mathcal{X}}\varphi(\mathbf{x})$.
* (2)
$\inf_{\mathbf{x}\in\mathcal{X}}\varphi_{T}(\mathbf{x})\rightarrow\inf_{\mathbf{x}\in\mathcal{X}}\varphi(\mathbf{x})$
as $T\rightarrow\infty$.
###### Proof.
In the following, we first state the key steps for proving convergence
properties in the global scenario and then demonstrate how to obtain the local
convergence properties accordingly.
Step 1. We should first verify that for $\bar{\mathbf{x}}\in\mathcal{X}$,
$\psi$ satisfies
$\limsup_{\mathbf{x}\rightarrow\bar{\mathbf{x}}}\psi(\mathbf{x})=\psi(\bar{\mathbf{x}}).$
Step 2. Then for any limit point $\bar{\mathbf{x}}$ of the sequence
$\\{\mathbf{x}_{T}\\}$, there exist
$\mathbf{y}_{m}(\mathbf{x}_{m})\rightarrow\bar{\mathbf{y}}$ for a subsequence
$\\{\mathbf{x}_{m}\\}$ and some $\bar{\mathbf{y}}$. Thus we can obtain
$\bar{\mathbf{y}}\in\mathcal{S}(\bar{\mathbf{x}})$.
Step 3. Next, we verify the convergence property of the UL objective as
follows:
$\lim_{T\rightarrow\infty}\varphi_{T}(\mathbf{x})=\varphi(\mathbf{x}).$
Step 4. For any $\epsilon>0$, we verify the following inequality:
$\varphi(\bar{\mathbf{x}})\leq
F(\mathbf{x}_{m},\mathbf{y}_{m}(\mathbf{x}_{m}))+\epsilon\leq\lim_{m\rightarrow\infty}\varphi_{m}(\mathbf{x})+\epsilon,\
\forall\mathbf{x}\in\mathcal{X}.$
Step 5. Finally, we verify the following inequality
$\limsup_{T\rightarrow\infty}\left\\{\inf_{\mathbf{x}\in\mathcal{X}}\varphi_{T}(\mathbf{x})\right\\}\leq\inf_{\mathbf{x}\in\mathcal{X}}\varphi(\mathbf{x}).$
Thus we can obtain convergence results stated in Theorem 1.
For convergence to the local minimum, we actually consider $\mathbf{x}_{T}$ as
a local minimum of $\varphi_{T}(\mathbf{x})$ with uniform neighborhood modulus
$\delta>0$. Then any limit point $\bar{\mathbf{x}}$ of the sequence
$\\{\mathbf{x}_{T}\\}$ is a local minimum of $\varphi(\mathbf{x})$, i.e.,
there exists $\tilde{\delta}>0$ such that
$\varphi(\bar{\mathbf{x}})\leq\varphi(\mathbf{x}),\forall\mathbf{x}\in\mathbb{B}_{\delta}(\bar{\mathbf{x}})\cap\mathcal{X}$.
According to the neighborhood property to spread out the analysis, the result
of convergence towards local minimum can also be proved by the same steps. ∎
The above theoretical results actually provide us a general recipe to analyze
the iteration behaviors and convergence properties of gradient-based BLOs,
especially for EGBRs. In other words, we can understand that these existing
numerical schemes and their required assumptions on the UL and LL subproblems
are just to meet the above elementary iteration properties.
It can be observed that classical EGBRs (e.g., [12, 13]) require to first
enforce the LLS assumption on the BLO problem. The work in [12] assumes that
the UL and LL objectives are continuously differentiable and also enforces the
restrictive (local) strong convexity assumption on the LL objective. In fact,
such properties can ensure the uniform convergence of
$\\{\mathbf{y}_{T}(\mathbf{x})\\}$ towards $\mathbf{y}^{*}(\mathbf{x})$, thus
lead to the two elementary properties. In fact, the LLS assumption considered
in [12] is more strict than that required in the proof template. The works in
[32, 189] also consider that the UL and LL objectives are continuously
differentiable, but make a weaker assumption on the LL objective, i.e.,
$f(\mathbf{x},\mathbf{y})$ is level-bounded in $\mathbf{y}$ and locally
uniform in $\mathbf{x}\in\mathcal{X}$ (or the gradient of
$f(\mathbf{x},\cdot)$ is Lipschitz continuous). Indeed, it can be verified
that the conditions in [32, 189] can also ensure two elementary properties
required by our proof template. Therefore, we have that the above two
elementary convergence properties hold and we can obtain the convergence
results stated in Theorem 1.
It has been verified in [32, 189] that these classical EGBRs [12, 13] may lead
to incorrect solutions if the LLS assumption is not satisfied. As stated in
the above Section 7, BDA [32, 189] has been proposed to extend the EGBR method
to address this issue. Theoretically, the work in [32] actually introduces the
LL solution set property and the UL objective convergence property.
Theoretical investigations in [189] further demonstrate that the iterative
gradient-aggregation dynamics can solve the ISB subproblem without the LL
singleton assumption and the UL strong convexity. Again, in order to remove
the restrictive singleton and convex assumptions on the LL objective, BVFIM
[195] further proves the same convergence results by introducing the strong
constraints on a series of positive decreasing parameters $(\mu,\theta,\tau)$
when $f(\mathbf{x},\mathbf{y})$ and $F(\mathbf{x},\mathbf{y})$ are level-
bounded in $\mathbf{y}$ and locally uniformly in $\mathbf{x}\in\mathcal{X}$.
### 8.3 Time and Space Complexity
In this subsection, we analyze the complexity of time and space for these
mainstream gradient-based BLO methods (i.e., EGBRs [51, 13, 32], IGBRs [74,
85, 72] and VFBR [195]), as summarized in Table IV. Please notice that here we
just follow most BLO literature (e.g., [51, 13, 14]) to only estimate the
complexity of computing the gradient of $\varphi$ w.r.t. $\mathbf{x}$ (defined
in Eq. (9)) with a fixed (e.g., $T$-step) LL iteration.
EGBR: As discussed in Section 5, EGBRs generally construct the BR mapping
$\mathbf{y}^{*}(\mathbf{x})$ or the indirect gradient $\mathbf{G}(\mathbf{x})$
with the implementation of an unrolled dynamic system (see Eq. (13)). In [51],
the dynamic system can be implemented in either a forward automatic
differential mode (i.e., FAD) or a reverse automatic differential mode (i.e.,
RAD). Especially, BDA implements a reverse aggregated gradient flow from the
UL and LL subproblems to approximate the BR mapping. More specifically, taking
into account the fact that the Hessian-matrix product is repeatedly calculated
(i.e.,
$\sum_{t=0}^{T}\left(\prod_{i=t+1}^{T}\mathbf{A}_{i}\right)\mathbf{B}_{t}$) in
the forward propagation, FAD requires the space complexity $O(mn)$ and the
time complexity $O(m^{2}nT)$. RAD in the backward pass needs to evaluate
Hessian- and Jacobian-vector products, and stores all the intermediate
variables $\\{\mathbf{y}_{t}\in\mathbb{R}^{n}\\}_{t=1}^{T}$ in memory. So we
have that the time and space costs are $O(n(m+n)T)$ and $O(m+nT)$,
respectively. By ignoring the long-term dependencies, TRAD uses the truncated
back-propagation trajectory with a smaller number of steps (i.e., $M<T$). As
for BDA, with the similar backward propagation manner, we have that the
complexity of time and space is the same as that for RAD.
IGBR: As for IGBRs, we have that they require to derive the indirect gradient
based on the implicit function theorem, which results in the overloaded
computation with respect to the inverse of Hessian (see Eq. (31)). To mitigate
this problem, IGBRs generally solve a linear system by Conjugate Gradient (CG)
[74, 85] or Neumann series [72], as stated in Section 6. Without loss of
generality, we uniformly assume that these methods perform $J$-step iterations
to solve the linear system. Each step contains a hessian-vector product
computation requiring the time cost $O(m+n^{2}J)$. Then with a $T$-step
gradient descent on the LL subproblem, we have that the overall time and space
complexities can be written as $O(m+nT+n^{2}J)$ and $O(m+n)$, respectively. It
should be noted that the iteration step $J$ generally relies on the properties
of Hessian-matrix, thus it should be set much larger than $T$.
VFBR: It has been stated in Section 7 that VFBR type method (i.e., BVFIM) does
not require to solve the unrolled dynamic system or approximate the inverse of
Hessian, thus can obtain lower time and space complexity than EGBRs and IGBRs,
especially on BLOs with high-dimensional LL subproblems. Specifically, we use
$Q_{1}$ and $Q_{2}$ to represent the number of gradient iterations for solving
the regularized subproblems in Eqs. (34) and (35), respectively. Then it can
be checked that the time costs of calculating each gradient descent for the LL
and UL value-functions are $O(nQ_{1})$ and $O(nQ_{2})$, respectively.
Moreover, we require additional $O(m)$ time to perform the UL gradient
updating. Thus the overall time cost of BVFIM is $O(m+n(Q_{1}+Q_{2}))$. As for
the space complexity, it is easy to check that BVFIM requires $O(m+n)$ space
cost and is the same as that in IGBRs.
TABLE IV: Comparison of the time and space complexity for several gradient-based mainstream BLOs. Category | Method | Time | Space
---|---|---|---
EGBR | FAD [51] | $O(m^{2}nT)$ | $O(mn)$
RAD [51] | $O(n(m+n)T)$ | $O(m+nT)$
BDA [32] | $O(n(m+n)T)$ | $O(m+nT)$
TRAD [13], | $O(n(m+n)M)$ | $O(m+nM)$
IGBR | CG [74, 85] | $O(m+nT+n^{2}J)$ | $O(m+n)$
Neumann [72] | $O(m+nT+n^{2}J)$ | $O(m+n)$
VFBR | BVFIM [195] | $O(m+n(Q_{1}+Q_{2}))$ | $O(m+n)$
It can be seen in Table IV that the reverse propagation methods (i.e., RAD,
TRAD and BDA) have benefited from the lightweight matrix-vector multiplication
(rather than the overweight Hessian-matrix), thus can obtain less
computational complexity in comparison to the forward propagation approach
(e.g., FAD). Especially for TRAD, the time and space complexity can be further
reduced by the truncated back-propagation strategy. Compared with EGBRs, IGBRs
maintain higher computational complexity due to the overloaded computation in
terms of the inverse of Hessian. In contrast, VFBR can obtain lower time
consuming than both EGBRs and IGBRs. It actually also outperforms EGBRs in
costing less memory, especially when solving the LL subproblem on high-
dimensional tasks (e.g., with extremely large $n$).
## 9 Potentials for New Algorithms Design
As the last but not least part of the survey, this section aims to demonstrate
the potentials of our general algorithmic framework for designing new gradient
schemes for challenging BLO formulations, such as pessimistic BLOs (stated in
Eq. (5)).
In fact, pessimistic BLO formulation can be naturally interpreted as a non-
cooperative game between two players and has been utilized to formulate
problems in the area of mathematical programming [215, 216, 217] and other
application fields, such as economics [218, 219] and biology [220]. However,
from the pessimistic viewpoint, the UL player (i.e., leader) cannot anticipate
the LL player (i.e., follower)’s decision, the constraint must be satisfied
for any rational decision of the follower, thus pessimistic BLO is perceived
to be very difficult to solve, especially in high-dimensional application
scenarios [196].
Now we demonstrate how to develop a practical algorithm within our BR mapping
based BLO algorithmic framework for pessimistic BLO formulations101010We
emphasize that we just present an example to demonstrate the potentials of our
framework for new algorithm design. Strict theoretical analysis and
evaluations are definitely out of the scope in this paper and will be
considered as the future work.. Concretely, based on Eq. (5) and pessimistic
BR mapping (defined in Eq. (7)), we can follow the similar idea in Eq. (32) to
aggregate the UL and LL gradients
$\widetilde{\mathbf{d}}(\mathbf{y}_{t-1};\mathbf{x})=-\rho_{t}\frac{\partial
F(\mathbf{x},\mathbf{y}_{t-1})}{\partial\mathbf{y}_{t-1}}+(1-\rho_{t})\frac{\partial
f(\mathbf{x},\mathbf{y}_{t-1})}{\partial\mathbf{y}_{t-1}}.$
With the above procedure, it can be seen that the only difference between
$\mathbf{d}$ and $\widetilde{\mathbf{d}}$ is just the sign of the UL gradient.
Thus we can adopt the same calculation scheme as that in [32, 53] to solve Eq.
(5). The corresponding roadmap is also illustrated in Fig. 9.
## 10 Conclusions and Future Prospects
Bi-Level Optimization (BLO) is an important mathematical tool for modeling and
solving machine learning and computer vision problems that have hierarchical
optimization structures, such as hyper-parameter optimization, multi-task and
meta learning, neural architecture search, adversarial learning and deep
reinforcement learning, etc. In the above sections, we first demonstrated how
to formulate different learning and vision tasks from a uniform BLO
perspective. We then established a value-function-based single-level
reformulation for different categories of BLO models and proposed a best-
response-based optimization platform to uniformly understand and formulate a
variety of existing gradient-based BLO methods. The convergence behaviors and
complexity properties of these BLO algorithms have also been discussed. We
also demonstrated potentials of our BLO platform for designing new algorithms
to solve the more challenging pessimistic BLOs tasks. The future research of
BLOs may focus but is not limited to the following aspects:
* •
Theoretical breakthrough: The convergence behaviors of gradient-based
algorithms on various challenging BLOs, such as pessimistic BLOs [215, 221,
222], BLOs with complex constraints [223, 224], nonconvex objectives [225] and
multiple followers [226], should be investigated.
* •
Computational improvement: It is also urgent to design efficient acceleration
techniques (e.g., momentum and its variations) to speed up gradient-based BLOs
in high-dimensional optimization scenario [227, 228, 229].
* •
Wider applications: Recent deep learning tasks (e.g., knowledge distillation
[230], self-supervised learning [231], and transformer [232]) are more and
more sophisticated. BLOs should be a promising tool to formulate and analyze
these complex learning paradigms.
## Acknowledgments
This work is partially supported by the National Key R&D Program of China
(2020YFB1313503), the National Natural Science Foundation of China (Nos.
61922019, 61733002, and 61672125), LiaoNing Revitalization Talents Program
(XLYC1807088), Shenzhen Science and Technology Program (No.
RCYX20200714114700072), and the Fundamental Research Funds for the Central
Universities.
## References
* [1] S. Dempe, “Bilevel optimization: Theory, algorithms, applications and a bibliography,” in _Bilevel Optimization_ , 2020, pp. 581–672.
* [2] H. Von Stackelberg and S. H. Von, _The theory of the market economy_. Oxford University Press, 1952\.
* [3] J. Fortuny-Amat and B. McCarl, “A representation and economic interpretation of a two-level programming problem,” _Journal of the Operational Research Society_ , vol. 32, no. 9, pp. 783–792, 1981.
* [4] S. Dempe, V. Kalashnikov, G. A. Pérez-Valdés, and N. Kalashnykova, “Bilevel programming problems,” _Energy Systems. Springer, Berlin_ , 2015\.
* [5] A. Sinha, P. Malo, and K. Deb, “A review on bilevel optimization: from classical to evolutionary approaches and applications,” _IEEE Transactions on Evolutionary Computation_ , vol. 22, no. 2, pp. 276–295, 2017\.
* [6] S. Wogrin, S. Pineda, and D. A. Tejada-Arango, “Applications of bilevel optimization in energy and electricity markets,” in _Bilevel Optimization_ , 2020, pp. 139–168.
* [7] J. Domke, “Generic methods for optimization-based modeling,” in _Artificial Intelligence and Statistics_ , 2012.
* [8] C.-s. Foo, C. B. Do, and A. Y. Ng, “Efficient multiple hyperparameter learning for log-linear models,” in _NeurPIS_ , 2008.
* [9] M. MacKay, P. Vicol, J. Lorraine, D. Duvenaud, and R. B. Grosse, “Self-tuning networks: Bilevel optimization of hyperparameters using structured best-response functions,” in _ICLR_ , 2019.
* [10] T. Okuno, A. Takeda, and A. Kawana, “Hyperparameter learning via bilevel nonsmooth optimization,” _arXiv:1806.01520_ , 2018.
* [11] L. Franceschi, M. Donini, P. Frasconi, and M. Pontil, “A bridge between hyperparameter optimization and learning-to-learn,” _arXiv:1712.06283_ , 2017\.
* [12] L. Franceschi, P. Frasconi, S. Salzo, R. Grazzi, and M. Pontil, “Bilevel programming for hyperparameter optimization and meta-learning,” in _ICML_ , 2018.
* [13] A. Shaban, C.-A. Cheng, N. Hatch, and B. Boots, “Truncated back-propagation for bilevel optimization,” in _AISTATS_ , 2019.
* [14] H. Liu, K. Simonyan, and Y. Yang, “DARTS: differentiable architecture search,” in _ICLR_ , 2019.
* [15] Y. Hu, X. Wu, and R. He, “TF-NAS: rethinking three search freedoms of latency-constrained differentiable neural architecture search,” in _ECCV_ , 2020.
* [16] D. Lian, Y. Zheng, Y. Xu, Y. Lu, L. Lin, P. Zhao, J. Huang, and S. Gao, “Towards fast adaptation of neural architectures with meta learning,” in _ICLR_ , 2019.
* [17] Y. Li, L. Song, X. Wu, R. He, and T. Tan, “Learning a bi-level adversarial network with global and local perception for makeup-invariant face verification,” _Pattern Recognition_ , vol. 90, pp. 99–108, 2019.
* [18] H. Jiang, Z. Chen, Y. Shi, B. Dai, and T. Zhao, “Learning to defense by learning to attack,” _arXiv:1811.01213_ , 2018.
* [19] Y. Tian, L. Shen, G. Su, Z. Li, and W. Liu, “Alphagan: Fully differentiable architecture search for generative adversarial networks,” _arXiv:2006.09134_ , 2020.
* [20] H. Zhang, W. Chen, Z. Huang, M. Li, Y. Yang, W. Zhang, and J. Wang, “Bi-level actor-critic for multi-agent coordination,” in _AAAI_ , vol. 34, no. 05, 2020, pp. 7325–7332.
* [21] D. Pfau and O. Vinyals, “Connecting generative adversarial networks and actor-critic methods,” _arXiv:1610.01945_ , 2016.
* [22] Z. Yang, Y. Chen, M. Hong, and Z. Wang, “Provably global convergence of actor-critic: A case for linear quadratic regulator with ergodic cost,” in _NeurIPS_ , 2019.
* [23] P. Ochs, R. Ranftl, T. Brox, and T. Pock, “Bilevel optimization with nonsmooth lower level problems,” in _SSVM_ , 2015.
* [24] J. Chen, P. Mu, R. Liu, X. Fan, and Z. Luo, “Flexible bilevel image layer modeling for robust deraining,” in _ICME_ , 2020.
* [25] R. Liu, Z. Li, Y. Zhang, X. Fan, and Z. Luo, “Bi-level probabilistic feature learning for deformable image registration,” in _IJCAI_ , 2020.
* [26] G. Kunapuli, K. P. Bennett, J. Hu, and J.-S. Pang, “Classification model selection via bilevel programming,” _Optimization Methods & Software_, vol. 23, no. 4, pp. 475–489, 2008.
* [27] S. Dempe and S. Franke, “On the solution of convex bilevel optimization problems,” _Computational Optimization and Applications_ , vol. 63, no. 3, pp. 685–703, 2016.
* [28] P. Hansen, B. Jaumard, and G. Savard, “New branch-and-bound rules for linear bilevel programming,” _SIAM Journal on scientific and Statistical Computing_ , vol. 13, no. 5, pp. 1194–1217, 1992.
* [29] A. B. Zemkoho, “Solving ill-posed bilevel programs,” _Set-Valued and Variational Analysis_ , vol. 24, no. 3, pp. 423–448, 2016.
* [30] L. Vicente, G. Savard, and J. Júdice, “Descent approaches for quadratic bilevel programming,” _Journal of Optimization Theory and Applications_ , vol. 81, no. 2, pp. 379–399, 1994.
* [31] H. I. Calvete and C. Galé, “Algorithms for linear bilevel optimization,” in _Bilevel Optimization_ , 2020.
* [32] R. Liu, P. Mu, X. Yuan, S. Zeng, and J. Zhang, “A generic first-order algorithmic framework for bi-level programming beyond lower-level singleton,” in _ICML_ , 2020.
* [33] A. Sharma, “Optimistic variants of single-objective bilevel optimization for evolutionary algorithms,” _International Journal of Computational Intelligence and Applications_ , no. 03, p. 2050020, 2020.
* [34] H. I. Calvete, C. Galé, S. Dempe, and S. Lohse, “Bilevel problems over polyhedra with extreme point optimal solutions,” _Journal of Global Optimization_ , vol. 53, no. 3, pp. 573–586, 2012.
* [35] J. Lu, C. Shi, G. Zhang, and D. Ruan, “An extended branch and bound algorithm for bilevel multi-follower decision making in a referential-uncooperative situation,” _International Journal of Information Technology Decision Making_ , vol. 6, no. 02, pp. 371–388, 2007.
* [36] G. Savard and J. Gauvin, “The steepest descent direction for the nonlinear bilevel programming problem,” _Operations Research Letters_ , vol. 15, no. 5, pp. 265–272, 1994.
* [37] E. S. H. Neto and Á. R. De Pierro, “On perturbed steepest descent methods with inexact line search for bilevel convex optimization,” _Optimization_ , vol. 60, no. 8-9, pp. 991–1008, 2011.
* [38] G. Anandalingam and D. White, “A solution method for the linear static stackelberg problem using penalty functions,” _IEEE Transactions on Automatic Control_ , vol. 35, no. 10, pp. 1170–1173, 1990.
* [39] Z. Wan, L. Mao, and G. Wang, “Estimation of distribution algorithm for a class of nonlinear bilevel programming problems,” _Information Sciences_ , vol. 256, pp. 184–196, 2014.
* [40] B. El-Sobky and Y. Abo-Elnaga, “A penalty method with trust-region mechanism for nonlinear bilevel optimization problem,” _Journal of Computational and Applied Mathematics_ , vol. 340, pp. 360–374, 2018.
* [41] S. Dempe and J. F. Bard, “Bundle trust-region algorithm for bilinear bilevel programming,” _Journal of Optimization Theory and Applications_ , vol. 110, no. 2, pp. 265–288, 2001.
* [42] G. B. Allende and G. Still, “Solving bilevel programs with the kkt-approach,” _Mathematical Programming_ , vol. 138, no. 1-2, pp. 309–332, 2013.
* [43] A. Sinha, T. Soun, and K. Deb, “Using karush-kuhn-tucker proximity measure for solving bilevel optimization problems,” _Swarm and Evolutionary Computation_ , vol. 44, pp. 496–510, 2019.
* [44] A. Sinha, S. Bedi, and K. Deb, “Bilevel optimization based on kriging approximations of lower level optimal value function,” in _2018 IEEE Congress on Evolutionary Computation (CEC)_ , 2018, pp. 1–8.
* [45] S. E. Yimer, P. Kumam, and A. G. Gebrie, “Proximal gradient method for solving bilevel optimization problems,” _Mathematical and Computational Applications_ , no. 4, p. 66, 2020.
* [46] A. U. Raghunathan and L. T. Biegler, “Mathematical programs with equilibrium constraints (mpecs) in process engineering,” _Computers & chemical engineering_, no. 10, pp. 1381–1392, 2003.
* [47] A. B. Zemkoho and S. Zhou, “Theoretical and numerical comparison of the karush–kuhn–tucker and value function reformulations in bilevel optimization,” _Computational Optimization and Applications_ , pp. 1–50, 2020.
* [48] D. Aussel and C. Lalitha, _Generalized Nash equilibrium problems, Bilevel programming and MPEC_. Springer, 2018.
* [49] G. Kunapuli, _A bilevel optimization approach to machine learning_. Citeseer, 2008.
* [50] S. Dempe and J. Dutta, “Is bilevel programming a special case of a mathematical program with complementarity constraints?” _Mathematical programming_ , vol. 131, no. 1, pp. 37–48, 2012.
* [51] L. Franceschi, M. Donini, P. Frasconi, and M. Pontil, “Forward and reverse gradient-based hyperparameter optimization,” _arXiv:1703.01785_ , 2017.
* [52] T. Okuno and A. Takeda, “Bilevel optimization of regularization hyperparameters in machine learning,” in _Bilevel Optimization_ , 2020.
* [53] Y. Liu and R. Liu, “Boml: A modularized bilevel optimization library in python for meta learning,” in _ICME_ , 2021.
* [54] K. Ji, J. Yang, and Y. Liang, “Multi-step model-agnostic meta-learning: Convergence and improved algorithms,” _arXiv:2002.07836_ , 2020.
* [55] Z. Li, F. Zhou, F. Chen, and H. Li, “Meta-sgd: Learning to learn quickly for few-shot learning,” _arXiv:1707.09835_ , 2017.
* [56] Y. Chen, M. W. Hoffman, S. G. Colmenarejo, M. Denil, T. P. Lillicrap, M. Botvinick, and N. Freitas, “Learning to learn without gradient descent by gradient descent,” in _ICML_ , 2017.
* [57] A. Antoniou, H. Edwards, and A. Storkey, “How to train your maml,” _arXiv:1810.09502_ , 2018.
* [58] H. Dong, B. Zou, L. Zhang, and S. Zhang, “Automatic design of cnns via differentiable neural architecture search for polsar image classification,” _IEEE Transactions on Geoscience and Remote Sensing_ , vol. 58, no. 9, pp. 6362–6375, 2020.
* [59] Y. Xu, L. Xie, X. Zhang, X. Chen, G.-J. Qi, Q. Tian, and H. Xiong, “Pc-darts: Partial channel connections for memory-efficient architecture search,” in _ICLR_ , 2019.
* [60] Y. Liu, W. Cai, X. Yuan, and J. Xiang, “Gl-gan: Adaptive global and local bilevel optimization model of image generation,” _arXiv:2008.02436_ , 2020\.
* [61] L. Wang, Q. Cai, Z. Yang, and Z. Wang, “On the global optimality of model-agnostic meta-learning: Reinforcement learning and supervised learning,” in _ICML_ , 2020.
* [62] M. Hong, H.-T. Wai, Z. Wang, and Z. Yang, “A two-timescale framework for bilevel optimization: Complexity analysis and application to actor-critic,” _arXiv:2007.05170_ , 2020.
* [63] G. Anandalingam and T. L. Friesz, “Hierarchical optimization: An introduction,” _Annals of Operations Research_ , vol. 34, no. 1, pp. 1–11, 1992.
* [64] U.-P. Wen and S.-T. Hsu, “Linear bi-level programming problems—a review,” _Journal of the Operational Research Society_ , vol. 42, no. 2, pp. 125–133, 1991.
* [65] S. Ukkusuri, K. Doan, and H. A. Aziz, “A bi-level formulation for the combined dynamic equilibrium based traffic signal control,” _Procedia-Social and Behavioral Sciences_ , vol. 80, no. 7, pp. 729–752, 2013.
* [66] K. Lachhwani and A. Dwivedi, “Bi-level and multi-level programming problems: taxonomy of literature review and research issues,” _Archives of Computational Methods in Engineering_ , vol. 25, no. 4, pp. 847–877, 2018.
* [67] A. Chinchuluun, P. M. Pardalos, and H.-X. Huang, “Multilevel (hierarchical) optimization: complexity issues, optimality conditions, algorithms,” in _Applied Mathematics and Global Optimization_ , 2009.
* [68] S. Gould, B. Fernando, A. Cherian, P. Anderson, R. S. Cruz, and E. Guo, “On differentiating parameterized argmin and argmax problems with application to bi-level optimization,” _arXiv:1607.05447_ , 2016.
* [69] C.-A. Deledalle, S. Vaiter, J. Fadili, and G. Peyré, “Stein unbiased gradient estimator of the risk (sugar) for multiple parameter selection,” _SIAM Journal on Imaging Sciences_ , vol. 7, no. 4, pp. 2448–2487, 2014.
* [70] W. Jiang and S. Siddiqui, “Hyper-parameter optimization for support vector machines using stochastic gradient descent and dual coordinate descent,” _EURO Journal on Computational Optimization_ , vol. 8, no. 1, pp. 85–101, 2020.
* [71] D. Maclaurin, D. Duvenaud, and R. Adams, “Gradient-based hyperparameter optimization through reversible learning,” in _ICML_ , 2015.
* [72] J. Lorraine, P. Vicol, and D. Duvenaud, “Optimizing millions of hyperparameters by implicit differentiation,” in _AISTATS_ , 2020.
* [73] L. Franceschi, R. Grazzi, M. Pontil, S. Salzo, and P. Frasconi, “Far-ho: A bilevel programming package for hyperparameter optimization and meta-learning,” _arXiv:1806.04941_ , 2018.
* [74] F. Pedregosa, “Hyperparameter optimization with approximate gradient,” _arXiv:1602.02355_ , 2016.
* [75] V. Likhosherstov, X. Song, K. Choromanski, J. Davis, and A. Weller, “Ufo-blo: Unbiased first-order bilevel optimization,” _arXiv:2006.03631_ , 2020.
* [76] B. Amos and J. Z. Kolter, “Optnet: Differentiable optimization as a layer in neural networks,” in _ICML_ , 2017.
* [77] J. Bae and R. B. Grosse, “Delta-stn: Efficient bilevel optimization for neural networks using structured response jacobians,” in _NeurIPS_ , 2020.
* [78] S. Flennerhag, A. A. Rusu, R. Pascanu, H. Yin, and R. Hadsell, “Meta-learning with warped gradient descent,” _arXiv:1909.00025_ , 2019.
* [79] S. Qiao, C. Liu, W. Shen, and A. L. Yuille, “Few-shot image recognition by predicting parameters from activations,” in _CVPR_ , 2018.
* [80] S. Gidaris and N. Komodakis, “Dynamic few-shot visual learning without forgetting,” in _CVPR_ , 2018.
* [81] N. Mishra, M. Rohaninejad, X. Chen, and P. Abbeel, “A simple neural attentive meta-learner,” _arXiv:1707.03141_ , 2017.
* [82] K. Li and J. Malik, “Learning to optimize,” _arXiv:1606.01885_ , 2016.
* [83] A. A. Rusu, D. Rao, J. Sygnowski, O. Vinyals, R. Pascanu, S. Osindero, and R. Hadsell, “Meta-learning with latent embedding optimization,” in _ICLR_ , 2018.
* [84] L. Zintgraf, K. Shiarli, V. Kurin, K. Hofmann, and S. Whiteson, “Fast context adaptation via meta-learning,” in _ICML_ , 2019.
* [85] A. Rajeswaran, C. Finn, S. M. Kakade, and S. Levine, “Meta-learning with implicit gradients,” in _NeurIPS_ , 2019.
* [86] L. Bertinetto, J. F. Henriques, P. H. Torr, and A. Vedaldi, “Meta-learning with differentiable closed-form solvers,” _arXiv:1805.08136_ , 2018.
* [87] E. Park and J. B. Oliva, “Meta-curvature,” in _NeurIPS_ , 2019.
* [88] S. Ravi and H. Larochelle, “Optimization as a model for few-shot learning,” in _ICLR_ , 2016.
* [89] C. Finn, P. Abbeel, and S. Levine, “Model-agnostic meta-learning for fast adaptation of deep networks,” in _ICML_ , 2017.
* [90] A. Nichol, J. Achiam, and J. Schulman, “On first-order meta-learning algorithms,” _arXiv:1803.02999_ , 2018.
* [91] H. S. Behl, A. G. Baydin, and P. H. Torr, “Alpha maml: Adaptive model-agnostic meta-learning,” _arXiv:1905.07435_ , 2019.
* [92] J. Guo, X. Zhu, C. Zhao, D. Cao, Z. Lei, and S. Z. Li, “Learning meta face recognition in unseen domains,” in _CVPR_ , 2020.
* [93] Q. Wu, Z. Lin, G. Wang, H. Chen, B. F. Karlsson, B. Huang, and C.-Y. Lin, “Enhanced meta-learning for cross-lingual named entity recognition with minimal resources.” in _AAAI_ , 2020.
* [94] A. Nichol and J. Schulman, “Reptile: a scalable metalearning algorithm,” _arXiv:1803.02999_ , 2018.
* [95] P. Zhou, X. Yuan, H. Xu, S. Yan, and J. Feng, “Efficient meta learning via minibatch proximal update,” in _NeurIPS_ , 2019.
* [96] X. Song, W. Gao, Y. Yang, K. Choromanski, A. Pacchiano, and Y. Tang, “Es-maml: Simple hessian-free meta learning,” _arXiv:1910.01215_ , 2019.
* [97] G. Denevi, C. Ciliberto, R. Grazzi, and M. Pontil, “Learning-to-learn stochastic gradient descent with biased regularization,” in _ICML_ , 2019\.
* [98] Y. Lee and S. Choi, “Gradient-based meta-learning with learned layerwise metric and subspace,” _arXiv:1801.05558_ , 2018.
* [99] J. W. Soh, S. Cho, and N. I. Cho, “Meta-transfer learning for zero-shot super-resolution,” in _CVPR_ , 2020.
* [100] P. Tian, Z. Wu, L. Qi, L. Wang, Y. Shi, and Y. Gao, “Differentiable meta-learning model for few-shot semantic segmentation.” in _AAAI_ , 2020\.
* [101] J.-Y. Hsu, Y.-J. Chen, and H.-y. Lee, “Meta learning for end-to-end low-resource speech recognition,” in _ICASSP_ , 2020.
* [102] C. Wong, N. Houlsby, Y. Lu, and A. Gesmundo, “Transfer learning with neural automl,” in _NeurIPS_ , 2018.
* [103] C. Jiang, H. Xu, W. Zhang, X. Liang, and Z. Li, “Sp-nas: Serial-to-parallel backbone search for object detection,” in _CVPR_ , 2020.
* [104] S. Xie, H. Zheng, C. Liu, and L. Lin, “Snas: stochastic neural architecture search,” _arXiv:1812.09926_ , 2018.
* [105] H. Pham, M. Y. Guan, B. Zoph, Q. V. Le, and J. Dean, “Efficient neural architecture search via parameter sharing,” _arXiv:1802.03268_ , 2018.
* [106] X. Chen, L. Xie, J. Wu, and Q. Tian, “Progressive differentiable architecture search: Bridging the depth gap between search and evaluation,” in _CVPR_ , 2019.
* [107] A. Noy, N. Nayman, T. Ridnik, N. Zamir, S. Doveh, I. Friedman, R. Giryes, and L. Zelnik, “Asap: Architecture search, anneal and prune,” in _AISTATS_ , 2020.
* [108] T. Elsken, B. Staffler, J. H. Metzen, and F. Hutter, “Meta-learning of neural architectures for few-shot learning,” in _CVPR_ , 2020.
* [109] Q. Yao, J. Xu, W.-W. Tu, and Z. Zhu, “Efficient neural architecture search via proximal iterations.” in _AAAI_ , 2020.
* [110] S. Hu, S. Xie, H. Zheng, C. Liu, J. Shi, X. Liu, and D. Lin, “Dsnas: Direct neural architecture search without parameter retraining,” in _CVPR_ , 2020\.
* [111] C. Liu, L.-C. Chen, F. Schroff, H. Adam, W. Hua, A. L. Yuille, and L. Fei-Fei, “Auto-deeplab: Hierarchical neural architecture search for semantic image segmentation,” in _CVPR_ , 2019.
* [112] Y. Chen, T. Yang, X. Zhang, G. Meng, X. Xiao, and J. Sun, “Detnas: Backbone search for object detection,” in _NeurIPS_ , 2019.
* [113] H. Xu, L. Yao, W. Zhang, X. Liang, and Z. Li, “Auto-fpn: Automatic network architecture adaptation for object detection beyond classification,” in _ICCV_ , 2019.
* [114] J. Chang, Y. Guo, G. MENG, S. XIANG, C. Pan _et al._ , “Data: Differentiable architecture approximation,” in _NeurIPS_ , 2019.
* [115] Q. Yu, D. Yang, H. Roth, Y. Bai, Y. Zhang, A. L. Yuille, and D. Xu, “C2fnas: Coarse-to-fine neural architecture search for 3d medical image segmentation,” in _CVPR_ , 2020.
* [116] K. Zhou, Q. Song, X. Huang, and X. Hu, “Auto-gnn: Neural architecture search of graph neural networks,” _arXiv:1909.03184_ , 2019.
* [117] T. Li, J. Zhang, K. Bao, Y. Liang, Y. Li, and Y. Zheng, “Autost: Efficient neural architecture search for spatio-temporal prediction,” in _KDD_ , 2020\.
* [118] J. Guo, K. Han, Y. Wang, C. Zhang, Z. Yang, H. Wu, X. Chen, and C. Xu, “Hit-detector: Hierarchical trinity architecture search for object detection,” in _CVPR_ , 2020.
* [119] C. He, H. Ye, L. Shen, and T. Zhang, “Milenas: Efficient neural architecture search via mixed-level reformulation,” in _CVPR_ , 2020.
* [120] W. Cheng, Y. Shen, and L. Huang, “Differentiable neural input search for recommender systems,” _arXiv:2006.04466_ , 2020.
* [121] L. Metz, B. Poole, D. Pfau, and J. Sohl-Dickstein, “Unrolled generative adversarial networks,” _arXiv:1611.02163_ , 2016.
* [122] H. Yin, D. Li, X. Li, and P. Li, “Meta-cotgan: A meta cooperative training paradigm for improving adversarial text generation.” in _AAAI_ , 2020.
* [123] C. Gao, Y. Chen, S. Liu, Z. Tan, and S. Yan, “Adversarialnas: Adversarial neural architecture search for gans,” in _CVPR_ , 2020.
* [124] C. Jin, P. Netrapalli, and M. Jordan, “What is local optimality in nonconvex-nonconcave minimax optimization?” in _ICML_ , 2020, pp. 4880–4889.
* [125] J. Hamm and Y.-K. Noh, “K-beam minimax: Efficient optimization for deep adversarial learning,” in _ICML_ , 2018.
* [126] Z. Chen, D. Liu, X. Wu, and X. Xu, “Research on distributed renewable energy transaction decision-making based on multi-agent bilevel cooperative reinforcement learning,” in _CIRED_ , 2019.
* [127] S. Tschiatschek, A. Ghosh, L. Haug, R. Devidze, and A. Singla, “Learner-aware teaching: Inverse reinforcement learning with preferences and constraints,” in _NeurIPS_ , 2019.
* [128] Y. Zhang, Q. Cai, Z. Yang, and Z. Wang, “Generative adversarial imitation learning with neural network parameterization: Global optimality and convergence rate,” in _ICML_ , 2020, pp. 11 044–11 054.
* [129] J. Yang, I. Borovikov, and H. Zha, “Hierarchical cooperative multi-agent reinforcement learning with skill discovery,” in _AAMAS_ , 2020.
* [130] Y. Gao, H. Yang, P. Zhang, C. Zhou, and Y. Hu, “Graphnas: Graph neural architecture search with reinforcement learning,” _arXiv:1904.09981_ , 2019\.
* [131] Y. Ye, D. Qiu, M. Sun, D. Papadaskalopoulos, and G. Strbac, “Deep reinforcement learning for strategic bidding in electricity markets,” _IEEE Transactions on Smart Grid_ , vol. 11, no. 2, pp. 1343–1355, 2019.
* [132] J. Ho and S. Ermon, “Generative adversarial imitation learning,” _NeurIPS_ , vol. 29, pp. 4565–4573, 2016.
* [133] Y. Li, J. Song, and S. Ermon, “Infogail: Interpretable imitation learning from visual demonstrations,” in _NeurIPS_ , 2017, pp. 3815–3825.
* [134] F. Torabi, G. Warnell, and P. Stone, “Generative adversarial imitation from observation,” _arXiv:1807.06158_ , 2018.
* [135] H. Liu, R. Socher, and C. Xiong, “Taming maml: Efficient unbiased meta-reinforcement learning,” in _ICML_ , 2019, pp. 4061–4071.
* [136] K. Kunisch and T. Pock, “A bilevel optimization approach for parameter learning in variational models,” _SIAM Journal on Imaging Sciences_ , vol. 6, no. 2, pp. 938–983, 2013.
* [137] R. Liu, J. Liu, Z. Jiang, X. Fan, and Z. Luo, “A bilevel integrated model with data-driven layer ensemble for multi-modality image fusion,” _IEEE Transactions on Image Processing_ , vol. 30, pp. 1261–1274, 2020.
* [138] H. Li and L. Zhang, “A bilevel learning model and algorithm for self-organizing feed-forward neural networks for pattern classification,” _IEEE Transactions on Neural Networks and Learning Systems_ , 2020.
* [139] P. Zhou, C. Zhang, and Z. Lin, “Bilevel model-based discriminative dictionary learning for recognition,” _IEEE transactions on image processing_ , vol. 26, no. 3, pp. 1173–1187, 2016.
* [140] B. Fernando and S. Gould, “Learning end-to-end video classification with rank-pooling,” in _ICML_ , 2016, pp. 1187–1196.
* [141] D. Pfau, S. Petersen, A. Agarwal, D. G. Barrett, and K. L. Stachenfeld, “Spectral inference networks: Unifying deep and spectral learning,” in _ICLR_ , 2019.
* [142] M. D’Elia, J. De los Reyes, and A. M. Trujillo, “Bilevel parameter optimization for nonlocal image denoising models,” _arXiv:1912.02347_ , 2019\.
* [143] B. Stadie, L. Zhang, and J. Ba, “Learning intrinsic rewards as a bi-level optimization problem,” in _UAI_ , 2020, pp. 111–120.
* [144] R. Liu, Z. Li, X. Fan, C. Zhao, H. Huang, and Z. Luo, “Learning deformable image registration from optimization: perspective, modules, bilevel training and beyond,” _arXiv:2004.14557_ , 2021.
* [145] Q. Pham, D. Sahoo, C. Liu, and S. C. Hoi, “Bilevel continual learning,” _arXiv:2007.15553_ , 2020.
* [146] R. Liu, P. Mu, J. Chen, X. Fan, and Z. Luo, “Investigating task-driven latent feasibility for nonconvex image modeling,” _IEEE Transactions on Image Processing_ , vol. 29, pp. 7629–7640, 2020.
* [147] R. Liu, L. Ma, X. Yuan, S. Zeng, and J. Zhang, “Bilevel integrative optimization for ill-posed inverse problems,” _arXiv:1907.03083_ , 2019.
* [148] R. Liu, L. Ma, J. Zhang, X. Fan, and Z. Luo, “Retinex-inspired unrolling with cooperative prior architecture search for low-light image enhancement,” _arXiv:2012.05609_ , 2020.
* [149] T. Stouraitis, I. Chatzinikolaidis, M. Gienger, and S. Vijayakumar, “Online hybrid motion planning for dyadic collaborative manipulation via bilevel optimization,” _IEEE Transactions on Robotics_ , vol. 36, no. 5, pp. 1452–1471, 2020.
* [150] S. Mounsaveng, I. Laradji, I. Ben Ayed, D. Vazquez, and M. Pedersoli, “Learning data augmentation with online bilevel optimization for image classification,” in _WACV_ , 2020.
* [151] Z. Borsos, M. Tagliasacchi, and A. Krause, “Semi-supervised batch active learning via bilevel optimization,” _arXiv:2010.09654_ , 2020.
* [152] Z. Borsos, M. Mutnỳ, and A. Krause, “Coresets via bilevel optimization for continual learning and streaming,” in _NeurIPS_ , 2020.
* [153] N. K. Chada, C. Schillings, X. T. Tong, and S. Weissmann, “Consistency analysis of bilevel data-driven learning in inverse problems,” _arXiv:2007.02677_ , 2020.
* [154] F. Yousefian, “Bilevel distributed optimization in directed networks,” _arXiv:2006.07564_ , 2020.
* [155] Y. Liu, Y. Su, A.-A. Liu, B. Schiele, and Q. Sun, “Mnemonics training: Multi-class incremental learning without forgetting,” in _CVPR_ , 2020.
* [156] O. Litany and D. Freedman, “Soseleto: A unified approach to transfer learning and training with noisy labels,” in _ICLR_ , 2018.
* [157] K. P. Bennett, G. Kunapuli, J. Hu, and J.-S. Pang, “Bilevel optimization and machine learning,” in _IEEE World Congress on Computational Intelligence_ , 2008, pp. 25–47.
* [158] N. Couellan and W. Wang, “Bi-level stochastic gradient for large scale support vector machine,” _Neurocomputing_ , vol. 153, 2015.
* [159] J. Bergstra, D. Yamins, and D. Cox, “Making a science of model search: Hyperparameter optimization in hundreds of dimensions for vision architectures,” in _ICML_ , 2013.
* [160] A. G. Baydin and B. A. Pearlmutter, “Automatic differentiation of algorithms for machine learning,” _arXiv:1404.7456_ , 2014.
* [161] O. Chapelle, V. Vapnik, O. Bousquet, and S. Mukherjee, “Choosing multiple parameters for support vector machines,” _Machine Learning_ , vol. 46, no. 1-3, pp. 131–159, 2002.
* [162] M. W. Seeger, “Cross-validation optimization for large scale structured classification kernel methods,” _The Journal of Machine Learning Research_ , vol. 9, no. Jun, pp. 1147–1178, 2008.
* [163] L. Calatroni, C. Cao, J. C. De Los Reyes, C.-B. Schönlieb, and T. Valkonen, “Bilevel approaches for learning of variational imaging models,” _Variational Methods: In Imaging and Geometric Control_ , vol. 18, no. 252, p. 2, 2017.
* [164] J. Lorraine and D. Duvenaud, “Stochastic hyperparameter optimization through hypernetworks,” _arXiv:1802.09419_ , 2018.
* [165] S. Thrun and L. Pratt, “Learning to learn: Introduction and overview,” in _Learning to Learn_ , 1998.
* [166] S. Ruder, “An overview of multi-task learning in deep neural networks,” _arXiv:1706.05098_ , 2017.
* [167] M. Zhao, B. An, Y. Yu, S. Liu, and S. J. Pan, “Data poisoning attacks on multi-task relationship learning.” in _AAAI_ , 2018.
* [168] F. Alesiani, S. Yu, A. Shaker, and W. Yin, “Towards interpretable multi-task learning using bilevel programming,” _arXiv:2009.05483_ , 2020.
* [169] M. Andrychowicz, M. Denil, S. Gomez, M. W. Hoffman, D. Pfau, T. Schaul, B. Shillingford, and N. De Freitas, “Learning to learn by gradient descent by gradient descent,” in _NeurIPS_ , 2016.
* [170] M.-F. Balcan, M. Khodak, and A. Talwalkar, “Provable guarantees for gradient-based meta-learning,” in _ICML_ , 2019.
* [171] Y. Lee and S. Choi, “Meta-learning with adaptive layerwise metric and subspace,” in _ICML_ , 2017.
* [172] T. Elsken, J. H. Metzen, and F. Hutter, “Neural architecture search: A survey,” _arXiv:1808.05377_ , 2018.
* [173] F. P. Casale, J. Gordon, and N. Fusi, “Probabilistic neural architecture search,” _arXiv:1902.05116_ , 2019.
* [174] H. Mendoza, A. Klein, M. Feurer, J. T. Springenberg, and F. Hutter, “Towards automatically-tuned neural networks,” in _AutoML_ , 2016.
* [175] J. Wang, J. Wu, H. Bai, and J. Cheng, “M-nas: Meta neural architecture search.” in _AAAI_ , 2020.
* [176] J. Kim, S. Lee, S. Kim, M. Cha, J. K. Lee, Y. Choi, Y. Choi, D.-Y. Cho, and J. Kim, “Auto-meta: Automated gradient based meta learner search,” _arXiv:1806.06927_ , 2018.
* [177] Z. Zhu, C. Liu, D. Yang, A. Yuille, and D. Xu, “V-nas: Neural architecture search for volumetric medical image segmentation,” in _3DV_ , 2019.
* [178] T. Pang, X. Yang, Y. Dong, K. Xu, H. Su, and J. Zhu, “Boosting adversarial training with hypersphere embedding,” _arXiv:2002.08619_ , 2020.
* [179] I. Goodfellow, J. Pouget-Abadie, M. Mirza, B. Xu, D. Warde-Farley, S. Ozair, A. Courville, and Y. Bengio, “Generative adversarial nets,” in _NeurIPS_ , 2014.
* [180] X. Wang, H. Chen, J. Wu, Y. Ding, Q. Lou, and S. Liu, “Bi-level multi-agents interactive decision-making model in regional integrated energy system,” in _Energy Internet and Energy System Integration_ , 2019, pp. 2103–2108.
* [181] Z. Xu, H. van Hasselt, and D. Silver, “Meta-gradient reinforcement learning,” _arXiv:1805.09801_ , 2018.
* [182] I. Clavera, J. Rothfuss, J. Schulman, Y. Fujita, T. Asfour, and P. Abbeel, “Model-based reinforcement learning via meta-policy optimization,” in _Conference on Robot Learning_ , 2018, pp. 617–629.
* [183] J. X. Wang, Z. Kurth-Nelson, D. Tirumala, H. Soyer, J. Z. Leibo, R. Munos, C. Blundell, D. Kumaran, and M. Botvinick, “Learning to reinforcement learn,” _arXiv:1611.05763_ , 2016.
* [184] S. Arora, S. S. Du, S. Kakade, Y. Luo, and N. Saunshi, “Provable representation learning for imitation learning via bi-level optimization,” _arXiv:2002.10544_ , 2020.
* [185] M. Pan, C. Jian, R. Liu, F. Xin, and Z. Luo, “Learning bilevel layer priors for single image rain streaks removal,” _IEEE Signal Processing Letters_ , vol. 26, no. 2, pp. 307–311, 2019.
* [186] L. Ma, Q. Sun, B. Schiele, and L. Van Gool, “A novel bilevel paradigm for image-to-image translation,” _arXiv:1904.09028_ , 2019.
* [187] J. Xu, L. Luo, C. Deng, and H. Huang, “Bilevel distance metric learning for robust image recognition,” in _NeurIPS_ , 2018, pp. 4198–4207.
* [188] S. Zha, T. N. Pappas, and D. L. Neuhoff, “Hierarchical bilevel image compression based on cutset sampling,” in _ICIP_ , 2012, pp. 2517–2520.
* [189] R. Liu, P. Mu, X. Yuan, S. Zeng, and J. Zhang, “A generic descent aggregation framework for gradient-based bi-level optimization,” in _ICML_ , 2021.
* [190] S. Jenni and P. Favaro, “Deep bilevel learning,” in _ECCV_ , 2018.
* [191] R. Grazzi, L. Franceschi, M. Pontil, and S. Salzo, “On the iteration complexity of hypergradient computation,” in _ICML_ , 2020.
* [192] S. Dempe, J. Dutta, and B. Mordukhovich, “New necessary optimality conditions in optimistic bilevel programming,” _Optimization_ , vol. 56, no. 5-6, pp. 577–604, 2007.
* [193] B. Kohli, “Optimality conditions for optimistic bilevel programming problem using convexifactors,” _Journal of Optimization Theory and Applications_ , vol. 152, no. 3, pp. 632–651, 2012.
* [194] L. Lampariello and S. Sagratella, “Numerically tractable optimistic bilevel problems,” _Computational Optimization and Applications_ , vol. 76, no. 2, pp. 277–303, 2020.
* [195] R. Liu, X. Liu, X. Yuan, S. Zeng, and J. Zhang, “A value-function-based interior-point method for non-convex bi-level optimization,” 2021.
* [196] W. Wiesemann, A. Tsoukalas, P.-M. Kleniati, and B. Rustem, “Pessimistic bilevel optimization,” _SIAM Journal on Optimization_ , vol. 23, no. 1, pp. 353–380, 2013.
* [197] P. Loridan and J. Morgan, “Weak via strong stackelberg problem: new results,” _Journal of Global Optimization_ , vol. 8, no. 3, pp. 263–287, 1996.
* [198] R. Kicsiny, Z. Varga, and A. Scarelli, “Backward induction algorithm for a class of closed-loop stackelberg games,” _European Journal of Operational Research_ , vol. 237, no. 3, pp. 1021–1036, 2014.
* [199] J. Dutta and T. Pandit, “Algorithms for simple bilevel programming,” in _Bilevel Optimization_ , 2020, pp. 253–291.
* [200] E. Weinan, “A proposal on machine learning via dynamical systems,” _Communications in Mathematics and Statistics_ , vol. 5, no. 1, pp. 1–11, 2017.
* [201] Y. Lu, A. Zhong, Q. Li, and B. Dong, “Beyond finite layer neural networks: Bridging deep architectures and numerical differential equations,” in _ICML_ , 2018.
* [202] A. G. Baydin, B. A. Pearlmutter, A. A. Radul, and J. M. Siskind, “Automatic differentiation in machine learning: a survey,” _The Journal of Machine Learning Research_ , vol. 18, no. 1, pp. 5595–5637, 2017.
* [203] B. A. Pearlmutter and J. M. Siskind, “Reverse-mode ad in a functional framework: Lambda the ultimate backpropagator,” _ACM Transactions on Programming Languages and Systems_ , vol. 30, no. 2, pp. 1–36, 2008.
* [204] A. N. Gomez, M. Ren, R. Urtasun, and R. B. Grosse, “The reversible residual network: Backpropagation without storing activations,” in _NeurIPS_ , 2017\.
* [205] K. A. Khan and P. I. Barton, “A vector forward mode of automatic differentiation for generalized derivative evaluation,” _Optimization Methods and Software_ , vol. 30, no. 6, pp. 1185–1212, 2015.
* [206] J. Revels, M. Lubin, and T. Papamarkou, “Forward-mode automatic differentiation in julia,” _arXiv:1607.07892_ , 2016.
* [207] J. Luketina, M. Berglund, K. Greff, and T. Raiko, “Scalable gradient-based tuning of continuous regularization hyperparameters,” in _ICML_ , 2016.
* [208] A. G. Baydin, R. Cornish, D. M. Rubio, M. Schmidt, and F. Wood, “Online learning rate adaptation with hypergradient descent,” _arXiv:1703.04782_ , 2017.
* [209] A. Raghu, M. Raghu, S. Bengio, and O. Vinyals, “Rapid learning or feature reuse? towards understanding the effectiveness of maml,” in _ICLR_ , 2020\.
* [210] K. Ji, J. D. Lee, Y. Liang, and H. V. Poor, “Convergence of meta-learning with task-specific adaptation over partial parameters,” _arXiv:2006.09486_ , 2020\.
* [211] R. T. Rockafellar and R. J.-B. Wets, _Variational analysis_. Springer Science & Business Media, 2009, vol. 317\.
* [212] M. Solodov, “An explicit descent method for bilevel convex optimization,” _Journal of Convex Analysis_ , vol. 14, no. 2, p. 227, 2007.
* [213] S. Sabach and S. Shtern, “A first order method for solving convex bilevel optimization problems,” _SIAM Journal on Optimization_ , vol. 27, no. 2, pp. 640–660, 2017.
* [214] A. Fallah, A. Mokhtari, and A. Ozdaglar, “On the convergence theory of gradient-based model-agnostic meta-learning algorithms,” in _SIATATS_ , 2020\.
* [215] S. Dempe, B. S. Mordukhovich, and A. B. Zemkoho, “Necessary optimality conditions in pessimistic bilevel programming,” _Optimization_ , vol. 63, no. 4, pp. 505–533, 2014.
* [216] A. Tsoukalas, W. Wiesemann, B. Rustem _et al._ , “Global optimisation of pessimistic bi-level problems,” _Lectures on Global Optimization_ , vol. 55, pp. 215–243, 2009.
* [217] A. V. Malyshev and A. S. Strekalovsky, “On global search for pessimistic solution in bilevel problems,” _International Journal of Biomedical Soft Computing and Human Sciences: the official journal of the Biomedical Fuzzy Systems Association_ , vol. 18, no. 1, pp. 57–61, 2013.
* [218] Y. Zheng, G. Zhang, J. Han, and J. Lu, “Pessimistic bilevel optimization model for risk-averse production-distribution planning,” _Information Sciences_ , vol. 372, pp. 677–689, 2016.
* [219] T. Kis, A. Kovács, and C. Mészáros, “On optimistic and pessimistic bilevel optimization models for demand response management,” _Energies_ , vol. 14, no. 8, p. 2095, 2021.
* [220] B. Zeng, “A practical scheme to compute the pessimistic bilevel optimization problem,” _INFORMS Journal on Computing_ , vol. 32, no. 4, pp. 1128–1142, 2020.
* [221] J. Liu, Y. Fan, Z. Chen, and Y. Zheng, “Methods for pessimistic bilevel optimization,” in _Bilevel Optimization_ , 2020, pp. 403–420.
* [222] S. Dempe and J. Dutta, “Is bilevel programming a special case of a mathematical program with complementarity constraints?” _Mathematical Programming_ , vol. 131, pp. 37–48, 2012.
* [223] D. Ralph and S. Dempe, “Directional derivatives of the solution of a parametric nonlinear program,” _Mathematical programming_ , vol. 70, no. 1, pp. 159–172, 1995.
* [224] J. Y. Jane, “Constraint qualifications and optimality conditions in bilevel optimization,” in _Bilevel Optimization_ , 2020, pp. 227–251.
* [225] P. Borges, C. Sagastizábal, and M. Solodov, “A regularized smoothing method for fully parameterized convex problems with applications to convex and nonconvex two-stage stochastic programming,” _Mathematical Programming_ , 2020.
* [226] X. Zhou, R. Luo, Y. Tu, B. Lev, and W. Pedrycz, “Data envelopment analysis for bi-level systems with multiple followers,” _Omega_ , vol. 77, pp. 180–188, 2018.
* [227] S. Ghadimi and M. Wang, “Approximation methods for bilevel programming,” _arXiv:1802.02246_ , 2018.
* [228] Y. Shehu, P. T. Vuong, and A. Zemkoho, “An inertial extrapolation method for convex simple bilevel optimization,” _Optimization Methods and Software_ , vol. 36, no. 1, pp. 1–19, 2021.
* [229] F. Huang and H. Huang, “Biadam: Fast adaptive bilevel optimization methods,” _arXiv:2106.11396_ , 2021.
* [230] M. Phuong and C. Lampert, “Towards understanding knowledge distillation,” in _ICML_ , 2019, pp. 5142–5151.
* [231] L. Jing and Y. Tian, “Self-supervised visual feature learning with deep neural networks: A survey,” _IEEE transactions on pattern analysis and machine intelligence_ , vol. PP, no. 99, pp. 1–1, 2020.
* [232] N. Parmar, A. Vaswani, J. Uszkoreit, L. Kaiser, N. Shazeer, A. Ku, and D. Tran, “Image transformer,” in _ICML_ , 2018, pp. 4055–4064.
| Risheng Liu received his B.Sc. (2007) and Ph.D. (2012) from Dalian
University of Technology, China. From 2010 to 2012, he was doing research as
joint Ph.D. in robotics institute at Carnegie Mellon University. From 2016 to
2018, He was doing research as Hong Kong Scholar at the Hong Kong Polytechnic
University. He is currently a full professor with the Digital Media Department
at International School of Information Science & Engineering, Dalian
University of Technology. He was awarded the “Outstanding Youth Science
Foundation” of the National Natural Science Foundation of China. His research
interests include optimization, computer vision and multimedia.
---|---
| Jiaxin Gao received the B.S. degree in Applied Mathematics from Dalian
University of Technology, China, in 2018. She is currently pursuing the PhD
degree in software engineering at Dalian University of Technology, Dalian,
China. She is with the Key Laboratory for Ubiquitous Network and Service
Software of Liaoning Province, Dalian University of Technology, Dalian, China.
Her research interests include computer vision, machine learning and
optimization.
---|---
| Jin Zhang received the B.A. degree in Journalism from the Dalian University
of Technology in 2007. He pursued a degree in mathematics and received the
M.S. degree in Operational Research and Cybernetics from the Dalian University
of Technology, China, in 2010, and the PhD degree in Applied Mathematics from
University of Victoria, Canada, in 2015. After working in Hong Kong Baptist
University for 3 years, he joined Southern University of Science and
Technology as a tenure-track assistant professor in the Department of
Mathematics. His broad research area is comprised of optimization, variational
analysis and their applications in economics, engineering and data science.
---|---
| Deyu Meng (Member, IEEE) received the B.Sc. degree in information science,
the M.Sc. degree in applied mathematics, and the Ph.D. degree in computer
science from Xi’an Jiaotong University, Xi’an, China, in 2001, 2004, and 2008,
respectively. He is currently a Professor with the School of Mathematics and
Statistics, Xi’an Jiaotong University, and adjunct Professor with the Faculty
of Information Technology, The Macau University of Science and Technology,
Macao. From 2012 to 2014, he took his two-year sabbatical leave at Carnegie
Mellon University, Pittsburgh, PA, USA. His current research interests include
model-based deep learning, variational networks, and meta learning.
---|---
| Zhouchen Lin (M’00-SM’08-F’18) received the PhD degree from Peking
University in 2000. He is currently a professor with the Key Laboratory of
Machine Perception, School of EECS, Peking University. His research interests
include computer vision, image processing, machine learning, pattern
recognition, and numerical optimization. He has been an area chair of CVPR,
ICCV, NIPS/NeurIPS, AAAI, IJCAI, ICLR and ICML many times, and is a Program
Co-Chair of ICPR 2022. He was an associate editor of the IEEE Transactions on
Pattern Analysis and Machine Intelligence and currently is an associate editor
of the International Journal of Computer Vision. He is a Fellow of IAPR and
IEEE.
---|---
|
# DESY 20–214 ISSN 0418-9833
Dezember 2020
Angular analysis of bottom-flavored hadron production in semileptonic decays
of polarized top quarks
Bernd A. Kniehl111E-mail<EMAIL_ADDRESS>
II. Institut für Theoretische Physik, Universität Hamburg,
Luruper Chaussee 149, 22761 Hamburg, Germany
S. Mohammad Moosavi Nejad222E-mail<EMAIL_ADDRESS>
Faculty of Physics, Yazd University, P.O. Box 89195–741, Yazd, Iran
###### Abstract
We study the inclusive production of bottom-flavored hadrons from semileptonic
decays of polarized top quarks at next-to-leading order in QCD using
fragmentation functions recently determined from a global fit to $e^{+}e^{-}$
data. We provide the relevant differential decay widths at parton level in
analytic form. These results fill an important gap in the theoretical
interpretation of recent measurements of the top-quark polarization and the
$t\bar{t}$ spin correlations using dilepton final states in proton-proton
collisions at the CERN Large Hadron Collider. As an illustration, we study the
distributions in the scaled bottom-hadron energy of the polarized-top-quark
decay widths for different $W$-boson helicities.
PACS numbers: 12.38.Bx, 13.85.Ni, 14.40.Nd, 14.65.Ha
## 1 Introduction
The top quark $t$ of the standard model (SM) is the heaviest known elementary
particle. Due to its high mass, it plays a crucial role in testing the
electroweak symmetry breaking mechanism and in searching for new physics
beyond the SM. The precise determination of its properties, including its mass
$m_{t}$ and total decay width $\Gamma_{t}$, and its process-dependent
features, like its polarization or the correlation of its spin with that of a
co-produced antitop quark, is of prime importance. The latter quantities are
particularly sensitive probes of deviations from the SM and allow us to
constrain, e.g., the anomalous chromoelectric and chromomagnetic dipole
moments of the top quark. The top-quark lifetime
$\tau_{t}=\hbar/\Gamma_{t}\approx 5\times 10^{-25}$ s [1] is shorter than the
typical time scale of quantum chromodynamics (QCD)
$\hbar/\Lambda_{\mathrm{QCD}}\approx 10^{-24}$ s and much shorter than the
spin correlation time scale $\hbar m_{t}/\Lambda_{\mathrm{QCD}}^{2}\approx
10^{-21}$ s, where $\Lambda_{\mathrm{QCD}}$ is the asymptotic scale parameter
of QCD. Therefore, the top quark decays before it can hadronize, and its full
spin information is preserved during its decay process and fully encoded in
the angular distribution of its decay products.
The top-quark polarization and the $t\bar{t}$ spin correlations have recently
been measured using dilepton final states in Run 2 at the CERN Large Hadron
Collider (LHC) with center-of-mass energy $\sqrt{s}=13$ TeV, by the ATLAS [2]
and CMS [3] Collaborations. Intriguingly, ATLAS found a deviation of 3.8
standard deviations ($\sigma$) from the SM prediction of the asymmetry
$A_{|\Delta\phi_{\ell\ell}|}$ of the distribution in the azimuthal angular
difference $\Delta\phi_{\ell\ell}$ of the decay leptons. This deviation was
confirmed by CMS, albeit with a smaller significance of about $2\sigma$. These
analyses are relying on the factorization of the squared matrix element of the
full process, $|\mathcal{M}(q\bar{q}/gg\to t\bar{t}\to
b\ell^{+}\nu\bar{b}\ell^{-}\bar{\nu})|^{2}\propto\rho\times
R\times\bar{\rho}$, into the spin density matrices $R$, $\rho$, and
$\bar{\rho}$ for on-shell $t\bar{t}$ hadroproduction and semileptonic $t$ and
$\bar{t}$ decays, respectively, via the narrow-width approximation. While $R$
is treated in Refs. [2, 3] at next-to-leading order (NLO) in QCD [4, 5],
$\rho$ and $\bar{\rho}$ are only modeled at leading order (LO) using the
program package madspin [6]. Moreover, the formation of bottom-flavored
hadrons is not taken into account within the rigorous framework of the QCD
parton model with fragmentation functions (FFs) whose universality is
guaranteed by the factorization theorem [7]. It is an urgent matter to clarify
in how far the observed deviations may be related to a lack of precision in
the theoretical treatment of the semileptonic $t$ and $\bar{t}$ decays.
It is the purpose of the present paper to provide theoretical input needed to
fill this gap. Specifically, we calculate the partial width of the inclusive
decay$t(\uparrow)\to bW^{+}(\uparrow)\to Bl^{+}\nu_{l}+X$, where $B$
generically denotes a bottom-flavored hadron, at NLO in QCD allowing for top-
quark polarization and definite $W$-boson helicity and properly accounting for
parton-to-hadron fragmentation and finite-hadron-mass effects. By doing so, we
generalize our previous work [8], where the top-quark spin was averaged over.
The general-mass variable-flavor-number scheme (GM-VFNS), which has been
elaborated for inclusive heavy-flavored-hadron production in $e^{+}e^{-}$
annihilation [9], two-photon collisions [10], photoproduction [11], and
hadroproduction [12, 13, 14, 15], provides an ideal theoretical framework also
here. However, owing to the large mass hierarchy $m_{b}\ll m_{t}$,
finite-$m_{b}$ corrections are expected to be negligible in the case at hand.
This expectation was actually confirmed in Ref. [8], by a comparative analysis
of the partial width of the decay $t\to B+W^{+}$ in the GM-VFNS and the zero-
mass variable-flavor-number scheme (ZM-VFNS), where bottom is included among
the massless quark flavors. In fact, the finite-$m_{b}$ corrections were found
to be much smaller than the contribution from gluon fragmentation. Therefore,
we will adopt the ZM-VFNS in the following. However, we will include
finite-$m_{B}$ effects, which modify the relations between partonic and
hadronic variables and reduce the available phase space, as explained in Sec.
2 of Ref. [8].
In Ref. [8], we adopted the $B$ FFs from Ref. [16], which were determined at
NLO in the ZM-VFNS through a joint fit to $e^{+}e^{-}$ annihilation data taken
by ALEPH [17] and OPAL [18] at CERN LEP1 and by SLD [19] at SLAC SLC.
Specifically, the power ansatz
$D_{b}(z,\mu_{F}^{\text{ini}})=Nz^{\alpha}(1-z)^{\beta}$ was used as the
initial condition for the $b\to B$ FF at factorization scale
$\mu_{F}^{\text{ini}}=m_{b}=4.5$ GeV, while the gluon and light-quark FFs were
generated via the Dokshitzer-Gribov-Lipatov-Altatelli-Parisi (DGLAP) [20, 21,
22] evolution. In Ref. [23], the analysis of Ref. [16] was updated by
including the data taken by DELPHI [24] at CERN LEP1, which were published
after Ref. [16], working both at NLO and next-to-next-to-leading order (NNLO)
with the same theoretical assumptions. In our numerical analysis, we employ
the new $B$ FFs from Ref. [23].
A similar analysis, albeit without fragmentation and finite-$m_{B}$ effects,
was reported in Refs. [25, 26]. Our analysis provides an independent check of
analytic results presented therein. A related NLO analysis with a different
treatment of the final state, for bottom jets instead of bottom hadrons, was
performed in Ref. [27], leading to results that, unfortunately, cannot be
compared with ours in any straightforward way. Recently, the analysis of Refs.
[25, 26] has been extended to NNLO in QCD using the optical theorem [28]. Due
to the totally inclusive treatment of the hadronic part of the final state,
this result does not allow for the implementation of FFs.
This paper is organized as follows. In Sec. 2, we list our parton-level
results in analytic form, relegating lengthy formulas to the Appendix. In Sec.
3, we present our numerical analysis. In Sec. 4, we summarize our conclusions.
## 2 Analytic results
We work at NLO in the ZM-VFNS, implemented in the modified minimal-subtraction
($\overline{\mathrm{MS}}$) scheme, and consider the decay process
$t(p_{t},s)\to b(p_{b})+W^{+}(p_{W},\lambda)({}+g(p_{g}))\to
B(p_{B})+\ell^{+}(p_{\ell})+\nu_{\ell}(p_{\nu})+X,$ (1)
where $X$ collectively denotes the unobserved final-state hadrons and the
four-momentum, spin, and helicity assignments are indicated in parentheses. We
have $s=\pm 1/2$ and $\lambda=0,\pm 1$. The gluon in Eq. (1) contributes to
the real radiation at NLO. Both the $b$ quark and the gluon may hadronize into
the $B$ hadron. For simplicity, we employ the narrow-width approximation,
where $p_{W}^{2}=m_{W}^{2}$ and small terms of order
$\mathcal{O}(\Gamma_{W}^{2}/m_{W}^{2})$ are neglected. As mentioned in Sec. 1,
we put $m_{b}=0$, but keep $m_{B}$ finite. In the top-quark rest frame, the
$b$ quark, gluon, and $B$ hadron have energies $E_{i}=p_{t}\cdot p_{i}/m_{t}$
($i=b,g,B$), which nominally range from
$E_{b}^{\text{min}}=E_{g}^{\text{min}}=0$ and $E_{B}^{\text{min}}=m_{B}$ to
$E_{b}^{\text{max}}=E_{g}^{\text{max}}=(m_{t}^{2}-m_{W}^{2})/(2m_{t})$ and
$E_{B}^{\text{max}}=(m_{t}^{2}+m_{B}^{2}-m_{W}^{2})/(2m_{t})$, respectively.
As in Ref. [8], we choose the scaling variable $z$ by setting $E_{B}=zE_{a}$,
with $a=b,g$, in the range $0\leq z\leq 1$. Introducing the scaled energies
$x_{i}=E_{i}/E_{b}^{\text{max}}$ ($i=b,g,B$), we then have
$m_{B}/E_{b}^{\text{max}}\leq x_{B}\leq x_{a}\leq 1$. An alternative
definition of the scaling variable, in terms of light-cone variables, is
discussed in Sec. 3 of Ref. [8], to where we refer the interested reader.
We implement the polarization of the top quark by writing its (average) spin
four-vector in its rest frame as
$s_{t}^{\mu}=P(0,\sin\theta_{P}\cos\phi_{P},\sin\theta_{P}\sin\phi_{P},\cos\theta_{P})$,
where $P$ is the magnitude of the top-quark polarization, taking values in the
range $0\leq P\leq 1$. Here, it is understood that the $z$ axis is chosen to
point along the $W$-boson flight direction, so that $s_{t}\cdot
p_{W}=-P|\vec{p}_{W}|\cos\theta_{P}$. To describe the leptonic decay of the
$W$ boson, we boost into the rest frame of the latter, which leaves the $z$
axis invariant, and define the charged-lepton four-momentum to be
$p_{\ell}^{\mu}=E_{\ell}(1,\sin\theta\cos\phi,\sin\theta\sin\phi,\cos\theta)$.
The polar angles $\theta_{P}$ and $\theta$, which appear in our final results,
are also illustrated in Fig. 1.
Figure 1: Definitions of the polar angle $\theta_{P}$ of the top-quark
polarization three-vector in the top-quark rest frame and of the polar angle
$\theta$ of the charged-lepton three-momentum in the $W$-boson rest frame. In
both cases, the $z$ axis is chosen to point along the $W$-boson three-momentum
in the top-quark rest frame.
We wish to calculate the triply differential partial decay width
$d^{3}\Gamma/(dx_{B}\,d\cos\theta\,d\cos\theta_{P})$ of process (1).
Analogously to Eq. (3) in Ref. [8], we have
$\frac{d^{3}\Gamma}{dx_{B}\,d\cos\theta\,d\cos\theta_{P}}=\sum_{a=b,g}\int_{x_{B}}^{1}\frac{dx_{a}}{x_{a}}\,\frac{d^{3}\hat{\Gamma}_{a}}{dx_{a}\,d\cos\theta\,d\cos\theta_{P}}(\mu_{R},\mu_{F})D_{a}\left(\frac{x_{B}}{x_{a}},\mu_{F}\right),$
(2)
where $\mu_{R}$ is the renormalization scale, $\mu_{R}$ is the factorization
scale, $D_{a}(z,\mu_{F})$ is the $a\to B$ FF, and
$\frac{d^{3}\hat{\Gamma}_{a}}{dx_{a}\,d\cos\theta\,d\cos\theta_{P}}=\frac{1}{2}\left(\frac{d^{2}\hat{\Gamma}_{a}^{\text{unpol}}}{dx_{a}\,d\cos\theta}+P\frac{d^{2}\hat{\Gamma}_{a}^{\text{pol}}}{dx_{a}\,d\cos\theta}\cos\theta_{P}\right).$
(3)
The factor $1/2$ on the right-hand side of Eq. (3) ensures that
$\int_{-1}^{1}d\cos\theta_{P}\,\frac{d^{3}\hat{\Gamma}}{dx_{a}\,d\cos\theta\,d\cos\theta_{P}}=\frac{d^{2}\hat{\Gamma}_{a}^{\text{unpol}}}{dx_{a}\,d\cos\theta}.$
(4)
Each $W$-boson helicity is featured by a characteristic $\theta$ dependence,
which is encoded in the structure
$\frac{d^{2}\hat{\Gamma}_{a}}{dx_{a}\,d\cos\theta}=\frac{3}{8}(1+\cos\theta)^{2}\frac{d\hat{\Gamma}_{a}^{+}}{dx_{a}}+\frac{3}{8}(1-\cos\theta)^{2}\frac{d\hat{\Gamma}_{a}^{-}}{dx_{a}}+\frac{3}{4}\sin^{2}\theta\frac{d\hat{\Gamma}_{a}^{0}}{dx_{a}}.$
(5)
This holds for both the unpolarized and polarized terms in Eq. (3). Notice
that the $\theta$-dependent coefficients on the right-hand side of Eq. (5) are
normalized so that, upon integration over $\cos\theta$, we have
$\frac{d\hat{\Gamma}_{a}}{dx_{a}}=\sum_{\lambda=-1}^{1}\frac{d\hat{\Gamma}_{a}^{\lambda}}{dx_{a}}.$
(6)
At LO, we only have $a=b$, and $x_{b}=1$ is fixed, i.e. the $x_{b}$ dependence
comes as a delta-function peak. Specifically, we have
$\displaystyle\hat{\Gamma}_{b,\text{LO}}^{0,\text{unpol}}$ $\displaystyle=$
$\displaystyle\hat{\Gamma}_{b,\text{LO}}^{0,\text{pol}}=F(1-\omega)^{2},$
$\displaystyle\hat{\Gamma}_{b,\text{LO}}^{-,\text{unpol}}$ $\displaystyle=$
$\displaystyle-\hat{\Gamma}_{b,\text{LO}}^{-,\text{pol}}=F(2\omega)(1-\omega)^{2},$
$\displaystyle\hat{\Gamma}_{b,\text{LO}}^{+,\text{unpol}}$ $\displaystyle=$
$\displaystyle\hat{\Gamma}_{b,\text{LO}}^{+,\text{pol}}=0,$ (7)
where $\omega=m_{W}^{2}/m_{t}^{2}$ and
$F=G_{F}m_{t}^{3}|V_{tb}|^{2}B(W^{+}\to\ell^{+}\nu_{\ell})/(8\pi\sqrt{2})$.
Here, $G_{F}$ is Fermi’s constant, $V_{ij}$ is the $ij$ element of the
Cabibbo-Kobayashi-Matrix quark mixing matrix [29, 30], and
$B(W^{+}\to\ell^{+}\nu_{\ell})$ is the branching ratio of the leptonic
$W$-boson decay mode considered. Neglecting the masses of the charged leptons
and the first five quark flavors, we have
$\displaystyle B(W^{+}\to\ell^{+}\nu_{\ell})$ $\displaystyle=$
$\displaystyle\frac{1}{3+N_{c}\sum_{\begin{subarray}{c}i=u,c\\\
j=d,s,b\end{subarray}}|V_{ij}|^{2}\left[1+3C_{F}\alpha_{s}(\mu_{R})/(4\pi)\right]}$
(8) $\displaystyle\approx$
$\displaystyle\frac{1}{9}\left[1-\frac{2}{3}\,\frac{\alpha_{s}(\mu_{R})}{\pi}\right],$
where we have included the NLO QCD correction, with color factors $N_{c}=3$
and $C_{F}=(N_{c}^{2}-1)/(2N_{c})=4/3$. In the last equality in Eq. (8), we
have approximated $V_{ij}\approx\delta_{ij}$. At LO, the top-quark spin is
passed on to the bottom quark for $\lambda=0$, while it is flipped for
$\lambda=-1$; the case $\lambda=+1$ is forbidden by angular-momentum
conservation in the limit $m_{b}\to 0$, as is reflected in Eq. (7). Using Eq.
(6), we have
$\displaystyle\hat{\Gamma}_{b,\text{LO}}^{\text{unpol}}$ $\displaystyle=$
$\displaystyle F(1+2\omega)(1-\omega)^{2},$
$\displaystyle\hat{\Gamma}_{b,\text{LO}}^{\text{pol}}$ $\displaystyle=$
$\displaystyle F(1-2\omega)(1-\omega)^{2}.$ (9)
Inserting Eq. (7) in Eq. (2), we obtain a very simple formula for the final LO
result:
$\displaystyle\frac{d^{3}\Gamma_{\text{LO}}}{dx_{B}\,d\cos\theta\,d\cos\theta_{P}}$
$\displaystyle=$
$\displaystyle\frac{3}{8}F(1-\omega)^{2}D_{b}(x_{B},\mu_{F})[\sin\theta(1+P\cos\theta_{P})$
(10) $\displaystyle{}+\omega(1-\cos\theta)^{2}(1-P\cos\theta_{P})].$
Figure 2: Feynman diagrams contributing to the partial decay width of the
process in Eq. (1) at NLO: (a) initial-state radiation; (b) final-state
radiation; (c) vertex correction; and (d) combination of wave function
renormalizations and vertex counterterm. The leptonic $W$-boson decay is not
shown.
The Feynman diagrams contributing to the partial decay width of the process in
Eq. (1) at NLO are depicted in Fig. 2. The NLO coefficient functions of the
unpolarized case may be found in Appendix A of Ref. [8] and those of the
polarized case are presented in the Appendix of this paper. At NLO, the case
$\lambda=+1$ is enabled by the presence of the additional spin-one gluon even
for $m_{b}=0$.
At this point, we compare our new analytic results with the literature. In
Refs. [25, 26], process (1) was also considered at NLO using the narrow-width
approximation and putting $m_{b}=0$, but treating the final state in a less
differential fashion, which does not allow for the convolution with $a\to B$
FFs on the basis of the ZM-VFNS. We can compare our results for
$\hat{\Gamma}_{b}^{\lambda,\text{pol}}$ with Refs. [25, 26] upon integration
over $x_{b}$ in the range $0\leq x_{b}\leq 1$. Specifically, the quantities
$\hat{\Gamma}_{U}^{P}$, $\hat{\Gamma}_{L}^{P}$, and $\hat{\Gamma}_{F}^{P}$
listed in Eqs. (18)–(20) of Ref. [25] (see also Eqs. (42)–(44) in Ref. [26])
are related to $\hat{\Gamma}_{b}^{\lambda,\text{pol}}$ as
$\hat{\Gamma}_{b}^{0,\text{pol}}/\hat{\Gamma}_{b,\text{LO}}^{\text{unpol}}=\hat{\Gamma}_{L}^{P}$
and
$\hat{\Gamma}_{b}^{\pm,\text{pol}}/\hat{\Gamma}_{b,\text{LO}}^{\text{unpol}}=(\hat{\Gamma}_{U}^{P}\pm\hat{\Gamma}_{F}^{P})/2$.
With these identifications, we fully agree with Refs. [25, 26].
## 3 Numerical results
We are now in a position to explore the phenomenological consequences of our
results by performing a numerical analysis. We adopt from Ref. [1] the input
parameter values $G_{F}=1.1663787\times 10^{-5}$ GeV-2, $m_{W}=80.379$ GeV,
$m_{t}=172.4$ GeV, $m_{B}=5.279$ GeV, $|V_{tb}|=1$, and
$B(W^{+}\to\ell^{+}\nu_{\ell})=10.86\%$. We evaluate
$\alpha_{s}^{(n_{f})}(\mu_{R})$ at NLO (NNLO) in the $\overline{\text{MS}}$
scheme using Eq. (4) of Ref. [31], retaining only the first two (three) terms
on the right-hand side, with $n_{f}=5$ active quark flavors and asymptotic
scale parameter $\Lambda_{\overline{\text{MS}}}^{(5)}=225$ MeV (207 MeV)
adjusted such that $\alpha_{s}^{(5)}(m_{Z})=0.1179$ for $m_{Z}=91.1876$ GeV
[1]. This yields $\alpha_{s}^{(5)}(m_{t})=0.1076$ (0.1076). As already
mentioned in Sec. 1, we use the up-to-date $B$ FFs from Ref. [23], both at NLO
and NNLO. For definiteness, we identify $\mu_{R}=\mu_{F}=\xi m_{t}$ and vary
$\xi$ from 1/2 to 2 about the default value 1 to estimate the theoretical
uncertainty due to the lack of knowledge of higher-order corrections.
The angular dependencies at the parton level, in Eqs. (3) and (5), are passed
on to the hadron level via Eq. (2). The hadron level counterparts,
$d\Gamma^{\lambda}/dx_{B}$, of the coefficient functions
$d\hat{\Gamma}_{a}^{\lambda}/dx_{a}$ in Eq. (5) may be projected out from the
measured $\theta$ distribution $d\Gamma/(dx_{B}\,d\cos\theta)$ as explained in
Sec. 4 of Ref. [8] (see Eqs. (21)–(23) therein) and thus represent physical
observables by themselves. In Ref. [8], the top quarks were taken to be
unpolarized, i.e., according to our present notation,
$d\Gamma^{\lambda,\text{unpol}}/dx_{B}$ were considered. In the following, we
complement the study of Ref. [8] by presenting predictions for
$d\Gamma^{\lambda,\text{pol}}/dx_{B}$. For the sake of a coherent treatment,
we also provide the analogous predictions for
$d\Gamma^{\lambda,\text{unpol}}/dx_{B}$, thus updating the analysis of Ref.
[8].
Our central predictions are of NLO. To assess their significance, we compare
them with the respective LO results. Our LO predictions are slightly
inconsistent because they are evaluated with NLO FFs. Unfortunately, Ref. [23]
does not provide a LO FF set. By the way, the same is true for Ref. [16], to
which we could have resorted otherwise. On the other hand, Ref. [23] also
supplies a NNLO set. While consistent NNLO predictions are unfeasible in the
absence of NNLO parton-level results, this still offers us the opportunity to
get a first impression of the typical magnitude of the NNLO effects. In our
pseudo-NNLO analysis, besides using NNLO FFs, we also evaluate
$\alpha_{s}(\mu_{R})$ at NNLO as explained above.
Figure 3: LO and NLO results for (a) $d\Gamma^{\lambda,\text{unpol}}/dx_{B}$
and (b) $d\Gamma^{\lambda,\text{pol}}/dx_{B}$ with $\lambda=0,\pm 1$ as
functions of $x_{B}$. $d\Gamma_{\text{NLO}}^{+,\text{unpol}}/dx_{B}$ and
$d\Gamma_{\text{NLO}}^{+,\text{pol}}/dx_{B}$ are rescaled by a factor of 10
for better visibility. The theoretical uncertainties are indicated by the
shaded bands.
Figure 4: NLO results for (a) $d\Gamma^{\lambda,\text{unpol}}/dx_{B}$ and (b)
$d\Gamma^{\lambda,\text{pol}}/dx_{B}$ with $\lambda=0,-1$, normalized to the
respective default LO results, as functions of $x_{B}$. The theoretical
uncertainties of the NLO results are indicated by the shaded bands.
In Fig. 3, we study the $x_{B}$ dependencies of
$d\Gamma^{\lambda,\text{unpol}}/dx_{B}$ and
$d\Gamma^{\lambda,\text{pol}}/dx_{B}$ for $\lambda=0,\pm 1$ at NLO and compare
them with the respective LO results for $\lambda=0,-1$. As explained in Sec.
2, $d\Gamma^{+,\text{unpol}}/dx_{B}$ and $d\Gamma^{+,\text{pol}}/dx_{B}$
vanish at LO in our approximation. The theoretical uncertainties are indicated
by the shaded bands. The slight inconsistency in our LO analysis mentioned
above, not only affects the default predictions, but also the estimation of
the theoretical uncertainty, which is expected to be slightly larger if the
$\mu_{F}$ dependence is subject to LO DGLAP evolution. Comparing Figs. 3(a)
and (b), we observe that $d\Gamma^{\lambda,\text{unpol}}/dx_{B}$ and
$d\Gamma^{\lambda,\text{pol}}/dx_{B}$ are very similar as for normalization
and line shape. Each $x_{B}$ distribution exhibits a maximum close to
$x_{B}=0.8$. Longitudinal $W$-boson helicity is favored, with
$d\Gamma_{\text{LO}/\text{NLO}}^{0,\text{unpol}/\text{pol}}/dx_{B}$ being more
than twice as large as
$d\Gamma_{\text{LO}/\text{NLO}}^{-,\text{unpol}/\text{pol}}/dx_{B}$ for
negative $W$-boson helicity $\lambda=-1$. On the other hand, positive
$W$-boson helicity $\lambda=+1$ is perturbatively suppressed; rescaling
$d\Gamma_{\text{NLO}}^{+,\text{unpol}/\text{pol}}/dx_{B}$ by the inverse
couplant, $2\pi/\alpha_{s}^{(5)}(m_{t})$, brings it up to the level of
$d\Gamma_{\text{NLO}}^{-,\text{unpol}/\text{pol}}/dx_{B}$. The NLO corrections
have a significant effect on the $x_{B}$ distributions for $\lambda=0,-1$, by
raising their peaks and lowering their small-$x_{B}$ tails. To render these
features more visible, we present, in Fig. 4, the QCD correction ($K$) factors
for $\lambda=0,-1$, which we evaluate by normalizing the NLO predictions
including their theoretical-uncertainty bands relative to the default LO
predictions. From Fig. 4(a) and (b), we observe that, both for unpolarized and
polarized top quarks, the $K$ factors steadily increase by one order of
magnitude, typically from 0.2 to 2, as $x_{B}$ runs across its range of
values. We conclude from Figs. 3 and 4 that the NLO corrections are quite
significant and should be taken into account in theoretical interpretations of
future top-quark polarization measurements.
Figure 5: NLO and pseudo-NNLO results for (a)
$d\Gamma^{\lambda,\text{unpol}}/dx_{B}$ and (b)
$d\Gamma^{\lambda,\text{pol}}/dx_{B}$ with $\lambda=0,\pm 1$ as functions of
$x_{B}$. $d\Gamma_{\text{NLO}/\text{NNLO}}^{+,\text{unpol}}/dx_{B}$ and
$d\Gamma_{\text{NLO}/\text{NNLO}}^{+,\text{pol}}/dx_{B}$ are rescaled by a
factor of 10 for better visibility. The theoretical uncertainties are
indicated by the shaded bands.
Figure 6: Pseudo-NNLO results for (a) $d\Gamma^{\lambda,\text{unpol}}/dx_{B}$
and (b) $d\Gamma^{\lambda,\text{pol}}/dx_{B}$ with $\lambda=0,\pm 1$,
normalized to the respective default NLO results, as functions of $x_{B}$. The
theoretical uncertainties of the pseudo-NNLO results are indicated by the
shaded bands.
In Fig. 5, compare our pseudo-NNLO evaluations of
$d\Gamma^{\lambda,\text{unpol}}/dx_{B}$ and
$d\Gamma^{\lambda,\text{pol}}/dx_{B}$ for $\lambda=0,\pm 1$ with the
respective NLO results already presented in Fig. 3. For better visibility, we
also plot, in Fig. 6, the ratios of the pseudo-NNLO results including their
theoretical-uncertainty bands and the default NLO predictions. We conclude
from Figs. 5 and 6 that the NNLO effects are likely to be relatively modest,
of the order of 10% or less.
## 4 Summary and Conclusions
We studied the inclusive production of bottom-flavored hadrons from
semileptonic decays of polarized top quarks at NLO in the ZM-FVNS using the
narrow-width approximation for the intermediate $W$ bosons, whose helicities
we distinguished. Specifically, we considered the partial decay width
differential in the scaled bottom-hadron energy $x_{B}$, the azimuthal angle
$\theta$ of the charged lepton in the $W$-boson rest frame, and the azimuthal
angle $\theta_{P}$ of the top-quark polarization in the top-quark rest frame.
In our numerical analysis, we employed up-to-date $B$ FFs, recently determined
from a global fit to all available $e^{+}e^{-}$ data [23]. We thus extended
our previous study in Ref. [8], which was restricted to unpolarized top
quarks. In Ref. [8], we had convinced ourselves, by comparing evaluations in
the GM-VFNS and ZM-VFNS, that finite-$m_{b}$ corrections may be safely
neglected. On the other hand, we retained finite-$m_{B}$ corrections, which
reduce the available phase space and lead to visible effects at small values
of $x_{B}$. We provided full analytic results, ready to be used by the
interested reader.
We found the NLO QCD corrections to be quite significant, inducing a reduction
in the lower $x_{B}$ range and an enhancement in the upper $x_{B}$ range, with
$K$ factors ranging from 0.2 to 2. On the other hand, including partial
information from NNLO, contained in the evaluation of the $B$ FFs and
$\alpha_{s}(\mu_{R})$, turned out to yield only mild modifications, with the
due caveat that such results suffer from a violation of renormalization group
invariance already in the considered order.
Our combined results, from Ref. [8] and this paper, help us to fill an
important gap in the theoretical interpretation of recent measurements of the
top-quark polarization and the $t\bar{t}$ spin correlations in dilepton final
states at the LHC [2, 3]. In particular, it will be interesting to see if
these theoretical improvements will contribute to reconciling the ATLAS [2]
and CMS [3] measurements of $A_{|\Delta\phi_{\ell\ell}|}$ with the SM
expectations.
## Acknowledgments
We thank G. Kramer for useful discussions at the initial stage of this work
and C. Schwanenberger for detailed information on Ref. [3]. The work of B.A.K.
was supported in part by the German Federal Ministry for Education and
Research BMBF through Grant No. 05H18GUCC1. The work of S.M.M.N. was supported
in part by the Iran National Science Foundation INSF through Grant No.
97005414.
## Appendix
In this appendix, we list the coefficient functions
$d\hat{\Gamma}_{a}^{\lambda,\text{pol}}/dx_{a}$ appearing in Eq. (5) at NLO in
the ZM-VFNS with $m_{b}=0$. They possess the following structure:
$\displaystyle\frac{1}{\hat{\Gamma}_{b,\text{LO}}^{\text{pol}}}\,\frac{d\hat{\Gamma}_{a}^{0,\text{pol}}}{dx_{a}}$
$\displaystyle=$
$\displaystyle\frac{1}{1-2\omega}\left[\delta_{ab}\delta(1-x_{a})+\frac{\alpha_{s}(\mu_{R})}{2\pi}\left(P_{ab}(x_{a})\ln\frac{m_{t}^{2}}{\mu_{F}^{2}}+C_{F}C_{a}^{0,\text{pol}}(x_{a})\right)\right],$
$\displaystyle\frac{1}{\hat{\Gamma}_{b,\text{LO}}^{\text{pol}}}\,\frac{d\hat{\Gamma}_{a}^{-,\text{pol}}}{dx_{a}}$
$\displaystyle=$
$\displaystyle\frac{-2\omega}{1-2\omega}\left[\delta_{ab}\delta(1-x_{a})+\frac{\alpha_{s}(\mu_{R})}{2\pi}\left(P_{ab}(x_{a})\ln\frac{m_{t}^{2}}{\mu_{F}^{2}}+C_{F}C_{a}^{-,\text{pol}}(x_{a})\right)\right],$
$\displaystyle\frac{1}{\hat{\Gamma}_{b,\text{LO}}^{\text{pol}}}\,\frac{d\hat{\Gamma}_{a}^{+,\text{pol}}}{dx_{a}}$
$\displaystyle=$
$\displaystyle\frac{4\omega}{(1-2\omega)(1-\omega)^{2}}\frac{\alpha_{s}(\mu_{R})}{2\pi}C_{F}C_{a}^{+,\text{pol}}(x_{a}),$
(11)
where
$\displaystyle P_{qq}(x)$ $\displaystyle=$ $\displaystyle
C_{F}\left(\frac{1+x^{2}}{1-x}\right)_{+},$ $\displaystyle P_{gq}(x)$
$\displaystyle=$ $\displaystyle C_{F}\frac{1+(1-x)^{2}}{x}$ (12)
are the timelike $q\to q$ and $q\to g$ splitting functions at LO.
For $a=b$, we have
$\displaystyle C_{b}^{0,\text{pol}}(x)$ $\displaystyle=$
$\displaystyle-\delta(1-x)\big{[}2\ln\omega\ln(1-\omega)+4\mathop{\mathrm{Li}}\nolimits_{2}(\omega)+\frac{2\omega}{1-\omega}\ln\omega+6\big{]}+2(1+x^{2})\bigg{(}\frac{\ln(1-x)}{1-x}\bigg{)}_{+}$
(13)
$\displaystyle{}+2\frac{1+x^{2}}{(1-x)_{+}}\ln(x(1-\omega))+\frac{8(x-2)(1-x)^{2}}{x^{2}(Sx^{2}-2x+2)}+2\omega(1-x)-\frac{12}{x}+\frac{8}{x^{2}}+\frac{6}{1-x}-\frac{(1+x)^{2}}{1-x}-\frac{1+x^{2}}{1-x}R_{1}$
$\displaystyle{}+\bigg{(}1-|1-2x+2Sx^{2}|\bigg{)}\bigg{(}6\frac{1-x}{x^{2}}-\frac{1}{(1-x)(1-\omega)}+6\frac{(1-x)^{2}(x-2)^{2}}{x^{2}(Sx^{2}-2x+2)^{2}}+\frac{1}{2x^{2}(Sx^{2}-2x+2)}\big{[}9x^{3}$
$\displaystyle{}-57x^{2}+97x-\frac{1}{1-x}-47\big{]}\bigg{)}+\frac{R_{2}}{S^{3/2}(Sx^{2}-2x+2)^{5/2}}\bigg{(}x^{3}S^{4}(-x^{3}+7x^{2}-22x+10+\frac{2}{1-x})+xS^{3}(x^{4}$
$\displaystyle{}+4x^{3}+25x^{2}-32x+12)-2(1-x)^{2}+S^{2}(1-x)(5x^{3}+21x^{2}-12x+4)-S(1-x)(9x^{2}-4x-4)\bigg{)},$
$\displaystyle C_{b}^{-,\text{pol}}(x)$ $\displaystyle=$
$\displaystyle-\delta(1-x)\big{[}2\ln\omega\ln(1-\omega)+4\mathop{\mathrm{Li}}\nolimits_{2}(\omega)+\frac{2\omega}{1-\omega}\ln\omega+6+\frac{1-\omega}{\omega}\ln(1-\omega)\big{]}+2(1+x^{2})\bigg{(}\frac{\ln(1-x)}{1-x}\bigg{)}_{+}$
(14)
$\displaystyle{}+2\frac{1+x^{2}}{(1-x)_{+}}\ln(x(1-\omega))-\frac{1+x^{2}}{2(1-x)}R_{1}+\frac{(1-S)^{2}}{\omega
S}+\frac{\omega+Sx}{\omega}-\frac{\omega}{2S(1-x)}+\frac{(1-S-Sx)}{S}\ln(1-2Sx)$
$\displaystyle{}-\frac{3S(2x-1)+6(1-x)}{2S(Sx^{2}-2x+2)^{2}}+\frac{S^{2}(1+3x)+4x\omega-3S}{2\omega
S(Sx^{2}-2x+2)}+B_{3}|1-2x+2Sx^{2}|+B_{1}\ln|2Sx^{2}-2x+1|$
$\displaystyle{}+\frac{R_{3}}{S^{2}\sqrt{\omega}}B_{2}-\frac{R_{2}}{2\sqrt{Sx^{2}-2x+2}}\bigg{(}x^{2}\sqrt{S}+\frac{(S-1)x}{\sqrt{S}}-\frac{2\sqrt{S}}{1-x}+\frac{2S^{2}-5S+7}{S^{\frac{3}{2}}}$
$\displaystyle{}+\frac{Sx(4S^{2}-23S+16)-S(28-19S)+6}{S^{\frac{5}{2}}(Sx^{2}-2x+2)}-\frac{3Sx(12-9S+S^{2})+15S(S-2)+12(1-x)}{S^{\frac{5}{2}}(Sx^{2}-2x+2)^{2}}\bigg{)},$
$\displaystyle C_{b}^{+,\text{pol}}(x)$ $\displaystyle=$
$\displaystyle\frac{(1+x^{2})R_{1}}{1-x}S^{2}-\frac{2(1+x)S^{3}}{\omega}+2S\big{[}1+Sx+(1-S-Sx)\ln(1-2Sx)\big{]}$
(15)
$\displaystyle{}-\frac{(17-21x)S^{3}+(20x-27)S^{2}+4(2-x)S}{\omega(Sx^{2}-2x+2)}+\frac{6(1-x)S-3(1-2x)S^{2}}{(Sx^{2}-2x+2)^{2}}+\frac{\omega
S}{1-x}+\frac{2R_{3}}{\sqrt{\omega}}B_{2}$
$\displaystyle{}-2S^{2}B_{3}|1-2x+2Sx^{2}|+2S^{2}B_{1}\ln|2Sx^{2}-2x+1|-\frac{R_{2}(1-x)\sqrt{S}}{x^{3}\sqrt{Sx^{2}-2x+2}}\bigg{(}11+7x^{2}-16x$
$\displaystyle{}+\frac{6(2-x)(1-x)^{3}}{(2-2x+Sx^{2})^{2}}+\frac{16x^{3}-58x^{2}+69x-28}{2-2x+Sx^{2}}+\frac{Sx^{2}(x^{2}+5x-4)}{1-x}+\frac{S^{2}x^{4}(1+x^{2})}{(1-x)^{2}}\bigg{)},$
where $S=(1-\omega)/2$,
$\displaystyle R_{1}$ $\displaystyle=$
$\displaystyle\ln\bigg{[}(1-S)x^{2}-x+\frac{1}{2}+\frac{1}{2}|2Sx^{2}-2x+1|\bigg{]},$
$\displaystyle R_{2}$ $\displaystyle=$
$\displaystyle\ln\bigg{[}(1-x)(1-3Sx)+Sx^{2}(1-2Sx)+\sqrt{S(Sx^{2}-2x+2)}|2Sx^{2}-2x+1|\bigg{]}-$
$\displaystyle\ln\bigg{[}1+(S-1)x+\sqrt{S(Sx^{2}-2x+2)}\bigg{]},$
$\displaystyle R_{3}$ $\displaystyle=$
$\displaystyle\ln\big{[}1-x(1-\sqrt{\omega})\big{]}-\ln\big{|}1-x(1+\sqrt{\omega})|,$
$\displaystyle B_{1}$ $\displaystyle=$ $\displaystyle
x-\frac{1-S}{S}-\frac{1-x}{2S(2-2x+Sx^{2})^{2}}\big{[}5S^{2}x^{2}-2(1-x)-2S(x^{2}+3x-3)\big{]},$
$\displaystyle B_{2}$ $\displaystyle=$ $\displaystyle-\omega
S+xS^{2}+\frac{Sx(-4S^{2}+7S-2)+S(8-10S)-2}{2(Sx^{2}-2x+2)}-\frac{\omega(S^{2}x-4Sx+3S+2x-2)}{(Sx^{2}-2x+2)^{2}},$
$\displaystyle B_{3}$ $\displaystyle=$
$\displaystyle\frac{1}{2S(1-x)}+\frac{1}{1-2Sx}+\frac{6(1-x)-3S(1-2x)}{2S(Sx^{2}-2x+2)^{2}}+\frac{S^{2}(3x-5)+S(11-2x)-4}{2\omega
S(Sx^{2}-2x+2)}.$ (16)
For $a=g$, we have
$\displaystyle C_{g}^{0,\text{pol}}(x)$ $\displaystyle=$
$\displaystyle\frac{1+(1-x)^{2}}{x}\bigg{(}-R_{1}+2\ln[x(1-\omega)(1-x)]\bigg{)}+\bigg{(}8S-x-26+\frac{13}{S}+\frac{26S^{2}-38S+13}{2S^{3}x^{2}}$
(17)
$\displaystyle{}+\frac{-44S^{2}+88S-35}{2xS^{2}}\bigg{)}R_{4}+\bigg{(}\frac{1-6S}{2(2Sx-1)}+9\frac{1-S}{2Sx}+\frac{-4S^{2}+25S-13}{2S^{2}x^{2}}+\frac{\omega}{2(1-2Sx)^{2}}\bigg{)}|2Sx^{2}-2x+1|$
$\displaystyle{}+2(7-6S)+(1-4S)x+7\frac{7S-5}{2Sx}+\frac{4S^{2}-25S+13}{2S^{2}x^{2}},$
$\displaystyle C_{g}^{-,\text{pol}}(x)$ $\displaystyle=$
$\displaystyle\frac{1+(1-x)^{2}}{2x}\bigg{(}4\ln[x(1-\omega)(1-x)]-R_{1}\bigg{)}+\frac{1-6S}{4\omega}x-\frac{\omega^{2}}{32S^{3}x^{2}(1-2Sx)^{2}}-\frac{(10S-7)\omega}{32S^{3}x^{2}(1-2Sx)}$
(18)
$\displaystyle{}+\frac{1+5S-8S^{2}}{2S\omega}-A_{3}R_{4}+\frac{|2Sx^{2}-2x+1|}{8\omega
S^{2}x^{2}}A_{5}+\frac{A_{2}}{2S^{3}x^{2}}\ln(1-2Sx)+\frac{R_{3}A_{4}}{2\sqrt{\omega}S^{3}x^{2}}-A_{1}\frac{\ln|2Sx^{2}-2x+1|}{2x^{2}S^{3}}$
$\displaystyle{}+\frac{1}{16\omega
xS^{2}}(48S^{3}+72S^{2}-50S-5)+\frac{1}{16\omega
S^{3}x^{2}}(32S^{3}-88S^{2}+44S-3),$ $\displaystyle C_{g}^{+,\text{pol}}(x)$
$\displaystyle=$
$\displaystyle\frac{1}{2Sx^{2}}\Big{[}\frac{2R_{3}}{\sqrt{\omega}}A_{4}+2x(1+(1-x)^{2})S^{3}R_{1}+4S^{3}x^{2}A_{3}R_{4}+2A_{2}\ln(1-2Sx)-2A_{1}\ln|2Sx^{2}-2x+1|$
(19)
$\displaystyle{}-\frac{SA_{5}}{2\omega}|2Sx^{2}-2x+1|+\frac{S}{4\omega}\bigg{(}4S^{2}(1+2S)x^{3}+8S(4S^{2}-7S+1)x^{2}+\frac{8\omega^{2}x}{S(1-4S)}-(96S^{2}-70S+5)x$
$\displaystyle{}+\frac{32S^{2}-12S-3}{S}-\frac{40S^{3}-68S^{2}+38S-7}{2S(1-2Sx)}+\frac{8S^{3}-12S^{2}+6S-1}{2S(1-2Sx)^{2}}\bigg{)}\Big{]},$
where
$\displaystyle R_{4}$ $\displaystyle=$
$\displaystyle\ln\bigg{[}1-S(-2Sx^{2}+2x+1-|2Sx^{2}-2x+1|)\bigg{]}-\ln[1-2Sx],$
$\displaystyle A_{1}$ $\displaystyle=$ $\displaystyle
2x^{2}(x-1)S^{3}+(-2x^{2}+5x+1)S^{2}-2(2+x)S+2,$ $\displaystyle A_{2}$
$\displaystyle=$ $\displaystyle
x(x^{2}-2)S^{3}+(-2x^{2}+5x+1)S^{2}-2(2+x)S+2,$ $\displaystyle A_{3}$
$\displaystyle=$
$\displaystyle\frac{2+x^{2}}{2x}-\frac{5x^{2}+5x+1}{2Sx^{2}}-\frac{7-S(8+15x)}{4x^{2}S^{3}},$
$\displaystyle A_{4}$ $\displaystyle=$ $\displaystyle
2x(2-x^{2})S^{3}+(2x^{2}-7x-4)S^{2}+2(x+3)S-2,$ $\displaystyle A_{5}$
$\displaystyle=$ $\displaystyle
20S-11-Sx(10S-7)-\frac{\omega(7-10S)}{2(1-2Sx)}+\frac{\omega^{2}}{2(1-2Sx)^{2}}.$
(20)
Adding Eqs. (13)–(15) for $a=b$ and Eqs. (17)–(19) for $a=g$ according to Eq.
(6), we obtain the respective coefficient functions pertaining to the case
where $\theta$ is integrated over. Specifically, we have
$\displaystyle\frac{1}{\hat{\Gamma}_{b,\text{LO}}^{\text{pol}}}\,\frac{d\hat{\Gamma}_{b}^{\text{pol}}}{dx_{b}}$
$\displaystyle=$
$\displaystyle\delta(1-x_{b})+\frac{\alpha_{s}(\mu_{R})}{2\pi}\left(P_{qq}(x_{b})\ln\frac{m_{t}^{2}}{\mu_{F}^{2}}+C_{F}C_{b}^{\text{pol}}(x_{b})\right),$
$\displaystyle\frac{1}{\hat{\Gamma}_{b,\text{LO}}^{\text{pol}}}\,\frac{d\hat{\Gamma}_{g}^{\text{pol}}}{dx_{g}}$
$\displaystyle=$
$\displaystyle\frac{\alpha_{s}(\mu_{R})}{2\pi}\left(P_{gq}(x_{g})\ln\frac{m_{t}^{2}}{\mu_{F}^{2}}+C_{F}C_{g}^{\text{pol}}(x_{g})\right),$
(21)
where
$\displaystyle C_{b}^{\text{pol}}(x)$ $\displaystyle=$
$\displaystyle\delta(1-x)\big{[}-2\ln\omega\ln(1-\omega)+\frac{2(1-\omega)}{1-2\omega}\ln(1-\omega)-4\mathop{\mathrm{Li}}\nolimits_{2}(\omega)-\frac{2\omega}{1-\omega}\ln\omega-6\big{]}$
$\displaystyle{}+2(1+x^{2})\bigg{(}\frac{\ln(1-x)}{1-x}\bigg{)}_{+}+2\frac{1+x^{2}}{(1-x)_{+}}\ln(x(1-\omega))+\frac{1-S}{S(Sx^{2}-2x+2)}-\frac{2}{1-4S}-\frac{\omega}{S(1-x)}-1-x$
$\displaystyle{}-\frac{1+x^{2}}{1-x}R_{1}+\bigg{(}\frac{1}{S(1-x)}+\frac{4\omega}{(1-4S)(1-2Sx)}-\frac{1-S}{S(Sx^{2}-2x+2)}\bigg{)}|2Sx^{2}-2x+1|$
$\displaystyle{}-\frac{R_{2}}{\sqrt{S(Sx^{2}-2x+2)}}\bigg{(}\frac{1}{S}+\frac{2}{1-4S}+\frac{(13S-5-4S^{2})x}{1-4S}-\frac{Sx(x^{2}-x+2)}{1-x}-\frac{(1-S)[1-(1-S)x]}{S(Sx^{2}-2x+2)}\bigg{)},$
$\displaystyle C_{g}^{\text{pol}}(x)$ $\displaystyle=$
$\displaystyle\frac{1+(1-x)^{2}}{x}\bigg{(}-R_{1}+2\ln[x(1-\omega)(1-x)]\bigg{)}-\bigg{(}x-\frac{1+\omega^{2}}{4S^{3}x^{2}}-\frac{8S^{2}-6S+3}{S(4S-1)}-\frac{2-\omega^{2}(2\omega-5)}{2S^{2}(1-4S)x}\bigg{)}R_{4}$
(22)
$\displaystyle{}+\bigg{(}\frac{S-1}{2S^{2}x^{2}}+\frac{\omega}{2(1-2Sx)^{2}}-\frac{12S^{2}-15S+7}{2S(1-4S)x}-\frac{24S^{2}-26S+9}{2(1-4S)(1-2Sx)}\bigg{)}|2Sx^{2}-2x+1|+x+\frac{1-S}{2S^{2}x^{2}}$
$\displaystyle{}-\frac{2}{1-4S}+\frac{12S^{2}-7S+5}{2S(1-4S)x}.$
## References
* [1] P. A. Zyla et al. [Particle Data Group], “Review of Particle Physics,” Prog. Theor. Exp. Phys. 2020 (2020) 083C01 (2092 pages) doi:10.1093/ptep/ptaa104
* [2] M. Aaboud et al. [ATLAS Collaboration], “Measurements of top-quark pair spin correlations in the $e\mu$ channel at $\sqrt{s}=13$ TeV using $pp$ collisions in the ATLAS detector,” Eur. Phys. J. C 80 (2020) 754 (43 pages) doi:10.1140/epjc/s10052-020-8181-6 [arXiv:1903.07570 [hep-ex]].
* [3] A. M. Sirunyan et al. [CMS Collaboration], “Measurement of the top quark polarization and $\mathrm{t\bar{t}}$ spin correlations using dilepton final states in proton-proton collisions at $\sqrt{s}=$ 13 TeV,” Phys. Rev. D 100 (2019) 072002 (39 pages) doi:10.1103/PhysRevD.100.072002 [arXiv:1907.03729 [hep-ex]].
* [4] W. Bernreuther and Z.-G. Si, “Top quark spin correlations and polarization at the LHC: Standard model predictions and effects of anomalous top chromo moments,” Phys. Lett. B 725 (2013) 115–122 Erratum: [Phys. Lett. B 744 (2015) 413] doi:10.1016/j.physletb.2013.06.051, 10.1016/j.physletb.2015.03.035 [arXiv:1305.2066 [hep-ph]].
* [5] W. Bernreuther, D. Heisler and Z.-G. Si, “A set of top quark spin correlation and polarization observables for the LHC: Standard Model predictions and new physics contributions,” J. High Energy Phys. 12 (2015) 026 (36 pages) doi:10.1007/JHEP12(2015)026 [arXiv:1508.05271 [hep-ph]].
* [6] P. Artoisenet, R. Frederix, O. Mattelaer and R. Rietkerk, “Automatic spin-entangled decays of heavy resonances in Monte Carlo simulations,” J. High Energy Phys. 03 (2013) 015 (19 pages) doi:10.1007/JHEP03(2013)015 [arXiv:1212.3460 [hep-ph]].
* [7] J. C. Collins, “Hard-scattering factorization with heavy quarks: A general treatment,” Phys. Rev. D 58 (1998) 094002 (29 pages) doi:10.1103/PhysRevD.58.094002 [hep-ph/9806259].
* [8] B. A. Kniehl, G. Kramer and S. M. Moosavi Nejad, “Bottom-flavored hadrons from top-quark decay at next-to-leading order in the general-mass variable-flavor-number scheme,” Nucl. Phys. B 862 (2012) 720–736 doi:10.1016/j.nuclphysb.2012.05.008 [arXiv:1205.2528 [hep-ph]].
* [9] T. Kneesch, B. A. Kniehl, G. Kramer and I. Schienbein, “Charmed-meson fragmentation functions with finite-mass corrections,” Nucl. Phys. B 799 (2008) 34–59 doi:10.1016/j.nuclphysb.2008.02.015 [arXiv:0712.0481 [hep-ph]].
* [10] G. Kramer and H. Spiesberger, “Inclusive $D^{*}$ production in photon-photon collisions at next-to-leading order QCD,” Eur. Phys. J. C 22 (2001) 289–301 doi:10.1007/s100520100805 [hep-ph/0109167].
* [11] G. Kramer and H. Spiesberger, “Inclusive photoproduction of $D^{*}$ mesons with massive charm quarks,” Eur. Phys. J. C 38 (2004) 309–318 doi:10.1140/epjc/s2004-02054-5 [hep-ph/0311062].
* [12] B. A. Kniehl, G. Kramer, I. Schienbein and H. Spiesberger, “Inclusive $D^{*\pm}$ production in $p\bar{p}$ collisions with massive charm quarks,” Phys. Rev. D 71 (2005) 014018 (23 pages) doi:10.1103/PhysRevD.71.014018 [hep-ph/0410289].
* [13] B. A. Kniehl, G. Kramer, I. Schienbein and H. Spiesberger, “Collinear subtractions in hadroproduction of heavy quarks,” Eur. Phys. J. C 41 (2005) 199–212 doi:10.1140/epjc/s2005-02200-7 [hep-ph/0502194].
* [14] B. A. Kniehl, G. Kramer, I. Schienbein and H. Spiesberger, “Inclusive $B$-meson production at small $p_{T}$ in the general-mass variable-flavor-number scheme,” Eur. Phys. J. C 75 (2015) 140 (9 pages) doi:10.1140/epjc/s10052-015-3360-6 [arXiv:1502.01001 [hep-ph]].
* [15] M. Benzke, M. V. Garzelli, B. A. Kniehl, G. Kramer, S. Moch and G. Sigl, “Prompt neutrinos from atmospheric charm in the general-mass variable-flavor-number scheme,” J. High Energy Phys. 12 (2017) 021 (41 pages) doi:10.1007/JHEP12(2017)021 [arXiv:1705.10386 [hep-ph]].
* [16] B. A. Kniehl, G. Kramer, I. Schienbein and H. Spiesberger, “Finite-mass effects on inclusive $B$-meson hadroproduction,” Phys. Rev. D 77 (2008) 014011 (11 pages) doi:10.1103/PhysRevD.77.014011 [arXiv:0705.4392 [hep-ph]].
* [17] A. Heister et al. [ALEPH Collaboration], “Study of the fragmentation of $b$ quarks into $B$ mesons at the $Z$ peak,” Phys. Lett. B 512 (2001) 30–48 doi:10.1016/S0370-2693(01)00690-6 [hep-ex/0106051].
* [18] G. Abbiendi et al. [OPAL Collaboration], “Inclusive analysis of the $b$ quark fragmentation function in $Z$ decays at LEP,” Eur. Phys. J. C 29 (2003) 463–478 doi:10.1140/epjc/s2003-01229-x [hep-ex/0210031].
* [19] K. Abe et al. [SLD Collaboration], “Precise Measurement of the $b$-Quark Fragmentation Function in $Z^{0}$ Boson Decays,” Phys. Rev. Lett. 84 (2000) 4300–4304 doi:10.1103/PhysRevLett.84.4300 [hep-ex/9912058].
* [20] V. N. Gribov and L. N. Lipatov, “Deep inelastic $ep$ scattering in perturbation theory,” Sov. J. Nucl. Phys. 15 (1972) 438-450 [Yad. Fiz. 15 (1972) 781-807].
* [21] G. Altarelli and G. Parisi, “Asymptotic freedom in parton language,” Nucl. Phys. B 126 (1977) 298–318. doi:10.1016/0550-3213(77)90384-4
* [22] Yu. L. Dokshitzer, “Calculation of the structure functions for deep inelastic scattering and $e^{+}e^{-}$ annihilation by perturbation theory in quantum chromodynamics.,” Sov. Phys. JETP 46 (1977) 641-653 [Zh. Eksp. Teor. Fiz. 73 (1977) 1216-1240].
* [23] M. Salajegheh, S. M. Moosavi Nejad, H. Khanpour, B. A. Kniehl and M. Soleymaninia, “$B$-hadron fragmentation functions at next-to-next-to-leading order from a global analysis of $e^{+}e^{-}$ annihilation data,” Phys. Rev. D 99 (2019) 114001 (8 pages) doi:10.1103/PhysRevD.99.114001 [arXiv:1904.08718 [hep-ph]].
* [24] J. Abdallah et al. [DELPHI Collaboration], “A study of the $b$-quark fragmentation function with the DELPHI detector at LEP I and an averaged distribution obtained at the $Z$ Pole,” Eur. Phys. J. C 71 (2011) 1557 (29 pages) doi:10.1140/epjc/s10052-011-1557-x [arXiv:1102.4748 [hep-ex]].
* [25] M. Fischer, S. Groote, J. G. Korner, M. C. Mauser and B. Lampe, “Polarized top decay into polarized $W$: $t(\uparrow)\to W(\uparrow)+b$ at $O(\alpha_{s})$,” Phys. Lett. B 451 (1999) 406–413 doi:10.1016/S0370-2693(99)00194-X [hep-ph/9811482].
* [26] M. Fischer, S. Groote, J. G. Körner and M. C. Mauser, “Complete angular analysis of polarized top decay at $O(\alpha_{s})$,” Phys. Rev. D 65 (2002) 054036 (28 pages) doi:10.1103/PhysRevD.65.054036 [hep-ph/0101322].
* [27] W. Bernreuther, P. González and C. Mellein, “Decays of polarized top quarks to lepton, neutrino, and jets at NLO QCD,” Eur. Phys. J. C 74 (2014) 2815 (12 pages) doi:10.1140/epjc/s10052-014-2815-5 [arXiv:1401.5930 [hep-ph]].
* [28] A. Czarnecki, S. Groote, J. G. Körner and J. H. Piclum, “NNLO QCD corrections to the polarized top quark decay $t(\uparrow)\to X_{b}+W^{+}$,” Phys. Rev. D 97 (2018) 094008 (8 pages) doi:10.1103/PhysRevD.97.094008 [arXiv:1803.03658 [hep-ph]].
* [29] N. Cabibbo, “Unitary Symmetry and Leptonic Decays,” Phys. Rev. Lett. 10 (1963) 531–533 doi:10.1103/PhysRevLett.10.531
* [30] M. Kobayashi and T. Maskawa, “CP-Violation in the Renormalizable Theory of Weak Interaction,” Prog. Theor. Phys. 49 (1973) 652–657 doi:10.1143/PTP.49.652
* [31] B. A. Kniehl, A. V. Kotikov, A. I. Onishchenko and O. L. Veretin, “Strong-Coupling Constant with Flavor Thresholds at Five Loops in the Modified Minimal-Subtraction Scheme,” Phys. Rev. Lett. 97 (2006) 042001 (4 pages) doi:10.1103/PhysRevLett.97.042001 [hep-ph/0607202].
|
# Practical Utility PV Multilevel Inverter Solutions
John Buczek and Viktor Ivankevych
<EMAIL_ADDRESS>
<EMAIL_ADDRESS>
_Abstract–_ Multilevel inverters are used to improve power quality and reduce
component stresses. This paper describes and compares two multilevel cascaded
three phase inverter implementations with two different modulation techniques:
Phase Shifted Pulse Width Modulation, and Nearest Level Control. Further
analysis will show required number of inverter levels with respect to
modulation techniques to provide desired power and power quality to resistive
load or grid. Cascaded inverter will be designed and simulated to draw power
from PV cells.
## I Introduction
A multi-level inverter is a power electronic system that synthesizes a desired
voltage output from several levels of DC voltages as inputs [9]. Today, there
are many different topologies of multilevel converters including, but not
limited to, Diode-Clamped, Flying Capacitor, and Cascade H-bridge (CHB). While
the topologies may be different, they all offer similar beneficial features.
For sinusoidal outputs, multilevel converters improve their output voltage in
quality as the number of levels of the converter increase, thus decreasing the
Total Harmonic Distortion (THD) [6]. For this reason and others, multilevel
converters have been used for high power photovoltaic (PV) inversion, electric
motor drivers in electric vehicles, and other research and commercial
applications [6, 3, 2, 3, 4]. Although, technological problems such as
reliability, efficiency, the increase of the control complexity, and the
design of simple modulation methods have slowed down the application of
multilevel converters [6].
Figure 1: 5 Level Cascade H-bridge Converter [2]
Figure 1 shows a 5 level CHB converter. As can be seen, CHB converters consist
of multiple MOSFET (or equivalent) H-bridges that are connected in series.
Each H-bridge having its own isolated DC voltage source. For the shown 5 level
case, it requires two H-bridges that can be configured to output the 5 levels:
+V, +V/2, 0, -V/2, and -V. While the theory of adding and subtracting isolated
voltage sources is simple, such is harder to do in practice. Common methods
include using isolated DC-DC converters such as flyback and forward converters
that have transformers with multiple secondary windings [6]. Others use
individual, or multiple isolated sets of PVs that power individual H-bridges
[9, 2].
This paper uses two of the common modulation techniques for CHB converters,
Phase Shifted PWM (PSPWM) and Nearest Level Control (NLS), to propose designs
for a utility 3 phase PV inverter. Our designs will display many of the common
advantages and disadvantages of the different modulation techniques for CHB.
Our design requirements were to develop a 3 phase utility PV CHB inverter to
supply 125kW at $480V_{L-L}^{RMS}$ with a THD below 5%.
## II Nearest Level Switching
The Nearest-Level Switching control for a multilevel converter compares a
control sine wave to DC voltage levels. For a Cascading Multi-level Converter
consisting of _N_ H-bridges, each trigger when the voltage control sinusoidal
is greater than their respective threshold given by the equation:
$\displaystyle V_{thresh}(i)=V_{pk}\frac{2i-1}{2N}$ (1)
Where _N_ is the number of H-bridges equal to $N=\frac{L-1}{2}$ for an _L_
level Cascade H-bridge (CHB), and _i_ is the switch number which ranges from 0
to _N-1_.
Figure 2: Nearest Level Switching Waveform Synthesis [6]
Figure 2 shows the waveform for the Nearest Level Switching. It should be
noted that for every DC voltage source, the threshold voltage is at half of
the DC value. Additionally, the peak output voltage is equal to $N*V_{DC}$.
Figure 3: 9-Level NLS CHB PWM Waveforms
Figure 3 shows the PWM waveforms for the top H-bridge of a 9-level NLS CHB.
The PWM for switches 1 and 3 are not the same, as is the case for switches 2
and 4. For a single H-bridge inverter, the diagonal switches could have the
same PWM input, but that is not possible for multilevel converters. In order
to achieve all of the voltage steps, either voltages need to be added and
subtracted, or the outputs of H-bridges need to be shorted [4]. Our approach
was to use the shorting technique. This essentially adds an equivalent of a
zero state in Space Vector Modulation [6]. This output shorting was
implemented by making the PWM for switch 3 the inverse of PWM 2 and PWM 4 the
inverse of PWM 1. This shorted the outputs through switches 4 and 3. In
practice, both zero states of switches 3,4 as well as switches 1,2 should be
used in order to have more even wear on individual switches.
## III Nearest Level Switching Resistive Load RMS and THD Calculations
For a cascading H-Bridge Multilevel Inverter with $L$ levels, using the
Nearest-Level-Switching technique, the switching point $\alpha_{i}$ for level
$i=0$ to $i=\frac{L-1}{2}$, are given by the equation:
$\displaystyle\alpha_{i}=sin^{-1}\left(\frac{2i+1}{L-1}\right)$ (2)
### III-A Root Mean Squared
From [8], the equation for calculating the RMS ($X^{RMS}$) of a function
$x(t)$ is given as:
$\displaystyle X^{RMS}=\sqrt{\frac{1}{T}\int_{0}^{T}(x(t))^{2}dt}$ (3)
#### III-A1 3-Level
For the simple 3-level Inverter with NLS, the RMS voltage of a resistive load
can be calculated to be:
$\displaystyle V_{3-L}^{RMS}$
$\displaystyle=\sqrt{\frac{1}{\pi}\int_{\alpha_{0}}^{\pi-\alpha_{0}}(V_{m})^{2}dt}$
$\displaystyle=V_{m}\sqrt{\frac{1}{\pi}\left[t\right]_{\alpha_{0}}^{\pi-\alpha_{0}}}$
$\displaystyle=V_{m}\sqrt{\frac{1}{\pi}\left(\pi-\alpha_{0}-\alpha_{0}\right)}$
$\displaystyle=V_{m}\sqrt{1-\frac{2}{\pi}\alpha_{0}}$
For $\alpha_{0}=sin^{-1}\left(\frac{1}{2}\right)=\frac{\pi}{6}$
$\displaystyle V_{3-L}^{RMS}$ $\displaystyle=V_{m}\sqrt{\frac{2}{3}}$
### III-B 5-Level
For the simple 5-level Inverter with NLS, the RMS voltage of a resistive load
can be calculated to be:
$\displaystyle V_{5-L}^{RMS}$
$\displaystyle=\biggl{\\{}\frac{1}{\pi}\biggl{(}\int_{\alpha_{0}}^{\alpha_{1}}\biggl{(}\frac{V_{m}}{2}\biggl{)}^{2}dt+\int_{\alpha_{1}}^{\pi-\alpha_{1}}\biggl{(}V_{m}\biggl{)}^{2}dt$
$\displaystyle+\int_{\pi-\alpha_{1}}^{\pi-\alpha_{0}}\biggl{(}\frac{V_{m}}{2}\biggl{)}^{2}dt\biggl{)}\biggr{\\}}^{1/2}$
$\displaystyle=V_{m}\sqrt{\frac{1}{\pi}\left(\left[\frac{t}{4}\right]_{\alpha_{0}}^{\alpha_{1}}+\left[t\right]_{\alpha_{1}}^{\pi-\alpha_{1}}+\left[\frac{t}{4}\right]_{\pi-\alpha_{1}}^{\pi-\alpha_{0}}\right)}$
$\displaystyle=V_{m}\sqrt{\frac{1}{\pi}\left(\pi-\frac{1}{2}\alpha_{0}-\frac{3}{2}\alpha_{1}\right)}$
$\displaystyle=V_{m}\sqrt{1-\frac{1}{2\pi}\alpha_{0}+\frac{3}{2\pi}\alpha_{1}}$
For $\alpha_{0}=sin^{-1}\left(\frac{1}{4}\right)$ and
$\alpha_{1}=sin^{-1}\left(\frac{3}{4}\right)$
$\displaystyle V_{5-L}^{RMS}$ $\displaystyle\approx V_{m}0.7449$
### III-C 7-Level
For the simple 7-level Inverter with NLS, the RMS voltage of a resistive load
can be calculated to be:
$\displaystyle V_{7-L}^{RMS}$
$\displaystyle=\biggl{\\{}\frac{1}{\pi}\biggl{(}\int_{\alpha_{0}}^{\alpha_{1}}\left(\frac{V_{m}}{3}\right)^{2}dt+\int_{\alpha_{1}}^{\alpha_{2}}\left(\frac{2V_{m}}{3}\right)^{2}dt$
$\displaystyle+\int_{\alpha_{2}}^{\pi-\alpha_{2}}\left(V_{m}\right)^{2}dt+\int_{\pi-\alpha_{2}}^{\pi-\alpha_{1}}\left(\frac{2V_{m}}{3}\right)^{2}dt$
$\displaystyle+\int_{\pi-\alpha_{1}}^{\pi-\alpha_{0}}\left(\frac{V_{m}}{3}\right)^{2}dt\biggl{)}\biggl{\\}}^{1/2}$
$\displaystyle=V_{m}\biggl{\\{}\frac{1}{\pi}\biggl{(}\left[\frac{t}{3}\right]_{\alpha_{0}}^{\alpha_{1}}+\left[\frac{2t}{3}\right]_{\alpha_{1}}^{\alpha_{2}}$
$\displaystyle+\left[t\right]_{\pi-\alpha_{2}}^{\alpha_{2}}+\left[\frac{2t}{3}\right]_{\pi-\alpha_{1}}^{\pi-\alpha_{2}}+\left[\frac{t}{3}\right]_{\pi-\alpha_{1}}^{\pi-\alpha_{0}}\biggl{)}\biggl{\\}}^{1/2}$
$\displaystyle=V_{m}\sqrt{\frac{1}{\pi}\left(\pi-\frac{2}{9}\alpha_{0}-\frac{6}{9}\alpha_{1}-\frac{10}{9}\alpha_{2}\right)}$
$\displaystyle=V_{m}\sqrt{1-\frac{2}{9\pi}\alpha_{0}-\frac{6}{9\pi}\alpha_{1}-\frac{10}{9\pi}\alpha_{2}}$
For $\alpha_{0}=sin^{-1}\left(\frac{1}{6}\right)$,
$\alpha_{1}=sin^{-1}\left(\frac{3}{6}\right)$, and
$\alpha_{2}=sin^{-1}\left(\frac{5}{6}\right)$
$\displaystyle V_{7-L}^{RMS}$ $\displaystyle\approx V_{m}0.7217$
### III-D L-Level
From the previous derivations and Equation 2, a pattern for the RMS voltage
can be seen. For a multilevel cascade H-Bridge Inverter with L levels, the
equation for the RMS voltage of a resistive load can be given by:
$\displaystyle V_{L}^{RMS}$
$\displaystyle=V_{m}\sqrt{1-\sum^{N-1}_{i=0}\left(\frac{2(2i+1)}{\pi*N^{2}}*\alpha_{i}\right)}$
$\displaystyle=V_{m}\sqrt{1-\frac{2}{\pi
N^{2}}\sum^{N-1}_{i=0}\left(sin^{-1}\left(\frac{2i+1}{L-1}\right)\left(2i+1\right)\right)}$
(4)
### III-E Fourier Series Expansion
The Fourier Series Expansion was also performed on the ideal cascade H-bridge
Inverter. In the ideal case with a purely resistive load, the output waveform
has odd symmetry and quarter-wavelength symmetry inherently. From [8], the
Fourier Series Expansion Coefficients for such symmetry are given as:
$\displaystyle a_{h}$ $\displaystyle=0$ $\displaystyle b_{h}$
$\displaystyle=\frac{4}{\pi}\int_{0}^{\frac{\pi}{2}}\left(x(t)sin(h\omega
t)\right)d\omega t$
#### III-E1 3-Level
For the simple 3-level Inverter with NLS, the Fourier Series Expansion
Coefficients of a resistive load can be expressed as:
$\displaystyle b_{h}(3)$
$\displaystyle=\frac{4}{\pi}\int_{\alpha_{0}}^{\frac{\pi}{2}}\left(V_{m}sin(h\omega
t)\right)d\omega t$ $\displaystyle=\frac{4V_{m}}{\pi h}\left[-cos(h\omega
t)\right]^{\frac{\pi}{2}}_{\alpha_{0}}$ $\displaystyle=\frac{4V_{m}}{\pi
h}cos(h\alpha_{0})$
### III-F 5-Level
For the simple 3-level Inverter with NLS, the Fourier Series Expansion
Coefficients of a resistive load can be expressed as:
$\displaystyle b_{h}(5)$
$\displaystyle=\frac{4}{\pi}\biggl{(}\int_{\alpha_{0}}^{\frac{\pi}{2}}\left(\frac{V_{m}}{2}sin(h\omega
t)\right)d\omega t$
$\displaystyle+\int_{\alpha_{1}}^{\frac{\pi}{2}}\left(\frac{V_{m}}{2}sin(h\omega
t)\right)d\omega t\biggl{)}$ $\displaystyle=\frac{2V_{m}}{\pi
h}\left(\left[-cos(h\omega
t)\right]_{\alpha_{0}}^{\frac{\pi}{2}}+\left[-cos(h\omega
t)\right]_{\alpha_{1}}^{\frac{\pi}{2}}\right)$
$\displaystyle=\frac{2V_{m}}{\pi
h}\left(cos(h\alpha_{0})+cos(h\alpha_{1})\right)$
$\displaystyle=\frac{2V_{m}}{\pi
h}\left(cos(h\alpha_{0})+cos(h\alpha_{1})\right)$
### III-G L-Level
For a multilevel cascade H-Bridge Inverter with L levels, the Fourier Series
Expansion Coefficients of a resistive load can be expressed as:
$\displaystyle b_{h}(L)$
$\displaystyle=\frac{4}{\pi}\left(\sum^{N-1}_{i=0}\int_{\alpha_{i}}^{\frac{\pi}{2}}\left(\frac{V_{m}}{N}sin(h\omega
t)\right)d\omega t\right)$ $\displaystyle=\frac{4V_{m}}{\pi
hN}\sum^{N-1}_{i=0}\left(cos\left(h\alpha_{i}\right)\right)$
This can be simplified using the trigonometry identity:
$\displaystyle cos(sin^{-1}(x))=\sqrt{1-x^{2}}$
From Equation 2, the Fourier Series first Coefficient can be simplified to be:
$\displaystyle b_{1}(L)$ $\displaystyle=\frac{4V_{m}}{\pi
N}\sum^{N-1}_{i=0}\left(\sqrt{1-\left(\frac{(2i+1)}{L-1}\right)^{2}}\right)$
(5)
### III-H Total Harmonic Distortion
From [8], the equation for the Total Harmonic Distortion (THD) is given by:
$\displaystyle THD=\frac{\sqrt{(X^{RMS})^{2}-(X^{RMS}_{1})^{2}}}{X^{RMS}_{1}}$
(6)
From Equation 5, the equation for the RMS magnitude of the first harmonic can
be calculated as:
$\displaystyle V^{RMS}_{1}(L)=\frac{4V_{m}}{\pi
N\sqrt{2}}\sum^{N-1}_{i=0}\left(\sqrt{1-\left(\frac{(2i+1)}{L-1}\right)^{2}}\right)$
(7)
Equations 7, 4, and 6 were combined and plotted using python (Appendix A) to
calculate the THD of an L-level CHB inverter with a resistive load. At the
same time, PSIM simulations for the same number of levels were run with
resistive loads and compared against one another.
Table I: PSIM Simulated and Theoretically Calculated THD Levels (L) | PSIM THD (%) | Calculated THD(%)
---|---|---
3 | 31.0512 | 31.08419
5 | 17.5799 | 17.6012
7 | 12.2126 | 12.2272
9 | 9.35322 | 9.363669
11 | 7.58321 | 7.587252
13 | 6.3712 | 6.378124
15 | 5.49467 | 5.502021
17 | 4.83621 | 4.837995
19 | 4.31314 | 4.317328
21 | 3.89612 | 3.89809
23 | 3.55342 | 3.553263
25 | 3.26193 | 3.264629
27 | 3.017 | 3.01947
Table I shows the THD results from both the Theoretical Calculated THD and the
PSIM simulated THD for 3 levels to 27 levels. The two data sets are consistent
with each other. From the Table, as the number of levels increases, the THD
decreases. This quantitatively confirms that the output sinusoidal quality
improves with more and more additional levels of the CHB.
## IV Nearest Level Switching Simulation Design
As previously mentioned, the design goals were to produce a 60 Hz 3 phase
inverter at $480V_{L-L}^{RMS}$ with a real power output of 125 kW. The THD of
our inverter also needed to be below 5 %. These specifications were consistent
with other commercially available PV inverters [5].
Figure 4: Nearest Level Switching Simulated H-bridge and Buck Converter
Figure 4 shows the PSIM simulation for the NLS-CHB inverter individual
H-bridge and buck converter. From the previous section, it was found that for
a resistive load, the theoretical THD of a 27 level NLS-CHB was approximately
3%, so a 27 level inverter was chosen. From the number of levels and the peak
grid voltage of $480V_{L-L}^{RMS}$, the individual H-bridge voltage was
calculated to be $V_{dc}=480\sqrt{2/3}/13\approx 30.15V$. Our simulations did
not include PV maximum power point tracking, so our simulations used a
constant voltage source equal to the maximum voltage output of a single PV
from the datasheet in [10].
From [7], for a buck converter, we know:
$\displaystyle V_{o}$ $\displaystyle=DVs$ (8) $\displaystyle\Delta i_{L}$
$\displaystyle\approx\frac{V_{l}*\Delta t}{L}$
$\displaystyle\approx\frac{DV_{s}(1-D)T}{L}$ $\displaystyle L$
$\displaystyle\approx\frac{DV_{s}(1-D)T}{\Delta i_{L}}$ (9)
$\displaystyle\Delta V_{c}$ $\displaystyle\approx\frac{T\Delta i_{L}}{8C}$
(10)
We expect $I_{Lpk}\approx 30A$. Let $f_{s}=200kHz$, $V_{s}=48.9V$,
$V_{o}=480\sqrt{2/3}/13\approx 30.15V$,$\Delta V_{c}=4V$ , and $\Delta
i_{L}=5\%=6A$. Then:
$\displaystyle D$ $\displaystyle=0.6165$ $\displaystyle L$
$\displaystyle=9.633\mu H\approx 10\mu H$ $\displaystyle C$
$\displaystyle=937nF\approx 1\mu F$
The capacitor and inductor were intentionally kept small, but above critical
values such to reduce their equivalent impedance when placed in series. Once
the buck converters were designed, a single phase was simulated against a grid
ac voltage source with a phase shift of -2.5 degrees in order to determine the
desired cutoff frequency for the output filter. From our simulations, a cutoff
frequency of approximately 700Hz was selected based on the output current
harmonics. A simple LC low pass filter was chosen, with a cutoff frequency
given by [7] as:
$\displaystyle f_{c}=\frac{1}{2\pi\sqrt{LC}}$ (11)
An inductor $L_{f}=1mH$ and a capacitor $C_{f}=50\mu F$ were selected.
## V Nearest Level Switching Simulation Simulation
### V-A NLS Results
The NLS-CHB was first simulated with a single phase using ideal switches and a
phase angle of -2.5 degrees from the grid voltage source. The simulation was
able to reach a power level of 8.5 kW per phase at the desired line to neutral
voltage level. The output current RMS was 30.66A and the current THD was
3.02%. At this current level, using the PV arrays from [10], we would need two
in parallel going connected to each buck converter.
The NLS-CHB was then simulated with a single phase using the PSIM default
lossy switching models for NMOS MOSFETs. In addition, all reactive components
were given series resistance values of 50m $\Omega$. The circuit was simulated
for 0.5s and the output THD increased to 4.486% at a power factor of 99.56%.
Figure 5: Nearest Level Switching PSIM 3-Phases (Y Connection) Figure 6:
Nearest Level Switching 3 Phase Voltage and Current Waveforms Figure 7:
Nearest Level Switching 3 Phase Voltage and Current Characteristics
The NLS-CHB was finally simulated with all three phases, each with a
respective phase shift of -2.5 degrees. Figure 5 shows the PSIM file of the
NLS-CHB three phase simulation. Figures 6 and 7 show the output waveforms and
characteristics respectively. The desired output voltage of $480V_{L-L}^{RMS}$
was achieved. The output real power was approximately 25kW with an output
current THD of approximately 3.12%. From our simulation, in order to achieve
our design requirements, 5 of such three phase inverters would need to be
placed in parallel.
Figure 8: Nearest Level Switching H-bridge Switch Stresses
Figure 8 displays the individual switch stresses for the H-Bridge MOSFETs.
Each MOSFET in each H-bridge needed to conduct a peak of 50A and block a peak
of 40V. Base on design experience, voltage and current ratings are desired to
be increased by 150 % to maintain safe operation at extreme performance cases.
For a $480V_{L-L}^{RMS}$ system, these are comparatively small values [4].
While in this topology there are more switches, each switch experiences less
stress. This leads to the possibility of using more efficient switches that
have lower voltage and current ratings. Our simulation did kept each H-bridge
at a constant duty ratio. The duty cycle for individual H-Bridges would need
to be different, or a more complex controller would need to be used in order
to evenly apply duty cycles to all the H-bridges.
### V-B NLS Comments
The benefits that the Nearest Level Switching brings include greater
efficiency and the ability to use additional active elements (additional
H-bridges) in order to improve output waveform quality [3, 4]. From our
analysis it has been shown that by increasing the number of H-bridges and
levels, the THD of the output waveform decreases, thus making the output
closer to a pure sinusoid. Due to the slow switching nature of the NLS
technique, all switches turn on and off only once for the fundamental period
of the output waveform, reducing the commutation losses of the switches but
increasing the conduction losses. Also, as the number of H-bridges increases,
the voltages across each of the switches in the H-bridges decreases (Equation
1), thus reducing individual switch losses and stresses. At the same time, as
the number of H-bridges increases, the number of total switches will increase,
thus increasing the total losses as well as the overall complexity of the
control for all of the switches [4]. Thus the optimized number of switches
depends on the specifications of the switches used as well as the output
voltage and power. As a result, NLS CHB circuits are better suited for some
applications more than others [3, 6].
## VI Phase Shifted PWM
The Phase Shifted PWM control for a multilevel converter applies a triangular
waveform in comparison to a control sinusoidal function in order to obtain the
desired PWM for each H-bridge. Each H-bridge’s triangle waveform has a phase
shift depending on the number of levels:
$\displaystyle\theta_{shift}=\frac{360^{\circ}}{L-1}=\frac{360^{\circ}}{2N}$
(12)
Where N is the number of H-bridges in the cascade. The DC voltage for each
H-bridge level is defined as:
$\displaystyle V_{DC}=\frac{V_{DC,0}}{N}$ (13)
Where $V_{DC,0}$ is the DC voltage required to generate desired AC output
voltage in case of a single level inverter.
## VII Phase Shifted PWM Inverter Design
For this project, it has been decided to choose cascade consisting of 6 level
H-bridge inverters with PSPWM control. The carrier waveforms for all of the 6
levels were triangle waves of 100kHz. The overall 3 phase circuit with load,
and level circuitry are shown in Figures 9 and 10. Note that the inverter is
connected to the grid and the grid has an associated inductance is 1 mH. The
grid voltage was given a -2.5° phase shift with respect to inverter output
voltage to facilitate current flow from inverter to grid.
From our project requirements, the required output voltage was
$480V_{L-L}^{RMS}$ , or $\approx 277V_{L-N}^{RMS}$. From [8] and Equation 13,
the total required DC voltage $V_{DC,0}$ and the individual H-bridge DC
voltages $V_{DC,level}$ were found as:
$\displaystyle V_{DC,0}=\frac{V_{rms,LL}}{m_{a}|_{m_{a}=0.8}}$
$\displaystyle*\sqrt{\frac{2}{3}}=\sqrt{\frac{2}{3}}\frac{480}{0.8}\approx
490V$ $\displaystyle V_{DC,level}$ $\displaystyle=\frac{490}{6}\approx 81.67V$
Similarly, from equation 12, the individual carrier wave phase shift can be
found as:
$\displaystyle\theta_{shift}=\frac{360^{\circ}}{2*6}=30^{\circ}$
Figure 9: 3 phase 6 level cascaded H-bridge inverter with grid as a load
Figure 10: PSPWM carriers (leg A, leg B) for 6 level cascaded H-bridge
inverter. Phase shift is $30^{\circ}$
Figure 10 shows the carrier signals for a 6 level cascaded H-bridge inverter
(positive leg control signals are shown in top part, and negative leg control
signals are shown in bottom part)
Figure 11: Buck converter required to facilitate 81.67 VDC for inverter level
Figure 11 displays the PS-PWM buck converter used. The inductor and capacitor
values were calculated using Equations 9 and 10 (Final inductor and capacitor
values were chosen to be $100\mu H$ and $100\mu F$).This circuit will
facilitate power flow from the PV network to the cascaded inverter. Based on
the single phase inverter, current draw from the circuit was $\approx
13.7ADC$. That leads to conclude that the inverter’s levels will behave as
resistive load of $6\Omega$ for the buck converter. This figure has been used
to design the appropriate buck converter (Equation 11).
Based on the data for PV cells in [10], it was estimated that two sets of
three series PV cells in parallel connected to each buck converter would be
needed in order to provide sufficient voltage and current for our inverter.
Hence, the input voltage and available current for buck converter are: 120.6
VDC @ 19.42 ADC. This defines duty cycle of buck converter to be 0.677 from
Equation 8.
Figure 12: Filtered and unfiltered voltage and current output waveforms
Based on the unfiltered voltage and current simulation data, the voltage high
frequency harmonics at frequencies above 2 kHz. Therefore, in order to have a
low filtered THD with reasonably high L and C values for filter, it was
decided to set cut-off frequency of LC filter at $\approx 1453Hz$. In addition
to LC filter, the equivalent grid line inductance of 1 mH also acts as a
filter. From Equation 11 the calculated filter values were:
$\displaystyle L_{f}=200\mu H$ $\displaystyle C_{f}=60\mu F$
## VIII PS-PWM Simulation Results
### VIII-A $3\phi$ Simulation Data Analysis
First, the PSIM simulations for a single phase using the default PSIM lossy
MOSFET models and lossy reactive elements ($R_{series}=50m\Omega$) were
examined.
Figure 13: Phase A filtered/unfiltered voltage and current waveforms Figure
14: Phase A filtered/unfiltered voltage/current waveform characteristic
Figures
Figures 13 and 14 display the output voltage and current filtered and
unfiltered waveforms as well as the wave characteristics respectively for the
single phase lossy simulation. As can be seen, the output voltage met the
required line-line voltage of $480V_{L-L}^{RMS}$. In addition, all of the THD
values were below 5%. The single phase output power was approximately 6.8kW.
### VIII-B $3\phi$ Simulation Data Analysis
Next, the PSIM simulations for a three phase circuit using the default PSIM
lossy MOSFET models and lossy reactive elements ($R_{series}=50m\Omega$) were
examined.
Figure 15: Three phase filtered/unfiltered voltage and current waveforms
Figure 16: Three phase filtered/unfiltered voltage/current waveform
characteristic Figures
Figures 15 and 16 display the output voltage and current filtered and
unfiltered waveforms as well as wave their characteristics for the single
phase lossy simulation. As can be seen, the output voltage and output current
THD values are below the required 5%. The total output power from the three
phase simulation was calculated to be approximately 20.3kW. Thus in order to
meet the design requirement of 125kW, at least seven three phase H-bridge
inverters of this topology would need to be used.
Figure 17: 6 level cascaded H-bridge inverter. Switch voltage/current stresses
Figure 17 shows the voltages and currents experienced by the H-Bridges of the
PS-PWM inverter. Based on the simulation data, the voltage stress on MOSFETs
was approximately 85 VDC, and current stress was approximately 40 ADC. Base on
design experience, voltage and current ratings are desired to be increased by
150 % to maintain safe operation at extreme performance cases. That implies
that actual switch ratings should be 150 VDC / 60 ADC.
## IX Conclusions
Multilevel converters inherently provide desired characteristics for high
powered applications; but with them come inherent issues such as more complex
structure and operation. The CHB in particular has a structure that allows for
very high power applications due to their series connections of isolated power
supplies. The drawback of this structure is that the if a single power source
is used to supply each of the levels, then the isolation transformer would
require currently non-standard transformers with large numbers of secondary
windings[4]. Our solutions, as well as [9, 2], solve this problem by having
isolated PV cells. This method assumes that all of the PV’s output the same
current, but in practice would require more complex control in order to
achieve the desired output voltage and power from PV cells that are not
providing equal power. In addition, the gate control of each H-bridge would
require to be isolated.
This paper proposes two solutions for creating CHB inverters capable of
outputting 125kW at $480V_{L-L}^{RMS}$. Our simulation results show that such
is possible while maintaining a current THD below 5% as required for IEEE-519
[1]. Multilevel converters such as the CHB have unique features for power
quality and modularity. Although they are not commonly used in industry now,
they have great potential for the future.
## References
* [1] IEEE Recommended Practice and Requirements for Harmonic Control in Electric Power Systems. IEEE Std 519-2014 (Revision of IEEE Std 519-1992), pages 1–29, 2014\.
* [2] A. R. Beig, U. R. Y. Kumar, and V. T. Ranganathan. A novel fifteen level inverter for photovoltaic power supply system. In Conference Record of the 2004 IEEE Industry Applications Conference, 2004. 39th IAS Annual Meeting., volume 2, pages 1165–1171 vol.2, 2004.
* [3] Z. Du, L. M. Tolbert, J. N. Chiasson, B. Ozpineci, H. Li, and A. Q. Huang. Hybrid cascaded H-bridges multilevel motor drive control for electric vehicles. In 2006 37th IEEE Power Electronics Specialists Conference, pages 1–6, 2006.
* [4] L. G. Franquelo, J. Rodriguez, J. I. Leon, S. Kouro, R. Portillo, and M. A. M. Prats. The age of multilevel converters arrives. IEEE Industrial Electronics Magazine, 2(2):28–39, 2008.
* [5] Kaco. blueplanet 87.0 – 150 TL3 Transformerless, three-phase string inverters. 125TL3.
* [6] S. Kouro, R. Bernal, H. Miranda, C. A. Silva, and J. Rodriguez. High-Performance Torque and Flux Control for Multilevel Inverter Fed Induction Motors. IEEE Transactions on Power Electronics, 22(6):2116–2123, 2007.
* [7] Philip T. Krein. Elements of Power Electronics. Oxford University Press, 1998.
* [8] Ned Mohan, Tore M. Undeland, and Willian P. Robbins. Power Electronics: Converters, Applications, and Designs. John Wiley & Sons, Inc., 3rd edition, 2003.
* [9] R. L. Naik and K. R. Y. Udaya. A novel technique for control of cascaded multilevel inverter for photovoltaic power supplies. In 2005 European Conference on Power Electronics and Applications, pages 9 pp.–P.9, 2005.
* [10] Trina Solar. Bifacial Dual Glass 144 Cell Multi Busbar Module, 2019. TSM-DEG15MC.20(II).
| John Buczek currently pursuing a BSMS in Electrical Engineering with
concentration in Power Systems at Northeastern University, Boston, MA. His
research interests include power electronics, UAVs, and the Wireless Internet
of Things.
---|---
| Viktor Ivankevych received B.S (2016) and M.S (2017) degrees in Electrical
Engineering from New York University, Brooklyn, NY. He is currently pursuing
the Ph.D. degree in Electrical Engineering at Northeastern University, Boston,
MA. His research interests include power electronics and power converters.
---|---
### Appendix A: THD Python3 Code
from math import pi, asin, sqrt
def rms_value(L):
’’’
Function to return the rms value for the ’L’ level cascade H-bridge inverter
:param L: int for the number of levels (needs to be odd)
:retutn: float of the magnitude of rms value
’’’
N = (L-1)/2 # Number of H-bridges
sum_steps = 0 # summation variable
for ii in range(0, int(N), 1): # loop from 0 to N-1
sum_steps = sum_steps + asin((2*ii+1)/(L-1))*(2*ii + 1)
return sqrt(1 - (2/(pi*N*N))*sum_steps)
def first_harmonic_rms(L):
’’’
Funciton to return the rms value of the first harmonic for
the ’L’ level cascade H-bridge inverter
:param L: int for the number of levels (needs to be odd)
:return: float of the magnitude of the rms of the first harmonic
’’’
N = (L-1)/2 # Number of H-bridges
sum_steps = 0 # summation variable
for ii in range(0, int(N), 1): # loop from 0 to N-1
sum_steps = sum_steps + sqrt(1-((2*ii+1)/(L-1))**2)
return (8/(pi*1.0*(L-1)))*sum_steps/sqrt(2)
def main():
’’’
Main Method
’’’
L_array = range(3, 29, 2) # L from 3 to 27 odd integers
for L in L_array:
rms = rms_value(L) # get rms
f_harm = first_harmonic_rms(L) # get first harmonic rms
thd = sqrt(rms**2 - f_harm**2) / f_harm # calc THD
print("Levels: ",L, "Vrms: " ,rms, "V,1,rms: ",f_harm , "THD: ", thd)
if __name__ == "__main__":
main()
|
TCID_50 PFU infectious doses SIN
# Time to revisit the endpoint dilution assay and to replace and as measures
of a virus sample’s infection concentration
Daniel Cresta Department of Physics, Ryerson University, Toronto, ON, M5B
2K3, Canada Donald C. Warren Interdisciplinary Theoretical and Mathematical
Sciences (iTHEMS) program, RIKEN, Wako-shi, Saitama, 351-0198, Japan
Christian Quirouette Department of Physics, Ryerson University, Toronto, ON,
M5B 2K3, Canada Amanda P. Smith Department of Pediatrics, University of
Tennessee Health Science Center, Memphis, TN, 38163, USA Lindey C. Lane
Department of Pediatrics, University of Tennessee Health Science Center,
Memphis, TN, 38163, USA Amber M. Smith Department of Pediatrics, University
of Tennessee Health Science Center, Memphis, TN, 38163, USA Catherine A. A.
Beauchemin Department of Physics, Ryerson University, Toronto, ON, M5B 2K3,
Canada Interdisciplinary Theoretical and Mathematical Sciences (iTHEMS)
program, RIKEN, Wako-shi, Saitama, 351-0198, Japan
###### Abstract
The infectivity of a virus sample is measured by the infections it causes, via
a plaque or focus forming assay ( or FFU) or an endpoint dilution (ED) assay
(, CCID50, EID50, etc., hereafter collectively $\text{ID}_{50}$). The counting
of plaques or foci at a given dilution intuitively and directly provides the
concentration of infectious doses in the undiluted sample. However, it has
many technical and experimental limitations. For example, it is subjective as
it relies on one’s judgement in distinguishing between two merged plaques and
a larger one, or between small plaques and staining artifacts. In this regard,
ED assays are more robust because one need only determine whether or not
infection occurred. The output of the ED assay, the 50% infectious dose
($\text{ID}_{50}$), is calculated using either the Spearman-Kärber (1908|1931)
or Reed-Muench (1938) mathematical approximations. However, these are often
miscalculated and their approximation of the $\text{ID}_{50}$ cannot be
reliably related to the infectious dose. Herein, we propose that the plaque
and focus forming assays be abandoned, and that the measured output of the ED
assay, the $\text{ID}_{50}$, be replaced by a more useful measure we coined
_specific infections_ (). We introduce a free, open-source web-application,
midSIN, that computes the concentration in a virus sample from a standard ED
assay, requiring no changes to current experimental protocols. We use midSIN
to analyze sets of influenza and respiratory syncytial virus samples, and
demonstrate that the /mL of a sample reliably corresponds to the number of
infections a sample will cause per unit volume. The /mL concentration of a
virus sample estimated by midSIN, unlike the $\text{ID}_{50}$/mL, can be used
directly to achieve the desired multiplicity of infection. Estimates obtained
with midSIN are shown to be more accurate and robust than those obtained using
the Reed-Muench and Spearman-Kärber approximations. The impact of ED plate
design choices (dilution factor, replicates per dilution) on measurement
accuracy is also explored. The simplicity of as a measure and the greater
accuracy provided by midSIN make them an easy and superior replacement for the
, FFU, and other ID50 measures. We hope to see their universal adoption to
measure the infectivity of virus samples.
## 1 Introduction
The progression of a virus infection _in vivo_ or _in vitro_ , or the
effectiveness of therapeutic interventions in reducing viral loads, are
monitored over time through sample collections to measure changes (increases
or decreases) in virus concentrations. As such, accurate measurement of the
concentration in a sample is critical to study and manage virus infections.
The most direct method is to count individual virions as observed under an
electron microscope. However, this technique is costly, time consuming, and
largely destructive of the samples, and is thus almost never used. Viral RNA
can be counted via quantitative polymerase chain reaction (qPCR), a method
that amplifies a specific virus genome segment (RNA or DNA) within the sample
over multiple cycles. The growth curve resulting from successive amplification
cycles, compared against the standard curve for a sample of known
concentration, provides an estimate of the number of viral segments in the
sample. The major limitation of this method is that it measures not only viral
RNA from intact virions, only some of which are infection-competent, but also
debris from apoptotic or lysed cells, and antibody- or antiviral-neutralized
virions, which misrepresents the effective virion concentration. For this
reason, a count of infectious particles rather than, or in addition to, total
viral genome segments is preferred.
Infectious virions do not systematically differ in any observable way from
replication-defective virions, nor do they differ in a physical way that would
allow for their mechanical or chemical separation. For this reason, methods to
count infectious virions are based on counting the infections they cause,
rather than the particles themselves. In practice, however, not all infection-
competent virions contained in a sample will go on to successfully cause
infection. Certain experimental conditions, such as temperature or acidity of
the medium, can hasten the rate at which virions that were infection-competent
in the sample lose infectivity before they can cause infection. This is why,
hereafter, we will refer to the quantity measured by infectivity assays as the
_infection concentration_ or the number of infections the sample will cause
per unit volume, rather than its concentration of infectious virions, which is
not a measurable quantity. Two main types of assays are used to quantify the
infection concentration within a virus sample: (1) the plaque forming and
focus forming assays; and (2) assays we will collectively refer to as endpoint
dilution (ED) assays111Technically, the plaque and focus forming assays are
also endpoint dilution assays because they rely on the counting of plaques or
foci (the endpoint) as a function of dilutions. However, herein, we will refer
to them as plaque or foci forming assays rather than endpoint dilution
assays., which include the 50% tissue culture infectious dose () or cell
culture infectious dose (CCID50) or egg infectious dose (EID50) assays, etc.
The plaque forming assay was introduced by Renato Dulbecco in 1952 [3], as an
improvement over the ED assay. The plaque forming assay and the focus forming
assay, which rely on the same principles, suffer from a number of critical
issues that cannot be overcome. For example, the liquid accumulation
(meniscus) that forms around well edges means some infectious doses will not
get quantified correctly or at all. It can be hard to distinguish two merged
plaques from a single large plaque, or to decide how small a plaque one should
consider when counting. Some of the difficulties in establishing a robust,
unambiguous plaque or focus count for a given well are illustrated in Figure
1. For these reasons, different researchers will count a different number of
plaques or foci when observing the same well. This subjectivity in the count
means there is opportunity to (sub)consciously count a few more plaques or
foci, for example, when expecting a virus strain to be more severe than
another or in the absence of an antiviral compound. Ideally, there would be no
discretion involved in the counting process of a quantification assay. Indeed,
the decision process should be made by a physical, automated measurement,
without the possibility of post-facto adjustments of any kind, for any reason.
Figure 1: Examples of challenges in establishing a robust count of infection
plaques or foci. MDCK cells were infected with a sample containing influenza A
(H3N2) virions, and cell infection was visualized via staining by antibodies
against the matrix (M) viral protein. The uneven liquid distribution along the
well’s edges means some infectivity is lost or miscounted. It can be hard to
distinguish between two merged foci and a single larger uneven focus. It is
difficult to determine how small a focus should be counted, and doing so to
decide on a focus size threshold to be used consistently for all wells and all
samples within a particular experiment. As a result of these difficulties,
different individuals will commonly count a different number of foci in the
same well. Stained well image graciously provided by Frederick Koster
(Lovelace Research Institute, NM, USA).
In contrast, the ED assay offers a more decisive and robust binary
determination as to whether or not infection has taken place in each well (or
egg, animal, etc.). This determination is insensitive to small, spatially
localized irregularities and is typically unanimously agreed upon by all
observers. Therefore, it is less subject to (sub)conscious bias. In fact, this
feature of the ED assay makes it ideal for systematic, machine-based
determination of positive wells (or eggs or other culture types), eliminating
subjectivity. Furthermore, infection of wells in the ED assay can be carried
out in exactly the same way as planned infection experiments where they will
make up the inoculum, e.g., in the same cell type, reproducing whether the
inoculum is rinsed or not post-inoculation, and the duration of incubation
with the inoculum. In contrast, plaque and focus forming assays can require
the use of a semi-solid cellular overlay (e.g., agarose) to restrict the
spread of virus beyond cells neighbouring those initially infected by the
inoculum. The need to rinse or remove the inoculum to add the semi-solid
overlay imposes strict constraints on the timing of this rinse. Because a
longer incubation provides more opportunities for infectious virus to cause
infection, the number of infectious doses counted via a plaque or focus assay
can underestimate the true number of infections that will result when the
quantified sample is later used to infect cells under longer incubation
periods. The plaque assay can also require the use of different cells than
those used in the infection experiments whenever the latter fail to die or
detach (form clearly visible plaques) post-infection, making it difficult to
predict the number of infections that will result when the quantified sample
is later used to infect different cells.
For all its many advantages, the ED assay currently has one key, remediable
weakness: its output quantity, the (or CCID50 or EID50), does not directly
correspond, or trivially relate, to one infectious dose. The simplistic
calculations, introduced by Spearman-Kärber (SK) [9, 5] and Reed-Muench (RM)
[8] nearly a century ago, remain the primary methods to quantify a virus
sample’s infectivity using the ED assay. Many research groups rely on
spreadsheet calculators that are passed down through generations of trainees
or found on the internet, and can contain errors222For example, versions 2 and
3 of the Excel spreadsheet calculator provided by the Lindenbach Lab at Yale
University (http://lindenbachlab.org/resources.html), which have since been
removed.. While, theoretically, a dose of $1\text{\,}\tcid$ is expected to
cause
$-1/\ln(50\%)=$1.44\text{\,}\mathrm{i}\mathrm{n}\mathrm{f}\mathrm{e}\mathrm{c}\mathrm{t}\mathrm{i}\mathrm{o}\mathrm{n}\mathrm{s}$$
[2], the approximation used by the SK and RM methods introduces an often
overlooked bias where
$$1\text{\,}\tcid$\approx$1.781\text{\,}\mathrm{i}\mathrm{n}\mathrm{f}\mathrm{e}\mathrm{c}\mathrm{t}\mathrm{i}\mathrm{o}\mathrm{n}\mathrm{s}$$
where $1.781=\mathrm{e}^{\gamma}$ and $\gamma=0.5772$ is the Euler-Mascheroni
constant [10, 4]. This makes it problematic to experimentally achieve the
desired multiplicity of infection when inoculating from a sample quantified
via the SK or RM methods. Many have proposed replacements for the RM and SK
calculations with some based on logit or probit transforms of the data [4, 6,
2] and others on statistical analysis of the ED assay output [6, 7]. Sadly,
none of these improvements were widely adopted, possibly due to a lack of
visibility of these publications, or the lack of widespread awareness of the
limitations of the RM and SK methods.
Thus, the one issue with the ED assay is not with the assay itself but with
the calculation of the . We submit that for all the reasons outlined above,
the ED assay is experimentally more robust and reliable than the plaque and
focus forming assays, and should be preferred over the latter. We propose to:
1. 1.
Continue the use, or encourage the adoption, of the ED assay (e.g., assay),
but to replace its output, the TCID50/mL (or CCID50/mL, EID50/mL, etc.), with
a new quantity in units of Specific INfections or /mL corresponding to the
number of infections the sample will cause per mL. The word _specific_
highlights the fact that the infectivity of a sample is specific to the
particulars of the experimental conditions (temperature, medium, cell type,
incubation time, etc.).
2. 2.
Replace the Reed-Muench and Spearman-Kärber approximations with a computer
software, midSIN (measure of infectious dose in SIN), that relies on Bayesian
inference to measure the /mL of a virus sample. To avoid calculation errors
and make the new method widely accessible, midSIN is maintained and
distributed as free, open-source software on GitHub
(https://github.com/cbeauc/midSIN) for user installation, but also via a free-
to-use website application (https://midsin.physics.ryerson.ca) with an
intuitive user interface.
Here, we present examples of midSIN being used to analyze influenza and
respiratory syncytial virus samples. We demonstrate that midSIN’s output, /mL,
is an accurate estimate of the number of infections the sample will cause per
unit volume. We show how the accuracy of the concentration estimate is
affected by experimental choice of plate layout, including the dilution
factor, and the number of replicates per dilution. We compare midSIN’s
performance to that of the RM and SK methods, and demonstrate how the latter
estimators are inaccurate under various circumstances, underlining the need to
adopt midSIN to quantify virus samples via the ED assay.
## 2 Results
### 2.1 Key features of midSIN’s output
Let us consider a fictitious ED experiment, with 11 dilutions and 8 replicate
wells per dilutions, in which the minimum sample dilution,
$\mathcal{D}_{1}=1/100=10^{-2}$, is serially diluted by a factor of
$10^{-0.5}\approx 0.32$ ($\mathcal{D}_{2}=10^{-2.5}$,
$\mathcal{D}_{3}=10^{-3}$, …, $\mathcal{D}_{11}=10^{-7}$), and the total
volume of inoculum (diluted virus sample + dilutant) placed in each well is
$V_{\text{inoc}}=$0.1\text{\,}\mathrm{mL}$$. Now, consider that a virus sample
is measured using this ED experiment and one observes (8,8,8,8,8,7,7,5,2,0,0)
infected wells out of 8 replicates at each of the 11 dilutions, as illustrated
in Figure 2A.
Figure 2: Visual representation of midSIN’s output for the example ED plate.
(A) Illustration of the example ED plate where $\mathcal{D}_{i}$ are the
chosen serial dilutions of the sample. For the example described in the text,
$\mathcal{D}_{1}=10^{-2}$, $\mathcal{D}_{2}=10^{-2.5}$, …,
$\mathcal{D}_{11}=10^{-7}$, with 8 replicates per dilution. The number of
infected wells (# inf) is indicated at the bottom of each dilution column. (B)
The midSIN-estimated likelihood distribution of the $\log_{10}$ infection
concentration, $\log_{10}(\idnew/\mathrm{m}\mathrm{L})$, for the example ED
experiment. The vertical lines correspond to
$\log_{10}(\idnew/\mathrm{m}\mathrm{L})$, based on the most likely value
(mode) of midSIN’s likelihood distribution (solid blue), or computed from the
RM (solid orange) and SK (dashed green) approximations of the
$\log_{10}(\tcid)$ (see Methods). The $x$-value of the white and light grey
region on either sides of the mode indicate the edges of the 68% and 95%
credible interval (CI), respectively. The midSIN-estimated
$\log_{10}(\idnew/\mathrm{m}\mathrm{L})$ $\text{mode}\pm 68\%\left[\pm
95\%\right]$ CI are indicated numerically above the graph. (C) The number of
infected wells (black circles) out of the 8 replicates, as a function of the
11 serial dilutions of the example ED plate, from the least (leftmost) to the
most (rightmost) diluted. For example, $x=3.0$ corresponds to a sample
dilution of $10^{-3}$ or 1/1,000. The average (expected) number of infected
wells, as a function of sample dilution, is shown for the most likely value of
$\log_{10}(\idnew/\mathrm{m}\mathrm{L})$ (blue curve) or its 68% and 95% CI
(inner and outer edge of the grey bands, respectively). The sample dilution
($x$-value) at which the blue curve crosses the horizontal dotted line (50%
infected wells) corresponds to a concentration of $1\text{\,}\tcid$ per ED
well volume. The vertical lines indicate the sample dilution that yields a
concentration of $1\text{\,}\tcid$ according to the RM and SK approximations.
midSIN provides a graphical output of its results, shown in Figure 2B,C for
this example. Note how the likelihood distribution for
$\log_{10}(\idnew/\mathrm{m}\mathrm{L})$ (Fig. 2B) is approximately a normal
distribution. This is why $\log_{10}$ of the infection concentration should be
used and reported, rather than the concentration itself. midSIN also
graphically compares the number of infected wells observed experimentally
(Fig. 2C, black dots) against the theoretically expected values (blue curve
and grey CI bands). This graphical representation makes it easy to identify
issues with the data entered or with the experiment itself.
Importantly, midSIN provides a more useful quantity to the user than the : an
estimate of the concentration of infections the sample will cause, /mL. For
this example, the concentration is $10^{6.2\pm
0.1}\text{\,}\idnew\mathrm{/}\mathrm{m}\mathrm{L}$, where 6.2 is the mode
(most likely value) of $\log_{10}(\idnew/\mathrm{m}\mathrm{L})$, and $\pm 0.1$
is its 68% credible interval (CI). The /mL corresponds to the number of
infections that will be caused per m L of the sample, which can be directly
used to determine the sample dilution required to obtain a desired
multiplicity of infection (MOI).
Figure 3: Quantification of RSV sampled from _in vitro_ infections. Each row
corresponds to a different experiment (mock-yield [my] or single-cycle [sc])
and sampling time point (e.g., $8\text{\,}\mathrm{h}$,
$36\text{\,}\mathrm{h}$), and each sample was measured in duplicate (rep1,
rep2). These data were collected from in vitro infections with the RSV A Long
strain, and were previously reported in [1]. The ED measurement experiment
were conducted using a plate layout of 11 dilutions, with 8 replicates per
dilution, an inoculum volume of $V_{\text{inoc}}=$0.1\text{\,}\mathrm{mL}$$,
serial dilutions from $\mathcal{D}_{1}=10^{-1}$ to $\mathcal{D}_{11}=10^{-6}$,
separated by a dilution factor of $10^{-0.5}$.
In a laboratory setting, ED experiments can be performed in batches, such as
to quantify the infectious concentration in samples collected at several time
points over the course of a cell culture infection. For such applications,
midSIN provides a comma separated value (csv) template file readily editable
in a spreadsheet program, to collect and submit the results for batch
processing. Details on the format of the template file are available on
midSIN’s website (https://midsin.physics.ryerson.ca). Figure 3 illustrates the
output for a subset of measurements for _in vitro_ infection with the
respiratory syncytial virus (RSV). Each sample was measured twice, and
midSIN’s estimates are in good agreement with one another (within 95% CI).
The $y$-axis in the left graph panels of midSIN’s graphical output is the non-
normalized scale of the likelihood distribution for
$\log_{10}(\idnew/\mathrm{m}\mathrm{L})$, which ranges between $10^{-7}$ and
$10^{-2}$. The scale loosely relates to the likelihood of observing a
particular ED experimental outcome (see Methods). Unlikely ED outcomes appear
as large departures of the observed number of infected wells (right panels,
black dots) from what is theoretically expected (right panels, curve). It is
interesting that the uncertainty (CI) of midSIN’s estimated
$\log_{10}(\idnew/\mathrm{m}\mathrm{L})$ appears to be independent of how much
the ED outcome deviates from theoretical expectations. That is, the accuracy
of midSIN is not strongly affected even when it is provided more unlikely,
noisy experimental data. This robustness is explored further below.
### 2.2 Comparing to and virus sample concentrations
The midSIN calculator provides an estimate of the number of infections that
will be caused per mL of a virus sample (/mL). In principle, a plaque assay
also measures the number of infections a sample will cause, with each
infection expected to develop into a plaque. If a plaque assay is performed
under experimental conditions and protocols as similar as possible to those of
the ED assay (i.e., using the same cells, medium, period of incubation,
rinsing method, etc.), midSIN’s /mL estimate is expected to be comparable, in
theory, to the number of /mL observed in the plaque assay. In practice,
however, the plaque assay likely provides a biased estimate of the
concentration of infections in a sample due to its many experimental issues,
discussed in the Introduction. To evaluate midSIN’s performance compared to
existing methods, the infection concentration in two influenza A (H1N1) virus
strain samples were measured via both plaque and ED assays, and their
concentration in units of , , and were compared (Fig. 4). Details regarding
the samples, and how the plaque and ED assays were performed are provided in
Methods.
Figure 4: Comparing to and for influenza A virus samples. (A,B) The infection
concentration in two influenza A (H1N1) virus strain samples was measured via
both an ED assay and a plaque assay (x, PFU). The ED assay was quantified in
$\log_{10}(\tcid)$ using the RM (square) or SK (triangle) methods, or in
$\log_{10}(\idnew)$ using midSIN (circle with 68%,95% CI). Each of the 2
strain samples was measured over 2 separate experiments (Exp. #1, #2),
performed each time by 2 different researchers (Researcher A or B), with 5
biological replicates each. The grey bars indicate the range of
$\log_{10}(\idnew)$ values across the 5 replicates. The RM, SK, and SIN
measures were estimated for each replicate based on the same ED plate. The
experimental details are provided in Methods. (C,D) The $\log_{10}$ of the
ratio between either the via the RM or SK method or the , over the via midSIN.
The ratios were computed for each replicate ($5\times 5$ replicates), per
experiment, per researcher (25 replicates $\times$ 2 researchers $\times$ 2
experiments = 100 ratios) shown as individual symbols (dots) for each method
(RM, SK, ). The mean and 68% CI of the 100 ratios are indicated numerically
and as black circles with error bars.
The concentrations estimated via the RM and SK methods are $\sim$1.5–1.7 times
larger (Fig. 4C,D) than the concentration, and the set of ratios are
statistically inconsistent with the assumption of equality ($p$-value:
0.01–0.03). Theoretically, $1\text{\,}\tcid$ is expected to cause
$1.44\text{\,}\mathrm{i}\mathrm{n}\mathrm{f}\mathrm{e}\mathrm{c}\mathrm{t}\mathrm{i}\mathrm{o}\mathrm{n}\mathrm{s}$
($=1/\ln(2)$) [2]. However, the RM or SK approximations are known to introduce
a bias such that $1\text{\,}\tcid$ estimated by these methods is expected to
cause
$1.781\text{\,}\mathrm{i}\mathrm{n}\mathrm{f}\mathrm{e}\mathrm{c}\mathrm{t}\mathrm{i}\mathrm{o}\mathrm{n}\mathrm{s}$
($=\mathrm{e}^{\gamma}$ where $\gamma=0.5772$ is the Euler-Mascheroni
constant) [10, 4]. Using the RM, SK, and measurements presented in Figure
4A,B, we confirmed333The mean $\log_{10}(\mathrm{ratio})$ was re-computed for
$\mathrm{ratio}=(\mathrm{RM}/1.781)/\idnew$ and (SK/1.781)/, and found to be
0.85–0.93. This is statistically consistent ($p$-value: 0.1–0.3) with the
assumption of equality, i.e., $\mathrm{ratio}=10^{0}=1$. that
$$1.781\text{\,}\idnew$\approx$1\text{\,}\tcid$$ when the latter is estimated
via the RM or SK approximations, as expected theoretically if is indeed
measuring the infection concentration in a sample.
Similarly, the ratio of the concentration determined via the plaque assay and
the concentrations estimated by midSIN is $\sim$0.89–0.93, which is
statistically consistent with the assumption of equality ($p$-value: 0.2–0.5).
These results confirm the theoretical expectation that
$$1\text{\,}\pfu$\approx$1\text{\,}\idnew$$ when the plaque and ED assays are
performed in the same manner, as was the case here. This provides further
support, via two independent assays, that the concentration estimated by
midSIN from the ED assay is a robust measure of the infection concentration of
a virus sample.
### 2.3 Comparing midSIN’s performance to that of the RM and SK methods
Figure 5: midSIN’s estimate of a sample’s infection concentration based on a
single dilution. This is a simulated example of an ED plate with an
inoculation volume of $V_{\text{inoc}}=$0.1\text{\,}\mathrm{mL}$$. Instead of
serial dilutions, a single dilution ($\mathcal{D}_{1}=0.01$) is used, and
either 1, 2 or 3 well(s) out of the 4 replicate wells are infected. As the
fraction of infected wells increases, the uncertainty on the estimate (68% and
95% CIs) decreases, and the likelihood distribution becomes more symmetric
(Normal-like).
The RM and SK methods rely on the number of infected wells decreasing as
dilution increases. Their estimates are affected when the number of infected
wells remains unchanged or even increases as dilution increases, which
statistics tell us can reasonably occur experimentally. The RM and SK methods
also mostly require that at the lowest and highest sample dilutions, all wells
be infected and uninfected, respectively. In contrast, midSIN is robust to
these issues. Figure 5 demonstrates how midSIN can provide an estimate for the
$\log_{10}(\idnew/\mathrm{m}\mathrm{L})$ in a sample using the number of
infected wells at a single dilution, as long as at least one well is
uninfected if all others are infected or vice-versa. This is because midSIN
relies on Bayesian inference, i.e., when more than one column is available, it
uses information from each column successively to revise and improve its
estimate. This allows midSIN to correct for even large deviations from
theoretical expectations, and thus improves its accuracy.
Figure 6: Comparing known input to estimated output concentrations. For each
input concentration between $10^{2.2}$ and $10^{9.4}$, one million random ED
experiment outcomes (# of positive wells in each dilution column) were
generated. For each ED outcome, either (A) midSIN was used to determine the
most likely $\log_{10}(\idnew/\mathrm{m}\mathrm{L})$; or the (B) RM or (C) SK
method was used to estimate the $\log_{10}(\tcid/\mathrm{m}\mathrm{L})$.
Vertically stacked grey bands at each input concentration are sideways
histograms, proportional to the number of ED outcomes that yield a given
$y$-axis value. The black curves join the median (thick), 68${}^{\text{th}}$
(thin) and 95${}^{\text{th}}$ (dashed) percentile of the histograms,
determined at (but not between) each input concentration. A plate layout of 11
dilutions, with 8 replicates per dilution, an inoculum volume of
$V_{\text{inoc}}=$0.1\text{\,}\mathrm{mL}$$, serial dilutions from
$\mathcal{D}_{1}=10^{-2}$ to $\mathcal{D}_{11}=10^{-8}$, separated by a
dilution factor of $10^{-0.6}\approx 1/4$ were used in the simulated ED
experiments.
Figure 6 illustrates how well the midSIN, RM, and SK methods recover a known
input sample concentration in simulated ED experiments, based on a plate
layout consisting of 11 dilutions ($\mathcal{D}_{1}=10^{-2}$ to
$\mathcal{D}_{11}=10^{-8}$), a dilution factor of $1/4$, and 8 replicates per
dilutions. The infection concentration estimated by midSIN is in excellent
agreement with the input concentration. For the RM and SK methods, which
estimate the $\log_{10}(\tcid/\mathrm{m}\mathrm{L})$ rather than the
$\log_{10}(\idnew/\mathrm{m}\mathrm{L})$, the agreement is generally poor due
to the bias they introduce. Furthermore, the RM and SK predictions are more
variable (wavy pattern), and lose accuracy dramatically as the sample
concentration approaches the limits of detection (the 2 ends) which, for the
example plate layout simulated here, is around
$10^{3}\text{\,}\idnew\mathrm{/}\mathrm{m}\mathrm{L}$ and
$10^{9}\text{\,}\idnew\mathrm{/}\mathrm{m}\mathrm{L}$. Interestingly, the
basic calculations behind the RM and SK methods constrain the set of values
they can return (sparsely populated grey histograms), compared to the more
continuous range returned by midSIN, which contributes to its increased
accuracy.
### 2.4 Estimate accuracy as a function of plate layout
In Figure 3, we observed that even for large discrepancies between the
expected (right panels, blue curve) and observed (right panels, black dots) ED
assay outcome, the uncertainty (CI) of midSIN’s estimate remains relatively
unchanged. This apparent robustness is because the uncertainty is primarily
determined by the experimental design, namely the change in dilution between
columns (dilution factor) and the number of replicate wells per dilution.
Figure 7 explores the impact of varying either only the dilution factor, or
only the number of replicates at each dilution, or varying one at the expense
of the other by using a fixed number of wells (96 wells). When using midSIN,
smaller changes in dilution (e.g., going from a dilution factor of 2.2/100 to
61/100) or more replicates per dilution (4 to 24) each improves the measure’s
accuracy (narrower CIs) by comparable amounts, but only when the total number
of wells is allowed to increase to accommodate the change. When the total
number of wells used is fixed, changing one at the expense of the other leaves
the accuracy (CI) unchanged. This is somewhat also true for the
$\log_{10}(\tcid)$ output concentration estimated by the RM and SK methods.
However, at the smallest dilution factors (10/100 and 2.2/100), the bias
introduced by the RM and SK methods becomes even larger and more
unpredictable. For the input concentration considered in Figure 7
($10^{5}\text{\,}\idnew\mathrm{/}\mathrm{m}\mathrm{L}$), the dilution at which
50% of wells are infected is near the middle dilution. For sample
concentrations such that 50% infected wells occur near or at the lowest or
highest dilution chosen, the effect is even more significant.
Figure 7: Comparing the effect of the dilution factor and number of replicates
per dilution. The effect of either (A,D,G) decreasing the change in dilution
(from a dilution factor of $2.2/100$ to $61/100$) while keeping 8 replicates
per dilution; or (B,E,H) increasing the number of replicates per dilutions (4
to 24) while keeping a fixed dilution factor ($\approx 35/100$); or (C,F,I)
increasing the dilution factor while decreasing the number of replicates,
keeping a fixed number of 96 wells used in total to titer one virus sample.
Different rows represent the ratio of the estimated output concentration using
(A–C) midSIN in /mL, (D–F) RM or (G–I) SK in /mL, and the input concentration.
In all cases (A–I), the input concentration was
$10^{5}\text{\,}\idnew\mathrm{/}\mathrm{m}\mathrm{L}$, and as the dilution
factor was varied, the highest and lowest dilutions in the simulated ED plate
were held fixed to $\mathcal{D}_{1}=10^{-2}$ and
$\mathcal{D}_{\text{last}}=10^{-7}$, respectively, by changing the total # of
dilutions performed (simulated). Everything else is generated or computed as
described in the caption of Figure 6.
Figure 7 also demonstrates that varying the dilution by smaller increments
(e.g., a dilution factor of 61/100 rather than 10/100) provides greater
granularity (uniqueness) of ED plate outcomes, and thus, greater accuracy of
the $\log_{10}$ infection concentration estimates. Here, a distinct plate
outcome means a distinct number of infected wells at each dilution, with no
distinction as to exactly which of the replicate wells (e.g., the second
versus the fourth) is infected at each dilution. An ED plate with serial
dilutions ranging over 6 orders of magnitude (e.g., $10^{-2}$ to $10^{-7}$),
with 4 different dilutions and 24 replicates/dilution (i.e., dilution factor
of 2.2/100) provides $\sim 10^{6}$ ($[24+1]^{4}$) possible, distinct ED plate
outcomes. In contrast, a plate with the same serial dilution range, but with
24 different dilutions and 4 replicates/dilution (i.e., dilution factor of
61/100) yields $\sim 10^{17}$ ($[4+1]^{24}$) distinct outcomes. More
generally, $[\text{reps}+1]^{\text{dils}}$ is the number of distinct plate
outcomes for a chosen number of dilutions (dils) and replicates (reps). Having
fewer possible plate outcomes means that a larger range of concentrations
would share the same most-likely ED plate outcome, yet each plate outcome only
maps to one (the most likely) concentration estimate. This means that with
fewer dilutions, the concentration estimate is forced to take on the nearest
possible value it can take (Fig. 7, the next grey bar), and the accuracy of
the concentration estimate is therefore reduced. So although having a greater
number of dilutions is more labour intensive, it should be preferred over
having a greater number of replicates per dilution.
## 3 Discussion
We have introduced a new calculator tool called midSIN to replace the Reed-
Muench (RM) and Spearman-Kärber (SK) calculations to quantify the infectivity
of a virus sample based on a endpoint dilution (hereafter ED) assay. Rather
than estimating the of a virus sample, midSIN calculates the number of
infections the sample will cause, reported in units of specific infections ().
It does so without requiring any changes to current ED assay protocols, and
can be accessed for free via an open-source web-application
(https://midsin.physics.ryerson.ca). Importantly, since the of a virus sample
corresponds to the number of infections it will cause, it can be used directly
to determine what dilution of the sample will achieve the desired multiplicity
of infection (MOI).
We showed that midSIN provides more accurate and robust estimates than the
biased RM and SK approximations. We confirmed that the RM and SK
approximations overestimate the by 23.5%, such that $1\text{\,}\tcid$
estimated by these methods will cause 1.781 rather than
$1.44\text{\,}\mathrm{i}\mathrm{n}\mathrm{f}\mathrm{e}\mathrm{c}\mathrm{t}\mathrm{i}\mathrm{o}\mathrm{n}\mathrm{s}$
[10, 4]. While in theory one can obtain the intended MOI by multiplying the by
0.7 (or rather $\ln(2)=0.693$), one should instead multiply by 0.561 to
account for the overestimation by RM and SK. Even when accounting for the
overestimation, we showed that these methods perform particularly poorly when
too few replicate wells per dilutions are used or when the change in dilution
is large between successive serial dilutions. The two methods perform
especially poorly when quantifying samples whose infection concentration
approaches, but is still well within, the detection limit of the ED assay. In
such cases, the bias introduced by these methods becomes even larger and more
significant. For example, if the minimum and maximum dilutions of an ED plate
are $10^{-2}$ and $10^{-8}$, virus samples with a concentration less than
$10^{2.2}\text{\,}\idnew$ or greater than $10^{7.6}\text{\,}\idnew$ per
inoculated well volume (typically $0.1\text{\,}\mathrm{m}\mathrm{L}$), will
see their concentration estimated with an even larger bias by the RM and SK
methods.
Using midSIN, rather than RM or SK, to measure the infectivity of a virus
sample based on an ED assay does not require any change to ED experimental
protocols and methods currently in use in one’s laboratory (e.g., dilution
factor, replicate per dilution, minimum dilution). Indeed, we demonstrated
that midSIN can estimate a virus sample’s concentration based on even just a
single dilution, as long as only a fraction of the replicate wells are
infected at that dilution. For a given number of ED wells used to titrate the
sample and fixed minimum and maximum dilutions (ED detection range), we showed
that having smaller changes between dilutions (a larger number of serial
dilutions) is better than having more replicates per dilution. So those
wishing to improve the accuracy in estimating the infectivity of their virus
samples should consider using more wells in titrating each virus sample, and
favouring smaller dilution changes over more replicates. For example, using 11
dilutions, with a 4-fold dilution factor between dilutions and 8 replicate
wells per dilution uses up 88 wells, leaving 8 wells of a 96-well plate for
controls. This ED plate design, analyzed using midSIN, accurately measures
virus sample concentrations ranging over $\sim$6 orders of magnitude (e.g.,
[$10^{1}$–$10^{7}$] /mL, or [$10^{6}$–$10^{12}$] /mL, etc.) with an accuracy
of $\sim$1.6-fold ($\times 10^{\pm 0.2}$, 95% CI). In comparison, using 7
dilutions, with a 10-fold dilution factor, and 4 replicates (which uses 28
rather than 88 wells) would also span 6 orders of magnitude, but with an
accuracy of $\sim$3.2-fold ($\times 10^{\pm 0.5}$, 95% CI). To put these 2
accuracies in perspective: $1\text{\,}\mathrm{mL}$ of a sample measured to
contain $10\text{\,}\idnew\mathrm{/}\mathrm{m}\mathrm{L}$, is expected to
yield either 6–16 or 3–31 infections 95% of the time, given an accuracy of
either $\times 10^{\pm 0.2}$ or $\times 10^{\pm
0.5}\text{\,}\idnew\mathrm{/}\mathrm{m}\mathrm{L}$, respectively. Such an
important decrease in accuracy means a reduced ability to detect experimental
changes as statistically signficant, with the $\times 10^{\pm 0.5}$ accuracy
requiring a $>$10-fold change for statistical significance. Failing to
identify a change as statistically significant as part of a study is far more
costly than using a few more wells for each sample to increase measurement
accuracy, and thus the statistical power of the study.
The midSIN-estimated obtained from an ED assay was also compared to the from a
plaque assay for a set of influenza A virus samples. When the plaque and ED
assays are performed as identically as possible (cell type, incubation time,
etc.), as was the case here, $$1\text{\,}\idnew$\approx$1\text{\,}\pfu$$. This
demonstrates that indeed midSIN’s is a measure of the number of infections a
virus sample will cause. However, as mentioned, the plaque and focus forming
assays often impose experimental requirements (e.g., an early rinse of the
inoculum to add agarose, use of cells with pronounced CPE). Such constraints
on the plaque or focus assay inoculation protocol make it nearly impossible to
relate the number of plaques or foci observed to the number of infections the
virus sample will cause under the intended, experimental infection conditions
(e.g., late or no inoculum rinse, no agarose, to infect cells exhibiting no
significant CPE). Adding to this the subjectivity of counting plaques or foci,
it is clear the ED assay combined with midSIN to estimate the concentration of
a virus sample is more accessible, accurate, and predictive.
Beyond the work presented herein, the development of midSIN will continue
online, as we implement new features and inputs for integration with various
colorimetric and fluorescence instruments. The ease of use of midSIN and the
greater usefulness and relevance of as a measure of a virus sample’s
infectivity make them far superior to all currently available alternatives,
including the PFU, FFU, , and other $\text{ID}_{50}$ measures. We hope to see
them adopted widely.
### Acknowledgements
The authors wish to thank Frederick Koster (Lovelace Respiratory Research
Institute, NM, USA) for providing the antibody stained well image, and Evan
Williams (UTHSC, UT, USA) for technical assistance.
### Funding
This work was supported in part by Discovery Grant 355837-2013 (CAAB) from the
Natural Sciences and Engineering Research Council of Canada (www.nserc-
crsng.gc.ca), Early Researcher Award ER13-09-040 (CAAB) from the Ministry of
Research and Innovation of the Government of Ontario
(www.ontario.ca/page/early-researcher-awards), by the Interdisciplinary
Theoretical and Mathematical Sciences programme (iTHEMS, ithems.riken.jp) at
RIKEN (CAAB), and by R01 AI139088 (AMS, APS, LCL) from the NIH NIAID
(www.niaid.nih.gov). The funders had no role in study design, data collection
and analysis, or decision to publish.
### Authors contribution
CAAB was responsible for study conceptualization, and project administration,
DC, DCW, CQ, CAAB all contributed to the development of the methodology, data
analysis, software, visualization. AMS, APS, LCL carried out experimentation,
CAAB, AMS were responsible for funding acquisition, experimental planning, and
supervision. DC and CAAB contributed to the original manuscript draft, and all
authors contributed to its review and editing.
## 4 Methods
### 4.1 The mathematics of the dose-response assay
#### 4.1.1 Considering a single well
Consider a virus sample of volume $V_{\text{sample}}$ which contains an
unknown concentration of infectious virions, $C_{\mathrm{inf}}$, which we aim
to determine. Drawing a small volume, $V_{\text{inoc}}<V_{\text{sample}}$,
from the sample of volume $V_{\text{sample}}$, is analogous to drawing balls
out of a bag containing green and yellow balls, and considering green balls a
success, and yellow ones a failure. It is a series of Bernoulli trials where
$n=V_{\text{inoc}}/V_{\text{vir}}$
is the number of draws, i.e., the number of virion-size volumes
($V_{\text{vir}}$) drawn from the sample to form the inoculum volume
($V_{\text{inoc}}$), analogous to the number of balls drawn.
$k$
is the number of successes, i.e., the number of infectious virions drawn from
the sample to form the inoculum, analogous to the number of green balls drawn.
$p$
is the probability of success, i.e., the fraction of virion-size volumes in
the sample that are occupied by infectious virions, analogous to the
probability of drawing a green ball.
The probability of success, $p$, is related to the concentration of infectious
virus in the sample, $C_{\mathrm{inf}}$, as
$p=\frac{\text{Number of virions in sample}}{\text{Number of virion-size
volumes in the
sample}}=\frac{C_{\mathrm{inf}}V_{\text{sample}}}{V_{\text{sample}}/V_{\text{vir}}}=C_{\mathrm{inf}}V_{\text{vir}}\
,$
where $C_{\mathrm{inf}}$ is the quantity we aim to estimate. Unlike the ball
analogy where it is easy to count how many green balls $k$ were drawn, after
having drawn $n$ virion-size volumes from the sample into our inoculum, we
cannot count how many infectious virions were drawn into the inoculum.
However, if this inoculum is deposited onto a susceptible cell culture, we can
observe whether or not infection occurs, and this would indicate that the
inoculum contained at least one or more infectious virions. Note that, as
explained in the Introduction, even a productively infectious virion, i.e.,
one capable of completing the full virus replication from attachment to
progeny release, might not result in a productive infection. As such, from
hereon, $C_{\mathrm{inf}}$ is used to designate the concentration of specific
infections in the sample, which is smaller or equal to the concentration of
infectious virions, i.e., measures a subset of the infectious virions.
Having deposited the inoculum into one well of the 96-well plate of our ED
experiment, the likelihood that the well will _not_ become infected
corresponds to the likelihood of having drawn $k=0$ infectious virions (or
rather, specific infections) out of the $n$ virion volumes that make up our
inoculum, namely
$\displaystyle q_{\text{noinf}}$
$\displaystyle=\text{Binomial}(k=0|n=V_{\text{inoc}}/V_{\text{vir}},p=C_{\mathrm{inf}}V_{\text{vir}})$
(1) $\displaystyle=\frac{n!}{0!(n-0)!}p^{0}(1-p)^{n-0}=(1-p)^{n}$
$\displaystyle q_{\text{noinf}}$
$\displaystyle=\left(1-C_{\mathrm{inf}}\,V_{\text{vir}}\right)^{V_{\text{inoc}}/V_{\text{vir}}}$
where $q_{\text{noinf}}$ can be simplified by realizing that
$\displaystyle\ln(1-x)$
$\displaystyle\stackrel{{\scriptstyle|x|<1}}{{=}}-x-\frac{x^{2}}{2}-\frac{x^{3}}{3}-...\stackrel{{\scriptstyle|x|\ll
1}}{{\approx}}-x$ $\displaystyle\ln(q_{\text{noinf}})$
$\displaystyle=\frac{V_{\text{inoc}}}{V_{\text{vir}}}\
\ln(1-C_{\mathrm{inf}}\,V_{\text{vir}})\approx\frac{V_{\text{inoc}}}{V_{\text{vir}}}\
(-C_{\mathrm{inf}}\,V_{\text{vir}})=-C_{\mathrm{inf}}\,V_{\text{inoc}}\ .$
As such,
$q_{\text{noinf}}=(1-C_{\mathrm{inf}}\,V_{\text{vir}})^{V_{\text{inoc}}/V_{\text{vir}}}\approx\exp\left[-C_{\mathrm{inf}}\,V_{\text{inoc}}\right]$
(2)
where $q_{\text{noinf}}$ and $(C_{\mathrm{inf}}V_{\text{vir}})\in[0,1]$
because $C_{\mathrm{inf}}=N_{\text{vir}}/V_{\text{sample}}$ and the number of
specific infections in the sample, $N_{\text{vir}}$, is at a minimum zero, and
at most the maximum number of virion-size volumes that can physically fit in
the sample volume, namely $V_{\text{sample}}/V_{\text{vir}}$. As such, the
maximum possible infection concentration, given a sample of volume
$V_{\text{sample}}$, is
$C_{\mathrm{inf}}=(V_{\text{sample}}/V_{\text{vir}})/V_{\text{sample}}=1/V_{\text{vir}}$,
and $C_{\mathrm{inf}}\in[0,1/V_{\text{vir}}]$.
#### 4.1.2 Considering replicate wells at a given dilution
The ED assay is based on serial dilutions of the sample, with each dilution
separated by a fixed dilution factor. We define the dilution factor $\in(0,1)$
as the fraction of the inoculum volume drawn from the previous dilution. For
example, if the inoculum for a well,
$V_{\text{inoc}}=$100\text{\,}\mathrm{\SIUnitSymbolMicro L}$$, comprises
$10\text{\,}\mathrm{\SIUnitSymbolMicro L}$ drawn from the previous dilution
and $90\text{\,}\mathrm{\SIUnitSymbolMicro L}$ of dilution media, the dilution
factor is $10/100=0.1$. If the serial dilution begins with a dilution of
$\mathcal{D}_{1}=0.2$, then the following dilution will be
$\mathcal{D}_{2}=0.02$. In Eqn. (1), the dilution under consideration,
$\mathcal{D}_{i}$, will affect $n$, the number of virion-sized volumes drawn
from the sample and deposited into the wells of the $i^{\mathrm{th}}$
dilution, such that $n=\mathcal{D}_{i}V_{\text{inoc}}/V_{\text{vir}}$.
Therefore, the probability that a well at the $i^{\mathrm{th}}$ dilution will
_not_ become infected is given by
$q_{i}\equiv
q_{\text{noinf}}^{\mathcal{D}_{i}}=(1-C_{\mathrm{inf}}V_{\text{vir}})^{\mathcal{D}_{i}V_{\text{inoc}}/V_{\text{vir}}}\approx\exp\left[-C_{\mathrm{inf}}V_{\text{inoc}}\mathcal{D}_{i}\right]$
(3)
where $1-q_{i}$ is the probability of infection for a well at the
$i^{\mathrm{th}}$ dilution, where $\mathcal{D}_{i}\in[0,1]$.
When conducting an ED assay, each dilution in the assay contains a number of
independent infection wells (replicates), all inoculated with the same
dilution, $\mathcal{D}_{i}$. This is analogous again to drawing balls out of a
bag, but this time there are $n_{i}$ draws (replicate wells), and the
probability of success (i.e., that a well becomes infected) is simply one
minus the probability of failure (i.e., that a well does not become infected,
$q_{i}$). The probability that $k_{i}$ out of the $n_{i}$ wells become
infected at dilution $\mathcal{D}_{i}$, is described by the Binomial
distribution
$\text{Binomial}(k=k_{i}|n=n_{i},p=1-q_{i})=\frac{n_{i}!}{k_{i}!(n_{i}-k_{i})!}\,(1-q_{i})^{k_{i}}\,q_{i}^{n_{i}-k_{i}}\propto(1-q_{\text{noinf}}^{\mathcal{D}_{i}})^{k_{i}}\,q_{\text{noinf}}^{\mathcal{D}_{i}(n_{i}-k_{i})}$
where $n_{i}$ is the number of replicate wells at each dilution, but could be
less if any well at dilution $\mathcal{D}_{i}$ are spoiled or contaminated.
However, our interest is not in determining $k_{1}$ given $q_{\text{noinf}}$,
but rather in determining $q_{\text{noinf}}$ given that we observed $k_{1}$
infected wells out of $n_{1}$ wells in the first column. To this aim, we can
make use of Bayes’ theorem which, in our context, can be expressed as
$\mathcal{P}(p|\text{data})=\frac{\mathcal{P}(\text{data}|p)\
\mathcal{P}(p)}{\int_{0}^{1}\mathcal{P}(\text{data}|p)\,\mathcal{P}(p)\,\mathrm{d}p}$
or rather
$\displaystyle\mathcal{P}_{\text{post,1}}(q_{\text{noinf}}|k_{1})$
$\displaystyle=\frac{\mathcal{P}(k_{1}|q_{\text{noinf}})\
\mathcal{P}_{\text{prior}}(q_{\text{noinf}})}{\int_{0}^{1}\mathcal{P}(k_{1}|q_{\text{noinf}})\,\mathcal{P}_{\text{prior}}(q_{\text{noinf}})\,\mathrm{d}q_{\text{noinf}}}$
$\displaystyle=\frac{\left[(1-q_{\text{noinf}}^{\mathcal{D}_{1}})^{k_{1}}\,q_{\text{noinf}}^{\mathcal{D}_{1}(n_{1}-k_{1})}\right]\,\mathcal{P}_{\text{prior}}(q_{\text{noinf}})}{\int_{0}^{1}\mathcal{P}(k_{1}|q_{\text{noinf}})\,\mathcal{P}(q_{\text{noinf}})\,\mathrm{d}q_{\text{noinf}}}$
$\displaystyle\mathcal{P}_{\text{post,1}}(q_{\text{noinf}}|k_{1})$
$\displaystyle\propto\left[(1-q_{\text{noinf}}^{\mathcal{D}_{1}})^{k_{1}}\,q_{\text{noinf}}^{\mathcal{D}_{1}(n_{1}-k_{1})}\right]\,\mathcal{P}_{\text{prior}}(q_{\text{noinf}})$
where $\mathcal{P}_{\text{post,1}}(q_{\text{noinf}}|k_{1})$ is our updated,
posterior belief about $q_{\text{noinf}}$ after having observed $k_{1}$
successes out of $n_{1}$ trials in the first column ($i=1$), and given our
prior belief, $\mathcal{P}_{\text{prior}}(q_{\text{noinf}})$, about
$q_{\text{noinf}}$ before making this observation.
#### 4.1.3 Considering all dilutions of the ED assay
As mentioned above, in the 96-well ED assay, each dilution contains a number
of independent infection wells (replicates) inoculated with the same sample
concentration. This process is then repeated over a series of dilutions, each
separated from the previous by a fixed dilution factor. Having observed the
fraction of wells infected at the first dilution considered,
$\mathcal{D}_{1}$, we have updated our posterior belief about
$q_{\text{noinf}}$. We will now use this updated belief as our new prior as we
observe our second dilution ($\mathcal{D}_{2}$), such that
$\displaystyle\mathcal{P}_{\text{post,2}}(q_{\text{noinf}}|\vec{k}_{2})$
$\displaystyle\propto\mathcal{P}(k_{2}|q_{\text{noinf}})\
\mathcal{P}_{\text{post,1}}(q_{\text{noinf}}|k_{1})$
$\displaystyle\mathcal{P}_{\text{post,2}}(q_{\text{noinf}}|\vec{k}_{2})$
$\displaystyle\propto\left[(1-q_{\text{noinf}}^{D_{2}})^{k_{2}}\,q_{\text{noinf}}^{D_{2}(n_{2}-k_{2})}\right]\,\left[(1-q_{\text{noinf}}^{\mathcal{D}_{1}})^{k_{1}}\,q_{\text{noinf}}^{\mathcal{D}_{1}(n_{1}-k_{1})}\right]\mathcal{P}_{\text{prior}}(q_{\text{noinf}})$
$\displaystyle\mathcal{P}_{\text{post,2}}(q_{\text{noinf}}|\vec{k}_{2})$
$\displaystyle\propto\mathcal{Q}(\vec{k}_{2}|q_{\text{noinf}})\,\mathcal{P}_{\text{prior}}(q_{\text{noinf}})\
,$
where we introduce $\vec{k}_{2}=\\{k_{1},k_{2}\\}$ and
$\mathcal{Q}(\vec{k}_{2}|q_{\text{noinf}})=\left[(1-q_{\text{noinf}}^{D_{2}})^{k_{2}}\,q_{\text{noinf}}^{D_{2}(n_{2}-k_{2})}\right]\,\left[(1-q_{\text{noinf}}^{\mathcal{D}_{1}})^{k_{1}}\,q_{\text{noinf}}^{\mathcal{D}_{1}(n_{1}-k_{1})}\right]$
as short-hands for convenience. From this, it is easy to extrapolate the
posterior likelihood distribution (pPLD) after having observed all $J$
dilutions ($\mathcal{D}_{1}$, $\mathcal{D}_{2}$, …, $\mathcal{D}_{J}$) of the
ED assay, namely
$\mathcal{P}_{\text{post,J}}(q_{\text{noinf}}|\vec{k}_{J})\propto\mathcal{Q}(\vec{k}_{J}|q_{\text{noinf}})\,\mathcal{P}_{\text{prior}}(q_{\text{noinf}})$
(4)
where
$\mathcal{Q}(\vec{k}_{J}|q_{\text{noinf}})=\left[\prod_{j=1}^{J}(1-q_{\text{noinf}}^{\mathcal{D}_{j}})^{k_{j}}\right]q_{\text{noinf}}^{\sum_{j=1}^{J}\mathcal{D}_{j}(n_{j}-k_{j})}\
.$ (5)
Note that this expression is largely equivalent to that obtained by Mistry et
al. [7].
#### 4.1.4 Considering the choice of prior
In Eqn. (4), we obtained a pPLD for $q_{\text{noinf}}$. Our objective,
however, is to estimate the pPLD of $C_{\mathrm{inf}}$, the specific infection
concentration in our sample, rather than $q_{\text{noinf}}$. In fact, because
both the plaque and ED assays provide an accuracy that is normally distributed
in $\log_{10}(C_{\mathrm{inf}})$ rather than $C_{\mathrm{inf}}$, it follows
that $\log_{10}(C_{\mathrm{inf}})$ (hereafter $\ell_{\mathrm{Cinf}}$) rather
than $C_{\mathrm{inf}}$ is the quantity of interest. We note that
$\mathcal{Q}(\vec{k}_{J}|q_{\text{noinf}})$ in Eqn. (4) is a probability
density function in $\vec{k}_{J}$ rather than in $q_{\text{noinf}}$. As such,
a change of variables, say from $q_{\text{noinf}}$ to
$\ell_{\mathrm{Cinf}}(q_{\text{noinf}})$, would affect only the prior because
$\mathcal{Q}(\vec{k}_{J}|q_{\text{noinf}})=\mathcal{Q}(\vec{k}_{J}|q_{\text{noinf}}(\ell_{\mathrm{Cinf}}))=\mathcal{Q}(\vec{k}_{J}|\ell_{\mathrm{Cinf}})$.
Thus, the pPLD for $\ell_{\mathrm{Cinf}}$ is given by
$\mathcal{P}_{\text{post},J}(\ell_{\mathrm{Cinf}}|\vec{k}_{J})\propto\mathcal{Q}(\vec{k}_{J}|q_{\text{noinf}}(\ell_{\mathrm{Cinf}}))\
\mathcal{P}_{\text{prior}}(\ell_{\mathrm{Cinf}})\ ,$ (6)
where
$\mathcal{Q}(\vec{k}_{J}|q_{\text{noinf}})=\mathcal{Q}(\vec{k}_{J}|C_{\mathrm{inf}})=\mathcal{Q}(\vec{k}_{J}|\ell_{\mathrm{Cinf}})$
because $\mathcal{Q}(\vec{k}_{J}|q_{\text{noinf}}(...))$ can be written in
terms of $q_{\text{noinf}}$, $C_{\mathrm{inf}}$, or $\ell_{\mathrm{Cinf}}$,
because it is a probability density function in
$\vec{k}_{J}=\\{k_{1},k_{2},...,k_{J}\\}$ rather than in $q_{\text{noinf}}$.
To complete this expression, we need to choose a physically and biologically
appropriate prior belief regarding $\ell_{\mathrm{Cinf}}$.
Prior to conducting the ED assay, we know at least that
$C_{\mathrm{inf}}\in[1/V_{\text{Earth}},1/V_{\text{vir}}]$, where
$1/V_{\text{vir}}$ is the maximum possible concentration, namely that if the
entire volume of the sample is constituted solely of infectious virions, and
$1/V_{\text{Earth}}$ is the minimum possible concentration, namely that if
there was only one infectious virion left on Earth. As we explain below, these
limits are not important; only the fact that they are convincingly physically
bounded both from above and below, i.e., $\in(0,\infty)$, is relevant.
If we choose our prior to be uniform in
$C_{\mathrm{inf}}\in[1/V_{\text{Earth}},1/V_{\text{vir}}]$, namely
$\mathcal{P}_{\text{prior}}(C_{\mathrm{inf}})=1/(1/V_{\text{vir}}-1/V_{\text{Earth}})\approx
V_{\text{vir}}$, and using the fact that
$\mathcal{P}_{\text{prior}}(C_{\mathrm{inf}})\,\mathrm{d}C_{\mathrm{inf}}=\mathcal{P}_{\text{prior}}(\ell_{\mathrm{Cinf}})\,\mathrm{d}\ell_{\mathrm{Cinf}}$,
we can write
$\mathcal{P}_{\text{prior}}(\ell_{\mathrm{Cinf}})=\mathcal{P}_{\text{prior}}(C_{\mathrm{inf}})\frac{\mathrm{d}C_{\mathrm{inf}}}{\mathrm{d}\ell_{\mathrm{Cinf}}}=V_{\text{vir}}\frac{\mathrm{d}\left[10^{\ell_{\mathrm{Cinf}}}\right]}{\mathrm{d}\ell_{\mathrm{Cinf}}}=V_{\text{vir}}\ln(10)10^{\ell_{\mathrm{Cinf}}}\propto
10^{\ell_{\mathrm{Cinf}}}$
which yields
$\mathcal{P}_{\text{post},J}(\ell_{\mathrm{Cinf}}|\vec{k}_{J})\propto\mathcal{Q}(\vec{k}_{J}|q_{\text{noinf}}(\ell_{\mathrm{Cinf}}))\
10^{\ell_{\mathrm{Cinf}}}\ .$ (7)
We see here that the range chosen for the uniform prior in $C_{\mathrm{inf}}$
is not important because it only contributes a constant to our proportionality
Eqn. (6).
Alternatively, because the ED assay estimates $\ell_{\mathrm{Cinf}}$ rather
than $C_{\mathrm{inf}}$, our prior belief about the virus concentration is
more appropriately expressed in $\ell_{\mathrm{Cinf}}$ rather than
$C_{\mathrm{inf}}$. Again, the bounds of the uniform distribution in
$\ell_{\mathrm{Cinf}}$ is unimportant, provided that it is finite in extent
such that
$\ell_{\mathrm{Cinf}}\in[{\ell_{\mathrm{Cinf}}}_{\text{min}},\log_{10}(1/V_{\text{vir}})]$
where ${\ell_{\mathrm{Cinf}}}_{\text{min}}>-\infty$, such that we can write
$\mathcal{P}_{\text{post},J}(\ell_{\mathrm{Cinf}}|\vec{k}_{J})\propto\mathcal{Q}(\vec{k}_{J}|q_{\text{noinf}}(\ell_{\mathrm{Cinf}}))\
.$ (8)
Figure 8 illustrates the two distinct priors assumed to arrive at Eqns. (7)
and (8) and their impact on the posterior
$\mathcal{P}_{\text{post},J}(\ell_{\mathrm{Cinf}}|\vec{k}_{J})$ for the
example ED experiment described in Section 2.1. Figure 8A illustrates the
consequence of choosing a prior uniform in $C_{\mathrm{inf}}$, i.e., a bias
towards higher virus concentrations. This is because a uniform prior in
$C_{\mathrm{inf}}$ corresponds to a belief that one is as likely to measure a
set of virus concentrations in the range $[0.001,\,0.002]$ as in the range
$[1,000,000.001,\,1,000,000.002]$. When plotted on a log-scale, there are
$100\times$ more intervals of width 0.001 in $[10^{4},10^{5}]$ than in
$[10^{2},10^{3}]$. Thus, this prior corresponds to a belief that the
likelihood of measuring a certain virus concentration increases exponentially
as $\ell_{\mathrm{Cinf}}$ increases linearly. In contrast, a prior uniform in
$\ell_{\mathrm{Cinf}}$ corresponds to a belief that one is as likely to
measure a set of virus concentrations in the range $[0.001,\,0.002]$ than in
the range $[1,000,000,\,2,000,000]$, or rather in the range $[1,2]\times
10^{-3}$ than in the range $[1,2]\times 10^{6}$. As such, a uniform
distribution in $\ell_{\mathrm{Cinf}}$ is more physically and biologically
sensible and therefore was chosen for our estimation method.
Figure 8: Impact of the choice of prior on the posterior distribution for
$\ell_{\mathrm{Cinf}}$. (A) Non-normalized priors for $\log_{10}$(specific
infections, /mL)$=\ell_{\mathrm{Cinf}}$ that are uniform in either
$C_{\mathrm{inf}}$ or $\ell_{\mathrm{Cinf}}$ are shown. A prior uniform in
$C_{\mathrm{inf}}$ is biased towards larger values of $\ell_{\mathrm{Cinf}}$.
(B) Updated posterior belief about $\ell_{\mathrm{Cinf}}$ for each of the two
prior beliefs shown in (A), as per Eqns. (7) and (8), after having observed
the ED assay example provided in Section 2.1. While the prior uniform in
$C_{\mathrm{inf}}$ yields a pPLD with a mode of $\ell_{\mathrm{Cinf}}=6.21$,
that for a prior uniform in $\ell_{\mathrm{Cinf}}$ yields a mode of
$\ell_{\mathrm{Cinf}}=6.18$.
### 4.2 Calculation of midSIN’s outputs
One of the graphical outputs of midSIN is the non-normalized PLD of
$\ell_{\mathrm{Cinf}}$ given the number of wells that were infected at each
dilution, $\vec{k}_{J}$, like that shown in Figure 2(left panel), computed as
$\displaystyle\mathcal{U}_{\text{post}}(\,\ell_{\mathrm{Cinf}}\,|\,\vec{k}_{J}\,)$
$\displaystyle=\prod_{j=1}^{J}\ \frac{n_{j}!}{k_{j}!\ (n_{j}-k_{j})!}\ \cdot\
p_{j}^{k_{j}}\ \cdot\ (1-p_{j})^{n_{j}-k_{j}}$ (9) $\displaystyle\text{where}\
\ p_{j}$ $\displaystyle=1-\exp\left[\ -10^{\ell_{\mathrm{Cinf}}}\cdot
V_{\text{inoc}}\cdot\mathcal{D}_{j}\ \right]\ .$ (10)
While $\mathcal{U}_{\text{post}}$ is not the normalized likelihood of
$\ell_{\mathrm{Cinf}}$, its maximum value at its mode
(${\ell_{\mathrm{Cinf}}}_{\text{,mode}}$) is the normalized probability of
observing this particular ED plate outcome ($\vec{k}_{J}$) out of all other
possible plate outcomes, assuming the true, specific infection concentration
in the sample is ${\ell_{\mathrm{Cinf}}}_{\text{,mode}}$.
Another visual output of midSIN is a graphical representation of the
theoretical number of wells that would be infected given the most likely
$\ell_{\mathrm{Cinf}}$, like that shown in Figure 2(right panel). It is
computed following
$N_{\text{wells infected}}(x)=N_{\text{wells
total}}\left[1-\exp\left(-10^{{\ell_{\mathrm{Cinf}}}_{\text{,mode}}}\,V_{\text{inoc}}\,10^{-x}\right)\right]\
,$ (11)
where $x$ is the $\log_{10}$ of the dilution such that $\mathcal{D}=10^{-x}$
is the dilution. It corresponds to the continuous equivalent of this quantity
which is discrete in the ED assay, namely $\mathcal{D}_{i}=10^{-x_{i}}$ which
is the $i^{\mathrm{th}}$ dilution of the sample. As such, $\mathcal{D}_{i}=$
(minimum dilution) $\cdot$ $\text{(dilution factor between columns)}^{i-1}$
where $i\in{[1,J]}$. For example, if the dilution of the least diluted column
is $0.1=10^{-1}$ and the dilution factor between dilutions in the ED assay is
such that it halves the concentration between each dilution, i.e.,
$1/2=2^{-1}=10^{-\log_{10}(2)}\approx 10^{-0.301}$, then
$\mathcal{D}_{i}=10^{-1}\cdot 10^{-0.301\cdot(i-1)}$ such that
$\mathcal{D}_{1}=10^{-1}$, $\mathcal{D}_{2}=10^{-1.301}$,
$\mathcal{D}_{3}=10^{-1.602}$, and so on, such that $x_{1}=1$, $x_{2}=1.301$,
$x_{3}=1.602$, and so on.
In the graphical representation of the ED assay, the edges of the grey bands
flanking the theoretical blue curve correspond to Eqn. (11) wherein
${\ell_{\mathrm{Cinf}}}_{\text{,mode}}$ has been replace by the 68% and 95% CI
values for $\ell_{\mathrm{Cinf}}$. These CI bands _do not_ correspond to the
68% and 95% CI of the expected number of infected wells at each dilution given
${\ell_{\mathrm{Cinf}}}_{\text{,mode}}$.
The sample dilution corresponding to $1\text{\,}\tcid$ estimated based on the
biased RM and SK approximations (right panels) are converted to (left panels)
based on
$$1\text{\,}\tcid$=$\mathrm{e}^{\gamma=0.5772}\text{\,}\idnew$=$1.781\text{\,}\idnew$$
[10, 4]. In contrast, the $\log_{10}(\idnew/\mathrm{m}\mathrm{L})$ computed by
midSIN can be converted to a true (unbiased) estimate of $\log_{10}(\tcid)$
using $$1\text{\,}\tcid$=$1/\ln(2)\text{\,}\idnew$=$1.44\text{\,}\idnew$$ [2].
### 4.3 Infection concentration measures of influenza A virus samples
#### 4.3.1 Cell culture
Madin-Darby canine kidney cells (MDCKs) were cultured in growth media
(complete MEM media with 5% heat-inactivated FBS), in tissue culture treated
T75 flasks, at 37 °C with 5% CO2 and 95% relative humidity. Cells were split
1/10 every 3–4 days or upon reaching approximately 95% confluency. One passage
of cells was expanded for use by both researchers in one experiment to
quantify the 50% tissue culture infectious dose () and plaque forming units ()
of one viral strain.
#### 4.3.2 Viral stocks
Stocks of influenza A/Puerto Rico/8/34 (H1N1) (PR8) and influenza
A/California/4/09 (Cali/09) were stored at -80 °C and thawed on ice
immediately before use. The and of stock viruses was known to both researchers
prior to this study. Serial dilutions were made in MDCK infection media
(complete MEM media with 4.25% BSA) and dilutions were made by each researcher
independently for titering. ‘Researcher A’ and ‘Researcher B’ independently
performed the and assays of one viral strain for one experiment on the same
day using the same viral stock, reagents, and passage of cells. Each
experiment was performed on a separate day (Fig. 4).
#### 4.3.3 Plaque assay
MDCKs were seeded in six-well plates ($5.5\times
10^{5}\text{\,}\mathrm{c}\mathrm{e}\mathrm{l}\mathrm{l}\mathrm{s}\mathrm{/}\mathrm{m}\mathrm{L}$,
$2\text{\,}\mathrm{m}\mathrm{L}\mathrm{/}\mathrm{w}\mathrm{e}\mathrm{l}\mathrm{l}$)
and grown to 90% confluency overnight (37 °C, 5% CO2, 95% relative humidity).
Each six-well plate contained 10-fold serial dilutions plated in singlet as
well as a negative control and five 6-well plates were carried out per
experiment. Cells were washed twice with PBS w/ Ca2+Mg2+ before the addition
of $500\text{\,}\mathrm{\SIUnitSymbolMicro L}$ of viral dilutions per well.
After $1\text{\,}\mathrm{h}$ at room temperature on a rocker, the inoculum was
aspirated, cells were washed with PBS containing Ca2+Mg2+ (PBS w/ Ca2+Mg2+)
(Gibco), and gently covered with $2\text{\,}\mathrm{mL}$ of agarose overlay
(complete media, 4.25% BSA, 0.9% agarose,
$1\text{\,}\mathrm{\SIUnitSymbolMicro}\mathrm{g}\mathrm{/}\mathrm{m}\mathrm{L}$
TPCK-Trypsin). After drying the overlay at room temperature, plates were
inverted and incubated (37 °C, 5% CO2, 95% relative humidity) for
$3\text{\,}\mathrm{d}$ (PR8) or $4\text{\,}\mathrm{d}$ (Cali/09). Plaques were
visualized by staining cells with 0.1% crystal violet solution in 37%
formaldehyde for $30\text{\,}\mathrm{m}\mathrm{i}\mathrm{n}$ and counted by
‘Researcher A’ or ‘Researcher B’ on their respective experiments (Fig. 4).
#### 4.3.4 assay
MDCKs were seeded in 96-well flat bottom plates ($5\times 10^{4}$
cells/$100\text{\,}\mathrm{\SIUnitSymbolMicro L}$,
$100\text{\,}\mathrm{\SIUnitSymbolMicro}\mathrm{L}\mathrm{/}\mathrm{w}\mathrm{e}\mathrm{l}\mathrm{l}$)
and grown to 80% confluency overnight (37 °C, 5% CO2, 95% relative humidity).
For each experiment, 4 replicate wells, at each of 7 different dilutions
separated by a 10-fold dilution, were infected, and the dilution series was
performed 5 times. Cells were washed with PBS w/ Ca2+Mg2+ before the addition
of $100\text{\,}\mathrm{\SIUnitSymbolMicro L}$ of viral dilutions per well.
After $1\text{\,}\mathrm{h}$ at room temperature on a rocker, the inoculum was
aspirated and replaced with $100\text{\,}\mathrm{\SIUnitSymbolMicro L}$ of
infection media containing
$1\text{\,}\mathrm{\SIUnitSymbolMicro}\mathrm{g}\mathrm{/}\mathrm{m}\mathrm{L}$
TPCK-Trypsin. Cells were incubated (37 °C, 5% CO2, 95% relative humidity) for
$3\text{\,}\mathrm{d}$ (PR8) or $4\text{\,}\mathrm{d}$ (Cali/09). Supernatants
were used to do a hemagglutination (HA) assay with chicken red blood cells. HA
assays were performed and read by ‘Researcher A’ or ‘Researcher B’ on their
respective experiments.
#### 4.3.5 Statistical analysis
The data points reported in Figure 4C,D were computed by taking each of the 5
replicates measured with either the PFU, RM, or SK and the 5 replicates
measured via (5 replicates $\times$ 5 replicates = 25 pairs) for each of the 2
experiments by each of the 2 researchers, yielding 100 pairs. For each pair,
the $\log_{10}$ of ratio of either PFU, RM or SK over SIN was computed. The
mean and standard deviation of the resulting 100 $\log_{10}(\mathrm{ratio})$
were computed and are reported in Figure 4C,D. The statistical significance
($p$-value) of the differences between (PFU,RM,SK) and () was computed using
the Mann-Whitney U test (scipy.stats.mannwhitneyu).
## References
* [1] C. A. A. Beauchemin, Y.-I. Kim, Q. Yu, G. Ciaramella, and J. P. DeVincenzo. Uncovering critical properties of the human respiratory syncytial virus by combining in vitro assays and in silico analyses. PLOS ONE, 14(4):e0214708, 15 April 2019. doi:10.1371/journal.pone.0214708.
* [2] W. R. Bryan. Interpretation of host response in quantitative studies on animal viruses. Ann. N. Y. Acad. Sci., 69(4):698–728, 16 December 1957. doi:10.1111/j.1749-6632.1957.tb49710.x.
* [3] R. Dulbecco. Production of plaques in monolayer tissue cultures by single particles of an animal virus. Proc. Natl. Acad. Sci. U.S.A., 38(8):747–752, August 1952. doi:10.1073/pnas.38.8.747.
* [4] Z. Govindarajulu. Statistical techniques in bioassay, chapter 4. The Logit Approach, pages 35–90. Karger, Basel; New York, 2nd edition, 2001. doi:10.1159/isbn.978-3-318-00617-9.
* [5] G. Kärber. Beitrag zur kollecktiven behandlung pharmakologischer reihenversuche. Archiv f. Experiment. Pathol. u. Pharmakol., 162(4):480–483, July 1931. doi:10.1007/BF01863914.
* [6] D. D. LaBarre and R. J. Lowy. Improvements in methods for calculating virus titer estimates from TCID50 and plaque assays. J. Virol. Methods, 96(2):107–126, August 2001. doi:10.1016/S0166-0934(01)00316-0.
* [7] B. A. Mistry, M. R. D’Orsogna, and T. Chou. The effects of statistical multiplicity of infection on virus quantification and infectivity assays. Biophys. J., 114(12):2974–2985, 19 June 2018. doi:10.1016/j.bpj.2018.05.005.
* [8] L. J. Reed and H. Muench. A simple method of estimating fifty per cent endpoints. Am. J. Hygiene, 27(3):493–497, 1 May 1938. doi:10.1093/oxfordjournals.aje.a118408.
* [9] C. Spearman. The method of “right and wrong cases” (constant stimuli) without Gauss’s formula. Br. J. Psychol., II(Part 3):227–242, January 1908. doi:10.1111/j.2044-8295.1908.tb00176.x.
* [10] N. H. Wulff, M. Tzatzaris, and P. J. Young. Monte Carlo simulation of the Spearman-Kaerber TCID50. J. Clin. Bioinforma., 2(1):5, 13 February 2012. doi:10.1186/2043-9113-2-5.
|
# SYNTACTICALLY GUIDED GENERATIVE EMBEDDINGS FOR ZERO-SHOT SKELETON ACTION
RECOGNITION
###### Abstract
We introduce SynSE, a novel syntactically guided generative approach for Zero-
Shot Learning (ZSL). Our end-to-end approach learns progressively refined
generative embedding spaces constrained within and across the involved
modalities (visual, language). The inter-modal constraints are defined between
action sequence embedding and embeddings of Parts of Speech (PoS) tagged words
in the corresponding action description. We deploy SynSE for the task of
skeleton-based action sequence recognition. Our design choices enable SynSE to
generalize compositionally, i.e., recognize sequences whose action
descriptions contain words not encountered during training. We also extend our
approach to the more challenging Generalized Zero-Shot Learning (GZSL) problem
via a confidence-based gating mechanism. We are the first to present zero-shot
skeleton action recognition results on the large-scale NTU-60 and NTU-120
skeleton action datasets with multiple splits. Our results demonstrate SynSE’s
state of the art performance in both ZSL and GZSL settings compared to strong
baselines on the NTU-60 and NTU-120 datasets.
Index Terms— ZSL, skeleton action recognition, VAE, deep learning, language
and vision
## 1 Introduction
Advances in human action recognition have been predominantly driven by the
abundance of online RGB videos. However, with the advent of accurate depth
sensing technologies (e.g. Microsoft Kinect, Intel Real Sense), action
recognition from 3D human skeleton data has also gained traction. Skeleton
representations can be advantageous since they are compact, robustly separate
the action subject (human) from background and enable privacy-preserving
action capture.
The introduction of large scale skeleton action datasets such as NTU-60 [1]
and NTU-120 [2], have allowed researchers to develop high-performance
approaches for skeleton action recognition [3, 4, 5, 6]. However, these
approaches are resource intensive, prone to overfitting and fail to generalize
on classes outside the training set. Therefore, there is a strong motivation
for Zero-Shot Learning (ZSL) approaches in an attempt to readily generalize
across actions outside the training set.
In ZSL, visual representations and corresponding labels for seen classes are
assumed to be available. During test time, the model is evaluated using data
from unseen classes which are not present during training. Typically, side
information (e.g. class attributes) is leveraged to transfer knowledge from
the seen to unseen classes. As a popular approach, ZSL approaches employ a
shared embedding strategy wherein the visual (image or video) features and
semantic attribute features of the corresponding class labels are projected
into a common embedding space [7, 8, 9, 10, 11, 12, 13]. Generative ZSL
approaches present an alternative strategy wherein unseen samples [14] or
features from unseen samples [11] are generated using Generative Adversarial
Networks(GANs). Owing to the instability in training GANs, Variational Auto-
Encoders(VAEs) [15, 16, 17] have also been used for feature generation.
ZSL has been previously explored for skeleton action recognition. In the only
available work [18] (arXiv), embedding based methods [8, 19] are used to align
visual feature embedding of skeleton action sequence with the text embedding
of the descriptive action phrase (e.g ‘take off jacket’, ‘put on glasses’).
The visual features are represented by the final layer features of a skeleton
action recognition model and the action phrase embedding is typically obtained
by pooling the individual embeddings of words comprising the phrase. However,
this approach does not enable alignment of visual embedding with respect to
the individual contributors of phrase semantics - the verb (‘action’) and the
noun(s) (‘participating entities’). This inability is a major shortcoming
since it does not enable generalization, i.e., being able to map the test
action sequences to a description containing novel combinations of verbs and
nouns, some of which might be from training action descriptions themselves.
To address these shortcomings, we propose an approach wherein the visual
embedding is aligned based on the Parts of Speech (PoS) tags (verb, noun) of
the phrasal words. Instead of directly mapping the visual and PoS-wise
embeddings in a discriminative setting [20], we use group (per-PoS, visual)
specific generative models with cross-group latent objective [17] for improved
ZSL performance (Section 2). We also extend our approach to the Generalized
Zero-Shot Learning (GZSL) problem, a more challenging and realistic variant of
ZSL wherein good performance is required from seen and unseen classes. We do
so by incorporating a confidence-based gating mechanism. (Section 2.5). Our
approach enables state-of-the-art performance for ZSL and GZSL compared to
strong baselines on the NTU-60 and the much larger NTU-120 dataset. (Section
4).
The source code and pre-trained models can be accessed at
https://github.com/skelemoa/synse-zsl.
Fig. 1: Architectural diagram for our approach (SynSE). (left) The dotted path
represents the process flow for the GZSL setting while the solid arrows
represent the flow for ZSL. The Generative Multimodal Alignment Module is
detailed on right side. It contains modality VAEs, where Part-of-Speech (PoS)
specific latent generative embeddings $z_{v}$ (verb), $z_{n}$ (noun) are
jointly aligned with segments ($z_{s,v},z_{s,n}$) of latent generative
skeleton embedding $z_{s}$ via cross-modal alignment - refer Section 2 for
more details. Note that the RGB images have been included for reference. Only
the skeleton sequence is provided as input to the network.
## 2 SynSE
### 2.1 Problem definition
Let $D_{tr}=\\{(x_{s}^{tr},y_{s}^{tr})\\}$ denote the set of $N_{tr}$ training
samples where $x_{s}^{tr}$ denotes visual feature embedding of a skeleton
action sequence, $y_{s}^{tr}\in Y_{s}$ is the corresponding member from the
label set of seen classes. On similar lines,
$D_{u}=\\{(x_{s}^{u},y_{s}^{u})\\}$ denotes the set of test samples with the
subscript $u$ standing for unseen. Suppose $\hat{y}$ represents the test time
class prediction. For ZSL, we have $\hat{y}\in Y_{u},Y_{s}\cap
Y_{u}=\emptyset$ while for GZSL, we have $\hat{y}\in Y_{u}\cup Y_{s},Y_{s}\cap
Y_{u}=\emptyset$. For simplicity, we drop the subscript for seen, unseen and
refer to the class names as $y$ and the visual feature embedding as $x_{s}$.
### 2.2 Learning modality-wise latent generative spaces
A crucial requirement for a ZSL approach is the ability to correctly map novel
inputs. For this, we employ a Variational Auto Encoder (VAE) [21] as the base
architecture to learn the generative space of latent representations. A VAE is
trained by maximizing the Evidence Lower Bound (ELBO):
$\mathcal{L}=\mathbb{E}_{q_{\phi}(z|x)}[\log p_{\theta}(x|z)]-\beta
D_{KL}(q_{\phi}(z|x)||p_{\theta}(z))$ (1)
Here, the first term on the right hand side is the reconstruction error and
the second term is the Kullback-Leibler divergence between likelihood
$p_{\theta}(z)$ and the prior $q_{\phi}(z|x)$. $\beta$ is a hyperparameter
which acts as a trade-off between the two error terms. A popular choice for
the prior is the multivariate Gaussian distribution,
$q_{\phi}(z|x)=\mathcal{N}(\mu,\Sigma)$. The VAE maps the input $x$ initially
to representations for $\mu,\Sigma$ and eventually to the randomized latent
representation $z$ via the reparameterization trick [21].
The first stage in our approach involves learning individual latent generative
latent spaces for visual and linguistic representations. This is achieved by
using a VAE for each space. To enable semantically aware compositional
generalization, the text description for class label $y$ is tokenized into
constituent Part-of-Speech (PoS) specific sets - $y_{v}$ for verb and $y_{n}$
for noun. The tokens are encoded using a natural language encoder module to
obtain the corresponding PoS-wise embeddings $e_{v}$ and $e_{n}$ (see Figure
1). Since our approach employs independent VAEs for skeleton ($s$) and
linguistic ($v,n$) representations, the overall cost function for a single
sample can be written as:
$\displaystyle\mathcal{L}_{VAE}=\sum_{m\in\\{s,v,n\\}}\mathbb{E}_{q_{\phi}(z_{m}|x_{m})}[\log
p_{\theta}(x_{m}|z_{m})]-$ (2) $\displaystyle\beta
D_{KL}(q_{\phi}(z_{m}|x_{m})||p_{\theta}(z_{m}))$
### 2.3 Cross-modal alignment
The VAEs optimize latent representations for individual modalities. To achieve
alignment between the skeleton sequence and linguistic latent representations,
a cross-modal reconstruction objective is formulated [17]. First, the latent
embeddings from the PoS embeddings ($z_{v}$, $z_{n}$) are concatenated (see
$z_{l}$ in Figure 1) and the result $x_{l}$ is used to reconstruct the visual
representation via the skeleton representation VAE’s posterior decoder
$D_{s}$. Next, the skeleton sequence latent embedding ($z_{s}$) is uniformly
mapped to as many embeddings ($z_{s,v}$, $z_{s,n}$) as the number of PoS tags.
Complementary to the processing of $z_{l}$, each of the split embedding is
used to reconstruct the corresponding PoS token embedding ($e_{n}$, $e_{v}$)
via the corresponding PoS token embedding’s decoder ($D_{v}$, $D_{n}$).
Overall, the cross-modal reconstruction objective for a training sample is
formulated as:
$\mathcal{L}_{CMR}=|x_{s}-D_{s}(z_{l})|_{2}+\sum_{m\in\\{v,n\\}}|e_{m}-D_{i}(z_{s,m})|_{2}$
(3)
Finally, the VAE loss and the cross-modal reconstruction loss are optimized
together as:
$\mathcal{L}=\mathcal{L}_{VAE}+\alpha\mathcal{L}_{CMR}$ (4)
where $\alpha$ is a trade-off weight factor.
### 2.4 Zero-shot Classification using Latent Embedding
The PoS tag embeddings of each unseen class are respectively transformed by
the PoS encoders ($E_{v},E_{n}$) and used to obtain samples from the latent
generative space ($z_{l}$ \- see Figure 1). A softmax classifier
$f_{u}:z_{l}\rightarrow Y_{u}$ is trained to classify these latent samples
into the unseen classes.
The cross-modal VAE setup described earlier aims to align the visual features
with the language features in the common latent generative space. In other
words, $z_{s}$ and $z_{l}$ are optimized to be interchangeable. Taking
advantage of this, during inference, the unseen class skeleton sequence
representation $x_{s}$ is first obtained. Supplying $x_{s}$ to the visual VAE
encoder ($E_{s}$) enables us to obtain the mean visual latent embedding
($\mu_{s}$) of the sequence111Note that
$z_{s}=\mu_{s}+\Sigma_{s}\odot\mathcal{E}$, where
$\mathcal{E}=\mathcal{N}(0,I)$ by the VAE reparameterization trick.. The
corresponding class prediction is obtained using $\mu_{s}$ and the classifier
$f_{u}$ mentioned previously.
### 2.5 Gating Module for GZSL
For a given skeleton sequence representation $x_{s}$, the probability
distribution $c_{s}$ over seen classes is obtained from the skeleton action
recognition model $f_{{s}}:x_{s}\rightarrow Y_{s}$ from which the action
sequence embedding has been sourced all along. The unseen class classifier
$f_{{u}}:E_{s}(x_{s})\rightarrow Y_{u}$, is a part of our ZSL approach
described in the previous section which provides the unseen class
probabilities $c_{u}$. The probability distribution over all the classes can
be written as:
$\displaystyle
p(y|x)=c_{s}p^{gate}(s;c_{s},c_{u})+c_{u}p^{gate}(u;c_{s},c_{u})$ (5)
Further, we use a gating model (due to its superior performance for GZSL in
other domains) to first decide whether the sample belongs to a seen class or
an unseen class [22]. For this, the seen and unseen class probabilities are
used as features to train a probabilistic binary classifier
$p^{gate}(s;c_{s},c_{u})$ [22]. The resulting outputs are used to determine
the probability distribution over all ($Y_{s}\cup Y_{u}$) classes (Equation
5).
### 2.6 Implementation Details
Method | NTU-60 | NTU-120
---|---|---
$55/5$ split | $48/12$ split | $110/10$ split | $96/24$ split
Jasani [18] (preprint) | $65.53$ | - | - | -
ReViSE [12] | $53.91$ | $17.49$ | $55.04$ | $32.38$
JPoSE [20] | $64.82$ | $28.75$ | $51.93$ | $32.44$
CADA-VAE [17] | $\mathbf{76.84}$ | $28.96$ | $59.53$ | $35.77$
SynSE (ours) | $75.81$ | $\mathbf{33.30}$ | $\mathbf{62.69}$ | $\mathbf{38.70}$
Table 1: ZSL accuracy (%) on the NTU-60 and NTU-120 datasets.
| NTU-60 | NTU-120
---|---|---
Method | ($55/5$) random split | ($48/12$) random split | ($110/10$) random split | ($96/24$) random split
| s | u | h | s | u | h | s | u | h | s | u | h
ReViSE [12] | $74.23$ | $34.73$ | $29.22$ | $62.36$ | $20.77$ | $31.16$ | $48.69$ | $44.84$ | $46.68$ | $49.66$ | $25.06$ | $33.31$
JPoSE [20] | $64.44$ | $50.29$ | $56.49$ | $60.49$ | $20.62$ | $30.75$ | $47.66$ | $46.40$ | $47.05$ | $38.62$ | $22.79$ | $28.67$
CADA-VAE [17] | $69.38$ | $61.79$ | $\mathbf{65.37}$ | $51.32$ | $27.03$ | $35.41$ | $47.16$ | $49.78$ | $48.44$ | $41.11$ | $34.14$ | $37.31$
SynSE | $61.27$ | $56.93$ | $59.02$ | $52.21$ | $27.85$ | $\mathbf{36.33}$ | $52.51$ | $57.60$ | $\mathbf{54.94}$ | $56.39$ | $32.25$ | $\mathbf{41.04}$
SynSE (+ softgating) | $65.17$ | $59.51$ | $62.21$ | $69.23$ | $21.74$ | $33.09$ | $74.76$ | $37.68$ | $50.10$ | $72.54$ | $21.09$ | $32.67$
SynSE (- temp. scaling) | $74.45$ | $37.46$ | $49.84$ | $45.74$ | $25.87$ | $33.05$ | $67.87$ | $38.05$ | $48.77$ | $66.97$ | $25.55$ | $36.99$
SynSE (+ CADA-VAE’s GZSL) | $82.70$ | $0$ | $0$ | $87.63$ | $0$ | $0$ | $80.43$ | $0$ | $0$ | $82.46$ | $0$ | $0$
Table 2: GZSL Accuracy (%) for seen (s) classes, unseen (u) classes and their
harmonic mean (h) on NTU-60 and NTU-120 datasets
Visual and Textual features: The visual features $x_{s}$, are realised using
the $256$ dimensional penultimate layer feature from 4s-ShiftGCN [3], a state-
of-the-art deep network for skeleton action recognition. To maintain the zero-
shot assumption, we train 4s-ShiftGCN only on the seen classes. We use the
Sentence BERT model [23] to obtain 1024-dimensional PoS-wise word embeddings.
Before splitting into verbs and nouns, the class names are modified to fill
the missing PoS tag, e.g. ‘reading’ is changed to ‘reading book’, ‘drop’ to
‘drop object’, ‘headache’ to ‘have headache’. For actions where adding the
missing tag (usually a noun) would be unreasonable (e.g. ‘jump up’, ‘stand
up’), the average of all noun embeddings is used as a placeholder.
Architectural Details: We have a single dense layer as the encoder ($E_{s}$)
and decoder ($D_{s}$), which map the input features ($x_{s},e_{v},e_{n}$) to
the latent space ($z_{s},z_{v},z_{n}$) and vice versa. $x_{s}$ is
$256$-dimensional and $e_{v},e_{n}$ are $1024$-dimensional. The size of the
latent dimension is based on the number of unseen classes. For small number
($5$) of unseen classes, the skeleton latent dimension is set as $100$ and the
latent dimension for the PoS tags is $50$. For larger number of unseen
classes, the latent dimensions are doubled to $200$ and $100$ for the skeleton
latent dimension and PoS tags respectively. The ZSL classifier has a single
dense layer which takes latent features as input and returns the softmax
probabilities for unseen classes.
Training Details: The VAEs within the Generative Multimodal Alignment Module
are optimized using the Adam optimizer with a learning rate of $1e^{-4}$ and a
batch size of $64$. The VAEs are trained using a cyclic annealing schedule
[24] for multiple cycles to mitigate the vanishing KL divergence problem. The
$\beta$ hyperparameter for the KL divergence is turned on after $1000$ epochs,
starting with $0$ and is increased with a rate of $0.0021$ per epoch in each
cycle. Similarly, the $\alpha$ parameter for the cross modal reconstruction is
turned on after $1400$ epochs for experiments on NTU-60 and $1500$ epochs on
NTU-120 and is kept constant with a value of $1$. One cycle is completed in
$1700$ epochs for NTU-60 and $1900$ epochs for NTU-120 dataset. The zero-shot
classifier (Section 2.4) is also optimized using Adam with a learning rate for
$1e^{-3}$. $500$ features per unseen class are generated and the classifier is
trained for $300$ epochs.
The input to the gating model (Section 2.5) is the concatenation of the top
$k$ softmax probabilities from the outputs of the seen and unseen classifiers.
We set $k$ equal to the number of unseen classes and we temperature scale [25]
the seen classifier probabilities as well. The gating model is implemented as
a binary logistic regression classifier and optimized using LBFGS solver from
the scikit-learn library with the default aggressiveness hyperparameter
($C=1$). For training the gating model, we set aside a few samples from the
training set and refer to them as the gating train set. Similarly, we set
aside a few samples from validation set (gating validation set). We train the
gating model using the gating train set and determine the hyperparameters
(temperature coefficient, threshold) using the gating validation set. The
gating module is configured for use in ‘hard’ gating mode wherein
$p^{gate}(s;c_{s},c_{u})$ and $p^{gate}(u;c_{u},c_{u})$ in Equation 5 take
binary values [22].
## 3 Experiments
### 3.1 Datasets
NTU-60 [1]: This is a large-scale dataset curated for 3D human action
analysis. It contains $56{,}880$ samples belonging to $60$ action classes,
with $40$ different subjects captured from $80$ distinct camera viewpoints.
The action sequences of skeleton representations are in the form of 3D
coordinates for $25$ human body joints. We create two splits for ZSL
evaluation - a $55/5$ split with $55$ seen classes, $5$ randomly chosen unseen
classes and a more challenging $48/12$ split.
NTU-120 [2]: NTU-120 builds upon NTU-60 and contains $60$ additional fine-
grained action classes. It contains a total of $114{,}480$ samples spread
across $120$ actions performed by $106$ different subjects captured from $155$
different camera viewpoints. Analogous to the NTU-60 ZSL evaluation setup, we
create two splits - $110$ (seen)/$10$ (unseen) and $96$ (seen)/$24$ (unseen).
### 3.2 Experimental Details
We perform ZSL and GZSL experiments on the NTU-60 and NTU-120 datasets on the
described splits. Since no previous works for skeleton ZSL exist, we modify
representative state-of-the-art approaches from other problem domains and
implement from scratch.
CADA-VAE [17] learns a generative latent space under a cross aligned and
distribution aligned objective. Since we found the distribution alignment
objective to induce instability in training, we omit it during optimization.
ReViSE [12] aims to align the latent embeddings realised via autoencoders
using a Maximum Mean Discrepancy criterion. JPose [20] attempts to learn PoS
aware embeddings of word2vec representations for video retrieval tasks. It
learns a series of progressively refined embeddings under inter/intra modal
constraints in a discriminative setting. For fair comparison, the visual
features and PoS embeddings are the same as ones used in our approach (Section
2.6).
Component | Default in SynSE | Ablation | Accuracy
---|---|---|---
Language Embedding | Sentence-BERT [23] | Word2Vec [26] | $60.76$
Visual Features | 4s-ShiftGCN [3] | MS-G3D [4] | $68.80$
Latent Dimension | $100$ | $50$ | $73.83$
Latent Dimension | $100$ | $200$ | $74.67$
Latent features | $500$ | $250$ | $73.89$
Latent features | $500$ | $1000$ | $73.82$
| original | | $\mathbf{75.81}$
Table 3: SynSE ZSL accuracy (%) on the NTU-60 dataset for various ablations
(55/5 split).
## 4 Results
### 4.1 ZSL results
Table 1 shows the ZSL results of the various approaches on the NTU-60 and
NTU-120 datasets. For the $55/5$ split of NTU-60, the VAE-based generative
approaches significantly outperform the discriminative embedding based
approaches. SynSE’s performance is comparable to that of CADA-VAE.
Predictably, results on the more challenging $48/12$ split show that having a
larger number of unseen classes impacts performance across the board. However,
SynSE offers significant improvement over other baseline approaches, including
CADA-VAE. On the larger NTU-120 dataset, SynSE outperforms other methods on
both the splits.
### 4.2 GZSL results
Since we use a gating-based strategy for GZSL in SynSE (Section 2.5), we
compare against other baselines by incorporating the same strategy.
Specifically, the seen class classifier is kept the same while the specific
baseline approach provides the corresponding unseen class probabilities.
Following standard convention for GZSL, we report the average seen class
accuracy (s), the average unseen class accuracy (u) and their harmonic mean
(h). Table 2 shows the results for datasets and the associated pre-defined
splits. Similar to the trend in ZSL for the $55/5$ split of NTU-60, SynSE
performs poorer compared to CADA-VAE on the harmonic scale for the 55/5 NTU-60
split. However, it outperforms other approaches on the harmonic scale for
other splits of NTU-60 and NTU-120. We also compare our hard gating strategy
[27] with the soft gating based strategy [22]. The results in Table 2 show
that soft gating is biased towards seen classes, resulting in poor harmonic
accuracy. Additionally, Table 2 also shows the significant performance hit
when temperature scaling (Sec. 2.6) is removed [25].
To further demonstrate the effectiveness of our GZSL strategy (i.e. gating
model), we explored an alternative based on the approach used for CADA-VAE
[17], which does not involve gating. As Table 2 shows, the resulting setup
ends up too heavily skewed for seen classes and is unable to classify the
unseen classes.
### 4.3 Ablations
We perform ablation experiments on the 55/5 split of the NTU-60 dataset to
analyse the importance of the building blocks of our alignment module and
design choices affecting its ZSL performance. As the results show (Table 3),
Sentence-BERT is a superior choice to Word2Vec [26] for embedding PoS-tagged
words. Similarly, 4s-ShiftGCN provides better visual embeddings compared to
another state-of-the-art skeleton action recognition model MS-G3D [4]. We
further ablate on the architectural choices by varying the size of the latent
dimension. As shown in Table 3, we see that both an increase and decrease in
the size of the latent embedding causes reduction in ZSL performance. In order
to validate our choice of $500$ latent features per class, we experiment with
varying number of latent features with results as shown in Table 3.
## 5 Conclusion
In this work, we have presented SynSE, a compositional approach for infusing
latent visual representations of skeleton-based human actions with syntactic
information derived from corresponding textual descriptions. We present the
first set of zero-shot skeleton action recognition results on the large-scale
NTU-60 and NTU-120 datasets. Our experiments show that SynSE outperforms
strong baselines for ZSL and the more challenging GZSL setup. Going forward,
we would like to explore the viability of SynSE for zero-shot RGB video action
recognition.
## References
* [1] Amir Shahroudy, Jun Liu, Tian-Tsong Ng, and Gang Wang, “Ntu rgb+ d: A large scale dataset for 3d human activity analysis,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2016, pp. 1010–1019.
* [2] Jun Liu, Amir Shahroudy, Mauricio Lisboa Perez, Gang Wang, Ling-Yu Duan, and Alex Kot Chichung, “Ntu rgb+ d 120: A large-scale benchmark for 3d human activity understanding,” IEEE transactions on pattern analysis and machine intelligence, 2019\.
* [3] Ke Cheng, Yifan Zhang, Xiangyu He, Weihan Chen, Jian Cheng, and Hanqing Lu, “Skeleton-based action recognition with shift graph convolutional network,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2020, pp. 183–192.
* [4] Ziyu Liu, Hongwen Zhang, Zhenghao Chen, Zhiyong Wang, and Wanli Ouyang, “Disentangling and unifying graph convolutions for skeleton-based action recognition,” in The IEEE Conference on Computer Vision and Pattern Recognition (CVPR), June 2020.
* [5] Pengfei Zhang, Cuiling Lan, Junliang Xing, Wenjun Zeng, Jianru Xue, and Nanning Zheng, “View adaptive neural networks for high performance skeleton-based human action recognition,” IEEE Transactions on Pattern Analysis and Machine Intelligence, 2019\.
* [6] Lei Shi, Yifan Zhang, Jian Cheng, and Hanqing Lu, “Two-stream adaptive graph convolutional networks for skeleton-based action recognition,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2019, pp. 12026–12035.
* [7] Zeynep Akata, Florent Perronnin, Zaid Harchaoui, and Cordelia Schmid, “Label-embedding for image classification,” IEEE transactions on pattern analysis and machine intelligence, vol. 38, no. 7, pp. 1425–1438, 2015.
* [8] Andrea Frome, Greg S Corrado, Jon Shlens, Samy Bengio, Jeff Dean, Marc’Aurelio Ranzato, and Tomas Mikolov, “Devise: A deep visual-semantic embedding model,” in Advances in neural information processing systems, 2013, pp. 2121–2129.
* [9] Zeynep Akata, Scott Reed, Daniel Walter, Honglak Lee, and Bernt Schiele, “Evaluation of output embeddings for fine-grained image classification,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2015, pp. 2927–2936.
* [10] Mohammad Norouzi, Tomas Mikolov, Samy Bengio, Yoram Singer, Jonathon Shlens, Andrea Frome, Greg S Corrado, and Jeffrey Dean, “Zero-shot learning by convex combination of semantic embeddings,” arXiv preprint arXiv:1312.5650, 2013.
* [11] Yongqin Xian, Tobias Lorenz, Bernt Schiele, and Zeynep Akata, “Feature generating networks for zero-shot learning,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2018, pp. 5542–5551.
* [12] Yao-Hung Hubert Tsai, Liang-Kang Huang, and Ruslan Salakhutdinov, “Learning robust visual-semantic embeddings,” in Proceedings of the IEEE International Conference on Computer Vision, 2017, pp. 3571–3580.
* [13] Tanmoy Mukherjee, Makoto Yamada, and Timothy M Hospedales, “Deep matching autoencoders,” arXiv preprint arXiv:1711.06047, 2017.
* [14] Scott Reed, Zeynep Akata, Honglak Lee, and Bernt Schiele, “Learning deep representations of fine-grained visual descriptions,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2016, pp. 49–58.
* [15] Vinay Kumar Verma, Gundeep Arora, Ashish Mishra, and Piyush Rai, “Generalized zero-shot learning via synthesized examples,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2018, pp. 4281–4289.
* [16] Ashish Mishra, Shiva Krishna Reddy, Anurag Mittal, and Hema A Murthy, “A generative model for zero shot learning using conditional variational autoencoders,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, 2018, pp. 2188–2196.
* [17] Edgar Schönfeld, Sayna Ebrahimi, Samarth Sinha, Trevor Darrell, and Zeynep Akata, “Generalized zero-shot learning via aligned variational autoencoders,” red, vol. 2, pp. D2, 2019.
* [18] Bhavan Jasani and Afshaan Mazagonwalla, “Skeleton based zero shot action recognition in joint pose-language semantic space,” arXiv preprint arXiv:1911.11344, 2019.
* [19] Flood Sung, Yongxin Yang, Li Zhang, Tao Xiang, Philip HS Torr, and Timothy M Hospedales, “Learning to compare: Relation network for few-shot learning,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2018, pp. 1199–1208.
* [20] Michael Wray, Diane Larlus, Gabriela Csurka, and Dima Damen, “Fine-grained action retrieval through multiple parts-of-speech embeddings,” in IEEE/CVF International Conference on Computer Vision (ICCV), 2019\.
* [21] Diederik P Kingma and Max Welling, “Auto-encoding variational bayes,” arXiv preprint arXiv:1312.6114, 2013.
* [22] Yuval Atzmon and Gal Chechik, “Adaptive confidence smoothing for generalized zero-shot learning,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2019, pp. 11671–11680.
* [23] Nils Reimers and Iryna Gurevych, “Sentence-bert: Sentence embeddings using siamese bert-networks,” arXiv preprint arXiv:1908.10084, 2019.
* [24] Hao Fu, Chunyuan Li, Xiaodong Liu, Jianfeng Gao, Asli Celikyilmaz, and Lawrence Carin, “Cyclical annealing schedule: A simple approach to mitigating kl vanishing,” arXiv preprint arXiv:1903.10145, 2019.
* [25] Geoffrey Hinton, Oriol Vinyals, and Jeff Dean, “Distilling the knowledge in a neural network,” arXiv preprint arXiv:1503.02531, 2015.
* [26] Tomas Mikolov, Kai Chen, Greg Corrado, and Jeffrey Dean, “Efficient estimation of word representations in vector space,” arXiv preprint arXiv:1301.3781, 2013.
* [27] Richard Socher, Milind Ganjoo, Christopher D Manning, and Andrew Ng, “Zero-shot learning through cross-modal transfer,” in Advances in neural information processing systems, 2013, pp. 935–943.
|
[0]RY is partially funded by NSF DMS 1916037. HM is partially funded by JSPS
KAKENHI 19H04987, 19H05024 and 19K06957. KM is partially funded by JSPS
KAKENHI 18K11485
url]polytopes.net
1]organization=Department of Operations Research, Naval Postgraduate School,
city=Monterey, postcode=93943, state=CA, country=USA
2]organization=Graduate School of Science and Technology, Kwansei Gakuin
University, city=Sanda, postcode=669-1337, state=Hyogo, country=Japan
3]organization=Graduate School of Medicine, Osaka City University,
postcode=558-8585, state=Osaka, country=Japan
[orcid=0000-0002-9258-6541]
[1] [1]Corresponding author
# Tropical Support Vector Machines: Evaluations and Extension to Function
Spaces
Ruriko Yoshida<EMAIL_ADDRESS>[ Misaki Takamori [ Hideyuki Matsumoto [
Keiji Miura<EMAIL_ADDRESS>
###### Abstract
Support Vector Machines (SVMs) are one of the most popular supervised learning
models to classify using a hyperplane in an Euclidean space. Similar to SVMs,
tropical SVMs classify data points using a tropical hyperplane under the
tropical metric with the max-plus algebra. In this paper, first we show
generalization error bounds of tropical SVMs over the tropical projective
torus. While the generalization error bounds attained via Vapnik-Chervonenkis
(VC) dimensions in a distribution-free manner still depend on the dimension,
we also show numerically and theoretically by extreme value statistics that
the tropical SVMs for classifying data points from two Gaussian distributions
as well as empirical data sets of different neuron types are fairly robust
against the curse of dimensionality. Extreme value statistics also underlie
the anomalous scaling behaviors of the tropical distance between random
vectors with additional noise dimensions. Finally, we define tropical SVMs
over a function space with the tropical metric.
###### keywords:
Extreme Value Statistics Function Spaces Max-plus Algebra Supervised Learning
Tropical Geometry
We obtained generalization error bounds of tropical Support Vector Machines
(SVMs) via the Vapnik-Chervonenkis dimensions of tropical hyperplanes using
Tropical Radon Lemma by Jaggi et al. (2008).
We demonstrated theoretically by extreme value statistics that the tropical
SVMs are robust against the curse of dimensionality for classifying data
points from two Gaussian distributions as well as experimentally recorded
activities from different neuron types.
We defined tropical SVMs over a function space to enable the classification
of curves, which is a task we encounter frequently in practice, such as
classifications on neuronal tuning curves.
## 1 Introduction
In data science, one of the well-known challenges we face is to classify data
points with a large number of predictors. For example, in image processing, in
order to discriminate one object, such as missiles or face recognition, from
others in images, these images are described as a vector in pixels which are
typically in a very high dimensional vector space (Hung-Chih Chiang et al.
(2000)). In bioinformatics, researchers try to classify particular diseases
using high dimensional data sets such as micorarrays or SNPs (Fan and Ren
(2006)). Fan and Ren (2006) summarize challenges and difficulties with high
dimensionality in classification.
Support Vector Machines (SVMs) are one of well-known supervised learning
models to classify data points using a hyperplane and introduced by Boser et
al. (1992) and Cortes and Vapnik (1995). The classical SVMs introduced by
Boser et al. (1992) can be written as the $L_{2}$ norm SVMs, that is a hinge
loss plus the $L_{2}$ norm penalty formulation. Zhang and Zhou (2010) showed
that including many redundant features can cause difficulties in the
performance of $L_{2}$ norm SVM. Bradley and Mangasarian (1998) showed that
the SVM with the $L_{1}$ norm penalty instead of the $L_{2}$ norm penalty
works well with variable selection as well as classification at the same time.
This is called $L_{1}$ norm SVMs. While there are several frameworks to
analyze the generalization performance, such as covering numbers (Mohri et
al., 2018), real log canonical thresholds (Hayashi and Watanabe, 2017), mean-
field regime (Nitanda et al., 2021), and Langevin dynamics regime (Suzuki,
2020), the SVMs especially allow us to derive the upper bounds for the
generalization error via Vapnik-Chervonenkis (VC) dimensions (Vapnik, 2000).
Although the bounds via VC dimensions may not necessarily be tight, it is
still significant to have the first bound for a new variant of SVMs. Peng et
al. (2016) showed the generalization bound on the probability of errors of the
$L_{1}$ norm SVMs and it still depends on the number of predictors, i.e., the
dimension of the normal vector of the hyperplane.
Because of advances in computational algebraic geometry (Ren (2015)), tropical
geometry finds applications in data science. For example, it can be applied to
principal component analysis (Page et al. (2020); Yoshida et al. (2019)) and
Bayesian Networks (Améndola et al. (2020)). With these statistical
methodologies, tropical geometry is applied to phylogenomics analyses on
Apicomplexa, African coelacanth whole genome data sets, and 1089 full length
sequences of hemagglutinin (HA) for influenza A H3N2 from 1993 to 2017 in the
state of New York obtained from the GI-SAID EpiFlu (Page et al. (2020);
Yoshida et al. (2019)). Gärtner and Jaggi (2006) applied tropical geometry to
SVMs. Instead of using a hyperplane defined by the $L_{2}$ metric, Gärtner and
Jaggi (2006) used a tropical hyperplane with the tropical metric with the min-
plus algebra to classify the data points. Same as the $L_{2}$ norm SVMs,
tropical SVMs are the tropicalization of the $L_{2}$ norm hard margin SVMs
which maximizes the margin, the distance from the tropical hyperplane to the
closest data points in terms of the min-plus algebra. Gärtner and Jaggi (2006)
also showed that a hard margin tropical SVM can be formulated as a linear
programming problem. Then, Tang et al. (2020) showed necessary and sufficient
conditions of the optimal solutions of the linear programming problem for
finding a hard margin tropical SVM if it is feasible in terms of the max-plus
algebra. They also introduced soft margin tropical SVMs and showed that
finding a soft margin tropical SVM can be formulated as a linear programming
problem.
While there are some developments in computational sides of tropical SVMs,
there has not been much in its theoretical evaluation and statistical
analysis. For example, it is unclear how the tropical metric with the max-plus
algebra and tropical SVMs handle the curse of dimensionality. Whether tropical
SVMs are robust against the curse of dimensionality is an important question
to answer, because we can extend tropical SVMs to various types of data
including a function space. In fact, there are many classification problems
with functions as predictors, such as neuronal tuning curves (Dayan and Abbott
(2001)), instead of vectors.
In this paper we focus on hard margin tropical SVMs with the max-plus algebra.
Like the $L_{2}$ norm and the $L_{1}$ norm SVMs, we assume that there exists
the optimal tropical hyperplane defined by the normal vector
$\omega\in\mathbb{R}^{d}\\!/\mathbb{R}{\bf 1}$ such that the probability of
the loss function being positive equals to zero. In addition, in order for the
tropical metric to be well defined, we consider the _tropical projective
torus_ , that is, the projective torus $\mathbb{R}^{d}\\!/\mathbb{R}{\bf 1}$,
where ${\bf 1}:=(1,1,\ldots,1)$ is defined as the all-one vector. Note that
any vector $(v_{1},\ldots,v_{d})\in\mathbb{R}^{d}\\!/\mathbb{R}{\bf 1}$ is
equal to $(v_{1}+c,\ldots,v_{d}+c)$ with any scalar $c\in\mathbb{R}$ and
$\mathbb{R}^{d}\\!/\mathbb{R}{\bf 1}$ is isometric to $\mathbb{R}^{d-1}$. In
applications, the tropical projective torus is useful for subtracting the
baseline $\mathbb{R}{\bf 1}$ components from feature vectors, as we will see.
Our main contribution of this paper primarily consists of two items: (1)
evaluation of generalization errors for tropical SVMs using VC dimensions for
tropical hyperplanes over $\mathbb{R}^{d}\\!/\mathbb{R}{\bf 1}$; and (2)
extension to tropical SVMs on a function space, which consists of the tropical
metric and a set of functions from a multi-dimensional vector space to a real
number.
While generalization error bounds of tropical SVMs over
$\mathbb{R}^{d}\\!/\mathbb{R}{\bf 1}$ still depends on the dimension, our
simulations with a set of Gaussian distribution functions show that errors
rates of tropical SVMs on $\mathbb{R}^{d}\\!/\mathbb{R}{\bf 1}$ grow much
slower than ones with $L_{2}$ norm SVMs when we increase the number of
predictors while we fix the sample sizes. Then we extend a notion of tropical
SVMs to a function space with the tropical metric. In fact, we show that a set
of all functions from a multi-dimensional vector space to a real number with
the tropical metric is a normed vector space.
This manuscript is organized as follows: In Section 2, we set up basics in
tropical geometry under the max-plus algebra. In Section 3, as an application,
we show anomalous scaling behaviors of tropical distances. In Section 4, we
set up tropical SVMs with the tropical metric over the tropical projective
torus as linear programming problems and show the generalization error bounds
via the VC dimension of a tropical hyperplane. In Section 5, as an
application, we show theoretically by the extreme value statistics the
robustness of the tropical SVMs against the curse of dimensionality in a case
of two different multivariate Gaussian distributions and empirical data set of
neuron types. In Section 6, we first define the tropical distance over a
function space. Then we show that the tropical distance between functions is
metric and we also show that a function space with the tropical metric forms a
normed vector space. In addition, we define tropical SVMs over a function
space with the tropical metric.
## 2 Definitions of Tropical Distances and Hyperplanes
In this section, we remind readers some basics in tropical arithmetic and
algebra using the max-plus algebra. Through this paper, we consider the
tropical projective torus $\mathbb{R}^{d}\\!/\mathbb{R}{\bf 1}$ which is
isometric to $\mathbb{R}^{d-1}$. For more details, see Joswig (2022); Maclagan
and Sturmfels (2015); Tang et al. (2020).
###### Definition 2.1 (Tropical Arithmetic Operations).
In this tropical semiring $(\,\mathbb{R}\cup\\{-\infty\\},\boxplus,\odot)\,$,
the basic tropical arithmetic operations of addition and multiplication are
defined as:
$a\boxplus b:=\max\\{a,b\\},~{}~{}~{}~{}a\odot b:=a+b~{}~{}~{}~{}\mbox{ where
}a,b\in\mathbb{R}\cup\\{-\infty\\}.$
Over this semiring, $-\infty$ is the identity element under addition and 0 is
the identity element under multiplication.
###### Definition 2.2 (Tropical Scalar Multiplication and Vector Addition).
For any scalars $a,b\in\mathbb{R}\cup\\{-\infty\\}$ and for any vectors
$v=(v_{1},\ldots,v_{d}),w=(w_{1},\ldots,w_{d})\in(\mathbb{R}\cup-\\{\infty\\})^{d}$,
we define tropical scalar multiplication and tropical vector addition as
follows:
$a\odot v\boxplus b\odot
w:=(\max\\{a+v_{1},b+w_{1}\\},\ldots,\max\\{a+v_{d},b+w_{d}\\}).$
###### Definition 2.3.
Suppose $S\subset\mathbb{R}^{d}\\!/\mathbb{R}{\bf 1}$. If
$a\odot v\boxplus b\odot w\in S$
for any $a,b\in\mathbb{R}$ and for any $v,w\in S$, then $S$ is called
tropically convex.
###### Definition 2.4 (Tropical Convex Hull).
Suppose we have a finite subset
$V=\\{v^{1},\ldots,v^{s}\\}\subset\mathbb{R}^{d}\\!/\mathbb{R}{\bf 1}$. The
tropical convex hull or tropical polytope of $V$ is the smallest tropically-
convex subset containing $V$. It can be written as the set of all tropical
linear combinations of $V$ such that:
$\mathrm{tconv}(V)=\\{a_{1}\odot v^{1}\boxplus a_{2}\odot
v^{2}\boxplus\cdots\boxplus a_{s}\odot v^{s}\mid
a_{1},\ldots,a_{s}\in\mathbb{R}\\}.$
A tropical polytope of a set of two points
$\\{v^{1},\,v^{2}\\}\subset\mathbb{R}^{d}\\!/\mathbb{R}{\bf 1}$ is called a
tropical line segment between two points $v^{1},\,v^{2}$.
Note that over the tropical projective torus $\mathbb{R}^{d}\\!/\mathbb{R}{\bf
1}$, any point $x=(x_{1},\ldots,x_{d})\in\mathbb{R}^{d}\\!/\mathbb{R}{\bf 1}$
can be also written as
$(x_{1},\ldots,x_{d})=(x_{1}+c,\ldots,x_{d}+c)$
for any $c\in\mathbb{R}$ by the definition of taking mod by ${\bf
1}:=(1,\ldots,1)$. Therefore, we can assume that
$(x_{1},\ldots,x_{d})=(x_{1}-x_{d},\ldots,x_{d-1}-x_{d},0)$
for any point $x:=(x_{1},\ldots,x_{d})\in\mathbb{R}^{d}\\!/\mathbb{R}{\bf 1}$.
Therefore over this manuscript we assume that the last coordinate of any point
in the tropical projective torus $\mathbb{R}^{d}\\!/\mathbb{R}{\bf 1}$ is
equal to $0$ by taking this normalization.
Figure 1: The tropical line segments between the origin and the three example
points in $\mathbb{R}^{3}\\!/\mathbb{R}{\bf 1}$ are represented by the blue,
green, and red lines. The tropical distances from the origin to the three
example points are $3$ (blue), $2$ (green), and $3$ (red).
###### Example 2.5 (Examples of Tropical Line Segments).
Suppose we have
$(2,3,0),\,(2,1,0),\,(2,-1,0)\in\mathbb{R}^{d}\\!/\mathbb{R}{\bf 1}$. Then,
the tropical line segments between the origin
$(0,0,0)\in\mathbb{R}^{d}\\!/\mathbb{R}{\bf 1}$ and these three points in
$\mathbb{R}^{d}\\!/\mathbb{R}{\bf 1}$ are drawn by blue, green, and red lines
in Figure 1.
###### Definition 2.6 (Generalized Hilbert Projective Metric).
For any $v,\,w\in\mathbb{R}^{d}\\!/\mathbb{R}{\bf 1}$ such that
$v=(v_{1},\ldots,v_{d})$ and $w=(w_{1},\ldots,w_{d})$, the tropical distance
$d_{\rm tr}$ between them is defined such that:
$d_{\rm
tr}(v,w):=\max_{i}\bigl{\\{}v_{i}-w_{i}\bigr{\\}}-\min_{i}\bigl{\\{}v_{i}-w_{i}\bigr{\\}}.$
###### Remark 2.7.
The tropical metric $d_{\rm tr}$ is a metric over
$\mathbb{R}^{d}\\!/\mathbb{R}{\bf 1}$.
###### Example 2.8.
The tropical distances from the origin $(0,0,0)$ to the three example points
in Figure1 in $\mathbb{R}^{3}\\!/\mathbb{R}{\bf 1}$ are
$d_{\rm tr}((2,3,0),(0,0,0))=3-0=3,$ $d_{\rm
tr}((2,1,0),(0,0,0))=2-{\color[rgb]{0,0,0}0=2},$ $d_{\rm
tr}((2,-1,0),(0,0,0))=2-(-1)=3.$
As one of many geodesics (shortest paths) toward the blue (green, red) point,
the tropical line segment is denoted by the blue (green, red) line.
Intuitively, a tropical distance is a shortest path length as far as you are
allowed to go only parallel to the dotted lines.
###### Definition 2.9 (Tropical Hyperplane, Joswig (2005)).
For any
$\omega:=(\omega_{1},\ldots,\omega_{d})\in\mathbb{R}^{d}\\!/\mathbb{R}{\bf
1}$, the tropical hyperplane defined by $\omega$ is the set of points
$x\in\mathbb{R}^{d}\\!/\mathbb{R}{\bf 1}$ such that
$\begin{array}[]{l}H_{\omega}:=\Big{\\{}(x_{1},\ldots,x_{d})\in\mathbb{R}^{d}\\!/\mathbb{R}{\bf
1}|\exists i,j\in\\{1,\ldots,d\\}\mbox{ such that }\\\
\omega_{i}+x_{i}=\omega_{j}+x_{j}=\max\\{\omega_{1}+x_{1},\ldots\omega_{d}+x_{d}\\}\Big{\\}}.\\\
\end{array}$
In addition, we call $\omega$ the normal vector of the tropical hyperplane
$H_{\omega}$.
###### Remark 2.10.
In terms of tropical geometry, $H_{\omega}$ is the solutions of the tropical
linear function with unknown $x_{1},\ldots,x_{d}$ such that for a fixed
$\omega=(\omega_{1},\ldots,\omega_{d})\in\mathbb{R}^{d}\\!/\mathbb{R}{\bf 1}$,
$\displaystyle\omega_{1}\odot x_{1}\boxplus\ldots\boxplus\omega_{d}\odot
x_{d}\mbox{ for }x=(x_{1},\ldots,x_{d})\in\mathbb{R}^{d}\\!/\mathbb{R}{\bf
1},$
which is a special case of tropical polynomials. For more details on finding
solutions of a tropical polynomial, see Maclagan and Sturmfels (2015).
###### Example 2.11 (Tropical Hyperplane in $\mathbb{R}^{3}\\!/\mathbb{R}{\bf
1}$).
Suppose we have $\omega=(1,2,0)\in\mathbb{R}^{3}\\!/\mathbb{R}{\bf 1}$. Then
$\begin{array}[]{rcl}H_{\omega}:&=&\Big{\\{}(x_{1},x_{2},x_{3})\in\mathbb{R}^{3}\\!/\mathbb{R}{\bf
1}|1+x_{1}=2+x_{2}\geq x_{3}\mbox{ or}\\\ &&1+x_{1}=x_{3}\geq 2+x_{2}\mbox{ or
}2+x_{2}=x_{3}\geq 1+x_{1}\Big{\\}}.\\\ \end{array}$
This means that we have four cases:
1. 1.
Case 1: The first term and the second term are equal and max.
$1+x_{1}=2+x_{2}>x_{3}(=0)$.
2. 2.
Case 2: The first term and the third term are equal and max.
$2+x_{2}<1+x_{1}=x_{3}(=0)$.
3. 3.
Case 3: The second term and the third term are equal and max.
$1+x_{1}<2+x_{2}=x_{3}(=0)$.
4. 4.
Case 4: All terms are equal and max. $1+x_{1}=2+x_{2}=x_{3}(=0)$.
We can set $x_{3}=0$ without loss of generality, since $(x_{1},x_{2},x_{3})$
and $(x_{1}-x_{3},x_{2}-x_{3},0)$ represent the same point in
$\mathbb{R}^{3}\\!/\mathbb{R}{\bf 1}$. Thus $H_{(1,2,0)}$ is the set of
$(x_{1},x_{2},0)$ on the three half lines in Figure 2.
Figure 2: Tropical hyperplane defined by $\omega=(1,2,0)$ over
$\mathbb{R}^{3}\\!/\mathbb{R}{\bf 1}$.
###### Definition 2.12 (Sectors of Tropical Hyperplane, Joswig (2005)).
Each tropical hyperplane $H_{\omega}$ divides the tropical projective torus
$\mathbb{R}^{d}\\!/\mathbb{R}{\bf 1}$ into $n$ connected components. These
connected components are open sectors defined as:
$S_{\omega}^{i}~{}:=~{}\\{x\in\mathbb{R}^{d}\\!/\mathbb{R}{\bf
1}\;|\;\omega_{i}+x_{i}>\omega_{j}+x_{j},\;\forall j\neq
i\;\\},\;\;\\!i\\!=\\!1,\\!\ldots\\!,d.$
###### Definition 2.13 (Tropical Distance to a Tropical Hyperplane).
The tropical distance $d_{\rm tr}$ from a point
$x\in\mathbb{R}^{d}\\!/\mathbb{R}{\bf 1}$ to a tropical hyperplane
$H_{\omega}$ is:
$d_{\rm tr}(x,H_{\omega})\;:=\;\min\\{d_{\rm tr}(x,y)\;|\;y\in H_{\omega}\\}.$
###### Proposition 2.14 (Lemma 2.1 in Gärtner and Jaggi (2006)).
Let $H_{\bf 0}$ be the tropical hyperplane defined by the zero vector ${\bf
0}=(0,0,\ldots,0)\in{\mathbb{R}}^{d}\\!/\mathbb{R}{\bf 1}$. For any
$x=(x_{1},\ldots,x_{d})\in\mathbb{R}^{d}\\!/\mathbb{R}{\bf 1}$,
$d_{\rm tr}(x,H_{\bf 0})={\rm max}(x)-{\rm second\;max}(x).$
###### Corollary 2.15 (Corollary 2.3 in Gärtner and Jaggi (2006)).
For any $\omega\in{\mathbb{R}}^{d}\\!/\mathbb{R}{\bf 1}$ and for any
$x\in\mathbb{R}^{d}\\!/\mathbb{R}{\bf 1}$,
$d_{\rm tr}(x,H_{\omega})=d_{\rm tr}(\omega+x,H_{\bf 0}).$
###### Example 2.16.
We use the sample $H_{\omega}$ from Example 2.11. Let $d=3$ and suppose
$\omega=(1,2,0)\in\mathbb{R}^{3}\\!/\mathbb{R}{\bf 1}$. In addition, suppose
we have a point $x=(1,1,0)\in\mathbb{R}^{3}\\!/\mathbb{R}{\bf 1}$. By
Corollary 2.15,
$d_{\rm tr}(x,H_{\bf\omega})=d_{\rm tr}(x+\omega,H_{\bf 0})=d_{\rm
tr}((2,3,0),H_{\bf 0})=3-2=1.$
Figure 3: Tropical and Euclidean distances between two bell-shaped tuning
curves which are horizontally shifted (left column) or vertically lifted
(right column). The tropical distance is insensitive to the vertical lift of
tuning curves.
## 3 Application 1: Valuable Scaling Behaviors of Tropical Distances
### 3.1 Bell-shaped Tuning Curves
To characterize the properties of the tropical distances $d_{\rm tr}(v,w)$
proposed in this paper, we computed the tropical distances for the
mathematical models of neuronal tuning curves as benchmarks. The tuning curves
are the vectors consisting of neural responses to different stimuli. For
example, the responses of neurons in the primary visual cortex to oriented
bars on the screens as visual stimuli are known to show famous bell-shaped
curves as a function of the orientation.
We modeled neuronal tuning curves basically as standard Gaussian distribution
functions and examined how the distances scale when one of two curves is
horizontally shifted or vertically lifted. Both the tropical and Euclidean
distances between two bell-shaped tuning curves increased with the increasing
horizontal shift between them (Figure 3, left). Meanwhile, the tropical
distances between the two bell-shaped tuning curves was insensitive to the
vertical lift of tuning curves, while the Euclidean distance was not (Figure
3, right). The insensitivity to the vertical shift originates from the
definition of the tropical distances $d_{\rm tr}(v,w)$ where tuning curves are
considered in $\mathbb{R}^{d}\\!/\mathbb{R}{\bf 1}$ and
$(1,1,\ldots,1)$-direction is collapsed and ignored in Def.2.6. That is,
$d_{\rm tr}(v+c{\bf 1},w+d{\bf 1})=d_{\rm tr}(v,w)$ for any $c$ and $d$ in
$\mathbb{R}$.
Figure 4: Tropical and Euclidean distances between random and flat tuning
curves. The tropical distance scales with $\log n$ as a result of extreme
value statistics. The circles denote simulation results and the lines denote
theoretical results.
Note that the entire lifting of tuning curves are often observed as a result
of drifting background neural activities (Okun et al. (2015); Stringer et al.
(2019); Miura et al. (2005, 2006, 2007); Miura (2013); Takahashi et al.
(2019)). Thus, practically, the tropical distance can be useful for the sake
of ignoring the temporal drift of tuning curves.
### 3.2 Random Tuning Curves
Next we considered a random tuning curve consisting of noisy responses. When
we computed the distances between a random tuning curve
$v=(5,-x_{2},-x_{3},\ldots,-x_{n}),$ (1)
where $x_{i}\sim\textrm{Exp(1)}$ and a flat tuning curve,
$w=(0,0,0,\ldots,0),$ (2)
the tropical distances was relatively insensitive to the additional noise
dimensions. That is, the tropical distance scaled with $\log n$ while the
Euclidean distance scaled with $\sqrt{n}$. The anomalous scaling for the
tropical distance can be explained by the extreme value statistics (Gumbel
(2004)) as,
$\displaystyle d_{\rm tr}(v,w)$ $\displaystyle:=$
$\displaystyle\max_{i}\bigl{\\{}v_{i}-w_{i}\bigr{\\}}-\min_{i}\bigl{\\{}v_{i}-w_{i}\bigr{\\}}$
(3) $\displaystyle\approx$ $\displaystyle
5+\max_{i}\bigl{\\{}x_{i}\bigr{\\}}.$
Then, because $\max_{i}\bigl{\\{}x_{i}\bigr{\\}}-\log
n\sim\textrm{Gumbel(0,1)}$ (Gumbel (2004)), the expectation is given as
$\overline{d_{\rm tr}(v,w)}=5+\gamma+\log n.$ (4)
This theoretical result explains the simulation result in Figure 4. In this
way, the extreme value statistics plays the crucial role in the computation of
the tropical distance between random vectors.
## 4 Evaluation of Tropical Support Vector Machines (SVMs)
### 4.1 $L_{2}$-norm SVMs
In this section we provide an overview of the $L_{2}$ norm SVMs. Let
$\mathcal{D}_{2}$ be the distribution on the joint random variable $(X,Y)$ for
$X\in\mathbb{R}^{d}$ and $Y\in\\{-1,1\\}$ and let $\mathcal{S}_{2}$ be the
sample $\mathcal{S}_{2}:=\\{(X^{1},Y^{1}),\ldots,(X^{n},Y^{n})\\}$. Then the
standard linear SVM is a statistical model by solving the following
regularization problem:
$\min_{\omega}\left(\underbrace{\lambda||\omega||^{2}}_{\text{regularizer}}+\underbrace{n^{-1}\sum_{i=1}^{n}(1-Y^{i}((X^{i})^{T}\omega+\omega_{0}))_{+}}_{\text{error}}\right),$
where $||x||$ is the $L_{2}$ norm of a vector $x\in\mathbb{R}^{d}$,
$(1-u)_{+}=\max\\{1-u,0\\}$ is the hinge loss function, $\lambda$ is a tuning
parameter, $\omega_{0}$ is a constant term of the hyperplane, and $\omega$ is
the normal vector of the hyperplane to separate the data points. Vapnik (2000)
discussed the generalization bounds on the $L_{2}$ norm SVMs using Theorem
4.11 and the VC dimension of the $L_{2}$ norm SVMs. In addition, using the
Rademacher complexity of $L_{2}$ norm of unit vectors (Bartlett and Mendelson
(2003)), one can show the following generalization bound of the error rate:
###### Theorem 4.1.
Let $w^{s}$ be an output from the classical hard margin SVM from a sample
$\mathcal{S}$. Here $Y\in\\{-1,1\\}$ with the distribution $\pi_{+}=P(Y=1)$
and $\pi_{-}=P(Y=-1)$. Let $X\in\mathbb{R}^{d}$ be a random variable and
$\mathcal{D}_{2}$ be the distribution for a random variable $(X,Y)$. Let
$||x||$ be the $l_{2}$ norm of a vector $x\in\mathbb{R}^{d}$. Assuming that
$||X||\leq R$ with probability 1 and there exists $\omega^{*}$ with
$P_{\mathcal{D}_{2}}(Y^{i}((X^{i})^{T}\omega+\omega_{0})\geq 1)=1$. Then, for
any $\eta>0$ with the probability greater than or equal to $1-\eta$ we have
$P_{\mathcal{D}_{2}}\left(Y\not=sign(X\cdot\omega^{s})\right)\leq\frac{4R||\omega^{s}||}{\sqrt{n}}+(1+2R||\omega^{s}||)\sqrt{\frac{2\log(4||\omega^{s}||/\eta)}{n}}.$
(5)
There has been much work done in tighter generalization bounds on the $L_{2}$
norm SVMs using the VC dimension. As it is not this paper’s focus to discuss
the details on the generalization bounds for the $L_{2}$ norm SVMs, for more
details, see Burges (1998); Guermeur (2007) and the references therein.
###### Remark 4.2.
Recall that the $L_{p}$ norm SVMs, for $0\leq p\leq\infty$, can be written as
the following optimization problem:
$\min_{\omega}\left(\underbrace{\lambda||\omega||_{p}^{2}}_{\text{regularizer}}+\underbrace{n^{-1}\sum_{i=1}^{n}(1-Y^{i}((X^{i})^{T}\omega+\omega_{0}))_{+}}_{\text{error}}\right),$
where $||x||_{p}$ is the $L_{1}$ norm of a vector $x\in R^{d}$,
$(1-u)_{+}=\max\\{1-u,0\\}$ is the hinge loss function, $\lambda$ is a tuning
parameter, and $\omega$ is the normal vector of the hyperplane to separate the
data points.
### 4.2 Tropical SVMs
In this section we provide an overview the hard margin tropical SVMs
originally developed by Gärtner and Jaggi (2006) with the random variable
$X\in\mathbb{R}^{d}\\!/\mathbb{R}{\bf 1}$ given the response variable
$Y\in\\{0,1\\}^{d}$ for one-hot encoding. Note that tropical SVMs can perform
multiclass classification naturally. Before we formally define the hard margin
tropical SVMs, we need to define some notation.
Let $S(x)$ be a set of indices of nonzero elements in a vector
$x\in\mathbb{R}^{d}\\!/\mathbb{R}{\bf 1}$, i.e., $S(x)\subset\\{1,\ldots,d\\}$
where $x_{i}\not=0$. Let $I_{\omega}(x)\in\\{0,1\\}^{d}$ be a vector of
indicator functions of index set $\\{1,\ldots,d\\}$ of a vector $x$ with a
vector $\omega\in\mathbb{R}^{d}\\!/\mathbb{R}{\bf 1}$ where
$I_{\omega}(x)_{i}=\begin{cases}1&\text{if }x_{i}+\omega_{i}=\max(x+\omega)\\\
0&\text{otherwise.}\end{cases}$
Let $J_{\omega}{\color[rgb]{0,0,0}(x)}\in\\{0,1\\}^{d}$ be also a vector of
indicator functions of index set $\\{1,\ldots,d\\}$ of a vector $x$ such that
$J_{\omega}(x)_{i}=\begin{cases}1&\text{if }x_{i}+\omega_{i}=\text{second
max}(x+\omega)\\\ 0&\text{otherwise.}\end{cases}$
Suppose we have a categorical response variable
$Y^{\prime}\in\\{g_{1},\ldots,g_{q}\\}$ with $q\leq d$ many levels. Then we
set a vector of indicator functions $Y=(Y_{1},\ldots,Y_{d})$ such that
$Y_{i}=\begin{cases}1&\mbox{if }Y^{\prime}=g_{i}\\\ 0&\mbox{else.}\end{cases}$
###### Example 4.3.
Suppose that we have a binary categorical response variable
$Y^{\prime}\in\\{\mbox{"yes" , "no"}\\}$. Then we can set as a vector of
indicator functions $Y=(Y_{1},\ldots,Y_{d})$ such that
$Y_{1}=\begin{cases}1&\mbox{if }Y^{\prime}=\mbox{yes}\\\
0&\mbox{else,}\end{cases}$ $Y_{2}=\begin{cases}1&\mbox{if
}Y^{\prime}=\mbox{no}\\\ 0&\mbox{else,}\end{cases}$
and
$Y_{3}=Y_{4}=\ldots=Y_{d}=0.$
In order to simplify the problem, here we consider a case that there are only
two classes $Y^{A}$ and $Y^{B}$ in the response variable. More precisely, we
have a random variable
$Y^{A}:=(Y^{A}_{1},\ldots,Y^{A}_{d}),Y^{B}:=(Y^{B}_{1},\ldots,Y^{B}_{d})\in\\{0,1\\}^{d}$
such that for fixed $i,k\in\\{1,\ldots,d\\}$ with $i\neq k$,
$Y^{A}_{j}=\begin{cases}1&\mbox{if }j=i\\\ 0&\mbox{else,}\end{cases}$
and
$Y^{B}_{l}=\begin{cases}1&\mbox{if }l=k\\\ 0&\mbox{else}\end{cases}$
with a discrete probability $\pi_{A}:=P(Y=Y^{A})$ and $\pi_{B}:=P(Y=Y^{B})$ .
Then, suppose we have a multivariate random variable
$X\in\mathbb{R}^{d}\\!/\mathbb{R}{\bf 1}$ given $Y$ with the probability
density function $f_{A}$ if $Y=Y^{A}$ and the probability density function
$f_{B}$ if $Y=Y^{B}$ such that there exists a tropical hyperplane
$H_{\omega^{*}}$ with a normal vector
$\omega^{*}\in\mathbb{R}^{d}\\!/\mathbb{R}{\bf 1}$ with the following
properties:
* (i)
there exists an index $i\in\\{1,\ldots,d\\}$ such that
$\text{for
any}\;j\in\\{1,\ldots,d\\}\backslash\\{i\\},\;\;\omega^{*}_{i}+X_{i}\;>\;\omega^{*}_{j}+X_{j},\;\;\;\text{and}$
* (ii)
$\max\left\\{I_{\omega^{*}}(X)-Y\right\\}=0,$
with probability $1$.
Let $\mathcal{D}$ be the distribution on the joint random variable $(X,Y)$ and
let $\mathcal{S}$ be the sample
$\mathcal{S}:=\\{(X^{1},Y^{1}),\ldots,(X^{n},Y^{n})\\}$. Then we formulate an
optimization problem for solving the normal vector $\omega$ of an optimal
tropical separating hyperplane $H_{\omega}$ for random variables $X$ given
$Y$: For some cost $C\in\mathbb{R}$
$\begin{matrix}\displaystyle\left[\max\limits_{\omega\in\mathbb{R}^{d}\\!/\mathbb{R}{\bf
1}}\;\min\limits_{X\in\mathcal{S},i\in S(I_{\omega}(X)),j\in
S(J_{\omega}(X))}\left\\{\underbrace{\left(X_{i}+\omega_{i}-X_{j}-\omega_{j}\right)}_{\text{margin}}+\underbrace{\frac{C}{n}\sum_{k=1}^{n}\min\left\\{Y^{k}-I_{\omega}(X^{k})\right\\}}_{\text{error}}\right\\}\right].\\\
\end{matrix}$ (6)
Here, the expectation of the random variable
$\max\left\\{I_{\omega}(X)-Y\right\\}$ is the $0-1$ loss function. Also note
that
$d_{\rm tr}(X,H_{\omega})\;=\;X_{i}+\omega_{i}-X_{j}-\omega_{j},$
where $i\in S(I_{\omega}(X)),j\in S(J_{\omega}(X))$. Thus, this optimization
problem can be explicitly written as a linear programming problem (7)–(10)
below, where the optimal solution $z$ means the margin of the tropical SVM:
For some cost $C\in\mathbb{R}$
$\displaystyle\max\limits_{(z,\omega)\in\mathbb{R}\times\mathbb{R}^{d}\\!/\mathbb{R}{\bf
1}}\;\left(z+\frac{C}{n}\sum_{k=1}^{n}\min\left\\{Y^{k}-I_{\omega}(X^{k})\right\\}\right)$
(7) $\displaystyle\textrm{s.t.}\;\;\forall X$
$\displaystyle\in\mathcal{S},\forall i\in S(I_{\omega}(X)),\forall j\in
S(J_{\omega}(X)),\;\;z+{\color[rgb]{0,0,0}X_{j}}+\omega_{j}{\color[rgb]{0,0,0}-X_{i}}-\omega_{i}\leq
0,$ (8) $\displaystyle\forall X$ $\displaystyle\in\mathcal{S},\forall i\in
S(I_{\omega}(X)),\forall j\in S(J_{\omega}(X)),\;\;\omega_{j}-\omega_{i}\leq
X_{i}-X_{j},$ (9) $\displaystyle\forall X$
$\displaystyle\in\mathcal{S},\forall l\not\in S(I_{\omega}(X))\cup
S(J_{\omega}(X)),j\in S(J_{\omega}(X))\;\;\omega_{l}-\omega_{j}\leq
X_{j}-X_{l}.$ (10)
###### Remark 4.4.
With the regularization with the tropical metric to the tropical hyperplane,
it minimizes the coefficients of the tropical hyperplane $\omega_{i}$ for
$i=1,\ldots d$ and tries to minimize the difference between the coordinate of
the maximum and second maximum of $X+\omega$. This means that this
regularization behaves similar to the $L_{1}$ or $L_{\infty}$ norm
regularization, that is, to select features which contribute to discriminate
$X|Y^{A}$ and $X|Y^{B}$ .
### 4.3 Bound on Generalization Errors via VC Dimensions
In this section we show the generalization bound for the error rate of the
hard margin tropical SVM. In order to compute them, we define the loss
function in terms of the distribution of $X|Y$ defined in the previous section
and the sample loss function. In order to prove the generalization bound, we
use the VC dimension (Vapnik, 1995; Mohri et al., 2018)
Let $\mathcal{D}$ be the distribution of $X|Y$ defined in Section 4. In
addition, let
$L_{\mathcal{D}}(\omega)=\mathbb{E}_{\mathcal{D}}\left(\max\left\\{I_{\omega}(X)-Y\right\\}\right)$
be the loss function with respect to $\mathcal{D}$ and let
$L_{\mathcal{S}}(\omega)=\mathbb{E}_{\mathcal{S}}\left(\max\left\\{I_{\omega}(X)-Y\right\\}\right)$
be the loss function with respect to the sample $\mathcal{S}$.
Let $(Z_{s},\omega^{s})$ be an output of the tropical SVM. Then we want to
find the upper bound for
$P\left(\max\left\\{I_{\omega^{s}}(X)-Y\right\\}>0\right)=L_{\mathcal{D}}(\omega^{s}).$
(11)
In order to prove our bound, we have to define VC dimension.
###### Definition 4.5.
Suppose $M$ is a set and a family of a subset $R\subset 2^{M}$. A set
$T\subset M$ is called shattered by $R$ if
$|\left\\{r\cap T|r\in R\right\\}|=2^{|T|}.$
The pair $(M,R)$ is called range space. The VC dimension of a range space,
$VC-dim(M,R)$ is the maximal cardinality of a subset of $M$ which can be
shattered by $R$.
###### Example 4.6 (Linear Classifiers).
In $(d-1)$-dimensional Euclidean space, a linear classifier can shatter or
successfully classify $d$ points in general positions with arbitrary labels.
Thus, the VC dimension of the linear classifier is $d$.
###### Lemma 4.7 (Tropical Radon Lemma in Jaggi et al. (2008)).
For $d+1$ points in $\mathbb{R}^{d}\\!/\mathbb{R}{\bf 1}$, there exists a
partition into two sets with intersecting convex hulls, which are therefore
not separable by any tropical hyperplane $H_{\omega}$.
###### Example 4.8.
Any tropical hyperplane in $\mathbb{R}^{3}\\!/\mathbb{R}{\bf 1}$ cannot
shatter the four points shown in Figure 5. The fact that the two line segments
intersect to each other suggests that it is impossible to classify them by
tropical hyperplanes. In fact, ANY four points cannot be shattered by a
tropical hyperplane in $\mathbb{R}^{3}\\!/\mathbb{R}{\bf 1}$. This is because,
given four or more points, there always exist ways of partitioning so that the
tropical convex hull of the points in each class intersects.
Figure 5: Example positions where four points cannot be shattered by tropical
hyperplanes in $\mathbb{R}^{3}\\!/\mathbb{R}{\bf 1}$. The gray lines represent
the tropical convex hulls of the two points in the same class. The fact that
the line segments intersect suggests that it is impossible to classify them by
tropical hyperplanes.
###### Lemma 4.9 (Lemma 27 in Jaggi et al. (2008)).
There exists a set of $d$ points in $\mathbb{R}^{d}\\!/\mathbb{R}{\bf 1}$ that
can be shattered by single sectors defined by a tropical hyperplane
$H_{\omega}$.
While Tropical Radon Lemma 4.7 shows that the VC dimension of a tropical
hyperplane is smaller than $d+1$, (trivial) Lemma 4.9 shows that it is larger
than or equal to $d$ (as single sectors should have less ability than
hyperplanes).
###### Theorem 4.10 (Jaggi et al. (2008)).
Let $\mathbf{H}$ be a family of tropical hyperplanes in
$\mathbb{R}^{d}\\!/\mathbb{R}{\bf 1}$. Then the VC dimension of a range space
$(\mathbb{R}^{d}\\!/\mathbb{R}{\bf 1},\mathbf{H})$ is $d$.
###### Theorem 4.11 (Vapnik (2000)).
Let $\mathcal{H}$ be a family of classification models whose VC dimension is
$h$. Let $n$ be the sample size of the sample $\mathcal{S}$. Then, for any
$\eta>0$ and for a sufficiently large sample size $n$ with the probability
greater than or equal to $1-\eta$, we have
$\ell_{\mathcal{D}^{\prime}}(\mathcal{H})-\ell_{\mathcal{S}}(\mathcal{H})\leq\sqrt{\frac{h(\log(2n/h)+1)-\log({\color[rgb]{0,0,0}\eta/4})}{n}},$
where $\ell_{\mathcal{D}^{\prime}}(\mathcal{H})$ is the $0-1$ loss for the
distribution $\mathcal{D}^{\prime}$ with $\mathcal{H}$ and
$\ell_{\mathcal{S}}(\mathcal{H})$ is the $0-1$ loss for the sample
$\mathcal{S}$ with $\mathcal{H}$.
The generalization error bound for tropical SVMs is obtained by plugging in
the VC dimension of tropical hyperplanes in Theorem 4.10 for $h$ in Theorem
4.11.
###### Theorem 4.12.
Let $n$ be the sample size of the sample $\mathcal{S}$. Then, for any $\eta>0$
and for a sufficiently large sample size $n$ with the probability greater than
or equal to $1-\eta$, we have
$P_{\mathcal{D}}\left(\max\left\\{I_{\omega^{s}}(X)-Y\right\\}>0\right)\leq\sqrt{\frac{d(\log(2n/d)+1)-\log({\color[rgb]{0,0,0}\eta/4})}{n}}.$
(12)
###### Proof.
Let $\mathcal{S}$ be a sample for a test set. Since there is a tropical hard
margin and $(z_{s},\omega^{s})$ is a feasible solution on the test set
$\mathcal{S}$, $L_{\mathcal{S}}(\omega^{s})=0$. With Theorem 4.10 and Theorem
4.11, we have
$L_{\mathcal{D}}(\omega^{s})-L_{\mathcal{S}}(\omega^{s})\leq\sqrt{\frac{d(\log(2n/d)+1)-\log({\color[rgb]{0,0,0}\eta/4})}{n}}.$
Since $L_{\mathcal{S}}(\omega^{s})=0$ by the assumption and
$L_{\mathcal{D}}(\omega^{s})=P_{\mathcal{D}}\left(\max\left\\{I_{\omega^{s}}(X)-Y\right\\}>0\right),$
we have the result. ∎
## 5 Application 2: Robustness of Tropical SVMs against Curse of
Dimensionality
In the previous section, we derive the generalization error bound for tropical
SVMs. The bound is stated in a distribution-free manner as a general feature
applicable to all data. Specifically, the bound is derived purely
combinatorially only based on the shapes of the hyperplanes. Algorithmic
details of tropical SVMs are not yet taken into account there. Therefore it is
possible that tropical SVMs outperform in specific situations, for example,
when only a few features are prominently informative among many. Thus the
evaluation of the performance of tropical SVMs by computational experiments
can give complementary information. Here we applied tropical SVMs to specific
problems of computational experiments and real data analyses to evaluate its
performance originating from the max-plus algebra.
### 5.1 Implementation of Tropical SVMs
Soft margin tropical SVMs are defined in Tang et al. (2020) which allow some
data points in the wrong open sectors with the hinge loss function as similar
to the $L_{2}$ norm SVMs. Since the loss function deals with $L_{0}$ norm of
$I_{\omega}(X)-Y$, we have to run exponentially many linear programming
problems in order to find the optimal tropical hyperplane to fit the data set.
Therefore, Tang et al. introduced several heuristic algorithms to estimate the
optimal tropical hyperplane to fit the data: Algorithms 1 – 4. In the
computational experiments in this paper, we applied Algorithms 3, the simplest
one, from Tang et al. (2020) to fit each data to the model, as it always
performed the best for our simple examples. We obtained the R code from
https://github.com/HoujieWang/Tropical-SVM.
### 5.2 Computational Experiments
As we observe in Section 3.2, the tropical distance can ignore the nuisance
dimensions that are purely noises. Therefore it is expected that tropical SVMs
can show a good performance in the curse-of-dimensionality regime where extra
dimensions are all uninformative.
Here we performed a simple computational experiment of the two-class
classification using the uncorrelated unit-variance Gaussian distributions
whose means are ${\color[rgb]{0,0,0}m_{1}=}$ $(s,-s,0,0,\ldots,0)$ for class
$1$ and ${\color[rgb]{0,0,0}m_{2}=}$ $(-s,s,0,0,\ldots,0)$ for class $2$,
where $s=\sqrt{2}$ or $5$. To be specific, the unit covariance matrices are
common for the two distributions as
$\Sigma_{1}=\Sigma_{2}=\begin{pmatrix}1&0&\ldots&0\\\ 0&1&\ldots&0\\\
\vdots&\vdots&\vdots&\vdots\\\ 0&0&\ldots&1\\\ \end{pmatrix}$. Note that the
signal or the difference of the means, $m_{2}-m_{1}=(2s,-2s,0,0,\ldots,0)$, is
orthogonal to the irrelevant direction $(1,1,1,1,\ldots,1)$. As a measure of
separation, the d-prime or the Euclidean distance between $m_{1}$ and $m_{2}$
divided by the S.D. (=1) is given by $d^{\prime}=2s\sqrt{2}$.
Before we look at the curse-of-dimensionality regime or small samples, we
first consider the case when the numbers of training and test samples $N$ are
fairly large to compare the numerically obtained generalization errors with
the theoretical bounds in Figure 6. We draw the lower bound of the
classification success rates by subtracting the penalty term in the right hand
side of the inequality in Theorem 4.11 from the hit rate for the training set.
The gaps between the numerical computations and the theoretical bounds suggest
that the bounds are not so tight. In fact, the lower bounds are mostly smaller
than the chance level $50\%$ and thus not informative. Although we solely use
$\eta=0.1$, the bounds are rather insensitive to the value of $\eta$. Even if
we have as large as $N=1000$ or $10000$ observations, the gap is still quite
noticeable as expected from the penalty term in the right hand side of the
inequality in Theorem 4.11.
Figure 6: Numerically obtained classification success rates (thick lines) and
theoretical lower bounds (dotted lines) for the classification of two
multivariate Gaussian distributions whose means are separated by
$d^{\prime}=4$ and covariances are the unit matrices. The results for tropical
SVMs are red colored while those for classical SVMs are black colored. The
numbers of training and test samples for each class are $100$ (left column)
and $200$ (right column). The cross-validated performance averaged over $1000$
realizations of data were plotted. Note that the lower bounds of the hit rates
(=upper bounds of error rates) are not tight.
Next, we consider small samples as a curse-of-dimensionality regime. Figure 7
shows the classification success rates of two multivariate Gaussian
distributions separated by ${\color[rgb]{0,0,0}d^{\prime}=}4$ (top) or
${\color[rgb]{0,0,0}d^{\prime}=}10\sqrt{2}$ (bottom). The numbers of training
and test samples for each class are 5 (left) and 10 (right). In all four
cases, the hit rate decreases significantly for classical SVMs, while it
decreases only slightly for tropical SVMs. In fact, it almost stays constant
for the ${\color[rgb]{0,0,0}d^{\prime}=}10\sqrt{2}$ cases. Thus the tropical
SVM is robust against the curse of dimensionality in the case for separating
two Gaussian distributions.
Figure 7: Classification success rates of two multivariate Gaussian
distributions whose means are separated by $d^{\prime}=4$ (top row) or
$10\sqrt{2}$ (bottom row) and covariances are the unit matrices. The numbers
of training and test samples for each class are $5$ (left column) and $10$
(right column). The cross-validated performance averaged over $1000$
realizations of data were plotted. The errorbars denote the standard
deviations for the $1000$ realizations. Note that the tropical SVM is robust
against the curse of dimensionality.
As our setting of classifying two Gaussians was so simple, it is expected that
the results can be quite general. The only assumption here is that the
difference of the two means is not parallel to ${\bf 1}=(1,1,\ldots,1)$. We
believe that in real data the difference of the means rarely becomes by chance
parallel to ${\bf 1}$. We will examine the robustness against the curse of
dimensionality for the real data in what follows. But before doing that, let
us explain theoretically why tropical SVMs are robust against the curse of
dimensionality.
### 5.3 Understanding Robustness by Numerical Examples
First, as a basic example, let us consider the simplest case with two points
in three dimensions: $x_{1}=(5,-5,0)$ for class $1$ and $x_{2}=(-5,5,0)$ for
class $2$. As our problem is to find
$\omega=(\omega_{1},\omega_{2},\omega_{3})$ in
$\mathbb{R}^{3}\\!/\mathbb{R}{\bf 1}$, we can limit to
$\omega=(\omega_{1},\omega_{2},0)$ without loss of generality. By maximizing
the following margin over $\omega$ in $\mathbb{R}^{3}\\!/\mathbb{R}{\bf 1}$
under the condition with no classification error,
$M(\omega):=\min\left(d_{\rm tr}(x_{1},H_{\omega}),d_{\rm
tr}(x_{2},H_{\omega})\right),$ (13)
we obtain $\omega^{*}=(5+c,5+c,0)$ for $c>0$ as in Figure 8.
Figure 8: (left) Margin $M(\omega)$ for $\omega=(\omega_{1},\omega_{2},0)$.
For the regions A, it is $10-|\omega_{1}-\omega_{2}|$. For the region $\times$
containing $(0,0)$, it is
$5$+min$(\omega_{1},\omega_{2})=\frac{\omega_{1}+\omega_{2}+10}{2}-\frac{|\omega_{1}-\omega_{2}|}{2}$.
For the regions B, it is
$\frac{|\omega_{1}-\omega_{2}|}{2}-\frac{|\omega_{1}+\omega_{2}+10|}{2}$. For
the regions C, it is $5-|\max(\omega_{1},\omega_{2})|$. The other non-marked
regions have non-zero classification errors and are ineligible. Note that on
the verge of misclassification, it is always $0$. The maximum is attained at
$\omega^{*}=(5+c,5+c,0)$ for $c>0$. (right) Configuration in optimal solutions
where margin is $10$. The red crosses denote the data points to be classified
when $\omega^{*}=(5,5,0)$. The orange crosses denote the data points to be
classified when $\omega^{*}=(25,25,0)$. Note that both configurations attain
the maximum margin.
In fact, $x_{1}+\omega^{*}=(10+c,c,0)$ and $x_{2}+\omega^{*}=(c,10+c,0)$ lead
to $d_{\rm tr}(x_{1},H_{\omega^{*}})=d_{\rm tr}(x_{2},H_{\omega^{*}})=10$. In
Figure 8 (left) for the margin $M(\omega)$, error-free regions for $\omega$
are only A, B and C. The other regions have classification errors. The entire
shape of the error-free region reflect the Y-shape of the hyperplane $H_{0}$.
For example, for $\omega=(-10,-10,0)$, $x_{1}+\omega$ and $x_{2}+\omega$
belong to the same bottom left sector of $H_{0}$, meaning misclassification.
Note that as the margin function $M(\omega)$ is convex unless $x_{1}+\omega$
or $x_{2}+\omega$ crosses the sector borders, the linear programming is needed
only once for each possible combination of labels and sectors (Tang et al.
(2020)).
Here we can impose the regularization term, $d_{\rm tr}(\omega,0)$, so that we
choose the solution with the minimum norm even if the evaluation function
outputs ties. In this tradition, the unique solution is $\omega^{*}=(5,5,0)$.
Second, let us consider the general case with $N$ points for each class in $d$
dimensions, whose coordinates are perturbed from $x_{1}$ or $x_{2}$ by
standard Gaussian noises $\xi$ or $\eta$ (Figure 9) as in the previous
computational experiments. We again need to maximize the margin:
$M(\omega):=\min_{1\leq i\leq N,~{}1\leq l\leq 2}d_{\rm
tr}(x_{Y=l}^{i},H_{\omega}),$ (14)
with
$\displaystyle
x_{Y=1}^{i}=(5+\xi_{1}^{i},-5+\xi_{2}^{i},\xi_{3}^{i},\xi_{4}^{i},\ldots,\xi_{d}^{i}),$
$\displaystyle
x_{Y=2}^{j}=(-5+\eta_{1}^{j},5+\eta_{2}^{j},\eta_{3}^{j},\eta_{4}^{j},\ldots,\eta_{d}^{j}),$
(15)
where $\xi_{p}^{i}\sim N(0,1)$ and $\eta_{p}^{i}\sim N(0,1)$ for $1\leq p\leq
d$ and $1\leq i\leq 2$. Let us try to find the solution near
$\omega^{*}=(5+c,5+c,0,0,\ldots,0)$ and $d\rightarrow\infty$.
In the previous computational experiments, we observed that for all cases of
$1000$ data realizations the best classifiers used the first and the second
sectors as in Figure 9. Furthermore, among $x_{p}+\omega_{p}$ for $1\leq p\leq
d$, the maximum and the second maximum were attained by $x_{1}+\omega_{1}$ and
$x_{2}+\omega_{2}$ when the training data were classified. Therefore we focus
on this situation, i.e. classification by using only sector $1$ and sector
$2$, in what follows.
Figure 9: A solution of tropical SVM that uses first and second sectors.
$\omega$ is adjusted so that the tropical hyperplane is located right in the
middle of the two support vectors.
Under the assumption of $\omega_{2}\gg 5$, which will be justified later, the
tropical distance to $H_{\omega}$ is
$\displaystyle d_{\rm tr}(x_{Y=1}^{i},H_{\omega})=d_{\rm
tr}(x_{Y=1}^{i}+\omega,H_{0})$
$\displaystyle={\rm\\{max-2nd~{}max~{}of\\}}~{}(5+\xi_{1}^{i}+\omega_{1},-5+\xi_{2}^{i}+\omega_{2},\xi_{3}^{i}+\omega_{3},$
$\displaystyle\ldots,\xi_{d}^{i}+\omega_{d})$
$\displaystyle=10+\xi_{1}^{i}+\omega_{1}-\xi_{2}^{i}-\omega_{2}$ (16)
The minimum over $N$ points with class $1$ is
$\displaystyle\min_{1\leq i\leq N}d_{\rm tr}(x_{Y=1}^{i},H_{\omega})$
$\displaystyle=$ $\displaystyle 10+\omega_{1}-\omega_{2}+\min_{1\leq i\leq
N}\left(\xi_{1}^{i}-\xi_{2}^{i}\right)$ (17) $\displaystyle=$ $\displaystyle
10+\omega_{1}-\omega_{2}+\xi_{1}^{i_{*}}-\xi_{2}^{i_{*}}.$
where $i_{*}(\omega)$ denotes the minimizer. Similarly for the class $2$,
$\displaystyle\min_{1\leq j\leq N}d_{\rm tr}(x_{Y=2}^{j},H_{\omega})$
$\displaystyle=$ $\displaystyle 10+\omega_{2}-\omega_{1}+\min_{1\leq j\leq
N}\left(\eta_{2}^{j}-\eta_{1}^{j}\right)$ (18) $\displaystyle=$ $\displaystyle
10+\omega_{2}-\omega_{1}+\eta_{2}^{j_{*}}-\eta_{1}^{j_{*}},$
where $j_{*}(\omega)$ denotes the minimizer. By equating the minimum distances
for the two classes,
$10+\omega_{1}-\omega_{2}+\xi_{1}^{i_{*}}-\xi_{2}^{i_{*}}=10+\omega_{2}-\omega_{1}+\eta_{2}^{j_{*}}-\eta_{1}^{j_{*}},$
(19)
we get
$\omega_{2}=\omega_{1}+\frac{\xi_{1}^{i_{*}}-\xi_{2}^{i_{*}}+\eta_{1}^{j_{*}}-\eta_{2}^{j_{*}}}{2}.$
(20)
This theoretical prediction perfectly matched with the computational
experiments. This can be rewritten as,
$\displaystyle\omega_{1}$ $\displaystyle=$ $\displaystyle 5+c$
$\displaystyle\omega_{2}$ $\displaystyle=$ $\displaystyle
5+\frac{\xi_{1}^{i_{*}}-\xi_{2}^{i_{*}}+\eta_{1}^{j_{*}}-\eta_{2}^{j_{*}}}{2}+c,$
(21)
for $c>0$, which give the (evenly scored) maximizers of the margin under
$\omega_{2}\gg 5$. Note that $\omega$ is perturbed from $(5+c,5+c)$ only
slightly of order $O(1)$. This equation illustrates that the hyperplane just
listens to the positions of the support vectors. Adding $c$ might look strange
at first glance, but it is interpreted as the hyperplane’s freedom to shift
along the irrelevant coordinates of the support vectors. In fact, the
hyperplane for classical SVMs actually has the same freedom but you just
cannot distinguish the shifted planes.
Now we can explain why tropical SVMs are robust against the curse of
dimensionality. In the computational experiments, the other $\omega_{p}(p\geq
3)$ were smaller by around five or more. Therefore the misclassification never
happened as the maximum of $(x+\omega)$ always occurred at the first or second
coordinate appropriately.
Figure 10: Classification success rates of dopaminergic and nondopaminergic
neurons by using various numbers of features for tropical and classical SVMs.
All features are the firing rates, i.e. the number of spikes per second, in
$300$ms bins. The first $7$ features we preferentially used are the responses
to $7$ different reward stimuli, which are informative. The following $7$
features we appended are spontaneous activities after reward stimuli, which
are less informative. The other features are spontaneous activities before
reward stimuli, which are least informative. Note that the classification
success rates decreased slower for the tropical SVM than the classical SVM,
suggesting that the tropical SVM is relatively robust against the curse of
dimensionality.
### 5.4 Real Data Analysis
We now turn to empirical data from brain science. To summarize the data
acquisition in this paragraph, we recorded the spiking activity of $493$ total
neurons in the ventral tegmental area (VTA) and the substantia nigra (SN) of
DAT (dopamine transporter)-cre mice using tetrodes while these mice performed
various classical conditioning tasks (Matsumoto et al. (2016)). We
optogenetically identified $179$ dopaminergic neurons (Cohen et al. (2012);
Matsumoto et al. (2016)), which enabled us to assess the classification
success rates based on the neuronal tuning curves. In each trial of the task,
we delivered to a mouse one of seven stimuli, which were water rewards, air
puff, etc. The conventional tuning curve for these neurons consists of the
neural responses to the seven stimuli measured as the spike count within a
$300$ms time bin.
The tropical SVM for classifying dopaminergic and nondopaminergic neurons
based on their activities as features was also robust against the curse of
dimensionality (Figure 10). In order to use the same set of features, we
limited to $41$ neurons including $17$ dopaminergic and $24$ nondopaminergic
neurons. That is, the sample size is $41$. Both the training and test data
consist of, randomly selected, $5$ dopaminergic and $5$ nondopaminergic
neurons. The classification performance for the test data was used to select
the model or which two sectors to use in Algorithm 3 from Tang et al. (2020).
The cross validated classification success rates were computed for the
remaining $21$ neurons. The cross validated classification success rates
averaged over the $1000$ ways of selections of training and test data were
plotted in Figure 10.
Here we did not necessarily use the entire features but rather changed the
numbers of features used for the classification. As the order of adding
features can matter, we ordered and added the putatively informative features
first. After the classification success rates showed the peaks at the
beginning, they decreased slower for the tropical SVM than the classical SVM
(Figure 10).
In general, we believe that although the classification success rate is
enhanced by adding some most informative features at the beginning, soon it
will be deteriorated by less informative features, which can be a majority in
real world big data. The tropical SVM may be superior in the latter regime of
curse of dimensionality.
## 6 Extension of Tropical SVMs to Function Spaces
In this section, we define tropical SVMs over a function space to enable the
classification of curves, which we encounter frequently in practice. For
example, we consider neuronal tuning curves in Figure 3 as a feature vector.
That is, we can treat a tuning curve as a point in a function space. For
example, it is valuable for identifying dopaminergic neurons in clinical
researches (Ishikawa et al., 2018, 2019) to measure the distances between
neuronal tuning curves and classify neurons using a tropical hyperplane
boundary, which itself can be another curve in tropical SVMs.
### 6.1 Tropical Distances and Hyperplanes on Function Spaces
Suppose we have a set of functions mapping from $\mathbb{R}^{d}$ to
$\mathbb{R}$ denoted as $\mathcal{F}$ such that:
$\begin{array}[]{c}\mathcal{F}:=\\{f:\mathbb{R}^{s}\to\mathbb{R}:|f(x)|\mbox{
is bounded,}\,f(x)=f(x)+c,\\\ \forall x\in\mathbb{R}^{s},\,\forall
c\in\mathbb{R}\\}.\end{array}$
Let $d_{\rm tr}$ be a distance between functions in $\mathcal{F}$ such that
$d_{\rm tr}(f,g):=\max\\{f(x)-g(x):\forall
x\in\mathbb{R}^{s}\\}-\min\\{f(x)-g(x):\forall x\in\mathbb{R}^{s}\\}.$
###### Example 6.1 (Functional Tropical Distances, Figure 11).
Let $\mathcal{F}$ be a set of univariate Gaussian distribution functions with
$\mu$ and $\sigma$, and mixtures of these Gaussian distributions with real
valued coefficients. Suppose $F_{1}$, $F_{2}$, $F_{3}$ and $F_{4}$ are
univariate Gaussian distribution functions with $\mu_{1}=\mu_{2}=-2$,
$\mu_{3}=\mu_{4}=2$, $\sigma_{1}=\sigma_{3}=1$, and
$\sigma_{2}=\sigma_{4}=1/2$. Then, $d_{\rm tr}(F_{1},F_{2})=d_{\rm
tr}(F_{3},F_{4})=\frac{11}{8\sqrt{2\pi}}=0.549$, $d_{\rm
tr}(F_{1},F_{3})=0.798$, and $d_{\rm tr}(F_{1},F_{4})=d_{\rm
tr}(F_{2},F_{3})=1.197$.
Figure 11: (left) Four example functions, $F_{1}$, $F_{2}$, $F_{3}$ and
$F_{4}\in\mathcal{F}$, which are Gaussian distribution functions with $\mu=-2$
or $2$ and $\sigma=1$ or $0.5$. (middle) $f(x)-F_{1}(x)$ for the four
functions in the left figure as $f(x)$. $F_{1}(x)$ denotes the reference black
function with $\mu=-2$ and $\sigma=1$. (right) $d_{\rm tr}(f(x),F_{1}(x))$ is
gray-color coded for various $\mu$ and $\sigma$ of $f(x)$. Note that the
distance is $0$ for $(\mu,\sigma)=(-2,1)$ and increases with increasing $\mu$
or decreasing $\sigma$: $d_{\rm tr}(F_{2},F_{1})=0.549$, $d_{\rm
tr}(F_{3},F_{1})=0.798$, and $d_{\rm tr}(F_{4},F_{1})=1.197$.
###### Lemma 6.2.
$d_{\rm tr}(f(x),g(x))=d_{\rm tr}(-f(x),-g(x))=d_{\rm tr}(g(x),f(x))$
###### Proof.
$\begin{array}[]{cl}&d_{\rm tr}(f(x),g(x))\\\ =&-\min\\{f(x)-g(x):\forall
x\in\mathbb{R}^{s}\\}+\max\\{f(x)-g(x):\forall x\in\mathbb{R}^{s}\\}\\\
=&\max\\{-f(x)+g(x):\forall x\in\mathbb{R}^{s}\\}-\min\\{-f(x)+g(x):\forall
x\in\mathbb{R}^{s}\\}\\\ =&d_{\rm tr}(-f(x),-g(x))\\\ =&d_{\rm
tr}(g(x)-f(x),0)\\\ =&d_{\rm tr}(g(x),f(x)).\end{array}$
∎
###### Lemma 6.3.
$d_{\rm tr}(f(x),0)$ is a norm on $\mathcal{F}$.
###### Proof.
For the identity,
$d_{\rm tr}(f,0)=0$ if and only if $\max\\{f(x):\forall
x\in\mathbb{R}^{s}\\}=\min\\{f(x):\forall x\in\mathbb{R}^{s}\\},$ meaning that
$f(x)$ is constant. Next, by Lemma 6.2 with $g(x)=0$
$d_{\rm tr}(af,0)=d_{\rm tr}(|a|f,0)=|a|d_{\rm tr}(f,0)$
for $a\in\mathbb{R}$. For triangle inequality,
$\begin{array}[]{cl}&d_{\rm tr}(f(x)+g(x),0)\\\ =&\max\\{f(x)+g(x):\forall
x\in\mathbb{R}^{s}\\}-\min\\{f(x)+g(x):\forall x\in\mathbb{R}^{s}\\}\\\
=&\max\\{f(x)+g(x):\forall x\in\mathbb{R}^{s}\\}+\max\\{-f(x)-g(x):\forall
x\in\mathbb{R}^{s}\\}\\\ \leq&\max\\{f(x):\forall
x\in\mathbb{R}^{s}\\}+\max\\{g(x):\forall x\in\mathbb{R}^{s}\\}\\\
&+\max\\{-f(x):\forall x\in\mathbb{R}^{s}\\}+\max\\{-g(x):\forall
x\in\mathbb{R}^{s}\\}\\\ =&\max\\{f(x):\forall
x\in\mathbb{R}^{s}\\}+\max\\{g(x):\forall x\in\mathbb{R}^{s}\\}\\\
&-\min\\{f(x):\forall x\in\mathbb{R}^{s}\\}-\min\\{g(x):\forall
x\in\mathbb{R}^{s}\\}\\\ =&d_{\rm tr}(f(x),0)+d_{\rm tr}(g(x),0).\\\
\end{array}$
∎
###### Lemma 6.4.
$d_{\rm tr}$ is a metric on $\mathcal{F}$.
###### Proof.
$d_{\rm tr}$ is a metric because it is induced by the norm of the difference,
$d_{\rm tr}(f(x),g(x))=d_{\rm tr}(f(x)-g(x),0).$
∎
By the proof of Lemma 6.4, we have the following theorem:
###### Theorem 6.5.
$(\mathcal{F},d_{\rm tr})$ is a normed vector space.
###### Definition 6.6 (Tropical Hyperplane in $\mathcal{F}$).
Let $B_{\epsilon}(x)$ be an open ball around $x\in\mathbb{R}^{s}$ with its
radius $\epsilon>0$. For any $\omega\in\mathcal{F}$ and for $\epsilon>0$, the
tropical hyperplane defined by $\omega$ and $\epsilon$, denoted by
$H_{\omega,\epsilon}$, is the set of points $f\in\mathcal{F}$ such that
$|\\{\omega(y)+f(y)=\omega(x^{*})+f(x^{*}):y\in(\mathbb{R}^{s}-B_{\epsilon}(x^{*}))\\}|\geq
1,$
where
$x^{*}\in\operatorname*{arg\,max}_{x\in\mathbb{R}^{s}}\left(f(x)+\omega(x)\right).$
We call $\omega\in\mathcal{F}$ the normal vector of $H_{\omega,\epsilon}$.
###### Definition 6.7 (Tropical Distance to a Tropical Hyperplane in
$\mathcal{F}$).
The tropical distance from a point $f\in\mathcal{F}$ to a tropical hyperplane
$H_{\omega,\epsilon}$ is defined as
$d_{\rm tr}(f,H_{\omega,\epsilon})\;:=\;\min\\{d_{\rm tr}(f,g)\;|\;g\in
H_{\omega,\epsilon}\\}.$
###### Lemma 6.8.
$d_{\rm tr}(f,H_{\omega,\epsilon})\;=\;d_{\rm tr}(f+\omega,H_{0,\epsilon}).$
###### Proof.
$H_{\omega,\epsilon}$ is defined as:
$\begin{array}[]{c}H_{\omega,\epsilon}=\\{g\in\mathcal{F}:|\\{\omega(y)+g(y)=\omega(x^{*})+g(x^{*}):\\\
y\in(\mathbb{R}^{s}-B_{\epsilon}(x^{*}))\\}|\geq 1\\},\end{array}$
where
$x^{*}\in\operatorname*{arg\,max}_{x\in\mathbb{R}^{s}}(\omega(x)+g(x)).$
Thus
$\begin{array}[]{rl}&d_{\rm tr}(f,H_{\omega})\\\ =&\Bigl{\\{}d_{\rm
tr}(f,g):|\\{\omega(y)+g(y)=\omega(x^{*})+g(x^{*}):\\\
&y\in(\mathbb{R}^{s}-B_{\epsilon}(x^{*}))\\}|\geq 1\Bigr{\\}}\\\
=&\Bigl{\\{}\max\\{f(x)-g(x)\\}-\min\\{f(x)-g(x)\\}:\\\
&|\\{\omega(y)+g(y)=\omega(x^{*})+g(x^{*}):y\in(\mathbb{R}^{s}-B_{\epsilon}(x^{*}))\\}|\geq
1\Bigr{\\}}\\\
=&\Bigl{\\{}\max\\{f(x)-(h(x)-\omega(x))\\}-\min\\{f(x)-(h(x)-\omega(x))\\}:\\\
&|\\{h(y)=h(x^{*}):y\in(\mathbb{R}^{s}-B_{\epsilon}(x^{*}))\\}|\geq
1\Bigr{\\}}\\\
=&\Bigl{\\{}\max\\{(f(x)+\omega(x))-h(x)\\}-\min\\{(f(x)+\omega(x))-h(x)\\}:\\\
&|\\{h(y)=h(x^{*}):y\in(\mathbb{R}^{s}-B_{\epsilon}(x^{*}))\\}|\geq
1\Bigr{\\}}\\\ =&\left\\{d_{\rm
tr}(f+w,g):|\\{g(y)=g(x^{*}):y\in(\mathbb{R}^{s}-B_{\epsilon}(x^{*}))\\}|\geq
1\right\\}\\\ =&d_{\rm tr}(f+\omega,H_{0,\epsilon}).\end{array}$
∎
###### Lemma 6.9.
Let $x^{*}\in\operatorname*{arg\,max}_{x\in\mathbb{R}^{d}}f(x)$. Then $d_{\rm
tr}(f,H_{0,\epsilon})=\max\\{f(x):x\in\mathbb{R}^{d}\\}-\max\\{f(y):y\in(\mathbb{R}^{s}-B_{\epsilon}(x^{*}))\\}$.
###### Proof.
Let
$\alpha=\max\\{f(x):x\in\mathbb{R}^{d}\\}-\max\\{f(y):y\in(\mathbb{R}^{s}-B_{\epsilon}(x^{*}))\\}$.
Then, $g(x)=f(x)+\alpha\mathcal{I}(y)$, where
$y\in(\mathbb{R}^{s}-B_{\epsilon}(x^{*}))\\}$ and $\mathcal{I}(y)$ is the
indicator function. Notice that $g(x)$ is in $H_{0,\epsilon}$. ∎
###### Example 6.10 (Tropical Hyperplanes in $\mathcal{F}$).
We will use the same set up as in Example 6.1 and Figure 11. Let $\epsilon=1$.
Then by Lemma 6.9, we have
$\begin{array}[]{rl}&d_{\rm tr}(F_{1},H_{0,\epsilon})\\\
=&\max\\{F_{1}(x):x\in\mathbb{R}\\}-\max\\{F_{1}(y):y\in(\mathbb{R}-B_{\epsilon}(x^{*})),\\\
&x^{*}=\operatorname*{arg\,max}(F_{1}(x):x\in\mathbb{R})\\}\\\
=&F_{1}(-2)-F_{1}(-1)\\\ =&0.157.\end{array}$
Similarly,
$\begin{array}[]{rl}d_{\rm tr}(F_{3},H_{0,\epsilon})=&d_{\rm
tr}(F_{1},H_{0,\epsilon})=F_{1}(-2)-F_{1}(-1)=0.157,\\\ d_{\rm
tr}(F_{4},H_{0,\epsilon})=&d_{\rm
tr}(F_{2},H_{0,\epsilon})=F_{2}(-2)-F_{2}(-1)=0.690.\end{array}$
Suppose we have $\omega=F_{3}$. Then, by Lemma 6.8, we have
$\begin{array}[]{rl}&d_{\rm tr}(F_{1},H_{\omega,\epsilon})\\\ =&d_{\rm
tr}(F_{1}+\omega,H_{0,\epsilon})\\\ =&d_{\rm
tr}(F_{1}+F_{3},H_{0,\epsilon})\\\
=&(F_{1}(-2)+F_{3}(-2))-(F_{1}(2)+F_{3}(2))\\\ =&0.399-0.399\\\
=&0.\end{array}$
Thus, $F_{1}\in H_{F_{3},\epsilon}$. Similarly, $F_{3}\in H_{F_{1},\epsilon}$,
$F_{2}\in H_{F_{4},\epsilon}$ and $F_{4}\in H_{F_{2},\epsilon}$.
###### Definition 6.11 (Sectors of Tropical Hyperplane in $\mathcal{F}$).
Each tropical hyperplane $H_{\omega,\epsilon}$ divides $\mathcal{F}$ into
components, which are open sectors,
$\begin{array}[]{l}S_{\omega,\epsilon}^{x}:=\\{\;f\in\mathcal{F}|\\\
\omega(x_{0})+f(x_{0})>\omega(y)+f(y),\forall x_{0}\in B_{\epsilon}(x),\forall
y\not\in B_{\epsilon}(x)\\},x\in\mathbb{R}^{s}\\}.\end{array}$
###### Example 6.12.
We will use the same set up as in Example 6.1. Again, let $\mathcal{F}$ be a
set of univariate Gaussian distribution functions with $\mu$ and $\sigma$, and
mixtures of these Gaussian distributions with real valued coefficients.
Suppose $\omega\equiv 0$ which is a Gaussian distribution function with
$\mu=0$ and $\sigma\to\infty$, $\epsilon=1$, and $x=0$. Then
$\begin{array}[]{l}S_{\omega,1}^{0}:=\\{\mbox{Mixtures of univariate
Gaussian}\\\ \mbox{distributions with real coefficients whose argmax in
}[-1,1]]\\}.\end{array}$
Also note that a set of
$\begin{array}[]{l}\\{\mbox{Gaussian distribution functions with
}\mu\in[-1,1]\\\ \mbox{ and }\sigma>L\mbox{ for some }L>0\\}\end{array}$
is in this open sector.
### 6.2 Tropical SVMs on Function Spaces
For a given $\epsilon>0$, suppose $F\in\mathcal{F}$ and
$Y^{1},\,Y^{2}\in\\{0,1\\}$ are random variables such that there exist
$\omega^{*}\in\mathcal{F}$ with
$\Bigl{(}\bigcup_{x^{*}\in\operatorname*{arg\,max}_{x\in\mathbb{R}^{s}}(\omega^{*}(x)+F(x)|Y^{1})}B_{\epsilon}(x^{*})\Bigr{)}\bigcap\Bigl{(}\bigcup_{x^{*}\in\operatorname*{arg\,max}_{x\in\mathbb{R}^{s}}(\omega^{*}(x)+F(x)|Y^{2})}B_{\epsilon}(x^{*})\Bigr{)}=\emptyset.$
(22)
Now we set up a tropical SVM over $\mathcal{F}$, whose solution $\omega$
satisfies Equation 22. Let $\mathcal{D}_{\mathcal{F}}$ be the distribution on
the joint random variable $(F,Y)$ for $F\in\mathcal{F}$ and $Y\in\\{0,1\\}$,
and let $\mathcal{S}_{\mathcal{F}}$ be the sample
$\mathcal{S}_{\mathcal{F}}:=\\{(F^{1},Y^{1}),\ldots,(F^{n},Y^{n})\\}$. For a
given $\epsilon>0$, we formulate an optimization problem for solving the
normal vector $\omega\in\mathcal{F}$ of an optimal tropical separating
hyperplane $H_{\omega,\epsilon}$ for random variables $X\in\mathcal{F}$ given
$Y\in\\{0,1\\}$: For some cost $C\in\mathbb{R}$,
$\begin{matrix}\displaystyle\max\limits_{\omega\in\mathcal{F}}\min\limits_{F(x)\in\mathcal{S}_{\mathcal{F}}}\left(\underbrace{\left(\max_{x\in\mathbb{R}^{s}}(F(x)+\omega(x))-\max_{z\in\mathcal{Z}}(F(z)+\omega(z))\right)}_{\text{margin}}+\underbrace{\frac{C}{n}\sum_{k=1}^{n}\max\left\\{\mathcal{I}_{B_{\epsilon}(X^{*})}(x)-Y^{k}\right\\}}_{\text{error}}\right),\end{matrix}$
(23)
where $X^{*}\in\operatorname*{arg\,max}_{x\in\mathbb{R}^{s}}(\omega(x)+F(x))$
and
$\mathcal{Z}:=\operatorname*{arg\,max}_{x\in(\mathbb{R}^{s}-B_{\epsilon}(X^{*}))}(\omega(x)+F(x))$.
In practice, we approximate each function $F^{i}$, for $i=1,\ldots n$, by its
empirical function $\hat{F}^{i}$ by taking some point $x\in\mathbb{R}^{s}$ to
evaluate $F^{i}(x)$. In this paper we propose Algorithm 1 to heuristically
conduct a tropical SVM using empirical functions $\hat{F}^{i}$, for
$i=1,\ldots n$, using finite set of points
$\\{x^{1},\ldots,x^{k}\\}\subset\mathbb{R}^{s}$ with $x^{j}\not\in
B_{\epsilon}(x^{i})$ for $j\not=i$ and for $i=1,\ldots,k$. Let
$\hat{F}^{i}=(F^{i}(x^{1}),\ldots,F^{i}(x^{k}))$, for $i=1,\ldots n$.
Input: A train set
$\mathcal{S}=\\{(\hat{F}^{1},Y^{1}),\ldots,(\hat{F}^{n},Y^{n})\\}$
Output: Estimated normal vector $\hat{\omega}$ of a tropical hyperplane
Apply $\\{(\hat{F}^{1},Y^{1}),\ldots,(\hat{F}^{n},Y^{n})\\}$ to a tropical SVM
over $\mathbb{R}^{k}/\mathbb{R}{\bf 1}$
return _The output from the tropical SVM over $\mathbb{R}^{k}/\mathbb{R}{\bf
1}$_
Algorithm 1 Heuristic tropical SVM over $\mathcal{F}$
## 7 Discussion
We show the generalization error bounds for tropical SVMs over the tropical
projection space $\mathbb{R}^{d}/\mathbb{R}{\bf 1}$ which is isometric to
$\mathbb{R}^{d-1}$. These bounds still depend on the dimension $d$ and if we
fix the sample size $n$, these bounds do not make sense and we cannot extend
these bounds for tropical SVMs over a function space $\mathcal{F}$ with
tropical metric $d_{\rm tr}$. For future work it is interesting to obtain
generalization error bounds for tropical SVMs over a function space
$\mathcal{F}$ with tropical metric $d_{\rm tr}$. In addition, computational
experiments show that tropical SVMs over $\mathbb{R}^{d}/\mathbb{R}{\bf 1}$
have much lower error rates than ones of $L_{2}$ norm SVMs over
$\mathbb{R}^{d}$ when we fix the sample size and grow the dimension $d$. It
seems these error rates are bounded by some constant. We are interested in
tighter generalization error bounds for tropical SVMs over the tropical
projection space $\mathbb{R}^{d}/\mathbb{R}{\bf 1}$.
The generalization error bound for tropical SVMs was derived without any
assumptions on data distributions. Specifically, the bound was derived purely
combinatorially only based on the shapes of the hyperplanes. Algorithmic
details of tropical SVMs are not yet taken into account. Thus it is not
surprising that the tropical SVMs under the max-plus algebra outperform in the
case where a few axes are much more informative than the others. Thus the
evaluation of the tropical SVM in computational experiments gave complementary
information.
The max-plus algebra also leads to the anomalous $\log n$ scaling behaviors in
the tropical distance $d_{\rm tr}$ between random vectors of length $n$ and
its scaling behavior is the direct consequence of the extreme value
statistics. As it is also essential for the theoretical explanation of the
robustness of tropical SVMs against the curse of dimensionality, extreme value
statistics play the key role in the computation with the tropical metric.
For computational experiments, we used software developed by Tang et al.
(2020). Tang et al. (2020) developed heuristic methods for hard margin and
soft margin tropical SVMs over $\mathbb{R}^{d}/\mathbb{R}{\bf 1}$ since it is
infeasible to compute them in practice. We conjecture that computing optimal
separating tropical hyperplanes for tropical SVMs is NP-hard. However, we
still do not know the exact computational time complexity in terms of $n$ and
$d$.
## References
* Améndola et al. (2020) Améndola, C., Klüppelberg, C., Lauritzen, S., Tran, N., 2020\. Conditional independence in max-linear bayesian networks. Available at https://arxiv.org/abs/2002.09233.
* Bartlett and Mendelson (2003) Bartlett, P.L., Mendelson, S., 2003\. Rademacher and gaussian complexities: Risk bounds and structural results. J. Mach. Learn. Res. 3, 463–482.
* Boser et al. (1992) Boser, B.E., Guyon, I.M., Vapnik, V.N., 1992. A training algorithm for optimal margin classifiers, in: Proceedings of the 5th Annual ACM Workshop on Computational Learning Theory, ACM Press. pp. 144–152.
* Bradley and Mangasarian (1998) Bradley, P.S., Mangasarian, O.L., 1998\. Feature selection via concave minimization and support vector machines, in: Proceedings of the Fifteenth International Conference on Machine Learning, Morgan Kaufmann Publishers Inc., San Francisco, CA, USA. p. 82–90.
* Burges (1998) Burges, J.C., 1998. A tutorial on support vector machines for pattern recognition. Data Mining and Knowledge Discovery 2, 121–167.
* Cohen et al. (2012) Cohen, J.Y., Haesler, S., Vong, L., Lowell, B.B., Uchida, N., 2012. Neuron-type-specific signals for reward and punishment in the ventral tegmental area. Nature 482, 85–8.
* Cortes and Vapnik (1995) Cortes, C., Vapnik, V., 1995\. Support-vector networks, in: Machine Learning, pp. 273–297.
* Dayan and Abbott (2001) Dayan, P., Abbott, L.F., 2001\. Theoretical Neuroscience. MIT Press.
* Fan and Ren (2006) Fan, J., Ren, Y., 2006. Statistical analysis of dna microarray data in cancer research. Clin Cancer Res. 1;12, 4469–73. Doi: 10.1158/1078-0432.CCR-06-1033. PMID: 16899590.
* Gärtner and Jaggi (2006) Gärtner, B., Jaggi, M., 2006\. Tropical support vector machines. ACS Technical Report. No.: ACS-TR-362502-01.
* Guermeur (2007) Guermeur, Y., 2007. How sure are you that large margin implies low vc dimension. Journal of Machine Learning Research 8, 2551–2594.
* Gumbel (2004) Gumbel, E.J., 2004. Statistics of Extremes. Dover.
* Hayashi and Watanabe (2017) Hayashi, N., Watanabe, S., 2017\. Upper bound of bayesian generalization error in non-negative matrix factorization. Neurocomputing 266, 21–28.
* Hung-Chih Chiang et al. (2000) Hung-Chih Chiang, Moses, R.L., Potter, L.C., 2000. Model-based classification of radar images. IEEE Transactions on Information Theory 46, 1842–1854. doi:10.1109/18.857795.
* Ishikawa et al. (2018) Ishikawa, T., Matsumoto, H., Miura, K., 2018. Discriminating dopamine from non-dopamine neurons based on the spontaneous electrical activities. Proc of IEEE Brain Initiative Workshop on Advanced NeuroTechnologies .
* Ishikawa et al. (2019) Ishikawa, T., Matsumoto, H., Miura, K., 2019. Identification of midbrain dopamine neurons using features from spontaneous spike activity patterns. IEEE EMBC 2019 , 2990–2993.
* Jaggi et al. (2008) Jaggi, M., Katz, G., Wagner, U., 2008. New results in tropical discrete geometry.
* Joswig (2005) Joswig, M., 2005. Tropical halfspaces. Combinatorial and computational geometry 52, 409–431.
* Joswig (2022) Joswig, M., 2022. Essentials of tropical combinatorics. Graduate Studies in Mathematics, American Mathematical Society, Providence, RI.
* Maclagan and Sturmfels (2015) Maclagan, D., Sturmfels, B., 2015\. Introduction to Tropical Geometry. volume 161 of Graduate Studies in Mathematics. American Mathematical Society, Providence, RI.
* Matsumoto et al. (2016) Matsumoto, H., Tian, J., Uchida, N., Watabe-Uchida, M., 2016\. Midbrain dopamine neurons signal aversion in a reward-context-dependent manner. Elife 5, e17328.
* Miura (2013) Miura, K., 2013. A semiparametric covariance estimator immune to arbitrary signal drift. Interdisciplinary Information Sciences 19, 35–41.
* Miura et al. (2005) Miura, K., Okada, M., Amari, S., 2005. Unbiased estimator of shape parameter for spiking irregularities under changing environments. Advances in Neural Information Processing Systems 18, 891–8.
* Miura et al. (2006) Miura, K., Okada, M., Amari, S., 2006. Estimating spiking irregularities under changing environments. Neural Comput. 18, 2359–86.
* Miura et al. (2007) Miura, K., Tsubo, Y., Okada, M., Fukai, T., 2007\. Balanced excitatory and inhibitory inputs to cortical neurons decouple firing irregularity from rate modulations. J Neurosci. 27, 13802–12.
* Mohri et al. (2018) Mohri, M., Rostamizadeh, A., Talwalkar, A., 2018. Foundations of Machine Learning. Adaptive Computation and Machine Learning. 2 ed., MIT Press, Cambridge, MA.
* Nitanda et al. (2021) Nitanda, A., Wu, D., Suzuki, T., 2021. Particle dual averaging: Optimization of mean field neural network with global convergence rate analysis, in: Advances in Neural Information Processing Systems, Curran Associates, Inc. URL: https://proceedings.neurips.cc/paper/2021/file/a34e1ddbb4d329167f50992ba59fe45a-Paper.pdf.
* Okun et al. (2015) Okun, M., Steinmetz, N., Cossell, L., Iacaruso, M.F., Ko, H., Barthó, P., Moore, T., Hofer, S.B., Mrsic-Flogel, T.D., Carandini, M., Harris, K.D., 2015. Diverse coupling of neurons to populations in sensory cortex. Nature 521, 511–515.
* Page et al. (2020) Page, R., Yoshida, R., Zhang, L., 2020. Tropical principal component analysis on the space of phylogenetic trees. Bioinformatics 36, 4590–4598. URL: https://doi.org/10.1093/bioinformatics/btaa564, doi:10.1093/bioinformatics/btaa564.
* Peng et al. (2016) Peng, B., Wang, L., Wu, Y., 2016. An error bound for l1-norm support vector machine coefficients in ultra-high dimension. J. Mach. Learn. Res. 17, 8279–8304.
* Ren (2015) Ren, Y., 2015. Tropical geometry in singular, dissertation. Technische Universität Kaiserslautern, Germany.
* Stringer et al. (2019) Stringer, C., Pachitariu, M., Steinmetz, N., Reddy, C.B., Carandini, M., Harris, K.D., 2019\. Spontaneous behaviors drive multidimensional, brainwide activity. Science 364, 255–8.
* Suzuki (2020) Suzuki, T., 2020. Generalization bound of globally optimal non-convex neural network training: Transportation map estimation by infinite dimensional langevin dynamics, in: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (Eds.), Advances in Neural Information Processing Systems, Curran Associates, Inc.. pp. 19224–19237. URL: https://proceedings.neurips.cc/paper/2020/file/df1a336b7e0b0cb186de6e66800c43a9-Paper.pdf.
* Takahashi et al. (2019) Takahashi, T., Maruyama, Y., Ito, H., Miura, K., 2019. Assessing the impacts of correlated variability with dissociated timescales. eNeuro 6, ENEURO.0395–18.2019.
* Tang et al. (2020) Tang, X., Wang, H., Yoshida, R., 2020. Tropical support vector machines and its applications to phylogenomics. Available at https://arxiv.org/abs/2003.00677.
* Vapnik (1995) Vapnik, V.N., 1995. The nature of statistical learning theory. Springer-Verlag New York, Inc.
* Vapnik (2000) Vapnik, V.N., 2000. The Nature of Statistical Learning Theory. 2nd edition ed., Springer.
* Yoshida et al. (2019) Yoshida, R., Zhang, L., Zhang, X., 2019. Tropical principal component analysis and its application to phylogenetics. Bulletin of Mathematical Biology 81, 568–597.
* Zhang and Zhou (2010) Zhang, L., Zhou, W., 2010. On the sparseness of 1-norm support vector machines. Neural Netw. 23, 373–385. URL: https://doi.org/10.1016/j.neunet.2009.11.012, doi:10.1016/j.neunet.2009.11.012.
|
# A Characterization for Optimal Bundling of Products with Non-Additive Values
Soheil Ghili
Yale University E-mail<EMAIL_ADDRESS>Click here for the most current
version. I thank Dirk Bergemann, Nima Haghpanah, Johannes Horner, Peter
Klibanoff, Barry Nalebuff, Larry Samuelson, Kai Hao Yang, Jidong Zhou, and
seminar participants for helpful comments. All errors are mine.
###### Abstract
This paper studies optimal bundling of products with non-additive values.
Under monotonic preferences and single-peaked profits, I show a monopolist
finds pure bundling optimal if and only if the optimal sales volume for the
grand bundle is larger than the optimal sales volume for any smaller bundle. I
then (i) detail how my analysis relates to “ratio monotonicity” results on
bundling; and (ii) describe the implications for non-linear pricing.
## 1 Introduction
This paper studies optimal bundling decisions by a multi-product monopolist,
and carries out an analysis with two main features. First, I allow for non-
additive values: a consumer’s valuation for a given product can depend on
whether s/he has purchased other products as well. Second, I seek to obtain a
full characterization of when pure bundling (i.e., the act of selling only the
package of all available products together as one bundle) is optimal. Under
monotonic preferences and single-peaked profits, I prove that optimal bundling
admits a simple characterization: Pure bundling is optimal if the optimal
sales volume for the grand bundle (if sold alone) is strictly larger than that
for any other bundle. Conversely, if there is at least one bundle whose
optimal sales volume (if sold alone) is strictly larger than that of the grand
bundle, then pure bundling is sub-optimal.111Note that this is slightly short
of a full characterization given it does not determine the optimal bundling
strategy when the optimal sales volume for the grand bundle is larger than
those for smaller bundles but not always strictly. This small gap can be
closed but with stronger assumptions that I decided not to make. See section 3
for more details. In simpler terms, bundling is optimal if and only if it
helps sell more.
The rest of the paper is organized as follows. Section 2 reviews the related
literature. Section 3 sets up the model and formally presents the assumptions
and the main result. Section 4 provides the proof. Section 5 discusses the
assumptions and delves into the implications and interpretations of the main
result. It also discusses additional results which are provided in the
appendix. Section 6 concludes.
## 2 Related Literature
The study of bundling dates at least as far back as Stigler (1963). Most
papers in this literature focus on the case of “additive values,” meaning the
valuation by each consumer of any given product $i$ is not impacted by whether
she also possesses product $i^{\prime}\neq i$. Pioneering in this area was
Adams and Yellen (1976), pointing out that bundling can be more profitable
than unbundling when there is negative correlation among consumers in how they
value individual products. Other studies such as McAfee et al. (1989);
Menicucci et al. (2015); Pavlov (2011); Schmalensee (1984); Fang and Norman
(2006); Manelli and Vincent (2007); Palfrey (1983); Bergemann et al. (2021)
further develop results on optimal bundling (or optimal upgrade pricing) under
additive values. Many of these studies focus on a setting with two products
only. Also, with few exceptions– e.g., Daskalakis et al. (2017) who provide
necessary and sufficient conditions–most of these studies concentrate on
sufficient conditions for bundling. Although most of this literature examines
a monopolist seller (which is also the focus of this paper), some studies have
analyzed multiple sellers (McAfee et al. (1989); Zhou (2017, 2019)).
The literature on non-additive values, to which this paper belongs, is
considerably smaller. Part of this literature focuses directly on bundling
(e.g., Haghpanah and Hartline (2021); Armstrong (2013, 2016); Long (1984))
whereas some study price discrimination settings which have implications for
bundling (e.g., Anderson and Dana Jr (2009); Deneckere and Preston McAfee
(1996)). This paper complements the literature on non-additive values in that
it imposes a different set of assumptions (stronger only than those imposed by
Haghpanah and Hartline (2021)) and delivers simple-to-interpret but necessary
and sufficient conditions for bundling based on optimal quantities sold. I
also connect the interpretation of my results to those based on price
elasticities (such as Long (1984); Armstrong (2013)) as well as those based on
ratio monotonicity (such as Haghpanah and Hartline (2021); Anderson and Dana
Jr (2009); Deneckere and Preston McAfee (1996); Salant (1989)). Finally, I
describe implications of my bundling results for screening models and
nonlinear tariff design (e.g., Mussa and Rosen (1978); Maskin and Riley
(1984)).
## 3 Main Result
### 3.1 Setup and Notations
A monopolist has $n$ products to sell, indexed 1 through $n$. Possible bundles
of these products are denoted $b\subseteq\\{1,...,n\\}$. Set
$\mathcal{B}=\\{b|b\subseteq\\{1,...,n\\}\\}$ represents the set of all
possible bundles.222My notation, in part, follows Haghpanah and Hartline
(2021). By $\bar{b}$ denote the grand bundle $\\{1,...,n\\}$. For any bundle
$b$, denote $b^{C}=\bar{b}\setminus b$. There is a unit mass of customers
whose types are represented by $t\in T\subset\mathbb{R}^{m}$ where $T$ is
compact. As will be shown later, one of the model assumptions will imply that
types are one-dimensional (i.e., $m=1$.)333The reason why I start with general
$m$ and later show $m=1$ is that this exposition clarifies that $m=1$ is a an
implication of the rest of the model assumptions, rather than a separate
assumption itself. Probability distribution over types $f(\cdot)>0$ has no
atoms.444The main result should hold without these assumptions on $f$. But I
expect the proof to be less clean. The valuation by type $t$ for bundle $b$ is
denoted $v(b,t)$. Assume $v(\emptyset,t)=0$. Also, for all $b$, suppose that
$v(b,t)$ is continuous in $t$ except, possibly, for finitely many points. The
per-unit cost of production for each product $i$ is $c_{i}\geq 0$.
The problem the monopolist solves has two components. First, she makes a
bundling decision. She chooses the optimal set $B^{*}$ of bundles $b$ among
subsets $B$ of $\mathcal{B}$ that satisfy $\emptyset\notin B$. Note that there
are $2^{2^{n}-1}$ possible bundling strategies. Thus, characterizing the
conditions under which the monopolist can simply choose $B^{*}=\\{\bar{b}\\}$
should indeed be of value.
The second decision by the firm is choosing prices
$p(\cdot):B\rightarrow\mathbb{R}$ for the bundles offered.555Note that, in
principle, one could model the bundling decision through pricing; because not
offering a product would be equivalent to pricing it so high that no customer
would purchase it. As such, separating the bundling and pricing decisions in
the model is redundant. Nevertheless, I decided to carry out this separation
because it makes the notation easier. Denote by $\mathcal{P}_{B}$ the set of
all possible such pricing functions.
Once the firm decides on set $B$ and prices $p(\cdot)$, customers decide which
bundles to purchase (note that this model only considers deterministic selling
procedures.) Each customer $t$’s decision $\beta(t|B,p)\subseteq B$ is
determined by:
$\beta(t|B,p)=\arg\max_{\hat{\beta}\subseteq
B}v(\cup_{b\in\hat{\beta}}b,t)-\Sigma_{b\in\hat{\beta}}p(b)$ (1)
Throughout, I assume customers break ties in favor of the seller. Also, note
that equation 1 implies that customers want at most one unit of each product
$i$ and find additional units redundant.
Demand for bundle $b$ is given by the measure of customers $t$ who would
choose to purchase bundle $b$:
$D(b|B,p)=\int_{t}\mathds{1}_{b\in\beta(t|B,p)}f(t)dt$ (2)
Firm profit under strategy $(B,p)$ is:
$\pi(B,p)=\int_{t}\Sigma_{b\in\beta(t|B,p)}\bigg{(}p(b)-\Sigma_{i\in
b}c_{i}\bigg{)}f(t)dt$ (3)
The monopolist chooses $(B^{*},p^{*})$ to maximize profit:
$(B^{*},p^{*})=\arg\max_{B\in\mathcal{B},p\in\mathcal{P}_{B}}\pi(B,p)$ (4)
With the setup of the firm problem laid out, I now introduce a few more
definitions and notations. For disjoint bundles $b$ and $b^{\prime}$, denote
by $v(b,t|b^{\prime})$ the valuation by type $t$ for $b$ conditional on
possessing $b^{\prime}$. Formally:
$v(b,t|b^{\prime})\equiv v(b\cup b^{\prime},t)-v(b^{\prime},t)$
In a similar manner, denote
$\beta(t|B,p,b^{\prime})=\arg\max_{\hat{\beta}\subseteq(B\setminus\\{b^{\prime}\\})}v(\cup_{b\in\hat{\beta}}b,t|b^{\prime})-\Sigma_{b\in\hat{\beta}}p(b)$.
Also denote
$D(b|B,p,b^{\prime})=\int_{t}1_{b\in\beta(t|B,p,b^{\prime})}f(t)dt$. Moreover,
denote
$\pi(B,p|b^{\prime})=\int_{t}\Sigma_{b\in\beta(t|B,p,b^{\prime})}\bigg{(}p(b)-\Sigma_{i\in
b}c_{i}\bigg{)}f(t)dt.$
Finally, for any set $\beta$ of bundles, denote
$\beta^{\cup}=\cup_{b\in\beta}b$. Similarly, I write
$\beta^{\cup}(t|B,p,b^{\prime})$ instead of
$\cup_{b\in\beta(t|B,p,b^{\prime})}b$. I next turn to the assumptions and the
main result.
### 3.2 Assumptions and Characterization
The main results of the paper is about how optimal bundling decisions are
informed by the comparison among optimal sales volumes for different bundles.
I start with some necessary assumptions and definitions.
###### Assumption 1.
Monotonicity: For all $b$, $v(b,t)$ is increasing in $v(\bar{b},t)$, and
strictly so whenever $v(b,t)>0$. The same applies to $v(b,t|b^{C})$ for all
$b$.
###### Assumption 2.
Quasi-concavity: For any $b\in\mathcal{B}$, profit functions $\pi(b,p)$ and
$\pi(b,p|b^{C})$ are strictly quasiconcave in $p(b)$ for all values of $p(b)$
that yield strictly positive demand for $b$.666This simply means the profit
peaks only once as we vary each price.
###### Definition 1.
By $D^{*}(b)$ denote the “optimal quantity sold” of bundle $b$ if no other
bundle were offered by the firm. Formally, $D^{*}(b)$ is defined as
$D(b|\\{b\\},p_{b}^{*})$ where $p_{b}^{*}$ is the optimal price for bundle $b$
when $B=\\{b\\}$.
###### Definition 2.
A given firm strategy $(B,p)$ involves pure bundling if:
$\forall t:\beta^{\cup}(t|B,p)\in\\{\emptyset,\bar{b}\\}$
We are now ready to state the main result.
###### Theorem 1.
Under assumptions 1 and 2, the optimal strategy $(B^{*},p^{*})$ involves pure
bundling if:
$D^{*}(\bar{b})>\max_{b\in\mathcal{B}\setminus\\{\bar{b}\\}}D^{*}(b)$
Conversely, the optimal strategy does not involve pure bundling if:
$D^{*}(\bar{b})<\max_{b\in\mathcal{B}\setminus\\{\bar{b}\\}}D^{*}(b)$
In words, this result says that the firm should pure bundle if and only if it
helps “sell more.”777Note that this result is slightly short of a full
characterization because it does not specify whether pure bundling is optimal
when $D^{*}(\bar{b})=\max_{b\in\mathcal{B}\setminus\\{\bar{b}\\}}D^{*}(b)$.
One can show that under this last possibility, pure bundling is optimal if
instead of assuming profits are strictly quasi-concave in each price, we
assume they are strictly concave and differntiable at peak. Even though this
would yield a full characterization, I decided that the ability to speak to
the “measure-zero” case of
$D^{*}(\bar{b})=\max_{b\in\mathcal{B}\setminus\\{\bar{b}\\}}D^{*}(b)$ is too
small a return to justify such a restrictive assumption as strict concavity.
As such, I maintain the quasi-concavity assumption.
## 4 Proof of Theorem 1
I start by some remarks, definitions, and lemmas.
###### Lemma 1.
There is a mapping $\tau$ from the set $T$ of types $t$ on to the interval
$[0,1]$ such that:
1. 1.
$\forall t,t^{\prime}\in
T:v(\bar{b},t)>v(\bar{b},t^{\prime})\Leftrightarrow\tau(t)>\tau(t^{\prime})$.
2. 2.
$\tau$ is a sufficient statistic: Once $\tau(t)$ is known, one can fully pin
down all $v(b,t)$ without having to know $t$.
Proof of Lemma 1. Set $\tau(t)\triangleq\frac{v(\bar{b},t)-\min_{t^{\prime}\in
T}v(\bar{b},t^{\prime})}{\max_{t^{\prime}\in
T}v(\bar{b},t^{\prime})-\min_{t^{\prime}\in T}v(\bar{b},t^{\prime})}$. By
construction, it satisfies (1). To see why it satisfies (2), first note that
by monotonicity, for any $t,t^{\prime}$ such that
$v(\bar{b},t)=v(\bar{b},t^{\prime})$, we have $v(b,t)=v(b,t^{\prime})$ for any
other $b$. As a result, $v(\bar{b},t)$ is sufficient information for
determining $v(b,t)$ for all $b$. Next, observe that one can recover
$v(\bar{b},t)$ from $\tau(t)$: $v(\bar{b},t)=\min_{t^{\prime}\in
T}v(\bar{b},t^{\prime})+\tau(t)\times(\max_{t^{\prime}\in
T}v(\bar{b},t^{\prime})-\min_{t^{\prime}\in T}v(\bar{b},t^{\prime}))$. As a
result, once one knows $\tau(t)$, one would also know $v(b,t)$ for all $b$.
Q.E.D.
Based on this lemma, it is without loss to think of $t$ as $\tau(t)$ and,
hence, the set of all possible $t$ as $[0,1]$. Therefore, we can use
expressions such as $t\geq t^{\prime}$. Going forward, I assume $t\in[0,1]$.
###### Remark 1.
Suppose functions $f_{1}(x)$, $f_{2}(x)$ and $f_{1}(x)+f_{2}(x)$ are all
strictly quasi-concave over the interval $[a,b]$. Then either (i) $\arg\max
f_{1}\leq\arg\max f_{2}$ or (ii) $\arg\max f_{1}\leq\arg\max(f_{1}+f_{2})$
will imply:
$\arg\max f_{1}\leq\arg\max(f_{1}+f_{2})\leq\arg\max f_{2}.$
The proof is left to the reader.
###### Definition 3.
For disjoint bundles $b$ and $b^{\prime}$, denote by $D^{*}(b|b^{\prime})$ the
“optimal quantity sold” of bundle $b$ if all customers are already endowed
with $b^{\prime}$ and only $b$ is offered by the firm at optimal price.
Formally, $D^{*}(b|b^{\prime})\equiv
D(b|\\{b\\},p_{b|b^{\prime}}^{*},b^{\prime})$ where
$p_{b|b^{\prime}}^{*}:\\{b\\}\rightarrow\mathbb{R}$ is effectively one real
number, and it is chosen among other possible $p$ so that
$\pi(\\{b\\},p|b^{\prime})$ is maximized.
Next, I show that the problem of finding the optimal price for a bundle is
equivalent to the problem of finding the right type $t^{*}$ and sell to types
$t\geq t^{*}$.
###### Definition 4.
Define by $t^{*}(b|b^{\prime})$ the largest $t$ such that $1-F(t)\geq
D^{*}(b|b^{\prime})$. Also, for simplicity, denote $t^{*}(b|\emptyset)$ by
$t^{*}(b)$.
###### Lemma 2.
Consider bundles $b$ and $b^{C}=\bar{b}\setminus b$. Suppose that all types
are endowed with bundle $b^{C}$, and that the firm is selling only bundle $b$,
optimally choosing $p^{*}_{b|b^{C}}$. The set of types who will buy the
product at this price is the interval $[t^{*}(b|b^{C}),1]$.
Proof of Lemma 2. Follows directly from monotonicity. Monotonicity implies
that the optimal sales volume $D^{*}(b|b^{C})$ would be purchased by the
highest types $t$ with $t$ weakly above some cutoff $\hat{t}$. Definition 4
says that for the demand volume to equal $D^{*}(b|b^{C})$, the cutoff
$\hat{t}$ has to equal $t^{*}(b|b^{C})$. Q.E.D.
Lemma 2 is important in that it shows the problem of choosing
$p^{*}_{b|b^{C}}$ can equivalently be thought of as the problem of choosing
$t^{*}_{b|b^{C}}$. This allows us to set up the firm’s problem based on $t$.
Next definition introduces a necessary notation for this purpose.
###### Definition 5.
Consider disjoint bundles $b$ and $b^{\prime}$. Suppose that all types have
already been endowed with $b^{\prime}$, and that the firm is to sell only
bundle $b$. By $\pi_{b}(t|b^{\prime})$ denote the profit to the firm if it
chose a price for bundle $b$ such that all types $t^{\prime}\geq t$ would
purchase bundle $b$:
$\pi_{b}(t|b^{\prime})=(1-F(t))\times\big{(}(v(t,b|b^{\prime})-\Sigma_{i\in
b}c_{i}\big{)}$ $=\pi(\\{b\\},v(b,t|b^{\prime})|b^{\prime})$
###### Lemma 3.
$\pi_{b}(t|b^{C})$ is strictly quasi-concave in $t$.
Proof of Lemma 3. Suppose $\pi_{b}(t|b^{C})$ is not quasi-concave in $t$.
This means there are $t_{1}<t_{2}<t_{3}$ such that
$\pi_{b}(t_{2}|b^{C})\leq\min(\pi_{b}(t_{1}|b^{C}),\pi_{b}(t_{3}|b^{C}))$.
Then construct $p_{1},p_{2}$ and $p_{3}$ from $t_{1},t_{2}$ and $t_{3}$
according to the procedure in definition 5. That is, set $p_{i}=v(b,t|b^{C})$
for each $i$. Monotonicity puts $p_{2}$ strictly between $p_{1}$ and $p_{3}$.
Note that for these prices, we have:
$\pi(\\{b\\},p_{2}|b^{C})\leq\min(\pi(\\{b\\},p_{1}|b^{C}),\pi(\\{b\\},p_{3}|b^{C}))$
which violates the quasi-concavity assumption in $p$. Q.E.D.
With the above definitions and lemmas in hand, we are ready to prove the main
theorem. I start by the necessity condition (i.e., the condition that
$D^{*}(\bar{b})\geq D^{*}(b)$ for all $b$ is necessary for pure bundling to
optimal).
Proof of necessity. We want to show that if there is some $b$ such that
$D^{*}(b)>D^{*}(\bar{b})$, then pure bundling is sub-optimal. Specifically, I
show that offering bundles $b$ and $\bar{b}$ would be strictly more profitable
to the firm compared to offering $\bar{b}$ alone. The argument follows.
###### Lemma 4.
$D^{*}(b)>D^{*}(\bar{b})$ implies $D^{*}(b)>D^{*}(b^{C}|b)$.
Proof of Lemma 4. Suppose, on the contrary, that $D^{*}(b)\leq
D^{*}(b^{C}|b)$. This means $t^{*}(b)\geq t^{*}(b^{C}|b)$. We know:
$t^{*}(b)=\arg\max_{t}\pi_{b}(t)$
and
$t^{*}(b^{C}|b)=\arg\max_{t}\pi_{b^{C}}(t|b).$
Also, given definition 5, it is straightforward to verify that:
$\pi_{\bar{b}}(t)\equiv\pi_{b^{C}}(t|b)+\pi_{b}(t)$
By strict quasi-concavity of all profits in $t$ and by remark 1, it has to be
that the argmax of $\pi_{\bar{b}}(t)$ falls in between the argmax values
$t^{*}(b^{C}|b)$ and $t^{*}(b)$. Therefore, we get: $t^{*}(\bar{b})\leq
t^{*}(b)$, which implies $D^{*}(b)\leq D^{*}(\bar{b})$, contradicting a
premise of the lemma. Q.E.D.
###### Lemma 5.
Selling $D^{*}(b^{C}|b)$ units of the grand bundle $\bar{b}$ along with
$D^{*}(b)-D^{*}(b^{C}|b)$ units of bundle $b$ would be strictly more
profitable to the monopolist compared to selling $D^{*}(\bar{b})$ units of the
grand bundle alone.
Proof of Lemma 5. Note that given monotonicity and given Lemma 4, selling
$D^{*}(b^{C}|b)$ units of the grand bundle $\bar{b}$ along with
$D^{*}(b)-D^{*}(b^{C}|b)$ units of bundle $b$ would simply mean selling $b$ to
types $[t^{*}(b),t^{*}(b^{C}|b))$ and selling $\bar{b}$ to types
$[t^{*}(b^{C}|b),1]$. This can be implemented by offering bundles $b$ and
$\bar{b}$ and pricing them at $p^{*}_{b}$ and $p^{*}_{b}+p^{*}_{b^{C}|b}$
respectively.888Recall that I use the notation $p^{*}_{b}$ for the optimal
price of $b$ when only $b$ is offered. Similarly, $p^{*}_{b^{C}|b}$ is the
optimal price of $b^{C}$ when everyone is endowed by $b$ and the monopolist is
only selling $b^{C}$. This would deliver the following profit:
$\pi_{1}=\pi_{b^{C}}(t^{*}(b^{C}|b)|b)+\pi_{b}(t^{*}(b))$
Again, by monotonicity, selling $D^{*}(\bar{b})$ units of the grand bundle can
be thought of as selling $\bar{b}$ to types $t^{*}(\bar{b})$ and above. This
would deliver a profit of $\pi_{2}=\pi_{\bar{b}}(t^{*}(\bar{b}))$, which can
be expanded and written as:
$\pi_{2}=\pi_{b^{C}}(t^{*}(\bar{b})|b)+\pi_{b}(t^{*}(\bar{b}))$
Note that each term in $\pi_{2}$ is weakly less than its corresponding term in
$\pi_{1}$ (due to the optimality of the terms in $\pi_{1}$). Also by the fact
that $t^{*}(b^{C}|b)>t^{*}(b)$, then either $t^{*}(b^{C}|b)\neq
t^{*}(\bar{b})$ or $t^{*}(b)\neq t^{*}(\bar{b})$. Thus, by strict quasi-
concavity, at least one of the two inequalities between corresponding terms in
$\pi_{1}$ and $\pi_{2}$ has to be strict, yielding $\pi_{1}>\pi_{2}$. Q.E.D.
Given this lemma, the proof of the if side of the theorem is now complete.
Q.E.D.
Note that the proof of the first side did not use the constant marginal costs
assumption. Next, I turn to the proof of the sufficiency conditions (i.e.,
that $D^{*}(\bar{b})>\max_{b\in\mathcal{B}\setminus\\{\bar{b}\\}}D^{*}(b)$
implies that pure bundling is optimal).
Proof of sufficiency. I start with some lemmas.
###### Lemma 6.
Under assumptions 1 and 2, there is a firm optimal strategy $(B^{*},p^{*})$
such that non-measure-zero set of customers $t$ we have
$\beta^{\cup}(t|B^{*},p^{*})=\bar{b}$.
Proof of Lemma 6. I start by assuming that no optimal strategy $(B^{*},p^{*})$
involves selling $\bar{b}$ to a non-measure-zero set of consumers. Then I
reach a contradiction by constructing a weakly profitable deviation from an
assumed optimal $(B^{*},p^{*})$ such that the deviation sells $\bar{b}$ to a
non-measure-zero set of consumers.
By $D^{*}(\bar{b})>D^{*}(b)$ for all $b\neq\bar{b}$, we get:
$D^{*}(\bar{b})>0$, which in turn yields $t^{*}(\bar{b})<1$. Next, note that
the number of possibilities for $\beta(t|B^{*},p^{*})$ is finite. By this
finiteness and by piece-wise continuity of value functions, there is some
$\tilde{t}\geq t^{*}(\bar{b})$ such that all consumers with types higher than
$\tilde{t}$ have the same purchase behavior.999Perhaps with the exception of
type $t=1$; but that does not matter given its zero measure. Formally:
$\forall
t,t^{\prime}\in(\tilde{t},1):\beta^{\cup}(t|B^{*},p^{*})=\beta^{\cup}(t^{\prime}|B^{*},p^{*})$
Denote this commonly purchased bundle $\tilde{b}$. Given that our
contrapositive assumption is that no non-measure-zero set of types purchases
the grand bundle $\bar{b}$, it has to be that $\tilde{b}\neq\bar{b}$. Next, I
construct a profitable deviation for the monopolist from $(B^{*},p^{*})$.
First, note that given that currently no consumer purchases $\bar{b}$ or
constructs it from other bundles, it has to be that either $\bar{b}$ is not
part of $B^{*}$ or it is expensive enough for no customer to prefer to obtain
it. Next, assume that the monopolist deviates from $(B^{*},p^{*})$ by adding
$\bar{b}$ to the set of bundles and pricing it at
$p^{*}(\tilde{b})+\Sigma_{i\in\tilde{b}^{C}}c_{i}+\epsilon$ where $\epsilon$
is chosen so that (i) type $\tilde{t}$ would weakly prefer $\tilde{b}$ over
$\bar{b}$ but (ii) type 1 would weakly prefer $\bar{b}$ over $\tilde{b}$.
Next, I proceed to show two things. First: finding such an $\epsilon$ is
feasible. Second: with that $\epsilon$, the monopolist will see a weak profit
increase relative to $(B^{*},p^{*})$ and sell $\bar{b}$ to a positive-measure
set of types.
For type $\tilde{t}$ to weakly prefer $\tilde{b}$ over $\bar{b}$, it has to be
that
$v(\bar{b},\tilde{t})-p^{*}(\tilde{b})-\Sigma_{i\in\tilde{b}^{C}}c_{i}-\epsilon\leq
v(\tilde{b},\tilde{t})-p^{*}(\tilde{b})$
$\Leftrightarrow\epsilon\geq
v(\tilde{b}^{C},\tilde{t}|\tilde{b})-\Sigma_{i\in\tilde{b}^{C}}c_{i}$
Similarly, for type to 1 weakly to prefer to purchase $\bar{b}$, one can show
that $\epsilon$ must satisfy:
$\epsilon\leq v(\tilde{b}^{C},1|\tilde{b})-\Sigma_{i\in\tilde{b}^{C}}c_{i}$
But by monotonicity, we have $v(\tilde{b}^{C},\tilde{t}|\tilde{b})\leq
v(\tilde{b}^{C},1|\tilde{b})$. Therefore, $\epsilon$ may be chosen within the
following interval:
$[v(\tilde{b}^{C},\tilde{t}|\tilde{b})-\Sigma_{i\in\tilde{b}^{C}}c_{i},v(\tilde{b}^{C},1|\tilde{b})-\Sigma_{i\in\tilde{b}^{C}}c_{i}]$
If the interval is not a singleton, choose $\epsilon$ in the interior.
Next, I show that once such $\epsilon$ is chosen, the new bundling and pricing
strategy by the firm will weakly increase the profit to the insurer while
selling $\bar{b}$ to a non-measure zero set of consumers. To see the latter,
denote by $\tilde{t}^{\prime}$ the set of types who weakly prefer $\bar{b}$
over $\tilde{b}$ under this new strategy. By the choice of $\epsilon$ and by
monotonicity, we have $\tilde{t}\leq\tilde{t}^{\prime}<1$. Therefore the new
strategy will change the purchase behavior by types $t\geq\tilde{t}^{\prime}$
(and only those types.) Next note that this behavior-change is weakly
profitable to the monopolist. Prior to this change, the profit to the
monopolist from these types was:
$\pi_{1}=(1-\tilde{t}^{\prime})\times(p^{*}(\tilde{b})-\Sigma_{i\in\tilde{b}}c_{i})$
Under the new strategy (i.e., with the introduction of $\bar{b}$ at the price
of $p^{*}(\tilde{b})+\Sigma_{i\in\tilde{b}^{C}}c_{i}+\epsilon$,) the new
profit level from these types is:
$\pi_{2}=(1-\tilde{t}^{\prime})\times\bigg{(}(p^{*}(\tilde{b})+\Sigma_{i\in\tilde{b}^{C}}c_{i}+\epsilon)-\Sigma_{i\in\bar{b}}c_{i}\bigg{)}$
$=\pi_{1}+(1-\tilde{t}^{\prime})\times\epsilon$
Thus, it remains to show $\epsilon\geq 0$. To this end, note that by
$D^{*}(\bar{b})>D^{*}(\tilde{b})$ we have $t^{*}(\bar{b})\leq
t^{*}(\tilde{b})$. This, by monotonicity, quasi-concavity, and remark 1,
implies $t^{*}(\tilde{b}^{C}|\tilde{b})\leq t^{*}(\bar{b})$ which in turn
yields $t^{*}(\tilde{b}^{C}|\tilde{b})\leq\tilde{t}$. That is, all types
weakly above $\tilde{t}$ would purchase $\tilde{b}^{C}$ if (i) they were
offered it at the optimal price for the monopolist and (ii) they were pre-
endowed with $\tilde{b}$. This implies that $\forall
t\geq\tilde{t}:v(\tilde{b}^{C},t|\tilde{b})\geq
p^{*}_{\tilde{b}^{C}|\tilde{b}}\geq\Sigma_{i\in\tilde{b}^{C}}c_{i}$. But this
means
$\epsilon=v(\tilde{b}^{C},\tilde{t}|\tilde{b})-\Sigma_{i\in\tilde{b}^{C}}c_{i}\geq
0$. As a result, we get $\pi_{2}\geq\pi_{1}$, which completes the proof of the
lemma.Q.E.D.
Next, it is useful to observe that the monotonicity assumption imposes a
vertical relationship not only on consumers’ preferences, but also on their
purchase behaviors.
###### Lemma 7.
Consider bundling strategy $B$ and pricing strategy $p$. Consider types $t$
and $t^{\prime}$ such that
$\beta^{\cup}(t|B,p)\neq\beta^{\cup}(t^{\prime}|B,p)$. Then the following
statements hold:
1. 1.
If $\beta^{\cup}(t^{\prime}|B,p)=\emptyset$, we have $t^{\prime}<t$.
2. 2.
If $\beta^{\cup}(t^{\prime}|B,p)=\bar{b}$, we have $t^{\prime}>t$.
This lemma says that if there are types who buy $\bar{b}$, they are the
highest types. Also if there are types who buy nothing, they are the lowest
types.
Proof of Lemma 7. I prove the second statement in the lemma. The first
statement would be proved in a similar way. Suppose that
$\beta^{\cup}(t|B,p)\neq\beta^{\cup}(t^{\prime}|B,p)=\bar{b}.$ For simplicity,
denote $\beta^{\cup}(t|B,p)=\tilde{b}$. Now suppose, contrary to the statement
of the lemma, that $t^{\prime}\leq t$. Given
$\beta(t|B,p)^{\cup}\neq\beta^{\cup}(t^{\prime}|B,p)$, we know $t\neq
t^{\prime}$ which implies $t^{\prime}<t$. Next, observe the following two
inequalities:
First, note that under $(B,p)$, type $t$ prefers purchasing $\beta(t|B,p)$ and
forming $\tilde{b}$ over purchasing $\beta(t^{\prime}|B,p)$ and forming
$\bar{b}$. Formally:
$v(\tilde{b},t)-\Sigma_{b\in\beta(t|B,p)}p(b)\geq
v(\bar{b},t)-\Sigma_{b\in\beta(t^{\prime}|B,p)}p(b)$ (5)
Similarly, type $t^{\prime}$ prefers to purchase $\beta(t^{\prime}|B,p)$ and
forming $\bar{b}$ over purchasing $\beta(t|B,p)$ and forming $\tilde{b}$.
Formally:
$v(\bar{b},t^{\prime})-\Sigma_{b\in\beta(t^{\prime}|B,p)}p(b)\geq
v(\tilde{b},t^{\prime})-\Sigma_{b\in\beta(t|B,p)}p(b)$ (6)
At least one of the two inequalities above has to be strict (because if both
types were indifferent between $\bar{b}$ and $\tilde{b}$, they would break
this tie the same way.) Adding these two inequalities together, we get:
$v(\bar{b},t^{\prime})+v(\tilde{b},t)>v(\bar{b},t)+v(\tilde{b},t^{\prime})$
$\Leftrightarrow
v(\bar{b},t^{\prime})-v(\tilde{b},t^{\prime})>v(\bar{b},t)-v(\tilde{b},t)$
$v(\tilde{b}^{C},t^{\prime}|\tilde{b})>v(\tilde{b}^{C},t|\tilde{b})$
This latter statement, combined with $t^{\prime}<t$, contradicts monotonicity.
Q.E.D.
In light of lemma 6, the following two corollaries of lemma 7 are useful.
###### Corollary 1.
Under $(B^{*},p^{*})$, the set of types to for which
$\beta^{\cup}(t|B^{*},p^{*})=\bar{b}$ takes the form of $[t_{1},1]$ for some
$t_{1}<1$.
###### Corollary 2.
Under $(B^{*},p^{*})$, the set of types to for which
$\beta^{\cup}(t|B^{*},p^{*})=\emptyset$ takes the form of $[0,t_{2})$ for some
$t_{2}<1$.
With these lemmas in hand, I next turn to the proof of the sufficiency
conditions. The strategy is, again, contrapositive.
Assume on the contrary that we have, at the same time: (i) $\forall
b\neq\bar{b}:D^{*}(b)<D^{*}(\bar{b})$ and (ii) the firm’s optimal strategy
does not involve pure bundling. This latter statement implies that the set of
all distinct bundles chosen by customers under $(B^{*},t^{*})$ includes
members other than $\emptyset$ or $\bar{b}$. Formally, if we denote
$\beta^{*}=\\{b|\exists t:\beta^{\cup}(t|B^{*},p^{*})=b\\}$
then $\beta^{*}\setminus\\{\emptyset,\bar{b}\\}\neq\emptyset$. In other words,
our contrapositive assumption implies that $t_{1}$ in corollary 1 is strictly
larger than $t_{2}$ in corollary 2.
Then, note that by corollary 1 and the piece-wise continuity of values in $t$,
there is some bundle $b_{1}\in\beta^{*}\setminus\\{\emptyset,\bar{b}\\}$ such
that for $t_{1}^{\prime}$ close enough to but smaller than $t_{1}$, we have:
$\forall t\in[t_{1}^{\prime},t_{1}):\beta^{\cup}(t|B^{*},p^{*})=b_{1}$ (7)
Also, by corollary 2 and by piece-wise continuity of values in $t$, there is
some bundle $b_{2}\in\beta^{*}\setminus\\{\emptyset,\bar{b}\\}$ (which may or
may not be the same as $b_{1}$) such that for $t_{2}^{\prime}$ close enough to
but larger than $t_{2}$, we have:
$\forall t\in[t_{2},t_{2}^{\prime}]:\beta^{\cup}(t|B^{*},p^{*})=b_{2}$ (8)
The rest of the proof of the sufficiency conditions of the theorem is
organized as follows. I first make a series of claims (without proving them).
Next I use the claims to prove the sufficiency conditions of the theorem.
Finally, I will return to the proofs of the claims.
###### Claim 1.
$t^{*}(b_{1}^{C}|b_{1})=t_{1}$.
In words, claim 1 says that the set of customers who purchase the grand bundle
$\beta(t|B^{*},p^{*})=\bar{b}$ under the firm optimal strategy $(B^{*},p^{*})$
is the same as those who purchase $b_{1}^{C}$ and construct the grand bundle
if (i) everyone is endowed with $b_{1}$ and (ii) the firm offers only
$b_{1}^{C}$, pricing it optimally.
###### Claim 2.
$t^{*}(b_{2})=t_{2}$.
Claim 2 says that the set of customers who purchase $\emptyset$ under the firm
optimal strategy $(B^{*},p^{*})$ is the same as those who purchase $\emptyset$
if the firm offers only $b_{2}$ and prices it optimally.
Next, note that the assumption $D^{*}(\bar{b})>D^{*}(b_{2})$, combined with
monotonicity and claim 2, implies $t^{*}(\bar{b})\leq t_{2}$. By
$t_{1}>t_{2}$, we get $t^{*}(\bar{b})<t_{1}=t^{*}(b_{1}^{C}|b_{1})$. Also note
that:
$\forall t:\pi_{\bar{b}}(t)=\pi_{b_{1}^{C}}(t|b_{1})+\pi_{b_{1}}(t)$
As such, by strict quasi-concavity of profits, by
$t^{*}(\bar{b})<t^{*}(b_{1}^{C}|b_{1})$, and by remark 1, the peak of
$\pi_{\bar{b}}(t)$ should happen in between those of
$\pi_{b_{1}^{C}}(t|b_{1})$ and $\pi_{b_{1}}(t)$. Therefore, we should have:
$t^{*}(b_{1})\leq t^{*}(\bar{b})\leq t^{*}(b_{1}^{C}|b_{1})$. But
$t^{*}(b_{1})\leq t^{*}(\bar{b})$ implies:
$D^{*}(b_{1})\geq D^{*}(\bar{b})$
which is a contradiction. Therefore, the sufficiency part of the theorem is
true provided that claims 1 and 2 are true. I now turn to the proofs of these
claims.
Proof of Claim 1. Suppose on the contrary that $t^{*}(b_{1}^{C}|b_{1})\neq
t_{1}$. In that case, it can be shown that the firm can strictly improve its
profit upon the optimal strategy $(B^{*},p^{*})$. The proof of this claim
constructs such improvement. To this end, the following remark is useful to
state.
###### Remark 2.
Construct the bundling strategy $(\hat{B},\hat{p})$ from $(B^{*},p^{*})$ in
the following way:
* •
$\hat{B}=\\{\beta^{\cup}(t|B^{*},p^{*})\forall t\\}$
* •
For each $t$, or in other words for each
$\hat{b}=\beta^{\cup}(t|B^{*},p^{*})\in\hat{B}$, set
$\hat{p}(\hat{b})=\Sigma_{b\in\hat{b}}p^{*}(b)$.
For such $(\hat{B},\hat{p})$, we have:
1. 1.
$\forall t:\beta(t|\hat{B},\hat{p})=\beta(t|B^{*},p^{*})$
2. 2.
$\pi(\hat{B},\hat{p})=\pi(B^{*},p^{*})$
This remark simply states that there is an optimal strategy by the seller
under which each buyer type only purchases a single bundle instead of
combining different bundles to construct her/his desired one. I skip the proof
of this remark. Also, in order to save on notation, I assume from now on that
it is $(B^{*},p^{*})$ itself that has the feature of every
$\beta(t|B^{*},p^{*})$ being a singleton. I now return to the proof of claim 1
and construct a strict improvement upon the profit of $(B^{*},p^{*})$.
I do so by slightly adjusting the price of $\bar{b}$. That is, I show that
there is a pricing strategy $p$ with $p(b)=p^{*}(b)$ for all $b\neq\bar{b}$,
but with $p(\bar{b})\neq p^{*}(\bar{b})$, such that
$\pi(B^{*},p)>\pi(B^{*},p^{*})$.
To see why this is the case, construct bundling strategy $B^{\prime}$ in the
following way:
$B^{\prime}=\\{b_{1},\bar{b}\\}$ (9)
Also construct pricing strategy $p^{\prime}$ by fixing
$p^{\prime}(b_{1})=\min_{t}v(b_{1})$ and setting
$p^{\prime}(\bar{b})=p^{\prime}(b_{1})+\rho$ where $\rho$ is a parameter that
we will vary.
More specifically, I show that as long as
$\rho\in[p^{*}(\bar{b})-p^{*}(b_{1})-\epsilon,p^{*}(\bar{b})-p^{*}(b_{1})+\epsilon]$
for a small enough $\epsilon$, then $\pi(B^{*},p)$ and
$\pi(B^{\prime},p^{\prime})$ move in parallel if we set
$p(\bar{b})=p^{*}(b_{1})+\rho$ and
$p^{\prime}(\bar{b})=p^{\prime}(b_{1})+\rho$, and move $\rho$ (that is, as we
change $\rho$, the difference $\pi(B^{*},p)-\pi(B^{\prime},p^{\prime})$
remains constant). The range parameter $\epsilon$ should be chosen so that for
any pricing strategy $p$ constructed with a $\rho$ in this interval we have:
$D(\bar{b}|B^{*},p)<1-F(t_{1}^{\prime})$ where $t_{1}^{\prime}$ was
constructed in equation 7. In other words, $\epsilon$ should be small enough
(or, alternatively, $p(\bar{b})$ should be close enough to $p^{*}(\bar{b})$ )
so that as we change $\rho$, the only types who are affected are those around
$t_{1}$; and, hence, the only purchase decisions that are affected are choices
between $b_{1}$ and $\bar{b}$.
With this setup, note that if we set $\epsilon=0$, then demand for grand
bundle $\bar{b}$ under both strategies will be equal to demand for the grand
bundle under the optimal strategy:
$\epsilon=0\Rightarrow
D(\bar{b}|B^{*},p)=D(\bar{b}|B^{\prime},p^{\prime})=D(\bar{b}|B^{*},p^{*})$
Next, note that for small $\epsilon\neq 0$, we will still have
$D(\bar{b}|B^{*},p)=D(\bar{b}|B^{\prime},p^{\prime})$ because such changes in
$\rho$ will lead the exact same set of types to switch their purchase
decisions between $\bar{b}$ and $b_{1}$, under both strategies $(B^{*},p)$ and
$(B^{\prime},p^{\prime})$. This leads to the exact same revenue change between
the two strategies as a result of the change in $\epsilon$ (or, equivalently,
in $\rho$). Also the change in total costs are the same given the constant-
marginal-costs assumption. As a result, $\pi(B^{*},p)$ and
$\pi(B^{\prime},p^{\prime})$ change in the same way as a result of small
changes in $\rho$.101010One can show we do not need the constant marginal cost
assumption if $n=2$. or if the monotonicity condition is strengthened.
Now note that if claim 1 does not hold, then $\pi(B^{\prime},p^{\prime})$
under $\epsilon=0$ is not the global maximum for $\pi(B^{\prime},p^{\prime})$.
By strict quasi-concavity, it is not a local maximum either. As a result,
there is a small change in $\rho$ that would strictly increase
$\pi(B^{\prime},p^{\prime})$. Given that $\pi(B^{\prime},p^{\prime})$ and
$\pi(B^{*},p)$ move in parallel if we slightly change $\rho$, then
$\pi(B^{*},p)$ should also strictly increase relative to $\pi(B^{*},p^{*})$, a
contradiction.Q.E.D.
Proof of Claim 2. The proof of this claim is similar to that of the previous
claim. We start by assuming, on the contrary, that $t^{*}(b_{2})\neq t_{2}$
and reach a contradiction. Construct $(B^{\prime},p^{\prime})$ by assuming
$B^{\prime}=\\{b_{2}\\}$, which makes $p^{\prime}$ just one number (for the
price of $b_{2}$). Similar to the previous claim, one can show that for prices
$\rho$ for $b_{2}$ sufficiently close to $p^{*}(b_{2})$ the two profit
functions $\pi(B^{*},p)$ and $\pi(B^{\prime},p^{\prime})$ move in parallel as
we move $\rho$. Again, similarly to the previous claim, this implies that
$(B^{*},p^{*})$ can be improved upon if $t_{2}\neq t^{*}(b_{2})$. Q.E.D.
The completion of the proofs for claims 1 and 2 finishes the proof of the
sufficiency side of the theorem, and hence the theorem itself. Q.E.D.
## 5 Discussion
### 5.1 Discussion of the assumptions
Monotonicity: The most restrictive assumption in this model is monotonicity,
which, as Lemma 1 shows, makes the type space uni-dimensional. It is worth
noting that monotonicity is quite common in the literature. A prevalent
version of it in the screening literature is the single crossing condition
imposed by Maskin and Riley (1984).111111Single crossing in Maskin and Riley
(1984) is in fact stronger than my monotonicity condition. In order for
monotonicity to be as strong as the single crossing condition in Maskin and
Riley (1984), it has to be that
$v(\bar{b},t^{\prime})\geq(>)v(\bar{b},t)\Rightarrow\forall b\cap
b^{\prime}=\emptyset:v(b,t^{\prime}|b^{\prime})\geq(>)v(b,t|b^{\prime})$.
Within the bundling literature most of the papers that focus on products with
non-additive values focus on versions of the product-line pricing
problem–which can be thought of as a special case of the bundling problem–and
each impose a form of monotonicity (e.g., Anderson and Dana Jr (2009);
Deneckere and Preston McAfee (1996); Long (1984)). This usually comes in the
form of assumed increasing difference of values in the (unidimensional) type
and the product quality level. Also the seminal paper by Mussa and Rosen
(1978) on product line pricing assumes the valuation by each type of each
quality level is proportional to both type and quality, which implies
monotonicity.121212It is worth noting that unlike other papers on product line
pricing such as Anderson and Dana Jr (2009), the Mussa and Rosen (1978) study
does not focus on whether and when bundling is optimal. Mussa and Rosen (1978)
make a series of assumptions that imply pure bundling (i.e., offering only the
highest quality version) is always sub-optimal. One can indeed show that an
appropriate translation of the Mussa and Rosen (1978) problem into the setting
of this paper will satisfy the optimal sales volume conditions for sub-
optimality of pure bundling. See appendix for more details. To my knowledge,
the only studies of bundling of products with non-additive values that do not
impose a version of monotonicity are Haghpanah and Hartline (2021) and
Armstrong (2013). That said, the ratio-monotonicity conditions in Haghpanah
and Hartline (2021) –at least in one direction– have a similar implication to
my monotonicity condition.131313To be clear, there are many papers in the
bundling literature that do not impose a version of monotonicity as a model
assumption or as part of the conditions in their theorems. To the best of my
knowledge, however, all such papers study environments with additive values,
meaning they impose $\forall b,t:v(b,t)=\Sigma_{i\in b}v(\\{i\\},t)$.
Finally, note that the monotonicity assumption does not force consumers to
rank the products the same way. That is, it does not rule out
$v(b,t)>v(b^{\prime},t)$ co-existing with
$v(b,t^{\prime})<v(b^{\prime},t^{\prime})$. It, rather, rules out
$v(b,t)>v(b,t^{\prime})$ co-existing with
$v(b^{\prime},t)<v(b^{\prime},t^{\prime})$.
Quasi-concavity: Quasi-concavity simply requires that each relevant profit
function be single-peaked. This assumption has been made in the literature
before (e.g., see assumption 5 in Maskin and Riley (1984)). Without this
assumption, one can still prove a version of Theorem 1; but that version would
be weaker and less straightforward to state. Finally, it is worth specifying
what this assumption would look like if expressed based on the model
primitives rather than profit functions. This assumption would require that
$\frac{\partial\log(v(b,t)-\Sigma_{i\in b}c_{i})}{\partial
t}-\frac{f(t)}{1-F(t)}$ cross zero only once from above for all
$b\neq\emptyset$.
Other notes: Most papers on optimal bundling of products with non-additive
values assume there are only two products whereas this paper examines
arbitrarily many. Additionally, one direction of the results in Theorem 1 does
not use the assumption that marginal costs are constant. Furthermore, to my
knowledge, with the exception of this paper and Haghpanah and Hartline (2021)
which do not make assumptions on substitution patterns, other papers on
bundling with non-additive values impose relatively strong complementarity
assumptions.141414These assumptions are mostly implicit in the form of
product-line pricing which may be thought of as a “base product” plus
complementary add-ons. As an exception, Armstrong (2013) allows for
complementarity or substitution separately, but does not allow the same
products to be complementary for some types and substitutes for others.
### 5.2 Interpretation of the Result and Relation to Literature
In the literature on bundling of products with non-additive values, a commonly
mentioned condition for (sub-)optimality of pure bundling is ratio
monotonicity. According to ratio-monotonicity, pure bundling is optimal if
$\frac{v(b,t)}{v(\bar{b},t)}$ is everywhere weakly increasing in
$v(\bar{b},t)$ for all $b$. Also pure-bundling is sub-optimal if this fraction
is everywhere strictly increasing for at least one $b$. Versions of it have
been mentioned in Anderson and Dana Jr (2009); Salant (1989); Deneckere and
Preston McAfee (1996), and, most generally, Haghpanah and Hartline
(2021).151515Haghpanah and Hartline (2021) prove their results under weaker
underlying assumptions, use a stochastic version of ratio-monotonicity (which
is weaker than the deterministic version and allows for multi-dimensional
types,) and are to my knowledge the only paper that states both sides of the
condition. Theorem 1 shows that, at the cost of having to make assumptions, 1
and 2, optimality of pure bundling could be linked to local–as opposed to
global–ratio monotonicity.
###### Proposition 1.
Suppose that assumptions 1 and 2 hold and that production costs are zero. Also
suppose each $\pi(b,p)$ is differentiable in $p$ and that the derivative is
only zero at the peak. Then for any two bundles $b$ and $b^{\prime}$ we have
$D^{*}(b^{\prime})>D^{*}(b)$ if and only if for the unique $\tilde{t}$ with
$\frac{\partial\log(v(b^{\prime},\tilde{t}))}{\partial
t}=\frac{f(\tilde{t})}{1-F(\tilde{t})}$, we have:
$\frac{\partial\log(v(b,\tilde{t}))}{\partial
t}>\frac{f(\tilde{t})}{1-F(\tilde{t})}$.
The proof follows directly from the assumptions. To see why this is related to
local ratio monotonicity, note that according to Proposition 1, for
$D^{*}(\bar{b})>D^{*}(b)$ to hold, we need the $\tilde{t}$ that satisfies
$\frac{\partial\log(v(\bar{b},\tilde{t}))}{\partial
t}=\frac{f(\tilde{t})}{1-F(\tilde{t})}$ to also satisfy
$\frac{\partial\log(v(b,\tilde{t}))}{\partial
t}>\frac{f(\tilde{t})}{1-F(\tilde{t})}$. In other words:
$\frac{\partial\log(v(\bar{b},\tilde{t}))}{\partial
t}<\frac{\partial\log(v(b,\tilde{t}))}{\partial t}$. This means that for $t$
slightly larger than $\tilde{t}$, we get:
$\frac{v(\bar{b},t)}{v(\bar{b},\tilde{t})}<\frac{v(b,t)}{v(b,\tilde{t})}$.
Rearranging, we get
$\frac{v(b,\tilde{t})}{v(\bar{b},\tilde{t})}<\frac{v(b,t)}{v(\bar{b},t)}$.
This is exactly ratio monotonicity.161616Note that this also to some extent
resembles elasticity-comparison results such as those in Armstrong (2013) and
Long (1984).
Note that, according to this result, in order to check whether pure bundling
is optimal, one does not need to calculate $D^{*}(b)$ for all $b$. It would
suffice to calculate $D^{*}(\bar{b})$ and then check ratio-monotonicity at
that local point.
Parametric Examples: One can construct simple examples in which optimal sales
volumes/local ratio monotonicity can pin down optimal bundling strategy
whereas global ratio monotonicity cannot. Suppose $n=2$, and $t$ is uniformly
distributed between 0 and 1. Assume the firm can produce these products at no
cost. By $b$ denote the bundle $\\{1\\}$. For simplicity, assume the
complementary bundle $b^{c}=\\{2\\}$ is not valued by any type: $\forall
t:v(\\{2\\},t)=0$. A common example of this is when $\\{2\\}$ is an “add on,”
which is not of value by itself but can add value once the “base product” is
present (e.g., additional memory for a smart phone). Suppose $\forall
t:v(b,t)=t+k_{1}$ where $k_{1}$ is a fixed real number. Finally, assume
$\forall t:v(\bar{b},t)=t+k_{1}+t^{k_{2}}$ where $k_{2}$ is a positive real
number. That is, each type $t$’s valuation of the add-on on top of the
original product is $t^{k_{2}}$. One can verify that this setup satisfies
monotonicity and quasi-concavity.
Now fix $k_{1}=0.2$ and allow $k_{2}$ to vary. One can verify that for
$k_{2}\leq\frac{2}{3}$, pure bundling is optimal while for $k_{2}>\frac{2}{3}$
it is optimal to mixed-bundle. Also, once $k_{2}$ passes $\frac{2}{3}$, the
optimal sales volume for the grand bundle passes $0.6$ (which is the optimal
sales volume for partial bundle $b$) from above. Additionally,
$\frac{\partial\log\big{(}v(\bar{b},t)\big{)}}{\partial
t}-\frac{f(t)}{1-F(t)}$ evaluated at $t=t^{*}(b)=0.4$ passes 0 from below once
$k_{2}$ passes $\frac{2}{3}$. Finally, one can verify that for an interval
around $k_{2}=\frac{2}{3}$, (say $k_{2}\in[0.4,0.8]$) the ratio
$\frac{v(b,t)}{v(\bar{b}),t}$ first decreases and then increases as
$v(\bar{b})$ grows. To sum up, the optimal bundling decision is informed by
optimal sales volumes/local ratio monotonicity but not by global ratio
monotonicity. The appendix provides a visualization for this parametric class
of problems.
I finish this section by noting that the relation to ratio-monotonicity
implies that the intuition provided by Haghpanah and Hartline (2021) is
relevant to my framework as well: the correlation between WTP and perceived
complementarity/substitution levels has implications for bundling decisions.
Pure bundling is optimal if higher WTP customers see a lower degree of
complementarity (or higher degree of substitution) across products, compared
to lower WTP customers.
### 5.3 Implications for Non-linear Pricing
The traditional screening/non-linear pricing problem (a la Maskin and Riley
(1984) or Mussa and Rosen (1978)) can be cast as a form of bundling problem in
which each product $i$ represents the $i$-th quality/quantity unit. The
appendix shows that a version of the sales volumes result in Theorem 1 can be
used to fully characterize what the optimal tariff should look like in a non-
linear pricing problem (observe that this goes beyond the result in Theorem 1
which, instead of fully characterizing the bundling strategy, only specifies
when pure bundling is optimal.) Leaving the details to the appendix, here I
only discuss one important implication of it through an example:
Example. Suppose types $t$ are uniformly distributed within $[0,1]$. The
monopolist sells a single product with continuous quality levels $q\in[0,2]$.
The monopolist optimally prices each quality level at $p^{*}(q)$ where
$p^{*}(0)$ is fixed at 0. Production costs are zero. Valuations are given by
$v(q,t)=q\sqrt[q]{t}$,171717This function simply captures the idea that the
higher $q$, the more concave the value is across types $t$. For instance,
$v(0,t)\equiv 0$, $v(\frac{1}{2},t)\equiv\frac{t^{2}}{2}$, $v(1,t)\equiv t$,
and $v(2,t)\equiv 2\sqrt{t}$. and each type $t$ decides which quality level to
purchase. In this setting, one can show that it is optimal to charge a flat
fee of $\frac{2}{\sqrt{3}}=1.15$ for any strictly positive level of quality.
Under this tariff, types $t<\frac{1}{3}$ purchase nothing while the rest of
the types purchase the highest quality version of the product $q=1$. To see
why this is true, one can apply our main result and check local ratio
monotonicity at $t=\frac{1}{3}$: we have
$\frac{\partial\log\big{(}v(q,t)\big{)}}{\partial t}=\frac{f(t)}{1-F(t)}$ when
evaluated at $t=\frac{1}{3}$ and $q=2$. But for $t=\frac{1}{3}$ and any $q<2$,
we have $\frac{\partial\log\big{(}v(q,t)\big{)}}{\partial
t}=\frac{3}{q}>\frac{3}{2}=\frac{f(t)}{1-F(t)}$.181818Of course for our main
result, which was developed in the bundling domain, to apply in this non-
linear pricing domain, additional propositions are needed. The appendix
provides those.
This example shows a case where the optimal non-linear tariff leads to a
“jump” across types in the quality purchased (from $q=0$ to $q=1$ at
$t=\frac{1}{3}$.) The literature tends to assume this away. Mussa and Rosen
(1978) state that “[t]he economic rationale for the conclusion that jumps in
$q(\theta)$ are not optimal is that the monopolist would not be making full
use of his power to discriminate among different types of buyers.”191919They
denote type by $\theta$ instead of $t$ in this paper; and they denote optimal
quality purchased by type $\theta$ under optimal contract by $q(\theta)$. By
applying bundling results to non-linear pricing, this paper offers a different
perspective: Changing the tariff from one that induces a continuous range of
purchased quantities/qualities in an interval $[\underline{q},\bar{q}]$ to one
that induces a jump from $\underline{q}$ to $\bar{q}$ may indeed be optimal,
as long as this discontinuity leads sufficiently many consumers to “upgrade to
$\bar{q}$” as opposed to “downgrade to $\underline{q}$.” In other words,
inducing a jump in purchase behavior in tariff design may be optimal for
similar economic reasons to those that make pure bundling of products optimal.
## 6 Conclusion
This paper studied when pure bundling is optimal for a monopolist who sells
products with non-additive values (i.e., with no restriction on
complementarity/substitution patterns.) Under monotonicity and quasi-concavity
assumptions, I showed that pure bundling is optimal if and only if the grand
bundle, once sold on its own and optimally priced, would “sell more” than any
smaller bundle. The appendix provides additional results on the relation to
non-linear tariff design, and the implications of the model for the case of
additive values.
This paper provides the first if-and-only-if characterization for optimality
of pure bundling under non-additive values; but at the expense of a
monotonicity assumption that makes the type space single-dimensional. A full
characterization result under non-additive values and multi-dimensional types
would, in my view, be the most important way to extend the results in this
paper.
## References
* (1)
* Adams and Yellen (1976) William James Adams and Janet L Yellen. 1976. Commodity bundling and the burden of monopoly. _The quarterly journal of economics_ (1976), 475–498.
* Anderson and Dana Jr (2009) Eric T Anderson and James D Dana Jr. 2009. When is price discrimination profitable? _Management Science_ 55, 6 (2009), 980–989.
* Armstrong (2013) Mark Armstrong. 2013\. A more general theory of commodity bundling. _Journal of Economic Theory_ 148, 2 (2013), 448–472.
* Armstrong (2016) Mark Armstrong. 2016\. Nonlinear pricing. _Annual Review of Economics_ 8 (2016), 583–614.
* Bergemann et al. (2021) Dirk Bergemann, Alessandro Bonatti, Andreas Haupt, and Alex Smolin. 2021. The optimality of upgrade pricing. (2021).
* Daskalakis et al. (2017) Constantinos Daskalakis, Alan Deckelbaum, and Christos Tzamos. 2017\. Strong Duality for a Multiple-Good Monopolist. _Econometrica_ 85, 3 (2017), 735–767.
* Deneckere and Preston McAfee (1996) Raymond J Deneckere and R Preston McAfee. 1996. Damaged goods. _Journal of Economics & Management Strategy_ 5, 2 (1996), 149–174.
* Fang and Norman (2006) Hanming Fang and Peter Norman. 2006. To bundle or not to bundle. _The RAND Journal of Economics_ 37, 4 (2006), 946–963.
* Haghpanah and Hartline (2021) Nima Haghpanah and Jason Hartline. 2021. When is pure bundling optimal? _The Review of Economic Studies_ 88, 3 (2021), 1127–1156.
* Long (1984) John B Long. 1984\. Comments on” Gaussian Demand and Commodity Bundling”. _The Journal of Business_ 57, 1 (1984), S235–S246.
* Manelli and Vincent (2007) Alejandro M Manelli and Daniel R Vincent. 2007. Multidimensional mechanism design: Revenue maximization and the multiple-good monopoly. _Journal of Economic theory_ 137, 1 (2007), 153–185.
* Maskin and Riley (1984) Eric Maskin and John Riley. 1984. Monopoly with incomplete information. _RAND Journal of Economics_ 15, 2 (1984), 171–196.
* McAfee et al. (1989) R Preston McAfee, John McMillan, and Michael D Whinston. 1989\. Multiproduct monopoly, commodity bundling, and correlation of values. _The Quarterly Journal of Economics_ 104, 2 (1989), 371–383.
* Menicucci et al. (2015) Domenico Menicucci, Sjaak Hurkens, and Doh-Shin Jeon. 2015\. On the optimality of pure bundling for a monopolist. _Journal of Mathematical Economics_ 60 (2015), 33–42.
* Mussa and Rosen (1978) Michael Mussa and Sherwin Rosen. 1978. Monopoly and product quality. _Journal of Economic theory_ 18, 2 (1978), 301–317.
* Palfrey (1983) Thomas R Palfrey. 1983\. Bundling decisions by a multiproduct monopolist with incomplete information. _Econometrica: Journal of the Econometric Society_ (1983), 463–483.
* Pavlov (2011) Gregory Pavlov. 2011\. Optimal mechanism for selling two goods. _The BE Journal of Theoretical Economics_ 11, 1 (2011).
* Salant (1989) Stephen W Salant. 1989\. When is inducing self-selection suboptimal for a monopolist? _The Quarterly Journal of Economics_ 104, 2 (1989), 391–397.
* Schmalensee (1984) Richard Schmalensee. 1984\. Gaussian demand and commodity bundling. _Journal of business_ (1984), S211–S230.
* Stigler (1963) George J Stigler. 1963\. United States v. Loew’s Inc.: A note on block-booking. _The Supreme Court Review_ 1963 (1963), 152–157.
* Wilson (1993) Robert B Wilson. 1993\. _Nonlinear pricing_. Oxford University Press on Demand.
* Zhou (2017) Jidong Zhou. 2017\. Competitive bundling. _Econometrica_ 85, 1 (2017), 145–172.
* Zhou (2019) Jidong Zhou. 2019\. Mixed Bundling in Oligopoly Markets. (2019).
## Online Appendix
The online appendix provides several additions to the paper. Appendix A
provides visualizations for the examples in the paper. Appendix B formally
discusses the implications of the main results of the paper for non-linear
pricing. Finally, Appendix C studies the implications of our results (as well
as those of the ratio-monotonicity results from the literature) for
environments where values are additive.
## Appendix A Visualizations of Examples from the Main Text
The two panels of Figure 1 visualize the example from Section 5.2 of the
paper. As a reminder, the example was about a parametric class of value
functions in which $v(b,t)=t+k_{1}$ and $v(\bar{b},t)=t+k_{1}+t^{k_{2}}$. Each
panel of the figure has a fixed value for $k_{1}$ and examines a range of
values for $k_{2}$. For both cases of both panels, the ratio
$\frac{v(b,t)}{v(\bar{b},t)}$ is non-monotonic in $v(\bar{b},t)$, making ratio
monotonicity results less useful in determining the optimal bundling strategy.
However, optimal sales volume results are in line with the optimal bundling
decisions. In panel (a), pure bundling is sub-optimal in case 1 and optimal in
case 2. In panel (b), pure bundling is optimal in case 1 and sub-optimal in
case 2.
(a) $k_{1}=0.2$, Case 1: $k_{2}=0.4$, Case 2: $k_{2}=0.8$
(b) $k_{1}=-0.2$, Case 1: $k_{2}=2$, Case 2: $k_{2}=1.2$
Figure 1: Family of examples in which $v(b,t)=t+k_{1}$ and
$v(\bar{b},t)=t+k_{1}+t^{k_{2}}$.
Also, Figure 2 visualizes the value function $v(q,t)=q\sqrt[q]{t}$ from
Section 5.3 on the implications of our results for nonlinear tariff design. It
plots $v(q,t)$ as a function of $t$ for different values of $q$. As can be
seen and also algebraically verified, this function is increasing in both
arguments. The key feature of this value function is that it becomes more
concave in $t$ as $q$ increases (which implies the optimal sales volume
increases in $q$.) Other value functions that share this feature should also
yield a similar result, yielding a jump in consumer purchase behavior under
the optimal tariff.
Figure 2: Value function $v(q,t)=q\sqrt[q]{t}$ as a function of $t$ for
different values of $q$
## Appendix B Implications for Non-Linear Pricing
The main focus of this paper is the problem of optimal bundling. Nevertheless,
the main result also has implications for the non-linear pricing problem
studied, among others, by Maskin and Riley (1984) and Mussa and Rosen (1978).
Translated to the context of non-linear pricing, the analysis in this paper
will take the following form: A monopolist selling a single product faces the
problem of determining the optimal price schedule $T^{*}(q)$ for different
quantities $q\in\\{0,1,...,n\\}$. The monopolist’s objective is to maximize
total profit $\pi(T^{*})$. Valuations are denoted $v(q,t)$ with $v(0,t)=0$ for
all $t$. Conditional on a prior endowment of $q^{\prime}$, we denote the
valuations by $v(q,t|q^{\prime})$. Similarly, we use notation $\pi_{q}(p)$ to
denote the firm’s profit when it sells only $q$-size batches of the product,
pricing each batch at $p\in\mathbb{R}$. Also $\pi_{q}(p|q^{\prime})$ denotes
the same thing under the condition that all consumers have been pre-endowed
with $q^{\prime}\leq n-q$ unites of the product. $D^{*}(q)$, and
$D^{*}(q|q^{\prime})$ are defined in the expected way.
Proposition 2 shows that in the above setting, one can fully characterize the
optimal nonlinear tariff.
###### Proposition 2.
Consider the non-linear pricing setup described above. Assume the following:
(i) $v(q,t)$ is increasing in both arguments (and strictly so whenever
positive), and $\forall t^{\prime}>t,q>0$ we have
$v(q,t^{\prime})-v(q-1,t^{\prime})>v(q,t)-v(q-1,t)$. (ii) For all
$q+q^{\prime}\leq n$, the profit function $\pi_{q}(p|q^{\prime})$ is strictly
quasi-concave in $p$ over the range of $p$ that generates strictly positive
demand. (iii) $v(q,t)$ is continuous in $t$ except possibly for finitely many
points. (iv) For any $q^{\prime}<n$, the set
$\arg\max_{q\in\\{1,...,n-q^{\prime}\\}}D^{*}(q|q^{\prime})$ is a singleton
denoted $q^{*}(q^{\prime})$.202020That is, $q^{*}({q^{\prime}})$ is the batch-
size that, if sold and optimally priced in a market where all consumers have
already been endowed with ${q^{\prime}}$ unites, would generate the highest
demand. Similar to the main result, if instead of quasi-concavity we have
concavity, one need not impose the assumption that the argmax is unique. Then
the optimal price schedule will involve $m\leq n$ distinct quantities
$(q_{1}^{*},...,q_{m}^{*})$ such that $q_{1}^{*}=q^{*}(0)$ and $\forall
i\in\\{2,...,m\\}:q_{i}^{*}=q_{i-1}^{*}+q^{*}(q_{i-1}^{*})$. Precisely:
$\forall
q\in\\{q_{i-1}^{*}+1,...,q_{i}^{*}\\}:T^{*}(q)=p^{*}(q_{i}^{*}-q_{i-1}^{*}|q_{i-1}^{*})$
This proposition fully characterizes the optimal tariff $T^{*}$. It starts by
stating that the smallest quantity that any consumer can buy is
$q_{1}^{*}=q^{*}(0)\equiv\arg\max_{q\in\\{1,...,n\\}}D^{*}(q)$, i.e., the
batch size that would sell the most if sold alone and priced optimally. In
other words, at least $q_{1}^{*}$ units of the products are always bundled
together. The price of this $q^{*}_{1}$-bundle is the optimal price that the
monopolist would set if it were to only sell this bundle. From this point on,
the proposition takes a recursive structure and states that the second
distinct quantity sold of the product is
$q_{2}^{*}=q_{1}^{*}+\arg\max_{q\in\\{1,...,n-q_{1}^{*}\\}}D^{*}(q|q_{1}^{*})$
and so on. Though Proposition 2 in some ways resembles the demand profile
approach of Wilson (1993), it has fundamental differences. Before turning to
the proof of this results, I make a few notes.
First, as mentioned shortly before, Proposition 2 takes a step beyond the main
result in Theorem 1: it fully characterizes what the optimal tariff looks like
as opposed to only characterizing the conditions under which the optimal
tariff will involve a flat fee for selling all $n$ units of the
product.212121That condition would be
$n=\arg\max_{q\in\\{1,...,n\\}}D^{*}(q)$, mirroring the condition in Theorem
1. This property of Proposition 2 should naturally raise the question that
whether Theorem 1 can also be strengthened to fully characterize the optimal
bundling strategy as opposed to only characterizing when pure bundling is
optimal. Unfortunately the answer turns out to be no; a bundling strategy
constructed in a similar way to how the optimal tariff in Proposition 2 is
constructed may or may not be optimal.222222Counterexamples are available upon
request. Also, it is worth noting that the Maskin-Riley type monotonicity
condition used in Proposition 2 is stronger than that used in Theorem 1.
Nevertheless, this is not the reason why the optimal bundling strategy cannot
be fully characterized.
Second, there is a difference between Proposition 2 and the usual form in
which the nonlinear pricing problem is studied in the literature. The origin
of Proposition 2 being in bundling makes the domain of quantities $q$ by
construction discrete and bounded, whereas the literature (e.g., Maskin and
Riley (1984)) examines a continuous and possibly unbounded environment. That
said, I do not see this feature of Proposition 2 as too restrictive, given
that one an always examine the limit case as $n\rightarrow\infty$.
Fourth, $q$ in this proposition need not be interpreted as quantity. One can
also think of $q$ as quality in a similar fashion to Mussa and Rosen (1978).
In that case, the result holds if the cost function is non-linear in quality,
as long as it is still linear in the quantity of the consumers that purchase
each quality level.
Next, I turn to proving the proposition.
Proof of Proposition 2. First, let us introduce some notations. In a similar
spirit to the consumer choice notation $\beta(t|B,p)$ in the bundling problem,
define the following consumer choice function:
$\beta(t|T)\equiv\arg\max_{q}v(t,q)-T(q)$ (10)
I use the same notation $\beta$ as I did in the formulation of the bundling
problem in order for the parallels between the two settings to be clearer.
However, there are some differences. Most notably, the output of the $\beta$
function here is just an integer number, not a bundle. Suppose consumers break
ties in favor of higher quantities. I now proceed to state the following
lemma.
###### Lemma 8.
For any price schedule $T$ and any two types $t,t^{\prime}$ with
$t^{\prime}>t$ we have: $\beta(t^{\prime}|T)\geq\beta(t|T)$.
This lemma, whose proof is rather straightforward and is left to the reader,
simply says that the consumption quantity is weakly monotonic in type.
Also, in this proof, I will assume that the optimal schedule $T^{*}(q)$ is
weakly monotonic in $q$. This assumption is without loss, given that one can
show that any non-monotonic $T$ can be modified in a way that (i) makes it
weakly monotonic and (ii) delivers the same amount of profit to the
monopolist. Given this weak monotonicity, $T^{*}$ has to take the following
form for a strictly increasing sequence $q_{0}=0,q_{1},...,q_{m}=n$ and a
weakly increasing sequence $T_{1},...,T_{m}$:
$\forall q\in\\{q_{i-1}+1,...,q_{i}\\}:T^{*}(q)=T_{i}$ (11)
The following lemma will be useful for the proof:
###### Lemma 9.
$T_{1}=p^{*}(q_{1}|0)$. That is, the lowest type $t$ purchasing $q_{1}$ under
the optimal contract will satisfy $1-F(t)=D^{*}(q_{1})$.
Proof of Lemma 9. Suppose this lemma’s claim is not true: $T_{1}\neq
p^{*}(q_{1}|0)$. Then consider a scenario in which only $q_{1}$ is being sold
by the seller at the price of $T_{1}$. Given that $T_{1}$ is not the optimal
price, there is a small but nonzero $\epsilon$ such that if the seller prices
$q_{1}$ at $T_{1}+\epsilon$ instead of $T_{1}$, the seller will strictly
improve its profit when selling only $q_{1}$.
Next, I move from the scenario of selling only a batch of $q_{1}$ units back
to the full tariff design problem. I use the above deviation to construct a
similar deviation from the full schedule $T^{*}$ and show that the seller can
strictly improve its profit. Construct price schedule $T\equiv T^{*}+\epsilon$
for any positive $q$. I now claim that $T$ is strictly more profitable to the
seller than is $T^{*}$.
To see why this claim is true, note that for any $q\in\\{q_{2},...,q_{m}\\}$
and any type $t$ such that $\beta(t|T^{*})=q$, we have $\beta(t|T)=q$. This is
because (i) for those types it is the IC constraint (and not the IR) that is
binding; and (ii) the construction of $T$ from $T^{*}$ preserves all the IC
constraints. Therefore a move from $T^{*}$ to $T$ will lead to the exact same
revenue change that a move from $T_{1}$ to $T_{1}+\epsilon$ does in the
sceniario of selling only $q_{1}$: the exact same new types are added to (or
removed from) the set that purchases $q_{1}$, and the same change (i.e.,
$\epsilon$) has been made to the amount made off of each type that buys. In
addition to the change in the revenue, a move from $T^{*}$ to $T$ also leads
to the exact same change in total costs as would a change from $T_{1}$ to
$T_{1}+\epsilon$. This is because in both cases, the only change made in the
production is the number of $q_{1}$-size batches (or, in the Mussa and Rosen
(1978) interpretation, the number of $q_{1}$-quality products). Given that we
have assumed the cost function to be linear in this change, the changes in
total cost is the same between the two scenarios.
As a result, a move from $T^{*}$ to $T$ will lead to the exact same change in
the total profit as would a move from $T_{1}$ to $T_{1}+\epsilon$. Therefore,
a change from $T^{*}$ to the new schedule $T$ strictly profitable, finishing
the proof of the lemma. Q.E.D.
Next, I introduce a lemma which will be the building block of the proof of
this proposition.
###### Lemma 10.
In the presentation of $T^{*}$ in equation 11, it has to be that
$q_{1}=q_{1}^{*}$ where $q_{1}^{*}=q^{*}(0)$ as defined in the statement of
Proposition 2.
As a reminder, $q^{*}(0)=\arg\max_{q=1,...,n}D^{*}(q)$. Thus, the lemma simply
says that the smallest quantity that the optimal schedule $T^{*}$ offers to
consumers is the quantity that, if sold alone, would sell the highest volume.
Proof of Lemma 10. Suppose $q_{1}\neq q^{*}_{1}$. Then one can construct a
deviation from $T^{*}$ that would strictly improve the seller’s profit. This
will suffice to finish the proof of the lemma.
First, for convenience, assume that even though $q_{1}\neq q^{*}_{1}$, there
is some $k>1$ such that $q_{k}=q^{*}_{1}$. In other words, even though
$q^{*}_{1}$ is not the smallest package on the schedule, it is nonetheless
somewhere on the schedule. Later I will show this assumption is not necessary.
But for now, it will make the steps of the proof more straightforward.
For each $i\in\\{1,...,m\\}$ denote by $t_{i}$ the lowest type that buys
$q_{i}$ units under $T^{*}$. That is:
$t_{i}=\min\\{t:\,\beta(t|T^{*})=q_{i}\\}$. From Lemma 8 we know larger types
buy weakly more units of the product. That is: $t_{1}<t_{2}<...<t_{m}$.
From Lemma 9 we know that $1-F(t_{1})=D^{*}(q_{1})$. Given $t_{k-1}\geq
t_{1}$, we get $1-F(t_{k-1})\leq D^{*}(q_{1})$. But we also know, by
assumption, that $D^{*}(q^{*}_{1})>D^{*}(q_{1})$. Therefore:
$1-F(t_{k-1})<D^{*}(q^{*}_{1})\equiv D^{*}(q_{k})$ (12)
In addition, again by the definition of $q_{k}=q_{1}^{*}$, we know that
$D^{*}(q_{k-1})<D^{*}(q_{k})$. This inequality, along with a similar quasi-
concavity argument to that used in the proof of the main result yields:
$D^{*}(q_{k})\leq D^{*}(q_{k}-q_{k-1}|q_{k-1})$ (13)
Together, inequalities 12 and 13 yield:
$1-F(t_{k-1})<D^{*}(q_{k}-q_{k-1}|q_{k-1})$ (14)
Obviously, by $t_{k}>t_{k-1}$, we also know:
$1-F(t_{k})<D^{*}(q_{k}-q_{k-1}|q_{k-1})$ (15)
Note that equation 15 implies that if all consumers have already been endowed
with $q_{k-1}$ units of the product and the monopolist is selling only batches
of size $q_{k}-q_{k-1}$ and pricing them so that types $t_{k}$ and above
purchase, then the monopolist, by quasi-concavity, will strictly profit from a
small price reduction $\rho$.
Next, I move from the scenario of selling only $q_{k}-q_{k-1}$ packages under
a pre-endowment of $q_{k-1}$ to the main scenario of designing the full
schedule $T^{*}$. I argue that the monopolist will enjoy the same profit
increase as the one described in the previous paragraph if it modifies $T^{*}$
by reducing all $T_{j}$ for $j\geq k$ by $\rho$. The argument is similar to
that in the proof of Lemma 9. This modification does not alter the behavior of
any type $t$ that purchases $q_{k+1}$ or more units due to the fact that it
preserves all of the binding IC constraints for those types. As a result, the
impact of this change on the firm profit is (i) a cut in margin by $\rho$
across all consumers, combined by the change in the behavior of those who used
to purchase less than $q_{k}$ units under $T^{*}$ but will now switch to
$q_{k}$.232323None of these types would switch to buying more than $q_{k}$
units, due to the single-crossing condition. Note that if the price reduction
$\rho$ is small enough, these types will only consist of those who under
$T^{*}$ purchase $q_{k-1}$. From equation 14, we know there is a non-zero mass
of such types. Therefore, the effect of this price change parallels that of
the price change described in the previous paragraph, making it strictly
profitable.
But the above argument contradicts the optimality of $T^{*}$. Therefore, the
contrapositive assumption must have been incorrect. That is: it has to be that
$q_{1}=q^{*}_{1}$.
With the above argument, the proof is complete for the case where there is
some $k>1$ with $q_{k}=q^{*}_{1}$. That is, when $q_{1}^{*}$ is “on the price
schedule.” Thus, it remains to show that the proof also works when $\nexists
k:q_{k}=q^{*}_{1}$. Suppose this is the case, and take $k$ to be the smallest
index with $q_{k}\geq q^{*}_{1}$. Like before, denote by $t_{i}$ the smallest
type that purchases $q_{i}$ under $T^{*}$
I now construct schedule $T^{**}$ in the following way:
* •
For any $q\leq q_{k-1}$ or $q>q^{*}_{1}$, set $T^{**}(q)=T^{*}(q)$
* •
For all $q_{k-1}<q\leq q^{*}_{1}$, set
$T^{**}(q)=T^{*}(q_{k})+v(q_{1}^{*},t_{k})-v(q_{k},t_{k})$
Next, I take two steps. First, I show that $T^{**}$ delivers the same profit
to the monopolist as does $T^{*}$. Then, I will construct a deviation from
$T^{**}$ that yields a strict profit improvement.
As for the first step, note that by construction, $T^{**}(q^{*}_{1})$ is
designed to make the type $t_{k}$ consumer indifferent between purchasing
$q_{k}$ units and $q^{*}_{1}$ units. But we know, by construction of $t_{k}$,
that this type is also indifferent between buying $q_{k}$ units and buying
$q_{k-1}$ units. This makes this type indifferent among all three quantities
$q_{k-1}<q^{*}_{1}<q_{k}$ under the tariff $T^{**}$. But by the monotonicity
condition (i.e., single crossing,) any type $t<t_{k}$ will strictly prefer
$q_{k-1}$ to $q^{*}_{1}$ and any type $t>t_{k}$ will strictly prefer $q_{k}$
over $q^{*}_{1}$. In other words, this “addition of $q^{*}_{1}$ to the
schedule” will not change any consumer’s purchase behavior: $\forall
t:\beta(t|T^{*})=\beta(t|T^{**})$. Thus the two tariffs deliver the same
profit to the monopolist.
But now the structure of $T^{**}$ allows us to construct the same profit
enhancing modification that we applied to $T^{*}$ when we assumed it did have
$q^{*}_{1}$ on the schedule. All of the steps are the same. This finishes the
proof of the lemma. Q.E.D.
The rest of the proof is straightforward and involves recursive use of Lemma
10. First note that by quasi-concavity, and by $\forall
q\in\\{1,...,n-q^{*}_{1}\\}:D^{*}(q+q_{1}^{*})<D^{*}(q_{1}^{*})$, we have:
$\forall q\in\\{1,...,n-q^{*}_{1}\\}:D^{*}(q|q_{1}^{*})<D^{*}(q_{1}^{*})$
Thus, all of the optimal strategies will sell only to types $t_{1}$ and above.
This means that in order to construct “the rest of the optimal schedule”
conditional on having set $T^{*}(q)$ for all $q\leq q^{*}_{1}$ equal to
$p^{*}(q|0)$, one can just focus attention on designing the optimal schedule
for selling $n-q^{*}_{1}$ units when all consumers have been endowed with
$q^{*}_{1}$ units already. This means we are facing a version of the same
problem. It is straightforward to check that all of the conditions of the main
problem are satisfied for this “sub-problem.” Thus, one can apply Lemma 10
again and find that $q_{2}$ in the optimal schedule should be equal to
$q^{*}_{2}$ which was defined as $q^{*}_{1}+q^{*}(q^{*}_{1})$. Repeating this
procedure will fully characterize the optimal tariff and the outcome matches
what the statement of the proposition predicted. Q.E.D.
Note that Proposition 2 can now be used to characterize the optimal contract
in the non-linear pricing example given in the main text (i.e., the one in
Section 5.3 with $v(q,t)\equiv q\sqrt[q]{t}$.) That example could be cast as a
limit case of Proposition 2 as $n\rightarrow\infty$. It can be seen that the
quality level that would generate the highest sales volume would be $q=2$. To
check this, it would be sufficient to verify that ratio monotonicity holds at
$t=\frac{1}{3}$ which is the lowest type that would purchase if only quality
$q=2$ is offered and optimally priced.
## Appendix C A Brief Analysis of Environments with Additive Values
In this section, I explore the implications of Theorem 1 for environments with
additive values (i.e., environments in which $\forall b,t:v(b,t)=\Sigma_{i\in
b}v(\\{i\\},t)$), and compare that to implications of ratio monotonicity
conditions that are used in the literature.
Unfortunately, both ratio monotonicity conditions and conditions of Theorem 1
are of limited use when reduced to environments with additive values. This
point is especially pronounced about the former. The results below, further
clarify this matter.
###### Proposition 3.
Suppose values are additive. Do not impose assumptions 1-2 but instead assume
ratio monotonicity:
$\forall t,t^{\prime}:v(\bar{b},t^{\prime})\geq
v(\bar{b},t)\Rightarrow\frac{v(b,t^{\prime})}{v(\bar{b},t^{\prime})}\geq(\leq)\frac{v(b,t)}{v(\bar{b},t)}$
Under both of these weak forms of ratio-monotonicity (i.e., $\geq$ or $\leq$,)
all value functions are proportional. That is, for all $b,t,t^{\prime}$ we
have:
$\frac{v(b,t)}{v(\bar{b},t)}=\frac{v(b,t^{\prime})}{v(\bar{b},t^{\prime})}$.
As a corollary, strict ratio monotonicity under additive values is infeasible.
Proof. Define $\alpha(i,t)=\frac{v(\\{i\\},t)}{v(\bar{b},t)}$. By ratio
monotonicity, we have $\forall i:\alpha(i,t^{\prime})\geq\alpha(i,t)$. But we
also know by additivity of values that
$\Sigma_{i=1,...n}\alpha(i,t^{\prime})=\Sigma_{i=1,...n}\alpha(i,t)=1$. As a
result, it has to be that $\forall i:\alpha(i,t^{\prime})=\alpha(i,t)$. In
other words:
$\forall
i:\frac{v(\\{i\\},t^{\prime})}{v(\bar{b},t^{\prime})}=\frac{v(\\{i\\},t)}{v(\bar{b},t)}$
By additivity of values, we can say the same for all $b$:
$\forall
b:\frac{v(b,t^{\prime})}{v(\bar{b},t^{\prime})}=\frac{v(b,t)}{v(\bar{b},t)}$
which finishes the proof. The proof for the other direction of ratio
monotonicity is similar. Q.E.D.
In other words, all products and bundles will have proportional demand curves
and the exact same optimal quantity sold. As a result, the monopolist is
indifferent among all possible bundling strategies.242424Note that the above
results assumed a “deterministic” version of ratio monotonicity. I add
(without proving) that if we consider a “stochastic” version of ratio-
monotonicity (a la Haghpanah and Hartline (2021),) we can show that the two
random variables $v(\bar{b},t)$ and $\frac{v(b,t)}{v(\bar{b},t)}$ would have
to be independent of each other under weak ratio-monotonicity in either
direction. Similarly, a strict form of ratio monotonicity would be infeasible.
This implies that pure bundling is always optimal.
Next, I turn to the implications of this paper’s framework for additive
values.
###### Proposition 4.
Suppose values are additive and assumptions 1-2 hold. Then pure bundling is
optimal if and only if $\forall b:D^{*}(b)=D^{*}(b^{C})$.
Proof. If $\forall b:D^{*}(b)=D^{*}(b^{C})$, then the monopolist will be
indifferent among all possible bundling strategies. If $\exists
b:D^{*}(b)>D^{*}(b^{C})$, then one can use the fact that linearity implies
$D^{*}(b^{C}|b)=D^{*}(b^{C})$, in order to show that
$D^{*}(\bar{b})<D^{*}(b)$. Thus, pure bundling is not optimal. Q.E.D.
The fact that under assumptions 1-2 pure bundling is only optimal in such
“measure-zero” cases should not be surprising. Additive values, once combined
with monotonicity, will closely resemble s “positive-correlation” case in the
Adams and Yellen (1976) context. We know that under positive correlation mixed
bundling is preferable to pure bundling.
Overall, the tools originally developed to analyze bundling under non-additive
values (including this paper) seem to be less powerful when restricted to
environments with additive values.
|
# A rotational and vibrational investigation of phenylpropiolonitrile
(C6H5C3N)
Zachary Buchanan Kin Long Kelvin Lee Olivia Chitarra Michael C. McCarthy
Olivier Pirali Marie-Aline Martin-Drumel Université Paris-Saclay, CNRS,
Institut des Sciences Moléculaires d’Orsay, 91405 Orsay, France Department of
Chemistry, The University of California Davis, Davis, CA, USA Center for
Astrophysics $|$ Harvard & Smithsonian, Cambridge, Massachusetts 02138, United
States SOLEIL Synchrotron, AILES beamline, l’Orme des Merisiers, Saint-Aubin,
91190 Gif-sur-Yvette, France
###### Abstract
The evidence for benzonitrile (C6H5CN) in the starless cloud core TMC–1 makes
high-resolution studies of other aromatic nitriles and their ring-chain
derivatives especially timely. One such species is phenylpropiolonitrile
(3-phenyl-2-propynenitrile, C6H5C3N), whose spectroscopic characterization is
reported here for the first time. The low resolution (0.5 cm-1) vibrational
spectrum of C6H5C3N has been recorded at far- and mid-infrared wavelengths
(50–3500 cm-1) using a Fourier Transform interferometer, allowing for the
assignment of band centers of 14 fundamental vibrational bands. The pure
rotational spectrum of the species has been investigated using a chirped-pulse
Fourier transform microwave (FTMW) spectrometer (6–18 GHz), a cavity enhanced
FTMW instrument (6–20 GHz), and a millimeter-wave one (75–100 GHz, 140–214
GHz). Through the assignment of more than 6200 lines, accurate ground state
spectroscopic constants (rotational, centrifugal distortion up to octics, and
nuclear quadrupole hyperfine constants) have been derived from our
measurements, with a plausible prediction of the weaker bands through
calculations. Interstellar searches for this highly polar species can now be
undertaken with confidence since the astronomically most interesting radio
lines have either been measured or can be calculated to very high accuracy
below 300 GHz.
###### keywords:
pure rotation , vibration , astrophysical species , PAH derivative ,
phenylpropiolonitrile
††journal: Journal of Molecular Spectroscopy
## 1 Introduction
The ubiquity of aromatic molecules is closely-correlated to their stability
and lack of reactivity, with functionalized aromatics serving as a common
motif in biological chemistry. Polycyclic aromatic hydrocarbons are a
prominent class of aromatics; they are well-known constituents in the outflows
of certain evolved carbon stars and common byproducts in incomplete combustion
processes [1, 2, 3]. From a purely spectroscopic viewpoint, a number of simple
derivatives of benzene —the prototypical aromatic ring C6H6— have either not
been characterized at all or at insufficient resolution to undertake an
astronomical search in the coldest most quiescent molecular clouds in space.
The recent discovery of benzonitrile —the simplest aromatic nitrile (C6H5CN or
PhCN hereafter)— using radio observations towards Taurus Molecular Cloud
(TMC-1) [4] has reignited the interest in nitrogen-containing aromatics
generally and CN-functionalized aromatics specifically [5]. The subsequent
identification of cyanocyclopentadiene, C5H5CN, in the same cloud [6] has only
intensified this interest. Nitriles are also known to be important
constituents in the chemistry of Titan’s atmosphere [7], and they are very
prominent in the interstellar medium (ISM), accounting for roughly 20% of the
220 or so molecules (47) detected in the ISM to date, including cyanopolyynes
as long as HC11N [8].
The presence of a nitrile group often imparts a molecule with a large
permanent dipole moment and an intense rotational spectrum. In the case of
PhCN, replacing a single H atom in benzene with a CN group transforms an
otherwise highly symmetric ring into a highly polar species ($\mu_{a}=4.5$ D;
[9]) thereby greatly aiding its detection both in the laboratory and in space.
Whilst the rotational spectrum of PhCN has been known for more than half-a-
century [10], its interstellar detection was greatly aided by measurements at
very high accuracy ($i.e.$, at a resolving power $f/\Delta f>10^{6}$) at
centimeter wavelengths [4, 9]. In light of this finding, new —or in many cases
improved— high-resolution studies of molecules closely related in structure
and composition are worth pursuing.
In this work, we report a combined pure rotational and vibrational
investigation of a derivative of PhCN, phenylpropiolonitrile
(3-phenyl-2-propynenitrile, C6H5C3N, abbreviated as PhC3N in the following)
where the nitrile group is replaced by a longer chain variant (C3N) (Fig. 1).
This species has previously been identified as a possible product from the
reaction between the cyano radical and phenylacetylene (C6H5CCH) [11]. To the
best of our knowledge, however, spectroscopic investigations have been limited
to experimental and theoretical vibronic spectroscopy [12, 13] while high-
resolution, rotationally-resolved studies are apparently lacking. In light of
the large permanent electric dipole moment calculated here (5.9 D), laboratory
measurement of rotational frequencies would allow astronomical searches for
this ring-chain to be undertaken with little or no ambiguity. If found in
space, the abundance of PhC3N would provide a key test for models of aromatic
chemistry which are poorly constrained at present. The infrared spectrum of
PhC3N is also of interest as a point of comparison with other benzene
derivatives whose vibrational spectra are often plagued by a myriad of
perturbations and resonances, notably Fermi and Darling-Dennison in the C-H
and C#C stretching regions [14, 15, 16], in addition to Fermi and Coriolis
interactions for low frequency (${\sim}150$ cm-1), large amplitude modes that
are prominent at room temperature [17, 18, 19]—even under astronomical
conditions.
Figure 1: Molecular structure of PhC3N; $a$ and $b$ principal inertial axes
are indicated (in red and green, respectively); the $c$-axis extends out of
the molecular plane and is not shown.
## 2 Experimental and computational methods
### 2.1 Quantum chemical calculations
Calculations were performed using the Gaussian’16 suite of electronic
structure programs [20]. The goal of these calculations was to provide
accurate estimates of spectroscopic parameters, including rotational
constants, dipole moment, rotation-vibration corrections, and fundamental
vibrational frequencies using second-order vibrational perturbation theory
(VPT2). Geometry optimizations of PhC3N and four of its isomers—namely the
isocyanide analog PhCCNC and three variants where CN is substituted for a H
atom on the ring of phenyl acetylene to yield ortho-, meta-, and para-
cyanoethynylbenzene (CEB)—were carried out at the $\omega$B97X-D/cc-pVQZ level
on an ultrafine integration grid, in which optimized structures are those in
which convergence to better than $10^{-5}$ of the root-mean-squared (RMS)
value of the gradient has been achieved. For PhC3N, we also performed
(an)harmonic frequency analysis, obtaining both the harmonic and anharmonic
vibrational frequencies and intensities at the same level of theory; with the
latter, the rotation-vibration interaction constants $\alpha$ were computed.
Cartesian coordinates of the optimized equilibrium structures can be found in
Tables S1-S5 in the Supporting Information.
In addition to spectroscopic parameters, we have also performed rudimentary
thermochemical calculations on the relative energies of PhC3N and its isomers
using the G3//B3LYP composite method [21], which has been shown to provide
near-chemical accuracy at excellent computational cost [22]. Given that semi-
empirical methods typically perform best on closed-shell molecules with
relatively simpler electronic structure, as the ones studied here, we believe
these calculations provide a quantitative determination of relative
stabilities, accurate to $\pm 4$ kJ/mol.
### 2.2 Fourier-transform infrared measurements
The gas-phase vibrational spectrum of PhC3N was recorded in the far-infrared
(far-IR) and mid-infrared (mid-IR) using the Bruker IFS 125 FT interferometer
located at the AILES beamline of the SOLEIL synchrotron facility (no
synchrotron radiation was used in the present study) [23]. For the far-IR
measurements, the spectrometer was equipped with a $6$ µm mylar-silicon
composite beamsplitter and a liquid helium-cooled silicon bolometer. A KBr
beamsplitter and a sensitive HgCdTe detector, equipped with a cryogenically
cooled entrance iris and optical filters [24], were used in the mid-IR region.
Vapor of PhC3N was injected in a White-type multipass cell aligned to yield a
150 m optical path length [25] and isolated from the interferometer by 50 µm-
thick polypropylene windows in the far-IR and wedged ZnSe windows in the mid-
IR range. In both spectral regions, the interferometer was continuously
evacuated to $10^{-5}$ mbar to minimize absorption from residual water.
Spectra were recorded at a resolution of 0.5 cm-1 using a globar light source
and an entrance iris of 4 mm, and consist of 100 and 500 co-added
interferograms for the far-IR and mid-IR regions, respectively. Both spectra
were recorded at a sample pressure of 5 µbar. It is worth noting that the
rotational structure within the vibrational bands could not be resolved even
at the highest resolution of the spectrometer (0.001 cm-1).
### 2.3 Chirped-pulse Fourier-transform microwave measurements
Microwave measurements were performed using a chirped-pulse Fourier transform
microwave (CP-FTMW) spectrometer located at the Center for Astrophysics [26]
which operates between 8 and 18 GHz. About 0.3 g of solid PhC3N was introduced
into a reservoir located behind the pulsed nozzle and Ne carrier gas (at a
flow of $\sim$20 sccm at standard temperature and pressure) passed through the
reservoir; the resulting gas mixture was then injected into the vacuum chamber
by operating the pulsed valve at a very low repetition rate (5 Hz). Because
the vapor pressure of PhC3N is relatively high at room temperature (several
mbar), it was not necessary to heat the reservoir to observe rotational lines
in the CP-FTMW spectrum with good signal-to-noise ratios.
Approximately 17,000 nozzle pulses, each probed by 10 microwave chirps, were
acquired. Additional details on the experimental set-up are provided in Ref.
[26]. The resulting spectrum, with electronics artifacts and well-known
contaminant lines ($e.g.$, acetone) removed, is displayed in Fig. 2.
Transitions with $5\leq J^{\prime\prime}\leq 15$ and $K_{a}^{\prime\prime}\leq
4$ are visible.
Figure 2: Experimental CP-FTMW spectrum (in black) of PhC3N in comparison with
a simulation at $T_{\mathrm{rot}}=1$ K (in purple) using the calculated set of
rotational constants; the simulation is plotted here with negative intensity
to more easily compare the two spectra. Left panel: the full spectrum, noting
that a few lines originating from impurities are visible around 10 GHz; Right
panel: expanded view around the qR(7) transitions, as indicated by the gray
rectangle on the left panel. The simulation was performed using PGOPHER [27].
### 2.4 Cavity-based Fourier-transform microwave measurements
In parallel to the present investigation, pure rotational transitions of PhC3N
were also identified by several of the co-authors of this paper while
analyzing the discharge products of a benzene/nitrogen (N2) mixture [28] .
Using a cavity-enhanced FTMW (CE-FTMW) spectrometer operating between 6 and 40
GHz, transitions of PhC3N were measured at roughly ten times higher spectral
resolution than can be achieved with the CP-FTMW instrument. In the following,
we only briefly summarize the experimental conditions relevant to this work;
further details about the CE-FTMW instrument and the discharge mixture
experiments are presented in Refs. [26] and [28], respectively.
PhC3N was synthesized by subjecting a mixture of C6H6 and N2 diluted heavily
in Ne to a high-voltage discharge (800 V); typical flow rates were 14, 12, and
20 sccm respectively. Compared to the CP-FTMW measurements where PhC3N was the
precursor, the discharge experiment results in a slightly higher rotational
temperature, $i.e.$, $\sim 10$ K and produces rotational lines with
demonstratively lower signal-to-noise ratio. Nevertheless, at the highest
resolving power ($\sim 10^{7}$) the nitrogen nuclear-quadrupole structure for
several low-$K_{a}$ transitions ($K_{a}^{\prime\prime}=0-2$) in the 6–20 GHz
region was partially resolved.
### 2.5 Absorption millimeter-wave measurements
The room temperature spectrum of PhC3N was recorded at ISMO with an absorption
spectrometer in which a frequency multiplication scheme is used to generate
millimeter-wave radiation [29]. By combining the output of a radiofrequency
synthesizer (Rohde & Schwarz) operating in the 2–20 GHz region with one of two
amplifier / multiplier chains, it is possible to produce broadly tunable
radiation with modest power (of a few mW) throughout the millimeter band; a
Radiometer Physics GmbH (RPG) SMZ110 for 75–97 GHz, and a Virginia Diodes Inc.
(VDI) for 141–214 GHz.
The millimeter-wave radiation was collimated using a 10 mm focal length Teflon
lens into a 2 m long Pyrex absorption cell and further focused onto a Schottky
diode detector from VDI, using a second identical lens. The input radiation
was frequency modulated at a frequency of 48.2 kHz and a commercial lock-in
amplifier (Ametek) demodulated the signal at the second harmonic. The spectrum
was recorded using 50 kHz frequency steps, a time constant of 50/100 ms, and
an FM deviation of 200/250 kHz (where the two values refer the lower/higher
frequency measurements, respectively).
A flow of PhC3N, not exceeding a pressure of 2 µbar, was pumped through the
cell and evacuated by a turbomolecular pump. Above this pressure, significant
pressure broadening of the rotational lines was observed.
## 3 Results and discussion
### 3.1 Spectroscopic considerations
PhC3N belongs to the $C_{2\mathrm{v}}$ symmetry group and possesses a
$\tilde{\mathrm{X}}\,^{1}\mathrm{A}_{1}$ electronic ground state with 39 modes
of vibration following the irreducible representation $\Gamma=14\hskip
2.84526pt\mathrm{A}_{1}\oplus 3\hskip 2.84526pt\mathrm{A}_{2}\oplus 9\hskip
2.84526pt\mathrm{B}_{1}\oplus 13\hskip 2.84526pt\mathrm{B}_{2}$. All
vibrational modes except those of $\mathrm{A}_{2}$ symmetry are IR active;
$\mathrm{A}_{1}$ and $\mathrm{B}_{2}$ modes correspond to in-plane vibrations
(respectively $a$\- and $b$-type bands) while $\mathrm{B}_{1}$ modes are out-
of-plane vibrations ($c$-type bands). PhC3N is a prolate asymmetric top
molecule with a permanent dipole moment of 5.9 Debye along the $a$ axis (axis
of the C3N bonds, Fig. 1), according to our calculations. PhC3N also has
nitrogen nuclear quadrupole hyperfine structure, but this splitting is only
partially resolved at low frequency in the CE-FTMW measurements. Given that
the molecule is $C_{2\mathrm{v}}$ symmetry, we also include statistical
weights for equivalent exchangeable nuclei for correct transition intensities.
Identical to PhCN, there are two sets of equivalent hydrogen atoms
($I_{1}=I_{2}=\frac{1}{2}$) that give rise to Fermi-Dirac statistics for
symmetric (even $K$) and antisymmetric (odd $K$) rotational states with a
ratio of $10:6$ [4].
### 3.2 Vibrational spectroscopy
Figure 3: Experimental (top traces, in black) and simulated (bottom traces, at
300 K, where different colors correspond to different symmetries) vibrational
spectrum of PhC3N. The three bottom panels are expanded portions of the full
spectrum, as indicated by gray rectangles. Simulation performed using the
PGOPHER software [27] and the results of the anharmonic calculations (band
centers, intensities, and rotational constants) and normalized to the
strongest IR active mode ($\nu_{4}$). The simulations are inverted relative to
the observed spectrum solely for comparison purposes. Secure band assignments
are indicated by dash lines; additional labelled bands are those for which an
assignment is proposed; simulated bands without any labels remain unassigned.
The infrared spectrum of PhC3N is presented in Fig. 3 together with a
simulation of the vibrational fundamentals predicted by our anharmonic quantum
chemical calculations. At our experimental resolution, the rotational contour
of most bands is evident in the spectrum. Because the simulation is in very
good agreement with the experiment, numerous assignments can be made with
confidence. For ambiguous assignments, two criteria can be invoked: (i)
$c$-type bands (B1 symmetry) exhibit a sharp Q-branch; and (ii) $a$-type bands
(A1) are usually narrower than the others (as can be seen on the simulated
spectrum). Out of the 36 active infrared modes of PhC3N, 14 can be assigned
with little or no ambiguity while 9 other bands have tentative assignments at
this juncture; the remaining bands are either weak or are predicted in crowded
regions of the spectrum (see Table 1 for a detailed list of the proposed
assignments). Experimental band centers are taken as the frequency of the
$Q$-branch when one exists ($c$-type bands), equidistant between the $P$\- and
$R$-branches, or at the maximum of the envelope when no clear contour is
visible. Considering $c$-type $Q$-branches are several wavenumbers wide, an
accuracy of $\pm 2$ cm-1 can be expected for the band centers with B1
symmetry. For the others, a conservative value of $\pm 5$ cm-1 is proposed.
Table 1: Fundamental vibrational bands (position and intensity) of PhC3N from
the quantum chemical calculations performed in this work at the harmonic and
anharmonic levels of theory, and proposition of assignments. Modes are
numbered following the anharmonic calculations frequency order. Modes energy
are in wavenumbers, intensities are in km/mol, $\delta$ values are in %. The
experimental assignments are split in two categories, the relatively secure
ones (column “Assign.") and the tentative ones (column “Prop."); $\delta$
values of the latter are reported in italics.
Mode | | Harm. Calc. | | Anharm. Calc. | | Exp.
---|---|---|---|---|---|---
$\nu$ | Sym. | | Energy | Int. | | Energy | Int. | | Assign. | Prop. | $\delta\,^{a}$
1 | A1 | | 3214 | 6.7 | | 3055 | 13.5 | | 3078 | | 0.8
2 | A1 | | 3226 | 5.6 | | 3026 | 0.0 | | | 3047 | 0.7
3 | A1 | | 3197 | 0.0 | | 2999 | 1.7 | | | 3047 | 1.6
4 | A1 | | 2431 | 242.3 | | 2399 | 220.1 | | 2284 | | -4.8
5 | A1 | | 2292 | 10.7 | | 2265 | 7.5 | | | 2154/2198 | -4.9/-3.0
6 | A1 | | 1675 | 1.0 | | 1633 | 0.5 | | | |
7 | A1 | | 1545 | 7.7 | | 1510 | 4.9 | | 1495 | | -1.0
8 | A1 | | 1306 | 1.6 | | 1284 | 0.8 | | | |
9 | A1 | | 1214 | 2.1 | | 1193 | 2.0 | | 1182 | | -0.9
10 | A1 | | 1064 | 4.6 | | 1045 | 4.5 | | 1029 | | -1.5
11 | A1 | | 1027 | 1.3 | | 1012 | 1.0 | | | 1004/981 | -0.8/-3.1
12 | A1 | | 981 | 0.1 | | 960 | 0.0 | | | |
13 | A1 | | 706 | 3.4 | | 693 | 3.5 | | | |
14 | A1 | | 371 | 0.8 | | 362 | 0.5 | | | |
15 | A2 | | 1019 | 0.0 | | 1000 | 0.0 | | | |
16 | A2 | | 876 | 0.0 | | 862 | 0.0 | | | |
17 | A2 | | 413 | 0.0 | | 412 | 0.0 | | | |
18 | B1 | | 1042 | 0.0 | | 1023 | 0.0 | | | |
19 | B1 | | 967 | 2.8 | | 953 | 2.4 | | 920 | | -3.5
20 | B1 | | 792 | 38.7 | | 784 | 32.7 | | 757 | | -3.4
21 | B1 | | 716 | 36.7 | | 706 | 42.7 | | 687 | | -2.6
22 | B1 | | 576 | 13.2 | | 568 | 12.4 | | 532 | | -6.3
23 | B1 | | 523 | 0.2 | | 515 | 0.3 | | | |
24 | B1 | | 384 | 2.8 | | 377 | 2.8 | | 359 | | -4.9
25 | B1 | | 195 | 2.6 | | 192 | 2.5 | | | 185 | -3.4
26 | B1 | | 74 | 0.6 | | 71 | 0.7 | | 69 | | -3.3
27 | B2 | | 3206 | 1.9 | | 3030 | 9.5 | | | 3095 | 2.1
28 | B2 | | 3222 | 8.4 | | 3012 | 2.6 | | | |
29 | B2 | | 1646 | 0.7 | | 1608 | 0.5 | | | |
30 | B2 | | 1493 | 8.6 | | 1475 | 4.5 | | 1448 | | -1.8
31 | B2 | | 1364 | 1.2 | | 1331 | 0.9 | | | 1336 | 0.4
32 | B2 | | 1320 | 1.3 | | 1292 | 1.0 | | | 1280 | -0.9
33 | B2 | | 1193 | 0.4 | | 1183 | 0.5 | | | |
34 | B2 | | 1117 | 4.4 | | 1094 | 3.8 | | 1069 | | -2.3
35 | B2 | | 644 | 0.0 | | 635 | 0.0 | | | |
36 | B2 | | 584 | 0.5 | | 574 | 0.4 | | | |
37 | B2 | | 511 | 5.3 | | 504 | 5.6 | | 482 | | -4.3
38 | B2 | | 226 | 2.3 | | 221 | 2.4 | | | 209 | -5.6
39 | B2 | | 71 | 1.1 | | 71 | 1.2 | | | 69 | -2.7
${}^{a}\
\delta=(\mathrm{Exp.}-\mathrm{Anharm.Calc.})/\mathrm{Anharm.Calc.}\times 100$
A few of the assignments proposed in Fig. 3 and Table 1 warrant further
discussion. While the assignment of the $\nu_{26}$ band (out-of-plane backbone
motion) is relatively secure thanks to the presence of a sharp $Q$-branch, the
band appears stronger than expected from our calculations. A possible
explanation is that $\nu_{39}$ (C#C-C#N rocking), predicted roughly at the
same frequency as $\nu_{26}$, contributes to the experimental band profile,
thus enhancing absorption. In absence of any specific feature arising from
$\nu_{39}$ on the experimental spectrum, we have chosen to assign this band to
the same band center as $\nu_{26}$, at 69 cm-1. Slightly higher in frequency
are $\nu_{25}$ and $\nu_{38}$, which involve similar nuclear motion as
$\nu_{26}$ and $\nu_{39}$ and are also predicted to lie close in energy (at
192 and 221 cm-1, respectively), which in combination likely yield the feature
observed around 200 cm-1. It is not straightforward, however, to unambiguously
assign each band center, in part because $\nu_{25}$ is the only band of
$B_{1}$ symmetry with a sizeable transition moment that does not exhibit a
sharp Q-branch. Tentative assignments are proposed in Table 1, but their
accuracy should be taken with caution as the actual bands centers could lie in
a 10 to 20 cm-1 window from the proposed assignments. Concerning $\nu_{11}$
(ring symmetric breathing) and $\nu_{5}$ (C#C-C#N stretch)—predicted at 1012
and 2265 cm-1, respectively—two bands lie systematically close to the expected
energy, with reasonable band profiles, thus in each case both assignments are
reported in Table 1. In the case of $\nu_{5}$, the lowest frequency
assignment, $i.e.$, 2154 cm-1 appears most likely as the relatively strong
intensity of the observed band would indicate a fundamental that could
otherwise not be predicted. However the shape of the band lying at 2198 cm-1
is closer to that expected for an $a$-type band, assuming that our
aforementioned criterion remains valid at these frequencies (i.e. hot bands
and combination bands could significantly affect the simplistic fundamentals-
only picture).
The most difficult analysis lies in the ${\sim}3000-3200$ cm-1 region, where
C-H stretching motions are observed. From previous studies on similar
molecules like phenylacetylene [14] and naphthalene [16], it is well-known
that this frequency range is plagued by anharmonic resonances which can
significantly complicate assignment. For PhC3N, three modes of
$\mathrm{A}_{1}$ ($\nu_{1}$, $\nu_{2}$, $\nu_{3}$) and two of $\mathrm{B_{2}}$
($\nu_{27}$, $\nu_{28}$) symmetries (Table 1) very close in energy (of order
of ${\sim}$10 cm-1 for the harmonic frequencies) and can mix strongly, thus
qualitatively shifting the fundamental energies and altering band intensities.
In such cases, an approximate deperturbation analysis was performed using the
generalized VPT2 calculations to identify and treat anharmonic resonances,
e.g. those arising from Fermi (so-called “1–2” resonances) and Darling-
Dennison (“1–1” interactions in the present case). These resonances are
identified based on small differences in the energies of states and a model
variational calculation [30]; the former is a zeroth order estimate for
resonances, while the latter tests for the magnitude of the coupling [31].
The VPT2 routines in Gaussian treat the problem of state-to-state coupling as
effective $2\times 2$ Hamiltonians, where the diagonal elements correspond to
the state energies, and the off-diagonal elements represent coupling between
the two states. These values are shown in Table 2, and give rise to
significant deviation from the harmonic frequencies and intensities shown in
Table 1. Of particular note is the unintuitive complete loss of intensity in
$\nu_{2}$: $\nu_{3}$ is “dark” in the harmonic approximation, and it gains
intensity primarily through borrowing intensity from the stronger $\nu_{1}$
and $\nu_{2}$ bands. In a general resonance picture, this interaction shares
intensity, and to render a mode completely inactive is extremely rare if not
unheard of. Given the VPT2 treatment here is only approximate, our effective
deperturbation analysis is likely inadequate to properly treat these bands,
and a fully coupled model involving fundamentals, combination bands, and
overtones, is required instead.
In light of this preliminary analysis, however, as well as the fact that the
${\sim}3000$ cm-1 region is heavily congested, we assume all three
$\mathrm{A}_{1}$ modes are IR active. Measurement under cold conditions—either
in a supersonic jet or buffer gas cell—will help future analysis of this
molecule by eliminating the possibility of combination bands and overtones, in
addition to minimizing lineshape blending from rotational contours. Similarly,
selective deuteration might clarify the assignment of some features [14].
Table 2: Strong anharmonic resonances and their corresponding off-diagonal matrix elements for bands in the ${\sim}3100-3200$ cm-1 region. Darling-Dennison (DDR) and Fermi (FR) resonances are indicated; in this table, the former corresponds to $1-1$ type DDR, referring to the number of quanta for states involved. State 1 | State 2 | Type | Coupling
---|---|---|---
$v_{1}=1$ | $v_{3}=1$ | DD | -6.6
$v_{2}=1$ | $v_{3}=1$ | DD | 14.5
$v_{3}=1$ | $v_{6}=1+v_{7}=1$ | FR | -23.2
$v_{3}=1$ | $v_{29}=1+v_{30}=1$ | FR | -29.3
$v_{27}=1$ | $v_{28}=1$ | DD | 9.4
### 3.3 Rotational spectroscopy
Using the ground state rotational and quartic centrifugal distortion constants
from the anharmonic calculation, 65 strong lines of the CP-FTMW spectrum were
assigned in a straightforward fashion using the PGOPHER software [27] (Fig.
2). The derived values are extremely close to those predicted by the
calculation (the weighted frequency difference $\delta$ is less than 1 % for
the rotational constants, see Table 3). This initial set of constants was then
used to assign the millimeter-wave data. Loomis-Wood diagrams were produced by
means of the LWWa software from Lodyga et al. [32] to aid in the assignment of
the high-$J$ transitions. In total, 6151 $a$-type transitions (3780 different
frequencies as a result of unresolved asymmetric splitting) of PhC3N in its
ground vibrational state were assigned in the millimeter-wave spectrum, with
values of $J^{\prime\prime}$ up to 199 and $K_{a}^{\prime\prime}$ up to 42.
The SPFIT/SPCAT suite of programs [33] using the Watson $S$-reduced
Hamiltonian in the $I^{r}$ representation was used to determine best-fit
spectroscopic constants. All transitions were weighted accorded to their
expected experimental accuracy, $i.e.$, 2 kHz and 25 kHz for the CE-FTMW and
CP-FTMW transitions, respectively, and 50 kHz for the millimeter-wave
transitions. To reproduce the data to their experimental accuracy, inclusion
of several sextic and octic centrifugal distortion constants was required.
Finally, the CE-FTMW transitions—the only ones for which the nuclear
quadrupole splitting was resolved—were added to the fit to determine the
$\chi$(N) terms. All 57 hyperfine components were reproduced to the
measurement uncertainty by adjusting only $\chi_{aa}$(N) and $\chi_{bb}$(N).
Both parameters are close to those expected from calculation (to within about
15 %, Table 3). When a transition was observed by both CP-FTMW and CE-FTMW
spectroscopy, only the latter was retained in the fit, owing to the higher
frequency accuracy of the cavity instrument. The rotational constants derived
from a fit to all of the assigned rotational transitions are reported in Table
3. The final 49 kHz RMS value of the fit, corresponding to a reduced standard
deviation $\sigma=1.00$, indicates that our present model adequately
reproduces the ground state rotational spectrum of PhC3N.
Table 3: Spectroscopic constants (rotational, centrifugal distorsion, and
nuclear quadrupole constants) of PhC3N in its vibrational ground state (in
MHz) and relevant fit parameters. Numbers in parenthesis are $1\sigma$
uncertainties expressed in the unit of the last digit. Parameters in brackets
were kept fixed to the calculated values.
Constant | Calc.a | Exp. | $\delta\,^{b}$
---|---|---|---
$A_{0}$ | $5656.9$ | $5659.722\,(15)$ | 0.05
$B_{0}$ | $567.0$ | $569.582206\,(39)$ | 0.46
$C_{0}$ | $515.4$ | $517.404488\,(37)$ | 0.39
$D_{J}\times 10^{6}$ | $3.6$ | $3.85110\,(77)$ | 7.0
$D_{JK}\times 10^{3}$ | $0.78$ | $0.827177\,(85)$ | 6.1
$D_{K}\times 10^{3}$ | $0.42$ | $0.400\,(68)$ | -4.8
$d_{1}\times 10^{6}$ | $-0.50$ | $-0.55191\,(24)$ | 10
$d_{2}\times 10^{6}$ | $-0.52$ | $-0.57101\,(79)$ | 9.8
$H_{J}\times 10^{12}$ | | $-0.957\,(11)$ |
$H_{JK}\times 10^{9}$ | | $0.7916\,(18)$ |
$H_{KJ}\times 10^{6}$ | | $-0.03051\,(10)$ |
$h_{2}\times 10^{12}$ | | $0.575\,(15)$ |
$h_{3}\times 10^{12}$ | | $0.10329\,(55)$ |
$L_{JJK}\times 10^{15}$ | | $-1.295\,(20)$ |
$L_{JK}\times 10^{12}$ | | $0.06954\,(86)$ |
$L_{KKJ}\times 10^{12}$ | | $-3.512\,(36)$ |
$\chi_{aa}$ | $-4.96$ | $-4.219\,(77)$ | -15
$\chi_{bb}$ | $2.39$ | $2.114\,(68)$ | -12
$\chi_{cc}$ | $2.57$ | $2.143\,^{c}$ | -17
$N\,^{d}$ | | 6256/3877/57 |
$J^{\prime\prime}_{\mathrm{max}}$, $K^{\prime\prime}_{a\,\mathrm{max}}$ | | 199, 42 |
rms /kHz | | 0.049 |
$\sigma\,^{e}$ | | 1.00 |
a $\omega$B97XD/cc-pVQZ level of theory, Bayesian corrected for $A$, $B$, and
$C$, and anharmonic values for the centrifugal distortion constants, and
equilibrium values for the hyperfine constants
${}^{b}\delta=(B_{\mathrm{exp.}}-B_{\mathrm{calc.}})/B_{\mathrm{calc.}}\times
100$ (in %)
c Derived value
d Total number of lines in the fit / Number of different frequencies / Number
of lines with resolved nuclear quadrupole structure
e Reduced standard deviation, unitless
Figure 4: Portions of the millimeter-wave spectrum of PhC3N in comparison with
a simulation of the pure rotational transitions in the ground vibrational
state using the best-fit set of spectroscopic constants (Table 3). The
simulation has been performed using the PGOPHER software and the resulting
trace was then post-processed with a second derivative to allow a more
straightforward comparison with the experimental spectrum. The line density in
the experimental trace is far greater than our simulation, very likely because
of lines from vibrational satellites.
Figure 4 shows a portion of the millimeter-wave spectrum in comparison with a
simulation of PhC3N in its vibrational ground state using the experimentally
determined best-fit parameters from Table 3. As illustrated in this figure,
many lines remain unassigned, but most of these likely arise from vibrational
satellites, for which no attempt at assigning was made in the present study.
Although longer integration times would allow a more in-depth analysis of
these satellites, the spectrum is already fairly dense, implying that we may
be close to the confusion limit. Indeed, the 2 µbar pressure used in this
study, although quite low, was actually a compromise between reasonable
signal-to-noise ratio and pressure broadening. Even at this pressure many
lines are broader then expected from the effects of pressure broadening alone,
and consequently may in fact be a spectral superposition of several
transitions.
Regarding hyperfine splitting, its magnitude rapidly collapses with increasing
$J$, as expected from this type of interaction. No splitting is observed in
the CP-FTMW nor in millimeter-wave measurements, and in the cavity experiments
it is marginally resolved above $J^{\prime\prime}\approx 11$ (${\sim}12$ GHz).
Even at the lowest-$J$ transitions measured here ($J^{\prime\prime}=5$,
${\sim}6$ GHz), the splitting due to hyperfine and Doppler effects are
comparable (Fig. 5). The experimentally derived value of $\chi_{aa}$(N)
[$-4.219$ (77) MHz; Table 3] is very similar to that reported for PhCN
[-4.23738(36) MHz [9]], and the relative magnitudes are in agreement with the
theoretical values of $\chi_{aa}$(N) calculated at the $\omega$B97X-D/cc-pVQZ
level of theory for PhC3N (-4.959 MHz) and PhCN (-4.962 MHz). The small
changes in $\chi$(N) are an indication that the local electronic structure of
the nitrogen atom is relatively insensitive to the distance from the aromatic
ring. Equivalently, this finding implies that electron delocalization through
ring conjugation is very poorly coupled to the chain regardless of length.
Similar behavior is seen for cyanopolyyne chains, where the value of $eQq$
equivalent to $\chi_{aa}$(N) (around $-4.1$ MHz) is relatively invariant with
the length of the chain as well [34].
Figure 5: Example of a transition showing resolved hyperfine structure on the
CE-FTMW spectrum, and comparison with 10 K simulations using the final set of
spectroscopic parameters, with and without taking into account the Doppler
splitting (simulations performed using the PGOPHER software, assuming a
Lorentzian lineshape). As before, the simulation with Doppler splitting is
inverted relative to the observed spectrum solely for comparison purposes.
### 3.4 Astronomical considerations
Detection of molecules in space by radio astronomy is heavily dependent on the
magnitude of their permanent dipole moment. In comparison with PhCN, our
theoretical predictions suggest PhC3N is substantially more polar (5.9 D vs.
of 4.5 D, where the former value has statistical uncertainty of $\pm 0.25$ D
based on our prior benchmarking at the $\omega$B97X-D/cc-pVQZ level of theory
[35]). Currently, PhCN is theorized to form in cold, dark clouds via a
barrierless reaction between C6H6 and CN radical [4, 36]. We thus speculate
that PhC3N could be formed via C2 insertion to PhCN or through an analogous
C#N addition reaction between CN radical and phenylacetylene (PhC2H), or C3N
and C6H6. If the latter mode is operative, then the abundance ratio of
PhCN/PhC3N will be dependent on CN/C3N, assuming similar reaction cross-
sections. The CN + PhC2H process has been studied in crossed-beam experiments
by Bennett et al. [11], where the authors identify PhC3N as a potential
reaction product, albeit not definitively so due to the lack of isomer
specificity and at collision energies well in excess of interstellar cloud
conditions (${\sim}30$ kJ/mol). This finding suggests other isomers, namely
ortho, meta, para-CEB, might plausibly be formed from this reaction. While we
have not attempted to experimentally characterized these species, estimates of
their spectroscopic parameters are provided here. By scaling the experimental
parameters for PhC3N (Table 3) to correct for vibrational and electronic
effects (Table S6 in the Supporting Information), in conjunction to Bayesian
uncertainties obtained from benchmarking [35], reliable constraints of these
constants should aid future experimental searches for these species.
In terms of astronomical detection, although the hyperfine structure does not
take a large part the present rotational analysis, this splitting is partially
resolved with cavity measurements up to 16 GHz. In cold, dark clouds such as
TMC-1 where the linewidths are comparable to those measured with our CE-FTMW
spectrometer, it is thus necessary to consider hyperfine splitting, as recent
work on PhCN [4] demonstrates. This is particularly true for a relatively
heavy molecule like PhC3N whose strongest lines should lie at centimeter-
wavelengths at low temperatures (Fig. 6): at 10 K, the strongest features
correspond to $J^{\prime\prime}=21$ around 23 GHz, while at 6 K —a typical
temperature for molecules in TMC-1— the strongest features are near 15 GHz
($J^{\prime\prime}=14$). Thus, the $\mathrm{X/K_{u}}$ (8–12/12–18 GHz) bands
appear to be the most promising to detect PhC3N in TMC–1. In sources with
somewhat warmer temperatures, the intensity of individual transitions is
significantly decreased due to the larger partition function, and the peak
intensity, although relatively flat, falls in the W (75–110 GHz) and N
(100–200 GHz) bands.
Figure 6: Calculated rotational spectrum of PhC3N at 10 K (purple), 100 K
(blue), and 300 K (orange). For the purposes of display, these relative
intensities for the 100 K plot have increased by a factor of 5, while for the
300 K plot, the increase is 10.
Detection of multiple isomers is also a powerful tool for constraining
physical and chemical conditions in astrophysical environments. To assist in
this process, we have calculated the relative energy of the aforementioned
ortho, meta, and para-CEB isomers, along with the isocyanide isomer, PhCCNC.
As shown in Fig. 7, apart from the isocyanide isomer, placement of the
acetylenic unit on different parts of the ring produces isomers with
comparable stability to PhC3N, i.e., they are effectively degenerate at the
level of uncertainty afforded by G3//B3LYP ($\pm 4$ kJ/mol). As such, a
determination of their relative abundances would provide a sensitive test of
thermodynamics vs. kinetics in molecule formation. To aid further laboratory
and hopefully astronomical efforts, Table S6, in addition to providing
estimates of rotational constants, also reports calculated dipole moments at
the $\omega$B97X-D/cc-pVQZ level of theory using two methods of empirical
scaling to correct for vibrational effects and deficiencies in the electronic
structure method. We note that the Bayesian scaling factors obtained in Ref.
[35] applied to PhC3N —where we now have accurately determined parameters—
exceed the performance of the purely theoretical VPT2 corrections, and bring
the theoretical predictions within a few MHz of the experimentally measured
ones (see also Fig. S1). We expect a similar degree of precision and
uncertainty for other isomers.
Figure 7: 0 K energetics of isomers of interest calculated at the G3//B3LYP
level of theory, relative to PhC3N. Energies to the nearest kJ/mol are
annotated above each bar.
## 4 Conclusion
Using a combination of gas-phase measurements and quantum chemical
calculations, the fundamental vibrational frequencies of most of the strongest
IR-active fundamentals of PhC3N and over 6000 pure rotational transitions in
its ground state between 6.5 and 220 GHz have been measured. The assignment of
spectra in different regions was guided by theoretical predictions: in the
infrared, anharmonic calculations helped to assign most bands, with the
exception of the most congested and perturbed region around $3000$ cm-1; in
the radio domain, estimates of centrifugal distortion and hyperfine terms
proved useful. Comparisons with PhCN suggest that the local electronic
structure of the terminal nitrogen —as probed through its quadrupole moment
$\chi_{aa}$(N)— is relatively unaffected by chain lengthening, indicating a
similar electric field gradient for nitrogen in both PhCN and PhC3N. In
addition, we provide accurate and reliable predictions of the thermochemistry
and spectroscopic parameters for several isomers of PhC3N, which should prove
useful in guiding future laboratory experiments. These isomers include the
isocyanide isomer of PhC3N and the three cyanoethynylbenzenes alluded to in
previous work [11]; their discovery along with PhC3N in the ISM would provide
sensitive measurements of local chemical and physical interstellar
environments.
With highly precise measurements of the rest frequencies and corresponding
spectroscopic constants, a search for PhC3N can now be undertaken with
considerable confidence in the ISM. The large permanent dipole moment
(predicted to be 5.9 D), in addition to its chemical and structural similarity
to astronomical PhCN makes PhC3N an excellent candidate for detection towards
cold, dark molecular clouds such as TMC-1. Hyperfine-resolved measurements are
expected to be highly relevant in a potential discovery, given that at low
temperatures the strongest transitions lie in the $\mathrm{X/K_{u}}$ bands
where the splitting is comparable to the source linewidth.
## Acknowledgements
O.C., O.P., and M.-A.M.-D. acknowledge funding support from the Région Ile-de-
France through DIM-ACAV+, from the Agence Nationale de la Recherche
(ANR-19-CE30-0017-01), from the “Investissements d’Avenir” LabEx PALM
(ANR-10-LABX-0039-PALM), and from the Programme National “Physique et Chimie
du Milieu Interstellaire” (PCMI) of CNRS/INSU with INC/INP co-funded by CEA
and CNES. K.L.K.L. and M.C.M. acknowledge funding support from NSF grant
AST-1908576 and NASA grant 80NSSC18K0396. Z.S.B. acknowledges support from the
Chateaubriand Fellowship of the Office for Science & Technology of the Embassy
of France in the United States. The authors are thankful to the AILES beamline
staff for providing access to the FTIR interferometer.
## References
* Frenklach and Feigelson [1989] M. Frenklach, E. D. Feigelson, Formation of polycyclic aromatic hydrocarbons in circumstellar envelopes, The Astrophysical Journal 341 (1989) 372.
* Snow and Witt [1995] T. P. Snow, A. N. Witt, The Interstellar Carbon Budget and the Role of Carbon in Dust and Large Molecules, Science 270 (1995) 1455–1460.
* Tielens [2008] A. Tielens, Interstellar Polycyclic Aromatic Hydrocarbon Molecules, Annual Review of Astronomy and Astrophysics 46 (2008) 289–337.
* McGuire et al. [2018] B. A. McGuire, A. M. Burkhardt, S. Kalenskii, C. N. Shingledecker, A. J. Remijan, E. Herbst, M. C. McCarthy, Detection of the aromatic molecule benzonitrile (c-C6H5CN) in the interstellar medium, Science 359 (2018) 202–205.
* McNaughton et al. [2018] D. McNaughton, M. K. Jahn, M. J. Travers, D. Wachsmuth, P. D. Godfrey, J.-U. Grabow, Laboratory rotational spectroscopy of cyano substituted polycyclic aromatic hydrocarbons, Monthly Notices of the Royal Astronomical Society 476 (2018) 5268–5273.
* McCarthy et al. [2020] M. C. McCarthy, K. L. K. Lee, R. A. Loomis, A. M. Burkhardt, C. N. Shingledecker, S. B. Charnley, M. A. Cordiner, E. Herbst, S. Kalenskii, E. R. Willis, C. Xue, A. J. Remijan, B. A. McGuire, Interstellar detection of the highly polar five-membered ring cyanocyclopentadiene, Nature Astronomy (2020). In press.
* Gautier et al. [2011] T. Gautier, N. Carrasco, A. Buch, C. Szopa, E. Sciamma-O’Brien, G. Cernogora, Nitrile gas chemistry in Titan atmosphere, Icarus 213 (2011) 625.
* Loomis and et al. [2020] R. Loomis, et al., The third time’s a charm? a rigorous investigation of spectral line stacking techniques and application to the detection of HC11N, submitted (2020).
* Wohlfart et al. [2008] K. Wohlfart, M. Schnell, J.-U. Grabow, J. Küpper, Precise dipole moment and quadrupole coupling constants of benzonitrile, Journal of Molecular Spectroscopy 247 (2008) 119–121.
* Bak et al. [1962] B. Bak, D. Christensen, W. B. Dixon, L. Hansen–Nygaard, J. Rastrup–Andersen, Benzene Ring Distortion by One Substituent. Microwave Determination of the Complete Structure of Benzonitrile, The Journal of Chemical Physics 37 (1962) 2027–2031. Publisher: American Institute of Physics.
* Bennett et al. [2010] C. J. Bennett, S. B. Morales, S. D. L. Picard, A. Canosa, I. R. Sims, Y. H. Shih, A. H. H. Chang, X. Gu, F. Zhang, R. I. Kaiser, A chemical dynamics, kinetics, and theoretical study on the reaction of the cyano radical (CN; X${}^{2}\Sigma^{+}$) with phenylacetylene (C6H5CCH; X1A1), Phys. Chem. Chem. Phys. 12 (2010) 8737–8749.
* Jawad et al. [2017] K. M. Jawad, T. S. Zwier, L. V. Slipchenko, C. I. Viquez Rojas, The exotic excited state behavior of 3-phenyl-2-propynenitrile, http://hdl.handle.net/2142/97063, 2017\. International Symposium on Molecular Spectroscopy.
* Viquez Rojas et al. [2017] C. I. Viquez Rojas, L. V. Slipchenko, T. S. Zwier, K. M. Jawad, Computational modeling of electronic spectroscopy of 3-phenyl-2-propynenitrile, http://hdl.handle.net/2142/97102, 2017\. International Symposium on Molecular Spectroscopy.
* King and So [1970] G. King, S. So, Ethynylbenzene; The vibrational spectra of some deuterated isomers, Journal of Molecular Spectroscopy 36 (1970) 468–487. ZSCC: 0000063.
* Chang et al. [2010] C.-H. Chang, G. Lopez, T. J. Sears, P. M. Johnson, Vibronic Analysis of the S ${}_{\textrm{1}}$ —S ${}_{\textrm{0}}$ Transition of Phenylacetylene Using Photoelectron Imaging and Spectral Intensities Derived from Electronic Structure Calculations, The Journal of Physical Chemistry A 114 (2010) 8262–8270. ZSCC: NoCitationData[s0].
* Pirali et al. [2009] O. Pirali, M. Vervloet, G. Mulas, G. Malloci, C. Joblin, High-resolution infrared absorption spectroscopy of thermally excited naphthalene. Measurements and calculations of anharmonic parameters and vibrational interactions, Physical Chemistry Chemical Physics 11 (2009) 3443. ZSCC: 0000046.
* Pirali et al. [2015] O. Pirali, Z. Kisiel, M. Goubet, S. Gruet, M. A. Martin-Drumel, A. Cuisset, F. Hindle, G. Mouret, Rotation-vibration interactions in the spectra of polycyclic aromatic hydrocarbons: Quinoline as a test-case species, The Journal of Chemical Physics 142 (2015) 104310. Publisher: American Institute of Physics.
* Kisiel and Kraśnicki [2010] Z. Kisiel, A. Kraśnicki, The millimetre-wave rotational spectrum of phenylacetylene, Journal of Molecular Spectroscopy 262 (2010) 82–88.
* Zdanovskaia et al. [2018] M. A. Zdanovskaia, B. J. Esselman, H. S. Lau, D. M. Bates, R. C. Woods, R. J. McMahon, Z. Kisiel, The 103–360 GHz rotational spectrum of benzonitrile, the first interstellar benzene derivative detected by radioastronomy, Journal of Molecular Spectroscopy 351 (2018) 39–48.
* Frisch et al. [2016] M. J. Frisch, G. W. Trucks, H. B. Schlegel, G. E. Scuseria, M. A. Robb, J. R. Cheeseman, G. Scalmani, V. Barone, G. A. Petersson, H. Nakatsuji, X. Li, M. Caricato, A. V. Marenich, J. Bloino, B. G. Janesko, R. Gomperts, B. Mennucci, H. P. Hratchian, J. V. Ortiz, A. F. Izmaylov, J. L. Sonnenberg, D. Williams-Young, F. Ding, F. Lipparini, F. Egidi, J. Goings, B. Peng, A. Petrone, T. Henderson, D. Ranasinghe, V. G. Zakrzewski, J. Gao, N. Rega, G. Zheng, W. Liang, M. Hada, M. Ehara, K. Toyota, R. Fukuda, J. Hasegawa, M. Ishida, T. Nakajima, Y. Honda, O. Kitao, H. Nakai, T. Vreven, K. Throssell, J. A. Montgomery, Jr., J. E. Peralta, F. Ogliaro, M. J. Bearpark, J. J. Heyd, E. N. Brothers, K. N. Kudin, V. N. Staroverov, T. A. Keith, R. Kobayashi, J. Normand, K. Raghavachari, A. P. Rendell, J. C. Burant, S. S. Iyengar, J. Tomasi, M. Cossi, J. M. Millam, M. Klene, C. Adamo, R. Cammi, J. W. Ochterski, R. L. Martin, K. Morokuma, O. Farkas, J. B. Foresman, D. J. Fox, Gaussian 16 Revision A.01, 2016.
* Baboul et al. [1999] A. G. Baboul, L. A. Curtiss, P. C. Redfern, K. Raghavachari, Gaussian-3 theory using density functional geometries and zero-point energies, The Journal of Chemical Physics 110 (1999) 7650–7657.
* Simmie and Somers [2015] J. M. Simmie, K. P. Somers, Benchmarking Compound Methods (CBS-QB3, CBS-APNO, G3, G4, W1BD) against the Active Thermochemical Tables: A Litmus Test for Cost-Effective Molecular Formation Enthalpies, The Journal of Physical Chemistry A 119 (2015) 7235–7246.
* Brubach et al. [2010] J.-B. Brubach, L. Manceron, M. Rouzières, O. Pirali, D. Balcon, F. Kwabia-Tchana, V. Boudon, M. Tudorie, T. Huet, A. Cuisset, P. Roy, in: WIRMS 2009, volume 1214 of AIP Conference Proceedings, pp. 81–84.
* Faye et al. [2016] M. Faye, M. Bordessoule, B. Kanouté, J.-B. Brubach, P. Roy, L. Manceron, Improved mid infrared detector for high spectral or spatial resolution and synchrotron radiation use, Review of Scientific Instruments 87 (2016) 063119.
* Pirali et al. [2013] O. Pirali, M. Goubet, T. R. Huet, R. Georges, P. Soulard, P. Asselin, J. Courbe, P. Roy, M. Vervloet, The far infrared spectrum of naphthalene characterized by high resolution synchrotron FTIR spectroscopy and anharmonic DFT calculations, Physical Chemistry Chemical Physics 15 (2013) 10141–10150.
* Crabtree et al. [2016] K. N. Crabtree, M.-A. Martin-Drumel, G. G. Brown, S. A. Gaster, T. M. Hall, M. C. McCarthy, Microwave spectral taxonomy: A semi-automated combination of chirped-pulse and cavity fourier-transform microwave spectroscopy, The Journal of Chemical Physics 144 (2016) 124201.
* Western [2017] C. M. Western, PGOPHER: A Program for Simulating Rotational, Vibrational and Electronic Spectra, J. Quant. Spectrosc. Radiat. Transfer 186 (2017) 221–242.
* McCarthy et al. [2020] M. C. McCarthy, K. L. K. Lee, P. B. Carroll, J. P. Porterfield, B. Changala, J. H. Thorpe, J. F. Stanton, Exhaustive product analysis of three benzene discharges by microwave spectroscopy, J. Phys. Chem. A (2020).
* Pirali et al. [2017] O. Pirali, M. Goubet, V. Boudon, L. D’Accolti, C. Fusco, C. Annese, Characterization of isolated 1-aza-adamantan-4-one (C9H13NO) from microwave, millimeter-wave and infrared spectroscopy supported by electronic structure calculations, J. Mol. Spectrosc. 338 (2017) 6 – 14.
* Martin et al. [1995] J. M. L. Martin, T. J. Lee, P. R. Taylor, J. François, The anharmonic force field of ethylene, C ${}_{\textrm{2}}$ H ${}_{\textrm{4}}$ , by means of accurate ab initio calculations, The Journal of Chemical Physics 103 (1995) 2589–2602. ZSCC: NoCitationData[s0].
* Bloino et al. [2015] J. Bloino, M. Biczysko, V. Barone, Anharmonic Effects on Vibrational Spectra Intensities: Infrared, Raman, Vibrational Circular Dichroism, and Raman Optical Activity, The Journal of Physical Chemistry A 119 (2015) 11862–11874. ZSCC: 0000070.
* Lodyga et al. [2007] W. Lodyga, M. Kręglewski, P. Pracna, ǒ. Urban, Advanced graphical software for assignments of transitions in rovibrational spectra, Journal of Molecular Spectroscopy 243 (2007) 182 – 188. PRAHA2006, The 19th International Conference on High Resolution Molecular Spectroscopy.
* Pickett [1991] H. M. Pickett, The Fitting and Prediction of Vibration-rotation Spectra with Spin Interactions, J. Mol. Spectrosc. 148 (1991) 371–377.
* McCarthy et al. [2000] M. C. McCarthy, W. Chen, M. J. Travers, P. Thaddeus, Microwave Spectra of 11 Polyyne Carbon Chains, The Astrophysical Journal Supplement Series 129 (2000) 611.
* Lee and McCarthy [2020] K. L. K. Lee, M. McCarthy, Bayesian Analysis of Theoretical Rotational Constants from Low-Cost Electronic Structure Methods, The Journal of Physical Chemistry A 5 (2020) 898–910.
* Lee et al. [2019] K. L. K. Lee, B. A. McGuire, M. C. McCarthy, Gas-phase synthetic pathways to benzene and benzonitrile: a combined microwave and thermochemical investigation, Physical Chemistry Chemical Physics 21 (2019) 2946–2956.
|
# Constructing new APN functions through relative trace functions
Lijing Zheng, Haibin Kan, Yanjun Li, Jie Peng, Deng Tang
L. Zheng is with the School of Mathematics and Physics, University of South
China, Hengyang, Hunan, 421001, China, (E-mail: zhenglijing817@163.com).
H. Kan is with the School of Computer Sciences, Fudan University, Shanghai,
200433, China, (E-mail: hbkan@fudan.edu.cn).
Y. Li is with the Mathematics and Science College of Shanghai Normal
University, Shanghai, 200234, China, (yanjlmath90@163.com).
J. Peng is with the Mathematics and Science College of Shanghai Normal
University, Shanghai, 200234, China, (jpeng@shnu.edu.cn).
D. Tang is with the School of Electronic Information and Electrical
Engineering, Shanghai Jiao Tong University, Shanghai, 200240, China,
(dtang@foxmail.com).
Abstract: In 2020, Budaghyan, Helleseth and Kaleyski [IEEE TIT 66(11):
7081-7087, 2020] considered an infinite family of quadrinomials over
$\mathbb{F}_{2^{n}}$ of the form $x^{3}+a(x^{2^{s}+1})^{2^{k}}+bx^{3\cdot
2^{m}}+c(x^{2^{s+m}+2^{m}})^{2^{k}}$, where $n=2m$ with $m$ odd. They proved
that such kind of quadrinomials can provide new almost perfect nonlinear (APN)
functions when $\gcd(3,m)=1$, $k=0$, and $(s,a,b,c)=(m-2,\omega,\omega^{2},1)$
or $((m-2)^{-1}~{}{\rm mod}~{}n,\omega,\omega^{2},1)$ in which
$\omega\in\mathbb{F}_{4}\setminus\mathbb{F}_{2}$. By taking $a=\omega$ and
$b=c=\omega^{2}$, we observe that such kind of quadrinomials can be rewritten
as $a{\rm Tr}^{n}_{m}(bx^{3})+a^{q}{\rm Tr}^{n}_{m}(cx^{2^{s}+1})$, where
$q=2^{m}$ and ${\rm Tr}^{n}_{m}(x)=x+x^{2^{m}}$ for $n=2m$. Inspired by the
quadrinomials and our observation, in this paper we study a class of functions
with the form $f(x)=a{\rm Tr}^{n}_{m}(F(x))+a^{q}{\rm Tr}^{n}_{m}(G(x))$ and
determine the APN-ness of this new kind of functions, where
$a\in\mathbb{F}_{2^{n}}$ such that $a+a^{q}\neq 0$, and both $F$ and $G$ are
quadratic functions over $\mathbb{F}_{2^{n}}$. We first obtain a
characterization of the conditions for $f(x)$ such that $f(x)$ is an APN
function. With the help of this characterization, we obtain an infinite family
of APN functions for $n=2m$ with $m$ being an odd positive integer:
$f(x)=a{\rm Tr}^{n}_{m}(bx^{3})+a^{q}{\rm Tr}^{n}_{m}(b^{3}x^{9})$, where
$a\in\mathbb{F}_{2^{n}}$ such that $a+a^{q}\neq 0$ and $b$ is a non-cube in
$\mathbb{F}_{2^{n}}$. We verify that the aforementioned APN quadrinomials are
CCZ-inequivalent to any other known APN functions over $\mathbb{F}_{2^{10}}$.
We also obtain two infinite families of APN functions: $a{\rm
Tr}^{n}_{m}(bx^{3})+a^{q}{\rm Tr}^{n}_{m}(gx^{5}+ex^{4q+1})$, where
$b,~{}g,~{}e$ satisfy: $i)$ $b$ not a cube, $g=1$, $e=\frac{1}{b^{2q-2}}$; or
$ii)$ $b$ not a cube, and $g=e=b$. We can also find (at least) two new
sporadic instances of APN functions over $\mathbb{F}_{2^{10}}$ up to CCZ-
equivalence.
Keywords: APN functions; relative trace functions; quadratic functions; CCZ-
equivalence
## 1 Introduction
Throughout this paper, we often identify the finite field $\mathbb{F}_{2^{n}}$
with $\mathbb{F}^{n}_{2}$ which is the $n$-dimensional vector space over
$\mathbb{F}_{2}$. Any function
$F:\mathbb{F}_{2^{n}}\rightarrow\mathbb{F}_{2^{m}}$ is called an
$(n,m)$-function or vectorial Boolean functions if the values $n$ and $m$ are
omitted. Vectorial Boolean functions are of critical importance in the field
of symmetric cryptography, and the security of encryption algorithms heavily
depends on the cryptographic properties of the vectorial Boolean functions.
Researchers have proposed various properties to measure the resistance of a
vectorial Boolean function to different kinds of cryptanalysis, including
differential uniformity, nonlinearity, boomerang uniformity, algebraic degree,
and so on. The lower the differential uniformity of a vectorial Boolean
function, the better its security against differential cryptanalysis. In this
paper, we mainly focus on the $(n,n)$-functions. The differential uniformity
of any such functions is at least 2, and the functions achieving this bound
are called almost perfect nonlinear (APN).
It is difficult to find new infinite families of APN functions up to CCZ-
equivalence. Up to now, only 6 infinite families of APN monomials and 14
infinite families of APN polynomials are known, since the early 90’s. On the
other hand, in contrast to these facts, there are a lot of APN functions even
over “small” field: for example, thousands of CCZ-inequivalent APN functions
have been found over $\mathbb{F}_{2^{8}}$ [25]. Constructing new instances of
infinite families is an area of deep heading research. We present Tables I and
II including all currently known infinite families of APN functions. To Table
II, we add the new function found with Theorem 3.3 in Section 3 below. We
refer the readers to a recent nice work of Budaghyan et al. for more details
on the classification of the known families of APN functions [7].
TABLE I: Known infinite families of APN power functions over $\mathbb{F}_{2^{n}}$ Family | Exponent | Conditions | Algebraic degree | Source
---|---|---|---|---
Gold | $2^{i}+1$ | ${\rm gcd}(i,n)=1$ | 2 | [18]
Kasami | $2^{2i}-2^{i}+1$ | ${\rm gcd}(i,n)=1$ | $i+1$ | [19]
Welch | $2^{t}+3$ | $n=2t+1$ | $3$ | [14]
Niho | | $2^{t}+2^{t/2}-1$, $t$ even
---
$2^{t}+2^{(3t+1)/2}-1$, $t$ odd
$n=2t+1$ | | $t/2+1$
---
$t+1$
[15]
Inverse | $2^{2t}-1$ | $n=2t+1$ | $n-1$ | [1, 22]
Dobbertin | $2^{4i}+2^{3i}+2^{2i}+2^{i}-1$ | $n=5i$ | $i+3$ | [16]
Throughout this paper, let $\omega\in\mathbb{F}_{4}\backslash\\{0,1\\}.$ Very
recently, Budaghyan, Helleseth, and Kaleyski introduced an infinite family of
quadrinomials over $\mathbb{F}_{2^{n}}$ of the following form:
$g_{s}(x)=x^{3}+a(x^{2^{s}+1})^{2^{k}}+bx^{3\cdot
2^{m}}+c(x^{2^{s+m}+2^{m}})^{2^{k}},$
where $n=2m$. They showed that this family can provide new infinite families
of APN functions [12]. More precisely, they showed that $g_{s}(x)$ is a new
APN function if $k=0$, $(s,a,b,c)=(m-2,\omega,\omega^{2},1)$, or
$((m-2)^{-1}~{}{\rm mod}~{}n,\omega,\omega^{2},1)$, if $m$ is odd with ${\rm
gcd}(3,m)=1$. They also pointed out that when $k\geq 1$, $g_{s}(x)$ can also
be APN, however, CCZ-equivalent to some known ones.
Let $n=2m$ and $q=2^{m}$. In this paper, our motivation is to find new
infinite families of APN functions over $\mathbb{F}_{2^{n}}$. We revisit the
above-mentioned two infinite families of APN quadrionomials obtained in [12].
Observing that for any odd positive integer $s$, $\omega^{2^{s}}=\omega^{2}$,
the APN functions for $s=m-2$, or $(m-2)^{-1}{\rm mod}~{}n$ can be rewritten
as $g_{s}(x)=a{\rm Tr}^{n}_{m}(bx^{3})+a^{q}{\rm Tr}^{n}_{m}(cx^{2^{s}+1})$,
$a=\omega$, $b=c=\omega^{2}$. Here ${\rm Tr}^{n}_{m}(x):=x+x^{2^{m}}$ for
$n=2m$. Inspired by the quadrinomials and our observation, let
$a\in\mathbb{F}_{2^{n}}$, we study a class of functions with the following
form:
$f(x)=a{\rm Tr}^{n}_{m}(F(x))+a^{q}{\rm Tr}^{n}_{m}(G(x)),~{}a+a^{q}\neq 0,$
(1)
where $F$ and $G$ are quadratic functions with $F(0)=G(0)=0$.
Based on the framework (1), we carefully choose quadratic functions $F$ and
$G$ for finding APN functions. We mainly consider two kinds of functions in
(1) by setting $F$ and $G$ as follows.
$i)$ $F(x)=bx^{3}$, $G(x)=cx^{2^{s}+1}$;
$ii)$ $F(x)=bx^{2^{i}+1}+cx^{2^{i+m}+1}$, $G(x)=gx^{2^{s}+1}+ex^{2^{s+m}+1}$,
where $b,c,g,e\in\mathbb{F}_{2^{n}}$, and $i,s$ are positive integers.
Let $n=2m$ with $m$ odd. Let $a\in\mathbb{F}_{2^{n}}$, and
$f_{s}(x)=a{\rm Tr}^{n}_{m}(bx^{3})+a^{q}{\rm
Tr}^{n}_{m}(cx^{2^{s}+1}),~{}a+a^{q}\neq 0.$
We can find two more exponents $s=3$, or $m+2$, and the corresponding
conditions on the coefficients such that $f_{s}(x)$ is an APN function over
$\mathbb{F}_{2^{n}}$. Code isomorphism tests (see Sec. 2 below) indicate that
for the exponent $s=3$, the APN function found with Theorem 3.3:
$f_{3}(x)=a{\rm Tr}^{n}_{m}(bx^{3})+a^{q}{\rm Tr}^{n}_{m}(b^{3}x^{9}),$
where $b$ is a non-cube, is new up to CCZ-equivalence over
$\mathbb{F}_{2^{10}}$. We can also discover more coefficients for these two
exponents $s=m-2$, and $(m-2)^{-1}{\rm mod}~{}n$ discovered by Budaghyan et
al. such that $f_{s}(x)$ is APN without the assumption that ${\rm
gcd}(3,m)=1$. In this way, some new instances of APN functions over
$\mathbb{F}_{2^{10}}$ and $\mathbb{F}_{2^{14}}$ of the form $f_{s}(x)$ can
also be found.
Let $n=2m$, $q=2^{m}$, $a\in\mathbb{F}_{2^{n}}$, and
$\displaystyle h_{i,s,b,c,g,e}(x)=a{\rm
Tr}^{n}_{m}(bx^{2^{i}+1}+cx^{2^{i+m}+1})+a^{q}{\rm
Tr}^{n}_{m}(gx^{2^{s}+1}+ex^{2^{s+m}+1}),~{}a+a^{q}\neq 0.$
We can find two infinite families of APN functions as follows, by letting
$i=1$, $s=2$, $c=0$.
$\displaystyle h_{1,2,b,0,g,e}(x)=a{\rm Tr}^{n}_{m}(bx^{3})+a^{q}{\rm
Tr}^{n}_{m}(gx^{5}+ex^{4q+1}),$
where $a\in\mathbb{F}_{2^{n}}$ such that $a+a^{q}\neq 0$, $m$ is odd, and
$b,~{}g,~{}e$ satisfy: $i)$ $b$ not cube, $g=1$, $e=\frac{1}{b^{2q-2}}$; or
$ii)$ $b$ not cube in $\mathbb{F}^{\ast}_{2^{n}}$, and $g=e=b$. By means of
the code isomorphism test, we find that these two classes of APN functions are
CCZ-inequivalent to each other, however, CCZ-equivalent to some functions in
family F12 of Taniguchi over $\mathbb{F}_{2^{10}}$. The critical technique
needed in the proof is to forge links between the cube-ness of some certain
elements and the number of solutions to the equation of the following form:
$\displaystyle Ax^{3}+Bx^{2}+B^{q}x+A^{q}=0.$
The rest of the paper is organized as follows. Some basic definitions are
given in Section 2. We characterize the condition for $f(x)$ with the form (1)
such that $f(x)$ is an APN function over $\mathbb{F}_{2^{n}}$, $n=2m$. In
Section 3, we investigate the APN property of the functions with the form (1)
by letting $F$, $G$ are both Gold functions or both quadratic binomials. We
can find a new infinite family of APN quadrinomials, and generalize the two
infinite families of APN functions found by Budaghyan et al. in [12]. We can
find two infinite families of APN hexanomials, which computationally proved
that they belong to family F12 over $\mathbb{F}_{2^{10}}$. We can also find
(at least) two new APN instances over $\mathbb{F}_{2^{10}}$. A few concluding
remarks are given in Section 4.
## 2 Preliminaries
Let $\mathbb{F}_{2^{n}}$ be the finite field consisting of $2^{n}$ elements,
then the group of units of $\mathbb{F}_{2^{n}}$, denoted by
$\mathbb{F}^{\ast}_{2^{n}}$, is a cyclic group of order $2^{n}-1$. Let
$\alpha\in\mathbb{F}_{2^{n}}.$ It is called a cube in $\mathbb{F}_{2^{n}}$, if
$\alpha=\beta^{3}$ for some $\beta\in\mathbb{F}_{2^{n}}$; otherwise, it is
called a non-cube. Let $m$ and $n$ be two positive integers satisfying
$m~{}|~{}n$, we use ${\rm Tr}^{n}_{m}(\cdot)$ to denote the trace function
form $\mathbb{F}_{2^{n}}$ to $\mathbb{F}_{2^{m}}$, i.e., ${\rm
Tr}^{n}_{m}(x)=x+x^{2^{m}}+x^{2^{2m}}+\cdots+x^{2^{(n/m-1)m}}.$
Let $f(x)$ be a function over $\mathbb{F}_{2^{n}}$. Then it can be uniquely
represented as $f(x)=\sum^{2^{n}-1}_{i=0}a_{i}x^{i}$. This is the univariate
representation of $f$. Let $0\leq i\leq 2^{n}-1$. The binary weight of $i$ is
$w_{2}(i)=\sum^{n-1}_{s=0}i_{s}$, where $i=\sum^{n-1}_{s=0}i_{s}2^{s}$,
$i_{s}\in\\{0,1\\}$. The algebraic degree of $f$, denoted by ${\rm deg}(f)$,
is the largest binary weight of an exponent $i$ with $a_{i}\neq 0$ in the
univariate representation of $f$. Functions of algebraic degree one, and two
are called affine, quadratic, respectively.
Given an $(n,n)$-function $F$, we denote by $\Delta_{F}(a,b)$ the number of
solutions to the equation $D_{a}F(x)=b$, where $D_{a}F(x)=F(x)+F(x+a)$ is the
_derivative_ of $F$ in direction $a\in\mathbb{F}_{2^{n}}$. $F$ is called
_differentially $\delta$-uniform_ if the largest value of $\Delta_{F}(a,b)$
equals to $\delta$, for every nonzero $a$ and every $b$. If $F$ is
differentially 2-uniform, we say that $F$ is _almost perfect nonlinear_ (APN).
Two $(n,m)$-functions $F$ and $G$ are called extended affine equivalent (EA-
equivalent) if there exist some affine permutation $L_{1}$ over
$\mathbb{F}_{2^{n}}$ and some affine permutation $L_{2}$ over
$\mathbb{F}_{2^{m}}$, and some affine function $A$ such that $F=L_{2}\circ
G\circ L_{1}+A$. They are called Carlet-Charpin-Zinoviev equivalent (CCZ-
equivalent) if there exists some affine automorphism $L=(L_{1},L_{2})$ of
$\mathbb{F}_{2^{n}}\times\mathbb{F}_{2^{m}}$, where
$L_{1}:\mathbb{F}_{2^{n}}\times\mathbb{F}_{2^{m}}\rightarrow\mathbb{F}_{2^{n}}$
and
$L_{2}:\mathbb{F}_{2^{n}}\times\mathbb{F}_{2^{m}}\rightarrow\mathbb{F}_{2^{m}}$
are affine functions, such that $y=G(x)$ if and only if $L_{2}(x,y)=F\circ
L_{1}(x,y)$. It is well known that EA-equivalence is a special kind of CCZ-
equivalence, and that CCZ-equivalence preserves the differential uniformity
[13]. Proving CCZ-inequivalence of functions can be very difficult in general,
and this is resolved through code isomorphism. Let $\alpha$ be the primitive
element in $\mathbb{F}_{2^{n}}$. Then two $(n,n)$-functions functions $F$ and
$G$ are CCZ-equivalent if and only if $\mathcal{C}_{F}$, $\mathcal{C}_{G}$ are
isomorphic [3], where $\mathcal{C}_{F}$ is the linear code corresponding to
$F$ with the generating matrix as follows.
$\mathcal{C}_{F}=\left(\begin{array}[]{cccc}1&1&\cdots&1\\\
0&\alpha&\cdots&\alpha^{2^{n}-1}\\\
F(0)&F(\alpha)&\cdots&F(\alpha^{2^{n}-1})\\\ \end{array}\right)$
Let $f$ be a quadratic function over $\mathbb{F}_{2^{n}}$ with $f(0)=0$.
Denote
$\Delta_{d,f}(x):=f(dx)+f(dx+d)+f(d).$
Then it is well known that $f$ is APN if and only if for every $d\neq 0$,
$\Delta_{d,f}(x)=0$ only has trivial solutions in $x$, i.e., only
$x\in\mathbb{F}_{2}$ can be a solution to $\Delta_{d,f}(x)=0$.
In the following, we determine the APN-ness of the functions with the form
(1).
###### Lemma 2.1.
Let $n=2m$, and $q=2^{m}$. Let $F$, $G$ be quadratic functions over
$\mathbb{F}_{2^{n}}$ satisfying that $F(0)=0$, and $G(0)=0$. Let $f(x)=a{\rm
Tr}^{n}_{m}(F(x))+a^{q}{\rm Tr}^{n}_{m}(G(x)),$ where $a\in\mathbb{F}_{2^{n}}$
such that $a+a^{q}\neq 0$. Then $f(x)$ is APN over $\mathbb{F}_{2^{n}}$, if
and only if the following system
$\displaystyle\begin{cases}\Delta_{d,F}(x)\in\mathbb{F}_{2^{m}}&\\\
\Delta_{d,G}(x)\in\mathbb{F}_{2^{m}}&\end{cases}$ (2)
only has $x=0,1$ as its solutions for any $d\neq 0\in\mathbb{F}_{2^{n}}$.
###### Proof.
Since $f(x)$ is quadratic with $f(0)=0$, it is equivalent to showing that the
following equation only has $x=0,1$ as its solutions for any $d\neq 0$
$\Delta_{d,f}(x)=f(dx)+f(dx+d)+f(d)=0.$ (3)
We have
$\Delta_{d,f}(x)=a{\rm Tr}^{n}_{m}(\Delta_{d,F}(x))+a^{q}{\rm
Tr}^{n}_{m}(\Delta_{d,G}(x))=0.$ (4)
In the following, we shall show that (4) holds if and only if
${\rm Tr}^{n}_{m}(\Delta_{d,F}(x))={\rm Tr}^{n}_{m}(\Delta_{d,G}(x))=0.$
The sufficiency is clear. Let us show the necessity.
Raising (4) to its $q$-th power, we have
$a^{q}{\rm Tr}^{n}_{m}(\Delta_{d,F}(x))+a{\rm Tr}^{n}_{m}(\Delta_{d,G}(x))=0.$
(5)
Adding (4) and (5),
$(a+a^{q}){\rm Tr}^{n}_{m}(\Delta_{d,F}(x))+(a+a^{q}){\rm
Tr}^{n}_{m}(\Delta_{d,G}(x))=0,$
which infers, since $a+a^{q}\neq 0$, that
${\rm Tr}^{n}_{m}(\Delta_{d,F}(x))={\rm Tr}^{n}_{m}(\Delta_{d,G}(x)).$ (6)
Substituting (6) into (4), we can obtain
${\rm Tr}^{n}_{m}(\Delta_{d,F}(x))={\rm Tr}^{n}_{m}(\Delta_{d,G}(x))=0,$
which is exactly the system (2). Therefore, $f(x)$ is APN, if and only if the
system (2) only has trivial solutions $x=0,1$, for any $d\neq 0$. ∎
TABLE II: Known infinite families of quadratic APN polynomials over $\mathbb{F}_{2^{n}}$ ID | Functions | Conditions | Source
---|---|---|---
F1-F2 | $x^{2^{s}+1}+u^{2^{k}-1}x^{2^{ik}+2^{mk+s}}$ | $n=pk$, ${\rm gcd}(k,p)={\rm gcd}(s,pk)=1$, $p\in\\{3,4\\}$, $i=sk~{}{\rm mod}~{}p$, $m=p-i$, $n\geq 12$, $u$ primitive in $\mathbb{F}^{\ast}_{2^{n}}$ | [9]
F3 | $sx^{q+1}+x^{2^{i}+1}+x^{q(2^{i}+1)}+dx^{2^{i}q+1}+d^{q}x^{2^{i}+q}$ | $n=2m$, $q=2^{m}$, ${\rm gcd}(i,m)=1$, $d\in\mathbb{F}_{2^{n}}$, $s\in\mathbb{F}_{2^{n}}\backslash\mathbb{F}_{2^{m}}$, $X^{2^{i}+1}+dX^{2^{i}}+d^{q}X+1$ has no solution $x$ s.t. $x^{q+1}=1$ | [8, 7]
F4 | $x^{3}+a^{-1}{\rm Tr}^{n}_{1}(a^{3}x^{9})$ | $a\neq 0$ | [10]
F5 | $x^{3}+a^{-1}{\rm Tr}^{n}_{3}(a^{3}x^{9}+a^{6}x^{18})$ | $3~{}|~{}n$, $a\neq 0$ | [11]
F6 | $x^{3}+a^{-1}{\rm Tr}^{n}_{3}(a^{6}x^{18}+a^{12}x^{36})$ | $3~{}|~{}n$, $a\neq 0$ | [11]
F7-F9 | $ux^{2^{s}+1}+u^{2^{k}}x^{2^{-k}+2^{k+s}}+vx^{2^{-k}+1}+\omega u^{2^{k}+1}x^{2^{s}+2^{k+s}}$ | $n=3k$, ${\rm gcd}(k,3)={\rm gcd}(s,3k)=1$, $v$, $\omega\in\mathbb{F}_{2^{k}}$, $v\omega\neq 1$, $3~{}|~{}(k+s)$, $u$ primitive in $\mathbb{F}^{\ast}_{2^{n}}$ | [3, 4]
F10 | $cx^{q+1}+dx^{2^{i}+1}+d^{q}x^{q(2^{i}+1)}+\sum^{m-1}_{s=1}\gamma_{s}x^{2^{s}(q+1)}$ | $n=2m$, $q=2^{m}$, ${\rm gcd}(i,m)=1$, $i$, $m$ odd, $\gamma_{s}\in\mathbb{F}_{q}$, $c\notin\mathbb{F}_{q}$, $d$ not a cube | [3]
F11 | $(x+x^{q})^{2^{k}+1}+u^{\prime}(ux+u^{q}x^{q})^{(2^{k}+1)2^{i}}+u(x+x^{q})(ux+u^{q}x^{q})$ | $n=2m$, $m\geq 2$ even, ${\rm gcd}(k,m)=1$, $q=2^{m}$, and $i\geq 2$ even, $u$ primitive in $\mathbb{F}^{\ast}_{2^{n}}$, $u^{\prime}\in\mathbb{F}_{2^{m}}$ not a cube | [26]
F12 | $u(u^{q}x+ux^{q})(x+x^{q})+(u^{q}x+ux^{q})^{2^{2i}+2^{3i}}+\alpha(u^{q}x+ux^{q})^{2^{2i}}(x+x^{q})^{2^{i}}+\beta(x+x^{q})^{2^{i}+1}$ | $n=2m$, $q=2^{m}$, ${\rm gcd}(i,m)=1$, $u$ primitive in $\mathbb{F}^{\ast}_{2^{n}}$, $\alpha$, $\beta\in\mathbb{F}_{2^{m}}$, and $X^{2^{i}+1}+\alpha X+\beta$ has no solution in $\mathbb{F}_{2^{m}}$ | [23]
F13 | $L(x)^{2^{i}}x+L(x)x^{2^{i}}$ | $n=km$, $m\geq 2$, ${\rm gcd}(n,i)=1$, $L(x)=\sum^{k-1}_{j=0}a_{j}x^{2^{jm}}$ satisfies the conditions in Theorem 6.3 of [6] | [6]
F14 | $x^{3}+\omega x^{2^{s}+1}+\omega^{2}x^{3q}+x^{(2^{s}+1)q}$ | $n=2m$, $q=2^{m}$, $m$ odd, $3\nmid m$, $\omega$ primitive in $\mathbb{F}^{\ast}_{2^{2}}$, $s=m-2$, $(m-2)^{-1}~{}{\rm mod}~{}n$ | [12]
F15 | $a{\rm Tr}^{n}_{m}(bx^{3})+a^{q}{\rm Tr}^{n}_{m}(b^{3}x^{9})$ | $n=2m$, $m$ odd, $q=2^{m}$, $a\notin\mathbb{F}_{q}$, $b$ not a cube | new
## 3 Three infinite families of APN functions
We want to find new APN functions of the form (1). In the following two
subsections, the functions $F$ and $G$ were chosen very carefully to satisfy
the conditions characterized in Lemma 2.1. This will yield a new infinite
family of APN quadrinomails, two infinite families of APN hexanomials, and (at
least) two sporadic APN functions CCZ-inequivalent to any other known APN
functions over $\mathbb{F}_{2^{10}}$.
### 3-A F, G are both of Gold type
We need the following two lemmas, which will be used in the proof of Theorem
3.3.
###### Lemma 3.1.
Let $n=2m$ for $m$ odd, $q=2^{m}$. Suppose that for some
$c\in\mathbb{F}_{2^{n}}$ we have
$c^{3}(c+c^{2}+c^{4})^{q}\in\mathbb{F}_{2^{m}}.$
Then $c$ is a cube in $\mathbb{F}_{2^{n}}$.
###### Proof.
Since ${\rm gcd}(3,2^{m}-1)=1$, any element of $\mathbb{F}_{2^{m}}$ is a cube.
In the following, we assume that $c\notin\mathbb{F}_{2^{m}}$. Noting that
$c^{3}(c+c^{2}+c^{4})^{q}=c^{(q+1)+2}+c^{2(q+1)+1}+c^{3(q+1)+q}$, we have
$c^{q+1}(c+c^{q})^{2}+c^{2(q+1)}(c+c^{q})+c^{3(q+1)}(c+c^{q})=0$ by the
assumption that $c^{3}(c+c^{2}+c^{4})^{q}\in\mathbb{F}_{2^{m}}$. Since
$c+c^{q}\neq 0$, we have $c^{q+1}(c+c^{q})+c^{2(q+1)}+c^{3(q+1)}=0$, and hence
$c+c^{q}=c^{q+1}+c^{2(q+1)}$. Note that any nonzero element $c$ of
$\mathbb{F}_{2^{n}}$ has a unique polar decomposition of the form $c=vk$,
where $k^{q+1}=1$, and $v^{q-1}=1$. Substituting $c=vk$ into
$c+c^{q}=c^{q+1}+c^{2(q+1)}$, we have $k+k^{-1}=v+v^{3}$. By assumption that
$c\notin\mathbb{F}_{2^{m}}$, we have $k\neq 1$. Then according to [21, Theorem
7], we have that $k$ is a cube in
$U:=\\{x\in\mathbb{F}_{2^{n}}~{}|~{}x^{q+1}=1\\}$. Therefore, $c=vk$ is a cube
in $\mathbb{F}_{2^{n}}$. ∎
Let $s$ be a positive integer with ${\rm gcd}(s,n)=1$. Let
$x\in\mathbb{F}_{2^{n}}$. It is clear that $x+x^{2^{s}}\neq 0$, if and only if
$x\neq 0,1$. We have the following lemma.
###### Lemma 3.2.
Let $n=2m$ for $m$ odd with ${\rm gcd}(3,m)=1$. Let s be a positive integer
such that $3s\equiv 1~{}{\rm mod}~{}n$. Suppose that for some
$x\in\mathbb{F}_{2^{n}}\backslash\\{0,1\\}$, we have
$\frac{x+x^{2}}{(x+x^{2^{s}})^{2^{2s}-2^{s}+1}}\in\mathbb{F}_{2^{m}}.$
Then $x+x^{2^{s}}$ is a cube.
###### Proof.
Let $d=x+x^{2^{s}}$. Then $d\neq 0$, since $x\neq 0,1$, and ${\rm
gcd}(s,n)=1$. We can express $x+x^{2}=d+d^{2^{s}}+d^{2^{2s}}$. Then
$\frac{x+x^{2}}{(x+x^{2^{s}})^{2^{2s}-2^{s}+1}}=\frac{d+d^{2^{s}}+d^{2^{2s}}}{d^{2^{2s}-2^{s}+1}}=d^{-2^{s}(2^{s}-1)}+d^{-(2^{s}-1)^{2}}+d^{2^{s}-1}=A^{-2^{s}}+A^{-2^{s}+1}+A,$
where $A=d^{2^{s}-1}$. Then the condition of this lemma is equivalent to that
$A^{-2^{s}}+A^{-2^{s}+1}+A+1\in\mathbb{F}_{2^{m}},$ which is exaclty
$\frac{(A+1)^{2^{s}+1}}{A^{2^{s}}}\in\mathbb{F}_{2^{m}}.$
If $A=1$, i.e., $d^{2^{s}-1}=1$, then $d=1$, and hence $x+x^{2^{s}}=1$ is a
cube. In fact, since ${\rm gcd}(2^{s}-1,2^{n}-1)=1$, $g(x)=x^{2^{s}-1}$ is a
permutation of $\mathbb{F}_{2^{n}}$. Then by $g(d)=g(1)=1$, we have $d=1$. If
$A\neq 1$, then there exists some $\alpha\in\mathbb{F}^{\ast}_{2^{m}}$ such
that $A^{2^{s}}=(A+1)^{2^{s}+1}\alpha$. Since $s$ is odd, $3~{}|~{}2^{s}+1$,
we have $A^{2^{s}+1}\alpha$ is a cube, and hence $A^{2^{s}}$ is a cube, that
is, $A$ is a cube. However, note that ${\rm gcd}(3,2^{s}-1)=1$, we have that
$d$ is a cube, when $A=d^{2^{s}-1}$ is. ∎
In the following theorem, we investigate the APN property of the functions
with the form (1) by letting $F(x)=bx^{3}$, and $G(x)=cx^{2^{s}+1}$. This
allows us to find a new infinite family of APN quadrinomials $f(x)=a{\rm
Tr}^{n}_{m}(bx^{3})+a^{q}{\rm Tr}^{n}_{m}(b^{3}x^{9})$, where $b$ is a non-
cube in $\mathbb{F}_{2^{n}}$.
###### Theorem 3.3.
Let $n=2m$ with $m\geq 1$ odd, and $q=2^{m}$. Let $a\in\mathbb{F}_{2^{n}}$,
and $f_{s}(x)=a{\rm Tr}^{n}_{m}(bx^{3})+a^{q}{\rm Tr}^{n}_{m}(cx^{2^{s}+1})$
with $a\notin\mathbb{F}_{q}$, $bc\neq 0$, $s$ odd. Then $f_{s}(x)$ is APN over
$\mathbb{F}_{2^{n}}$, if $s,b,c$ satisfy the following
i) $s=m-2$, $b$ not a cube, $\frac{c^{4}}{b}\in\mathbb{F}_{2^{m}}$; or
ii) $s=(m-2)^{-1}~{}{\rm mod}~{}n$, $b$ not a cube,
$\frac{c^{2^{s}-1}}{b^{2^{2s}}}\in\mathbb{F}_{2^{m}}$; or
iii) $s=3$, $b$ not a cube, $\frac{c}{b^{3}}\in\mathbb{F}_{2^{m}}$; or
iv) ${\rm gcd}(3,m)=1$, $3s\equiv 1{\rm~{}mod~{}}n$, $b$ not a cube,
$\frac{c}{b^{2^{2s}-2^{s}+1}}\in\mathbb{F}_{2^{m}}$; or
v) $s=m$, $b$ not a cube, $c\notin\mathbb{F}_{2^{m}};$ or
vi) $s=m+2$, $b$ not a cube, $bc\in\mathbb{F}_{2^{m}}$; or
vii) $s=n-1$, $\frac{c^{2}}{b}\notin\mathbb{F}_{2^{m}}$.
###### Proof.
Let $F(x)=bx^{3}$, $G(x)=cx^{2^{s}+1}$. Then
$\Delta_{d,F}(x)=d^{3}b(x^{2}+x),~{}{\rm
and}~{}\Delta_{d,G}(x)=d^{2^{s}+1}c(x^{2^{s}}+x).$
According to Lemma 2.1, proving $f_{s}(x)$ is an APN function over
$\mathbb{F}_{2^{n}}$ is equivalent to showing that the system:
$\Delta_{d,F}(x)\in\mathbb{F}_{2^{m}}$, and
$\Delta_{d,G}(x)\in\mathbb{F}_{2^{m}}$ can only has trivial solutions $x=0,1$
for any $d\neq 0$. Assume, to the contrary, that $f_{s}(x)$ is not an APN
function, when $s,b,c$ satisfy the conditions of one item in this theorem.
Then the following system
$\displaystyle\begin{cases}d^{3}b(x^{2}+x)=\alpha,&\\\
d^{2^{s}+1}c(x^{2^{s}}+x)=\beta.&\end{cases}$ (7)
has a non-trivial solution $x\notin\mathbb{F}_{2}$ for some $d\neq 0$, where
$\alpha,~{}\beta\in\mathbb{F}_{2^{m}}$ with $\alpha\neq 0$.
Since $m$ is odd, ${\rm gcd}(3,2^{m}-1)=1$, we have that $\alpha=e^{3}$ for
some $e\in\mathbb{F}^{\ast}_{2^{n}}.$ Dividing both sides of the first
equation in (7) by $e^{3}$, we obtain that $(d/e)^{3}b(x^{2}+x)=1$. Dividing
both sides of the second equation in (7) by $e^{2^{s}+1}$, we have
$(d/e)^{2^{s}+1}c(x^{2^{s}}+x)=\beta e^{-(2^{s}+1)}$. Since $s$ is odd, we
have $3~{}|~{}2^{s}+1$, and $e^{2^{s}+1}\in\mathbb{F}_{2^{m}}.$ Therefore, the
system (7) has a non-trivial solution $x\notin\\{0,1\\}$ if and only if the
system
$\displaystyle\begin{cases}d^{3}b(x^{2}+x)=1,&\\\
d^{2^{s}+1}c(x^{2^{s}}+x)=\beta.&\end{cases}$ (8)
has a solution for some $d\in\mathbb{F}^{\ast}_{2^{n}}$ and
$\beta\in\mathbb{F}_{2^{m}}.$
$i)$ $s=m-2$, $b$ is a non-cube in $\mathbb{F}_{2^{n}}$ and
$\frac{c^{4}}{b}\in\mathbb{F}^{\ast}_{2^{m}}$.
Raising the second equation in (8) to its fourth power, we have
$d^{q+4}c^{4}(x^{q}+x^{4})=\beta^{4}$. From the first equation, we have
$d^{3}=\frac{1}{b(x^{2}+x)}$. Substituting this relation into the previous
equation, we have
$d^{q+1}\frac{c^{4}}{b}\frac{x^{q}+x^{4}}{x^{2}+x}\in\mathbb{F}_{2^{m}}$.
Since $d^{q+1}\in\mathbb{F}^{\ast}_{2^{m}}$, and
$\frac{c^{4}}{b}\in\mathbb{F}^{\ast}_{2^{m}}$ by assumption, we have
$\frac{x^{q}+x^{4}}{x+x^{2}}\in\mathbb{F}_{2^{m}}$. By [12, Lemma 1], we have
$x+x^{2}$ is a cube in $\mathbb{F}_{2^{n}}$, and hence $b$ is a cube by
$d^{3}b(x^{2}+x)=1$, a contradiction to the assumption that $b$ is a non-cube.
$ii)$ $s=(m-2)^{-1}~{}{\rm mod}~{}n$, $b$ is a non-cube in
$\mathbb{F}^{\ast}_{2^{n}}$ and
$\frac{c^{2^{s}-1}}{b^{2^{2s}}}\in\mathbb{F}^{\ast}_{2^{m}}$.
It can be seen from the proof of Theorem 2 in [12] that the critical
conditions ensuring the APN-ness of this $f_{s}(x)$ are exactly that $b$ is a
non-cube in $\mathbb{F}_{2^{n}}$ and
$\frac{c^{2^{s}-1}}{b^{2^{2s}}}\in\mathbb{F}^{\ast}_{2^{m}}$. We invite the
readers to check it, and we omit the arguments here.
$iii)$ $s=3$, $b$ is a non-cube in $\mathbb{F}_{2^{n}}$ and
$\frac{c}{b^{3}}\in\mathbb{F}^{\ast}_{2^{m}}.$
It can be seen that in this case (8) becomes
$\displaystyle\begin{cases}d^{3}b(x^{2}+x)=1,&\\\
d^{9}c(x^{8}+x)=\beta.&\end{cases}$
Substituting $d^{3}=\frac{1}{b(x+x^{2})}$ into the second equation of the
above system, we have
$\displaystyle\frac{c}{b^{3}}\cdot\frac{x+x^{8}}{(x+x^{2})^{3}}=\beta,$
which infers that $\frac{x+x^{8}}{(x+x^{2})^{3}}\in\mathbb{F}_{2^{m}}$, since
$\frac{c}{b^{3}}\in\mathbb{F}^{\ast}_{2^{m}}$ by assumption. It implies that
$(x+x^{2})^{3}(x+x^{8})^{q}\in\mathbb{F}_{2^{m}}$. Denoting $e=x+x^{2}$, we
have $x+x^{8}=e+e^{2}+e^{4}$, and hence
$e^{3}(e+e^{2}+e^{4})^{q}\in\mathbb{F}_{2^{m}}$. Now, according to Lemma 3.1,
$e=x+x^{2}$ is a cube. Then $b$ is a cube by $d^{3}b(x+x^{2})=1$, which
contradicts to the assumption that $b$ is a non-cube.
$iv)$ ${\rm gcd}(3,m)=1$, $3s\equiv 1{\rm~{}mod~{}}n$, $b$ is a non-cube in
$\mathbb{F}_{2^{n}}$ and
$\frac{c^{2^{2s}-2^{s}+1}}{b}\in\mathbb{F}^{\ast}_{2^{m}}$.
Since ${\rm gcd}(2^{s}-1,2^{n}-1)=2^{{\rm gcd}(s,n)}-1=1$, we have that
$x+x^{2^{s}}\neq 0$, when $x\neq 0,1$. Then (8) becomes
$\displaystyle\begin{cases}d^{2^{3s}+1}b(x+x^{2})=1,&\\\
d^{2^{s}+1}c(x+x^{2^{s}})=\beta,&\end{cases}$
where $\beta\in\mathbb{F}_{2^{m}}$ with $\beta\neq 0$, since $x+x^{2^{s}}\neq
0$. By the second equation, we have
$d^{2^{s}+1}=\frac{\beta}{c(x+x^{2^{s}})}$. Substituting this relation into
the first equation, noting that $2^{3s}+1=(2^{s}+1)(2^{2s}-2^{s}+1)$, we have
$\displaystyle\frac{b}{c^{2^{2s}-2^{s}+1}}\cdot\frac{x+x^{2}}{(x+x^{2^{s}})^{2^{2s}-2^{s}+1}}\in\mathbb{F}_{2^{m}},$
which infers, since $\frac{b}{c^{2^{2s}-2^{s}+1}}\in\mathbb{F}^{\ast}_{2^{m}}$
by assumption, that
$\displaystyle\frac{x+x^{2}}{(x+x^{2^{s}})^{2^{2s}-2^{s}+1}}\in\mathbb{F}^{\ast}_{2^{m}}.$
(9)
Now, by the assumption that $b$ is a non-cube in $\mathbb{F}_{2^{n}}$ and
$\frac{c^{2^{2s}-2^{s}+1}}{b}\in\mathbb{F}^{\ast}_{2^{m}}$, we have that $c$
is a non-cube. On the other hand, by (9) and Lemma 3.2, we have that
$x+x^{2^{s}}$ is a cube, which infers that $c$ is a cube from the second
equation $d^{2^{s}+1}c(x+x^{2^{s}})=\beta$ of the above system, a
contradiction.
$v)$ $s=m$, $b$ is a non-cube in $\mathbb{F}_{2^{n}}$, and
$c\notin\mathbb{F}_{2^{m}}.$
It can be seen that (8) becomes
$\displaystyle\begin{cases}d^{3}b(x+x^{2})=1,&\\\
d^{2^{m}+1}c(x+x^{2^{m}})=\beta,&\end{cases}$
where $\beta\in\mathbb{F}_{2^{m}}$. Since $c\notin\mathbb{F}_{2^{m}}$, and
$d^{2^{m}+1}\in\mathbb{F}^{\ast}_{2^{m}},$ $x+x^{2^{m}}\in\mathbb{F}_{2^{m}}$
for any $d\neq 0$, $x\in\mathbb{F}_{2^{n}}$, by the second equation, we have
$\beta$ must equal to zero, which infers that $x\in\mathbb{F}_{2^{m}}$. Then
by the fact that any element of $\mathbb{F}_{2^{m}}$ is a cube, we have
$d^{3}(x+x^{2})$ is a cube in $\mathbb{F}^{\ast}_{2^{n}}$, which implies that
$b$ is a cube in $\mathbb{F}^{\ast}_{2^{n}}$, a contradiction to the
assumption that $b$ is a non-cube.
$vi)$ $s=m+2$, $b$ is a non-cube in $\mathbb{F}_{2^{n}}$ and
$bc\in\mathbb{F}^{\ast}_{2^{m}}$. It can be seen (8) becomes
$\displaystyle\begin{cases}d^{3}b(x+x^{2})=1,&\\\
d^{4(q+1)-3}c(x+x^{4q})=\beta,&\end{cases}$
where $\beta\in\mathbb{F}_{2^{m}}$ with $\beta\neq 0$ since $x+x^{4q}\neq 0$
when $x\neq 0,1$. Since $d^{3}b(x+x^{2})=1$, we have
$d^{3}=\frac{1}{b(x+x^{2})}$. Substituting this relation into the second
equation, we have
$\displaystyle d^{4(q+1)}bc(x+x^{2})(x+x^{4q})=\beta.$
Then by the assumption that $bc\in\mathbb{F}^{\ast}_{2^{m}}$, we have
$(x+x^{2})(x+x^{4q})\in\mathbb{F}_{2^{m}}$. According to [12, Lemma 1], we
have $x+x^{2}\neq 0$ is a cube, which infers that $b$ is a cube by
$d^{3}b(x+x^{2})=1$, a contradiction to the assumption that $b$ is a non-cube.
$vii)$ $s=n-1$, $\frac{c^{2}}{b}\notin\mathbb{F}_{2^{m}}$.
Since ${\rm gcd}(2^{s}-1,2^{n}-1)=2^{{\rm gcd}(s,n)}-1=1$, we have that
$x+x^{2^{s}}\neq 0$, if $x\neq 0,1$. It can be seen that (8) becomes
$\displaystyle\begin{cases}d^{3}b(x+x^{2})=1,&\\\
d^{2^{s}+1}c(x+x^{2^{s}})=\beta,&\end{cases}$
where $\beta\in\mathbb{F}_{2^{m}}$ with $\beta\neq 0$. Squaring the second
equation, we have $d^{3}c^{2}(x+x^{2})=\beta^{2}$. Comparing with the first
equation, we have $\frac{c^{2}}{b}=\beta^{2}\in\mathbb{F}_{2^{m}}$, which
contradicts with the assumption that
$\frac{c^{2}}{b}\notin\mathbb{F}_{2^{m}}.$ ∎
###### Remark 3.4.
Code isomorphism tests described in Section 2 suggest that all the polynomials
from the same item of Theorem 3.3 are all CCZ-equivalent; the APN function
$x^{3}+\omega x^{2^{s}+1}+\omega^{2}x^{3q}+x^{(2^{s}+1)q}$ discovered in [12]
is CCZ-equivalent to all the functions in i), ii), respectively, for $s=m-2$,
and $s=(m-2)^{-1}~{}{\rm mod}~{}n$, if ${\rm gcd}(3,m)=1$; the polynomials
$f_{s}(x)$ for $s=m+2$ in vi) are equivalent to the ones for $s=m-2$ in i);
the polynomials $f_{s}(x)$ for $s=m$ in v) are equivalent to some functions in
family F10 from Table II, see also the arguments in Remark 3.7 below; the
polynomial $f_{s}(x)$ for $s=n-1$ in vii) is CCZ-equivalent to $x^{3}$.
The remaining value of $s=3$ in iii) yields APN quadrinomials $f_{3}(x)$,
which are CCZ-inequivalent to any currently known APN function over
$\mathbb{F}_{2^{10}}$. By the arguments above that all the polynomials in the
same item are all CCZ-equivalent, we only take a representative of iii). We
let $f_{3}(x)=\omega{\rm Tr}^{n}_{m}(bx^{3})+\omega^{2}{\rm
Tr}^{n}_{m}(b^{3}x^{9})$, where $b$ is a non-cube,
$\omega\in\mathbb{F}_{2^{2}}\backslash\mathbb{F}_{2}$. We use this $f_{3}(x)$
to compare against representatives from all the known infinite families
including $f_{s}(x)$, $s=m-2$, $(m-2)^{-1}~{}{\rm mod}~{}n$ in i), ii) which
are essentially due to Budaghyan, Helleseth, and Kaleyski ([12]). Note that,
Budaghyan et al. had presented a table listing all the representatives, except
family F12, of all the known CCZ-inequivalent APN functions over
$\mathbb{F}_{2^{10}}$, see Table III of [12]. To complete the work of code
isomorphism test, we have to find all the representatives of F12 over
$\mathbb{F}_{2^{10}}$. Thanks to the nice work [20], we can obtain these
representatives. In fact, let $\gamma$ be a primitive element in
$\mathbb{F}^{\ast}_{2^{5}}$, according to [20, Theorem 4.5], there are exactly
6 of CCZ-inequivalent Taniguchi APN functions from F12: $i=1$, take
$\alpha=1$, $\beta=1,~{}\gamma^{7},~{}\gamma^{11}$; $i=2$, take $\alpha=1$,
$\beta=1,~{}\gamma^{3},~{}\gamma^{15}$. The notations $i,~{}\alpha,~{}\beta$
used here are the same as the ones used in family F12 of Table II.
###### Remark 3.5.
Let $n=2m$ with $m$ odd, and ${\rm gcd}(m,3)=1$. Let $q=2^{m}$. Let $z$ be a
primitive element in $\mathbb{F}^{\ast}_{2^{n}}$, and
$\omega=z^{\frac{2^{n}-1}{3}}$. Then $\omega$ is a primitive element in
$\mathbb{F}_{2^{2}}$. Let $s=m-2$ or $(m-2)^{-1}{\rm~{}mod}~{}n$. Then
$g_{s}(x)=x^{3}+\omega x^{2^{s}+1}+\omega^{2}x^{3q}+x^{(2^{s}+1)q}$ is an APN
function ([12]). It can be seen that $g_{s}(x)$ can be covered by our theorem.
In fact, noting that $\omega^{2^{s}}=\omega^{2}$ for any odd $s$,
$g_{s}(x)=\omega{\rm Tr}^{n}_{m}(\omega^{2}x^{3})+\omega^{2}{\rm
Tr}^{n}_{m}(\omega^{2}x^{2^{s}+1})=a{\rm Tr}^{n}_{m}(bx^{3})+a^{q}{\rm
Tr}^{n}_{m}(cx^{2^{s}+1})$, where $a=\omega,b=c=\omega^{2}$. It is clear that
$a+a^{q}=1\neq 0$, and $b=\omega$ is a non-cube since ${\rm gcd}(m,3)=1$, and
$\frac{c^{4}}{b}=1=\frac{c^{2^{t}-1}}{b^{2^{2t}}}$, where
$t=(m-2)^{-1}{\rm~{}mod}~{}n$. Then by $i)$, $ii)$ of the above theorem, we
have that $g_{s}(x)$ is APN over $\mathbb{F}_{2^{n}}$, for $s=m-2$, and
$(m-2)^{-1}{\rm~{}mod}~{}n$, respectively.
###### Remark 3.6.
Let $n=2m$ with $m$ odd. Let us investigate the APN property of $f_{m-2}(x)$
further. A pair ($b,c$) is said to satisfy property $\mathbf{P}_{m-2}$, if $b$
is a cube in $\mathbb{F}^{\ast}_{2^{n}}$, and $c\in\mathbb{F}^{\ast}_{2^{n}}$
such that the following assertion holds:
For any $x\in\mathbb{F}_{2^{n}}$ with $x\neq 0,1$, $x+x^{2}$ is a non-cube in
$\mathbb{F}_{2^{n}}$, if
$\frac{c^{4}}{b}\cdot\frac{x^{q}+x^{4}}{x+x^{2}}\in\mathbb{F}_{2^{m}}$.
Then $f_{m-2}(x)$ is APN over $\mathbb{F}_{2^{n}}$ for these $b$, $c$. In
fact, this assertion can be seen from the proof of $i)$ in the above theorem.
With the help of computer, we find that when $m=5$, $7$, there exist a lot of
pairs ($b,c$) satisfying $\mathbf{P}_{m-2}$. More precisely, let $m=5$ or $7$,
$z$ be a primitive element in $\mathbb{F}^{\ast}_{2^{2m}}$,
$j=\frac{(2^{m}+1)}{3}$, and $U=\\{(z^{j})^{i}~{}|~{}{\rm gcd}(3,i)=1,~{}1\leq
i\leq 2^{n}-1\\}$. Then any pair ($b,c$) with $b\neq 0$ a cube, and
$\frac{c^{4}}{b}\in U$ satisfies $\mathbf{P}_{m-2}$. However, when $m=9$,
$11$, there does not exist such ($b,c$). We therefore propose the following:
Open Problem 1. Does there exist infinite odd integer $m\geq 1$ such that
$\mathbf{P}_{m-2}$ holds?
###### Remark 3.7.
Let $n=2m$ with $m$ odd, and $q=2^{m}$. Let us revisit the function
$f_{m}(x)=a{\rm Tr}^{n}_{m}(bx^{3})+a^{q}{\rm Tr}^{n}_{m}(cx^{2^{m}+1})$
investigated in $v)$. Replacing $bx^{3}$ by $bx^{2^{i}+1}$, we let $f(x)=a{\rm
Tr}^{n}_{m}(bx^{2^{i}+1})+a^{q}{\rm Tr}^{n}_{m}(cx^{2^{m}+1})$, where $i$ is
an odd positive integer with ${\rm gcd}(i,m)=1$. With similar arguments, by
$3~{}|~{}2^{i}+1$ and ${\rm gcd}(i,m)=1$, we can obtain that $f(x)$ is APN, if
$b$ is a non-cube in $\mathbb{F}_{2^{n}}$, and $c\notin\mathbb{F}_{2^{m}}$.
Note that $\frac{1}{a}f(x)=dx^{2^{m}+1}+{\rm Tr}^{n}_{m}(bx^{2^{i}+1})$, where
$d=a^{q-1}(c+c^{q})$ can be chosen as any element in
$\mathbb{F}_{2^{n}}\backslash\mathbb{F}_{2^{m}}$, since
$a,~{}c\notin\mathbb{F}_{q}$, we have that $f(x)$ in fact are exactly the
functions in family F10 up to EA-equivalence. This observation suggests that
it is worthy to finding APN functions with the following form:
$\displaystyle f_{i,s}(x)=a{\rm Tr}^{n}_{m}(bx^{2^{i}+1})+a^{q}{\rm
Tr}^{n}_{m}(cx^{2^{s}+1}),~{}\text{where~{}}a\in\mathbb{F}_{2^{n}}~{}such~{}that~{}a+a^{q}\neq
0,~{}n=2m~{}\text{is a positive integer}.$ (10)
###### Remark 3.8.
It is noted that there does not exist elements satisfying the conditions in
$iv)$. However, we decide to preserve this item, because we feel that the
technique used in the proof may provide some insights for the constructions of
APN functions.
### 3-B F, G are both quadratic binomials
Let us consider more general case. Let $n=2m$ with $m$ a positive integer. Let
$\displaystyle h_{i,s,b,c,g,e}(x)=a{\rm
Tr}^{n}_{m}(bx^{2^{i}+1}+cx^{2^{i+m}+1})+a^{q}{\rm
Tr}^{n}_{m}(gx^{2^{s}+1}+ex^{2^{s+m}+1}),$ (11)
where $a\in\mathbb{F}_{2^{n}}$ such that $a+a^{q}\neq 0$,
$b,c,g,e\in\mathbb{F}_{2^{n}}$.
In this subsection, we want to find APN functions of the form (11). We remark
first that the APN polynomials considered in family F3 can be covered by
$h_{i,s,b,c,g,e}(x)$. In fact, let $i=m$, $b\notin\mathbb{F}_{2^{m}}$, $c=0$,
$g=1$, then (11) becomes
$a^{q-1}(b+b^{q})x^{q+1}+x^{2^{s}+1}+x^{(2^{s}+1)q}+ex^{2^{s}q+1}+e^{q}x^{2^{s}+q}$,
which are exactly the functions in F3, since $a^{q-1}(b+b^{q})$ can be choosen
as any elements in $\mathbb{F}_{2^{n}}\backslash\mathbb{F}_{2^{m}}$.
We can find two infinite families of APN functions with the above form (11),
and computationally prove that they are CCZ-inequivalent to any APN power
functions over $\mathbb{F}_{2^{10}}$, and we can find a new sporadic instance
of APN functions over $\mathbb{F}_{2^{10}}$.
###### Theorem 3.9.
[24] Let $n=2m$, and $a\in\mathbb{F}^{\ast}_{2^{n}}$. Let $t_{1}$ be one
solution in $\mathbb{F}_{2^{n}}$ of $t^{2}+at+1=0$ (if ${\rm
Tr}^{n}_{1}\Big{(}\frac{1}{a^{2}}\Big{)}=0$). Let $f(x)=x^{3}+x+a$, then
$\bullet$ $f$ has no zeros in $\mathbb{F}_{2^{n}}$ if and only if ${\rm
Tr}^{n}_{1}\Big{(}\frac{1}{a^{2}}\Big{)}=0$, and $t_{1}$ is not a cube in
$\mathbb{F}_{2^{n}}$.
$\bullet$ $f$ has three zeros in $\mathbb{F}_{2^{n}}$ if and only if ${\rm
Tr}^{n}_{1}\Big{(}\frac{1}{a^{2}}\Big{)}=0$, and $t_{1}$ is a cube in
$\mathbb{F}_{2^{n}}$.
We need the following theorem, which will be used for generating APN functions
(see Corollary 1). Let $n=2m$ with $m$ being an odd positive integer, and
$q=2^{m}$. Let $x\in\mathbb{F}_{2^{n}}$ with $x\neq 0,1$. Then fix the
following notations for this given element $x$.
$\displaystyle r:=x^{q+1};~{}h:=x+x^{q};~{}c:=x+x^{2};$ $\displaystyle
D:=A(A^{q+1}+B^{q+1});~{}H:=A^{2}(A^{q}B^{3}+AB^{3q}+B^{2+2q}),$
where $A,B$ are some elements determined by $x$. By a routine work, we have
that
$\displaystyle h+h^{2}=c+c^{q}.$
The following result can not only give rise to APN functions of the form (11)
but can also yield Budaghyan-Carlet APN hexanomials (family F3), and hence it
has its own importance and we state it as a theorem. The proof can be seen in
the appendix.
###### Theorem 3.10.
Let $n=2m$ with $m$ being an odd positive integer. Let $x$ be any given
element in $\mathbb{F}_{2^{n}}\backslash\\{0,1\\}$. Use the notations given as
above. Let
$\displaystyle f(y)=Ay^{3}+By^{2}+B^{q}y+A^{q}=0.$ (12)
Then equation (12) has no solutions in $\mathbb{F}_{2^{n}}$, if A, B, c
satisfy
1) $A=c^{2-2q}(h+c+c^{2})$, $B=c+c^{2}$, and $c=x+x^{2}$ is a non-cube in
$\mathbb{F}_{2^{n}}$; or
2) $A=\frac{h+c+c^{2}}{c^{q}}$, $B=1+c$, and $c=x+x^{2}$ is a non-cube in
$\mathbb{F}_{2^{n}}$.
###### Remark 3.11.
Let $n=2m$, and $q=2^{m}$. Recall first that the condition needed in family F3
is that
$\displaystyle y^{2^{i}+1}+dy^{2^{i}}+d^{q}y+1=0$ (13)
has no solutions in $U=\\{x\in\mathbb{F}_{2^{n}}~{}|~{}x^{q+1}=1\\}$. Here $i$
is a positive integer with ${\rm gcd}(i,m)=1$. When $i=1$, this condition is
exactly that $y^{3}+dy^{2}+d^{q}y+1=0$ has no solutions in $U$.
With the same notations as in Theorem 3.10. Let $A$ be the elements given in
1) or 2). Let
$\Gamma=\\{A\in\mathbb{F}^{\ast}_{2^{m}}~{}|~{}x\in\mathbb{F}_{2^{n}}\backslash\mathbb{F}_{2^{m}},~{}c=x+x^{2}~{}\text{not
cube}\\}$. Numerical experiments suggest that $\Gamma$ is always nonempty for
any odd $m$. This can yield Budaghyan-Carlet APN functions in family F3. In
fact, let $A\in\Gamma$, then (12) becomes
$\displaystyle y^{3}+dy^{2}+d^{q}y+1=0,~{}d=\frac{B}{A}.$
According to Theorem 3.10, the above equation has no solutions in
$\mathbb{F}_{2^{n}}$. Therefore, this theorem can be used to yield APN
functions in family F3. It is noted that the existence of the coefficients $d$
such that the equation (13) has no solutions in $U$ (or $\mathbb{F}_{2^{n}}$)
for a given positive integer $i$ had also been studied in [2, 5].We expect
that $\Gamma$ does indeed empty for any odd positive integer $m$, and hence
propose the following:
Open problem 2. Let $n=2m$ with $m$ odd. Show that $\Gamma$ is always
nonempty.
It is also interesting and important to consider the following question.
Open problem 3. Let $n=2m$ with $m$ a positive integer, $q=2^{m}$. Let $i$ be
a positive with ${\rm gcd}(m,i)=1$. Find more exponents $i$, and elements
$A,B$ such that the following equation has no solutions in
$\mathbb{F}_{2^{n}}$.
$\displaystyle Ay^{2^{i}+1}+By^{2^{i}}+B^{q}y+A^{q}=0.$
In the following, we investigate the APN property of the functions with the
form (11) by letting $i=1,c=0$. We does indeed find two infinite families of
APN functions. But, astonishingly enough, the function obtained happened to be
CCZ-equivalent to some functions in family F12 with a completely different
from that of Taniguchi.
###### Corollary 1.
Let $n=2m$ be a positive integer with $m$ odd, and $q=2^{m}$. Let
$h_{s}(x)=a{\rm Tr}^{n}_{m}(bx^{3})+a^{q}{\rm
Tr}^{n}_{m}(gx^{2^{s}+1}+ex^{2^{s+m}+1})$ with $a\notin\mathbb{F}_{q}$,
$bge\neq 0$. Then $h_{s}(x)$ is APN over $\mathbb{F}_{2^{n}}$, if $s,b,g,e$
satisfy
$\displaystyle 1)$
$\displaystyle~{}~{}~{}s=2,~{}b{\rm~{}is~{}not~{}a~{}cube},~{}g=1,~{}e=\frac{1}{b^{2q-2}};{~{}~{}\rm
or}$ $\displaystyle 2)$
$\displaystyle~{}~{}~{}s=2,~{}b{\rm~{}is~{}not~{}a~{}cube},~{}g=e=b.$
###### Proof.
1) $s=2$, $b$ is not a cube, $g=1$, $e=\frac{1}{b^{2q-2}}$.
Let $F(x)=bx^{3}$, $G(x)=x^{2^{s}+1}+ex^{2^{s+m}+1}$. Then we have
$\displaystyle\Delta_{d,F}=d^{3}b(x+x^{2}),~{}\Delta_{d,G}=d^{2^{s}+1}(x+x^{2^{s}})+d^{2^{s+m}+1}e(x+x^{2^{s+m}}).$
According to Lemma 2.1, we have that $h_{s}(x)$ is APN if the following system
$\displaystyle\begin{cases}d^{3}b(x+x^{2})=\alpha&\\\
d^{2^{s}+1}(x+x^{2^{s}})+d^{2^{s+m}+1}e(x+x^{2^{s+m}})=\beta&\end{cases}$
only has $x=0,1$ as its solutions for any $d\neq 0$, where $\alpha$,
$\beta\in\mathbb{F}_{2^{m}}.$ Assume, to the contrary, that there exists some
$d\neq 0$, $x\neq 0,1$ such that the above system holds. Now let $s=2$, $b$ is
a non-cube, $e=\frac{1}{b^{2q-2}}$. Then $\alpha\neq 0$,
$b=\frac{\alpha}{d^{3}(x+x^{2})}$, $e=b^{-(2q-2)}=d^{6q-6}(x+x^{2})^{2q-2}$
(note that $\alpha^{2q-2}=1$). Substituting it into the second equation of the
above system, we have
$\displaystyle d^{5}(x+x^{4})+d^{10q-5}(x+x^{2})^{2q-2}(x+x^{4q})=\beta,$
which is equivalent to
$\displaystyle
d^{5}(x+x^{4})+d^{10q-5}(x+x^{2})^{2q-2}(x+x^{4q})+\Big{(}d^{5}(x+x^{4})+d^{10q-5}(x+x^{2})^{2q-2}(x+x^{4q})\Big{)}^{q}=0.$
(14)
Let $u=d^{5}$. Then the above equation becomes
$\displaystyle
u(x+x^{4})+u^{2q-1}(x+x^{2})^{2q-2}(x+x^{4q})+\Big{(}u(x+x^{4})+u^{2q-1}(x+x^{2})^{2q-2}(x+x^{4q})\Big{)}^{q}=0.$
(15)
Note that any nonzero element $u$ of $\mathbb{F}_{2^{n}}$ has a unique polar
decomposition of the form $u=vk$, where $v^{q+1}=1$, and $k^{q-1}=1$.
Substituting $u=vk$ into (15), then (15) can be reduced as
$\displaystyle
v(x+x^{4})+v^{2q-1}(x+x^{2})^{2q-2}(x+x^{4q})+\Big{(}v(x+x^{4})+v^{2q-1}(x+x^{2})^{2q-2}(x+x^{4q})\Big{)}^{q}=0.$
Multiplying both sides by $v^{3}$ of the above equation, by the fact that
$v^{q}=v^{-1}$, we have
$\displaystyle Ay^{3}+By^{2}+B^{q}y+A^{q}=0,$
where $y=v^{2}\in\mathbb{F}_{2^{n}}$, and $A$, $B$ are given in 1) of Theorem
3.10. Now, according to 1) of Theorem 3.10, we obtian that the element
$x+x^{2}$ is a cube, and hence $b$ is a cube from the first equation
$d^{3}b(x+x^{2})=\alpha$ of the system, since
$\alpha\in\mathbb{F}^{\ast}_{2^{m}}$ is a cube. This derives a contradiction
to the assumption that $b$ is a non-cube.
2) $s=2$, $b$ is not a cube, $g=e=b$.
Let $F(x)=bx^{3}$ and $G(x)=bx^{5}+bx^{4q+1}$. We have
$\displaystyle\Delta_{d,F}(x)=d^{3}b(x+x^{2})\hskip 5.69046pt{\rm and}\hskip
5.69046pt\Delta_{d,G}(x)=d^{5}b(x+x^{4})+d^{4q+1}b(x+x^{4q}).$
By Lemma 2.1, $h_{s}(x)$ is APN if and only if the following system
$\displaystyle\begin{cases}d^{3}b(x+x^{2})=\alpha\\\
d^{5}b(x+x^{4})+d^{4q+1}b(x+x^{4q})=\beta\end{cases}$
only has trivial solutions $x\in\mathbb{F}_{2}$ for any
$d\in\mathbb{F}_{2^{n}}^{*}$ and $\alpha,\beta\in\mathbb{F}_{2^{m}}$. Assume
now that there exist some $d\in\mathbb{F}_{2^{n}}^{*}$,
$\alpha\in\mathbb{F}_{2^{m}}$, $\beta\in\mathbb{F}_{2^{m}}$ such that the
system has non-trivial solutions
$x\in\mathbb{F}_{2^{n}}\backslash\mathbb{F}_{2}$. Then $\alpha\neq 0$. By the
first equation, we have $b=\frac{\alpha}{d^{3}(x+x^{2})}$. Substituting this
relation into the second equation, we have
$\displaystyle\frac{d^{2}(x+x^{4})}{x+x^{2}}+\frac{d^{4q-2}(x+x^{4q})}{x+x^{2}}=\frac{\beta}{\alpha},$
which implies that
$\displaystyle\frac{d^{2}(x+x^{4})}{x+x^{2}}+\frac{d^{4q-2}(x+x^{4q})}{x+x^{2}}+\bigg{(}\frac{d^{2}(x+x^{4})}{x+x^{2}}+\frac{d^{4q-2}(x+x^{4q})}{x+x^{2}}\bigg{)}^{q}=0,$
since $\alpha,~{}\beta\in\mathbb{F}_{2^{m}}$. Let $\mu=d^{2}$. We have
$\displaystyle\frac{\mu(x+x^{4})}{x+x^{2}}+\frac{\mu^{2q-1}(x+x^{4q})}{x+x^{2}}+\bigg{(}\frac{\mu(x+x^{4})}{x+x^{2}}+\frac{\mu^{2q-1}(x+x^{4q})}{x+x^{2}}\bigg{)}^{q}=0.$
(16)
To complete the proof, it suffices to show that $x+x^{2}$ is a cube of
$\mathbb{F}_{2^{n}}$, which will derive that $b$ is a cube from the first
equation of the above system and this will yield a contradiction to the
assumption that $b$ is a non-cube. Let $\mu=\nu k$, where $\nu^{q+1}=1$ and
$k\in\mathbb{F}_{2^{m}}^{*}$, and substitute $\mu=\nu k$ into (16), we have
$\displaystyle\frac{\nu(x+x^{4})}{x+x^{2}}+\frac{\nu^{2q-1}(x+x^{4q})}{x+x^{2}}+\bigg{(}\frac{\nu(x+x^{4})}{x+x^{2}}+\frac{\nu^{2q-1}(x+x^{4q})}{x+x^{2}}\bigg{)}^{q}=0.$
Multiplying both sides of the above equation by $\nu^{3}$, we have
$\displaystyle Ay^{3}+By^{2}+B^{q}y+A^{q}=0,$
where $y=\nu^{2}$, $A=\Big{(}\frac{x+x^{4q}}{x+x^{2}}\Big{)}^{q}$ and
$B=\frac{x+x^{4}}{x+x^{2}}=1+x+x^{2}$. According to 2) of Theorem 3.10,
$x+x^{2}$ is a cube in $\mathbb{F}_{2^{n}}$, otherwise, the above equation has
no solutions in $\mathbb{F}_{2^{n}}$. ∎
Example 1. Besides the two infinite classes of APN functions presented in
Corollary 1, we can also find a new instance of APN functions over
$\mathbb{F}_{2^{10}}$ CCZ-inequivalent to any other known APN functions. Let
$z$ be a primitive element in $\mathbb{F}^{\ast}_{2^{10}}$. Then
$\displaystyle h_{s}(x)=a{\rm Tr}^{n}_{m}(bx^{3})+a^{q}{\rm
Tr}^{n}_{m}(gx^{5}+ex^{4q+1})$
is an APN function over $\mathbb{F}_{2^{10}}$, where $b=1$, $g=z$,
$e=z^{369}$.
TABLE III: All Known CCZ-inequivalent APN functions over $\mathbb{F}_{2^{10}}$, $q=2^{5}$ Function | Conditions | Family
---|---|---
$x^{2^{i}+1}$ | $i=1,3$ | Gold
$x^{57}$ | $-$ | Kasami
$x^{339}$ | $-$ | Dobbertin
$x^{6}+x^{33}+\alpha^{31}x^{192}$ | $\alpha$ primitive in $\mathbb{F}^{\ast}_{2^{10}}$ | F3
$x^{33}+x^{72}+\alpha^{31}x^{258}$ | $\alpha$ primitive in $\mathbb{F}^{\ast}_{2^{10}}$ | F3
$x^{3}+{\rm Tr}^{10}_{1}(x^{9})$ | $-$ | F4
$x^{3}+\alpha^{-1}{\rm Tr}^{10}_{1}(\alpha^{3}x^{9})$ | $\alpha$ primitive in $\mathbb{F}^{\ast}_{2^{10}}$ | F4
| $u(u^{q}x+ux^{q})(x+x^{q})+$
---
$(u^{q}x+ux^{q})^{2^{2i}+2^{3i}}+$
$\alpha(u^{q}x+ux^{q})^{2^{2i}}(x+x^{q})^{2^{i}}+$
$\beta(x+x^{q})^{2^{i}+1}$
| $u$ primitive in $\mathbb{F}^{\ast}_{2^{10}}$,
---
$z$ primitive in $\mathbb{F}^{\ast}_{2^{5}}$,
$i=1$, $\alpha=1$, $\beta=1,z^{7},z^{11}$;
$i=2$, $\alpha=1$, $\beta=1,z^{3},z^{15}$
F12
$B(x)=x^{3}+\alpha^{341}x^{36}$ | $-$ | sporadic, see [17]
| $x^{3}+\omega x^{2^{s}+1}+$$\omega^{2}x^{3q}+x^{(2^{s}+1)q}$
---
| $s=3,7,$ $\omega$ primitive in $\mathbb{F}^{\ast}_{2^{2}}$
---
F14
| $\alpha{\rm Tr}^{n}_{m}(\alpha x^{3})+\alpha^{q}{\rm
Tr}^{n}_{m}(\alpha^{3}x^{9})$
---
| $\alpha$ primitive in $\mathbb{F}^{\ast}_{2^{10}}$
---
F15
| $\alpha{\rm Tr}^{n}_{m}(x^{3})+\alpha^{q}{\rm Tr}^{n}_{m}(\alpha^{11}x^{9})$
---
| $\alpha$ primitive in $\mathbb{F}^{\ast}_{2^{10}}$
---
| sporadic, see
---
Remark 3.6
| $\alpha{\rm Tr}^{n}_{m}(x^{3})+\alpha^{q}{\rm Tr}^{n}_{m}(\alpha
x^{5}+\alpha^{369}x^{4q+1})$
---
| $\alpha$ primitive in $\mathbb{F}^{\ast}_{2^{10}}$
---
| sporadic, see
---
Example 1
## 4 Conclusions
Let $n=2m$, and $q=2^{m}$. We studied a class of quadratic functions with the
form $f(x)=a{\rm Tr}^{n}_{m}(F(x))+a^{q}{\rm Tr}^{n}_{m}(G(x))$, where $F$,
$G$ are quadratic functions. We found a new infinite family of APN
quadrinomials over $\mathbb{F}_{2^{n}}$, $a\in\mathbb{F}_{2^{n}}$, $n=2m$ with
$m$ odd as follows.
$\displaystyle f_{1}(x)=a{\rm Tr}^{n}_{m}(bx^{3})+a^{q}{\rm
Tr}^{n}_{m}(b^{3}x^{9}),~{}b\text{~{}not~{}a~{}cube},~{}a\notin\mathbb{F}_{q}.$
We generalized the two infinite families of APN functions obtained in [12] to
a broader condition on $m$, that is, the assumption that ${\rm gcd}(3,m)=1$
needed in [12] can be removed, up to CCZ-equivalence. We also found two
infinite families of APN functions over $\mathbb{F}_{2^{2m}}$ for odd $m$,
which turned out to be in family F12, that is, the the Taniguchi APN functions
when $m=5$, as follows.
$\displaystyle f_{2}(x)=a{\rm Tr}^{n}_{m}(bx^{3})+a^{q}{\rm
Tr}^{n}_{m}(x^{5}+\frac{1}{b^{2q-2}}x^{4q+1}),~{}b\text{~{}not~{}a~{}cube},~{}a\in\mathbb{F}_{2^{n}}\backslash\mathbb{F}_{2^{m}},$
and
$\displaystyle f_{3}(x)=a{\rm Tr}^{n}_{m}(bx^{3})+a^{q}{\rm
Tr}^{n}_{m}(bx^{5}+bx^{4q+1}),~{}b\text{~{}not~{}a~{}cube},~{}a\in\mathbb{F}_{2^{n}}\backslash\mathbb{F}_{2^{m}}.$
Code isomorphism tests showed that $f_{2}$ and $f_{3}$ are CCZ-inequivalent to
each other over $\mathbb{F}_{2^{10}}$. We found two new instances of APN
functions over $\mathbb{F}_{2^{10}}$. We also proposed three open problems,
and we cordially invite the readers to attack these open problems.
## References
* [1] T. Beth., C. Ding., On almost perfect nonlinear permutations, Workshop on the Theory and Application of Cryptographic Techniques, Springer, pp. 65-76, 1993.
* [2] A. W. Bluher., On existence of Budaghyan-Carlet APN hexanomials, Finite fields and their applications, vol. 24, pp. 118-123, 2013.
* [3] C. Bracken., E. Byrne., N. Markin., G. McGuire., New families of quadratic almost perfect nonlinear trinomials and multinomias, Finite fields and their applications, vol. 14, no. 3, pp. 703-714, 2008.
* [4] C. Bracken., E. Byrne., N. Markin., G. McGuire., A few more quadratic APN functions, Cryptography and Communications, vol. 3, no. 1, pp. 43-53, 2011.
* [5] C. Bracken., C. H. Tan., Y. Tan., On a class of quadratic polynomials with no zeros and its application to APN functions, Finite fields and their applications, vol. 24, pp. 26-36, 2014.
* [6] L. Budaghyan., M. Calderini., C. Carlet., R. Coutter., I. Villa., Constructing APN functions through isotopic shift, IEEE Transactions on Information Theory, vol. 66, no. 8, pp. 5299-5309, 2020.
* [7] L. Budaghyan., M. Calderini., I. Villa., On equivalence between known families of quadratic APN functions, Finite fields and their applications, vol. 66, 101704, 2020.
* [8] L. Budaghyan., C. Carlet., Classes of quadratic APN trinomials and hexanomials and related structures, IEEE Transactions on Information Theory, vol. 54, no. 5, pp. 2354-2357, 2008.
* [9] L. Budaghyan., C. Carlet., G. Leander., Two classes of quadratic APN binomials inequivalent to power functions, IEEE Transactions on Information Theory, vol. 54, no. 9, pp. 4218-4229, 2008.
* [10] L. Budaghyan., C. Carlet., G. Leander., Constructing new APN functions from known ones, Finite fields and their applications, vol. 15, no. 2, pp. 150-159, 2009.
* [11] L. Budaghyan., C. Carlet., G. Leander., On a construction of quadratic APN functions, in Proceedings of IEEE Information Theory Workshop, ITW’09, pp. 374-378, 2009.
* [12] L. Budaghyan., T. Helleseth., N. Kaleyski., A new family of APN quadrinomials, IEEE Transactions on Information Theory, vol. 66, no. 11, pp. 7081-7087, 2020.
* [13] C. Carlet., P. Charpin., V. Zinoviev., _Codes, bent functions and permutations suitable for DES-like cryptosystems_ , Designs, Codes and Cryptography, vol. 15, no. 2, pp. 125-156, 1998.
* [14] H. Dobbertion., Almost perfect nonlinear power functions on GF($2^{n}$): the Welch case, IEEE Transactions on Information Theory, vol. 45, no. 4, pp. 1271-1275, 1999.
* [15] H. Dobbertion., Almost perfect nonlinear power functions on GF($2^{n}$): the Niho case, Information and Computation, vol. 151, no. 1, pp. 57-72, 1999.
* [16] H. Dobbertin., Almost perfect nonlinear power functions on GF($2^{n}$): A new case for $n$ divisible by 5, International Conference on Finite Fields and Applications, pp. 113-121, 2001.
* [17] Y, Edel., G. Kyureghyan., A. Pott., A new APN functions which is not equivalent to a power mapping, IEEE Transactions on Information Theory, vol. 52, no. 2, pp. 744-747, 2006.
* [18] R. Gold., Maximal recursive sequences with 3-valued recursive cross-correlation functions, IEEE Transactions on Information Theory, vol. 14, no. 1, pp. 154-156, 1968.
* [19] T. Kasami., The weight enumerators for several classes of subcodes of the 2nd order binary Reed-Muller codes, Information and Control, vol. 18, no. 4, pp. 369-394, 1971.
* [20] C. Kaspers., Y. Zhou., The number of almost perfect nonlinear functions grows exponentially, Journal of Cryptography, In Press.
* [21] H. K. Kim., S. Mesnager., Solving $x^{2^{k}+1}+x+a=0$ in $\mathbb{F}_{2^{n}}$ with ${\rm gcd}(n,k)=1$, Finite fields and their applications, vol. 63, 101630, 2020.
* [22] K. Nyberg., Differetially uniform mappings for cryptography, Lecture Notes in Computer Science, vol. 765, pp. 55-64, 1994.
* [23] H. Taniguchi., On some quadratic APN functions, Designs, codes and cryptography, vol. 87, pp. 1973-1983, 2019.
* [24] K. S. Williams., Note on Cubics over GF($2^{n}$) and GF($3^{n}$)∗, Journal of Number Theory, vol. 7, pp. 361-365, 1975.
* [25] Y, Yu., M. Wang., Y. Li., A matrix approach for constructing quadratic APN functions, Designs, codes and cryptography, vol. 73, no. 2, pp. 587-600, 2014.
* [26] Y. Zhou., A. Pott., A new family of semifields with 2 parameters, Advances in Mathematics, vol. 234, pp. 43-60, 2013.
## 5 Appendix
### 5-A Proof of 1) in Theorem 3.10
###### Proof.
It can be checked that $A^{q+1}+B^{q+1}=(x+x^{q})^{5}=h^{5}$ in this case. In
the following, we assume that $c$ is a non-cube in $\mathbb{F}_{2^{n}}$. Note
that $A\neq 0$. In fact, if $A=0$, then $h+c+c^{2}=x^{4}+x^{q}=0$, which
implies that $x\in\mathbb{F}_{2^{n}}\cap\mathbb{F}_{2^{m-2}}=\mathbb{F}_{2}$,
since $m$ is odd, and ${\rm gcd}(n,m-2)=1$, a contradiction to the assumption
that $x\neq 0,1$. Let $y:=y+\frac{B}{A}$. Then equation $(\ref{key-1})$
becomes
$\displaystyle
y^{3}+\frac{AB^{q}+B^{2}}{A^{2}}y+\frac{A^{q+1}+B^{q+1}}{A^{2}}=0.$
Let $y=Ez$, where $E$ satisfies that $E^{2}=\frac{AB^{q}+B^{2}}{A^{2}}$. Note
that $E\neq 0$. In fact, this would imply that $AB^{q}=B^{2}$, and hence
$A^{q}B=B^{2q}$, $A^{q+1}B^{q+1}=B^{2(q+1)}$. However, by the fact that $B\neq
0$ (if $B=0$, then $c=0,1$, a contradiction to the assumption that $c$ is a
non-cube), we have $A^{q+1}+B^{q+1}=0$, which implies that $(x+x^{q})^{5}=0$,
i.e., $x\in\mathbb{F}_{q}$, and then $c=x+x^{2}\in\mathbb{F}_{2^{m}}$ is a
cube in $\mathbb{F}_{2^{n}}$, since every element in $\mathbb{F}_{2^{m}}$ is a
cube by the fact that ${\rm gcd}(3,2^{m}-1)=1$ (since $m$ is odd), a
contradiction.
Then the above equation becomes
$\displaystyle z^{3}+z+a=0,$ (17)
where $a\neq 0$ satisfies that
$\displaystyle a=\frac{A^{q+1}+B^{q+1}}{A^{2}E^{3}}.$
It can be checked that
$\displaystyle\frac{1}{a^{2}}=\frac{(AB^{q}+B^{2})^{3}}{A^{2}(A^{q+1}+B^{q+1})^{2}}.$
(18)
It is clear that equation (12) has no solutions in $\mathbb{F}_{2^{n}}$ if and
only if (17) has no solutions. To complete the proof, according to Theorem
3.9, we have to show that ${\rm Tr}^{n}_{1}\Big{(}\frac{1}{a^{2}}\Big{)}=0$,
and $t_{1}$ is a non-cube in $\mathbb{F}_{2^{n}}$, where $t_{1}$ is one
solution in $\mathbb{F}_{2^{n}}$ of $t^{2}+at+1=0$.
Claim 1. ${\rm Tr}^{n}_{1}\Big{(}\frac{1}{a^{2}}\Big{)}=0$.
In fact, we have
$\displaystyle\frac{1}{a^{2}}=\frac{(AB^{q}+B^{2})^{3}}{A^{2}(A^{q+1}+B^{q+1})^{2}}=\frac{B^{3}+M}{A(A^{q+1}+B^{q+1})}+\Bigg{(}\frac{B^{3}+M}{A(A^{q+1}+B^{q+1})}\Bigg{)}^{2},$
(19)
where $M$ is one solution of the following equation
$\displaystyle M^{2}+DM+H=0.$ (20)
Recall the notations that $D=A(A^{q+1}+B^{q+1})$,
$H=A^{2}(A^{q}B^{3}+AB^{3q}+B^{2+2q})$, $A=c^{2-2q}(h+c+c^{2})$, $B=c+c^{2}$,
$c=x+x^{2}$. We need only to show that the above equation in $M$ has solutions
in $\mathbb{F}_{2^{n}}$, i.e., ${\rm
Tr}^{n}_{1}\Big{(}{\frac{H}{D^{2}}}\Big{)}=0$. This can be seen from the
following fact.
$\displaystyle\frac{H}{D^{2}}=\frac{A^{q}B^{3}+AB^{3q}+B^{2+2q}}{(A^{q+1}+B^{q+1})^{2}}$
is an element in $\mathbb{F}_{2^{m}}$, since $A^{q}B^{3}+AB^{3q}={\rm
Tr}^{n}_{m}(A^{q}B^{3})$, $A^{q+1}$, $B^{q+1}\in\mathbb{F}_{2^{m}}$.
Next, we need to find one solution $t_{1}$ in $\mathbb{F}_{2^{n}}$ of
$t^{2}+at+1=0$, and show that $t_{1}$ is a non-cube. It is clear that $t_{1}$
can be represented as $av$, where $v=\frac{B^{3}+M}{A(A^{q+1}+B^{q+1})}$,
since $\frac{1}{a^{2}}=v+v^{2}$ according to (19). Note that $t_{1}=av$
satisfies that
$\displaystyle t^{2}_{1}=a^{2}v^{2}=\frac{(B^{3}+M)^{2}}{(AB^{q}+B^{2})^{3}}.$
Therefore, to show $t_{1}$ is a non-cube in $\mathbb{F}_{2^{n}}$, we have to
show that $B^{3}+M$ is a non-cube.
Claim 2. $B^{3}+M$ is a non-cube in $\mathbb{F}_{2^{n}}$.
Our strategy is to find the explicit expression of $M$, and then show that
$B^{3}+M$ is a non-cube. To this end, we have to revisit equation (20), and
explore more information on the element $\frac{H}{D^{2}}$ (it is in
$\mathbb{F}_{2^{m}}$). Very fortunately, we find that ${\rm
Tr}^{m}_{1}\Big{(}\frac{H}{D^{2}}\Big{)}=0.$ In fact, recall the notations
that $h=x+x^{q}$, and $r=x^{q+1}$, we find (with computer assistance) that (a
surprise)
$\displaystyle\frac{H}{D^{2}}=u+u^{2},$ (21)
where
$\displaystyle u=\frac{h^{2}(r+r^{2})+r+r^{4}+hr^{2}}{h^{5}}.$
Then $M$ can be chosen as $Du$ (this is because it suffices to find one
solution of $M^{2}+DM+H=0$). We find that
$\displaystyle M=Du=A(A^{q+1}+B^{q+1})u$ $\displaystyle=$ $\displaystyle
c^{2-2q}(h+c+c^{2})h^{5}\cdot\frac{h^{2}(r+r^{2})+r+r^{4}+hr^{2}}{h^{5}}$
$\displaystyle=$
$\displaystyle\frac{c^{2}(h+c+c^{2})({h^{2}(r+r^{2})+r+r^{4}+hr^{2})}}{c^{2q}}.$
Then, recall the notation that $B=c+c^{2}$, we can obtain the expression of
$B^{3}+M$ as follows.
$\displaystyle\begin{aligned}
B^{3}+M&=&\frac{h(c^{2q+4}+c^{q+5}+c^{q+4})+c^{2q+4}+c^{q+5}}{c^{2q}}\\\
&=&\frac{c^{2}\Big{(}h(c^{2q+4}+c^{q+5}+c^{q+4})+c^{2q+4}+c^{q+5}\Big{)}}{c^{2+2q}}.\end{aligned}$
(22)
The above expression can be deduced from
$\displaystyle h+h^{2}=c+c^{q},~{}c^{q+1}=r+r^{2}+hr.$
Note that $h,~{}c^{2+2q}\in\mathbb{F}^{\ast}_{2^{m}}$ is a cube, it suffices
to show that
$\displaystyle
hc^{2}\Big{(}h(c^{2q+4}+c^{q+5}+c^{q+4})+c^{2q+4}+c^{q+5}\Big{)}$
is a non-cube. By the fact that $h+h^{2}=c+c^{q}$, we have
$\displaystyle
hc^{2}\Big{(}h(c^{2q+4}+c^{q+5}+c^{q+4})+c^{2q+4}+c^{q+5}\Big{)}=c^{5}c^{q+1}((c+c^{q})^{2}+h^{2}).$
Since $c^{q+1}$, $c+c^{q}$, $h\in\mathbb{F}^{\ast}_{2^{m}}$ are all cubes in
$\mathbb{F}_{2^{n}}$, we have that the above element is a non-cube, when $c$
is a non-cube. ∎
### 5-B Proof of 2) in Theorem 3.10
###### Proof.
The proof is similar to that of 1) in Theorem 3.10. Recall the following
notations: $r=x^{q+1};h=x+x^{q};c=x+x^{2};$$A=\frac{h+c+c^{2}}{c^{q}};B=1+c,$
from which we can obtain that $h+h^{2}=c+c^{q}$ and
$A^{q+1}+B^{q+1}=\frac{(x+x^{q})^{5}}{(x+x^{2})^{q+1}}=\frac{h^{5}}{c^{q+1}}$.
Note that $A\neq 0$, otherwise, we have $x+x^{4q}=0$ that means that
$x\in\mathbb{F}_{2^{n}}\cap\mathbb{F}_{4q}=\mathbb{F}_{2}$, since
$\gcd(m+2,n)=1$. Then setting $y:=y+\frac{B}{A}$, this can transform (12) into
$\displaystyle
y^{3}+\frac{AB^{q}+B^{2}}{A^{2}}y+\frac{A^{q+1}+B^{q+1}}{A^{2}}=0.$ (23)
Observe that $B\neq 0$ (otherwise $c=1$ is a cube) and $AB^{q}+B^{2}\neq 0$,
otherwise, we have $A^{q+1}+B^{q+1}=0$, that is, $h=0$, which implies that
$c\in\mathbb{F}_{q}$ contracting to the assumption that $c$ is a non-cube,
since $\gcd(3,2^{m}-1)=1$ for any odd $m$. Thus we can transform the equation
(23) into
$\displaystyle z^{3}+z+a=0$ (24)
by setting $y=Ez$, where $a,E\in\mathbb{F}_{2^{n}}^{*}$ such that
$\displaystyle E^{2}=\frac{AB^{q}+B^{2}}{A^{2}}\hskip 5.69046pt{\rm and}\hskip
5.69046pta^{2}=\frac{A^{2}(A^{q+1}+B^{q+1})^{2}}{(AB^{q}+B^{2})^{3}}.$
We need now to prove that equation (24) has no solutions in
$\mathbb{F}_{2^{n}}$. According to Theorem 3.9, we have to show that ${\rm
Tr}_{1}^{n}\Big{(}\frac{1}{a^{2}}\Big{)}=0$ and the solutions in
$\mathbb{F}_{2^{n}}$ of equation $t^{2}+at+1=0$ are not cubes of
$\mathbb{F}_{2^{n}}$.
Firstly, we prove that ${\rm Tr}_{1}^{n}\Big{(}\frac{1}{a^{2}}\Big{)}=0$. Note
that $\frac{1}{a^{2}}$ can be written as
$\displaystyle\frac{1}{a^{2}}=\frac{B^{3}+M}{A(A^{q+1}+B^{q+1})}+\bigg{(}\frac{B^{3}+M}{A(A^{q+1}+B^{q+1})}\bigg{)}^{2},$
(25)
where $M$ is a solution of
$\displaystyle M^{2}+DM+H=0,$ (26)
where $D=A(A^{q+1}+B^{q+1})$ and $H=A^{2}(AB^{3q}+A^{q}B^{3}+B^{2(q+1)})$.
Then we transform the problem into showing that equation (26) has solutions in
$\mathbb{F}_{2^{n}}$, which is equivalent to ${\rm
Tr}_{1}^{n}\Big{(}\frac{H}{D^{2}}\Big{)}=0$. Indeed, it can be seen that
$\displaystyle\frac{H}{D^{2}}=\frac{AB^{3q}+A^{q}B^{3}+B^{2(q+1)}}{(A^{q+1}+B^{q+1})^{2}}=\frac{{\rm
Tr}_{m}^{n}(AB^{3q})+B^{2(q+1)}}{(A^{q+1}+B^{q+1})^{2}},$
which is clearly in $\mathbb{F}_{q}$. Thus, ${\rm
Tr}_{1}^{n}\Big{(}\frac{H}{D^{2}}\Big{)}=0$.
Then, we show that the solutions of $t^{2}+at+1=0$ are not cubes in
$\mathbb{F}_{2^{n}}$. Assume that $t_{1}$ is a solution of $t^{2}+at+1=0$.
Then by (25), it can be represented by $t_{1}=a\nu$, where
$\nu=\frac{B^{3}+M}{A(A^{q+1}+B^{q+1})}$, and thus
$\displaystyle
t_{1}^{2}=a^{2}\nu^{2}=\frac{(B^{3}+M)^{2}}{(AB^{q}+B^{2})^{3}}.$
Therefore, to show $t_{1}$ is not a cube, it suffices to show $(B^{3}+M)^{2}$
and thus $B^{3}+M$ is not a cube of $\mathbb{F}_{2^{n}}$. In the following, we
show this fact by giving the explicit expression of $M$ by revisiting (26)
again.
By the above discussion, we have obtained that
$\frac{H}{D^{2}}\in\mathbb{F}_{q}$. We further want to show that ${\rm
Tr}_{1}^{m}\Big{(}\frac{H}{D^{2}}\Big{)}=0$, which is equivalent to showing
$\displaystyle\frac{H}{D^{2}}=\mu+\mu^{2}$ (27)
for some $\mu\in\mathbb{F}_{2^{m}}$. Recall that $A=\frac{h+c+c^{2}}{c^{q}}$,
$B=1+c$ and $A^{q+1}+B^{q+1}=\frac{h^{5}}{c^{q+1}}$, we have
$\displaystyle A^{q}B^{3}+AB^{3q}=$
$\displaystyle\frac{(h+c^{q}+c^{2q})B^{3}}{c}+\frac{(h+c+c^{2})B^{3q}}{c^{q}}$
$\displaystyle=$
$\displaystyle\frac{c^{q}(h+c^{q}+c^{2q})B^{3}+c(h+c+c^{2})B^{3q}}{c^{q+1}}$
$\displaystyle=$
$\displaystyle\frac{h(c^{q}B^{3}+cB^{3q})+c^{q}B^{3}(c^{q}+c^{2q})+cB^{3q}(c+c^{2})}{c^{q+1}}.$
While
$\displaystyle h(c^{q}B^{3}+cB^{3q})=$ $\displaystyle
h\big{(}c^{q}(1+c+c^{2}+c^{3})+c(1+c^{q}+c^{2q}+c^{3q})\big{)}$
$\displaystyle=$ $\displaystyle
h(c+c^{q}+c^{q+1}(c+c^{q})+c^{q+1}(c+c^{q})^{2})$
and
$\displaystyle c^{q}B^{3}(c^{q}+c^{2q})+cB^{3q}(c+c^{2})=$ $\displaystyle
c^{q}(1+c+c^{2}+c^{3})(c^{q}+c^{2q})+(c^{q}(1+c+c^{2}+c^{3})(c^{q}+c^{2q}))^{q}$
$\displaystyle=$ $\displaystyle
c^{2q}+c^{3q}+c^{2q+1}+c^{3q+1}+c^{2q+2}+c^{3q+2}+c^{2q+3}+c^{3q+3}+$
$\displaystyle~{}~{}~{}~{}~{}~{}~{}~{}(c^{2q}+c^{3q}+c^{2q+1}+c^{3q+1}+c^{2q+2}+c^{3q+2}+c^{2q+3}+c^{3q+3})^{q}$
$\displaystyle=$
$\displaystyle(c+c^{q})^{2}+c^{3}+c^{3q}+c^{q+1}(c+c^{q})+c^{q+1}(c+c^{q})^{2}.$
We have
$\displaystyle
c+c^{q}=x+x^{q}+(x+x^{q})^{2},c^{q+1}=x^{q+1}+x^{2(q+1)}+x^{q+1}(x+x^{q}),$
from which we can obtain that
$\displaystyle h(c^{q}B^{3}+cB^{3q})=$
$\displaystyle(x+x^{q})^{2}+(x+x^{q})^{3}+x^{q+1}(x+x^{q})^{2}+x^{q+1}(x+x^{q})^{3}+x^{q+1}(x+x^{q})^{5}$
$\displaystyle+x^{q+1}(x+x^{q})^{6}+x^{(2q+1)}(x+x^{q})^{2}+x^{(2q+1)}(x+x^{q})^{5}$
and
$\displaystyle c^{q}B^{3}(c^{q}+c^{2q})+cB^{3q}(c+c^{2})=$
$\displaystyle(x+x^{q})^{2}+(x+x^{q})^{3}+(x+x^{q})^{5}+(x+x^{q})^{6}+x^{q+1}(x+x^{q})^{2}$
$\displaystyle+x^{q+1}(x+x^{q})^{3}+x^{q+1}(x+x^{q})^{4}+x^{q+1}(x+x^{q})^{5}$
$\displaystyle+x^{2(q+1)}(x+x^{q})^{2}+x^{2(q+1)}(x+x^{q})^{4}.$
Thus we have
$\displaystyle c^{q+1}(A^{q}B^{3}+AB^{3q})=$
$\displaystyle(x+x^{q})^{5}+(x+x^{q})^{6}+x^{q+1}(x+x^{q})^{4}+x^{q+1}(x+x^{q})^{6}$
$\displaystyle+x^{2(q+1)}(x+x^{q})^{4}+x^{2(q+1)}(x+x^{q})^{5}$
and
$\displaystyle c^{2(q+1)}(A^{q}B^{3}+AB^{3q})=$ $\displaystyle
x^{q+1}(x+x^{q})^{5}+x^{q+1}(x+x^{q})^{7}+x^{2(q+1)}(x+x^{q})^{4}+x^{2(q+1)}(x+x^{q})^{7}$
$\displaystyle+x^{4(q+1)}(x+x^{q})^{4}+x^{4(q+1)}(x+x^{q})^{5}.$
We further have
$\displaystyle c^{2(q+1)}B^{2(q+1)}=$ $\displaystyle c^{2(q+1)}(1+c)^{2(q+2)}$
$\displaystyle=$ $\displaystyle c^{2(q+1)}+c^{2(q+1)}(c+c^{q})^{2}+c^{4(q+1)}$
$\displaystyle=$ $\displaystyle
x^{2(q+1)}+x^{2(q+1)}(x+x^{q})^{6}+x^{4(q+1)}(x+x^{q})^{2}+x^{8(q+1)}.$
Recall that $h=x+x^{q}$, $r=x^{q+1}$. Thus, we have
$\displaystyle c^{2(q+1)}(A^{q}B^{3}+AB^{3q}+B^{2(q+1)})=$ $\displaystyle
rh^{5}+rh^{7}+r^{2}+r^{2}h^{4}+r^{2}h^{6}+r^{2}h^{7}+r^{4}h^{2}+r^{4}h^{4}+r^{4}h^{5}+r^{8},$
and
$\displaystyle\frac{H}{D^{2}}=\frac{r}{h^{5}}+\frac{r}{h^{3}}+\frac{r^{2}}{h^{10}}+\frac{r^{2}}{h^{6}}+\frac{r^{2}}{h^{4}}+\frac{r^{2}}{h^{3}}+\frac{r^{4}}{h^{8}}+\frac{r^{4}}{h^{6}}+\frac{r^{4}}{h^{5}}+\frac{r^{8}}{h^{10}}=\mu+\mu^{2}$
where $\mu=\frac{r+r^{4}+r^{2}h+(r+r^{2})h^{2}}{h^{5}}$. The rest of this
proof is similar to that of Theorem 3.10, so we omit it here.∎
|
# Tractable higher-order under-approximating AE extensions for non-linear
systems
Eric Goubault Sylvie Putot LIX, CNRS, Ecole Polytechnique, Institut
Polytechnique de Paris, France (email<EMAIL_ADDRESS>
###### Abstract
We consider the problem of under and over-approximating the image of general
vector-valued functions over bounded sets, and apply the proposed solution to
the estimation of reachable sets of uncertain non-linear discrete-time
dynamical systems. Such a combination of under and over-approximations is very
valuable for the verification of properties of embedded and cyber-physical
controlled systems. Over-approximations prove properties correct, while under-
approximations can be used for falsification. Coupled, they provide a measure
of the conservatism of the analysis. This work introduces a general framework
relying on computations of robust ranges of vector-valued functions. This
framework allows us to extend for under-approximation many precision
refinements that are classically used for over-approximations, such as affine
approximations, Taylor models, quadrature formulae and preconditioning
methods. We end by evaluating the efficiency and precision of our approach,
focusing on the application to the analysis of discrete-time dynamical systems
with inputs and disturbances, on different examples from the literature.
###### keywords:
Uncertain systems, Computer-aided control design
## 1 Introduction
Guaranteed state estimation and reachability analysis are central to many
problems in control, such as robust and optimal control of dynamical systems,
set invariance, safety verification, or control synthesis. This ultimately
relies on computing ranges of functions over a domain, that we have to
approximate since this is an intractable problem.
Much of the existing work focuses on over-approximations of images of
functions, or of reachable sets, generally based on convex set representations
(in particular intervals, ellipsoids, polyhedra). We are interested here in
the much less studied problem of computing under-approximations, that is, sets
of states guaranteed to be reached. Combining over and under approximations is
fundamental for the validation of control systems.When the over-approximation
is not sufficient to prove a property, an under-approximation is helpful to
state the quality of the over-approximation. Additionally, when an under-
approximation of the reachable set intersects the set of error states, it
provides a proof of falsification of the property.
For general controlled systems, the reachability properties will depend on the
initial conditions of the system, but also on the sensitivity of the system to
some control inputs and external disturbances, as reflected by the notions of
minimal and maximal reachability Mitchell (2007). We generalize these notions
here to robust reachability, when both control inputs and adversarial
disturbances are present. The robust image will be the intersection, for all
possible disturbances, of the images of a function or reachable sets of a
system.
#### Contents and contributions
The computation of the reachable set of a dynamical system can be reduced to a
series of images of sets by some vector-valued function. In previous work
Goubault and Putot (2020), we introduced mean-value extensions allowing us to
compute under approximations (also called inner-approximations) of such images
in a very efficient way. In this article, we generalize this approach to
higher-order extensions, and develop quadrature formulas for more precise
under-approximations. We also address many questions that are to be solved for
an accurate and efficient implementation:
* •
Section 2 recaps the necessary background from previous work. Section 3
generalizes the mean-value extension of Goubault and Putot (2020) for the
under and over-approximation of the robust range of sufficiently smooth real-
valued functions $f:\mathbb{R}^{m}\rightarrow\mathbb{R}$; this generalization
allows us to propose higher-order Taylor extensions for under-approximating
robust ranges of functions. These new extensions, just as the mean-value
extensions, is the basis for under-approximations of elementary vector-valued
functions from $\mathbb{R}^{m}$ to $\mathbb{R}^{n}$, as detailed in Section
2.2, which is instrumental in the reachability analysis of dicrete-time
nonlinear systems proposed later;
* •
Section 4 proposes a novel approach to subdivisions for mean-value and Taylor
extensions, based on the idea of quadratures (in numerical calculations of
integrals): we show that this improves precision of the computation of under
and over-approximations, while still scaling with the dimension of the system;
* •
Section 5 applies this work to the approximation of robust reachable sets of
discrete-time dynamical systems; we present results on representative systems
from the literature, demonstrating the tractability and precision of our
approach.
#### Related work
Our approach is related to and partially relies on work on modal intervals and
mean-value extensions, which applications include the computation of under-
approximations of function images Goldsztejn (2012a, b). It is also related to
over-approximations of nonlinear functions and dynamical systems, on which we
rely to compute under-approximations. Many methods for over-approximating
reachable sets for non-linear systems have been developed, among which Taylor
methods Makino and Berz (2003) or polytopes-based methods Guernic and Girard
(2009); Dreossi et al. (2016).
There exist less methods for the harder problem of under-approximating images
of functions or sets of reachable states. Some approaches have been proposed
for linear discrete-time systems Kurzhanski and Varaiya (2000); Girard et al.
(2006); Raković and Fiacchini (2008). Interval-based methods, relying on space
discretization, have been used for under-approximating the image of nonlinear
functions Goldsztejn and Jaulin (2010). They were also used to over and under
approximate solutions of differential systems with uncertain initial
conditions Mézo et al. (2018). Tight approximations for reachable sets of
nonlinear continuous systems can be found via expensive Eulerian methods: the
zero sub-level set of the Lipschitz viscosity solution to a Hamilton-Jacobi
(HJB) partial differential equation gives the (backward) reachable set Chen et
al. (2016). Other approaches, using SoS methods and LMI relaxations have been
proposed for inner approximations, see e.g. Korda et al. (2013). In Xue et al.
(2020), under-approximations for polynomial systems are obtained by solving
semi-definite programs. Taylor models are used on the inverse flow map to
derive under-approximations Chen et al. (2014), but using topological
conditions that are checked with interval constraints solving, which have
difficulties to scale up with dimension. In Xue et al. (2016), the computation
of the under-approximated reachable set is based on a costly analysis of the
boundary of the reachable sets and polytopic approximations. In Kochdumper and
Althoff (2020), some non-convex under-approximations are computed with
polynomial zonotopes, relying on a computation of the outer-approximation of
the reachable set, of an enclosure of the boundary of the reachable set, and a
reduction of the outer-approximation until it is fully included in the region
delimited by the boundary.
#### Notations
For a continuously differentiable vector-valued function
$f:\mathbb{R}^{m}\rightarrow\mathbb{R}^{n}$, we note $f_{i}$ its $i$-th
component and $\nabla f=(\nabla_{j}f_{i})_{ij}=(\frac{\partial f_{i}}{\partial
x_{j}})_{1\leq i\leq n,1\leq j\leq m}$ its Jacobian matrix. We note $\langle
x,y\rangle$ the scalar product of vectors $x$ and $y$, and $\lvert x\rvert$
the absolute value extended componentwise.
Intervals are used in many situations to rigorously compute with interval
domains instead of reals, usually leading to over-approximations of function
ranges over boxes. Set valued quantities, whether scalar or vector-valued,
will be noted with bold letters, e.g $x$. We denote
${\mathbb{I}\mathbb{R}}=\\{{\makebox{\boldmath$x$}}=[\underline{x},\overline{x}],\>\underline{x}\in\mathbb{R},\>\overline{x}\in\mathbb{R}\\}$
the set of intervals with real-valued bounds. If $\overline{x}<\underline{x}$,
the interval represents the empty set. For a (possibly vector-valued) interval
${\makebox{\boldmath$x$}}\in{\mathbb{I}\mathbb{R}}^{m}$, we note
$c({{\makebox{\boldmath$x$}}})=(\underline{x}+\overline{x})/2$ its center and
$r({{\makebox{\boldmath$x$}}})=(\overline{x}-\underline{x})/2$ its radius. The
operators over real numbers are lifted in intervals using the same notation.
## 2 Background: AE extensions for computing function images
We recall in this section the notations and results of Goubault and Putot
(2020): we state mean-value over and under-approximating extensions for scalar
and vector-valued functions.
An _over-approximating extension_ , also called _outer-approximating
extension_ , of a function $f:\mathbb{R}^{m}\rightarrow\mathbb{R}^{n}$ is a
function
${\makebox{\boldmath$f$}}:\mathcal{P}(\mathbb{R}^{m})\rightarrow\mathcal{P}(\mathbb{R}^{n})$,
such that for all $x$ in $\mathcal{P}(\mathbb{R}^{m})$,
$\mbox{range}(f,{\makebox{\boldmath$x$}})=\\{f(x),x\in{\makebox{\boldmath$x$}}\\}\subseteq{\makebox{\boldmath$f$}}({\makebox{\boldmath$x$}})$.
Dually, under-approximations determine a set of values proved to belong to the
range of the function over some input set. An _under-approximating extension_
, also called _inner-approximating extension_ , of $f$ is a function
${\makebox{\boldmath$f$}}:\mathcal{P}(\mathbb{R}^{m})\rightarrow\mathcal{P}(\mathbb{R}^{n})$,
such that for all $x$ in $\mathcal{P}(\mathbb{R}^{m})$,
${\makebox{\boldmath$f$}}({\makebox{\boldmath$x$}})\subseteq\mbox{range}(f,{\makebox{\boldmath$x$}})$.
Under- and over-approximations can be interpreted as quantified propositions:
$\operatorname{range}(f,{\makebox{\boldmath$x$}})\subseteq{\makebox{\boldmath$z$}}$
can be written $\forall x\in{\makebox{\boldmath$x$}},\,\exists
z\in{\makebox{\boldmath$z$}},\,f(x)=z,$ while
${\makebox{\boldmath$z$}}\,\subseteq\operatorname{range}(f,{\makebox{\boldmath$x$}})$
can be written $\forall z\in\,{\makebox{\boldmath$z$}},\,\exists
x\in{\makebox{\boldmath$x$}},\,f(x)=z.$ Both these propositions are what we
will call _AE propositions_ , for quantified propositions where universal
quantifiers (A) precede existential quantifiers (E).
### 2.1 Mean-value AE extensions for scalar-valued functions
We consider a function $f:\mathbb{R}^{m}\rightarrow\mathbb{R}$. The natural
interval extension consists in replacing real operations by their interval
counterparts in the expression of the function. A generally more accurate
extension relies on a linearization by the mean-value theorem.
#### 2.1.1 Mean-value AE extensions
Suppose $f$ is differentiable over the box $x$. The mean-value theorem implies
that
$\forall x^{0}\in{\makebox{\boldmath$x$}},\,\forall
x\in{\makebox{\boldmath$x$}},\exists\xi\in{\makebox{\boldmath$x$}},\,f(x)=f(x^{0})+\langle\nabla
f(\xi),x-x^{0}\rangle.$
If we can bound the range of the gradient of $f$ over $x$, by
${\makebox{\boldmath$\nabla f$}}({\makebox{\boldmath$x$}})$, then we can
derive an interval enclosure, called the mean-value extension. Let us choose
$x^{0}$ to be the center $c({{\makebox{\boldmath$x$}}})$ of $x$ and recall we
note $r({{\makebox{\boldmath$x$}}})=(\overline{x}-\underline{x})/2$ its
radius.
###### Theorem 1 (Thm. 1, Goubault and Putot (2020))
Let $f$ be a continuously differentiable function from $\mathbb{R}^{m}$ to
$\mathbb{R}$ and ${\makebox{\boldmath$x$}}\in{\mathbb{I}\mathbb{R}}^{m}$. Let
${\makebox{\boldmath$f$}}^{0}=[\underline{f^{0}},\overline{f^{0}}]$ include
$f(c({{\makebox{\boldmath$x$}}}))$ and $\nabla$ a vector of intervals
${\makebox{\boldmath$\nabla_{i}$}}=[\underline{\nabla}_{i},\overline{\nabla}_{i}]$
for $i\in\\{1,\ldots,m\\}$ such that
$\left\\{\left\lvert\nabla_{i}f(c({\makebox{\boldmath$x$}}_{1}),\ldots,c({\makebox{\boldmath$x$}}_{i-1}),x_{i},\ldots,x_{m})\right\rvert,x\in{\makebox{\boldmath$x$}}\right\\}\subseteq{\makebox{\boldmath$\nabla$}}_{i}.$
We have the over- and under-approximating extensions
$\displaystyle\mbox{range}(f,{\makebox{\boldmath$x$}})\subseteq[\underline{f^{0}},\overline{f^{0}}]+\langle\overline{\nabla},r({{\makebox{\boldmath$x$}}})\rangle[-1,1]$
(1)
$\displaystyle[\overline{f^{0}}-\langle\underline{\nabla},r({\makebox{\boldmath$x$}})\rangle,\underline{f^{0}}+\langle\underline{\nabla},r({\makebox{\boldmath$x$}})\rangle]\subseteq\mbox{range}(f,{\makebox{\boldmath$x$}})$
(2)
###### Example 2.1
Let us consider the range of $f$ defined by $f(x)=x^{2}-x$ over
${\makebox{\boldmath$x$}}=[2,3]$. We can compute $f(2.5)=3.75$ and
$\lvert\nabla f([2,3])\rvert\subseteq[3,5]$. Then (1) and (2) yield
$3.75+1.5[-1,1]\subseteq\mbox{range}(f,[2,3])\subseteq 3.75+2.5[-1,1],$ from
which we deduce
$[2.25,5.25]\subseteq\mbox{range}(f,[2,3])\subseteq[1.25,6.25]$.
We refer to extensions (1) and (2) as _AE extensions_ , as they can be
interpreted as _AE propositions_. Note that the wider, lesser quality, are the
over-approximations of $f$ and its derivatives, the tighter, less quality, are
the under-approximations. The under-approximation can even become empty if the
width $\overline{f_{0}}-\underline{f_{0}}$ of the approximation of
$f(c({{\makebox{\boldmath$x$}}}))$ exceeds
$2\langle\overline{\nabla},r({{\makebox{\boldmath$x$}}})\rangle$: in this case
the lower bound of the resulting interval is larger than the upper bound,
which by convention we identify with the empty interval. Note also that when
$0\in{\makebox{\boldmath$\nabla_{i}{\makebox{\boldmath$f$}}$}}$, then
$\underline{\nabla_{i}}=0$ and if this is the case for all $i$, the under-
approximation is empty or reduced to a point. A special attention to the
practical evaluation of these extensions over the region $x$ of interest is
thus crucial, this is the object of Section 4.
#### 2.1.2 Mean-value AE extensions of the robust range
Mean-value AE extensions can be generalized to compute ranges that are robust
to disturbances, identified as some input components. Let us partition the
indices of the input space in two subsets $I_{\mathcal{A}}$ and
$I_{\mathcal{E}}$, where $I_{\mathcal{A}}$ defines the indices of the inputs
that correspond to disturbances, and $I_{\mathcal{E}}$ the remaining
dimensions. We decompose the input box $x$ accordingly by
${\makebox{\boldmath$x$}}={\makebox{\boldmath$x$}}_{\mathcal{A}}\times{\makebox{\boldmath$x$}}_{\mathcal{E}}$.
We define the robust range of function $f$ on $x$, robust on
${\makebox{\boldmath$x$}}_{\mathcal{E}}$ with respect to disturbances
${\makebox{\boldmath$x$}}_{\mathcal{A}}$, as
$\mbox{range}(f,{\makebox{\boldmath$x$}},I_{\mathcal{A}},I_{\mathcal{E}})=\\{z\,|\,\forall
w\in{\makebox{\boldmath$x$}}_{\mathcal{A}},\,\exists
u\in{\makebox{\boldmath$x$}}_{\mathcal{E}},\,z=f(w,u)\\}$. Intuitively, $u$
will be control components, $w$ disturbances to which the output range should
be robust.
###### Theorem 2 (Thm. 2, Goubault and Putot (2020))
Let $f$ be continuously differentiable function from $\mathbb{R}^{m}$ to
$\mathbb{R}$ and
${\makebox{\boldmath$x$}}={\makebox{\boldmath$x$}}_{\mathcal{A}}\times{\makebox{\boldmath$x$}}_{\mathcal{E}}\in{\mathbb{I}\mathbb{R}}^{m}$.
Let ${\makebox{\boldmath$f$}}^{0}$, ${\makebox{\boldmath$\nabla$}}_{w}$ and
${\makebox{\boldmath$\nabla$}}_{u}$ be vectors of intervals such that
$f(c({\makebox{\boldmath$x$}}))\subseteq{\makebox{\boldmath$f$}}^{0}$,
$\\{\left|\nabla_{w}f(w,c({\makebox{\boldmath$x$}}_{\mathcal{E}}))\right|\,,\,w\in{\makebox{\boldmath$x$}}_{\mathcal{A}}\\}\subseteq{\makebox{\boldmath$\nabla$}}_{w}$
and
$\\{\left|\nabla_{u}f(w,u)\right|\,,w\in{\makebox{\boldmath$x$}}_{\mathcal{A}},\,u\in{\makebox{\boldmath$x$}}_{\mathcal{E}}\\}\subseteq{\makebox{\boldmath$\nabla$}}_{u}$.
We have:
$\mbox{range}(f,{\makebox{\boldmath$x$}},I_{\mathcal{A}},I_{\mathcal{E}})\subseteq[\underline{f^{0}}-\langle\overline{\nabla}_{u},r({{\makebox{\boldmath$x$}}_{\mathcal{E}}})\rangle+\langle\underline{\nabla}_{w},r({{\makebox{\boldmath$x$}}_{\mathcal{A}}})\rangle,\\\
\overline{f^{0}}+\langle\overline{\nabla}_{u},r({{\makebox{\boldmath$x$}}_{\mathcal{E}}})\rangle-\langle\underline{\nabla}_{w},r({{\makebox{\boldmath$x$}}_{\mathcal{A}}})\rangle]$
(3)
$[\overline{f^{0}}-\langle\underline{\nabla}_{u},r({{\makebox{\boldmath$x$}}_{\mathcal{E}}})\rangle+\langle\overline{\nabla}_{w},r({{\makebox{\boldmath$x$}}_{\mathcal{A}}})\rangle,\underline{f^{0}}+\\\
\langle\underline{\nabla}_{u},r({{\makebox{\boldmath$x$}}_{\mathcal{E}}})\rangle-\langle\overline{\nabla}_{w},r({{\makebox{\boldmath$x$}}_{\mathcal{A}}})\rangle]\subseteq\mbox{range}(f,{\makebox{\boldmath$x$}},I_{\mathcal{A}},I_{\mathcal{E}})$
(4)
We refer to Example 2 of Goubault and Putot (2020) for a sample computation.
### 2.2 AE extensions for vector-valued functions
Following Goubault and Putot (2020), we now detail how full n-dimensional
boxes can be included in the image of vector-valued functions
$f:\mathbb{R}^{m}\rightarrow\mathbb{R}^{n}$, for $m\geq n$, using AE
extensions of robust ranges. Theorem 3 and Definition 2.2 will be instrumental
in Algorithm 1 for discrete-time reachability of Section 5.
The mean-value extensions of Theorem 1 or the generalization of Theorem 5 give
us under and over-approximations of projections of the image of the function.
The Cartesian product of the over-approximations of each component provides an
over-approximation of a vector-valued function
$f:\mathbb{R}^{m}\rightarrow\mathbb{R}^{n}$. This is however not the case for
under-approximation. Suppose for example that we have $\forall
z_{1}\in{\makebox{\boldmath$z$}}_{1},\exists
x_{1}\in{\makebox{\boldmath$x$}}_{1},\,\exists
x_{2}\in{\makebox{\boldmath$x$}}_{2},\>z_{1}=f_{1}(x)$ and $\forall
z_{2}\in{\makebox{\boldmath$z$}}_{2},\exists
x_{1}\in{\makebox{\boldmath$x$}}_{1},\,\exists
x_{2}\in{\makebox{\boldmath$x$}}_{2},\>z_{2}=f_{2}(x)$. We cannot deduce
directly that for all $\forall z_{1}\in{\makebox{\boldmath$z$}}_{1}$ and
$\forall z_{2}\in{\makebox{\boldmath$z$}}_{2}$ there exists $x_{1}$ and
$x_{2}$ such that $z=f(x)$.
Suppose now that we have: $\forall
z_{1}\in{\makebox{\boldmath$z$}}_{1},\forall
x_{1}\in{\makebox{\boldmath$x$}}_{1},\,\exists
x_{2}\in{\makebox{\boldmath$x$}}_{2},\>z_{1}=f_{1}(x)$ and $\forall
z_{2}\in{\makebox{\boldmath$z$}}_{2},\forall
x_{2}\in{\makebox{\boldmath$x$}}_{2},\,\exists
x_{1}\in{\makebox{\boldmath$x$}}_{1},\>z_{2}=f_{2}(x)$ with continuous
selections $x_{2}$ and $x_{1}$. Then there exists functions
$g_{2}(z_{1},x_{1}):{\makebox{\boldmath$z$}}_{1}\times{\makebox{\boldmath$x$}}_{1}\rightarrow{\makebox{\boldmath$x$}}_{2}$
and
$g_{1}(z_{2},x_{2}):{\makebox{\boldmath$z$}}_{2}\times{\makebox{\boldmath$x$}}_{2}\rightarrow{\makebox{\boldmath$x$}}_{1}$
that are continuous in $x_{1}$ (resp. $x_{2}$), and such that
$\forall(z_{1},z_{2})\in{\makebox{\boldmath$z$}}$,
$\forall(x_{1},x_{2})\in{\makebox{\boldmath$x$}}$,
$z_{1}=f_{1}(x_{1},g_{2}(z_{1},x_{1}))$ and
$z_{2}=f_{2}(g_{1}(z_{2},x_{2}),x_{2})$. Using the Brouwer fixed point theorem
on the continuous map
$g~{}:(x_{1},x_{2})\rightarrow(g_{1}(z_{2},x_{2}),g_{2}(z_{1},x_{1}))$ on the
compact set ${\makebox{\boldmath$x$}}_{1}\times{\makebox{\boldmath$x$}}_{2}$,
then $\forall(z_{1},z_{2})\in{\makebox{\boldmath$z$}}$,
$\exists(x^{z}_{1},x^{z}_{2})\in{\makebox{\boldmath$x$}}$ such that
$(z_{1},z_{2})=f(x^{z}_{1},x^{z}_{2})$.
This result can be generalized to functions
$f~{}:\mathbb{R}^{m}\rightarrow\mathbb{R}^{n}$ for any $n$, as stated in
Theorem 3.
###### Theorem 3 (Theorem 3 in Goubault and Putot (2020))
Let $f:\mathbb{R}^{m}\rightarrow\mathbb{R}^{n}$ be an elementary function and
$\pi:[1\ldots m]\rightarrow[1\ldots n]$ Let us note, for all $i\in[1\ldots
n]$, $J_{E}^{(z_{i})}=\\{j\in[1\ldots m],\;\pi(j)=i\\}$ and
$J_{A}^{(z_{i})}=\\{j\in[1\ldots m]\\}\setminus J_{E}^{(z_{i})}$. Consider the
$n$ AE-extensions $i\in[1\ldots n]$ built from Theorems 2, 4 or 5 and such
that
$\forall z_{i}\in{\makebox{\boldmath$z$}}_{i},\>(\forall
x_{j}\in{\makebox{\boldmath$x$}}_{j})_{j\in J_{A}^{(z_{i})}},\>(\exists
x_{j}\in{\makebox{\boldmath$x$}}_{j})_{j\in J_{E}^{(z_{i})}},\>z_{i}=f_{i}(x)$
(5)
Then
${\makebox{\boldmath$z$}}={\makebox{\boldmath$z$}}_{1}\times{\makebox{\boldmath$z$}}_{2}\times\ldots\times{\makebox{\boldmath$z$}}_{n}$,
if non-empty, is an under-approximation of the image of $f$: $\forall
z\in{\makebox{\boldmath$z$}},\,\exists x\in{\makebox{\boldmath$x$}},\>z=f(x).$
Theorem 3 gives us directly a computation of an under-approximation of
$\mbox{range}(f,{\makebox{\boldmath$x$}})$ for
$f:\mathbb{R}^{m}\rightarrow\mathbb{R}^{n}$. It can also be used to compute an
under-approximation of the robust range
$\mbox{range}(f,{\makebox{\boldmath$x$}},I_{\mathcal{A}},I_{\mathcal{E}})$.
For this, we need to choose $\pi:[1\ldots m]\rightarrow([1\ldots n]\setminus
I_{\mathcal{A}})$, which corresponds to the fact that the disturbance part of
the input components will always be quantified universally. We define below
the result of this process, which will be later used in reachability
algorithms for discrete-time dynamical systems.
###### Definition 2.2
Let $f:\mathbb{R}^{m}\rightarrow\mathbb{R}^{n}$ and $\pi:[1\ldots
m]\rightarrow([1\ldots n]\setminus I_{\mathcal{A}})$. We define
${\mathcal{I}}(f,{\makebox{\boldmath$x$}},I_{\mathcal{A}},I_{\mathcal{E}},\pi)$
an under-approximation of
$\mbox{range}(f,{\makebox{\boldmath$x$}},I_{\mathcal{A}},I_{\mathcal{E}})$
obtained using Theorem 3 with function $\pi$, in which the under-approximation
of each component is obtained using Theorem 2 (or Corollary 3.2). We define
${\mathcal{O}}(f,{\makebox{\boldmath$x$}},I_{\mathcal{A}},I_{\mathcal{E}},\pi)$
the over-approximation of the robust range obtained using Theorem 2 component-
wise.
## 3 Generalization to new AE extensions
We now introduce new robust AE extensions for a function
$f:\mathbb{R}^{m}\rightarrow\mathbb{R}$, which are no longer necessarily based
on the mean-value theorem.
###### Theorem 4
Suppose we have an approximation function $g$ for $f$, which is an
elementary111Elementary functions are compositions of +, -, $\times$, /, sine,
cosine, log, exp functions in particular. function in the sense of Goldsztejn
(2012a), satisfying $\forall w\in{\bf x}_{\mathcal{A}},\ \forall u\in{\bf
x}_{\mathcal{E}},\ \exists\xi\in{\bf x},\ f(w,u)=g(w,u,\xi)$. Then any under-
approximation (resp. over-approximation) of the robust range of $g$ with
respect to $x_{\mathcal{A}}$ and $\xi$,
${\mathcal{I}}_{g}\subseteq\mbox{range}(g,{\bf x}\times{\bf
x},I_{\mathcal{A}}\cup\\{m+1,\ldots,2m\\},I_{\mathcal{E}})$ is an under-
approximation (resp. over-approximation) of the robust range of $f$ with
respect to $x_{\mathcal{A}}$, i.e.
${\mathcal{I}}_{g}\subseteq\mbox{range}(f,{\bf
x},I_{\mathcal{A}},I_{\mathcal{E}})$.
For instance, for a continuously $(n+1)$-differentiable $f$, the following
$g$, obtained by a Taylor-Lagrange expansion and noting $x=(w,u)$, is an
approximation function for $f$
$g(x,\xi)=f(x^{0})+\sum\limits_{i=1}^{n}\frac{(x-x^{0})^{i}}{i!}D^{i}f(x^{0})\\\
+D^{n+1}f(\xi)\frac{(x-x^{0})^{n+1}}{(n+1)!}$ (6)
where $D^{\alpha}f$ denotes the higher order partial derivative of $f$. For
$n=0$, $g$ is the mean-value approximation.
We focus on the under-approximation. As $g$ is elementary, by Proposition 10.1
of Goldsztejn (2012a), we have a continuous Skolem function
$(w,u)\rightarrow\xi(w,u)$, i.e. a function such that for all
$x=(w,u)\in\mathbb{R}^{m}$, $f(w,u)=g(w,u,\xi(x))$. Consider an under-
approximation ${\mathcal{I}}_{g}$ of the robust range of $g$ with respect to
$\xi$ and $x_{\mathcal{A}}$. It satisfies $\forall z\in{\mathcal{I}}_{g},\
\forall w\in{\bf x}_{\mathcal{A}},\ \forall\xi\in{\bf x},\ \exists u\in{\bf
x}_{\mathcal{E}},\ z=g(w,u,\xi)$. Let $z\in{\mathcal{I}}_{g}$ and $w\in{\bf
x}_{\mathcal{A}}$ be fixed. As $g$ is elementary, we have a corresponding
continuous Skolem function $\xi\rightarrow u(\xi)$, i.e. a function such that
for all $\xi\in{\bf x}$, $z=g(w,u(\xi),\xi)$. For this $z\in{\mathcal{I}}_{g}$
and $w\in{\bf x}_{\mathcal{A}}$, the continuous map $u\rightarrow u(\xi(w,u))$
defined from ${\bf x}_{\mathcal{E}}$ over itself, has a fixed point
$u^{\infty}$, by Brouwer’s theorem. It is such that
$z=f(w,u^{\infty})=g(w,u^{\infty},\xi(w,u^{\infty}))$. Hence $z$ is in the
robust range of $f$ with respect to $x_{\mathcal{A}}$.
###### Example 3.1
Consider function $f(x)=x^{3}+x^{2}+x+1$ on $[-\frac{1}{4},\frac{1}{4}]$. The
exact range is $[0.796875,1.328125]$. Let us approximate $f$ by a quadratic
function, using an order 2 Taylor-Lagrange expansion. We compute
$f^{(1)}(x)=3x^{2}+2x+1$ and $f^{(2)}(x)=6x+2$. By Theorem 4, the range of $f$
over $[-\frac{1}{4},\frac{1}{4}]$ is under (resp. over) approximated by any
under (resp. over) approximation of the robust range with respect to $\xi$ of
$\begin{array}[]{rcl}g(x,\xi)&=&f(x^{0})+(x-x^{0})f^{(1)}(x^{0})+f^{(2)}(\xi)\frac{(x-x^{0})^{2}}{2}\\\
&=&1+x+x^{2}(3\xi+1)\end{array}$
In the general case, it may still be difficult to compute the under-
approximated robust range of $g$. However, Theorem 5 gives a simple way which
is well suited in particular for quadratic Taylor-based approximations.
###### Theorem 5
Let $g$ be an elementary function $g(w,u,\xi)=\alpha(w,u)+\beta(w,u,\xi)$ over
$x=(w,u)\in{\bf x}\subseteq{\mathbb{I}\mathbb{R}}^{m}$ and $\xi\in{\bf x}$.
Let ${\mathcal{I}}_{\alpha}$ be an under-approximation of the robust range of
$\alpha$ with respect to $w$, i.e. $\mbox{range}(\alpha,{\bf
x},I_{\mathcal{A}},I_{\mathcal{E}})$, and ${\mathcal{O}}_{\beta}$ an over-
approximation of the range of $\beta$, i.e. $\mbox{range}(\beta,{\bf
x}\times{\bf x},\emptyset,\\{1,\ldots,2m\\})$.
The robust range of $g$ with respect to $w\in{\bf x}_{\mathcal{A}}$ and
$\xi\in{\bf x}$, i.e. $\mbox{range}(g,{\bf x}\times{\bf
x},I_{\mathcal{A}}\cup\\{m+1,\ldots,2m\\},I_{\mathcal{E}})$, is under-
approximated by
${\mathcal{I}}_{g}=[\underline{\mathcal{I}}_{\alpha}+\overline{\mathcal{O}}_{\beta},\overline{\mathcal{I}}_{\alpha}+\underline{\mathcal{O}}_{\beta}].$
This is the case of Taylor expansions (6), where $\alpha$ is the degree $n$
polynomial, and $\beta$ the degree $n+1$ remainder.
${\mathcal{I}}_{\alpha}$ under-approximating $\mbox{range}(\alpha,{\bf
x},I_{\mathcal{A}},I_{\mathcal{E}})$ means $\forall
a\in{\mathcal{I}}_{\alpha},\ \forall w\in{\bf x}_{\mathcal{A}},\ \exists
u\in{\bf x}_{\mathcal{E}},\ a=\alpha(w,u)$. As $\alpha$ is elementary, we have
a continuous Skolem function $(w,a)\rightarrow u(w,a)$. Moreover, for all $z$
in ${\mathcal{I}}_{g}$, for all $b\in{\mathcal{O}}_{\beta}$, we have
$z-b\in[\underline{\mathcal{I}}_{\alpha}+\overline{\mathcal{O}}_{\beta}-\overline{\mathcal{O}}_{\beta},\overline{\mathcal{I}}_{\alpha}+\underline{\mathcal{O}}_{\beta}-\underline{\mathcal{O}}_{\beta}]={\mathcal{I}}_{\alpha}$.
For given $z\in{\mathcal{I}}_{g}$, $w\in{\bf x}_{\mathcal{A}}$ and $\xi\in{\bf
x}$, consider the continuous function $b\rightarrow\beta(w,u(w,z-b),\xi)$ over
${\mathcal{O}}_{\beta}$. By Brouwer fixed point theorem we have
$b^{\infty}=\beta(w,u(w,z-b^{\infty}),\xi)$. Therefore, for any
$z\in{\mathcal{I}}_{g}$, $\xi\in{\bf x}$, there exist
$a=z-b^{\infty}\in{\mathcal{I}}_{\alpha}$,
$b=b^{\infty}\in{\mathcal{O}}_{\beta}$ and $x=(w,u(w,a))\in{\bf x}$ such that
$z=a+b$. This implies that ${\mathcal{I}}_{g}$ is a robust under-approximation
of $g(x,\xi)=\alpha(x)+\beta(x,\xi)$ with respect to $w$ and $\xi$.
A direct consequence is a simple order 2 under-approximating Taylor method:
###### Corollary 3.2
Consider $f:\mathbb{R}^{m}\rightarrow\mathbb{R}$ a function in $C^{2}$. Let
${\makebox{\boldmath$f$}}^{0}$, ${\makebox{\boldmath$\nabla$}}_{w}^{0}$ and
${\makebox{\boldmath$\nabla$}}_{u}^{0}$ be such that
$f(x^{0})\subseteq{\makebox{\boldmath$f$}}^{0}$,
$\left|\nabla_{w}f(x^{0})\right|\subseteq{\makebox{\boldmath$\nabla$}}_{w}^{0}$
and
$\left|\nabla_{u}f(x^{0})\right|\subseteq{\makebox{\boldmath$\nabla$}}_{u}^{0}$
with $x^{0}=c({\makebox{\boldmath$x$}})$.
Then
$\mbox{range}(f,{\makebox{\boldmath$x$}},I_{\mathcal{A}},I_{\mathcal{E}})$ is
under-approximated by
$[\underline{\mathcal{I}}_{\alpha}+\overline{\mathcal{O}}_{\beta},\overline{\mathcal{I}}_{\alpha}+\underline{\mathcal{O}}_{\beta}]$
where
${\mathcal{I}}_{\alpha}=[\overline{f^{0}}-\langle\underline{\nabla}_{u}^{0},r({{\makebox{\boldmath$x$}}_{\mathcal{E}}})\rangle+\langle\overline{\nabla}_{w}^{0},r({{\makebox{\boldmath$x$}}_{\mathcal{A}}})\rangle,$
$\underline{f^{0}}+\langle\underline{\nabla}_{u}^{0},r({{\makebox{\boldmath$x$}}_{\mathcal{E}}})\rangle-\langle\overline{\nabla}_{w}^{0},r({{\makebox{\boldmath$x$}}_{\mathcal{A}}})\rangle]$
and ${\mathcal{O}}_{\beta}$ is any over-approximation of
$\\{\frac{1}{2}D^{2}f(x)(r({\makebox{\boldmath$x$}}))^{2},x\in{\makebox{\boldmath$x$}}\\}$.
This is a direct application of Theorems 4 and 5 where $g$ is the 2nd order
Taylor approximant, i.e. given by Equation (6) for $n=1$, combined with
Theorem 2 to compute the robust under-approximation of the order 1
approximation. We use a particular case of Theorem 2: the under-approximation
of an order 1 polynomial is almost trivial, we can compute it exactly if the
computation is performed in real numbers with exact evaluation of $f$ and its
gradient at a point. The expression we give here accounts for computation
errors.
###### Example 3.3
We carry on with Example 3.1. The computation of the under approximation of
the range of $1+x$ over $[-\frac{1}{4},\frac{1}{4}]$ yields
$[\frac{3}{4},\frac{5}{4}]$. We also need an over approximation of the range
of $x^{2}(3\xi+1)$ for $x$ and $\xi$ in $[-\frac{1}{4},\frac{1}{4}]$. Standard
interval computation yields
$[0,\frac{1}{16}][\frac{1}{4},\frac{7}{4}]=[0,\frac{7}{64}]$. Overall, we
deduce $[0.859375,1.25]\subseteq\mbox{range}(f,{\makebox{\boldmath$x$}})$. In
comparison, the mean-value AE extension of Theorem 2 would have given us the
less precise under-approximation $[0.875,1.125]$.
## 4 Refinements: preconditioning and quadrature formulae
The n-dimensional inner boxes that we compute with the techniques of Section
2.2 can sometimes be small or empty, even when the projected inner-
approximations on each component are tight. There are different reasons, for
which we propose solutions in this section.
### 4.1 Preconditioning for computing inner skewed boxes
The first difficulty is when the image of the vector-valued function cannot be
precisely approximated by a centered box.
###### Example 4.1
We consider $f(x)=(2x_{1}^{2}-x_{1}x_{2}-1,x_{1}^{2}+x_{2}^{2}-2)^{\intercal}$
with ${\makebox{\boldmath$x$}}=[0.9,1.1]^{2}$. The under-approximated
projections on the two components (respectively $[-0.38,0.38]$ and
$[-0.38,0.38]$) are close to the over-approximated ranges
($[-0.42,0.42]^{2}$), but we only find empty inner boxes.
This problem can be partly solved, as already described in Goubault and Putot
(2020) by computing a skewed box as under-approximation, that is the image of
a box by a linear map, instead of a box. This can be achieved by combining
preconditioning to the mean-value theorem. Let $C\in\mathbb{R}^{n\times n}$ be
a non-singular matrix. If $z$ is an interval vector such that
${\makebox{\boldmath$z$}}\subseteq\mbox{range}(Cf,{\makebox{\boldmath$x$}})$ ,
we can deduce a skewed box to be in the range of $f$, that is
$\\{C^{-1}z|z\in{\makebox{\boldmath$z$}}\\}\subseteq\mbox{range}(f,{\makebox{\boldmath$x$}})$.
A natural choice for $C$ is the inverse of the center of the interval Jacobian
matrix $C=(c({\makebox{\boldmath$\nabla$}}))^{-1}$.
###### Example 4.2
On Example 4.1, using this preconditioning and $pi:(1\rightarrow
1,2\rightarrow 2)$, we obtain for $f_{2}$ as a function of $f_{1}$ the yellow
under-approximating parallelotope of Figure 1(a). We estimate the image
$\mbox{range}(f,{\makebox{\boldmath$x$}})$ by sampling points in the input
domain. This sampling-based estimation is represented as the dark dots-filled
region. The green parallelotope and box are the over-approximations with and
without preconditioning.
(a) Example 4.2: under- (dotted lines) and over-approximation (plain lines)
$x^{0}$$x^{1}$${\makebox{\boldmath$x$}}^{k}={\makebox{\boldmath$x$}}\setminus{\makebox{\boldmath$x$}}^{k-1},\nabla^{k}=[\lvert\nabla
f\rvert]({\makebox{\boldmath$x$}}^{k})$$x_{1}^{-k}$$x_{1}^{-1}$$x_{1}^{0}$$x_{1}^{1}$$x_{1}^{k}$$x_{2}^{-k}$$x_{2}^{0}$$x_{2}^{k}$${\makebox{\boldmath$x$}}^{1},\nabla^{1}$${\makebox{\boldmath$x$}}^{2},\nabla^{2}$
(b) Partitioning the input domain
(c) Example 4.3: approximations for quadrature and order 2 extensions
Figure 1: Illustrations for Sections 2.2 and 4
### 4.2 Quadrature formulae for the mean-value extension
The mean-value interval extension can yield a rough approximation. This is
especially the case when the variation of the gradient is important over the
input range, the extreme case for under-approximation being when this
variation contains zero: the under-approximation is empty or reduced to a
point. Using simple quadrature formulae partially solves this problem.
Let $f:\mathbb{R}^{m}\rightarrow\mathbb{R}$. We partition each dimension
$j=[1\ldots m]$ of the $m$-dimensional input box
${\makebox{\boldmath$x$}}={\makebox{\boldmath$x$}}_{1}\times\ldots\times{\makebox{\boldmath$x$}}_{m}$
in $2k$ sub-intervals and define, for all $j=[1\ldots m]$, $x_{j}^{-k}\leq
x_{j}^{-(k-1)}\leq\ldots\leq x_{j}^{0}\leq\ldots\leq x_{j}^{k}$, with
$x_{j}^{-k}=\underline{{\makebox{\boldmath$x$}}_{j}}$,
$x_{j}^{0}=\mbox{mid}({{\makebox{\boldmath$x$}}_{j}})$,
$x_{j}^{k}=\overline{{\makebox{\boldmath$x$}}_{j}}$. We note
$dx^{i}=x^{i}-x^{i-1}$ the vector-valued deviation.
Let us refine the mean-value AE extensions using such a partition. The first
natural idea is to compute an under-approximation for each sub-box obtained as
product of sub-intervals in each dimension. But in general, the under-
approximating boxes will be non-contiguous, and their convex union is in
general not an under-approximation of
$\mbox{range}(f,{\makebox{\boldmath$x$}})$. Moreover, this approach would not
scale well.
We now propose a scheme that avoids these unions, and remains linear in $k$
with respect to the non-partitioned case. We note
${\makebox{\boldmath$x$}}^{1}=[x_{1}^{-1},x_{1}^{1}]\times[x_{2}^{-1},x_{2}^{1}]\times\ldots\times[x_{m}^{-1},x_{m}^{1}]$,
and for all $i$ between $2$ and $k$,
${\makebox{\boldmath$x$}}^{i}=[x_{1}^{-i},x_{1}^{i}]\times\ldots\times[x_{m}^{-i},x_{m}^{i}]\setminus\mathring{{\makebox{\boldmath$x$}}}^{i-1}$,
where $\setminus$ denotes the set difference and
$\mathring{{\makebox{\boldmath$x$}}}$ the interior of $x$. This partition is
represented in Figure 1(b) for a two-dimensional input space. In practice,
each ”square ring” ${\makebox{\boldmath$x$}}^{i}$ will be decomposed in $2n$
sub-boxes for the Jacobian evaluation.
By the mean-value theorem, $\forall x\in[x^{-1},x^{1}]$,
$\exists\xi^{1}\in[x^{-1},x^{1}],$ $f(x)=f(x^{0})+\langle\nabla
f(\xi^{1}),x-x^{0}\rangle$. Suppose we can compute
${\makebox{\boldmath$f$}}^{0}\supseteq f(x^{0})$ and
${\makebox{\boldmath$\nabla$}}^{i}$ for $i$ in $[1,k]$ such that
$\\{\lvert\nabla
f(x)\rvert,x\in{\makebox{\boldmath$x$}}^{i}\\}\subseteq{\makebox{\boldmath$\nabla$}}^{i}$.
We have
$\mbox{range}(f,{\makebox{\boldmath$x$}}^{1})\subseteq{\makebox{\boldmath$f$}}^{0}+\langle\overline{\nabla}^{1},dx^{1}\rangle[-1,1]$
and
$[\overline{f^{0}}-\langle\underline{\nabla}^{1},dx^{1}\rangle,\underline{f^{0}}+\langle\underline{\nabla^{1}},dx^{1}\rangle]\subseteq\mbox{range}(f,[x^{-1},x^{1}])$.
Let us now take $x\in{\makebox{\boldmath$x$}}^{2}$. We can iterate the mean-
value theorem on the adjacent input subdivision and write that for all
$x\in{\makebox{\boldmath$x$}}^{2}$, there exists
$x^{1}\in{\makebox{\boldmath$x$}}^{1}\cap{\makebox{\boldmath$x$}}^{2}$ (that
is on the border between ${\makebox{\boldmath$x$}}^{1}$ and
${\makebox{\boldmath$x$}}^{2}$), there exists
$\xi^{2}\in{\makebox{\boldmath$x$}}^{2}$ such that
$f(x)=f(x^{1})+\langle\nabla f(\xi^{2}),x-x^{1}\rangle$ and $\lvert
x_{1}-x^{1}_{1}\rvert\leq dx_{1}^{2}$ and $\lvert x_{2}-x^{1}_{2}\rvert\leq
dx_{2}^{2}$. (take for example for $x^{1}$ the intersection of the line from
$x^{0}$ to $x$ with the border between ${\makebox{\boldmath$x$}}^{1}$ and
${\makebox{\boldmath$x$}}^{2}$). We have
$\mbox{range}(f,{\makebox{\boldmath$x$}}^{1}\cup{\makebox{\boldmath$x$}}^{2})\subseteq{\makebox{\boldmath$f$}}^{0}+\langle\overline{\nabla}^{1},dx^{1}\rangle[-1,1]+\langle\overline{\nabla}^{2},dx^{2}\rangle[-1,1].$
There also exists
$(x,x^{1})\in{\makebox{\boldmath$x$}}^{2}\times{\makebox{\boldmath$x$}}^{1}$
such that $\lvert x_{1}-x^{1}_{1}\rvert=dx^{2}_{1}$ and $\lvert
x_{2}-x^{1}_{2}\rvert=dx^{2}_{2}$ (take the corners of the boxes
${\makebox{\boldmath$x$}}^{1}$ and ${\makebox{\boldmath$x$}}^{2}$), so that we
also have
$[\overline{f^{0}}-\langle\underline{\nabla}^{1},dx^{1}\rangle-\langle\underline{\nabla}^{2},dx^{2}\rangle,\underline{f^{0}}+\langle\underline{\nabla}^{1},dx^{1}\rangle+\langle\underline{\nabla}^{2},dx^{2}\rangle]\subseteq\mbox{range}(f,{\makebox{\boldmath$x$}}^{1}\cup{\makebox{\boldmath$x$}}^{2}).$
This generalizes to the $k$ subdivisions:
$\displaystyle\mbox{range}(f,{\makebox{\boldmath$x$}})\subseteq{\makebox{\boldmath$f$}}^{0}+\sum_{i=1}^{k}\langle\overline{\nabla}^{i},dx^{i}\rangle[-1,1]$
(7)
$\displaystyle[\overline{f^{0}}-\sum_{i=1}^{k}\langle\underline{\nabla}^{i},dx\rangle,\underline{f^{0}}+\sum_{i=1}^{k}\langle\underline{\nabla}^{i},dx\rangle]\subseteq\mbox{range}(f,{\makebox{\boldmath$x$}})$
(8)
The same idea applies to the estimation of robust ranges.
Naturally, other schemes can be proposed, relying on the idea that this
technique can be seen as using a quadrature formula for integrating the
Jacobian of a function.
###### Example 4.3
We consider
$f(x)=(2x_{1}^{2}+2x_{2}^{2}-2x_{1}x_{2}-2,x_{1}^{3}-x_{2}^{3}+4x_{1}x_{2}-3)^{\intercal}$
with ${\makebox{\boldmath$x$}}=[0.9,1.1]^{2}$. The results are represented in
Figure 1(c). The sampling-based estimation of the image is the dark dots-
filled region. We choose $pi:(1\rightarrow 1,2\rightarrow 2)$. Using the
preconditioned mean-value extension without partitioning, the over-
approximation is the largest green parallelotope and the under-approximation
for the joint range is empty. The quadrature formula for the mean-value
extension with $k=10$ partitions on one hand, and the order 2 extension of
Corollary 3.2 on the other hand, yield two very similar under-approximating
yellow parallelotopes. They also yield two very similar green over-
approximating parallelotopes. Both approaches actually have comparable
precision on the different examples tested. The light green box is the order 2
over-approximation without preconditioning.
###### Remark 4.4
One could be tempted to use more partitions to improve the quality of the
approximation. However, the computations yield approximations that are
centered at $f(x^{0})$. We can observe that the under-approximating skewed box
is actually already very close to being the largest skewed box entirely
included in the image, given a fixed skewing and a center at $f(x^{0})$. The
order 2 estimation allows for slight decentering, but it would also be
possible to use less basic quadrature formulae for that purpose. Quadrature
can also be combined with order 2 extensions.
Quadrature can also be combined with some classical partition of the inputs,
or with different center choices. However, a possibly disjoint union of
approximations (corresponding to a classical partition of inputs) is
inconvenient, so we would recommend this use only as a property-driven
refinement. Finally, as a sound under-approximation is still obtained by
considering a sub-region of the input set, refinements can be obtained by
detecting and removing sub-regions where Jacobian coefficients are either very
sensitive to the inputs or close to zero.
### 4.3 Bounding the Jacobian matrix
The approach relies on being able to compute over-approximations of $\nabla
f(x)$ over some sub-sets of input box $x$, namely
${\makebox{\boldmath$\nabla$}}^{i}$ for $i$ in $[1,k]$ such that
$\\{\lvert\nabla
f(x)\rvert,x\in{\makebox{\boldmath$x$}}^{i}\\}\subseteq{\makebox{\boldmath$\nabla$}}^{i}$.
Automatic differentiation allows to compute the derivatives, but need to be
combined with set-membership methods to handle uncertainties. We have found
the combination of automatic differentiation with an evaluation in affine
arithmetic to provide a good trade-off between efficiency and precision. All
the experiments presented in this work were performed with an implementation
relying on this combination. Affine forms provide an interesting combination
of parameterization and set-based estimation: a parametric approximate form
for $\nabla f(x)$ valid on all box $x$ is computed, that can be instantiated
on ${\makebox{\boldmath$x$}}^{i}$ to yield tight over-approximations
$\nabla^{i}$, without the need for performing several evaluations of the
differentiation. The full description is out of scope, we give below a flavor
of the use of affine arithmetic for Jacobian estimation on a simple example.
###### Example 4.5
Consider in Example 4.3 the derivative of
$f_{1}(x_{1},x_{2})=2x_{1}^{2}+2x_{2}^{2}-2x_{1}x_{2}-2$ with respect to
$x_{1}$, with $(x_{1},x_{2})\in[0.9,1.1]^{2}$. This derivative is
$\nabla_{1}f_{1}(x_{1},x_{2})=4x_{1}-2x_{2}$. Evaluation with affine
arithmetic first consists in creating a centered form with a fresh noise
symbol for each input: $\hat{x}_{1}=1.+0.1\varepsilon_{1}$ and
$\hat{x}_{2}=1.+0.1\varepsilon_{2}$, with
$(\varepsilon_{1},\varepsilon_{2})\in[-1,1]^{2}$. The gradient evaluated on
these affine forms is
$\hat{\nabla}_{1}\hat{f}_{1}(x)=2+0.4\varepsilon_{1}-0.2\varepsilon_{2}.$ Here
the abstraction is exact. In the general case, affine arithmetic will compute
an approximate affine form and bound the approximation error in a new noise
term. Let us now use this affine form to over-approximate $\nabla_{1}f_{1}(x)$
over all $x$. This amounts to computing the interval bounds when
$\varepsilon_{1}$ and $\varepsilon_{2}$ range in $[-1,1]$, for which we get
${\makebox{\boldmath$\nabla$}}_{1,1}=[1.4,2.6]$. Now we use the same affine
form to over-approximate $\nabla_{1}f_{1}(x)$ over
${\makebox{\boldmath$x$}}^{1}$. This amounts to take $\varepsilon_{1}$ and
$\varepsilon_{2}$ both ranging in $[-\frac{1}{k},\frac{1}{k}]$. Let us
consider $k=10$ subdivisions, we obtain
${\makebox{\boldmath$\nabla$}}^{1}_{1,1}=2+0.4[-0.1,0.1]-0.2[-0.1,0.1]=[1.94,2.06]$.
The process can be iterated to compute
${\makebox{\boldmath$\nabla$}}^{i}_{1,1}$ for $i\in[2\ldots k]$, expressing
the membership to a subset ${\makebox{\boldmath$x$}}^{i}$ of $x$ as
constraints on $\varepsilon_{1}$ and $\varepsilon_{2}$, used to instantiate to
${\makebox{\boldmath$x$}}^{i}$ the estimation of $\nabla_{1}f_{1}(x)$. For
instance, in order to compute ${\makebox{\boldmath$\nabla$}}^{2}_{1,1}$, we
decompose the computation on the 4 rectangles that define
${\makebox{\boldmath$x$}}_{2}$ (see Figure 1(b)):
$(\varepsilon_{1},\varepsilon_{2})\in([\frac{1}{k},\frac{2}{k}]\times[-\frac{2}{k},\frac{2}{k}])\cup([-\frac{2}{k},-\frac{1}{k}]\times[-\frac{2}{k},\frac{2}{k}])\cup([-\frac{1}{k},\frac{1}{k}]\times[-\frac{2}{k},-\frac{1}{k}])\cup([-\frac{1}{k},\frac{1}{k}]\times[\frac{1}{k},\frac{2}{k}])$.
## 5 Application to the reachability of discrete-time systems
We consider discrete-time non-linear dynamical systems with inputs of the form
$\begin{cases}z^{k+1}=f(z^{k},u^{k})\\\
z^{0}\in{\makebox{\boldmath$z$}}^{0}\end{cases}$ (9)
where $f:\mathbb{R}^{m}\rightarrow\mathbb{R}^{n}$ is a vector-valued non-
linear function with $m\geq n$, $z\in\mathbb{R}^{n}$ the vector of state
variables, $u\in{{\makebox{\boldmath$u$}}}\subseteq\mathbb{R}^{m-n}$ the input
signal, and ${\makebox{\boldmath$z$}}^{0}$ the initial set.
Given an initial set ${\makebox{\boldmath$z$}}^{0}$, we want to compute the
bounded time reachable set of the dynamical system, i.e, the set of states
visited by the dynamical system up to a fixed time horizon $K\in\mathbb{N}$.
The reachable set can be obtained as the solution of the recursion
${\makebox{\boldmath$z$}}^{k+1}=\\{f(z^{k},u^{k})|z^{k}\in{\makebox{\boldmath$z$}}^{k},u^{k}\in{{\makebox{\boldmath$u$}}}\\}$,
for $k\in[0,K]$. The computation of the reachable set can be seen as a series
of images of sets by vector-valued function $f$. We thus can use the results
of Sections 2 to 4 to compute approximations of these reachable sets.
For conciseness, we consider systems without disturbances and compute maximal
(or classical) reachable sets. The algorithms can be straightforwardly
extended to robust reach set of systems with disturbances, basically replacing
ranges by robust ranges. This allows us to use the lighter notations
${\mathcal{I}}(f,{\makebox{\boldmath$x$}},\pi)$ and
${\mathcal{O}}(f,{\makebox{\boldmath$x$}},\pi)$ to note the under and over-
approximating sets introduced in Definition 2.2.
#### 5.0.1 Method 1
the first method consists in iteratively using function image, independently
for under and over-approximation, taking as input the previously computed
approximation of the image. We compute under and over-approximations $I^{k}$
and $O^{k}$ of the reachable set ${\makebox{\boldmath$z$}}^{k}$ by
$\begin{cases}I^{0}={\makebox{\boldmath$z$}}^{0},\;O^{0}=z^{0}\\\
I^{k+1}={\mathcal{I}}(f,I^{k},\pi),\;O^{k+1}={\mathcal{O}}(f,O^{k},\pi)\end{cases}$
(10)
Indeed, at each step $k$, we have
$I^{k+1}\subseteq\mbox{range}(f,I^{k})\subseteq\mbox{range}(f,{\makebox{\boldmath$z$}}^{k})={\makebox{\boldmath$z$}}^{k+1}\subseteq\mbox{range}(f,O^{k})\subseteq
O^{k+1}$.
With this approach, at each step $k$, under and over-approximations $I^{k}$
and $O^{k}$ of the joint range are used as input for the next step. It is thus
particularly important to compute tight under and over-approximations of this
joint range. In particular, using the preconditioning of Section 4.1 will
often be crucial, both for under and over approximation. This yields Algorithm
1.
Algorithm 1 Iterated discrete-time reachability
0: $f:\mathbb{R}^{n}\rightarrow\mathbb{R}^{n}$,
${\makebox{\boldmath$z$}}^{0}\subseteq{\mathbb{I}\mathbb{R}}^{n}$ initial
state, $K\in\mathbb{N}^{+}$, an over-approximating extension $[\nabla f]$ (see
Section 4)
0: $I^{k}$ and $O^{k}$: under and over-approximations of the reachable set
$\mbox{range}(f^{k},{\makebox{\boldmath$z$}}^{0})$ for $k\in[1,K]$
$I^{0}\coloneqq{\makebox{\boldmath$z$}}^{0},O^{0}\coloneqq{\makebox{\boldmath$z$}}^{0}$;
choose $\pi:[1\ldots n]\mapsto[1\ldots n]$
for $k$ from 0 to $K-1$ do
${\makebox{\boldmath$\nabla$}}_{I}^{k}\coloneqq|[\nabla f](I^{k})|$,
${\makebox{\boldmath$\nabla$}}_{O}^{k}\coloneqq|[\nabla f](O^{k})|$
$A_{I}^{k}\coloneqq c({\makebox{\boldmath$\nabla$}}_{I}^{k})$,
$A_{O}^{k}\coloneqq c({\makebox{\boldmath$\nabla$}}_{O}^{k})$ (supposed non-
singular, otherwise taken to identity matrix)
$C_{I}^{k}\coloneqq(A_{I}^{k})^{-1}$, $C_{O}^{k}\coloneqq(A_{O}^{k})^{-1}$
${\makebox{\boldmath$z$}}_{I}^{k+1}\coloneqq{\mathcal{I}}(C_{I}^{k}f,I^{k},\pi)$,
${\makebox{\boldmath$z$}}_{O}^{k+1}\coloneqq{\mathcal{O}}(C_{O}^{k}f,O^{k},\pi)$
if ${\makebox{\boldmath$z$}}_{I}^{k}=\emptyset$ then
return
end
$I^{k+1}\coloneqq A_{I}^{k}{\makebox{\boldmath$z$}}_{I}^{k+1}$,
$O^{k+1}\coloneqq A_{O}^{k}{\makebox{\boldmath$z$}}_{O}^{k+1}$
end for
At each step $k=0\ldots K-1$, ${\makebox{\boldmath$z$}}_{I}^{k+1}$ is an
interval vector such that, if it is non empty,
$I^{k+1}=A_{I}^{k}{\makebox{\boldmath$z$}}_{I}^{k+1}\subseteq\mbox{range}(f,I^{k})\subseteq\mbox{range}(f^{k+1},{\makebox{\boldmath$z$}}^{0})$.
The over-approximation is computed similarly, and is fully decoupled.
#### 5.0.2 Method 2
the second method consists in computing the sensitivity to initial state by
approximating the gradient of the iterated function. At each step $k$, we
compute the under and over-approximation of
$\mbox{range}(f^{k},{\makebox{\boldmath$z$}}^{0})$, i.e. the loop body $f$
iterated $k$ times, starting from the initial state
${\makebox{\boldmath$z$}}^{0}$. This yields the schematic Algorithm 2, with
same inputs and hypotheses as in Algorithm 1.
Algorithm 2 Discrete-time reachability computed on $f^{k}$
for $k$ from 0 to $K-1$ do
$I^{k+1}\coloneqq{\mathcal{I}}(f^{k+1},{\makebox{\boldmath$z$}}^{0},\pi)$,
$O^{k+1}\coloneqq{\mathcal{O}}(f^{k+1},{\makebox{\boldmath$z$}}^{O},\pi)$
end for
Here, at each step $k$, the under- and over-approximation are both obtained
from an over-approximation of $f^{k+1}$ evaluated at the center of
${\makebox{\boldmath$z$}}^{0}$ and an over-approximation of the gradient of
$f^{k+1}$ over ${\makebox{\boldmath$z$}}^{0}$: at each step, the gradient can
be obtained by differentiating the gradient from the previous step. Of course,
this can also be combined with preconditioning.
#### 5.0.3 Discussion
While relying on the same techniques for range estimation, Algorithm 1 and
Algorithm 2 are different in spirit: Algorithm 2 relies only on the
propagation of over-approximations to deduce under-approximations. In
particular, the under-approximation may be empty at some step $k$, and become
non-empty again at further steps (a similar remark was made in the context of
continuous systems in Goubault and Putot (2017)). In comparison, Algorithm 1
needs at each step an under-approximating box or skew box that is non-empty
for all components. On the other hand, Algorithm 2 is more costly as it
requires a differentiation of the iterated function.
## 6 Implementation and examples
We now present results on small systems from the literature. The approach is
implemented as part of the RINO C++ prototype, available from
https://github.com/cosynus-lix/RINO. The prototype allows to experiment the
function range estimation, but its actual target is discrete and continuous-
time reachability, combining for continuous-time the techniques presented here
with Taylor model methods Goubault and Putot (2019, 2020). The prototype uses
the fadbad++ (http://www.fadbad.com/fadbad.html) automatic differentiation
library and the aaflib (http://aaflib.sourceforge.net/) affine arithmetic
library. The timings are given on a Macbook Pro 2.6GHz Intel Core i7 and 32Gb
of RAM.
##### Test Model
We consider the test model Dreossi et al. (2016):
$\begin{split}x_{1}^{k+1}&=x_{1}^{k}+(0.5(x_{1}^{k})^{2}-0.5(x_{2}^{k})^{2})\Delta\\\
x_{2}^{k+1}&=x_{2}^{k}+2x_{1}^{k}x_{2}^{k}\Delta\end{split}$
with as initial set a box $x_{1}\in[0.05,0.1]$ and $x_{2}\in[0.99,1.00]$, and
$\Delta=0.01$. Figure 2 shows the under and over-approximated reachable sets
(respectively the filled yellow region and green parallelotope) over time up
to 25 steps with Algorithm 1. They are obtained in 0.02 seconds. We can
observe that the under and over-approximations are very close one to another,
confirming the accuracy of the results.
Figure 2: Skewed box under and over-approximations for 25 steps (Algorithm 1)
for the test model
##### SIR Epidemic Model
We now consider the SIR epidemic model with the parameters of Dreossi et al.
(2016).
$\begin{split}x_{1}^{k+1}&=x_{1}^{k}-\beta x_{1}^{k}x_{2}^{k}\Delta\\\
x_{2}^{k+1}&=x_{2}^{k}+(\beta x_{1}^{k}x_{2}^{k}-\gamma x_{2}^{k})\Delta\\\
x_{3}^{k+1}&=x_{3}^{k}+\gamma x_{2}^{k}\Delta\end{split}$
We compute the reachable set up to 60 steps from the initial box
$(x_{1},x_{2},x_{3})\in[0.79,0.80]\times[0.19,0.20]\times[0,0.1]$. The
parameter values are $\beta=0.34$, $\gamma=0.05$, and $\Delta=0.5$.
The reachable sets computed in 0.05 seconds with Algorithm 1 up to 60 steps
are represented for $(x_{1},x_{2})$ in Figure 3.
Figure 3: Skewed box under and over-approximations for 60 steps (Algorithm 1)
of the SIR epidemic model
We can note in particular from the zoomed reachable set in the Figure, which
corresponds to last step (60), that the under-approximation (in yellow) is
still of good quality (the purple dots correspond to sample executions).
However, only Algorithm 2 is able to compute non-empty approximations when
taking as initial condition $x_{3}=0$ which is of empty interior, instead of
$x_{3}\in[0,0.1]$. We obtain in 0.05 seconds the very tight approximations of
the projections of the components represented Figure4(a).
(a) SIR epidemic model, 60 steps
(b) Bees model, 1500 steps
Figure 4: Projected under and over-approximations, Algorithm 2
##### Honeybees Site Choice Model
We consider the reachable sets up to 1500 steps of the model studied in
Dreossi et al. (2016):
$\begin{split}x_{1}^{k+1}&=x_{1}^{k}-(\beta_{1}x_{1}^{k}x_{2}^{k}+\beta_{2}x_{1}^{k}x_{3}^{k})\Delta\\\
x_{2}^{k+1}&=x_{2}^{k}+(\beta_{1}x_{1}^{k}x_{2}^{k}-\gamma
x_{2}^{k}+\delta\beta_{1}x_{2}^{k}x_{4}^{k}+\alpha\beta_{1}x_{2}^{k}x_{5}^{k})\Delta\\\
x_{3}^{k+1}&=x_{3}^{k}+(\beta_{2}x_{1}^{k}x_{3}^{k}-\gamma
x_{3}^{k}+\delta\beta_{2}x_{3}^{k}x_{5}^{k}+\alpha\beta_{2}x_{3}^{k}x_{4}^{k})\Delta\\\
x_{4}^{k+1}&=x_{4}^{k}+(\gamma
x_{2}^{k}-\delta\beta_{1}x_{2}^{k}x_{4}^{k}-\alpha\beta_{2}x_{3}^{k}x_{4}^{k})\Delta\\\
x_{5}^{k+1}&=x_{5}^{k}+(\gamma
x_{3}^{k}-\delta\beta_{2}x_{3}^{k}x_{5}^{k}-\alpha\beta_{1}x_{2}^{k}x_{5}^{k})\Delta\\\
\end{split}$
with as initial set the box $x_{1}=500$, $x_{2}\in[390,400]$,
$x_{3}\in[90,100]$, $x_{4}=x_{5}=0$ and the parameter values
$\beta_{1}=\beta_{2}=0.001$, $\gamma=0.3$, $\delta=0.5$, $\alpha=0.7$, and
$\Delta=0.01$.
Algorithm 1 runs very fast, taking only 1.7 seconds for the 1500 steps. But
the under-approximation is very soon empty and the over-approximation tends to
strongly widen after 800 steps. In comparison, Algorithm 2 takes 57 seconds to
complete the 1500 reachability steps. But the projected under-approximations
are very tight, close to the over-approximations, as can be seen on Figure
4(b) which represents under and over-approximations for all components as
functions of steps. These results should also be compared to the much wider
over-approximation of Dreossi et al. (2016) (Figure 7, where time is number of
steps divided by 100), obtained in 81 seconds.
## 7 Conclusion and future work
We focused on new AE under-approximating extensions and their accurate
practical evaluation for non-linear vector-valued functions, and exemplified
their interest for the reachability of discrete-time systems. These techniques
can also be used for the reachability analysis of continuous-time systems,
improving for instance over Goubault and Putot (2019, 2020).
## References
* Chen et al. (2016) Chen, M., Herbert, S., and Tomlin, C.J. (2016). Exact and efficient Hamilton-Jacobi-based guaranteed safety analysis via system decomposition.
* Chen et al. (2014) Chen, X., Sankaranarayanan, S., and Ábrahám, E. (2014). Under-approximate flowpipes for non-linear continuous systems. In _FMCAD_. IEEE.
* Dreossi et al. (2016) Dreossi, T., Dang, T., and Piazza, C. (2016). Parallelotope bundles for polynomial reachability. In _Proceedings of the 19th International Conference on Hybrid Systems: Computation and Control_. ACM Press.
* Girard et al. (2006) Girard, A., Le Guernic, C., and Maler, O. (2006). Efficient computation of reachable sets of linear time-invariant systems with inputs. In J.P. Hespanha and A. Tiwari (eds.), _Hybrid Systems: Computation and Control_ , 257–271. Springer Berlin Heidelberg, Berlin, Heidelberg.
* Goldsztejn (2012a) Goldsztejn, A. (2012a). Modal intervals revisited, part 1: A generalized interval natural extension. _Reliable Computing_ , 16, 130–183.
* Goldsztejn (2012b) Goldsztejn, A. (2012b). Modal intervals revisited, part 2: A generalized interval mean value extension. _Reliable Computing_ , 16, 184–209.
* Goldsztejn and Jaulin (2010) Goldsztejn, A. and Jaulin, L. (2010). Inner approximation of the range of vector-valued functions. _Reliable Computing_ , 14.
* Goubault and Putot (2017) Goubault, E. and Putot, S. (2017). Forward inner-approximated reachability of non-linear continuous systems. In _HSCC_. ACM.
* Goubault and Putot (2019) Goubault, E. and Putot, S. (2019). Inner and outer reachability for the verification of control systems. In _HSCC_.
* Goubault and Putot (2020) Goubault, E. and Putot, S. (2020). Robust under-approximations and application to reachability of non-linear control systems with disturbances. _IEEE Control Systems Letters_ , 4(4), 928–933.
* Guernic and Girard (2009) Guernic, C.L. and Girard, A. (2009). Reachability analysis of hybrid systems using support functions. In _Proceedings of 21st International Conference on Computer Aided Verification_ , LNCS.
* Kochdumper and Althoff (2020) Kochdumper, N. and Althoff, M. (2020). Computing non-convex inner-approximations of reachable sets for nonlinear continuous systems. In _CDC_.
* Korda et al. (2013) Korda, M., Henrion, D., and Jones, C.N. (2013). Inner approximations of the region of attraction for polynomial dynamical systems. In _9th IFAC Symposium on Nonlinear Control Systems, NOLCOS 2013_.
* Kurzhanski and Varaiya (2000) Kurzhanski, A.B. and Varaiya, P. (2000). Ellipsoidal techniques for reachability analysis. In N. Lynch and B.H. Krogh (eds.), _HSCC_ , 202–214.
* Makino and Berz (2003) Makino, K. and Berz, M. (2003). Taylor models and other validated functional inclusion methods. _Int. J. Pure Appl. Math_.
* Mitchell (2007) Mitchell, I.M. (2007). Comparing forward and backward reachability as tools for safety analysis. In _HSCC_.
* Mézo et al. (2018) Mézo, T.L., Jaulin, L., and Zerr, B. (2018). Bracketing the solutions of an ordinary differential equation with uncertain initial conditions. _Applied Mathematics and Computation_ , 318.
* Raković and Fiacchini (2008) Raković, S.V. and Fiacchini, M. (2008). Approximate reachability analysis for linear discrete time systems using homothety and invariance. _IFAC Proceedings Volumes_ , 41(2), 15327 – 15332.
* Xue et al. (2020) Xue, B., Fränzle, M., and Zhan, N. (2020). Inner-approximating reachable sets for polynomial systems with time-varying uncertainties. _IEEE Transactions on Automatic Control_ , 65(4), 1468–1483. 10.1109/TAC.2019.2923049.
* Xue et al. (2016) Xue, B., She, Z., and Easwaran, A. (2016). Under-approximating backward reachable sets by polytopes. In _Computer Aided Verification_. Springer.
|
# Multi-agent simulation of voter’s behaviour
Albin Soutif Carole Adam Sylvain Bouveret
Univ. Grenoble-Alpes, Grenoble Informatics Laboratory
(_This is an internship report originally written in June 2018 by A. Soutif,
ENSIMAG intern_
_under the supervision of C. Adam and S. Bouveret_ )
###### Abstract
The goal of this paper is to simulate the voters’ behaviour given a voting
method. Our approach uses a multi-agent simulation in order to model a voting
process through many iterations, so that the voters can vote by taking into
account the results of polls. Here we only tried basic rules and a single
voting method, but further attempts could explore new features.
Keywords: Computational social choice, Iterative voting, multi-agent
simulation
## 1 INTRODUCTION
A voting process involves the participation of many people that interact
together in order to reach a common decision. In this paper, we focus on
voting processes in which a single person is elected. A voting method is
defined as the set of rules that determine the winner of the election, given
an input from each voter, for example their preferred candidate or an order
relation between all candidates. Social Choice Theory is the field that
studies the aggregation of individual preferences towards a collective choice,
like for example electing a candidate or choosing a movie. Computational
social choice is a recent field which aim is to apply computer science to
social choice problems [3]. Here we aim to design a multi-agent simulation in
order to study the collective decision process involved in an election.
A multi-agent simulation is a simulation in which we make several autonomous
agent interact with each other in a defined environment. The goal of this kind
of simulation is to study the global behaviour emerging from the local actions
of the agents [4, 6, 5]. There are only a few attempts to mix the field of
multi-agent simulation and computational social choice to model a voting
process, and that is the kind of approach we use in this paper. The goal of
this approach is to study the dynamics of voting methods through many
iterations. Here, we will focus on one particular voting method which is
plurality voting. In this method, voters vote for only one candidate and the
winner of the election is the candidate with the most votes.
Our approach was supported by the results of a study on the French
presidential election of 2017 [7]. In this study, people were asked to vote
with different voting methods, and to tell for whom they officially voted. The
goal of this study was to better understand the properties of several voting
methods. Among these voting methods, one in particular helped to build this
model, because it gave a precise insight of voters’ voting intentions (not
just their favourite candidate), it is the continuous evaluation method. In
this method, a voter gives a mark between 0 and 100 to each of the 11
candidates. Thanks to this study, we had access to over 30 000 of these
continuous evaluations, which helped us initialise our model.
A voting process can be modelled intuitively as a multi-agent system. Indeed,
voters and candidates can be modelled as autonomous agents, the real world, or
the internet, is the environment in which these agents evolve, and the
political debates, polls, conversations … Are the interactions which can
influence their final decision. In a multi-agent simulation, it is reasonable
to limit ourselves to a simpler model, in which the agents evolve in a limited
environment and have much simpler interactions with each other. Otherwise it
would be too difficult to model, to interpret, and too computationally
expensive.
First we will review the state of the art in this field (Section 2), then we
will present the way we built our model (Section 3), then our results along
with our interpretation (Section 4), and eventually we will conclude and
discuss some possible extensions of the model (Section 5).
## 2 STATE OF THE ART
Some interesting attempts have been made at building voting process models.
Here we will review the ones that inspired us the most in building our model.
### 2.1 Static models
A static simulation has been made by Ka-Ping Yee [10] in which the voters are
modelled by 2-dimensional points. In this simulation a normal distribution of
voters is drawn around every point, and each point is coloured according to
the candidate that won the election with that distribution of voters. The
conclusion of these simulations highlighted the different behaviours of some
voting rules.
### 2.2 Iterative models
Airiau et al. [1] attempted to model the outcome of an election where voters
could react strategically by learning from the others and trying to react in
order to maximise an utility function. The authors found out that for a given
set of rule and a precise methodology, the outcome was convergent and had good
mathematical properties. The drawback is that for the convergence to be
certain, the voters had to vote one by one and know at each step which
candidate was winning, which is not a realistic procedure.
Another attempt has been made for educational purpose by a network of
researchers in multi-agent simulation (MAPS) [2]. They made a multi-agent
simulation called “the Iznogoud model”, with simple interactions between
2-dimensional agents. The interactions used were local (each voter directly
influenced its surroundings), whereas in our model there are no local
interactions because we only focused on the global effect of polls, and not on
the effect of local interactions such as political conversations between
individuals. Also in the Iznogoud model, the voters were initialised randomly,
making it impossible to compare the results to any real case scenario.
In both the approaches from Ka-Ping Yee [10] and the Iznogoud simulation [2],
the voters are placed in a 2-dimensional space. In Ka-Ping Yee simulation the
model is not dynamic, so such a model would not help us study the dynamics of
voting methods. In the Iznogoud simulation, the interactions between voters
are local and very simple, and the model is initialised with a random
population of voters. In our model, we will try to integrate interactions that
are more realistic, and to initialise the population of agents with the help
of the results of the ”voter autrement” experimentation [7].
## 3 OUR MODEL
We choose to model each voter as a point in a 11 dimensional space, and we
initialise the model with the study’s continuous evaluation results, giving as
coordinates to a voter the marks he has given to the different candidates.
Because we also know the official vote of each voter, we integrate the
candidates into that space by taking the centroids of their voter’s continuous
evaluation. The model is dynamic and at each iteration, the voters move and
then a plurality vote is computed. To compute this vote, we make each voter
vote for the closest candidate from him and the candidate with the most votes
is the winner. In order to make the agents interact between each other, we use
polls that give information to everyone on the current plurality vote outcome.
The voter compute their move taking into account the results of the last poll.
That’s the way the agents interact between each other.
The model ran with different set of utility functions and agent behaviour.
Basically, we make each agent move with a velocity vector computed at each
iteration, and several methods were used to compute this one. We will discuss
those methods and their expected properties in this section.
### 3.1 Formal notations
Formally, we note as $V_{i}$ the set of voters at iteration $i$, and $C$ the
(static) set of candidates. Both of these sets contain a finite number of
11-dimension vectors. Thus $v_{k,i}$ is the position of voter $v_{k}$ at
iteration $i$. $s_{i}$ is the last poll available at iteration $i$, also
represented as an 11-dimension vector, containing the number of votes for each
candidate in the poll’s sample.
$V(v_{k,i},s_{i})$ is the velocity vector of voter $v$, knowing its current
position and the result of the last poll. At each iteration, the next position
vector of each voter $v_{k,i+1}\in V_{i+1}$ is given by the equation (1), as
the sum of its previous position and its velocity vector computed from this
position and the last poll.
$v_{k,i+1}=v_{k,i}+V(v_{k,i},s_{i})$ (1)
### 3.2 Utility functions
Before introducing the movement rules of our model, we introduce the concept
of utility function that we will later use in these rules. A utility function
is defined for each voter, as the utility a candidate would have for this
voter if he was elected. We define a utility function
$u:\mathbb{R}^{11}\times\mathbb{R}^{11}\rightarrow\mathbb{R}$, giving for the
first point (a voter) the utility that the second point (a candidate) would
have if winning the election. In practice, we will use $u(v_{k,i},c)$ for
$v_{k,i}\in V_{i}$ and $c\in C$.
In our model, we want that utility function to be in line with the fact that
the voters vote for the closest candidate from them. So intuitively, the
closer a voter is to a candidate, the more he likes him. The main property
that we want from these functions is that they are decreasing functions of the
distance between the voter and the candidates, and that they are not linear
(to model that some voters are strongly attached to a candidate).
Noting $dist(v,c)$ the Euclidean distance between voter $v$ and candidate $c$,
we designed 2 different utility functions representing different reasoning.
The first utility function given by equation (2) below simply expresses
utility as inversely proportional to the distance with the candidate.
$u:v,c\rightarrow\frac{1}{dist(v,c)}$ (2)
The second utility function that we designed also models the fact that a voter
can be strongly opposed to a candidate when he is too far away from him
(further than a parameter $\alpha$). In this case this utility function
becomes negative, ensuring that the voter will not get any closer to this
candidate (repulsion). This second function is given by equation (3) below.
The drawback is that some voters just flee away from the whole set of
candidates, which we model as abstention in the vote.
$u:v,c\rightarrow\frac{\alpha-dist(v,c)}{(1+dist(v,c))^{2}}$ (3)
### 3.3 Velocity computation
In this section we describe three techniques that can be used to compute the
velocity vector of the voters.
#### First approach: 3-pragmatist rule
A first approach is to use a rule described in [1] which is the three
pragmatist rule. This rule makes the voter vote for his favourite candidate
amongst the first three candidates of the election, which represent a
strategic vote for a candidate who has a chance to win. We integrate that in
our model by making each voter move towards this candidate, once given the
results of the last poll. The 3-pragmatist velocity vector towards candidate
$c$, the closest to voter $v_{k,i}$ among the first 3 candidates in results of
poll $s_{i}$, is given by the following equation:
$V(v_{k,i},s_{i})=e_{v_{k,i},c}$ (4)
where $e_{a,b}$ is the unit vector from point $a$ to $b$.
#### Second approach: maximum expected utility
A second approach is to use a utility function and to move toward the
candidate that maximises the expected utility. In order to approximate this
expectation, we use the poll results to compute a probability that each
candidate wins, given by the following equation as the proportion of expressed
votes in favour of this candidate.
$p(wins(c)|s_{i})=\frac{numberofvotesforcins_{i}}{totalnumberofvotesins_{i}}$
(5)
Once we have the probability that each candidate wins, computed from the
polls, the expected utility of their victory is the product of this
probability by the utility of them winning. This approximation gives a chance
even to candidates who have a small number of votes, provided that the utility
of them winning is high enough to compensate. On the contrary, other
approaches would just dismiss these candidates altogether, considering that
voting for them is useless because they have no chance to win in the end.
Thus we can compute the candidate selected by the voter as the one that
maximises this expected utility, as expressed by the following equation:
$c=argmax_{c\in C}[u(v_{k,i},c)*p(wins(c)|s_{i})]$ (6)
In the end the velocity vector is computed as the unit vector from voter
$v_{k,i}$ to that candidate $c$:
$V(v_{k,i},s_{i})=e_{v_{k,i},c}$ (7)
#### Third approach: personalised opinion centre
The third approach is different in the sense that rather than moving towards a
single candidate, it makes the voter move towards a personalised opinion
centre. In this approach we compute velocity as a linear combination of the
expected utilities of the different candidates, as expressed by the following
equation.
$V(v_{k,i},s_{i})=\sum_{c\in C}e_{v_{k,i},c}*u(v_{k,i},c)*p(wins(c)|s_{i})$
(8)
Since movement towards each candidate is weighed by the expected utility of
that candidate, we can expect that this vector will somehow be pointing in a
direction that will make the voter closer to the candidate that maximises its
expected utility.
## 4 RESULTS
### 4.1 Visualization
In order to visualise the multi-agent model behaviour, we choose to use
dimensionality reduction techniques. Two of them were tested, PCA (Principal
Component Analysis, [9]) and T-SNE (t-distributed stochastic neighbor
embedding, [8]). We choose PCA because it is less computationally expensive
and is not stochastic. The PCA results on initialisation of the model are
shown on Figure 1.
Figure 1: Principal Component Analysis to initialise the model
### 4.2 Parameters
Before presenting the results, we will review the set of parameters used and
the circumstances of the simulations. The simulations were ran on a sample of
1000 voters which are drawn among the 30000 voters who took part in the “Voter
Autrement” experiment. We will use the same seed for the random generator for
this sample to allow reproducibility of the experiments. We have defined above
two utility functions and three velocity formulas that can be tested. We are
not going to present nor simulate all the combinations, but we will rather
focus on the most interesting ones, and try to identify the impact of the
parameters on the simulation.
Below are the official votes given by the sample’s voters. Compared to the
actual results of the 2017 presidential election, it is clearly not
representative, but it will still be interesting to compare those results to
the outputs of our model.
* •
Jean-Luc Melenchon 403
* •
Jean Lassalle 7
* •
François Asselineau 5
* •
Nicolas Dupont-Aignan 11
* •
François Fillon 34
* •
Benoit Hamon 144
* •
Philippe Poutou 12
* •
Emmanuel Macron 200
* •
Marine Le Pen 18
* •
Nathalie Arthaud 3
* •
Jacques Cheminade 2
### 4.3 Three-pragmatist rule
We tried the three-pragmatist velocity rule with different poll results. We
can see that this rule is very sensitive to the order of candidates given by
the poll (because it only draws electors towards the first three candidates in
the results of the poll).
Indeed, with a poll giving Nathalie Arthaud above Emmanuel Macron (Figure 2,
the voters closer to Macron (who is not in the top 3) move toward their
closest candidate among the three winners, in this case Benoit Hamon. This
leads to Hamon winning the election against Jean-Luc Melenchon.
Figure 2: With a poll where Nathalie Arthaud is above Emmanuel Macron
On the contrary, if the order in the poll results is the actual order (Macron
above Arthaud), the real order is just confirmed by the next iterations
(Figure 3).
Figure 3: With a poll where Emmanuel Macron is above Nathalie Arthaud
This rule gives accurate results regarding the actual official votes of our
sample, but only for the first three candidates of the poll. It also gives
totally wrong results if the poll order is not the same as the actual order.
In fact, some people behave in such a way that they will not change their
vote, even if their candidate has no chance to win at all. Explanations can be
a desire to show contestation, or a strong attachment to this candidate and
their ideas.
### 4.4 Maximum expected utility
In our next experiment, we used our second expected utility function and one
poll, and computed velocity with the second technique, of maximising expected
utility. Results are shown on Figure 4.
Figure 4: with utility function (2) and one poll
In this experiment, Emmanuel Macron has nearly as many votes as expected in
the sample. However, the utility function still needs to be tuned, because all
the voters who initially preferred Benoit Hamon, reported their votes (after
the poll) towards Jean-Luc Melenchon, which is not the case in the real sample
results. One explanation is that our expected utility model is imperfect;
indeed, people have a non-objective way of thinking in this situation, and
this is hard to model in a homogeneous way.
### 4.5 Utility speed (c)
In the next scenario, we used the third velocity formula, that moves voters
towards a personalised opinion centre. We compared the results obtained with
our 2 utility functions, and also with a single poll or regular polls along
the campaign.
Figure 5 illustrates the results with utility directly inversely proportional
to distance (first formula), and one single poll. In this figure, the number
of votes for Jean-Luc Melenchon is accurate, but Emmanuel Macron’s votes are
too low.
Figure 5: With utility function (1) and one poll
With that same utility function, regular polls do not greatly change the
results (Figure 6). We can observe nearly the same behaviour as in the
previous figure, even though we updated the information available to the
agents every 20 iterations.
Figure 6: with utility function (1) and one poll every 20 iteration
Finally we also tested the second utility function introducing repulsion
towards candidates that are too far, still with regular update polls. In
Figure 7 the results are much more tied, but they are also less accurate with
respect to the actual official votes of the sample. It is also hard to
conclude on a final order in this case, given that the rule does not converge.
Figure 7: With utility function (2) and one poll every 20 iteration
### 4.6 Periodicity of polls
The simulations run show that periodical polls do not fundamentally change the
result of the election. Indeed, quite often the results of the polls (here
made on 10% of the total population) are consistent with the order of the real
results, so the movement tendency and the order of candidates are just
confirmed a bit more by each new poll. And whenever a poll gives a different
order, it is often counter-balanced by the next poll.
However, in a real situation, other events can have an effect on the voters
preferences, such as media scandals (for instance the Penelope gate scandal
involving François Fillon during the French 2017 presidential elections, which
probably made him lose that election). In this case, even if some people are
not directly deterred from voting for their favourite candidate by the
scandal, they may be indirectly led to do so by strategy, if the subsequent
polls show that this candidate’s chance to win has dropped too much.
### 4.7 Notes about properties
The movement rules that we presented above were mainly built to make the
voters in the simulation behave somehow close to how the actual voters behave.
Because we focused on those properties, and did not make any mathematical
analysis of the rules, they do not necessarily have good mathematical
properties, in particular convergence.
It is hard to interpret the results if the simulations are not converging
towards at least a fixed order of the candidates, if not a fixed number of
votes. Indeed, we have no way to link the number of iterations in the
simulation to a real duration (in days), so any interpretation would remain a
strong supposition. The results presented here have been obtained after a
large number of iterations, when the electors are ”caught in a pit” and mainly
move towards one candidate, so that the order of the candidates does converge,
but the convergence of the number of votes is not guaranteed.
## 5 CONCLUSION AND OPENING
The models presented above have shown different interesting behaviours. The
one which gave the most accurate results is the simplest one, using the three-
pragmatist rule, even though it does not allow more than 3 candidates to
receive votes. To extend this approach it would be interesting to experiment a
k-pragmatist rule for different values of k, and play around it (maybe compute
a velocity vector using different k-pragmatist results).
The maximum expected utility rule should have really good properties, but the
utility functions were chosen quite arbitrarily. We struggled to accurately
compute an expected utility, because people do not really conduce complex
probabilistic calculus, and it is hard to model what really happens in
someones’ mind when making a voting decision. Many factors might play a role
such as emotions or cognitive biases. Also, the second utility function uses a
customisable $\alpha$ parameter to represent the tolerance before a voter is
repulsed by a candidate. Here we used the same $\alpha$ for each voter, but we
could also compute an individual value for each voter, as a function of his
starting position.
Finally, in this paper we only gave information to the voters through polls,
unique or recurrent, but more complex events could be modelled, as for example
media coverage of scandals, the use of social networks to communicate, or the
influence of other voters on each voter’s opinions.
## References
* [1] Stéphane Airiau, Umberto Grandi, and Filipo Studzinski Perotto. Learning agents for iterative voting. In Rothe J., editor, International Conference on Algorithmic Decision Theory, volume 10576 of Lecture Notes in Computer Science, pages 139–152, Cham, 2017. Springer.
* [2] Frédéric Amblard, Thomas Louail, Romain Reulier, Paul Salze, and Patrick Taillandier. Modèle iznogoud : fiche pédagogique. https://maps.hypotheses.org/production-pedagogique-de-maps/modeles-mapsiens/modele-iznogoud, 2014\.
* [3] Felix Brandt, Vincent Conitzer, Ulle Endriss, Jérôme Lang, and Ariel D Procaccia. Handbook of computational social choice. Cambridge University Press, New York, NY, USA, 1st edition, 2016.
* [4] Joshua M Epstein and Robert Axtell. Growing artificial societies: social science from the bottom up. Brookings Institution Press, 1996.
* [5] Charles M. Macal. Everything you need to know about agent-based modelling and simulation. Journal of Simulation, 10(2):144–156, 2016.
* [6] Charles M. Macal and Michael J. North. Tutorial on agent-based modelling and simulation. Journal of Simulation, 4(3):151–162, 2010.
* [7] CoCoRICo-CoDec project. Experimentation voter autrement. https://vote.imag.fr/, 2017.
* [8] Laurens Van der Maaten and Geoffrey Hinton. Visualizing data using t-sne. Journal of machine learning research, 9(11), 2008.
* [9] Svante Wold, Kim Esbensen, and Paul Geladi. Principal component analysis. Chemometrics and intelligent laboratory systems, 2(1-3):37–52, 1987\.
* [10] Ka-Ping Yee. Voting simulation visualizations. http://zesty.ca/voting/sim/, 2005.
|
# Autoencoder-based Condition Monitoring and Anomaly Detection Method for
Rotating Machines
Sabtain Ahmad, Kevin Styp-Rekowski, Sasho Nedelkoski, Odej Kao Distributed
and Operating Systems
TU Berlin
Berlin, Germany
<EMAIL_ADDRESS><EMAIL_ADDRESS>
###### Abstract
Rotating machines like engines, pumps, or turbines are ubiquitous in modern
day societies. Their mechanical parts such as electrical engines, rotors, or
bearings are the major components and any failure in them may result in their
total shutdown. Anomaly detection in such critical systems is very important
to monitor the system’s health. As the requirement to obtain a dataset from
rotating machines where all possible faults are explicitly labeled is
difficult to satisfy, we propose a method that focuses on the normal behavior
of the machine instead. We propose an autoencoder model-based method for
condition monitoring of rotating machines by using an anomaly detection
approach. The method learns the characteristics of a rotating machine using
the normal vibration signals to model the healthy state of the machine. A
threshold-based approach is then applied to the reconstruction error of unseen
data, thus enabling the detection of unseen anomalies. The proposed method can
directly extract the salient features from raw vibration signals and eliminate
the need for manually engineered features. We demonstrate the effectiveness of
the proposed method by employing two rotating machine datasets and the quality
of the automatically learned features is compared with a set of handcrafted
features by training an Isolation Forest model on either of these two sets.
Experimental results on two real-world datasets indicate that our proposed
solution gives promising results, achieving an average F1-score of 99.6%.
###### Index Terms:
LSTM autoencoder, feature extraction, anomaly detection, condition monitoring,
rotating machines
## I Introduction
In the modern day industry, machine systems are becoming more complex and
fulfill critical tasks. To enhance their reliability, the condition of the
system should be monitored. Any rotating machine, e.g. a pump, compressor, or
steam turbine, will eventually reach a point of poor health. One effective
strategy for enhancing their reliability and cost-effective maintenance is to
utilize Condition Monitoring (CM) and Prognostics and Health Management (PHM).
The aim is to identify unexpected anomalies, faults, and failures [1].
Prognostics systems built on a data-driven approach acquire data in-situ using
a network of sensors that monitor the system [3]. The dataset evolving from
the measurements of the sensors usually has a high dimensionality and may also
contain unwanted interference and noise. This dimensionality of the data has a
direct impact on the training time as well as the accuracy of neural network-
based models. A workaround for dealing with such a problem is to reduce the
dimensionality of the input signal by extracting features that carry the
health information of the system. Several methods have been proposed for
achieving this, for instance, extracting temporal features by computing the
root mean square (RMS), Skewness, Kurtosis, or Peak to Peak distance [2].
The main attributes of an ideal anomaly detection and condition monitoring
system include the ability to collect useful features and the utilization of
these features to identify the deteriorating condition of the machine by
observing the deviation from the normal (healthy) behavior. Manual feature
engineering methods require domain knowledge combined with “trial and error”
strategies. However, the advent of deep learning and recent progress in
autoencoder based models has provided an alternative way for feature
extraction and dimensionality reduction. By stacking up layers to form deep
autoencoders and by reducing the number of units in the hidden layers, it is
expected that hidden units will extract features that will represent the data.
The best features for the task at hand are learned directly from the data,
thus avoiding ad-hoc trial and error strategies. Since the autoencoder creates
a reduced representation of the data, it seems intuitive to use this
representation for anomaly detection. The assumption is that the autoencoder
only learns to map normal data points or inliers and does not include
anomalies in the trained representation. Hence, trying to reconstruct
anomalous data points will fail and incur huge losses. Finally, this
reconstruction error can be used as a basis for calculating anomaly scores and
labeling of unseen data points.
For rotating machine (RM) anomaly detection in general, it is much easier to
collect large amounts of data than to accurately obtain their corresponding
labels. Particularly in cases where faults or degradation evolve naturally
over time. Correctly assigning labels is susceptible to the data ambiguity
issue, especially at a pivotal stage when the machines exhibit early signs of
failure but are far from obvious when equated with the fully developed faults.
Moreover, anomalies are rare events, having prior knowledge about all possible
anomalies is almost impossible, so the attention is shifted from anomalous to
normal states for which data is available in large quantities. The motivation
is that by modeling the normal behavior of the machine the system will also be
able to detect anomalies that have not been observed previously.
We propose a framework for RM condition monitoring and anomaly detection based
on Long short-term memory (LSTM) autoencoder networks. We demonstrate its
applicability in both anomaly detection and condition monitoring by evaluating
it on two real-world datasets. The advantages of the autoencoder based
approach include the ability to work without any preprocessing, without any
predetermined transformations such as FFT, without any manual feature
engineering, without any feature selection, and the fact that it does not
limit itself to the preidentified anomalies; it has the potential to detect
new anomalies which have never been seen before. We also show that the time-
domain features learned in the process can be used to enhance the performance
of a simpler detection model such as Isolation Forest and thus can obviate the
need for manual feature extraction. The flowchart of the proposed method is
displayed in Fig. 1. The measured vibration signals will be preprocessed and
afterward, the training of the model is completed that is able to extract
features and reconstruct the signal for the final anomaly detection.
Figure 1: Framework flowchart of the proposed method.
The remainder of this paper is organized as follows. We will discuss related
works in Section II. Section III discusses our proposed LSTM-based autoencoder
method for robust feature extraction and anomaly detection. Sections IV and V
discuss the experimental results, while the conclusions and future works are
presented in Section VI.
## II Related Work
A plethora of studies has been conducted in the field of RM fault/anomaly
detection and time-series anomaly detection in general [4]. Features and their
generation are important concepts in data analysis and anomaly detection. The
use and selection of features are crucial for measuring differences in data
thus detecting anomalies. Traditionally, features extracted from time or
frequency domain have been used for monitoring the condition or computing the
remaining useful life (RUL) of the machine [5, 6]. Many extracted features are
influenced by operating conditions and are insensitive to anomalies.
Researchers have tried to increase the performance of anomaly detection
methods for RMs by performing multivariate analysis while using multiple
features together, [7] computed 21 bearing features using signal processing
techniques. Others have tried to achieve this by constructing features for
specific failure modes and performing multivariate analysis based on them as
did the authors in [8] by constructing 5 features each of which was sensitive
to a different failure mode. However, such manually extracted features are not
generalizable and fail to provide useful machine health information in
especially unknown cases.
More recently, [9] tried to classify various bearing classes by using a set of
time-frequency domain features and artificial neural networks. [10] replaced
the manual feature extraction step by applying 1-D convolutional neural
networks (CNN) to raw motor signals. The evaluation performed on bearing fault
detection demonstrated the superiority of their approach compared to
conventional feature extraction methods. [11] trained an autoencoder for
feature extraction and used these features to train a supervised fault
detection classification model. [12] studied several one-class classifiers
such as nearest neighbors and k-means for detecting faulty rotor bars in an
induction motor. They concluded that the k-nearest neighbor method stood out
among all the tested methods.
Deep neural network-based architectures in particular autoencoders are
successfully employed for supervised classification of faults into different
fault categories by using time or frequency domain features extracted using
prior knowledge [13]. A probabilistic framework for anomaly detection in
natural gas consumption time series is introduced in [14]. However, the
prediction method predicts the consumption levels using other independent
variables and does not incorporate the temporal information that the data had
to offer. [15] employed an vanilla autoencoder to detect anomalies in the
electric power system by embedding the temporal information using sliding
windows. Reconstruction errors obtained on sliding windows were used to
compute anomaly scores. The Inclusion of temporal information using a sliding
window works well in some cases but is not scalable usually.
The current RM anomaly detection methods, as discussed above, face one or
several of the following limitations: labor-intensive manual feature
extraction, the requirement of accurately labeled datasets, or failure to
incorporate temporal information. Time series data such as sensor data is best
modeled as a sequence where the data point at each timestep is dependent on
the previous data points. LSTM-based autoencoders are capable of dealing with
the time-series sequences and can take variable length input. To overcome
issues described above, in this study, we train an LSTM-based autoencoder over
the vibration signals to autonomously monitor the condition of RMs and extract
time-domain features to provide an alternative for manual feature extraction.
To the best of our knowledge, this is the first time that an unsupervised
method based on LSTM-autoencoders is used for identifying fault/anomalies in
accelerometer vibration signals.
## III Rotating Machine Anomaly Detection
We address the problem of anomaly detection in abnormal vibratory phenomena
captured through the accelerometer sensors mounted on RMs which should
indicate a deterioration of the system. As opposed to traditional health
prognostic systems that usually encapsulate feature extraction and anomaly
detection as distinct blocks, the proposed system takes directly raw time-
series vibration signals as input and it can efficiently learn optimal
features and based on these features determine the system’s health.
In the case of multi-dimensional input, the set of M sensors
$\\{m_{1},...,m_{M}\\}$ (also called generators) are used to capture the
behavior of a RM. This measured data is fed as input to an LSTM-based
autoencoder (LSTM-AE) that is trained over the vibration signals via batch
gradient descent to minimize a reconstruction error term between an original
signal and its reconstruction. Specifically, the encoder maps an input vector
x to a lower-dimensional hidden representation h by an affine mapping
following a nonlinearity and the decoder correspondingly generates an
estimation x’ of the input vector x.
AEs belong to the unsupervised representation learning class which try to
model the data distribution through the discovery of a set of latent
representations, also called embeddings, whose variations capture most of the
structure of the original data distribution [16]. These hidden-layer units or
low-dimensional embeddings force the AE model to learn the key representations
from the original vibration signal. The encoder generates a rich non-linear
set of features from the sensory data and the decoder learns to reconstruct
the original signal using these features. The motivation to use autoencoders
is their ability to detect anomalies based on the fact that anomalies are
quite rare and deviate greatly from the general pattern in normal healthy
data. The model is trained with the aim to learn the normal behavior of a RM,
thus not recognizing anomalies.
Figure 2: Condition monitoring of RMs through diagnosis of anomalies using
LSTM-AE.
The encoder and decoder are two main components of the network and both are
based on Long short-term memory network (LSTM) units. LSTMs are a type of
recurrent neural network (RNN) that can integrate the temporal information
into the network and maintain a hidden state vector which acts as a memory for
the past information [17].
##### Encoder
We observe the input data sequence denoted by
${\mathbf{X=(x^{(1)},x^{(2)},...,x^{(N)})}}$, where
${\mathbf{x^{N}=(x_{1}^{(N)},x_{2}^{(N)},...,x_{T}^{(N)})}}$ ${\in{R^{Txd}}}$,
meaning that for each index N there is a d-dimensional time-series sequence
with T timesteps each. We use the RNN to process the variable input sequence
and to extract the sequential information from the time-series data. For each
sequence this is given by;
$C_{t}^{\prime}=tanh(W_{C}.x_{t}+R_{C}h_{t-1}+b_{C})$ (1)
where ${C_{t}}^{\prime}$, $x_{t}$ and $h_{t-1}$ are memory state, input and
output vectors from last step and $W_{C}$, $R_{C}$ and $b_{C}$ are input
weights, recurrent weights and the bias. Tanh is used as a non-linear
activation function, whose output range lies in the interval $[-1,1]$.
The input sequences are passed through the encoder part of the LSTM network
which encodes an input sequence or batch of sequences using LSTM units and
updates its hidden state. The output of the encoder is given by;
$h_{t}=\sigma^{e}_{\phi}(x_{t},h_{t-1})$ (2)
where $h_{t}$ is the output of the $i^{th}$ LSTM-encoder, $\phi$ represents
the parameter set of the encoder and to avoid the vanishing and exploding
gradients issue, $\sigma$ for both encoder and decoder is chosen as ReLU
activation function.
##### Decoder:
The representations obtained from mapping the vibration signal to lower-
dimensional embeddings through the encoder are used by the decoder to
reproduce the original signal. The Output from the last encoder of the network
becomes the input to the LSTM-decoder network:
$h_{t}{{}^{\prime}}=\sigma^{d}_{\varphi}(h_{t},h^{\prime}_{t-1});\;\;\;x^{\prime}_{t}=\sigma(h^{\prime}_{t})$
(3)
where the set of parameters of the decoder is represented by $\varphi$ and
$x^{\prime}$ is the reconstructed input which is used to compute the
reconstruction error (RE) = ${||x_{t}-x^{\prime}_{t}||^{2}}$, also known as
the Mean Squared Error. The error is required to update the network’s encoder
and decoder parameters and later, compute the anomaly scores.
The AE model should be sensitive enough to reproduce the original signal but
insensitive enough to the training data and noise, such that the model learns
a generalizable representation of the data. During training, the focus of the
model lies in learning the normal behavior of the machine, hence anomalies are
not included in the training data. Consequently, during prediction, a slight
deviation from the normal behavior would increase the reconstruction error
(RE). So, monitoring the increase in reconstruction error gives the
possibility to detect anomalies but also anticipate the fault in advance by
identifying the degradation point, the timestamp after which the RE starts to
increase.
### III-A Model Training
The raw vibration data is normalized first to have zero mean and unit
variance. We divide the RM’s vibration data into four sets: a training set
($T_{N}$), two validation sets ($V_{N}$ & $V_{A}$), and a test set ($T_{A}$).
The distribution of data among these four sets is as follows: $T_{N}$ 70%,
$V_{N}$ & $V_{A}$ 5% (each) and $T_{A}$ 20%. $T_{N}$ consists only of normal
sequences and is used for training the LSTM-AE model. NASA bearing dataset
(dataset-1, details in section IV ) does not provide explicit labels for each
sequence, so we assume initial 70% of the data to be normal as machines have a
low probability of being in a faulty state, meaning that we assume anomalies
to occur rarely compared to the normal data. According to the ground truth in
the provided datasets, failures occurred only at the end of each run-to-
failure experiment and thus the first 70% is used for the training of the
model. The architecture of the LSTM-AE is designed as such to allow the model
to have enough capability for feature learning and secondly, the number of
units of the next hidden layer is set smaller than that of the previous layer
so that feature learning can be viewed as a signal compression process, as
shown in Fig. 2 (a). Here, the training process with the train set $T_{n}$ is
depicted, together with the compression process of the raw signal within the
autoencoder. The weights of the LSTM-AE are updated via stochastic gradient
descent using mini-batches and the Adam [24] optimizer is utilized to speed up
the training process. Batch normalization allows faster and stable training of
deep neural networks. To restrict the model from overfitting, dropout is used
and to avoid exploding gradients, gradient clipping is applied. Model
hyperparameters are learned using the Bayesian optimization method, including
the size of mini-batches, learning rate, and weight decay.
### III-B Anomaly Monitoring Process
The detection process of anomalies consists of converting reconstruction
errors into anomaly scores(AS) for each input sequence and using these scores
to obtain a threshold, characterizing the normal behavior of the machine.
After the model has been trained, sequences in $V_{N}$ are passed through the
model to get the reconstruction errors which in turn are used to estimate the
parameters of a Normal distribution ($\mu$ $\Sigma$) using maximum likelihood
estimation (MLE), similar to [18]. The probability $p_{i}$ of obtaining the
reconstruction error $e_{i}$ is given by the value of the Normal distribution
at $e_{i}$. Using the $\mu$ & $\Sigma$, the anomaly score for a datapoint
${\mathbf{x_{t}^{(N)}}}$ is computed as follows:
$a_{i}=(e_{i}-\mu)^{T}\Sigma^{-1}(e_{i}-\mu)$ (4)
where $a_{i}$ is the desired anomaly score, $e_{i}$ is RE obtained for a
sequence ${x_{t}^{(i)}}$ and $\mu$ $\Sigma$ are the mean and variance of a
multivariate Gaussian distribution.
During the initialization phase, an anomaly score threshold $\tau$ is also
learned using a validation set $V_{A}$ that may contain examples from
anomalous sequences alongside with the normal sequences. Unseen sequences
within $T_{A}$ are classified as follows: if a sequence has an anomaly score
$>$ $\tau$ it will be labeled as an anomaly, otherwise as normal. This is also
depicted in Fig. 2 (b) where the raw measured signal is input to the trained
autoencoder model that generates reconstruction errors and anomaly scores.
These are then used within the described decision process to determine whether
the signal was normal or anomalous.
The assumption here is, that as the monitored equipment degrades or faces a
failure, this disrupts the normal working of the machine and affects the
interaction between different variables which can be measured by
accelerometers, especially in form of vibrations. As the sensor values start
deviating from the normal working condition of the machine, it is expected to
see an increased error in the reconstruction of the input. By monitoring the
reconstruction error and the anomaly score, an indication of the health of the
monitored machine can be derived.
### III-C Feature Extraction
Feature learning is a critical step in improving the performance of anomaly
detection models due to the multidimensionality of data that is input into the
model. In general, machine vibration signals comprise a stationary vibration
part, a random vibration part, and noise [19]. We study two methods for the
inspection of vibration signals: automated feature extraction based on LSTM-AE
and manual feature extraction based on classic signal processing methods.
Isolation Forest is used to evaluate the effectiveness of these two feature
extraction methods.
Isolation Forest is a random forest-based anomaly detection algorithm that
utilizes isolation to determine anomalies in data [20]. The Isolation Forest
(IF) algorithm is based on Decision Forests, an ensemble method that uses the
averages of outputs from many different trees.
#### III-C1 LSTM-AE based Feature Extraction
One of the main characteristics of an autoencoder that is more powerful for
finding intrinsic data structures by reducing data dimensionality through non-
linear transformations than Principal Component Analysis (PCA) [21].
The encoder from the trained LSTM-AE is used to perform feature extraction.
The process is done by reducing the number of units in the hidden layer, it is
expected that the hidden units in the encoder network will extract features
that will represent the data. To learn more abstract features, multiple AEs
are stacked together to form a stacked AE, in which the output of each hidden
layer is connected to the input of successive hidden layers. A stacked AE
applies dimensionality reduction in a hierarchical manner, obtaining more
abstract features in higher hidden layers which lead to a better
reconstruction of the data [22].
#### III-C2 Manual Feature Extraction
Following time-domain statistical features are generally used to detect
incipient machine faults/anomalies: mean ($\mu_{x}$), root mean square (RMS),
percentiles (25th, 50th & 75th ), max absolute value, standard deviation
($\sigma_{x}$),peek-to-peek, skewness, kurtosis, entropy, and AR-coefficients.
The last five features are described through the following equations:
$Peak{-}peak(p{-}p)=abs(Max(x))+abs(Min(x))$ (5)
$Skewness=\frac{\sum_{i=1}^{N}(x_{i}-\mu_{x})^{3}}{N\sigma_{x}^{3}}$ (6)
$Kurtosis=\frac{\sum_{i=1}^{N}(x_{i}-\mu_{x})^{4}}{N\sigma_{x}^{4}}$ (7)
$Entropy=\sum_{i}p_{i}\log p_{i}$ (8)
$AR{-}coefficients=\sum_{k=1}^{p}a_{k}x[n-k]+e[n]$ (9)
where in equation 9; p is the degree of the AR model, x[n] is a signal
composed of b data points, $a_{k}$ is real-values AR coefficient and e[n] is
white noise.
## IV Evaluation
There are two datasets considered for the evaluation of the proposed method.
We evaluate and test the applicability of the proposed method in Condition
Monitoring using the IMS bearing dataset [23](Dataset-1). In addition, the
effectiveness to find anomalies in unsupervised settings is evaluated using a
private industry dataset from vibration data of RMs. Evaluations performed on
these two datasets demonstrate the ability of the method to detect anomalies
within the industry dataset and perform condition monitoring on the bearing
dataset, as the IMS bearing dataset doesn’t provide explicit labels but
instead contains a degrading health state scenario of the bearing under
experiment. Below, we introduce the datasets and illustrate the performance of
the proposed method on these datasets.
### IV-A Dataset-1
This dataset was gathered from a run-to-failure experimental setting,
involving four bearings and is subdivided into three datasets, each of which
consists of the vibration signals from these four bearings [23]. These sets
were collected from three test-to-failure experiments which were performed
independently, and failures occurred at the end of the test. Table I
represents the properties of the collected data from these experiments.
TABLE I: Dataset-1: NASA Bearing dataset description Set # | Batches | Batch Size | Anomaly
---|---|---|---
Set1 | 2156 | 4 x 20480 | B3 and B4
Set2 | 984 | 4 x 20480 | B1
Set3 | 6324 | 4 x 20480 | B3
TABLE II: Dataset-2: Industry dataset description Machine | Batches | Batch Size | Anomalies
---|---|---|---
RM-1 | 1176 | 6144 | 53
RM-2 | 1463 | 2401 | 4
RM-3 | 2204 | 2401 | 4
RM-4 | 1452 | 2401 | 3
RM-5 | 1452 | 2401 | 4
### IV-B Dataset-2
This dataset consists of data from five different RMs (RM-1 to RM-5) which are
of the same kind but build-wise unique. Data from each RM contains
3-dimensional vibration signals measured over time, captured using the
accelerometer sensors attached to the housing of the RM. The signal recordings
were taken in batches and the batch sizes vary, depending upon the RM.
Considering the condition of the machine at the time of recording, labels are
assigned to each batch, 0 representing normal and 1 anomalous, respectively.
Table II presents the summarized statistics of the industry dataset. The
difference between normal and anomalous signals is shown in Fig. 3, vibrations
in all three directions (x, y, z) differ significantly for anomalous signals
compared to normal vibration signals. These differences can have a variety of
characteristics as the amplitude of different frequencies differs between the
two signals.
Figure 3: Example of a normal (Left) and anomalous (Right) signal from RM-1,
rows corresponding to the x, y and z-axis of the accelerometer.
## V Experiments and Results
We demonstrate our method’s applicability in condition monitoring & anomaly
detection with three general settings where in the first one, the autoencoder
is trained with 70% of the data, 10% (split between two validation sets) is
used to calculate the threshold, and remaining 20% of it is kept for
evaluation to determine the prediction ability of the proposed approach for an
anomaly detection task. In the second experimental setting, each run-to-
failure experiment is simulated by training the LSTM-AE model with the
available data from the non-failing RMs to monitor the condition of the
machine in which failure occurred during the run-to-failure experiment. The
trained model is used to monitor the status of the faulty machines to
demonstrate that once the model is trained, it is capable of being applied to
different RMs of a similar kind, enabling the transfer of rare anomaly
knowledge from RMs to other RMs of a similar kind.
The third setting consists of a method for performing automated feature
extraction from vibration signals. Features extracted using this setting are
compared against a set of handcrafted features by training Isolation Forest on
these two features sets separately.
### V-A Setting 1: Online Prognostic
In this setting, as discussed, the autoencoder learns to reconstruct the
normal behavior of an RM using only 70% of the available healthy/normal data
in order to evaluate the prediction performance of the model in an online
monitoring phase. The threshold $\tau$ is obtained using a validation set of
size 5% to classify the samples as normal/healthy or anomalous. Anomaly scores
for each data sample are calculated using the reconstruction errors and data
points with anomaly scores larger than $\tau$ are labeled as anomalies. The
point at which the anomaly score crosses the threshold and starts to increase
gradually is considered as the degradation point.
##### Result: Dataset-1
Here we study 4 cases of failing bearings (B1 to B4) of the different sub-
datasets (S1 to S3), S1-B3, S1-B4, S2-B1, and S3-B3, for early fault
detection. The model trained on individual bearings is able to predict the
degradation point and capture the propagation of fault in the simulated run-
to-failure experiment as is depicted in Fig. 4. As the fault in every failing
bearing occurs at the end, Fig. 4 displays the anomaly scores of the last 620,
450, and 780 batch samples for S1-B3 & S1-B4, S2-B1, and S3-B3, respectively.
For every failing bearing in each dataset, the proposed method generates a
trend corresponding to the health status of the bearing based on the vibration
sensory data. As the bearing health starts to degrade, the anomaly score tends
to increase and once it passes the degradation point (marked with the filled
circles) it increases gradually. By following this trend and abrupt changes in
the scores one can feasibly detect the failure. The degradation point for each
bearing was calculated by identifying the abrupt change and continuous
increasing trend in anomaly scores. Sample no. 2039, 1704, 667, and 5241 were
identified as degradation points for S1-B3, S1-B4, S2-B1, and S3-B3,
respectively. The circle on the score line indicates the degradation starting
point.
Figure 4: Deteriorating condition of the failing bearings captured by the
model. High anomaly scores indicate that system is more likely to be in faulty
state. Circles mark the start of the degradation for each bearing.
Figure 5: Anomaly scores for RM-1 from the second dataset: Red straight line indicates the threshold, data points above this red line are labeled as anomalous, otherwise normal. Black circles indicate false positives. TABLE III: Setting 1 results: individual RM anomaly detection scores. Machine | Precision | Recall | TPR | FPR | F1-score
---|---|---|---|---|---
RM-1 | 0.993 | 0.993 | 0.981 | 0.010 | 0.993
RM-2 | 1.0 | 1.0 | 1.0 | 0.0 | 1.0
RM-3 | 1.0 | 1.0 | 1.0 | 0.0 | 1.0
RM-4 | 0.994 | 0.991 | 1.0 | 0.007 | 0.993
RM-5 | 0.965 | 0.972 | 0.25 | 0.007 | 0.967
##### Result: Dataset-2
The second dataset consists of 5 different RMs with anomalies, for each
machine there is a different number of anomalies (see Table II). To detect
anomalies from each individual machine, we conduct five different experiments
(one for each machine) using only the non-anomalous (normal) data to train the
autoencoder network. The anomaly detection results are evaluated using,
Precision, Recall, True positive rate (TPR), False positive rate (FPR), and
F1-score. These evaluation metrics are weighted per class.
As described, model performance is evaluated over the test set consisting of
20% of the data. For instance, the test set for RM-1 consisted of 145 (92
normal, 53 anomalous) sequences in total. Each sample from the test set was
reconstructed using the trained autoencoder network and was labeled as
anomalous if its anomaly score was larger than the threshold, normal
otherwise. The trained model was able to detect 52/53 anomalies without
raising many false alarms (just one false positive). Fig. 5 displays the
anomaly scores for each data point; sequences below the threshold (marked by
the red line) are classified as normal and data points with anomaly scores
larger than the threshold are labeled as anomalies. Evaluation scores
(precision, recall, TPR, FPR, F1-score) for individual RMs from the second
dataset, averaged over 10 repetitions of the experiments are presented in
Table III in which the model performs very well with F1-scores close to 1 and
is able to detect anomalies without generating many false positives which is
indicated by the high TPR and low FPR. Based on the results, the LSTM-AE can
effectively extract discriminative features directly from the raw vibration
data and achieve a competitive anomaly detection rate. For the test data, an
overall detection F1-score of 99% and TPR over 93% for each machine is
obtained.
### V-B Setting 2: Condition Monitoring
In dataset-1, during each run-to-failure experiment, only one out of 4
bearings faced failure while the other 3 remained in healthy condition except
for experiment-1 in which a fault occurred in two bearings at the end. The
question of interest is whether the knowledge extracted from the healthy
bearings can be applied to detect the deteriorating condition of a faulty
bearing and raise alarm well before the total failure actually happens. To
validate this, in this setting only the data from the healthy bearings is used
to train the model and this trained model is then used to evaluate the
condition of the previously unseen faulty bearings.
Three simulations, one per sub-dataset are performed as follows: The Set1
model is trained using only training data from the non-failing bearings S1-B1
& S1-B2 with the aim to learn the functioning of healthy bearing and apply
this learned knowledge to monitor and detect the deteriorating condition of
the failing bearings S1-B3 & S1-B4. Similarly, for Set2, the training data
comprised of S2-B2, S2-B3, and S2-B4 and while the test set consisted of S2-B1
and for Set3, training data consists of S3-B1, S3-B2, S3-B4 and test set of
S3-B3. Fig. 6 displays the output anomaly scores of the four failing bearings
from these experiments. The trend it generates for each bearing is in
accordance with the ground truth, from the beginning until near the end the
bearings remain in healthy condition, thus low anomaly scores are correctly
calculated. The fault starts to appear only at the very end, which is captured
by the model by a continuous increase in the anomaly scores.
A low anomaly score corresponds to healthy behavior while an upward trend
(high anomaly score) highlights the abnormal behavior of the machine. Arrival
and propagation of fault of for every failing bearing is shown in Fig. 6. It
is clearly visible that the proposed approach is able to identify the
initiation of the faulty trend as well as the increasing effect of the
deterioration for all four bearings. Fig. 7 visualizes the anomaly scores and
visually indicates the health status of the four faulty bearings S1-B3, S1-B4,
and S2-B1 and S3-B3 in parts B3, B4, B1, and B3, respectively. The color bar
represents the anomaly score, from 0 (blue) to 1 (red).
TABLE IV: Model trained on combined healthy data from four RMs is able to detect anomalies without raising any false alarms across all four RMs. Machine | Precision | Recall | TPR | FPR | F1-score
---|---|---|---|---|---
RM-2 | 1.0 | 1.0 | 1.0 | 0.0 | 1.0
RM-3 | 0.997 | 0.997 | 0.75 | 0.0 | 0.997
RM-4 | 1.0 | 1.0 | 1.0 | 0.0 | 1.0
RM-5 | 1.0 | 1.0 | 1.0 | 0.0 | 1.0
Figure 6: Anomaly scores for the condition monitoring of the faulty bearings
by modeling the normal working behavior of the remaining bearings using the
other non-failing bearings from the run-to-failure experiments.
Figure 7: Visualization of anomaly scores from the run-to-failure experiments,
where blue color indicates that the system is less likely to be in a faulty
state while an abnormal behavior is indicated by the red color bars.
The same setting was applied to our second dataset where the results were
obtained by combining the normal data points from four RMs (RM-2 to RM-5). We
divided each RM’s data into three sets: training ${\mathit{Train_{Pi}}}$
(70%), validation ${\mathit{V_{Pi}}}$ (10%) and test ${\mathit{Test_{Pi}}}$
(20%). Train set T was created by combining the four ${{Train_{Pi}}}$ and was
used to train the model. The validation sets ${\mathit{V_{Pi}}}$ were used to
calculate the threshold ${\tau}$ using the anomaly scores for each RM
${\mathit{P{i}}}$. The final anomaly detection model trained on T was then
evaluated on each ${\mathit{Test_{Pi}}}$ set and the obtained results are
presented in Table IV, in which the model performs very well in detecting
anomalies across four different machines without raising any false alarms. It
is worthwhile to notice that combining the training data from multiple RMs
increases the performance of the model compared to the performance of the
model on individual RMs. The trained LSTM-AE is able to model the healthy
working behavior of the RMs and is sensitive to anomalies, as it can be seen
in Fig. 8 which shows that the identified anomalies match well with the ground
truth. The model can reconstruct the normal signals very well, indicated by a
low anomaly score, while it fails to do so whenever it encounters anomalies.
Figure 8: Anomaly detection results obtained from mixing the normal signals
from four machines to train the LSTM-AE model and then using the validation
data from each machine to calculate the threshold to detect the anomalous
samples. All identified anomalies except for one false negative in case of
RM-3 were identified correctly by the model.
### V-C Setting 3: Feature Extraction
We employ the same LSTM-based autoencoder model to automatically extract time-
domain features from vibration signals. The automatically extracted features
are compared with a set of manually crafted features, by training an Isolation
Forest on both of these two sets and monitoring the performance of the model.
Results obtained from the IMS bearing dataset using the proposed method are
compared with the IF-based anomaly detection trained only on manually
extracted features. The comparison of results between two Isolation Forest
based models is shown in Table V, which shows the total number of anomalies
detected by the models in each bearing. It’s evident that the S1-B1, S1-B4,
S2-B1, and S3-B3 are having comparatively more anomalous samples than the rest
of the bearings as these four bearings result in failure according to the
ground truth provided. The comparison illustrates that the latent features
extracted by the LSTM-AE perform better than the manually engineered features
demonstrating the effectiveness of the proposed approach in providing a robust
set of features that boost the performance of the anomaly detection model.
This can be confirmed by looking at the number of anomalies detected by each
model; for each failing bearing, the IF model based on automatically extracted
features detects a greater number of anomalies than the model using the
handcrafted set of features. One interpretation of this could be that LSTM-AE
provides a set of features that are more sensitive and thus can detect the
subtle changes in the vibrations when the fault appears initially and starts
developing gradually until the failure finally happens and bearing stops
functioning. While on the other hand manually handcrafted features only start
labeling signals as anomalous when the fault has fully developed already, and
vibrations deviate significantly from the normal signals.
TABLE V: Number of anomalies detected by Isolation Forest using automatically extracted features (A_FE) and manually extracted features (M_FE): Failing bearings are marked in red and results of the best performing model on these bearings is highlighted with bold entries. Bearing | No. of Anomalies (A_FE) | No. of Anomalies (M_FE)
---|---|---
S1-B1 | 04 | 06
S1-B2 | 07 | 05
S1-B3 | 39 | 26
S1-B4 | 441 | 213
S2-B1 | 269 | 133
S2-B2 | 09 | 14
S2-B3 | 07 | 16
S2-B4 | 13 | 11
S3-B1 | 24 | 57
S3-B2 | 41 | 35
S3-B3 | 1016 | 693
S3-B4 | 39 | 51
Multi-step anomaly detection results, where the first step consists of
automated features extraction using LSTM-AE and in the second step these
features are utilized to train an anomaly detection model (Isolation Forest),
for the second dataset are summarized in Table VI. It is important to mention
that each RM except for RM-1 in dataset-2 contains very few anomalies (see
Table II). IF based on features provided by LSTM-AE (AFE$\\_$IF) outperforms
IF trained using the manually engineered features (MFE$\\_$IF) for all RMs in
terms of precision. Although, for RM-3 and RM-5 MFE$\\_$IF performed better in
terms of TPR. However, it is worthwhile to notice that the model based on
automatically engineered features results in less false positives in
comparison to the model trained on handcrafted features which had
significantly more false positives. As previously stated, the studies that
have used handcrafted features obviously are not able to carry the complete
system health representation and may not represent the characteristics of the
underlying vibration signal under all circumstances. As seen from the results,
fault/anomaly detection performance of the conventional methods such as
Isolation Forest depends highly on the carefully crafted features.
Consequently, this limits the general applicability of these methods.
TABLE VI: Comparison of Isolation Forest trained on automatically extracted features (AFE_If) and the same model trained using the handcrafted features (MFE_IF) using the evaluation metrics, row-wise for each RM in Dataset-2. | AFE_IF | MFE_IF
---|---|---
| P | TPR | FPR | P | TPR | FPR
RM-1 | 0.972 | 0.943 | 0.01 | 0.91 | 1.0 | 0.19
RM-2 | 1.0 | 1.0 | 0.0 | 0.976 | 1.0 | 0.23
RM-3 | 0.991 | 0.75 | 0.004 | 0.984 | 1.0 | 0.15
RM-4 | 1.0 | 1.0 | 0.0 | 0.983 | 1.0 | 0.09
RM-5 | 0.979 | 0.25 | 0.0 | 0.977 | 1.0 | 0.14
The experimental results obtained in all three settings are able to support
the earlier claims that LSTM-AE trained only on the normal/healthy signals is
not just able to monitor the health condition of an individual RM but
knowledge learned can also be effectively used to detect anomalies in similar
yet different RMs. Additionally, it is also shown that latent representations
obtained from LSTM-AE can be used as an alternative to manual feature
extraction, which requires prior domain and signal processing knowledge.
## VI Conclusion and Future Work
In conclusion, we have demonstrated an unsupervised method for automated
feature extraction from raw vibration signals that can be used for detecting
faults and anomalies in rotating machines. A typical condition monitoring
system requires feature extraction and decision about the health of the
system. The feature extraction component of such systems involves the
implementation of signal processing methods for preprocessing the data before
using it for anomaly detection. The proposed method fuses the features
extraction and anomaly detection modules within one condition monitoring
system. Encoding the time-dependent as well as inherent features of the
measured vibration time series into a hidden state of an LSTM-based
autoencoder enabled the usage of the resulting reconstruction error to be used
in an anomaly detection setup. The experimental results on two real-world
datasets illustrate the effectiveness of the proposed model and show that
LSTM-based autoencoders can extract salient features from vibration signals
and achieve high accuracy in fault diagnosis. The model was able to detect the
deteriorating condition of all four failing bearings for IMS dataset and
achieved an overall F1 score of 99.6% for the second dataset. As demonstrated
in the experiments, the latent representations obtained from the LSMT-AE carry
system health information and achieve a higher F1 score than the manually
extracted features (97% and 93% respectively), indicating a better
representation of the characteristics of the vibration signal for the proposed
method.
For future work, other alternatives for choosing a threshold value shall be
investigated. Also more advanced types of autoencoders such as variational
autoencoder can be tried out, which alternatively try to model the data
distribution instead of learning to re-create the normal data points, which
means datasets with mixed examples (anomalous $\&$ non-anomalous) could also
be used for training the model. This paper will help design further advanced
feature extraction methods from rotating machines and help build models to
monitor the condition of such machines with no or very minimal human
supervision.
## References
* [1] Marcos E Orchard and George J Vachtsevanos. A particle-filtering approach for on-line fault diagnosis and failure prognosis. Transactions of the Institute of Measurement and Control, 31(3-4):221–246, 2009.
* [2] Ruey S Tsay, Daniel Pena, and Alan E Pankratz. Outliers in multivariate time series. Biometrika, 87(4):789–804, 2000.
* [3] Patrick Nectoux, Rafael Gouriveau, Kamal Medjaher, Emmanuel Ramasso, Brigitte Chebel-Morello, Noureddine Zerhouni, and Christophe Varnier. Pronostia: An experimental platform for bearings accelerated degradation tests. 2012\.
* [4] Charu C Aggarwal and Saket Sathe. Outlier ensembles: An introduction. Springer, 2017.
* [5] Hai Qiu, Jay Lee, Jing Lin, and Gang Yu. Wavelet filter-based weak signature detection method and its application on rolling element bearing prognostics. Journal of sound and vibration, 289(4-5):1066–1090, 2006.
* [6] Diego Alejandro Tobon-Mejia, Kamal Medjaher, Noureddine Zerhouni, and Gerard Tripot. A data-driven failure prognostics method based on mixture of gaussians hidden markov models. IEEE Transactions on reliability, 61(2):491–503, 2012.
* [7] Zhanguo Xia, Shixiong Xia, Ling Wan, and Shiyu Cai. Spectral regression based fault feature extraction for bearing accelerometer sensor signals. Sensors, 12(10):13694–13719, 2012.
* [8] Jing Tian, Carlos Morillo, Michael H Azarian, and Michael Pecht. Motor bearing fault detection using spectral kurtosis-based feature extraction coupled with k-nearest neighbor distance analysis. IEEE Transactions on Industrial Electronics, 63(3):1793–1803, 2015\.
* [9] Jaouher Ben Ali, Nader Fnaiech, Lotfi Saidi, Brigitte Chebel-Morello, and Farhat Fnaiech. Application of empirical mode decomposition and artificial neural network for automatic bearing fault diagnosis based on vibration signals. Applied Acoustics, 89:16–27, 2015.
* [10] Turker Ince, Serkan Kiranyaz, Levent Eren, Murat Askar, and Moncef Gabbouj. Real-time motor fault detection by 1-d convolutional neural networks. IEEE Transactions on Industrial Electronics, 63(11):7067–7075, 2016\.
* [11] Jiedi Sun, Changhong Yan, and Jiangtao Wen. Intelligent bearing fault diagnosis method combining compressed data acquisition and deep learning. IEEE Transactions on Instrumentation and Measurement, 67(1):185–195, 2017.
* [12] Roozbeh Razavi-Far, Maryam Farajzadeh-Zanjani, Shokoofeh Zare, Mehrdad Saif, and Jafar Zarei. One-class classifiers for detecting faults in induction motors. In 2017 IEEE 30th Canadian Conference on Electrical and Computer Engineering (CCECE), pages 1–5. IEEE, 2017.
* [13] Feng Jia, Yaguo Lei, Jing Lin, Xin Zhou, and Na Lu. Deep neural networks: A promising tool for fault characteristic mining and intelligent diagnosis of rotating machinery with massive data. Mechanical Systems and Signal Processing, 72:303–315, 2016.
* [14] Hermine N Akouemo and Richard J Povinelli. Probabilistic anomaly detection in natural gas time series data. International Journal of Forecasting, 32(3):948–956, 2016.
* [15] Marco Martinelli, Enrico Tronci, Giovanni Dipoppa, and Claudio Balducelli. Electric power system anomaly detection using neural networks. In International Conference on Knowledge-Based and Intelligent Information and Engineering Systems, pages 1242–1248. Springer, 2004.
* [16] Guillaume Alain and Yoshua Bengio. What regularized auto-encoders learn from the data-generating distribution. The Journal of Machine Learning Research, 15(1):3563–3593, 2014\.
* [17] Klaus Greff, Rupesh K Srivastava, Jan Koutník, Bas R Steunebrink, and Jürgen Schmidhuber. Lstm: A search space odyssey. IEEE transactions on neural networks and learning systems, 28(10):2222–2232, 2016.
* [18] Pankaj Malhotra, Anusha Ramakrishnan, Gaurangi Anand, Lovekesh Vig, Puneet Agarwal, and Gautam Shroff. Lstm-based encoder-decoder for multi-sensor anomaly detection. arXiv preprint arXiv:1607.00148, 2016.
* [19] GK Singh and Saleh Al Kazzaz Sa’ad Ahmed. Vibration signal analysis using wavelet transform for isolation and identification of electrical faults in induction machine. Electric Power Systems Research, 68(2):119–136, 2004.
* [20] Fei Tony Liu, Kai Ming Ting, and Zhi-Hua Zhou. Isolation forest. In 2008 Eighth IEEE International Conference on Data Mining, pages 413–422. IEEE, 2008.
* [21] Tyler Manning-Dahan. Pca and autoencoders, 2018.
* [22] Pascal Vincent, Hugo Larochelle, Isabelle Lajoie, Yoshua Bengio, Pierre-Antoine Manzagol, and Léon Bottou. Stacked denoising autoencoders: Learning useful representations in a deep network with a local denoising criterion. Journal of machine learning research, 11(12), 2010.
* [23] J Lee, H Qiu, G Yu, and J Lin. Rexnord technical services (2007).’bearing data set’, ims, university of cincinnati. nasa ames prognostics data repository, 2009.
* [24] John Duchi, Elad Hazan, and Yoram Singer. Adaptive subgradient methods for online learning and stochastic optimization. Journal of machine learning research, 12(7), 2011.
|
# On Erdős’s Method for Bounding the Partition Function
Asaf Cohen Antonir School of Mathematical Sciences, Tel Aviv University, Tel
Aviv, 6997801, Israel. Email<EMAIL_ADDRESS>Asaf Shapira School of
Mathematics, Tel Aviv University, Tel Aviv 69978, Israel. Email:
<EMAIL_ADDRESS>Supported in part by ISF Grant 1028/16, ERC Starting
Grant 633509 and NSF-BSF Grant 2019679.
###### Abstract
For fixed $m$ and $R\subseteq\\{0,1,\ldots,m-1\\}$, take $A$ to be the set of
positive integers congruent modulo $m$ to one of the elements of $R$, and let
$p_{A}(n)$ be the number of ways to write $n$ as a sum of elements of $A$.
Nathanson proved that $\log p_{A}(n)\leq(1+o(1))\pi\sqrt{2n|R|/3m}$ using a
variant of a remarkably simple method devised by Erdős in order to bound the
partition function. In this short note we describe a simpler and shorter proof
of Nathanson’s bound.
## 1 Introduction.
A partition of an integer $n$ is a sequence of positive integers $a_{1}\leq
a_{2}\leq\cdots$ whose sum is $n$. Let $p(n)$ denote the classical partition
function of $n$, namely, the number of ways to write $n$ as a sum of positive
integers. The celebrated Hardy–Ramanujan formula [2] (discovered independently
by Uspensky [6]) states that
$p(n)\sim\frac{1}{4n\sqrt{3}}\exp(\pi\sqrt{2n/3})$. Erdős [1] later devised a
remarkably simple proof of the slightly weaker upper bound
$\log p(n)\leq\pi\sqrt{2n/3}\;.$ (1)
Let $\mathbb{N}$ denote the set of positive integers, and suppose
$S\subseteq\mathbb{N}$. We define $p_{S}(n)$ to be the number of partitions of
$n$ with all summands in $S$. For a fixed positive integer $m$ and
$R\subseteq\\{0,1,\ldots,m-1\\}$, we take $A=A(m,R)$ to be the set of all
positive integers $a$ with $a\;(\mathrm{mod}\;m)\in R$. Nathanson [4] used
Erdős’s method for proving (1) to obtain111Nathanson [4] also proves that
$\log p_{A}(n)\geq(1-o(1))\pi\sqrt{2n|R|/3m}\;$.
$\log p_{A}(n)\leq(1+o(1))\pi\sqrt{2n|R|/3m}\;.$ (2)
The argument in [4] was more complicated than Erdős’s due to the need to
control various error parameters (but was still simpler than the original
proof of this result [3]); see the remark at the end of the proof.
Our goal in this short note is to give a proof of (2) which is as simple as
Erdős’s proof of (1). The main trick is that, instead of directly bounding
$p_{A}(n)$, we will instead bound $p_{A^{+}}(n)$, where given $m$ and $R$ as
above, we take $A^{+}=A\setminus R$, that is, the set of all integers $a\geq
m$ with $a\;(\mathrm{mod}\;m)\in R$. Our main result here is the following
generalization222Indeed, when $m=1$ and $R=\\{0\\}$, we have
$p_{A^{+}}(n)=p(n)\;$. of (1).
###### Theorem 1.
For every $A^{+}$ as above, $\log p_{A^{+}}(n)\leq\pi\sqrt{2n|R|/3m}$ .
It is easy to obtain (2) from the upper bound given by Theorem 1. Indeed, we
first note that for every $n^{\prime}$ we have
$p_{R^{+}}(n^{\prime})\leq(n^{\prime}+1)^{|R|}$, where
$R^{+}=R\setminus\\{0\\}$. This follows immediately from the fact that in
every partition of $n^{\prime}$, each of the integers of $R^{+}$ is used at
most $n^{\prime}$ times. We thus infer that
$p_{A}(n)=\sum_{0\leq n^{\prime}\leq n}p_{R^{+}}(n^{\prime})\cdot
p_{A^{+}}(n-n^{\prime})\leq(n+1)^{|R|}\sum_{0\leq n^{\prime}\leq
n}e^{c\sqrt{n-n^{\prime}}}\leq(n+1)^{|R|+1}e^{c\sqrt{n}}\;,$
where $c=\pi\sqrt{2|R|/3m}$. Taking logs from both sides, we obtain (2).
The proof of Theorem 1 appears in the next section. At the end of that section
we briefly explain why our proof is simpler than that of [4].
## 2 Proof of Theorem 1.
For a given fixed integer $m\geq 1$ and $R\subseteq\\{0,1,\ldots,m-1\\}$, let
$A^{+}$ denote the set of all integers $a\geq m$ with $a\;(\mathrm{mod}\;m)\in
R$. We start with a few observations that extend those used in [1]. We first
note that, for every $0<t<1$, we have
$\sum_{a\in A^{+}}at^{a}=\sum_{r\in
R}\frac{(r+m)t^{r+m}-rt^{2m+r}}{(1-t^{m})^{2}}\;.$ (3)
Indeed, $\sum_{a\in A^{+}}at^{a}=\sum_{r\in R}\sum_{a\in A_{r}^{+}}at^{a}$
where $A_{r}^{+}$ is the set of all integers $a\geq m$ with
$a=r\;(\mathrm{mod}\;m)$ (i.e., $A_{r}^{+}=\\{r+m,r+2m,r+3m,\ldots\\}$).
Hence, without loss of generality we may assume $|R|=1$. Letting $r\in R$, we
have
$\displaystyle\sum_{a\in A_{r}^{+}}at^{a}$ $\displaystyle=t\sum_{a\in
A_{r}^{+}}\frac{d}{dt}t^{a}=t\cdot\frac{d}{dt}\sum_{a\in
A_{r}^{+}}t^{a}=t\cdot\frac{d}{dt}\frac{t^{r+m}}{1-t^{m}}=\frac{(r+m)t^{r+m}-rt^{2m+r}}{(1-t^{m})^{2}}\;.$
This proves (3). We next claim that, if $0\leq r\leq m-1$ is an integer, then
for all $x>0$, we have
$\frac{(r+m)e^{-(r+m)x}-re^{-(2m+r)x}}{(1-e^{-mx})^{2}}\leq\frac{1}{mx^{2}}\;.$
(4)
Indeed, since $x>0$, the power series expansion of $e^{x}$ gives
$e^{x/2}-e^{-x/2}=2\sum_{k=0}^{\infty}\frac{1}{(2k+1)!}\left(\frac{x}{2}\right)^{2k+1}=x+x^{3}\sum_{k=1}^{\infty}\frac{x^{2k-2}}{(2k+1)!\cdot
2^{2k}}>x\;,$
implying that
$\frac{e^{-x}}{(1-e^{-x})^{2}}=\frac{1}{(e^{x/2}-e^{-x/2})^{2}}<1/x^{2}\;.$
We can thus infer that
$\displaystyle\frac{(r+m)e^{-(r+m)x}-re^{-(2m+r)x}}{(1-e^{-mx})^{2}}$
$\displaystyle={((r+m)e^{-rx}-re^{-(m+r)x}})\frac{e^{-mx}}{(1-e^{-mx})^{2}}$
$\displaystyle\leq({(r+m)e^{-rx}-re^{-(m+r)x})}\frac{1}{m^{2}x^{2}}\;.$
It remains to check that the expression in parentheses is bounded by $m$.
Since the derivative of ${(r+m)e^{-rx}-re^{-(m+r)x}}$ (which is
$r(r+m)(e^{-(m+r)x}-e^{-rx})$) is always nonpositive for $x\geq 0$, it is
enough to check its value at $x=0$ where it attains the value $m$. This proves
(4).
We now note that (3) and (4) imply that, for every $x>0$,
$\sum_{a\in A^{+}}ae^{-ax}\leq\frac{|R|}{mx^{2}}\;.$ (5)
The final observation we will need is the well-known fact that, for every set
of positive integers $S$, we have
$n\cdot p_{S}(n)=\sum_{s\in S\cap[n]}s\sum_{1\leq k\leq n/s}p_{S}(n-sk)\;,$
(6)
where we use $[n]$ for the integers $\\{1,\ldots,n\\}$. To see this, let
$p_{S}(n,s,t)$ and $p_{S}^{\prime}(n,s,t)$ be the number of partitions of $n$
with summands in $S$ where $s$ appears exactly $t$ times, and at least $t$
times, respectively. Then by double counting,333The two sides of the first
equality count the sum of all integers that appear in all partitions of $n$
using integers from $S$ (there are $p_{S}(n)$ such partitions). As to the
third equality, it follows by observing that each partition of $n$ with
exactly $t$ occurrences of $s$ contributes $1$ to $t$ of the summands
$p_{S}^{\prime}(n,s,t)$, namely
$p_{S}^{\prime}(n,s,1),p_{S}^{\prime}(n,s,2),\ldots,p_{S}^{\prime}(n,s,t)$.
See Theorem 15.1 in [5] for a full detailed proof. we have
$\displaystyle n\cdot p_{S}(n)$ $\displaystyle=\sum_{s\in
S,t\in\mathbb{N}}s\cdot t\cdot p_{S}(n,s,t)=\sum_{s\in
S\cap[n]}s\sum_{t\in\mathbb{N}}t\cdot p_{S}(n,s,t)$ $\displaystyle=\sum_{s\in
S\cap[n]}s\sum_{t\in\mathbb{N}}p_{S}^{\prime}(n,s,t)=\sum_{s\in
S\cap[n]}s\sum_{1\leq k\leq n/s}p_{S}(n-sk)\;.$
This proves (6).
We are now ready to complete the proof of Theorem 1. We use induction on $n$,
with the base case trivially holding. We have
$\displaystyle n\cdot p_{A^{+}}(n)$ $\displaystyle=\sum_{a\in
A^{+}\cap[n]}a\sum_{1\leq k\leq n/a}p_{A^{+}}(n-ak)\leq\sum_{a\in
A^{+}\cap[n]}a\sum_{1\leq k\leq n/a}e^{c\sqrt{n-ak}}$ $\displaystyle\leq
e^{c\sqrt{n}}\sum_{a\in A^{+}\cap[n]}a\sum_{1\leq k\leq
n/a}e^{-\frac{cak}{2\sqrt{n}}}\leq e^{c\sqrt{n}}\sum_{k=1}^{\infty}\sum_{a\in
A^{+}}ae^{-\frac{cak}{2\sqrt{n}}}$ $\displaystyle\leq
e^{c\sqrt{n}}\sum_{k=1}^{\infty}\frac{4|R|n}{mc^{2}k^{2}}=ne^{c\sqrt{n}}\frac{4|R|}{mc^{2}}\sum_{k=1}^{\infty}\frac{1}{k^{2}}=n\cdot
e^{c\sqrt{n}}\;,$
where the first equality is (6), the first inequality is by the induction
hypothesis, the second inequality uses the elementary fact
$\sqrt{n-rk}\leq\sqrt{n}-\frac{rk}{2\sqrt{n}}$, and in the last inequality we
applied (5) with $x=\frac{ck}{2\sqrt{n}}$. Dividing both sides by $n$ we
obtain the theorem.
### Bounding $p_{A^{+}}(n)$ vs. bounding $p_{A}(n)$.
The reader might be wondering why bounding $p_{A^{+}}(n)$ is so much easier
than bounding $p_{A}(n)$. The answer is that the former gives us inequality
(4) from which we obtain the clean inequality (5). To illustrate the
complication that arises when working with $p_{A}(n)$, let us take $A$ to be
the set of odd integers. Then, running the same argument, instead of (4), one
would have liked to use the inequality
$\frac{e^{-x}+e^{-3x}}{(1-e^{-2x})^{2}}\leq\frac{1}{2x^{2}}$, which is false.
To overcome this, one then needs to use the fact that this inequality is
approximately correct for small $x$, which significantly complicates the
proof.
Acknowledgement: We would like to thank the referees for their detailed and
helpful comments.
## References
* [1] Erdős, P. (1942). On an elementary proof of some asymptotic formulas in the theory of partitions. Ann. Math. 43(3): 437–450.
* [2] Hardy, G. H., Ramanujan S. (1917). Asymptotic formulae for the distribution of integers of various types. Proc. London Math. Soc. 16(2): 112–132.
* [3] Meinardus, G. (1954). Asymptotische Aussagen über Partitionen. Math. Z. 61: 388–398.
* [4] Nathanson, M. B. (2002). On Erdős’s elementary method in the asymptotic theory of partitions. In: Halász, G., Lovász, L., Simonovits, M., Sós, V. T., eds. Paul Erdős and his Mathematics, I. Bolyai Soc. Math. Stud., volume 11. Budapest: János Bolyai Mathematical Society, pp. 515–531.
* [5] Nathanson, M. B. (2000). Elementary Methods in Number Theory. Graduate Texts in Mathematics, vol. 195. Berlin: Springer.
* [6] Uspensky, J. V. (1920). Asymptotic expressions of numerical functions occurring in problems concerning the partition of numbers into summands. Bull. Acad. Sci. de Russie. 14(6): 199–218.
|
∎ ††thanks: A.R.Usha Devi and Sudha acknowledge financial support from the
Department of Science and Technology, India through Project No.
DST/ICPS/QuST/Theme-2/2019/Project#107
11institutetext: Seeta Vasudevrao 22institutetext: Department of Physics,
Bangalore University, Bangalore. 22email<EMAIL_ADDRESS>33institutetext: I. Reena 44institutetext: Department of Physics, Bangalore
University, Bangalore. 44email<EMAIL_ADDRESS>55institutetext:
Sudha 66institutetext: Department of Physics, Kuvempu University,
Shankaraghatta.
Inspire Institute Inc., Alexandria, Virginia, 22303, USA. 66email:
<EMAIL_ADDRESS>77institutetext: A. R. Usha Devi 88institutetext:
Department of Physics, Bangalore University, Bangalore
Inspire Institute Inc., Alexandria, Virginia, 22303, USA. 88email:
<EMAIL_ADDRESS>99institutetext: A. K. Rajagopal 1010institutetext:
Inspire Institute Inc., Alexandria, Virginia, 22303, USA. 1010email:
<EMAIL_ADDRESS>
# Sum Uncertainty Relations: Uncertainty Regions for Qubits and Qutrits
Seeta Vasudevrao I. Reena Sudha A. R. Usha Devi A. K. Rajagopal
###### Abstract
We investigate the notion of uncertainty region using the variance based sum
uncertainty relation for qubits and qutrits. We compare uncertainty region of
the qubit (a 2-level system) with that of the qutrit (3-level system) by
considering sum uncertainty relation for two non-commuting Pauli-like
observables, acting on the two dimensional qubit Hilbert space. We identify
that physically valid uncertainty region of a qubit is smaller than that of a
qutrit. This implies that an enhanced precision can be achieved in the
measurement of incompatible Pauli-like observables acting on the 2-dimensional
subspace of a qutrit Hilbert space. We discuss the implication of the reduced
uncertainties in the steady states of $\Lambda$, V, $\Xi$ types of 3-level
atomic systems. Furthermore, we construct a two-qubit permutation symmetric
state, corresponding to a 3-level system and show that the reduction in the
sum uncertainty value - or equivalently, increased uncertainty region of a
qutrit system – is a consequence of quantum entanglement in the two-qubit
system. Our results suggest that uncertainty region can be used as a
dimensional witness.
###### Keywords:
Sum Uncertainty Relation Uncertainty Region Entanglement 3-Level System
Steady-state Population
††journal: International Journal of Theoretical Physics
## 1 Introduction
Quantum theory prevents assignment of precise values for two or more
incompatible observables simultaneously. Heisenberg’s heuristic argument
Heisenberg highlighted this uncertainty associated with the non-commuting
position ($Q$) and momentum ($P$) observables, in terms of the constraints
placed on the product of their standard deviations. If position of a particle
is measured, prediction of its momentum gets inaccurate and vice versa. A
mathematically formal version of the position-momentum uncertainty relation
$(\Delta Q)(\Delta P)\geq\frac{\hbar}{2}$ was subsequently formulated by
Kennard ken . Furthermore, Robertson rob (motivated by Weyl’s arguments Weyl
) extended the uncertainty relations to any arbitrary pairs of non-commuting
observables $A_{1}$, $A_{2}$.
Different forms of uncertainty relations have been formulated over the years
AK ; ozawa ; Hall ; BLW ; Hoffman ; sw1 ; arun ; kraus ; MU ; sw ; Berta ;
Coles ; Werner ; Li ; Abbot , capturing the trade-off between two or more non-
commuting observables. It has been shown that uncertainty relations play a
crucial role in quantum information processing tasks like quantum key
distribution Berta ; Coles ; Cerf1 ; Cerf2 ; Koshi . Non-trivial state-
dependent uncertainty relations which are experimentally verifiable are found
to be of importance in device-independent cryptography sw1 .
A broader perspective on uncertainty relations, based on the concept of
_uncertainty regions_ , is recently being explored Werner ; Li ; Abbot ; Busch
and provides a geometric visualization of the uncertainty relation. For any
uncertainty relation, the corresponding uncertainty region is the _legitimate
domain_ of standard deviation (or any other measure of uncertainties) of a
pair (or triple) of observables, in the entire range of their possible values
Busch . Points $(\Delta A_{1},\,\Delta A_{2})$ inside the uncertainty region
specify the uncertainty in the simultaneous measurement of a pair of
observables $A_{1}$, $A_{2}$. Different types of uncertainty relations can be
chosen for analysing the uncertainty regions Werner ; Li ; Abbot ; Busch . In
this work we have chosen variance based sum uncertainty relation, a state-
independent uncertainty relation, proposed by Hofmann and Takeuchi Hoffman ,
to analyze the uncertainty regions/minimum of the variance based sum
uncertainty relation for two non-commuting Pauli-like observables.
The structure of the paper is as follows: In Section 2, we outline the
geometry of uncertainty region corresponding to variance based sum-uncertainty
relation for a pair of incompatible Pauli-like observables, when they are
measured in quantum states of qubits (2-level systems) and qutrits (3-level
systems). We show that uncertainty region of qutrits is larger, containing
points with enhanced measurement precision for incompatible Pauli-like
observables, in comparison with that of qubits. In Section 3, we express the
minimum of the sum of variances of two noncommuting Pauli-like observables
$A^{(ij)}_{1}=\vec{\sigma}^{(ij)}\cdot\hat{a}$,
$A^{(ij)}_{2}=\vec{\sigma}^{(ij)}\cdot\hat{b},\ \ \hat{a}\cdot\hat{b}=0$ of a
3-level system, given that both the observables are restricted to the
2-dimensional (qubit) subspace (labelled by the pair of indices
$(ij),\,i<j=1,2,3$) of a 3-level (qutrit) system, in terms of the populations
$\rho_{ii}$, $\rho_{jj}$ in the $i^{\rm th}$ and $j^{\rm th}$ levels. We
discuss implication of the reduced uncertainties in the steady states of
$\Lambda$, V and $\Xi$ types of 3-level atomic systems. Section 4 details the
construction of permutation symmetric two-qubit system corresponding to a
qutrit state and explicit evaluation of equivalent sum uncertainty relation
for Pauli-like observables. In Section 5, we establish that separable two-
qubit states can never achieve utmost precision in the measurement of
incompatible Pauli-like observables. We also show that maximum precision in
measurement of the non-commuting Pauli observables is possible using entangled
symmetric two-qubit states. Section 6 provides concluding remarks.
## 2 Uncertainty regions for qubits and qutrits:
The well-known generalized uncertainty relation rob for observables $A_{1}$,
$A_{2}$ is given by
$(\Delta A_{1})(\Delta
A_{2})\geq\frac{1}{2}\left|\langle[A_{1},\,A_{2}]\rangle\right|$ (1)
where $[A_{1},\,A_{2}]=A_{1}A_{2}-A_{2}A_{1}$ is the commutator and $\Delta
A_{1}$, $\Delta A_{2}$ defined by
$\displaystyle\Delta A_{1}$ $\displaystyle=$
$\displaystyle\sqrt{\Delta^{2}\,A_{1}},\ \ \ \ \ \Delta^{2}A_{1}=\langle
A^{2}_{1}\rangle-\langle A_{1}\rangle^{2}$ $\displaystyle\Delta B$
$\displaystyle=$ $\displaystyle\sqrt{\Delta^{2}\,A_{2}},\ \ \ \ \
\Delta^{2}A_{2}=\langle A^{2}_{2}\rangle-\langle A_{2}\rangle^{2}.$ (2)
are the standard deviations of $A_{1}$, $A_{2}$ in any quantum state $\rho$.
Here, $\langle\cdots\rangle=\mbox{Tr}\,(\rho\cdots)$ is the expectation value
of any observable in the state $\rho$.
Hofmann and Takeuchi Hoffman reformulated the uncertainty relation (1) in the
form of sum of variances. Given a set of non-commuting operators $A_{i}$,
$i=1,\,2,\,\ldots,\ n$, they have shown that
$\sum_{i=1}^{n}\,\Delta^{2}\,A_{i}\geq\,k_{A},\ \ k_{A}\ \mbox{being a non-
negative real number.}$ (3)
It is a _state-independent_ uncertainty relation Hoffman with non-trivial
bound for incompatible observables.
We now set up the sum-uncertainty relation in (3) for a pair of observables
$A_{1}$, $A_{2}$ acting on the most general state of a qubit:
$\rho_{\rm
qubit}=\frac{1}{2}\left[I_{2}+r_{1}\sigma_{1}+r_{2}\sigma_{2}+r_{3}\sigma_{3}\right],\
\ \ r_{1}^{2}+r_{2}^{2}+r_{3}^{2}\leq 1$ (4)
where $\sigma_{i}$, $i=1,\,2,\,3$ are Pauli spin operators, $I_{2}$ is the
two-dimensional identity operator and $\vec{r}=(r_{1},\,r_{2},\,r_{3})$,
($|\vec{r}|\leq 1$) is a real three dimensional vector, the mean spin vector
of the state $\rho_{\rm qubit}$. On choosing $A_{1}=\sigma_{1}$ and
$A_{2}=\sigma_{2}$, we get $\langle A_{1}^{2}\rangle=\langle
A_{2}^{2}\rangle=1$, $\langle A_{1}\rangle=r_{1}$, $\langle
A_{2}\rangle=r_{2}$. With $\Delta^{2}\,A_{1}=1-r_{1}^{2}$,
$\Delta^{2}\,A_{2}=1-r_{2}^{2}$, the sum-uncertainty relation becomes
$\displaystyle\Delta^{2}\,A_{1}+\Delta^{2}\,A_{2}=2-(r_{1}^{2}+r_{2}^{2})\geq
1,$ (5) $\displaystyle 0\leq\Delta A_{1}\leq 1,\ \ \ \ \ \ 0\leq\Delta
A_{2}\leq 1.$
In general, we consider the _orthogonal_ Pauli-observables
$A_{1}=\vec{\sigma}\cdot\hat{a},\ \ \ A_{2}=\vec{\sigma}\cdot\hat{b},\ \
\vec{\sigma}=\left(\sigma_{1},\,\sigma_{2},\,\sigma_{3}\right),\ \ \
\hat{a}\cdot\hat{b}=0$ (6)
and we readily have $\langle A_{1}\rangle=\hat{a}\cdot\vec{r}$, $\langle
A_{2}\rangle=\hat{b}\cdot\vec{r}$, $\langle
A_{1}^{2}\rangle=\hat{a}\cdot\hat{a}=1$, $\langle
A_{2}^{2}\rangle=\hat{b}\cdot\hat{b}=1$ leading to
$\displaystyle\Delta^{2}\,A_{1}=1-\left(\hat{a}\cdot\vec{r}\right)^{2},\ \ \ \
\Delta^{2}\,A_{2}=1-\left(\hat{b}\cdot\vec{r}\right)^{2}$
$\displaystyle\Delta^{2}\,A_{1}+\Delta^{2}\,A_{2}=2-\left(\hat{a}\cdot\vec{r}\right)^{2}-\left(\hat{b}\cdot\vec{r}\right)^{2}.$
(7)
As $|\vec{r}|\leq 1$, $|\hat{a}|=|\hat{b}|=1$, we have
$\hat{a}\cdot\vec{r}\leq 1,\ \ \hat{b}\cdot\vec{r}\leq 1$ and from Eq. (2) we
obtain the following sum-uncertainty relation for orthogonal Pauli-observables
on a qubit:
$\Delta^{2}\,A_{1}+\Delta^{2}\,A_{2}\geq 1\ \ \ \mbox{with}\ \ \ 0\leq\Delta
A_{1}\leq 1,\ \ \ 0\leq\Delta A_{2}\leq 1.$ (8)
The sum uncertainty relation (8) implies that the points $(\Delta
A_{1},\,\Delta A_{2})$ lying outside the circular quadrant form the
uncertainty region for Pauli-observables $A_{1}$, $A_{2}$ measured on a qubit,
as can be seen in Fig. 1.
Figure 1: The uncertainty region (light-blue shaded) of a qubit for orthogonal
Pauli observables $A_{1}$, $A_{2}$ in (6):
Notice that the points $(\Delta A_{1},\,\Delta A_{2})$ close to the origin
correspond to measurements that result in better accuracy. But as the
uncertainty region does not contain points below the circular arc (See Fig.
1), precise joint measurements of Pauli-observables on a qubit are impossible
and Fig. 1 provides clear visualization of this fact.
We now consider a 3-level system,
$\rho_{\rm qutrit}=\omega|\psi\rangle\langle\psi|\oplus(1-\omega),\ \
0\leq\omega\leq 1,$ (9)
obtained by appending an ancillary level to a single-qubit pure state
$|\psi\rangle$. Here (9) corresponds to the state of a qutrit whose explicit
form is given by
$\rho_{\rm
qutrit}=\left(\begin{array}[]{ccc}\frac{\omega(1+r_{3})}{2}&\frac{\omega(r_{1}-ir_{2})}{2}&0\\\
\frac{\omega(r_{1}+ir_{2})}{2}&\frac{\omega(1-r_{3})}{2}&0\\\
0&0&1-\omega\end{array}\right),\ \ \ r_{1}^{2}+r_{2}^{2}+r_{3}^{2}=1$ (10)
Here, $r_{1}$, $r_{2}$, $r_{3}$ are the components of the _unit mean spin
vector_ $\hat{r}$, corresponding to the _pure state_ $|\psi\rangle$ (See (9))
and $\omega$ is a real parameter.
The uncertainty region of the qutrit in (9) has been examined in Ref. Busch ,
for orthogonal Pauli observables
$A_{1}=\vec{\sigma}\cdot\hat{a}\oplus 0,\ \ \ \
A_{2}=\vec{\sigma}\cdot\hat{b}\oplus 0,\ \ \hat{a}\cdot\hat{b}=0.$ (11)
It can be seen that Busch
$\langle A_{1}\rangle=\omega(\hat{a}\cdot\hat{r}),\ \ \ \langle
A_{2}\rangle=\omega(\hat{b}\cdot\hat{r}),\ \ \langle A_{1}^{2}\rangle=\langle
A_{2}^{2}\rangle=\omega,$ (12)
leading to
$\Delta^{2}A_{1}=\omega-\omega^{2}(\hat{a}\cdot\hat{r})^{2},\ \ \
\Delta^{2}A_{2}=\omega-\omega^{2}(\hat{b}\cdot\hat{r})^{2}.$ (13)
On fixing $\Delta^{2}A_{1}$ and minimizing $\Delta^{2}A_{2}$, one obtains
Busch
$\left(\Delta A_{2}\right)_{\rm min}=\Delta A_{1}\sqrt{(1-\Delta^{2}A_{1})}.$
(14)
Similarly, fixing $\Delta^{2}A_{2}$ and minimizing $\Delta^{2}A_{1}$ results
in Busch
$\left(\Delta A_{1}\right)_{\rm min}=\Delta A_{2}\sqrt{(1-\Delta^{2}A_{2})}.$
(15)
From (14), (15), we see that when $\Delta A_{1}=0$, $\Delta A_{2}$ can also
become zero and vice versa. This means, origin $(0,\,0)$ of the co-ordinate
system, a point corresponding to utmost precision in simultaneous measurement,
is physically realizable for measurement of orthogonal Pauli observables on
the qutrit state $\rho_{\rm qutrit}$ Busch . The uncertainty region of the
qutrit is larger in comparison with that of a qubit, containing points near
the origin and origin itself, as can be readily seen in Fig. 2.
Figure 2: The uncertainty region of a qutrit (light blue): The origin, with
$\Delta A_{1}=\Delta A_{2}=0$ and points closer to it lie in the admissible
region.
It is evident from the above observations that simultaneous measurements with
utmost precision are impossible when non-commuting Pauli measurements are
performed on a 2-dimensional Hilbert space (qubit) whereas a 3-dimensional
Hilbert space (qutrit) admits enhanced precision. In the following, we show
that the uncertainty sum for two Pauli-like observables in an arbitrary
3-level system reduces below that for a 2-level system.
## 3 Sum uncertainty relation for 3-level atomic systems
In this section we explore the sum uncertainty relation for two Pauli-like
observables (i.e., atomic operators acting on any 2-level subspace of a
3-level atomic system)
$\displaystyle A^{(ij)}_{1}$ $\displaystyle=$
$\displaystyle\vec{\sigma}^{(ij)}\cdot\hat{a},\ \
\vec{\sigma}^{(ij)}=\left(\sigma^{(ij)}_{1},\,\sigma^{(ij)}_{2},\,\sigma^{(ij)}_{3}\right)$
$\displaystyle A^{(ij)}_{2}$ $\displaystyle=$
$\displaystyle\vec{\sigma}^{(ij)}\cdot\hat{b},\ \ \ \ i<j=1,\,2,\,3,$ (16)
where $\hat{a}\cdot\hat{b}=0$, $\hat{a}\cdot\hat{a}=1=\hat{b}\cdot\hat{b}$ and
$\displaystyle\sigma^{(12)}_{1}$ $\displaystyle=$
$\displaystyle\left(\begin{array}[]{ccc}0&1&0\\\ 1&0&0\\\
0&0&0\end{array}\right),\ \ \ \
\sigma^{(13)}_{1}=\left(\begin{array}[]{ccc}0&0&1\\\ 0&0&0\\\
1&0&0\end{array}\right),\ \ \
\sigma^{(23)}_{1}=\left(\begin{array}[]{ccc}0&0&0\\\ 0&0&1\\\
0&1&0\end{array}\right)$ $\displaystyle\sigma^{(12)}_{2}$ $\displaystyle=$
$\displaystyle\left(\begin{array}[]{ccc}0&-i&0\\\ i&0&0\\\
0&0&0\end{array}\right),\ \
\sigma^{(13)}_{2}=\left(\begin{array}[]{ccc}0&0&-i\\\ 0&0&0\\\
i&0&0\end{array}\right),\ \
\sigma^{(23)}_{2}=\left(\begin{array}[]{ccc}0&0&0\\\ 0&0&-i\\\
0&i&0\end{array}\right)$ $\displaystyle\sigma^{(12)}_{3}$ $\displaystyle=$
$\displaystyle\left(\begin{array}[]{ccc}1&0&0\\\ 0&-1&0\\\
0&0&0\end{array}\right),\ \
\sigma^{(13)}_{3}=\left(\begin{array}[]{ccc}1&0&0\\\ 0&0&0\\\
0&0&-1\end{array}\right),\ \
\sigma^{(23)}_{3}=\left(\begin{array}[]{ccc}0&0&0\\\ 0&1&0\\\
0&0&-1\end{array}\right).$
In an arbitrary 3-level atomic system, characterized by the density matrix,
$\varrho_{{\rm
qutrit}}=\left(\begin{array}[]{ccc}\varrho_{11}&\varrho_{12}&\varrho_{13}\\\
\varrho^{*}_{12}&\varrho_{22}&\varrho_{23}\\\
\varrho^{*}_{13}&\varrho^{*}_{23}&\varrho_{33}\\\ \end{array}\right),$ (20)
we obtain
$\displaystyle\left\langle\left(A_{1}^{(ij)}\right)^{2}\right\rangle$
$\displaystyle=$ $\displaystyle{\rm Tr}\,\left[\varrho_{\rm
qutrit}\,\left(\vec{\sigma}^{(ij)}\cdot\hat{a}\right)^{2}\right]=\varrho_{ii}+\varrho_{jj},$
(21) $\displaystyle\left\langle\left(A_{2}^{(ij)}\right)^{2}\right\rangle$
$\displaystyle=$ $\displaystyle{\rm Tr}\,\left[\varrho_{\rm
qutrit}\,\left(\vec{\sigma}^{(ij)}\cdot\hat{b}\right)^{2}\right]=\varrho_{ii}+\varrho_{jj},$
(22)
and
$\displaystyle\left\langle A_{1}^{(ij)}\right\rangle$ $\displaystyle=$
$\displaystyle{\rm Tr}\,\left[\varrho_{\rm
qutrit}\,\left(\vec{\sigma}^{(ij)}\cdot\hat{a}\right)\right]=\vec{n}^{(ij)}\cdot\hat{a},$
(23) $\displaystyle\left\langle A_{1}^{(ij)}\right\rangle$ $\displaystyle=$
$\displaystyle{\rm Tr}\,\left[\varrho_{\rm
qutrit}\,\left(\vec{\sigma}^{(ij)}\cdot\hat{b}\right)\right]=\vec{n}^{(ij)}\cdot\hat{b},$
(24)
where
$\displaystyle\vec{n}^{(ij)}={\rm Tr}\,[\varrho_{\rm
qutrit}\,\vec{\sigma}^{(ij)}]=\left(2\,{\rm Re}\,\varrho_{ij},\,2\,{\rm
Im}\,\varrho_{ij},\,\varrho_{ii}-\varrho_{jj}\right).$ (25)
Choosing $\hat{a}=\hat{n}^{(ij)}=\vec{n}^{(ij)}/|\vec{n}^{(ij)}|$,
$\hat{b}=\hat{n}^{(ij)}_{\perp}$ and simplifying the sum of variances
$\Delta^{2}A_{1}^{(ij)}+\Delta^{2}A_{2}^{(ij)}$ in the 3-level system (20)
(with the help of (21), (22), (23), (24), (25))), we obtain
$\displaystyle\left[\Delta^{2}A_{1}^{(ij)}+\Delta^{2}A_{2}^{(ij)}\right]$
$\displaystyle=$ $\displaystyle
2\,(\varrho_{ii}+\varrho_{jj})-|\vec{n}^{(ij)}|^{2}$ (26) $\displaystyle=$
$\displaystyle
2\,(\varrho_{ii}+\varrho_{jj})-\left[4\,|\varrho_{ij}|^{2}+(\varrho_{ii}-\varrho_{jj})^{2}\right].$
Positive semidefiniteness of the $(ij)^{\rm th}$ $2\times 2$ block of
$\varrho_{\rm qutrit}$ imposes the condition
$|\vec{n}^{(ij)}|\leq\varrho_{ii}+\varrho_{jj}$ (27)
leading to the following minimum value for the uncertainty sum:
$\displaystyle\left[\Delta^{2}A_{1}^{(ij)}+\Delta^{2}A_{2}^{(ij)}\right]_{\rm
min}$ $\displaystyle=$ $\displaystyle
2\,\left(\varrho_{ii}+\varrho_{jj}\right)-\left(\varrho_{ii}+\varrho_{jj}\right)^{2}.$
(28)
It may be noted that if we restrict ourselves to the 2-level system i.e.,
$i,j=1,2$, we obtain
$\left[\Delta^{2}A_{1}^{(12)}+\Delta^{2}A_{2}^{(12)}\right]_{\rm min}=1$, as
$(\varrho_{11}+\varrho_{22})={\rm Tr}[\varrho]=1$. In other words, the
uncertainty sum is always greater than 1 in a 2-level system indicating that
joint measurement of the non-commuting Pauli observables
$\vec{\sigma}\cdot\hat{a},\ \vec{\sigma}\cdot\hat{b}$ is limited by the sum
uncertainty relation (8). On the other hand, when an additional level is
included, the populations $\varrho_{ii}$, $\varrho_{jj}$ of the $i^{\rm th}$
and $j^{\rm th}$ levels ($i>j=1,2,3$) play a crucial role in enhancing the
measurement precision of the atomic observables
$\vec{\sigma}^{(ij)}\cdot\hat{a}$, $\vec{\sigma}^{(ij)}\cdot\hat{b}$.
In Fig. 3 we have plotted the minimum value of the uncertainty sum
$\left[\Delta^{2}A_{1}^{(ij)}+\Delta^{2}A_{2}^{(ij)}\right]_{\rm min}$ as a
function of the populations $\varrho_{ii}$, $\varrho_{jj}$ (see (28)). It is
clearly seen that the minimum value of the uncertainty sum (28) can take
values smaller than 1. This establishes the advantage of the additional level
for improving precision in the measurement of atomic observables.
Figure 3: Minimum value of the uncertainty sum
$\left[\Delta^{2}A_{1}^{(ij)}+\Delta^{2}A_{2}^{(ij)}\right]_{\rm min}$ as a
function of the populations $\rho_{ii}$, $\rho_{jj}$ (see (28)). Enhanced
measurement precision of incompatible Pauli-like atomic observables is ensured
whenever $\left[\Delta^{2}A_{1}^{(ij)}+\Delta^{2}A_{2}^{(ij)}\right]_{\rm
min}<1$.
Considerable research interest has been evinced in exploring the response of
$\Lambda$, $\Xi$, V types of 3-level atomic systems to lasing radiation
3levelReview . In $\Lambda$-type system (see Fig. 4(a)) transition between the
two lower levels $|3\rangle$, $|2\rangle$ is forbidden and the upper level
$|1\rangle$ is commonly shared in atomic transitions with levels $|3\rangle$
and $|2\rangle$; in V-type system (Fig. 4(b)), transitions from the lower
level $|3\rangle$ with the two upper levels $|1\rangle$ and $|2\rangle$ are
allowed, but the transition $|1\rangle\leftrightarrow|2\rangle$ between the
upper levels is forbidden. While atomic transitions
$|1\rangle\leftrightarrow|2\rangle$ and $|2\rangle\leftrightarrow|3\rangle$
are allowed in $\Xi$ type system (Fig. 4(c)), the transition
$|1\rangle\leftrightarrow|3\rangle$ is forbidden. It is of interest to
consider any two levels of the atomic 3-level system, between which
transitions are allowed, as a qubit, and explore if there is any enhanced
precision in the measurement of two non-commuting qubit operators. To this
end, we consider the coherent population trapping state in a $\Lambda$ type
atomic system radmore82 ; pra95 ; pro_opt96 ; pra97 :
$\displaystyle\varrho^{\Lambda}_{33}=\frac{1}{2}=\varrho^{\Lambda}_{22},\ \
\varrho^{\Lambda}_{11}=0.$ (29)
It is clearly seen that the uncertainty sum (28) with $i=1,\,j=2$ and
$i=1,\,j=3$ is given by
$\displaystyle\left[\Delta^{2}A_{1}^{(12)}+\Delta^{2}A_{2}^{(12)}\right]^{\Lambda}_{\rm
min}=2\,(\varrho^{\Lambda}_{11}+\varrho^{\Lambda}_{22})-(\varrho^{\Lambda}_{11}+\varrho^{\Lambda}_{22})^{2}=0.75$
$\displaystyle\left[\Delta^{2}A_{1}^{(13)}+\Delta^{2}A_{2}^{(13)}\right]^{\Lambda}_{\rm
min}=2\,(\varrho^{\Lambda}_{11}+\varrho^{\Lambda}_{33})-(\varrho^{\Lambda}_{11}+\varrho^{\Lambda}_{33})^{2}=0.75$
revealing improved precision in the measurements of the qubit operator pairs
$\left\\{A_{1}^{(12)},\,A_{2}^{(12)}\right\\}$ and
$\left\\{A_{1}^{(13)},\,A_{2}^{(13)}\right\\}$.
In the case of V-type 3-level atom, with the transition
$|3\rangle\leftrightarrow|2\rangle$ driven by a strong-coupling laser field
and $|3\rangle\leftrightarrow|1\rangle$ transition driven by an incoherent
pump field the steady state populations are given by pra96
$\displaystyle\varrho^{V}_{11}\approx 0.2,\
\varrho^{V}_{22}\approx\varrho^{V}_{33}\approx 0.4.$ (30)
The uncertainty sum (28) of Pauli-like atomic observables associated with
$i=1,\,j=3$ and $i=2,\,j=3$ are given by
$\displaystyle\left[\Delta^{2}A_{1}^{(13)}+\Delta^{2}A_{2}^{(13)}\right]^{V}_{\rm
min}=2\,(\varrho^{V}_{11}+\varrho^{V}_{33})-(\varrho^{V}_{11}+\varrho^{V}_{33})^{2}=0.84$
$\displaystyle\left[\Delta^{2}A_{1}^{(23)}+\Delta^{2}A_{2}^{(23)}\right]^{V}_{\rm
min}=2\,(\varrho^{V}_{22}+\varrho^{V}_{33})-(\varrho^{V}_{22}+\varrho^{V}_{33})^{2}=0.96.$
Thus the steady state of V-type atomic qutrit (see (30)) offers advantage over
2-level atomic system in reducing the uncertainty sum of non-commuting Pauli-
like atomic observables.
Figure 4: Schematic diagrams of (a) $\Lambda$-type (b) V-type and (c)
$\Xi$-type 3-level atomic systems. Arrows between the energy levels indicate
allowed transitions.
Populations in the steady state of a 3-level $\Xi$ atomic system pra97 ; pra98
satisfy the condition
$\displaystyle\varrho^{\Xi}_{11}=\varrho^{\Xi}_{22}\leq\frac{1}{3},\ \ \
\frac{1}{3}\leq\varrho^{\Xi}_{33}\leq\frac{1}{2}.$ (31)
We thus obtain the limiting value of the uncertainty sum as,
$\displaystyle\left[\Delta^{2}A_{1}^{(12)}+\Delta^{2}A_{2}^{(12)}\right]^{\Xi}_{\rm
min}$ $\displaystyle=$ $\displaystyle
2\,\left(1-\varrho^{\Xi}_{33}\right)-\,\left(1-\varrho^{\Xi}_{33}\right)^{2};\
\ \ \frac{1}{3}\leq\varrho^{\Xi}_{33}\leq\frac{1}{2},$ $\displaystyle\implies$
$\displaystyle\hskip
14.45377pt\frac{3}{4}\leq\left[\Delta^{2}A_{1}^{(12)}+\Delta^{2}A_{2}^{(12)}\right]^{\Xi}_{\rm
min}\leq\frac{8}{9},$
$\displaystyle\left[\Delta^{2}A_{1}^{(23)}+\Delta^{2}A_{2}^{(23)}\right]^{\Xi}_{\rm
min}$ $\displaystyle=$ $\displaystyle
2\,\left(1-\varrho^{\Xi}_{11}\right)-\,\left(1-\varrho^{\Xi}_{11}\right)^{2};\
\ \ 0\leq\varrho^{\Xi}_{11}\leq\frac{1}{3},$ $\displaystyle\implies$
$\displaystyle\hskip
14.45377pt\frac{8}{9}\leq\left[\Delta^{2}A_{1}^{(12)}+\Delta^{2}A_{2}^{(12)}\right]^{\Xi}_{\rm
min}\leq 1,$
for the atomic Pauli-like observables associated with $i=1,\,j=2$ and
$i=2,j=3$ atomic levels of the $\Xi$-type atomic system. More detailed
investigations on improving measurement precision of non-commuting Pauli-like
atomic observables of a driven 3-level system, beyond what can be achieved in
a 2-level system, will be reported separately.
## 4 Transformation of qutrit state into a two-qubit symmetric state
We now analyze the enhanced accuracy of orthogonal Pauli measurements on
qutrit states from a perspective based on the separability/non-separability of
two-qubit symmetric states corresponding to qutrit states. In the following,
we detail the construction of two-qubit _symmetric states_ from qutrit states,
in particular the state in (9). We also carry out an analysis of the role, if
any, of two-qubit entanglement in the measurement precision possible in the
qutrit state (9).
A two-qubit symmetric state belongs to the 3-dimensional maximum multiplicity
space of the collective angular momentum $j=j_{1}+j_{2}$,
$j_{1}=j_{2}=\frac{1}{2}$. With the dimensions of the qutrit space and that of
the two-qubit symmetric state (expressed in angular momentum basis) being
equal, they have a one-one correspondence. Here, we outline this
correspondence and accomplish the construction of two-qubit symmetric states
corresponding to qutrit states.
The qutrit state in (9), expressed as a $3\times 3$ matrix in (10), can be
written equivalently as
$\rho_{\rm
qutrit}\equiv\left(\begin{array}[]{cccc}\frac{\omega(1+r_{3})}{2}&\frac{\omega(r_{1}-ir_{2})}{2}&0&0\\\
\frac{\omega(r_{1}+ir_{2})}{2}&\frac{\omega(1-r_{3})}{2}&0&0\\\
0&0&1-\omega&0\\\ 0&0&0&0\end{array}\right),$ (32)
The density matrix $\rho_{AB}$ of a two-qubit symmetric state is given by uma
$\rho_{AB}=\frac{1}{4}\left[I_{2}\otimes
I_{2}+\sum_{i=1}^{3}\,s_{i}\left(\sigma_{i}\otimes
I_{2}+I_{2}\otimes\sigma_{i}\right)+\sum_{i,j=1}^{3}\,t_{ij}\,(\sigma_{i}\otimes\sigma_{j})\right],$
(33)
where $t_{ij}=t_{ji}$. The elements $\rho_{ij}$, $i,\,j=1,\,2,\,3,\,4$ of the
$4\times 4$ matrix $\rho_{AB}$ are explicitly given by
$\displaystyle\rho_{11}$ $\displaystyle=$
$\displaystyle\frac{1}{4}\left(1+2s_{3}+t_{33}\right),\ \ \
\rho_{22}=\frac{1}{4}\left(1-t_{33}\right)=\rho_{33}$ $\displaystyle\rho_{12}$
$\displaystyle=$
$\displaystyle\rho_{13}=\frac{1}{4}\left(s_{1}-is_{2}+t_{13}-it_{23}\right)=\rho^{\ast}_{21}=\rho^{\ast}_{31}$
$\displaystyle\rho_{14}$ $\displaystyle=$
$\displaystyle\frac{1}{4}\left(t_{11}-t_{22}-2it_{12}\right)=\rho^{\ast}_{14}$
$\displaystyle\rho_{24}$ $\displaystyle=$
$\displaystyle\rho_{34}=\frac{1}{4}\left(s_{1}-is_{2}-t_{13}+it_{23}\right)=\rho^{\ast}_{42}=\rho^{\ast}_{43}$
$\displaystyle\rho_{23}$ $\displaystyle=$
$\displaystyle\frac{1}{4}\left(t_{11}+t_{22}\right)=\rho_{32},\ \ \
\rho_{44}=\frac{1}{4}\left(1-2s_{3}+t_{33}\right)$ (34)
The standard basis for a two-qubit state is given by the direct product basis
(uncoupled basis) consisting of orthonormal vectors
$|1/2;1/2\rangle,\ \ |1/2;-1/2\rangle,\ \ |-1/2;1/2\rangle,\ \
|-1/2;-1/2\rangle$ (35)
where $|m_{1};m_{2}\rangle=|m_{1}\rangle\otimes m_{2}\rangle$,
$m_{1},\,m_{2}=1/2,\,-1/2$,
One can readily express the direct product basis $\\{|m_{1};m_{2}\rangle\\}$
in terms of the collective angular momentum basis (coupled basis)
$\\{|jm\rangle\\}$, ($j=1,\,0$, $-j\leq m\leq j$ for each $j$) and vice versa,
through
$\displaystyle|11\rangle$ $\displaystyle=$ $\displaystyle|1/2;1/2\rangle,\ \ \
|10\rangle=\frac{1}{\sqrt{2}}\left(|1/2;-1/2\rangle+|-1/2;1/2\rangle\right)$
(36) $\displaystyle|1-1\rangle$ $\displaystyle=$
$\displaystyle|-1/2;-1/2\rangle,\ \ \
|00\rangle=\frac{1}{\sqrt{2}}\left(|1/2;-1/2\rangle-|-1/2;1/2\rangle\right).$
From (36), it follows that the unitary matrix
$U=\left(\begin{array}[]{cccc}1&0&0&0\\\ 0&0&0&1\\\
0&\frac{1}{\sqrt{2}}&\frac{1}{\sqrt{2}}&0\\\
0&\frac{1}{\sqrt{2}}&\frac{-1}{\sqrt{2}}&0\end{array}\right),\ \
U^{\dagger}\,U=U\,U^{\dagger}=I_{4}$ (37)
where † denotes hermitian conjugate, corresponds to the transformation from
coupled basis to uncoupled basis. The similarity transformation
$U^{\dagger}\rho_{\rm qutrit}U$ effects the transformation of the state
$\rho_{\rm qutrit}$ (See (9), (32)) to its equivalent two-qubit symmetric
state $\rho_{AB}$. Explicitly, we have
$\rho_{AB}=U^{\dagger}\rho_{\rm
qutrit}U=\frac{1}{2}\left(\begin{array}[]{cccc}(1+r_{3})\omega&0&0&(r_{1}-ir_{2})\omega\\\
0&1-\omega&1-\omega&0\\\ 0&1-\omega&1-\omega&0\\\
(r_{1}+ir_{2})\omega&0&0&(1-r_{3})\omega\end{array}\right).$ (38)
On comparing the elements (See (4)) of $\rho_{AB}$ with the corresponding
elements in Eq. (38), we obtain the following relation between the parameters
$\omega$, $r_{1}$, $r_{2}$, $r_{3}$, ($r_{1}^{2}+r_{2}^{2}+r_{3}^{2}=1$) of
the qutrit state in (9) and the parameters $s_{i}$, $t_{ij}$,
$i,\,j=1,\,2,\,3$ of the symmetric two-qubit state $\rho_{AB}$ in (33). That
is, the non-zero parameters $s_{i}$, $t_{ij}$, $i,\,j=1,\,2,\,3$ of
$\rho_{AB}$ (See (33)) are seen to be
$\displaystyle s_{3}$ $\displaystyle=$ $\displaystyle\omega r_{3},\ \ \
t_{12}=t_{21}=\omega r_{2}$ (39) $\displaystyle t_{11}$ $\displaystyle=$
$\displaystyle(1-\omega)+\omega r_{1},\ \ \ \ t_{22}=(1-\omega)-\omega r_{1},\
\ t_{33}=2\omega-1.$
Thus, a two-qubit symmetric state $\rho_{AB}$ in (33)) with its elements given
in (39) corresponds to the qutrit state $\rho_{\rm qutrit}$ in (9).
## 5 Sum uncertainty relation for two-qubit state $\rho_{AB}$
Here, we set up the sum uncertainty relation for the two-qubit state
$\rho_{AB}$ (See (38)) in order to establish the equivalence of its
uncertainty region with that of the qutrit state $\rho_{\rm qutrit}$ (See
(9)). In order to do this, we need to recognize the two-qubit observables
${\mathcal{A}}_{1}$, ${\mathcal{A}}_{2}$, which are equivalent to $A_{1}$,
$A_{2}$ (See (11)). For simplicity, and without loss of generality, we choose
the orthonormal vectors in (11) to be $\hat{a}=(1,\,0,\,0)$,
$\hat{b}=(0,\,1,\,0)$ so that $A_{1}=\sigma_{1}\oplus 0$,
$A_{2}=\sigma_{2}\oplus 0$. It is not difficult to see that we can express
$A_{1}$, $A_{2}$ as
$A_{1}=\sigma_{1}\oplus{\bf 0}_{2},\ \ A_{2}=\sigma_{2}\oplus{\bf 0}_{2},\ \
{\bf 0}_{2}=\left(\begin{array}[]{cc}0&0\\\ 0&0\end{array}\right),$ (40)
to facilitate their action on the qutrit state $\rho_{\rm qutrit}$ expressed
as a $4\times 4$ matrix in (32). Corresponding to the basis transformation
$\rho_{\rm qutrit}\longrightarrow\rho_{AB}$ in (38), the observables $A_{1}$,
$A_{2}$ (See (40)) undergo the similarity transformation
$\displaystyle{\mathcal{A}}_{1}$ $\displaystyle=$ $\displaystyle
U^{\dagger}\,A_{1}U=\frac{1}{2}\left[\sigma_{1}\otimes\sigma_{1}-\sigma_{2}\otimes\sigma_{2}\right],$
$\displaystyle{\mathcal{A}}_{2}$ $\displaystyle=$ $\displaystyle
U^{\dagger}\,A_{2}U=\frac{1}{2}\left[\sigma_{1}\otimes\sigma_{2}+\sigma_{2}\otimes\sigma_{1}\right].$
(41)
Here $U$ is the basis transformation matrix (See (37)) that takes $\rho_{\rm
qutrit}$ to $\rho_{AB}$ (See (38)). On explicit evaluation, we get
$\displaystyle\langle{\mathcal{A}}_{1}\rangle=\mbox{Tr}\,({\mathcal{A}}_{1}\rho_{AB})=\omega\,r_{1},\
\ \
\langle{\mathcal{A}}_{1}^{2}\rangle=\mbox{Tr}\,({\mathcal{A}}_{1}^{2}\rho_{AB})=\omega$
$\displaystyle\langle{\mathcal{A}}_{2}\rangle=\mbox{Tr}\,({\mathcal{A}}_{2}\rho_{AB})=\omega\,r_{2},\
\ \
\langle{\mathcal{A}}_{2}^{2}\rangle=\mbox{Tr}\,({\mathcal{A}}_{2}^{2}\rho_{AB})=\omega$
$\displaystyle\Delta^{2}{\mathcal{A}}_{1}=\omega-\omega^{2}\,r_{1}^{2},\ \ \ \
\ \ \ \ \ \ \ \Delta^{2}{\mathcal{A}}_{2}=\omega-\omega^{2}\,r_{2}^{2}.$ (42)
The expressions for $\Delta^{2}{\mathcal{A}}_{1}$,
$\Delta^{2}{\mathcal{A}}_{2}$ in (5) are the same as that obtained in (13) for
the qutrit state $\rho_{\rm qutrit}$ (See (9)). The uncertainty region of the
two-qubit state $\rho_{AB}$ is thus the same as that of the qutrit state
$\rho_{\rm qutrit}$, with the origin
$(\Delta{\mathcal{A}}_{1},\,\Delta{\mathcal{A}}_{2})=(0,\,0)$ being a
physically realizable point (See Fig. 2).
As $r_{1}^{2}+r_{2}^{2}+r_{3}^{2}=1$, the sum uncertainty relation of the two-
qubit state $\rho_{AB}$ can be simplified to (See (5))
$\Delta^{2}{\mathcal{A}}_{1}+\Delta^{2}{\mathcal{A}}_{2}=2\omega-\omega^{2}\,\kappa^{2},\
\ \ \kappa=\sqrt{r_{1}^{2}+r_{2}^{2}}=\sqrt{1-r_{3}^{2}}.$ (43)
Fig. 4 shows the variation of the uncertainty sum
$\Delta^{2}{\mathcal{A}}_{1}+\Delta^{2}{\mathcal{A}}_{2}$ as a function of the
parameters $0\leq\omega,\kappa\leq 1$.
Figure 5: The contour plot showing the variation of the uncertainty sum
$\Delta^{2}{\mathcal{A}}_{1}+\Delta^{2}{\mathcal{A}}_{2}$ with respect to the
parameters $\omega$ and $\kappa$. The dotted line corresponds to
$\Delta^{2}{\mathcal{A}}_{1}+\Delta^{2}{\mathcal{A}}_{2}=\frac{3}{4}$.
### 5.1 Sum uncertainty relation for symmetric two-qubit separable states
Our intention in obtaining the two-qubit counterpart $\rho_{AB}$ of the qutrit
state $\rho_{\rm qutrit}$ lies in utilizing its _composite nature_ and examine
whether separability/non-separability of $\rho_{AB}$ has any role in the
better precision observed in joint measurement of Pauli observables on the
_single_ party state $\rho_{\rm qutrit}$. In order to carry out this task, we
consider the most general bipartite, symmetric separable state
$\rho_{\rm sep}=\sum_{i}\,p_{i}\left(\rho_{i}\otimes\rho_{i}\right),\ \
i=1,\,2,\,3\ldots$ (44)
with $0\leq p_{i}\leq 1$, $\sum_{i}\,p_{i}=1$ being the probabilities. It has
been shown in Ref. pfun that the single qubit density operators $\rho_{i}$,
$i=1,\,2,\,3\ldots$ constituting any symmetric separable state $\rho_{\rm
sep}$ are necessarily _pure_. Thus,
$\rho_{i}=\frac{1}{2}\,\left(I_{2}+\vec{\sigma}\cdot\hat{s}_{i}\right),\ \
\hat{s}_{i}=\left(s_{1i},\,s_{2i},\,s_{3i}\right),\ \
s_{1i}^{2}+s_{2i}^{2}+s_{3i}^{2}=1.$ (45)
We evaluate the expectation values of ${\mathcal{A}}_{\alpha},$ and
${\mathcal{A}}^{2}_{\alpha}$, $\alpha=1,2$ (See (5)) in a product state
$\rho_{i}\otimes\rho_{i}$ (where $\rho_{i}$ is given by (45)):
$\displaystyle\langle{\mathcal{A}}_{1}\rangle_{i}$ $\displaystyle=$
$\displaystyle\mbox{Tr}\,\left[\left(\rho_{i}\otimes\rho_{i}\right){\mathcal{A}}_{1}\right]=\frac{1}{2}\,\left(s_{1i}^{2}-s_{2i}^{2}\right),$
$\displaystyle\langle{\mathcal{A}}_{1}^{2}\rangle_{i}$ $\displaystyle=$
$\displaystyle\mbox{Tr}\,\left[\left(\rho_{i}\otimes\rho_{i}\right){\mathcal{A}}_{1}^{2}\right]=\frac{1}{2}\,\left(1+s_{3i}^{2}\right),$
$\displaystyle\langle{\mathcal{A}}_{2}\rangle_{i}$ $\displaystyle=$
$\displaystyle\mbox{Tr}\,\left[\left(\rho_{i}\otimes\rho_{i}\right){\mathcal{A}}_{2}\right]=s_{1i}s_{2i},$
$\displaystyle\langle{\mathcal{A}}_{2}^{2}\rangle_{i}$ $\displaystyle=$
$\displaystyle\mbox{Tr}\,\left[\left(\rho_{i}\otimes\rho_{i}\right){\mathcal{A}}_{2}^{2}\right]=\frac{1}{2}\,\left(1+s_{3i}^{2}\right),$
leading to
$\displaystyle\left(\Delta^{2}\,{\mathcal{A}}_{1}\right)_{i}$ $\displaystyle=$
$\displaystyle\frac{1}{2}\left[1+s_{3i}^{2}-\frac{1}{2}(s_{1i}^{2}-s_{2i}^{2})^{2}\right]$
$\displaystyle\left(\Delta^{2}\,{\mathcal{A}}_{2}\right)_{i}$ $\displaystyle=$
$\displaystyle\frac{1}{2}\left[1+s_{3i}^{2}-2s_{1i}^{2}s_{2i}^{2}\right]$ (46)
Using $s_{1i}^{2}+s_{2i}^{2}+s_{3i}^{2}=1$ and on simplification, we get
$\left(\Delta^{2}\,{\mathcal{A}}_{1}\right)_{i}+\left(\Delta^{2}\,{\mathcal{A}}_{2}\right)_{i}=\frac{3}{4}+\frac{3}{2}s_{3i}^{2}-\frac{1}{4}s_{3i}^{4}$
(47)
From the structure of $\rho_{\rm{sep}}$ (See (44)), we readily have
$\Delta^{2}{\mathcal{A}}_{1}=\sum_{i}\,p_{i}\,\left(\Delta^{2}\,{\mathcal{A}}_{1}\right)_{i}$,
$\Delta^{2}{\mathcal{A}}_{2}=\sum_{i}\,p_{i}\,\left(\Delta^{2}\,{\mathcal{A}}_{2}\right)_{i}$
and hence we get (See (47))
$\Delta^{2}\,{\mathcal{A}}_{1}+\Delta^{2}\,{\mathcal{A}}_{2}=\frac{3}{4}+\frac{3}{2}\sum_{i}\,p_{i}\,s_{3i}^{2}-\frac{1}{4}\sum_{i}\,p_{i}\,s_{3i}^{4}$
(48)
As $0\leq s_{3i}\leq 1$, it readily follows that
$\left(\Delta^{2}\,{\mathcal{A}}_{1}+\Delta^{2}\,{\mathcal{A}}_{2}\right)_{\rm
min}=\frac{3}{4}$ (49)
which happens when $s_{3i}=0$ for all $i=1,\,2,\,3,\cdots$. In other words,
$\Delta^{2}\,{\mathcal{A}}_{1}+\Delta^{2}\,{\mathcal{A}}_{2}\geq\frac{3}{4}$
(50)
is the sum-uncertainty relation for symmetric separable two-qubit states, with
its lowest bound being $3/4$. Thus the uncertainty sum $\rho_{\rm sep}$ (See
(44)), set up for the two-qubit observables ${\mathcal{A}}_{1}$,
${\mathcal{A}}_{2}$ in (5) cannot even go close to zero. This implies that
symmetric separable states (See (44)) can never achieve maximum accuracy in
joint measurements by ${\mathcal{A}}_{1}$, ${\mathcal{A}}_{2}$ in (5).
We now wish to check whether entanglement in the two-qubit state $\rho_{AB}$
contributes to enhanced precision in the measurements of the observables
${\mathcal{A}}_{1}$, ${\mathcal{A}}_{2}$. To this end, we evaluate the
concurrence Wootters ; hill , a measure of two-qubit entanglement, of the
state $\rho_{AB}$. Concurrence of any arbitrary two-qubit state $\rho$ is
defined as Wootters ; hill
$C=\mbox{max}\,\left(0,\,\sqrt{\lambda_{1}}-\sqrt{\lambda_{2}}-\sqrt{\lambda_{3}}-\sqrt{\lambda_{4}}\right)$
(51)
where $\lambda_{k}$, $k=1,\,2,\,3,\,4$ are the eigenvalues of the matrix
$\rho(\sigma_{y}\otimes\sigma_{y})\rho^{\ast}(\sigma_{y}\otimes\sigma_{y})$,
arranged in the descending order (i.e.,
$\lambda_{1}\geq\lambda_{2}\geq\lambda_{3}\geq\lambda_{4}$).
The structure of $\rho_{AB}$ in (38) allows us to make use of the simplified
expression for concurrence given in Ref. wang , and leads to
$C_{AB}=\bigg{\\{}\begin{array}[]{cc}\omega(1+\kappa)-1\ \ \ \mbox{for}\ \ \
\omega(1+\kappa)\geq 1\\\ 1-\omega(1+\kappa)\ \ \ \mbox{for}\ \ \ \
\omega(1+\kappa)\leq 1\end{array}$ (52)
where $\kappa=\sqrt{1-r_{3}^{2}}$ and $0\leq r_{3}\leq 1$. In other words, we
have
$C_{AB}=\left|\omega(1+\kappa)-1\right|,\ \ \kappa=\sqrt{1-r_{3}^{2}},\ \
0\leq r_{3}\leq 1.$ (53)
A contour plot of $C_{AB}$ as a function of the parameters $\omega$ and
$\kappa$ is shown in Fig. 6.
Figure 6: The contour plot showing the variation of concurrence $C_{AB}$ with
respect to the parameters $\omega$ and $\kappa$
Based on Figs. 5 and 6, we reach the following conclusions:
1. 1.
The uncertainty sum
$\Delta^{2}\,{\mathcal{A}}_{1}+\Delta^{2}\,{\mathcal{A}}_{2}$ can be reduced
below the value $3/4$ only in an entangled state $\rho_{AB}$ (i.e., when
$C_{AB}\neq 0$). In particular,
$\Delta^{2}\,{\mathcal{A}}_{1}+\Delta^{2}\,{\mathcal{A}}_{2}\longrightarrow
0$, in maximally entangled two-qubit symmetric states (i.e., when
$C_{AB}\rightarrow 1$).
2. 2.
While no separable state can reduce the uncertainty sum
$\Delta^{2}\,{\mathcal{A}}_{1}+\Delta^{2}\,{\mathcal{A}}_{2}$ below the value
$3/4$, there indeed exist entangled states with
$\Delta^{2}\,{\mathcal{A}}_{1}+\Delta^{2}\,{\mathcal{A}}_{2}\geq 3/4$. This
implies that, while qutrit states $\rho_{\rm qutrit}$ that permit accurate
simultaneous measurements of the observables $A_{1},\ A_{2}$ are necessarily
associated with entangled two-qubit states, the converse is not always true.
It can thus be concluded that entanglement in a two-qubit state constructed
from a qutrit (a qubit appended with an additional level) plays a significant
role in the precise joint measurements by a pair of orthogonal Pauli
observables. It would be of interest to examine whether two-qubit states
constructed from a qudit (a qubit with $d-1$ ancillary levels) exhibit a
similar feature. A study of uncertainty region of such two-qubit states,
dimensional dependence of uncertainty sum and accuracy of simultaneous
measurements by _any_ incompatible pair of observables in $d$-dimensional
spaces form topics of further interest.
## 6 Conclusion
This work is a contribution to the ongoing study on uncertainty regions Werner
; Li ; Abbot ; Busch providing a different perspective in accounting for
better measurement precision seen in qutrits. For any arbitrary 3-level atomic
systems we have obtained an expression for minimum value of uncertainty sum
for Pauli-like observables in terms of atomic populations. This is useful to
study if enhanced measurement precision (reduction in the uncertainty sum) can
be realized in $\Lambda$, V and $\Xi$ types of 3-level atomic systems, which
are characterised by different schemes of allowed/forbidden atomic transitions
between any two levels. We have also examined whether entanglement in a two-
qubit state (which is constructed from a qutrit state of specific structure)
is responsible for the inclusion of more precisely determinable points in the
uncertainty region. We have shown that while simultaneous measurements of
orthogonal Pauli-like observables with utmost accuracy is not possible for
separable two-qubit states, entangled states allow for such precise
measurements. The technique we have proposed for the construction of two-qubit
symmetric states corresponding to qutrit states is helpful in such a
construction from multi-level states obtained by appending several ancillary
levels to a single qubit. It would be of interest to examine whether increase
in the dimension of Hilbert space leads to better measurement precision of
incompatible observables. Our work suggests that uncertainty region could be
employed as a dimensional witnesse.
## References
* (1) W. Heisenberg. Z. Phys. 43 (1927) 172-198
* (2) E. H. Kennard. Z. Phys. 44 (1927) 326-352.
* (3) H. P. Robertson. Phys. Rev. 34 (1929) 163-164.
* (4) H. Weyl. Gruppentheorie und Quantenmechanik Leipzig, (1928).
* (5) E. Arthurs and J. L. Kelly. Bell System Technical Journal 44 (1964) 725–729.
* (6) M. Ozawa. Phys. Rev. A 67 (2003) 042105 (1–6).
* (7) M. J. W. Hall. Phys. Rev. A 69 (2004) 052113 (1–12).
* (8) P. Busch, P. Lahti, and R. F. Werner. Phys. Rev. Lett. 111 (2013) 160405 (1-5).
* (9) H. F. Hofmann and S. Takeuchi. Phys. Rev. A 68 (2003) 032103 (1-6).
* (10) J. Kaniewski, M. Tomamichel and S. Wehner, Phys. Rev. A 90, (2014) 012332 (1-8).
* (11) L. Maccone and A. K. Pati. Phys. Rev. Lett. 113 (2014) 260401 (1-5).
* (12) K. Kraus, Phys. Rev. D. 35 (1987) 3070–3075.
* (13) H. Maassen and J. B. M. Uffink. Phys. Rev. Lett. 60 (1988) 1103-1106.
* (14) S. Wehner and A. Winter. New J. Phys. 12: (2010) 025009 (1-22).
* (15) M. Berta, M. Christandl, R. Colbeck, J. M. Renes, and R. Renner. Nature Physics 6 (2010) 659-662.
* (16) P. J. Coles, M. Berta, M. Tomamichel, and S. Wehner. Rev. Mod. Phys. 89 (2017) 015002 (1-58).
* (17) L. Dammeier, R. Schwonnek and R.F. Werner. New J. Phys. 17 (2015) 093046(1-31).
* (18) J. Li, and C. Qiao. Scientific Reports 5 (2015) 12708 (1-7).
* (19) A. A. Abbott, P. L. Alzieu, M. J. W. Hall and C. Branciard. Mathematics 4 (2016) 8 (1-17).
* (20) N. J. Cerf, M. Bourennane, A. Karlsson, and N. Gisin. Phys. Rev. Lett., 88 (2002) 127902 (1-4).
* (21) F. Grosshans, and N. J. Cerf. Phys. Rev. Lett. 92 (2004) 047905 (1-4).
* (22) M. Koashi. J. Phys.: Conf. Ser., 36 (2006) 98-102.
* (23) P. Busch and O. Reardon-Smith. arXiv:1901.03695 (quant-ph) (2019).
* (24) H.I. Yoo and J.H. Eberly. Phys. Rep. 118 (1985) 239-337.
* (25) P.M. Radmore and P.L. Knight. J. Phys. B. 15 (1982) 561-573.
* (26) I.V. Jyotsna and G.S. Agarwal. Phys. Rev. A. 52 (1995) 3147–3152.
* (27) E. Arimondo. Progress in Optics, 35 (1996) 257–-354.
* (28) M. Blaauboer. Phys. Rev. A., 55, (1997) 2459–2462.
* (29) Y. Zhu, Phys. Rev. A. 53, (1996) 2742–2747.
* (30) J. Mompart, C. Peters, and R. Corbalán. Phys. Rev. A. 57, (1998) 2163–2168.
* (31) A. R. Usha Devi, M. S. Uma, R. Prabhu and Sudha. J.Opt.B: Quantum Semiclass. Opt., 7 (2005) S740–S744.
* (32) A. R. Usha Devi, A. K. Rajagopal, Sudha, H. S. Karthik and J. Prabhu Tej, Quantum Inf Process 12 (2013) 3717–3723.
* (33) W. K. Wootters. Phys. Rev. Lett. 80 (1997) 2245–2248.
* (34) S. Hill and W. K. Wootters Phys. Rev. Lett. 78 (1997) 5022–5025.
* (35) X. Wang and K. Molmer Eur. Phys. J. D. 18 (2002) 385-391
|
# Modelling the Impact of Scandals: the case of the 2017 French Presidential
Election
Yassine Bouachrine and Carole Adam
(This is an ENSIMAG internship report originally written in June 2019 by
intern Yassine Bouachrine under the supervision of Carole Adam)
###### Abstract
This paper proposes an agent-based simulation of a presidential election,
inspired by the French 2017 presidential election. The simulation is based on
data extracted from polls, media coverage, and Twitter. The main contribution
is to consider the impact of scandals and media bashing on the result of the
election. In particular, it is shown that scandals can lead to higher
abstention at the election, as voters have no relevant candidate left to vote
for. The simulation is implemented in Unity 3D and is available to play
online. Keywords: agent-based simulation, computational social choice, voting
models
## 1 Introduction
During the 2017 French presidential election, the media had a very impactful
role in the shift of the opinion away from the election’s favorite François
Fillon. The seriousness of the accusations against the candidate led to Fillon
plummeting in the polls. We will try to model the impact of both conventional
and social media through scandal diffusion, in order to better understand the
dynamics underlying the voting process.
There are a variety of existing models for the voting process. However, most
of the models see voting as a discrete event, while it is actually a sample at
a given instant of a system in continuous evolution. This paper tries to shed
some light on improvements which could take into account the inherent dynamism
that comes with the interactions of the voters.
Another issue is that most computer voting models are adapted to the American
context. In France, there are multiple candidates participating in the first
round of the elections but in reality, few of them are actually considered by
the voters, despite being aligned with their ideals. This is due to the nature
of the voting process in France which takes place in two rounds. This leads
the voters to cast their vote strategically towards candidates who have a
chance of making it out of the first round.
The Voter Autrement111Voter Autrement : https://vote.imag.fr/ experiment
explored the effects of this strategic voting in 2017 by testing various
voting systems during the presidential election. The results showed that the
alternative voting methods yield vastly different results (in terms of who is
elected), especially for candidates such as J-L. Mélenchon and H. Hamon
(systematic improvement) or F. Fillon and M. Le Pen (systematic decline).
Other voting methods such as candidate ranking make the strategic vote
useless, since every candidate gets a chance to make it to the second turn.
Finally, the context for the 2017 presidential election is even more unique in
regard to the state of the French political scene. It is the first
presidential election since the split of the centre-right party (UMP), and it
is taking place amid the overall dissatisfaction of the population with the
French Socialist Party (PS, left wing). This will be further elaborated on in
the Voting models section as it is relevant to studying the impact of such a
context on the existing models.
Our goal here is to model the impact of scandals over the course of the
election and the change in the opinion of the voters. Last year, A. Soutif
[13] tried to model the election process using an agent-based model by feeding
the agents the results of the polls reported by the media. His goal in that
study was to model the impact of the polls on the votes and on the strategic
vote in particular, as polls give information on which candidates have chances
of making it out of the first round. Our approach aims to complement this work
by showing the additional impact of media through the diffusion of scandals
during the campaign.
The first part of the paper (Section 2) focuses on the data analysis upon
which the model is built. The second part of the paper (Section 3) addresses
voting models and our implementation of the suggested model.
## 2 Data analysis
Building a model requires the availability of sufficient and relevant data
about the phenomenon we are trying to model.
### 2.1 Comparing poll results and media trends
In our case, we took the aggregated poll results from various organisms [14]
shown in Figure 1, and interpolated linearly when we had missing data (mostly
B. Hamon in the early polls). We then compared the result of these polls with
the evolution of the presence of the candidates across traditional media and
Twitter.
Figure 1: Evolution of the polls during the elections (Source: Wikipedia [15])
Google Trends222https://trends.google.fr/trends/ is a Google website that
offers statistics on Google Search queries. We used it to evaluate the
searched queries associated with the the top five candidates in the News
section, the results are shown on Figure 2. An interesting point can be made
around how F. Fillon dominates the media presence in the early months of the
election, alongside B. Hamon which we can ignore since it is mostly due to the
French Socialist Party presidential primary.
Figure 2: Evolution of the search queries for the candidates
We can see that media coverage of this candidate spikes at multiple points in
time (Figure 3), which is likely due to the “Penelope gate”, which is a
scandal associated with the alleged fictitious employment of members of
Fillon’s family. Further, the media frenzy surrounding the “Penelope gate”
looks correlated with the evolution of the polls, as shown in Figure 3.
Figure 3: Evolution of the polls and search queries for Fillon
The takeaway from this comparison is that further media bashing does not seem
to impact polls as much. This can be attributed to the nature of the scandal
being a punctual event, and that people willing to cast away their vote
already did so with the initial media outlets. The biggest beneficiary of this
is definitely the closest candidate in the political spectrum: E. Macron.
Another interesting candidate is M. Le Pen who dropped significantly in the
polls (Figure 4) once the media picked up that she could actually finish first
at the first round.
Figure 4: Evolution of the polls and search queries for Le Pen
An hypothesis could be that there is actually a bidirectional relationship
between polls and media coverage, with each one affecting the other. For
example, in F. Fillon’s case it is the disclosure of the scandal by Le Canard
Enchaîné that led to him dropping significantly, whereas in M. Le Pen’s case,
it is the poll results that led to an increase in coverage, and then her poll
results dropped as a result of this increased coverage.
### 2.2 Processing Twitter data
The limits of the information we can leverage from Google Trends is that it
does not tell us about the nature or content of the coverage. We cannot know
for sure if the increasing number of articles are rather positive or negative
ones. We can deduce that a posteriori by observing the impact on the polls,
but it is what we are actually trying to model. Therefore, we used Twitter
data, and analyze the tweets during the primary round to try and visualize
opinion trends during the election.
#### 2.2.1 Data sets
Two datasets were used in our work:
* •
The first one is from Kaggle 333https://www.kaggle.com/jeanmidev/french-
presidential-election: Kaggle is an online exchange platform for
datascientist, users can publish datasets among other things. This Kaggle
dataset contains tweets sampled during the elections. It is very rich but the
data collection rate varies and some tweets appear to be truncated. More
details about this dataset can be found at the source.
* •
The other dataset is a courtesy of E. Duble, research engineer at LIG. It
contains an anonymized collection, sampling only geotagged tweets, that are
mentioning the top hashtags during the election. The importance of hashtags
has been shown for instance by the Politoscope project [10]
#### 2.2.2 Clustering
Building a vectorial representation of the tweets can be done in various ways.
M. Campr and K. Jezek [3] provided a performance evaluation of various methods
for paragraph vectorization. At first, we opted for Tweet2Vec [6], which
relies on character-based representations (as opposed to word representations
for Doc2Vec [8]) that perform better for content such as tweets. However,
Tweet2vec uses hashtag prediction to train the model, which is limited for our
use-case since we already have a restricted number of hashtags. It also takes
longer to train compared to word-based models.
We used Facebook fastText 444Facebook FastText: https://fasttext.cc/ to
generate embeddings, and enhanced them with their term frequency-inverse
document frequency (TF-IDF, [12]) in the corpus. We then computed the tweet
embeddings as the average of the word embeddings.
Performing Principal Component Analysis (PCA, [16]) on the 100-dimensional
tweet embeddings did not yield very good results as the embeddings are already
built to minimize colinearity.
## 3 Voting model
In this section, we provide an overview of existing models and their
limitations before presenting the model we built through the observation of
the data.
### 3.1 Existing models and their limitations
Doing a taxonomy of existing models is outside the scope of this paper, there
are good resources available for that [1, 9, 7]. We will focus on some
limitations of the existing models, starting with the psychological model. As
hinted to in the introduction, the context of the election makes partisan
identification hard to rely on: the schism of the centre-right party reshaped
the political scene entirely. Also, the overall dissatisfaction with F.
Hollande hurt the socialist party. Towards the end of his mandate, his
popularity rating was lower than Macron’s was during the Gilet Jaunes protests
[11]. On top of that, the appearance of new actors on the scene such as En
Marche further shook the scene. En Marche made retrospective voting irrelevant
as the party had never held responsibilities. The novelty Macron brings to the
table, and his ambition of uniting the political parties, gave him a
considerable advantage.
All of these circumstances made the elections very volatile. Even more
sophisticated models such as the funnel of causality have to be rethought.
Figure 5: Funnel of causality, source [4]
Media has to have a bigger role in this funnel, especially social media as it
has been shown to be a good indicator of standings [2], almost as good as
traditional polls. And that is, despite it being sensitive to social
engineering (cf. Cambridge Analytica’s impact on the American presidential
election).
### 3.2 Proposed model
Our model is an enhancement of proximity models, in order to take into account
the diffusion of scandals and the movement of neighboring agents.
#### 3.2.1 Simulation environment and initialization
The environment is a 100 X 100 units 2D plan. There are two types of active
agents: candidates and voters.
For the simulation, we define:
* •
The appeasement delta $\Delta\alpha\in[0,1]$, rate at which the repulsion of
the candidate diminishes
* •
The falloff rate for the potential of the scandals $\Delta\rho\in[0,1]$
* •
The maximum openness for the voters $\sigma_{max}\in[0,100]$, defines how far
a voter considers his surroundings
* •
The maximum tolerance for the voters $\theta_{max}\in[0,+\infty[$
#### 3.2.2 Agents and their attributes
For candidates $C$, we define:
* •
Position at time $t$ as $\psi_{t}\in[0,1]^{2}$ initialized manually
* •
Repulsion at time $t$ as $\gamma_{t}\in[0,1]$ with $\gamma_{0}=0$
* •
A list S of scandals with $S_{i}$ being the i-th one.
For voters $V$, we define:
* •
Position at time $t$ as $\psi_{t}\in[0,1]^{2}$ with
$\psi_{0}\sim\mathcal{U}^{2}(0,1)$ (distance to a candidate inversely
proportional to agreement with this candidate)
* •
Openness as $\sigma\in[0,\sigma_{max}]$ with $\sigma\sim
clamp_{[0,1]}(\mathcal{N}(0.5,0.2^{2}))\sigma_{max}$ (the radius is which a
voter considers agents around him)
* •
Charisma as $\kappa\in[0,1]$ with $\kappa\sim
clamp_{[0,1]}(\mathcal{N}(0.5,0.2^{2}))$ (the chance to influence others
around)
* •
Tolerance as $\theta\in[0,1]$ with $\theta\sim
clamp_{[0,1]}(\mathcal{N}(0.5,0.2^{2}))\theta_{max}$ (the threshold for
repulsion before dismissing a candidate completely)
* •
Conformity as $\eta\in[0,1]$ with $\eta\sim
clamp_{[0,1]}(\mathcal{N}(0.5,0.2^{2}))$
For scandals, we define:
* •
Potential at time $t$ as $\rho_{t}\in[0,1]$ with $\rho_{0}$ initialized
manually by the user (parameter in the simulation)
#### 3.2.3 Simulation update
At each time-step of the simulation, we update the entities. To simplify the
equations, we assume that the values are clamped to their domain.
For scandals, the potential decreases with time, at the falloff rate:
$\rho_{t+1}(x)=\rho_{t}(x)-\Delta\rho$ (1)
For candidates, the position is static, and the repulsion increases with each
scandal, and decreases with time at the pace set by the appeasement delta:
$\gamma_{t+1}(x)=\gamma_{t}(x)-\Delta\alpha+\sum_{y\in S(x)}\rho_{t+1}(y)$ (2)
For voters, the position evolves as their opinion about the different
candidates evolves based on their surroundings:
$\begin{split}\psi_{t+1}(x)&=\psi_{t}(x)\\\
&+\operatorname*{argmin}_{y}\\{||\psi_{t}(x)-\psi_{t}(y)||_{2}\mid y\in
C\land\gamma_{t+1}(y)<\theta(x)\\}\frac{1}{1+\gamma_{t+1}(y)}\\\
&+\eta\sum_{y\in
V\land||\psi_{t}(x)-\psi_{t}(y)||_{2}<\sigma(x)}\kappa(y)(\psi_{t}(x)-\psi_{t}(y))\end{split}$
(3)
When the simulation stops, each voter votes for the closest candidate still
considered in the openness radius around him. If there are none, the voter
withholds his vote (chooses abstention).
#### 3.2.4 Implementation details
The simulation is built in Unity 2018.3.5f1 555Unity 3D: https://unity.com/.
It allowed for faster prototyping and also supports a wide range of platforms
to run the simulation on. The simulation is available to play online at
http://lig-tdcge.imag.fr/votsim/ or to download as a WebGL export666WebGL
export of the simulator: https://ensiwiki.ensimag.fr/index.php?title=IRL_-
_Modélisation_de_la_dynamique_des_opinions_des_électeurs.
A first screen lets the user select the values of the global parameters of the
simulation (Figure 6): the number of voters and candidates, the appeasement
delta (rate at which the scandals decrease, which determines the duration of
their effect on opinions), and the maximum values for tolerance and openness
(individual values of all agents are then set randomly under this boundary).
Figure 6: Parameter selection before starting the simulation
On the next screen, the user can modify the simulation speed. Voters are
moving in the environment towards or away from the candidates. The user can
also trigger a scandal and choose its intensity and target candidate, by using
a button at the bottom of the window, in order to observe the influence on the
movements of the voters (see Figure 7). The intensity of the scandal then
decreases with time.
Figure 7: Screenshot of the simulation after triggering a scandal
#### 3.2.5 Discussion and results
Regarding the motivations behind the model, we wanted to enhance the existing
models with the observations made from our data analysis. First, regarding the
initialization, the justification behind the uniform distribution for the
voters’ positions is the unique context behind the 2017 election that we
detailed earlier, with many voters not knowing which parties to consider. A
fine-tuned Gaussian mixture model could also be explored.
Most of the reasoning behind the model is based on the reactions to the
Penelopegate, with some voters completely turning their backs on F. Fillon
(which we model by a scandal being above their tolerance threshold) and some
only showing hesitation (tolerance threshold not reached). A temporary dip
followed by a partial recovery in the polls supports the model: regardless of
the further media bashing around the event, voters have a threshold over which
further coverage has no effect.
Over the simulated scenarios, one of the most interesting observations is that
scandals tend to be tied with an increased abstention rate. In our model this
is represented by the voter moving too far from all candidates (due to
repulsion generated by scandals about their favourite candidates, or to
diverging opinions with the others), so that no valid candidate is still
present in the openness radius when the election comes; in that case the voter
prefers to choose abstention. Our model can therefore reproduce and explain
the 2017 presidential election’s high abstention rate in the first round, at
22.23% [5].
## 4 Conclusion
We have seen that media coverage of the campaign scandals can have a big
impact on the election results. The simulation showed that scandals can
totally shape the result of the election and that scandals profit to the
closer candidates on the political spectrum. The more interesting finding was
how scandals impact the abstention rate, which is in agreement with the
observations made in the context of the French 2017 presidential election and
the high abstention rate recorded.
There is still much to do to reach a unified model, a first step in that
direction would be enhancing the simulation with the results from A. Soutif’s
experiments regarding the impact of the polls [13]. We could then initialize
the model to match the French political scene at the beginning of the first
round and test if it corresponds to the observed election results. If the
model is validated, we could explore alternative scenarios for the election:
how different scandals could have led to different results and particularly
what would have happened if there were no scandals involving the pre-campaign
favorite F. Fillon.
## References
* [1] Rui Antunes. Theoretical models of voting behaviour. Exedra, 4(1):145–70, 2010.
* [2] David Anuta, Josh Churchin, and Jiebo Luo. Election bias: Comparing polls and twitter in the 2016 us election. arXiv preprint arXiv:1701.06232, 2017.
* [3] Michal Campr and Karel Ježek. Comparing semantic models for evaluating automatic document summarization. In International Conference on Text, Speech, and Dialogue, pages 252–260. Springer, 2015.
* [4] Russell J Dalton et al. Citizen politics in Western democracies: Public opinion and political parties in the United States, Great Britain, West Germany, and France. Chatham House Chatham, NJ, 1988.
* [5] Laurent de Boissieu. Participation et abstention aux élections. France politique, 2016.
* [6] Bhuwan Dhingra, Zhong Zhou, Dylan Fitzpatrick, Michael Muehl, and William W Cohen. Tweet2vec: Character-based distributed representations for social media. Technical report, Arxiv, 2016. https://arxiv.org/abs/1605.03481.
* [7] Marco Giugni. Theoretical models of voting behaviour. https://baripedia.org/wiki/Theoretical_models_of_voting_behaviour, 2011\.
* [8] Jey Han Lau and Timothy Baldwin. An empirical evaluation of doc2vec with practical insights into document embedding generation. arXiv preprint arXiv:1607.05368, 2016.
* [9] Michael S Lewis-Beck and Mary Stegmaier. Economic models of voting. In The Oxford handbook of political behavior. Oxford, 2007.
* [10] Maziyar Panahi, Noe Gaumont, and David Chavalarias. How political communities are using hashtags against their opponents in upcoming french presidential election: #presidentielle2017. Technical report, ISC-PIF, Politoscope, 2017. https://politoscope.org/2017/04/hashtags-matter-presidentielle2017/.
* [11] Opinion Publique. Popularité de François Hollande : Chiffres et analyses de la popularité de François Hollande et de son gouvernement. Opinion Publique: Sondages publiés et analyses sur les élections, la politique et les sujets de société, 2014. https://opinionpublique.wordpress.com/category/popularite-de-francois-hollande/.
* [12] Juan Ramos et al. Using tf-idf to determine word relevance in document queries. In Proceedings of the first instructional conference on machine learning, volume 242.1, pages 29–48, 2003.
* [13] Albin Soutif. Simulation multi-agent de systèmes de vote. Technical report, ENSIMAG. Projet d’Introduction à la Recherche en Laboratoire, encadré par Carole Adam et Sylvain Bouveret, 2017. https://ensiwiki.ensimag.fr/index.php?title=Albin_Soutif_:_Simulation_multi-agent_de_syst%C3%A8mes_de_vote.
* [14] Wikipedia. Opinion polling for the 2017 french presidential election. https://en.wikipedia.org/wiki/Opinion_polling_for_the_2017_French_presidential_election, 2017\.
* [15] Wikipedia. Évolution des intentions de vote au premier tour de l’élection. https://commons.wikimedia.org/wiki/File:Evolution_des_intentions_de_vote_à_l’élection_présidentielle_2017.png, 2017\.
* [16] Svante Wold, Kim Esbensen, and Paul Geladi. Principal component analysis. Chemometrics and intelligent laboratory systems, 2(1-3):37–52, 1987\.
|
A Study On Some Geometric and Physical Properties of
Hyper-Generalised Quasi-Einstein Spacetime
Kaushik Chattopadhyay1111This is the corresponding author., Arindam
Bhattacharyya2 and Dipankar Debnath3
1Department of Mathematics,
Jadavpur University, Kolkata-700032, India
<EMAIL_ADDRESS>
2Department of Mathematics,
Jadavpur University, Kolkata-700032, India
<EMAIL_ADDRESS>
3Department of Mathematics,
Bamanpukur High School(H.S), Nabadwip, India
<EMAIL_ADDRESS>
###### Abstract
In the present paper we discuss about a set of geometric and physical
properties of hyper-generalised quasi-Einstein spacetime. At the beginning we
discuss about pseudosymmetry over a hyper-generalised quasi-Einstein
spacetime. Here we discuss about $W_{2}$-Ricci pseudosymmetry, $Z$-Ricci
pseudosymmetry, Ricci pseudosymmetry and projective pseudosymmetry over a
hyper-generalised quasi-Einstein spacetime. Later on we take over Ricci
symmetric hyper-generalised quasi-Einstein spacetime and derive a set of
important geometric and physical theorems over it. Moving further we consider
some physical applications of the hyper-generalised quasi-Einstein spacetime.
Lastly we prove the existence of a hyper-generalised quasi-Einstein spacetime
by constructing a non-trivial example.
M.S.C.2010: 53C15, 53C25, 53C35.
Keywords: $W_{2}$-curvature tensor, $Z$ tensor, projective curvature tensor,
Riemannian curvature tensor, hyper generalised quasi-Einstein spacetime,
Einstein equation, heat flux, stress tensor.
## 1 Introduction
The General theory of Relativity is unarguably the most beautiful theory the
World of Physics has ever produced. It is the most powerful result of the
human intellect. This is an extremely important theory to study the nature of
this universe, cosmology and gravity. Three most important things that Modern
Scientists/ Mathematical Physicists can learn from special to general
relativity are as follows:
(i) The laws of Physics should be the same in every inertial reference frame,
i. e., the abandonment of the privileged states of inertial frame of
reference.
(ii) The acceptance of the dynamical role of the metric $g$, i.e., the study
of non-linear behaviour of nature.
and
(iii) The spacetime has to be considered as a class of semi-Riemannian
geometry.
The semi-Riemannian geometry has become more and more relevant and significant
in dealing with the nature of this universe with every passing day.
The theory of general relativity is mainly studied on a semi-Riemannian
manifold which sometimes is not an Einstein spacetime. Thus it was always
necessary to expand the concept of Einstein manifold to quasi-Einstein, then
generalised quasi-Einstein, mixed generalised quasi-Einstein and lastly to
hyper-generalised quasi-Einstein manifold. We demonstrate the introduction to
this procedure as follows:
An Einstein manifold is a Riemannian or pseudo-Riemannian manifold whose Ricci
tensor S of type $(0,2)$ is non-zero and proportional to the metric tensor.
Einstein manifolds form a natural subclass of various classes of Riemannian or
semi-Riemannian manifolds by a curvature condition imposed on their Ricci
tensor [4]. Also in Riemannian geometry as well as in general relativity
theory, the Einstein manifold plays a very important role.
M. C. Chaki and R. K. Maity had given the notion of quasi Einstein manifold
[18] in $2000$. A non flat $n$-dimensional Riemannian manifold $(M^{n},g)$,
$n(>2)$ is said to be a quasi-Einstein manifold if its nonzero Ricci tensor
$S$ of type $(0,2)$ satisfies the following condition
$S(X,Y)=\alpha g(X,Y)+\beta A(X)A(Y),$ (1.1)
where for all vector fields $X$,
$g(X,\xi_{1})=A(X),\leavevmode\nobreak\ g(\xi_{1},\xi_{1})=1.$ (1.2)
That is, $A$ being the associated $1$-form, $\xi_{1}$ is generally known as
the generator of the manifold. $\alpha$ and $\beta$ are associated nonzero
scalar functions. This manifold is denoted by $(QE)_{n}$. Clearly, for
$\beta=0$, this manifold reduces an Einstein manifold. We can note that
Robertson-Walker spacetimes are quasi-Einstein spacetimes. In the recent
papers [2], [19], the application of quasi-Einstein spacetime and generalised
quasi-Einstein spacetime in general relativity have been studied. Many more
works have been done in the spacetime of general relativity [3], [13], [14],
[21], [22], [23], [24], [25], [26].
Then M. C. Chaki initiated the notion of generalized quasi-Einstein manifold
[17] in 2001. A Riemannian manifold of dimension $n(>2)$ is said to be
generalized quasi Einstein manifold if its Ricci tensor $S$ of type $(0,2)$ is
not identically zero and satisfies the following condition
$S(X,Y)=\alpha g(X,Y)+\beta A(X)A(Y)+\gamma[A(X)B(Y)+A(Y)B(X)],$ (1.3)
where $\alpha$, $\beta$ and $\gamma$ are real valued, nonzero scalar functions
on $(M^{n},g)$, $A$ and $B$ are called two non zero $1$-forms such that
$g(X,\xi_{1})=A(X),g(X,\xi_{2})=B(X),g(\xi_{1},\xi_{2})=0,g(\xi_{1},\xi_{1})=1,g(\xi_{2},\xi_{2})=1.$
(1.4)
Here $\xi_{1}$ and $\xi_{2}$ are two unit vector fields which are orthogonal
to each other. $\alpha$, $\beta$ and $\gamma$ are called associated scalars,
$A$ and $B$ are called associated $1$-forms. $\xi_{1}$ and $\xi_{2}$ are two
generators of the manifold. This manifold is denoted by $(GQE)_{n}$. Clearly,
for $\gamma=0$, then it takes the form of a quasi-Einstein manifold and for
$\beta=\gamma=0$, it takes the form of an Einstein manifold.
The notion of hyper-generalized quasi-Einstein manifold has been introduced by
A. A. Shaikh, C. Özgür and A. Patra [1] in 2011. According to them, a
Riemannian manifold $(M^{n},g)$, $(n>2)$ is said to be a hyper-generalized
quasi-Einstein manifold if its Ricci tensor S of type $(0,2)$ is non-zero and
satisfies the following condition
$\displaystyle S(X,Y)=\alpha g(X,Y)+\beta A(X)A(Y)+\gamma[A(X)B(Y)+A(Y)B(X)]$
$\displaystyle\leavevmode\nobreak\ \leavevmode\nobreak\ \leavevmode\nobreak\
\leavevmode\nobreak\ \leavevmode\nobreak\ \leavevmode\nobreak\
\leavevmode\nobreak\ \leavevmode\nobreak\ \leavevmode\nobreak\
\leavevmode\nobreak\ \leavevmode\nobreak\ +\delta[A(X)D(Y)+A(Y)D(X)],$ (1.5)
for all
$X,Y,Z\in{\mathchoice{\raisebox{0.0pt}{$\displaystyle\chi$}}{\raisebox{0.0pt}{$\textstyle\chi$}}{\raisebox{0.0pt}{$\scriptstyle\chi$}}{\raisebox{0.0pt}{$\scriptscriptstyle\chi$}}}(M)$.
Here $\alpha$, $\beta$, $\gamma$, $\delta$ are non-zero scalar functions on
$(M^{n},g)$. $A$, $B$, $D$ are non-zero $1$-forms such that
$g(X,\xi_{1})=A(X),\leavevmode\nobreak\ g(X,\xi_{2})=B(X),\leavevmode\nobreak\
g(X,\xi_{3})=D(X),$ (1.6)
where $\xi_{1}$, $\xi_{2}$, $\xi_{3}$ are mutually orthogonal unit vector
fields. i. e.,
$g(\xi_{1},\xi_{2})=g(\xi_{2},\xi_{3})=g(\xi_{1},\xi_{3})=0;\leavevmode\nobreak\
g(\xi_{1},\xi_{1})=g(\xi_{2},\xi_{2})=g(\xi_{3},\xi_{3})=1.$ (1.7)
An $n$-dimensional hyper-generalized quasi-Einstein manifold is generally
denoted as $(HGQE)_{n}$. Shaikh, Özgür and Patra in [1] studied on hyper-
generalized quasi-Einstein manifolds with some geometric properties of it.
Güler and Demirbaǧ [20] dealt with some Ricci conditions on hyper-generalized
quasi-Einstein manifolds, D. Debnath [9] proved few theorems about the
properties of the hyper-generalized quasi-Einstein manifolds.
The concept of perfect fluid spacetime arose while discussing the structure of
this universe. Perfect fluids are often used in the general relativity to
model the idealised distribution of matter, such as the interior of a star or
isotropic pressure. In general relativity the matter content of the spacetime
is described by the energy-momentum tensor. The matter content is assumed to
be a fluid having density and pressure and possessing dynamical and
kinematical quantities like velocity, acceleration, vorticity, shear and
expansion. The energy-momentum tensor $T$ of a perfect fluid spacetime is
given by the following equation
$\cite[cite]{[\@@bibref{}{jmh12}{}{}]},\cite[cite]{[\@@bibref{}{jmh13}{}{}]}$
$T(X,Y)=(\sigma+p)A(X)A(Y)+pg(X,Y).$ (1.8)
Here $g(X,\xi_{1})=A(X),A(\xi_{1})=-1$, for any $X,Y$. $p$ and $\sigma$ are
called the isotropic pressure and the energy density respectively. $\xi_{1}$
being the unit timelike velocity vector field.
The Einstein field equation [6] is given by
$S(X,Y)-\frac{r}{2}g(X,Y)+\lambda g(X,Y)=kT(X,Y);\leavevmode\nobreak\ \forall
X,Y\in TM,$ (1.9)
here $r$ being the scalar curvature, $S$ being the Ricci tensor of type
$(0,2)$. $k$ and $\lambda$ are the gravitational constant and cosmological
constant respectively. From Einstein’s field equation it follows that energy
momentum tensor is a symmetric $(0,2)$ type tensor of divergence zero.
In the year of 2012, Mantica and Molinari[7] defined a new generalized
symmetric $(0,2)$ tensor called Z tensor. According to them it is given as,
$Z(X,Y)=S(X,Y)+\phi g(X,Y),$ (1.10)
where $\phi$ is an arbitrary scalar function. A set of properties of $Z$
tensor have been studied in the papers [7] and [8].
Like the $Z$ curvature tensor projective curvature tensor also plays a very
significant role in studying different properties of semi-Riemannian geometry.
Let $M^{n}(n\geq 3)$ be a semi-Riemannian manifold. The projective curvature
tensor[15] is defined by,
$P(X,Y)Z=R(X,Y)Z-\frac{1}{n-1}\\{S(Y,Z)X-S(X,Z)Y\\}.$ (1.11)
Hyper-generalized quasi-Einstein manifolds is considered as the base space of
general relativistic viscous fluid spacetime, which inspired us to take a look
on some geometric properties of the $(HGQE)_{4}$ spacetime under certain
conditions which we study on sections 2, 3, 4 and 5. Then we discuss about
Ricci symmetric hyper-generalized quasi-Einstein spacetimes in section 6. In
section 7 we derive a result about the energy-momentum tensor on hyper-
generalized quasi-Einstein spacetime of constant curvature with cyclic
parallel Ricci tensor. Also the spacetime has wide applications in general
relativistic viscous fluid spacetime admitting heat flux and stress, which
motivated us to discuss about some physical applications of an $(HGQE)_{4}$
spacetime in section 8. Finally in section 9 we construct a non-trivial
example of an $(HGQE)_{4}$ spacetime to prove the existence of such spacetime.
## 2 $W_{2}$-Ricci pseudosymmetric $(HGQE)_{4}$ spacetime
The $W_{2}$-curvature tensor was introduced by G. P. Pokhariyal and R. S.
Mishra [12] in $1970$ and they studied some properties of it. A
$W_{2}$-curvature tensor on a manifold $(M^{n},g)$, $n(>3)$ is defined by
$W_{2}(X,Y)Z=R(X,Y)Z-\frac{1}{n-1}[g(Y,Z)QX-g(X,Z)QY].$ (2.1)
Here $r$ being the curvature tensor and $Q$ is the Ricci operator defined by
$g(QX,Y)=S(X,Y),\forall\leavevmode\nobreak\ X,Y$.
For an $(HGQE)_{4}$ quasi-Einstein spacetime the $(0,4)$ $W_{2}$ curvature
tensor takes the following form,
$W_{2}(X,Y,Z,W)=R(X,Y,Z,W)-\frac{1}{3}[g(Y,Z)S(X,W)-g(X,Z)S(Y,W)].$ (2.2)
Firstly, we take a hyper generalized quasi-Einstein spacetime satisfying the
condition $W_{2}.S=F_{S}Q(g,S)$. Here $F_{S}$ being a certain function on the
set $U_{S}=\\{x\in M:S\neq\frac{r}{n}g$ at $x\\}$ and $Q(g,S)$ being the
Tachibana tensor working on the metric tensor and the Ricci tensor. This
spacetime is called $W_{2}$-Ricci pseudosymmetric $(HGQE)_{4}$. Now for all
$X,Y,Z\in{\mathchoice{\raisebox{0.0pt}{$\displaystyle\chi$}}{\raisebox{0.0pt}{$\textstyle\chi$}}{\raisebox{0.0pt}{$\scriptstyle\chi$}}{\raisebox{0.0pt}{$\scriptscriptstyle\chi$}}}(M^{4})$;
$\displaystyle S(W_{2}(X,Y)Z,W)+S(Z,W_{2}(X,Y)W)$
$\displaystyle=F_{S}[g(Y,Z)S(X,W)-g(X,Z)S(Y,W)$
$\displaystyle+g(Y,W)S(Z,X)-g(X,W)S(Y,Z)].$ (2.3)
From the equation (1) from the equation (2) we get,
$\displaystyle\alpha g(W_{2}(X,Y)Z,W)+\beta A(W_{2}(X,Y)Z)A(W)$
$\displaystyle+\gamma[A(W_{2}(X,Y)Z)B(W)+B(W_{2}(X,Y)Z)A(W)]$
$\displaystyle+\delta[A(W_{2}(X,Y)Z)D(W)+D(W_{2}(X,Y)Z)A(W)]$
$\displaystyle+\alpha g(W_{2}(X,Y)W,Z)+\beta A(W_{2}(X,Y)W)A(Z)$
$\displaystyle+\gamma[A(W_{2}(X,Y)W)B(Z)+B(W_{2}(X,Y)W)A(Z)]$
$\displaystyle+\delta[A(W_{2}(X,Y)W)D(Z)+D(W_{2}(X,Y)W)A(Z)]$
$\displaystyle=F_{S}[g(Y,Z)S(X,W)-g(X,Z)S(Y,W)$
$\displaystyle\leavevmode\nobreak\ \leavevmode\nobreak\
+g(Y,W)S(Z,X)-g(X,W)S(Y,Z)].$ (2.4)
Contracting the equation over $X$ and $W$ and putting $Z=\xi_{1}$ we get,
$\displaystyle\alpha[\frac{4}{3}\\{(\alpha-\beta)A(Y)-\gamma B(Y)-\delta
D(Y)\\}-\frac{r}{3}A(Y)]-\frac{1}{3}[-\gamma B(Y)-\delta D(Y)]$ (2.5)
$\displaystyle+\frac{\gamma^{2}}{3}A(Y)+\gamma
R(\xi_{1},Y,\xi_{1},\xi_{2})-\frac{\alpha\gamma}{3}B(Y)+\frac{\delta^{2}}{3}A(Y)$
$\displaystyle+\delta R(\xi_{1},Y,\xi_{1},\xi_{3})-\frac{\delta\alpha}{3}D(Y)$
$\displaystyle=F_{S}[rA(Y)-4\\{(\alpha-\beta)A(Y)-\gamma B(Y)-\delta
D(Y)\\}].$
Putting $X=\xi_{1},Z=\xi_{2},W=\xi_{3}$ in the equation (2) we get,
$\displaystyle\delta R(\xi_{1},Y,\xi_{2},\xi_{1})+\gamma
R(\xi_{1},Y,\xi_{3},\xi_{1})-\frac{\alpha}{3}B(Y)(-\delta)$ (2.6)
$\displaystyle-\frac{\alpha}{3}D(Y)(-\gamma)+\frac{\delta}{3}B(Y)(\alpha-\beta)+\frac{\gamma}{3}D(Y)(\alpha-\beta)$
$\displaystyle=F_{S}[B(Y)(-\delta)+D(Y)(-\gamma)].$
combining the equations (2.5) and (2.6) we get,
$\displaystyle\alpha[\frac{4}{3}\\{(\alpha-\beta)A(Y)-\gamma B(Y)-\delta
D(Y)\\}-\frac{r}{3}A(Y)]-\frac{1}{3}[-\gamma B(Y)-\delta D(Y)]$ (2.7)
$\displaystyle+\frac{\gamma^{2}}{3}A(Y)-\frac{\alpha\gamma}{3}B(Y)+\frac{\delta^{2}}{3}A(Y)$
$\displaystyle-\frac{\delta\alpha}{3}D(Y)-F_{S}[B(Y)(-\delta)+D(Y)(-\gamma)]-\frac{\alpha}{3}B(Y)(-\delta)$
$\displaystyle-\frac{\alpha}{3}D(Y)(-\gamma)+\frac{\delta}{3}B(Y)(\alpha-\beta)+\frac{\gamma}{3}D(Y)(\alpha-\beta)$
$\displaystyle=F_{S}[rA(Y)-4\\{(\alpha-\beta)A(Y)-\gamma B(Y)-\delta
D(Y)\\}].$
setting $\gamma=\delta$ yields,
$\displaystyle\alpha[\frac{4}{3}\\{(\alpha-\beta)A(Y)-\gamma B(Y)-\gamma
D(Y)\\}-\frac{r}{3}A(Y)]-\frac{1}{3}[-\gamma B(Y)-\gamma D(Y)]$ (2.8)
$\displaystyle+\frac{\gamma^{2}}{3}A(Y)-\frac{\alpha\gamma}{3}B(Y)+\frac{\gamma^{2}}{3}A(Y)$
$\displaystyle-\frac{\gamma\alpha}{3}D(Y)-F_{S}[B(Y)(-\gamma)+D(Y)(-\gamma)]-\frac{\alpha}{3}B(Y)(-\gamma)$
$\displaystyle-\frac{\alpha}{3}D(Y)(-\gamma)+\frac{\gamma}{3}B(Y)(\alpha-\beta)+\frac{\gamma}{3}D(Y)(\alpha-\beta)$
$\displaystyle=F_{S}[rA(Y)-4\\{(\alpha-\beta)A(Y)-\gamma B(Y)-\gamma
D(Y)\\}].$
Putting $Y=\xi_{1}$ in the above equation we get,
$-3\beta F_{S}=\alpha\beta-\frac{2\gamma^{2}}{3}.$ (2.9)
Again putting $Y=\xi_{2}$ in the equation (2.8) we get,
$\gamma(\alpha+3F_{S})=0,$ (2.10)
which gives either $\gamma=0$ or $F_{S}=\frac{-\alpha}{3}$. Now, if $\gamma=0$
then since $\gamma=\delta$ thus $\gamma=\delta=0$. Hence the spacetime becomes
a quasi-Einstein spacetime. On the other hand if $F_{S}=\frac{-\alpha}{3}$
then from the equation (2.9) we get $\gamma=0$. Again since $\gamma=\delta$
thus $\gamma=\delta=0$. Thus in both the cases the manifold is reduced to a
quasi-Einstein spacetime. Hence we conclude the following theorem as:
Theorem 2.1: A $W_{2}$-Ricci pseudosymmetric $(HGQE)_{4}$ spacetime is a
$(QE)_{4}$ spacetime if $\gamma=\delta$.
## 3 Z-Ricci pseudosymmetric $(HGQE)_{4}$ spacetime
A semi-Riemannian manifold $(M^{n},g),n\geq 3$, is called Z-Ricci
pseudosymmetric iff the following relation holds:
$Z\cdot K=F_{K}P(g,K),$ (3.1)
on the set $U_{K}=\\{x\in M:P(g,K)\neq 0\leavevmode\nobreak\
at\leavevmode\nobreak\ x\\}$, where $K$ is the Ricci operator defined by
$S(X,Y)=g(KX,Y)$ and $F_{K}$ is a smooth function on $U_{K}$. The operator $P$
is defined by the following way:
$P(g,K)(W;X,Y)=K((X\wedge_{g}Y)W)$ (3.2)
for all vector fields $X,Y,W$.
Now if the spacetime is a Z-Ricci pseudosymmetric then from the equation (3.1)
we get,
$(Z(X,Y)\cdot K)W=F_{K}P(g,K)(W;X,Y),$ (3.3)
which takes the following form,
$\displaystyle Z(Y,KW)X-Z(X,KW)Y-Z(Y,W)KX-Z(X,W)KY$
$\displaystyle=F_{K}\\{g(Y,W)KX-g(X,W)KY\\}.$ (3.4)
With the help of the equation $(\ref{e1.5})$ we demonstrate the Ricci operator
by the following equation:
$KX=\alpha X+\beta
A(X)\xi_{1}+\gamma[A(X)\xi_{2}+B(X)\xi_{1}]+\delta[A(X)\xi_{3}+D(X)\xi_{1}].$
(3.5)
Applying the equation (3.5) in equation (3) we get,
$\displaystyle Z(Y,\alpha W)X+\beta A(W)Z(Y,\xi_{1})X+\gamma
A(W)Z(Y,\xi_{2})X$ (3.6) $\displaystyle+\gamma B(W)Z(Y,\xi_{1})X+\delta
A(W)Z(Y,\xi_{3})X+\delta D(W)Z(Y,\xi_{1})X$ $\displaystyle-[Z(X,\alpha
W)Y+\beta A(W)Z(X,\xi_{1})Y+\gamma A(W)Z(X,\xi_{2})Y$ $\displaystyle+\gamma
B(W)Z(X,\xi_{1})Y+\delta A(W)Z(X,\xi_{3})Y+\delta D(W)Z(X,\xi_{1})Y]$
$\displaystyle=\\{F_{K}g(Y,W)+Z(Y,W)\\}KX-\\{F_{K}g(X,W)-Z(X,W)\\}KY.$
With the help of the equation (1.10) this further implies,
$\displaystyle\alpha Z(Y,W)X+\beta A(W)\\{(\alpha-\beta+\phi)A(Y)-\gamma
B(Y)-\delta D(Y)\\}X$ (3.7) $\displaystyle+\gamma
A(W)\\{(\alpha+\phi)B(Y)+\gamma A(Y)\\}X$ $\displaystyle+\gamma
B(W)\\{(\alpha-\beta+\phi)A(Y)-\gamma B(Y)-\delta D(Y)\\}X$
$\displaystyle+\delta A(W)\\{(\alpha+\phi)D(Y)+\gamma A(Y)\\}X$
$\displaystyle+\delta D(W)\\{(\alpha-\beta+\phi)A(Y)-\gamma B(Y)-\delta
D(Y)\\}X$ $\displaystyle-\alpha Z(X,W)Y-\beta
A(W)\\{(\alpha-\beta+\phi)A(X)-\gamma B(X)-\delta D(X)\\}Y$
$\displaystyle-\gamma A(W)\\{(\alpha+\phi)B(X)+\gamma A(X)\\}Y$
$\displaystyle-\gamma B(W)\\{(\alpha-\beta+\phi)A(X)-\gamma B(X)-\delta
D(X)\\}Y$ $\displaystyle-\delta A(W)\\{(\alpha+\phi)D(X)+\gamma A(X)\\}Y$
$\displaystyle-\delta D(W)\\{(\alpha-\beta+\phi)A(X)-\gamma B(X)-\delta
D(X)\\}Y$ $\displaystyle=\\{F_{K}g(Y,W)+S(Y,W)+\phi g(Y,W)\\}KX$
$\displaystyle-\\{F_{K}g(X,W)-S(X,W)-\phi g(X,W)\\}KY.$
Putting $X=\xi_{1},Y=\xi_{2}$ equation (3.7) yields,
$\displaystyle\alpha\\{(\alpha+\phi)B(W)+\gamma A(W)\\}\xi_{1}+\beta
A(W)\\{-\gamma\\}\xi_{1}+\gamma A(W)\\{\alpha+\phi\\}\xi_{1}$ (3.8)
$\displaystyle+\gamma B(W)\\{-\gamma\\}\xi_{1}+\delta
D(W)\\{-\gamma\\}\xi_{1}-\alpha\\{(\alpha-\beta+\phi)A(W)-\gamma B(W)-\delta
D(W)\\}\xi_{2}$ $\displaystyle+\beta
A(W)\\{\alpha-\beta+\phi)\\}\xi_{2}+\gamma^{2}A(W)\xi_{2}$
$\displaystyle+\gamma
B(W)\\{\alpha-\beta+\phi)\\}\xi_{2}+\delta^{2}A(W)\xi{2}+\delta
D(W)\\{\alpha-\beta+\phi)\\}\xi_{2}$ $\displaystyle=\\{F_{K}B(W)+\phi
B(W)+\alpha B(W)+\gamma A(W)\\}K\xi_{1}$ $\displaystyle-\\{F_{K}A(W)-\phi
A(W)-(\alpha-\beta)A(W)+\gamma B(W)+\delta D(W)\\}K\xi_{2}.$
Taking inner product with $\xi_{1}$ to both the sides of the equation (3.8) we
get,
$\displaystyle
A(W)\\{\alpha\gamma+\phi\gamma\\}+B(W)\\{-\gamma^{2}-F_{K}(\alpha-\beta)+\phi\beta+\alpha\beta\\}-D(W)\\{\delta\gamma\\}=0.$
(3.9)
Putting $W=\xi_{3}$ in equation (3.9) we get,
$-\gamma\delta=0.$ (3.10)
That means at least one of $\gamma$ or $\delta$ must be zero. Which means the
manifold is reduced to a generalized quasi-Einstein spacetime. This allows us
to derive the next theorem as:
Theorem 3.1: A Z-Ricci pseudosymmetric $(HGQE)_{4}$ spacetime is a $(GQE)_{4}$
spacetime.
## 4 Ricci pseudosymmetric $(HGQE)_{4}$ spacetime
A semi-Riemannian manifold $M^{n}(n\geq 3)$ is called Ricci-pseudosymmetric if
the following relation
$(R(X,Y)\cdot S)(Z,W)=F_{S}Q(g,S)$ (4.1)
holds on $U_{S}=\\{x\in M:S\neq\frac{r}{n}\leavevmode\nobreak\
at\leavevmode\nobreak\ x\\}$ and $L_{S}$ is a function on $U_{S}$. From the
equation (4.1) we get,
$\displaystyle\leavevmode\nobreak\ \leavevmode\nobreak\
S(R(X,Y)Z,W)+S(Z,R(X,Y)W)$ $\displaystyle=F_{S}[g(Y,Z)S(X,W)-g(X,Z)S(Y,W)$
$\displaystyle+g(Y,W)S(Z,X)-g(X,W)S(Y,Z)].$ (4.2)
Using the equation (1) we demonstrate the equation (4) as follows:
$\displaystyle\leavevmode\nobreak\ \leavevmode\nobreak\
\beta[A(R(X,Y)Z)A(W)+A(Z)A(R(X,Y)W)]$
$\displaystyle+\gamma[A(R(X,Y)Z)B(W)+A(W)B(R(X,Y)Z)+A(Z)B(R(X,Y)W)$
$\displaystyle+A(R(X,Y)W)B(Z)]+\delta[A(R(X,Y)Z)D(W)+A(W)D(R(X,Y)Z)$
$\displaystyle+A(Z)D(R(X,Y)W)+A(R(X,Y)W)D(Z)]$
$\displaystyle=F_{S}[\beta\\{g(Y,Z)A(X)A(W)-g(X,Z)A(Y)A(W)+g(Y,Z)A(Z)A(X)$
$\displaystyle-g(X,W)A(Y)A(Z)\\}+\gamma\\{g(Y,Z)[A(X)B(W)+A(W)B(X)]$
$\displaystyle-g(X,Z)[A(Y)B(W)+A(W)B(Y)]+g(Y,W)[A(X)B(Z)+A(Z)B(X)]$
$\displaystyle-g(X,W)[A(Y)B(Z)+A(Z)B(Y)]\\}+\delta\\{g(Y,Z)[A(X)D(W)+A(W)D(X)]$
$\displaystyle-g(X,Z)[A(Y)D(W)+A(W)D(Y)]+g(Y,Z)[A(X)D(Z)+A(Z)D(X)]$
$\displaystyle-g(X,W)[A(Y)D(Z)+A(Z)D(Y)]\\}].$ (4.3)
Contracting equation (4) over X and W we get,
$\displaystyle\leavevmode\nobreak\ \leavevmode\nobreak\
\beta[A(R(\xi_{1},Y)Z)-A(Z)S(Y,\xi_{1})]+\gamma[A(R(\xi_{2},Y)Z)$
$\displaystyle+B(R(\xi_{1},Y)Z)-A(Z)S(Y,\xi_{2})-B(Z)S(Y,\xi_{1})]$
$\displaystyle+\delta[A(R(X,Y)Z)D(W)+D(R(\xi_{1},Y)Z)$
$\displaystyle-A(Z)S(Y,\xi_{3})-S(Y,\xi_{1})D(Z)]$
$\displaystyle=F_{S}\\{\beta[-g(Y,Z)-4A(Y)A(Z)]-4\gamma[A(Y)B(Z)$
$\displaystyle A(Z)B(Y)]-4\delta[A(Y)D(Z)+A(Z)D(Y)]\\}.$ (4.4)
Setting $Z=\xi_{1}$ in equation (4) we get,
$\displaystyle\leavevmode\nobreak\ \leavevmode\nobreak\
\beta(Y,\xi_{1})+\gamma[R(\xi_{1},Y,\xi_{1},\xi_{2})+S(Y,\xi_{2})]$
$\displaystyle+\delta[R(\xi_{1},Y,\xi_{1},\xi_{3})+S(Y,\xi_{3})]$
$\displaystyle=F_{S}[3\beta A(Y)+4\gamma B(Y)+4\delta D(Y)].$ (4.5)
Now, putting $Z=\xi_{2},W=\xi_{3}$ in the equation (4) we have,
$\displaystyle-\gamma R(\xi_{1},Y,\xi_{1},\xi_{3})-\delta
R(\xi_{1},Y,\xi_{1},\xi_{2})$ (4.6)
$\displaystyle=F_{S}\\{\gamma[D(Y)A(X)-D(X)A(Y)]+\delta[A(X)B(Y)-A(Y)B(X)]\\}.$
If $\gamma=\delta$ then from the equations (4) and (4.6) we conclude,
$\displaystyle[\beta S(Y,\xi_{1})+\gamma S(Y,\xi_{2})+\gamma S(Y,\xi_{3})]$
$\displaystyle-F_{S}[3\beta A(Y)+4\gamma B(Y)+4\gamma D(Y)]$
$\displaystyle=F_{S}\\{\gamma[D(Y)A(X)-D(X)A(Y)]+\gamma[A(X)B(Y)-A(Y)B(X)]\\}.$
(4.7)
Putting $X=\xi_{1}$ we get,
$\displaystyle
A(Y)[\alpha\beta-\beta^{2}+2\gamma^{2}]+B(Y)[-\beta\gamma+\gamma\alpha]+D(Y)[-\beta\gamma+\gamma\alpha]$
$\displaystyle-F_{S}[3\beta A(Y)+4\gamma B(Y)+4\gamma
D(Y)]=-F_{S}\gamma[D(Y)+B(Y)].$ (4.8)
Putting $Y=\xi_{2}$ in equation (4) we get,
$\gamma[(\alpha-\beta)-3F_{S}]=0.$ (4.9)
Again putting $Y=\xi_{1}$ in equation (4) we get,
$F_{S}=\frac{\alpha\beta-\beta^{2}+2\gamma^{2}}{3\beta}.$ (4.10)
From the equation (4.9) we have either $\gamma=0$ or
$F_{S}=\frac{\alpha-\beta}{3}$. Now, if $\gamma=0$ then since $\gamma=\delta$,
thus $\gamma=\delta=0$, implying the manifold reduces to a quasi-Einstein
spacetime. Again if $F_{S}=\frac{\alpha-\beta}{3}$ then from the equation
(4.10) we get $\gamma=0$ and hence $\gamma=\delta=0$, which again implies the
manifold is reduced to a quasi-Einstein spacetime. This allows us to deduce
the following theorem:
Theorem 4.1: A Ricci pseudosymmetric $(HGQE)_{4}$ spacetime is a $(QE)_{4}$
spacetime if $\gamma=\delta$.
## 5 Projectively pseudosymmetric $(HGQE)_{4}$ spacetime
From the equation $(\ref{e1.11})$ we see that for an $(HGQE)_{4}$ quasi-
Einstein spacetime the $(0,4)$ projective curvature tensor takes the following
form,
$P(X,Y,Z,W)=R(X,Y,Z,W)-\frac{1}{3}[S(Y,Z)g(X,W)-S(X,Z)g(Y,W)].$ (5.1)
A semi-Riemannian manifold $M^{n}(n\geq 3)$ is called projectively
pseudosymmetric if the following relation
$(P(X,Y)\cdot S)(Z,W)=F_{S}Q(g,S)$ (5.2)
holds on $U_{S}=\\{x\in M:S\neq\frac{r}{n}\leavevmode\nobreak\
at\leavevmode\nobreak\ x\\}$ and $L_{S}$ is a function on $U_{S}$. From the
equation (5.2) we get,
$\displaystyle\leavevmode\nobreak\ \leavevmode\nobreak\
S(P(X,Y)Z,W)+S(Z,P(X,Y)W)$ $\displaystyle=F_{S}[g(Y,Z)S(X,W)-g(X,Z)S(Y,W)$
$\displaystyle+g(Y,W)S(Z,X)-g(X,W)S(Y,Z)].$ (5.3)
From the equation (1) from the equation (5) we get,
$\displaystyle\alpha g(P(X,Y)Z,W)+\beta A(P(X,Y)Z)A(W)$
$\displaystyle+\gamma[A(P(X,Y)Z)B(W)+B(P(X,Y)Z)A(W)]$
$\displaystyle+\delta[A(P(X,Y)Z)D(W)+D(P(X,Y)Z)A(W)]$ $\displaystyle+\alpha
g(P(X,Y)W,Z)+\beta A(P(X,Y)W)A(Z)$
$\displaystyle+\gamma[A(P(X,Y)W)B(Z)+B(P(X,Y)W)A(Z)]$
$\displaystyle+\delta[A(P(X,Y)W)D(Z)+D(P(X,Y)W)A(Z)]$
$\displaystyle=F_{S}[g(Y,Z)S(X,W)-g(X,Z)S(Y,W)$
$\displaystyle\leavevmode\nobreak\ \leavevmode\nobreak\
+g(Y,W)S(Z,X)-g(X,W)S(Y,Z)].$ (5.4)
Putting $X=\xi_{1},Z=\xi_{2},W=\xi_{3}$ in the equation (5) we get,
$\displaystyle-\delta R(\xi_{1},Y,\xi_{2},\xi_{1})-\gamma
R(\xi_{1},Y,\xi_{3},\xi_{1})$
$\displaystyle=-\frac{\alpha\gamma}{3}D(Y)-\frac{\alpha\delta}{3}B(Y)+\frac{\gamma}{3}(\alpha
D(Y)+\delta A(Y))$ $\displaystyle+\frac{\delta}{3}(\alpha B(Y)+\gamma
A(Y))-\frac{\gamma^{2}}{3}A(Y)-\frac{\delta^{2}}{3}A(Y)$
$\displaystyle+F_{S}\\{\gamma D(Y)+\delta B(Y)\\}.$ (5.5)
Contracting equation (5) over $X,W$ and putting $Z=\xi_{1}$ we get,
$\displaystyle-\frac{\beta\gamma}{3}B(Y)-\frac{\beta\delta}{3}D(Y)-\frac{\gamma^{2}}{3}A(Y)$
$\displaystyle+\frac{4\delta}{3}[\alpha D(Y)+\delta
A(Y)]+\gamma(\beta-\alpha)B(Y)-\frac{\delta^{2}}{3}A(Y)$
$\displaystyle-\frac{\delta r}{3}D(Y)+\delta(\beta-\alpha)D(Y)$
$\displaystyle-\frac{4\alpha}{3}[(\alpha-\beta)A(Y)-\gamma B(Y)-\delta
D(Y)]+\frac{\alpha r}{3}A(Y)$
$\displaystyle+\frac{4\beta}{3}[(\alpha-\beta)A(Y)-\gamma B(Y)-\delta
D(Y)]-\frac{\beta r}{3}A(Y)$ $\displaystyle+\frac{4\gamma}{3}[\alpha
B(Y)+\gamma A(Y)]-\frac{\gamma r}{3}B(Y)$ $\displaystyle-
F_{S}\\{rA(Y)-4[(\alpha-\beta)A(Y)-\gamma B(Y)-\delta D(Y)]\\}$
$\displaystyle=-\gamma R(\xi_{1},Y,\xi_{1},\xi_{2})-\delta
R(\xi_{1},Y,\xi_{1},\xi_{3}).$ (5.6)
If $\gamma=\delta$ then rom the equations $(\ref{e5.5}),(\ref{e5.6})$, by
using the property of $R$ we get,
$\displaystyle-\frac{\beta\gamma}{3}B(Y)-\frac{\beta\gamma}{3}D(Y)-\frac{\gamma^{2}}{3}A(Y)$
$\displaystyle+\frac{4\gamma}{3}[\alpha D(Y)+\gamma
A(Y)]+\gamma(\beta-\alpha)B(Y)-\frac{\gamma^{2}}{3}A(Y)$
$\displaystyle-\frac{\gamma r}{3}D(Y)+\gamma(\beta-\alpha)D(Y)$
$\displaystyle-\frac{4\alpha}{3}[(\alpha-\beta)A(Y)-\gamma B(Y)-\gamma
D(Y)]+\frac{\alpha r}{3}A(Y)$
$\displaystyle+\frac{4\beta}{3}[(\alpha-\beta)A(Y)-\gamma B(Y)-\gamma
D(Y)]-\frac{\beta r}{3}A(Y)$ $\displaystyle+\frac{4\gamma}{3}[\alpha
B(Y)+\gamma A(Y)]-\frac{\gamma r}{3}B(Y)$ $\displaystyle-
F_{S}\\{rA(Y)-4[(\alpha-\beta)A(Y)-\gamma B(Y)-\gamma D(Y)]\\}$
$\displaystyle=-\\{-\frac{\alpha\gamma}{3}D(Y)-\frac{\alpha\gamma}{3}B(Y)+\frac{\gamma}{3}(\alpha
D(Y)+\gamma A(Y))$ $\displaystyle+\frac{\gamma}{3}(\alpha B(Y)+\gamma
A(Y))-\frac{\gamma^{2}}{3}A(Y)-\frac{\gamma^{2}}{3}A(Y)$
$\displaystyle+F_{S}\\{\gamma D(Y)+\gamma B(Y)\\}\\}.$ (5.7)
Putting $Y=\xi_{2}$ in equation (5) we get,
$\frac{\gamma}{3}(-\beta+\alpha+3F_{S})=0.$ (5.8)
Again putting $Y=\xi_{1}$ in equation (5) we get,
$3\beta F_{S}=\frac{2\gamma^{2}}{3}+\alpha\beta-\beta^{2}.$ (5.9)
From the equation (5.8) we have either $\gamma=0$ or
$F_{S}=\frac{\alpha-\beta}{3}$. Now, if $\gamma=0$ then since $\gamma=\delta$,
thus $\gamma=\delta=0$, implying the manifold reduces to a quasi-Einstein
spacetime. Again if $F_{S}=\frac{\alpha-\beta}{3}$ then from the equation
(5.9) we get $\gamma=0$ and hence $\gamma=\delta=0$, which again implies the
manifold is reduced to a quasi-Einstein spacetime. This allows us to deduce
the following theorem:
Theorem 5.1: A projectively pseudosymmetric $(HGQE)_{4}$ spacetime is a
$(QE)_{4}$ spacetime if $\gamma=\delta$.
## 6 Ricci symmetric $(HGQE)_{4}$ spacetime
A semi-Riemannian manifold $M^{n}(n\geq 3)$ is called Ricci-symmetric if
$\nabla S=0$, where $S$ is the Ricci tensor of the manifold. Considering the
spacetime as a $(HGQE)_{4}$ spacetime we observe from equation $(\ref{e1.5})$
that the manifold becomes Ricci symmetric if it satisfies the following
relation:
$\displaystyle\nabla_{Z}S(X,Y)=d\alpha(Z)g(X,Y)+d\beta(Z)A(X)A(Y)$
$\displaystyle+\beta[(\nabla_{Z}A)(X)A(Y)+A(X)(\nabla_{Z}A)(Y)]$
$\displaystyle+d\gamma(Z)[A(X)B(Y)+A(Y)B(X)]$
$\displaystyle+\gamma[(\nabla_{Z}A)(X)B(Y)+A(X)(\nabla_{Z}B)(Y)]$
$\displaystyle+(\nabla_{Z}A)(Y)B(X)+A(Y)(\nabla_{Z}B)(X)]$
$\displaystyle+d\delta(Z)[A(X)D(Y)+A(Y)D(X)]$
$\displaystyle+\delta[(\nabla_{Z}A)(X)D(Y)+A(X)(\nabla_{Z}D)(Y)]$
$\displaystyle+(\nabla_{Z}A)(Y)D(X)+A(Y)(\nabla_{Z}D)(X)]=0.$ (6.1)
Putting $X=Y=\xi_{1}$ in $(\ref{e6.1})$ we get,
$-d\alpha(Z)+d\beta(Z)-2\gamma(\nabla_{Z}B)(\xi_{1})-2\delta(\nabla_{Z}D)(\xi_{1})=0.$
(6.2)
Putting $X=Y=\xi_{2}$ in $(\ref{e6.1})$ we get,
$d\alpha(Z)+2\gamma(\nabla_{Z}A)(\xi_{2})=0.$ (6.3)
Again putting $X=Y=\xi_{3}$ in $(\ref{e6.1})$ we get,
$d\alpha(Z)+2\delta(\nabla_{Z}A)(\xi_{3})=0.$ (6.4)
Since the vector fields $\xi_{1},\xi_{2},\xi_{3}$ are mutually orthogonal then
$g(\xi_{1},\xi_{2})=g(\xi_{1},\xi_{3})=0$, this implies that
$Z(g(\xi_{1},\xi_{2}))=Z(g(\xi_{1},\xi_{3}))=0$. Which further implies,
$(\nabla_{Z}B)(\xi_{1})=-(\nabla_{Z}A)(\xi_{2})$ (6.5)
and
$(\nabla_{Z}D)(\xi_{1})=-(\nabla_{Z}A)(\xi_{3}).$ (6.6)
Subtracting equations (6.3),(6.4) from (6.2) and using the relations (6.5) and
(6.6) we get,
$d(\beta-3\alpha)(Z)=0,\leavevmode\nobreak\ \forall\leavevmode\nobreak\
Z\in{\mathchoice{\raisebox{0.0pt}{$\displaystyle\chi$}}{\raisebox{0.0pt}{$\textstyle\chi$}}{\raisebox{0.0pt}{$\scriptstyle\chi$}}{\raisebox{0.0pt}{$\scriptscriptstyle\chi$}}}(M).$
(6.7)
That implies $\beta-3\alpha$ is a constant.
Now contracting the equation $(\ref{e6.1})$ over $X,W$ and using
$(\ref{6.5}),(\ref{6.6})$ we get,
$d(\beta-4\alpha)(Z)=0,\leavevmode\nobreak\ \forall\leavevmode\nobreak\
Z\in{\mathchoice{\raisebox{0.0pt}{$\displaystyle\chi$}}{\raisebox{0.0pt}{$\textstyle\chi$}}{\raisebox{0.0pt}{$\scriptstyle\chi$}}{\raisebox{0.0pt}{$\scriptscriptstyle\chi$}}}(M).$
(6.8)
From the equations (6.7) and (6.8) it is clear that $\alpha,\beta$ are
constants. So, from the equations (6.3) and (6.4) we get
$\gamma(\nabla_{Z})(\xi_{1})=0$ (6.9)
and
$\delta(\nabla_{Z})(\xi_{2})=0.$ (6.10)
The equation (6.9) shows $\gamma=0$ or $\gamma(\nabla_{Z})(\xi_{1})=0$. If
$\gamma=\delta$ then from (6) we get,
$\beta[(\nabla_{Z}A)(X)A(Y)+A(X)(\nabla_{Z}A)(Y)=0.$ (6.11)
Putting $x=\xi_{1}$ in (6.11) we get,
$\beta(\nabla_{Z}A)(Y)=0.$ (6.12)
If $\beta=0$ then since $\gamma=\delta=0$ the manifold reduces to an Einstein
manifold which is a contradiction. So, $\beta\neq 0$. This implies that,
$(\nabla_{Z}A)(Y)=0.$ (6.13)
Again if $\gamma\neq 0$ then from (6.9) we get,
$\beta(\nabla_{Z}A)(\xi_{2})=0.$ (6.14)
Using (6.14) and putting $X=\xi_{1},Y=\xi_{2}$ in the equation (6) we get,
$d\gamma(Z)=0.$ (6.15)
Which imply $\gamma$ is also a constant. Then, putting $X=\xi_{2}$ in (6) and
using (6.15) we get,
$\gamma(\nabla_{Z}A)(Y)=0.$ (6.16)
Since $\gamma\neq 0$ thus we get the equation (6.13) once again. Hence, we
always obtain $(\nabla_{Z}A)(Y)=0\leavevmode\nobreak\
\forall\leavevmode\nobreak\
Z,Y\in{\mathchoice{\raisebox{0.0pt}{$\displaystyle\chi$}}{\raisebox{0.0pt}{$\textstyle\chi$}}{\raisebox{0.0pt}{$\scriptstyle\chi$}}{\raisebox{0.0pt}{$\scriptscriptstyle\chi$}}}(M).$
Which can be written as
$g(\nabla_{Z}\xi_{1},Y)=0\leavevmode\nobreak\ \forall\leavevmode\nobreak\
Z,Y\in{\mathchoice{\raisebox{0.0pt}{$\displaystyle\chi$}}{\raisebox{0.0pt}{$\textstyle\chi$}}{\raisebox{0.0pt}{$\scriptstyle\chi$}}{\raisebox{0.0pt}{$\scriptscriptstyle\chi$}}}(M).$
(6.17)
Thus we obtain
$\nabla_{Z}\xi_{1}=0.$ (6.18)
Which implies that the generator vector field $\xi_{1}$ is always parallel.
Hence we obtain the following theorem as:
Theorem 6.1: In a Ricci symmetric $(HGQE)_{4}$ spacetime with $\gamma=\delta$
the generator vector field $\xi_{1}$ is always parallel.
Again putting $Z=\xi_{1}$ in (6.18) we get,
$\nabla_{\xi_{1}}\xi_{1}=0.$ (6.19)
Which implies that the integral curves of $\xi{1}$ are geodesics. This leads
us to the next theorem as:
Theorem 6.2: In a Ricci symmetric $(HGQE)_{4}$ spacetime with $\gamma=\delta$
the integral curves of the generator vector field $\xi_{1}$ are geodesics.
With the help of the theorems (6.1) and (6.2) we arrive at the following
condition
$R(X,Y)\xi_{1}=\nabla_{X}\nabla_{Y}\xi_{1}-\nabla_{Y}\nabla_{X}\xi_{1}-\nabla_{[X,Y]}\xi_{1}=0.$
(6.20)
Hence we obtain the following theorem as:
Theorem 6.3: In a Ricci symmetric $(HGQE)_{4}$ spacetime with $\gamma=\delta$
the Riemannian curvature tensor vanishes at the generator vector field.
Now contracting (6.20) we get
$S(X,\xi_{1})=0.$ (6.21)
Thus from (1) we get,
$(\alpha-\beta)A(X)-\gamma[B(X)-D(X)]=0.$ (6.22)
Since $\gamma=\delta$ thus putting $X=\xi_{1}$ in (6.22) we get,
$\alpha=\beta.$ (6.23)
Again putting $X=\xi_{2}$ in (6.22) we get,
$\gamma=0.$ (6.24)
Thus $\gamma=\delta=0$, hence from (1) we get,
$S(X,Y)=\alpha[g(X,Y)+A(X)A(Y)].$ (6.25)
This allows us to arrive at the next theorem as:
Theorem 6.4: Every Ricci symmetric $(HGQE)_{4}$ spacetime with $\gamma=\delta$
is a $(QE)_{4}$ spacetime with the scalar functions are constants and equal.
Using the equations (6.25) from the equation (1.9) we get,
$T(X,Y)=\frac{2\lambda-\alpha}{2k}g(X,Y)+\frac{\alpha}{k}A(X)A(Y).$ (6.26)
Since $\alpha,\lambda,k$ all are constants thus taking derivative to both the
sides the equation (6.26) we get,
$(\nabla_{Z}T)(X,Y)=0.$ (6.27)
Therefore we see in this case the energy-momentum tensor is covariantly
constant. This leads us to the following theorem:
Theorem 6.5: In a Ricci-symmetric $(HGQE)_{4}$ spacetime with $\gamma=\delta$
satisfying Einstein field equation with cosmological constant the energy-
momentum tensor is covariantly constant.
Again from (6.26) we observe that since the velocity vector field $\xi_{1}$ is
parallel and $\alpha$ is a constant thus the energy-momentum tensor is of
Codazzi type. Hence we derive the next theorem as:
Theorem 6.6: In a Ricci-symmetric $(HGQE)_{4}$ spacetime with $\gamma=\delta$
satisfying Einstein field equations with cosmological constant the energy-
momentum tensor is of Codazzi type.
Now from the equations (1.8), (1.9) and theorem (6.4) we conclude the values
of $\sigma$ and $p$ as,
$\sigma=\frac{3\alpha-2\lambda}{2k},\leavevmode\nobreak\
p=\frac{2\lambda-\alpha}{2k}.$ (6.28)
Since, $\alpha,\lambda,k$ all are constants thus we get $\sigma,p$ are also
constants. This leads us to the following theorem:
Theorem 6.7: In a Ricci-symmetric $(HGQE)_{4}$ spacetime with $\gamma=\delta$
satisfying Einstein field equation with cosmological constant the energy
density and the isotropic pressure are constants.
It is proved [2] that in a perfect fluid spacetime if the energy-momentum
tensor is of Codazzi type then the vorticity and shear of the spacetime
vanish. Hence we derive the next theorem:
Theorem 6.8: In a Ricci-symmetric $(HGQE)_{4}$ spacetime with $\gamma=\delta$
satisfying Einstein field equations with cosmological constant the vorticity
and the shear tensor vanish.
Here we see that the velocity vector $\xi_{1}$ is constant over the spacelike
hypersurface orthogonal to $\xi_{1}$. But it is described in [5] that perfect
fluid spacetime that is vorticity free and shear free is of petrov type $I,D$
or $O$. Thus we state the next theorem as:
Theorem 6.9: The local cosmological structure of a Ricci-symmetric
$(HGQE)_{4}$ spacetime with $\gamma=\delta$ satisfying Einstein field equation
with cosmological constant can be identified as petrov type $I,D$ or $O$.
## 7 $(HGQE)_{4}$ spacetime with cyclic parallel Ricci tensor
Consider an $(HGQE)_{4}$ with cyclic parallel Ricci tensor. Then we get the
following equation,
$(\nabla_{X}S)(Y,Z)+(\nabla_{Y}S)(X,Z)+(\nabla_{Z}S)(X,Y)=0.$ (7.1)
From the equation $(\ref{e1.9})$ we have
$(\nabla_{X}S)(Y,Z)=\frac{1}{2}dr(Z)g(X,Y)+k(\nabla_{Z}T)(X,Y).$ (7.2)
Now, if in an $(HGQE)_{4}$ with cyclic parallel Ricci tensor the scalar
curvature of the spacetime is constant then,
$dr(X)=0,$ (7.3)
for all
$X\in{\mathchoice{\raisebox{0.0pt}{$\displaystyle\chi$}}{\raisebox{0.0pt}{$\textstyle\chi$}}{\raisebox{0.0pt}{$\scriptstyle\chi$}}{\raisebox{0.0pt}{$\scriptscriptstyle\chi$}}}(M)$.
Using the equation $(\ref{e7.3})$ in the equation $(\ref{e8.2})$ we get,
$(\nabla_{X}S)(Y,Z)=k(\nabla_{Z}T)(X,Y).$ (7.4)
Now, if the $(HGQE)_{4}$ spacetime is cyclic parallel then,
$\displaystyle
k\\{(\nabla_{X}T)(Y,Z)+(\nabla_{Y}T)(X,Z)+(\nabla_{Z}T)(X,Y)\\}$
$\displaystyle=(\nabla_{X}S)(Y,Z)+(\nabla_{Y}S)(X,Z)+(\nabla_{Z}S)(X,Y)=0.$
(7.5)
Since $k$, being the gravitational constant is always nonzero, from the
equation (LABEL:e8.5) we have
$(\nabla_{X}T)(Y,Z)+(\nabla_{Y}T)(X,Z)+(\nabla_{Z}T)(X,Y)=0.$ (7.6)
This allows us to obtain the following theorem as:
Theorem 7.1: In a hyper-generalised quasi-Einstein spacetime with cyclic
parallel Ricci tensor if the scalar curvature is constant then the energy-
momentum tensor is also cyclic parallel.
## 8 On the physical applications of an $(HGQE)_{4}$ spacetime
Here we study some physical applications of the $(HGQE)_{4}$ spacetime. In
[10], [11] G. F. R. Ellis has given the energy momentum tensor of a fluid
matter distribution as follows:
$\displaystyle T(X,Y)=(\sigma+p)A(X)A(Y)+pg(X,Y)+A(X)B(Y)+A(Y)B(X)$
$\displaystyle\leavevmode\nobreak\ \leavevmode\nobreak\ \leavevmode\nobreak\
\leavevmode\nobreak\ \leavevmode\nobreak\ \leavevmode\nobreak\
\leavevmode\nobreak\ \leavevmode\nobreak\ \leavevmode\nobreak\
\leavevmode\nobreak\ \leavevmode\nobreak\ \leavevmode\nobreak\
\leavevmode\nobreak\ \leavevmode\nobreak\ +A(X)D(Y)+A(Y)D(X),$ (8.1)
where,
$\leavevmode\nobreak\ \leavevmode\nobreak\ \leavevmode\nobreak\
\leavevmode\nobreak\ \leavevmode\nobreak\ \leavevmode\nobreak\
\leavevmode\nobreak\ \leavevmode\nobreak\ \leavevmode\nobreak\
\leavevmode\nobreak\ \leavevmode\nobreak\ \leavevmode\nobreak\
\leavevmode\nobreak\ g(X,\xi_{1})=A(X),\leavevmode\nobreak\
g(X,\xi_{2})=B(X),\leavevmode\nobreak\ g(X,\xi_{3})=D(X),$
$\leavevmode\nobreak\ \leavevmode\nobreak\ \leavevmode\nobreak\
\leavevmode\nobreak\ \leavevmode\nobreak\ \leavevmode\nobreak\
\leavevmode\nobreak\ \leavevmode\nobreak\ \leavevmode\nobreak\
\leavevmode\nobreak\ \leavevmode\nobreak\ \leavevmode\nobreak\
\leavevmode\nobreak\ A(\xi_{1})=-1,\leavevmode\nobreak\
B(\xi_{2})=1,\leavevmode\nobreak\ D(\xi_{3})=1,$
$\leavevmode\nobreak\ \leavevmode\nobreak\ \leavevmode\nobreak\
\leavevmode\nobreak\ \leavevmode\nobreak\ \leavevmode\nobreak\
\leavevmode\nobreak\ \leavevmode\nobreak\ \leavevmode\nobreak\
\leavevmode\nobreak\ \leavevmode\nobreak\ \leavevmode\nobreak\
\leavevmode\nobreak\ g(\xi_{1},\leavevmode\nobreak\
\xi_{2})=0,\leavevmode\nobreak\ g(\xi_{2},\leavevmode\nobreak\
\xi_{3})=0,\leavevmode\nobreak\ g(\xi_{3},\leavevmode\nobreak\ \xi_{1})=0,$
and $\sigma$ is the matter density, $p$ is the isotropic pressure, $\xi_{1}$
is the timelike velocity vector field, $\xi_{2}$ is the heat conduction vector
field and $\xi_{3}$ is the stress vector field.
Combining equation (8) with equation (1.9) we get,
$\displaystyle S(X,Y)=(kp+\frac{r}{2}-\lambda)g(X,Y)+k(\sigma+p)A(X)A(Y)$
$\displaystyle\leavevmode\nobreak\ \leavevmode\nobreak\ \leavevmode\nobreak\
\leavevmode\nobreak\ \leavevmode\nobreak\ \leavevmode\nobreak\
\leavevmode\nobreak\ \leavevmode\nobreak\ \leavevmode\nobreak\
\leavevmode\nobreak\ \leavevmode\nobreak\ \leavevmode\nobreak\
\leavevmode\nobreak\ \leavevmode\nobreak\ +k[A(X)B(Y)+A(Y)B(X)]$
$\displaystyle\leavevmode\nobreak\ \leavevmode\nobreak\ \leavevmode\nobreak\
\leavevmode\nobreak\ \leavevmode\nobreak\ \leavevmode\nobreak\
\leavevmode\nobreak\ \leavevmode\nobreak\ \leavevmode\nobreak\
\leavevmode\nobreak\ \leavevmode\nobreak\ \leavevmode\nobreak\
\leavevmode\nobreak\ \leavevmode\nobreak\ +k[A(X)D(Y)+A(Y)D(X)].$ (8.2)
Comparing equation (8) with equation (1.9) we get that this spacetime is
clearly an $(HGQE)_{4}$ spacetime with the constants $\gamma=\delta=k.$ Hence
all the results that we derived in the earlier sections are absolutely
effective in this spacetime, and thus we derive the following set of theorems:
From theorem (2.1) we get,
Theorem 8.1: A $W_{2}$-Ricci pseudosymmetric viscous fluid $(HGQE)_{4}$
spacetime is a $(QE)_{4}$ spacetime.
From theorem (3.1) we get,
Theorem 8.2: A $Z$-Ricci pseudosymmetric viscous fluid $(HGQE)_{4}$ spacetime
is a $(GQE)_{4}$ spacetime.
From theorem (4.1) we get,
Theorem 8.3: A Ricci pseudosymmetric viscous fluid $(HGQE)_{4}$ spacetime is a
$(QE)_{4}$ spacetime.
From theorem (5.1) we get,
Theorem 8.4: A projectively pseudosymmetric viscous fluid $(HGQE)_{4}$
spacetime is a $(QE)_{4}$ spacetime.
From theorem (6.1) we get,
Theorem 8.5: In a Ricci symmetric $(HGQE)_{4}$ viscous fluid spacetime the
generator vector field $\xi_{1}$ is always parallel.
From theorem (6.2) we get,
Theorem 8.6: In a Ricci symmetric $(HGQE)_{4}$ viscous fluid spacetime the
integral curves of the generator vector field $\xi_{1}$ are geodesics.
From theorem (6.3) we get,
Theorem 8.7: In a Ricci symmetric $(HGQE)_{4}$ viscous fluid spacetime the
Riemannian curvature tensor vanishes at the generator vector field.
From theorem (6.4) we get,
Theorem 8.8: Every Ricci symmetric $(HGQE)_{4}$ viscous fluid spacetime is a
$(QE)_{4}$ spacetime with the scalar functions are constants and equal.
From theorem (6.5) we get,
Theorem 8.9: In a Ricci symmetric $(HGQE)_{4}$ viscous fluid spacetime
satisfying Einstein field equation with cosmological constant the energy-
momentum tensor is covariantly constant.
From theorem (6.6) we get,
Theorem 8.10: In a Ricci symmetric $(HGQE)_{4}$ viscous fluid spacetime
satisfying Einstein field equation with cosmological constant the energy-
momentum tensor is of Codazzi type.
From theorem (6.7) we get,
Theorem 8.11: In a Ricci symmetric $(HGQE)_{4}$ viscous fluid spacetime
satisfying Einstein field equation with cosmological constant the energy
density and the isotropic pressure are constants.
From theorem (6.8) we get,
Theorem 8.12: In a Ricci symmetric $(HGQE)_{4}$ viscous fluid spacetime
satisfying Einstein field equation with cosmological constant the vorticity
and the shear tensor vanish.
From theorem (6.9) we get,
Theorem 8.13: The local cosmological structure of a Ricci-symmetric
$(HGQE)_{4}$ viscous fluid spacetime satisfying Einstein field equation with
cosmological constant can be identified as petrov type $I,D$ or $O$.
## 9 Example of $(HGQE)_{4}$ spacetime
Finally we give a non-trivial example to establish the existence of
$(HGQE)_{4}$ spacetime non-trivially. For this we consider a metric known as
Lorentzian metric $g$ on $M^{4}$ given by
$ds^{2}=g_{ij}dx^{i}dx^{j}=-\frac{k}{r}(dt)^{2}+\frac{1}{\frac{c}{r}-4}(dr)^{2}+r^{2}(d\theta)^{2}+(r\sin\theta)^{2}(d\phi)^{2},$
where $i,j=1,2,3,4$ and $k,c$ are constants. Thus we obtain the nonzero
components of Christofell symbols, curvature tensors and Ricci tensors as
follows:
$\displaystyle\Gamma_{33}^{2}=4r-c,\Gamma_{12}^{1}=-\frac{1}{2r},\Gamma_{22}^{2}=\frac{c}{2r(c-4r)},\Gamma_{32}^{3}=\Gamma_{42}^{4}=\frac{1}{r},$
$\displaystyle\Gamma_{43}^{4}=\cot\theta,\Gamma_{44}^{2}=(4r-c)(\sin\theta)^{2},\Gamma_{44}^{3}=-\frac{\sin(2\theta)}{2}$
(9.1) $\displaystyle
R_{1221}=-\frac{k(c-3r)}{r^{3}(c-4r)},R_{1331}=\frac{k(c-4r)}{2r^{2}},R_{1441}=\frac{k(c-4r)(\sin\theta)^{2}}{2r^{2}}$
$\displaystyle
R_{2332}=\frac{c}{2(4r-c)},R_{2442}=\frac{c(\sin\theta)^{2}}{2(4r-c)},R_{3443}=r(c-5r)(\sin\theta)^{2}$
$\displaystyle
R_{11}=-\frac{k}{r^{3}},R_{22}=-\frac{3}{r(c-4r)},R_{33}=-3,R_{44}=-3(\sin\theta)^{2}$
(9.2)
From $(\ref{9.1})$ and $(\ref{9.2})$ it follows that $M^{4}$ is a Lorentzian
manifold of nonzero scalar curvature ($=-\frac{8}{r^{2}}$). Now we will prove
that this is an $(HGQE)_{4}$ manifold.
We consider $\alpha,\beta,\gamma$ and $\delta$ as the associated scalars and
we consider them as follows:
$\alpha=-\frac{5}{r^{2}},\beta=-\frac{12}{r^{2}},\gamma=\frac{3}{r^{2}},\delta=-\frac{4}{r^{2}}$
(9.3)
and the associated $1$-forms are as follows :
$\leavevmode\nobreak\ \leavevmode\nobreak\ \leavevmode\nobreak\
\leavevmode\nobreak\ \leavevmode\nobreak\ \leavevmode\nobreak\
\leavevmode\nobreak\ \leavevmode\nobreak\ \leavevmode\nobreak\
\leavevmode\nobreak\ \leavevmode\nobreak\ \leavevmode\nobreak\
A_{i}(x)=\left\\{\begin{array}[]{cccl}\sqrt{\frac{k}{2r}}&\mbox{for}&i=1\\\
\sqrt{\frac{r}{6(c-4r)}}&\mbox{for}&i=2\\\ 0&\mbox{for}&i=3,4\\\
\end{array}\right.$ ;
$B_{i}(x)=\left\\{\begin{array}[]{ccl}\sqrt{\frac{k}{r}}&\mbox{for}&i=1\\\
0&\mbox{for}&i=2,3,4\end{array}\right.$
$\leavevmode\nobreak\ \leavevmode\nobreak\ \leavevmode\nobreak\
\leavevmode\nobreak\ \leavevmode\nobreak\ \leavevmode\nobreak\
\leavevmode\nobreak\ \leavevmode\nobreak\ \leavevmode\nobreak\
\leavevmode\nobreak\
D_{i}(x)=\left\\{\begin{array}[]{ccl}\sqrt{\frac{k}{r}}&\mbox{for}&i=1\\\
0&\mbox{for}&i=2,3,4\end{array}\right.$
Hence we gain,
$(i)R_{11}=\alpha g_{11}+\beta
A_{1}A_{1}+\gamma[A_{1}B_{1}+B_{1}A_{1}]+\delta[A_{1}D_{1}+D_{1}A_{1}]$
$(ii)R_{22}=\alpha g_{22}+\beta
A_{2}A_{2}+\gamma[A_{2}B_{2}+B_{2}A_{2}]+\delta[A_{2}D_{2}+D_{2}A_{2}]$
$(iii)R_{33}=\alpha g_{33}+\beta
A_{3}A_{3}+\gamma[A_{3}B_{3}+B_{3}A_{3}]+\delta[A_{3}D_{3}+D_{3}A_{3}]$
$(iv)R_{44}=\alpha g_{44}+\beta
A_{4}A_{4}+\gamma[A_{4}B_{4}+B_{4}A_{4}]+\delta[A_{4}D_{4}+D_{4}A_{4}]$
As every Ricci tensor other than $R_{11},R_{22},R_{33}$ and $R_{44}$ are zero,
so we obtain
$R_{ij}=\alpha g_{ij}+\beta
A_{i}A_{j}+\gamma[A_{i}B_{j}+B_{i}A_{j}]+\delta[A_{i}D_{j}+D_{i}A_{j}],i,j=1,2,3,4.$
Consequently, scalar curvature $=4\alpha-\beta=-\frac{8}{r^{2}}$. Hence,
$(M^{4},g)$ is a hyper-generalized quasi Einstein manifold.
Conclusion: The general theory of relativity is the most prominent flagship of
modern physics. It deals with the curvature of spacetime. As hyper-generalized
quasi-Einstein spacetime is considered as the base space of the fluid matter
distribution. Thus it has been very necessary to study about the geometric and
physical applications of hyper-generalized quasi-Einstein spacetime. It deals
with the relativistic viscous fluid spacetime admitting heat flux and stress.
The general theory of relativity describes gravity as a geometric property of
spacetime. The curvature of spacetime is directly related to the energy-
momentum tensor. Also we know the cosmological constant to be of homogeneous
energy density which causes the expansion of the universe to accelerate. So,
here we obtain a set of geometric and physical properties of hyper-generalized
quasi-Einstein spacetimes and give a non-trivial example to establish the
existence of hyper-generalized quasi-Einstein spacetime non-trivially.
## References
* [1] A. A. Shaikh, C. Özgür, A. Patra, On hyper-generalized quasi-Einstein manifolds, Int. J. of Math. Sci. and Engg. Appl. 5 (2011), 189-206.
* [2] A. A. Shaikh, D. Y. Yoon and S. K. Hui, On quasi-Einstein spacetime, Tsukuba J. Math. 33(2) (2009), 305-326.
* [3] A. Bhattacharyya, T. De and D. Debnath, Space time with generalized covariant recurrent energy momentum tensor, Tamsui Oxford Journal of Mathematical Sciences 25(3) (2009), 269-276.
* [4] A. L. Besse, Einstein manifolds, Ergeb. Math. Grenzgeb., 3. Folge, Bd. 10. Berlin, Heidelberg, New York : Springer-Verlag. 1987.
* [5] A. Z. Petrov, Einstein Spaces, Oxford: Pergamon Press (1949).
* [6] B. O’Neill, Semi-Riemannian geometry and application to relativity, Pure and applied mathematics, Academic press. Inc. New York. (1983).
* [7] C. A. Mantica, L. G. Molinari, Weakly Z symmetric manifolds, Acta Math. Hunger., 135 (2012), 80-96.
* [8] C. A. Mantica, Y. J. Suh, Pseudo Z symmetric Riemannian manifolds with harmonic curvature tensors, Int. J. Geom. Methods. Mod. Physics., 9/1 (2012), 1250004.
* [9] D. Debnath, On hyper-generalised quasi-Einstein manifolds, International J. Math. Combin. Vol 3 (2018), 25-31.
* [10] G. F. R. Ellis, Relativistic cosmology, In: Sachs RK(ed), Academic Press, London, Course: 47 (1971), 104-182.
* [11] G. F. R. Ellis, R. Maartens, M. MacCallum Relativistic cosmology, Cambridge University Press, UK(1983).
* [12] G. P. Pokhariyal and R. S. Mishra, Curvature tensors and their relativistic significance, Yokohama Math. J. 19(2) (1971), 97-103.
* [13] K. Chattopadhyay, D. Debnath and A. Bhattacharyya, A study of spacetimes with vanishing $M$-projective curvature tensor, Journal of The Tensor Society of India Vol. 12 (2018), 23-31.
* [14] K. Chattopadhyay, N. Bhunia and A. Bhattacharyya, On Ricci-symmetric mixed generalized quasi-Einstein spacetime, Bull. Cal. Math. Soc. 110(6) (2018), 513-524.
* [15] K. Yano, M. Kon, Structures on manifolds, Series in pure mathematics, vol. 3. Singapore: World Scientific publishing Company(1984).
* [16] L. D. Landau, E. M. Lifshitz, Course of theoretical physics, Fluid Mechanics Vol. 6 (1963)
* [17] M. C. Chaki, On generalized quasi-Einstein manifolds, Publicationes Mathematicae 58(4) (2001).
* [18] M. C. Chaki and R. K. Maity, On quasi-Einstein manifolds, Publ. Math. Debrecen 57 (2000), 297-306.
* [19] S. Güler and S. A. Demirbaǧ, A study of generalized quasi-Einstein spacetime with application in general relativity, International Journal of Theoretical Physics 55 (2016), 548-562.
* [20] S. Güler, S. A. Demirbaǧ, Hyper-generalized quasi-Einstein manifolds satisfying certain Ricci conditions, Proceedings Book of International Workshop on Theory of Submanifolds 1 (2016), 205-215.
* [21] S. Mallick, Y. J. Suh and U. C. De, A spacetime with pseudo-projective curvature tensor, J. Math. Phys. 57 (2016), Paper ID 062501.
* [22] S. Ray-Guha, On perfect fluid pseudo Ricci symmetric spacetime, Tensor (N.S.) 67 (2006), 101-107.
* [23] Y. J. Suh, P. Majhi, U. C. De, On mixed quasi-Einstein spacetimes, Filomat 32:8 (2018), 2707-2719.
* [24] Z. Ahsan and M. Ali, Curvature tensor for the spacetime of general relativity, Int. J. of Geom. Methods Mod. Phys. 14(5) (2017), 1750078(13 pages).
* [25] Z. Ahsan and M. Ali, Quasi-conformal curvature tensor for the spacetime of general relativity, Palestine Journal of Mathematics 4(1) (2015), 234-241.
* [26] Z. Ahsan and S. A. Siddiqui, Concircular curvature tensor and fluid spacetimes, Internat. J. Theoret. Phys. 48 (2009), 3202-3212.
|
# Powering COVID-19 community Q&A with Curated Side Information
Manisha Verma, Kapil Thadani, and Shaunak Mishra Yahoo! Research,
VerizonMedia, New York<EMAIL_ADDRESS>
(2019)
###### Abstract.
Community question answering and discussion platforms such as Reddit, Yahoo!
answers or Quora provide users the flexibility of asking open ended questions
to a large audience, and replies to such questions maybe useful both to the
user and the community on certain topics such as health, sports or finance.
Given the recent events around COVID-19, some of these platforms have
attracted 2000+ questions from users about several aspects associated with the
disease. Given the impact of this disease on general public, in this work we
investigate ways to improve the ranking of user generated answers on COVID-19.
We specifically explore the utility of external technical sources of side
information (such as CDC guidelines or WHO FAQs) in improving answer ranking
on such platforms. We found that ranking user answers based on question-answer
similarity is not sufficient, and existing models cannot effectively exploit
external (side) information. In this work, we demonstrate the effectiveness of
different attention based neural models that can directly exploit side
information available in technical documents or verified forums (e.g.,
research publications on COVID-19 or WHO website). Augmented with a
temperature mechanism, the attention based neural models can selectively
determine the relevance of side information for a given user question, while
ranking answers.
††journalyear: 2019
## 1\. Introduction
Question answering systems are key to finding relevant and timely information
about several issues. Community question answering (cQ&A) platforms such as
Reddit, Yahoo! answers or Quora have been used to ask questions about wide
ranging topics. Most of these platforms let users ask, answer, vote or comment
on questions present on the platform. However, question answering platforms
are useful not only for getting public opinions or votes about areas such as
entertainment or sports but can also serve as information hot-spots for more
sensitive topics such as health, injuries or legal topics. Thus, it is
imperative that when the user visits _sensitive_ topics content, answer
ranking also takes into account curated side information from reliable
(external) sources. Most prior work on cQ&A has focused on incorporating
question-answer similarity (LTR_2008, ; semantic_sim2015, ), user reputation
(yang2016beyond, ; hong2009classification, ; user_interaction_2013, ),
integration of multi-modal content (medical_qa2019, ), community interaction
features (comm_interaction2018, ) associated with answers or just the question
answering network (comm_net2019, ) on the platform. However, there is very
limited work on incorporating _curated content_ from external sources.
Existing work only exploits knowledge bases (medical_qa2019, ) that consist of
different entities and relationships between these entities to score answers.
However, there are some limitations of knowledge bases that would make it
difficult to use them for community Q&A for rapidly evolving topics such as
disease outbreaks (e.g. ebola, COVID-19), wild-fires or earthquakes.
Figure 1. Illustrative example of COVID-19 community answer ranking powered by
side information in the form of research papers, and information from verified
sources (such as CDC, WHO, and NHS).
Firstly, knowledge bases contain information about established entities, and
do not rapidly evolve to incorporate new information which makes them
unreliable for novel disease outbreaks such as COVID-19 where information
rapidly changes and its verification is time sensitive. Secondly, it may be
hard to determine what even _constitutes an entity_ as new information arrives
about the topic. To overcome these limitation in this work, we posit that
external _curated_ free-text or semi-structured informational sources can also
be used effectively for cQ&A tasks.
In this work, we demonstrate that free text or semi structured external
information sources such as CDC111https://www.cdc.gov/,
WHO222https://www.who.int/ or NHS333https://www.nhs.uk/ can be very useful for
ranking answers on community Q&A platforms since they contain frequently
updated information about several topics such as ongoing disease outbreaks,
vaccines or resources about other topics such as surgeries, birth control or
historical numerical data about diseases across the world.
We argue that for sensitive topics such as COVID-19, it is useful to use
publicly available _vetted_ information for improving our ranking systems. In
this work, we explore the utility of publicly available information for
ranking answers for questions associated with COVID-19. We specifically focus
on ranking answers for questions in two publicly available _primary_ Q&A
datasets: a) Yahoo! Answers444https://answers.yahoo.com/ and b) recently
released annotated Q&A dataset (emnlp_data2020, ) in presence of two
_external_ semi-structured curated sources: a) TREC-COVID
(voorhees2020treccovid, ) and b) WHO questions and answers
555https://www.who.int/emergencies/diseases/novel-coronavirus-2019/question-
and-answers-hub on COVID-19. We explore the utility of deep learning models
with attention in this work to improve upon existing state-of-the-art systems.
More specifically, we propose a temperature regulated attention mechanism to
rank answers in presence of external (side) information. Our experiments on
10K+ questions from both _source_ datasets on COVID-19 show that our models
can improve ranking quality by a significant margin over question-answer
matching baselines in presence of external information. Figure 1 demonstrates
the overall design of our system. We specifically use attention based neural
architecture with temperature to automatically determine which components in
the external information are useful for ranking user answers with respect to a
question. Ranking performance, when evaluated with three metrics shows that
precision and recall for correct answer retrieval improves by $\sim$17% and
$\sim$9% for both source datasets respectively over several other cQ&A models.
## 2\. Related work
Community Question and Answering (cQ&A) systems is a well researched sub-field
both in information retrieval and NLP communities. Several systems have been
proposed to rank user submitted answers to questions on community platforms
such as Yahoo! answers, Reddit and Quora.
Ranking user submitted answers on community question-answering platforms has
been addressed with several approaches. Primary method is to determine the
relevance of the answer given an input question. Text based matching is one of
the most common approaches to rank answers. Researchers have used several
methods to compute _similarity_ between a question and user generated answers
to determine relevance. For instance, feature based question-answer matching
is used in (LTR_2008, ) with 17 features extracted from unigrams, bigrams and
web correlation features using unstructured user search logs to rank answers.
It is worth noting that user features and community features when incorporated
may still yield further improvements in the performance of these models but
this is not the focus of our work. The authors in (LTR_2008, ) used questions
extracted from Yahoo! answers for their experiments. Researchers have used
different approaches such representation learning, for instance, in
(LSTMans2017IWCS, ; cohen_2016, ) authors use LSTM to represent questions and
answers respectively. Convolutional networks have also been used in
(yang2016beyond, ; zhou2015answer, ) to rank answers. Other approaches such as
doc2vec (nie2017data, ), tree-kernels (severyn2012structural, ), adversarial
learning (yang2019adversarial, ), attention (attention_question2018, ;
medical_qa2019, ; LSTMans2017IWCS, ; attentive2017AAAI, ) or deep belief
networks (wang2010modeling, ) have been used to score question and answer
pairs. There have also been studies exploring community, user interaction or
question based features (yang2016beyond, ; hong2009classification, ;
user_interaction_2013, ; comm_interaction2018, ) to rank answers. While these
approaches are relevant, it is not always evident how one can incorporate
external information when it is either in free-text or semi-structured format
into these systems. We explore some question-answer based matching approaches
as baselines in this work and show that for rapidly evolving topics such as
COVID-19, inclusion of external _curated_ information can boost model
performance.
The line of work most closely related to ours is incorporation of knowledge
bases in Q&A systems. Existing work (rankQA_2019, ; kb_rank2018Sig, ;
medical_qa2019, ), however, approaches different tasks. For instance, authors
in (kb_rank2018Sig, ; rankQA_2019, ) focus on finding factual answers to
questions using a knowledge base. This does not extend easily to cQ&A where
neither the questions nor the answers may request or refer to any facts. Most
recent work is (medical_qa2019, ) on incorporating medical KB for ranking
answers on medical Q&A platforms. They propose to learn path based
representation of _entities (from KB)_ present in question and answers posted
by users. This approach relies on reliable detection of entities first, which
may be absent for emerging topics such as COVID-19 pandemic. Another
limitation of this work is that external knowledge may not always be present
in a _structured_ format. For example, CDC guidelines are usually simple
question-answer pairs posted on the website. This makes it difficult to apply
their approach to our problem. The proposed approach in this work incorporates
semi-structured information directly with help of temperature regulated
attention.
Finally, with the rise of COVID-19, researchers across disciplines are
actively publishing information and datasets to share understanding of the
virus and its impact on people. Researchers routinely organize dedicated
challenges such as SemEval (nakov2019semeval, ) with tasks such as ranking
answers on QA forums. One such initiative is TREC-COVID track
(voorhees2020treccovid, ) which released queries, documents and manual
relevance judgements to power search for COVID related information. Authors in
(cairecovid, ) also released COVID-19 related QA dataset with 100+ questions
and answers pairs extracted from TREC COVID
666https://ir.nist.gov/covidSubmit/data.html initiative. These
questions/answer pairs are not user generated content, hence, do not reflect
real user questions. We also rely on recently released Q&A dataset from
(emnlp_data2020, ) for our task. We also compile a dataset of 2000+ COVID-19
questions with 10K+ answers all submitted by users on Yahoo! answers for this
work.
## 3\. Method
### 3.1. Problem formulation
In this work, we focus on ranking answers for $n$ questions
$q_{1},\ldots,q_{n}$ related to an emerging topics such as COVID. Each $q_{i}$
is associated with a set of two or more answers $A_{i}=\\{a_{ij}:j\geq 2\\}$
and corresponding labels $Y_{i}=\\{y_{ij}:j\geq 2\\}$ representing answer
relevance. We use binary indicator for relevance where relevant judgments
(e.g., favorite, upvoted) are provided by the user, i.e., $y_{ij}\in\\{0,1\\}$
respectively.
We attempt to model the relevance of each answer $a_{ij}$ to its corresponding
question using an external source which may contain free text or semi-
structured information. For example, the _external_ source could consist of
information-seeking queries or questions $eq_{1},\ldots,eq_{m}$ related to a
topic, with each $eq_{k}$ linked to a set of relevant scientific articles or
answers $ED_{k}$, where each answer/document $ed_{1},\ldots,ed_{p}$ may be
judged for relevance by human judges (voorhees2020treccovid, ) or some
experts.
We hypothesize that this semi-structured or free-text information may be
valuable in identifying user answer quality for certain kinds of questions,
although not all. We investigate this with our model to recover the true
labels $y_{ij}$ for each user answer $a_{ij}\in A_{i}$ given its question
$q_{i}$, category information, and information from the _external_ source
$\langle eq_{k},ED_{k}\rangle_{k=1}^{m}$.
### 3.2. Proposed Model
Figure 2. External source augmentation model
In this work, we explore token-level matching mechanism to determine the
relevance of information in the external source that may inform the label
prediction task. Our model (_$\tau$ -att_) aims to match a given user question
with all the submitted answers in the presence of external information about
the same domain. First, the question $q_{i}$, an answer $a_{ij}$ and
additional metadata can be encoded into a $d$-dimensional vector $x_{i}$ using
a text encoder $f_{\text{input}}$. We use LSTM based encoder for both question
and answer in the _primary_ source which can handle input sequences of
variable length.
##### Question Encoding:
Each word $w_{i}^{q}$ in a question is represented as a $K$ dimensional vector
with pre-trained word embeddings. LSTM takes each token embedding as input and
updates hidden state $h_{i}^{q}$ based on previous state $h_{i-1}^{q}$.
Finally, the hidden state is input to a feed forward layer with smaller
dimension $F<K$ to compress question encoding as follows:
(1)
$h_{i}^{q}=LSTM(h_{i-1}^{q},w_{i}^{q}),\>f_{i}^{q}=RELU(h_{i}^{q}W_{q}+b_{q})$
##### Answer Encoding:
Each word $w_{j}^{a}$ in the answer is also represented as a $K$ dimensional
vector with pre-trained word embeddings. LSTM takes each token embedding as
input and updates hidden state $h_{j}^{a}$. We also reduce the dimension of
answer encoding with a feed forward layer with dimension $F<K$ as follows:
(2)
$h_{j}^{a}=LSTM(h_{j-1}^{a},w_{j}^{a}),\>f_{j}^{a}=RELU(h_{j}^{a}W_{a}+b_{a})$
We concatenate the question and answer representations for further processing.
(3) $f_{ij}=[f_{i}^{q},f_{j}^{a}]$
##### External source encoding:
External sources of information can vary from task-to-task. We encode each
segment of data individually. For instance, if there are two segments in the
source (e.g. question/answer or query/document), our system encodes both
segments individually. We use the same encoding architecture used for primary
source question/answer encoding above. Encoding example for two segment
_external_ source is given below.
(4)
$\begin{split}h_{t}^{eq}=LSTM(h_{t-1}^{eq},w_{t}^{eq}),\>f_{t}^{eq}=RELU(h_{t}^{eq}W_{eq}+b_{eq})\\\
h_{t}^{ed}=LSTM(h_{t-1}^{ed},w_{t}^{ed}),\>f_{t}^{ed}=RELU(h_{t}^{ed}W_{ed}+b_{ed})\\\
\end{split}$
Source | Question | Rel answer | Non-rel answer
---|---|---|---
Yahoo! Ans | 3cmI am really scared to go places for St. Patrick’s day because of the coronavirus. what do I do? | 7cmUnfortunately, there’s not enough people that care and will still go out and party despite the coronavirus epidemic. I’m proud of you in that you’re taking extra precautions … Good for you! | 5cmStop being scared of viruses. What’s the problem?
Infobot | 3cmCan corona live on cardboard? | 7cmA recent study shows that the virus can live in the air … On cardboard, it can live up to 24 hours (1 day) | 5cmThe risk is quite low for one to become infected with COVID19 through mail/packages - especially because…(over a period of a few days/weeks).
Table 1. Sample rel/non-rel answers from both sources
We incorporate external source encoding with a temperature $(\tau)$ based
variant of scaled dot-product attention, which provides a straightforward
conditioning approach over a set of query-document pairs. Question encoding
vector $f_{ij}$ serves as a query over keys $f_{t}^{eq}$. If two segments are
present in the external source such as query/document, the model uses the
attention weights over first segment (e.g. query) to determine the importance
of the second segment (e.g. document) respectively. It is easy to extend this
framework to _external_ sources with multiple segments. The two segment
attention is described below.
(5) $z_{it}=\frac{f_{ij}^{\top}f_{t}^{eq}}{\sqrt{d}}\\\
\alpha_{it}=\frac{e^{z_{it}/\tau}}{\sum_{l}e^{z_{l}/\tau}}\\\
s_{itd}^{\prime}=\sum_{d}\alpha_{it}f_{t}^{ed}$
To summarize, temperature $(\tau)$ based attention helps determine the
relevance of each $f_{t}^{ed}$ corresponding $f_{t}^{eq}$ with respect to the
question encoding. Temperature $(\tau)$ parameter helps us control the
uniformity of attention weights $\alpha_{it}$. Finally, labels are predicted
using a multi-layer perceptron over the input vector $f_{ij}$ and the learned
weighted average of side information $s_{itd}^{\prime}$. We use binary cross
entropy loss to train the proposed model.
(6) $\hat{y}_{ij}=F_{\text{output}}([f_{ij};s_{itd}^{\prime}])$
where $F_{\text{output}}$ uses sigmoid activation function. Since community
questions may often be entirely unrelated to external sources, a key aspect of
this approach is determining whether the _external source_ is useful, not
merely attending to its entries that are most relevant. Temperature based
attention mechanism is useful in controlling which external source entries are
useful for user questions. It is worth noting that one will have to experiment
and tune the value of temperature $\tau$ such that ranking performance
improves.
## 4\. Experimental Setup
Given the model architecture, in this section, we provide a detailed overview
of different datasets, metrics and baselines used in our experiments.
(a) Yahoo! ques length
(b) Yahoo! ans length
(c) Infobot ques length
(d) Infobot ans length
Figure 3. Token distribution in different sources
### 4.1. Data
We compiled two question answering datasets. The first was collected from
Yahoo! answers and the second was recently released in (emnlp_data2020, )
where both datasets have questions raised by real users. In this work we focus
specifically on questions associated with COVID-19. Different statistics about
the train and test split of both q&a datasets are given in Table 2
respectively. A pair of relevant and non-relevant answers for a question in
both datasets is also shown in Table 1 for reference. More details about them
is given below.
Stat | Yahoo! Ans | Infobot
---|---|---
Train Q-A | 9341 | 6354
Train ans/q | 6.25$\pm$2.9 | 4.40$\pm$0.77
Train #qwords | 12.71$\pm$5.8 | 6.55$\pm$3.93
Train #awords | 36.31$\pm$93.59 | 92.17$\pm$59.27
Test Q-A | 2232 | 1592
Test ans/q | 5.96$\pm$2.87 | 4.41$\pm$0.76
Test #qwords | 13.07$\pm$5.89 | 6.21$\pm$2.94
Test #awords | 35.64$\pm$80.31 | 92.39$\pm$59.47
Table 2. Train and test data from primary sources
##### Yahoo! Dataset
: We crawled COVID-19 related questions from Yahoo! answers
777https://answers.search.yahoo.com/search?p=coronavirus using several
keywords such as ‘coronavirus’, ‘covid-19’, ‘covid’, ‘sars-cov2’ and ‘corona
virus’ between the period of Jan 2020 to July 2020 to ensure we gather all
possible questions for our experiments. We keep only those questions have two
or more answers. In total, we obtained 1880 questions with 11500 answers. We
used favorite answers as positive labels (similar to previous work (LTR_2008,
)), assuming that users, over time rate answers (with upvotes/downvotes) that
are most relevant to the submitted question. We normalized the question and
answer text by removing a small list of stop words, numbers, links or any
symbols. Figure 3(a) and 3(b) show the distribution of question and answer
lengths respectively. Questions contain $12.7\pm 5.8$ _(qwords)_ words and
answers consist of $36.3\pm 93.5$ (mean$\pm$std) words _(awords)_ respectively
which indicates that user submitted answers can vary widely on Yahoo! answers.
On average, a question has about 6 answers _(ans/q)_ in Yahoo! ans dataset. We
spilt the data into three sets: train (64%, 1196 questions, 7435 answers),
validation (16%, 298 questions, 1858 answers) and test (20%, 374 questions,
2310 answers) set where questions for each set were uniformly sampled.
##### Infobot Dataset (emnlp_data2020, )
: Researchers at JHU (emnlp_data2020, ) have recently compiled a list of user
submitted questions on different platforms and manually labeled 22K+ question-
answer pairs. We cleaned this set by removing questions with less than two
answers or no relevant answers. In total, our dataset contains 8000+ question
answer pairs where each question may have _multiple_ relevant answers which is
not the same as Yahoo! answers dataset. Figure 3(c) and 3(d) show the
distribution of question and answer lengths respectively.
#### 4.1.1. External sources
We use two external datasets to rank answers. Details of each dataset are
given below:
##### TREC COVID (voorhees2020treccovid, ):
We use recently released TREC COVID-19 track data with 50 queries which also
contain manually drafted query descriptions and narratives. Expert judges have
labeled over 5000 scientific documents for these 50 queries from the CORD-19
dataset 888https://www.semanticscholar.org/cord19. These documents contain
coronavirus related research. Given the documents are scientific literature,
we initialize document embeddings using SPECTER (cohan2020specter, ).
##### WHO:
We use data released on question and answer hub of WHO
website999https://www.who.int/emergencies/diseases/novel-
coronavirus-2019/question-and-answers-hub to create a list of question-answer
pairs. There are 147 question and answer pairs in this dataset where questions
contain 13.28$\pm$5.36 words and answers contain 133.2$\pm$100.9 words
respectively.
### 4.2. Baselines
We evaluated our model against embedding similarity baseline. We computed four
baselines as follows:
##### Random:
An answer is chosen at random as relevant for a user question. This is
expected to provide a lower bound on retrieval performance.
##### Linear Attention (_att_)
: When $\tau=1.0$, our model defaults to simple linear attention over all the
information present in the external sources. This gives an indication of how
well the model performs when its forced to look at all the information in the
external source.
##### Linear combination (_$\lambda$ -sim_)
: We linearly combine similarities between Yahoo! question-answer and Trec
query-answer as shown below:
(7)
$\emph{$\lambda$-sim}=\lambda\;cos(ya,yq)+(1-\lambda)\;\max_{tq}(cos(ya,tq))$
where $ya$, $yq$ and $tq$ are Yahoo! answer, question and concatenated trec
query, narrative and description embeddings respectively. This is a more crude
version of temperature attention where $\lambda$ controls the contribution of
each component directly. We vary $\lambda$ to determine the optimal
combination. Question-Answer similarity (_qasim_) is similarity between
question and answer embedding i.e. $\lambda=1$. Both question and answer
embeddings are obtained by averaging over their individual token embeddings.
##### BERT Q&A (_bert_)
: Large scale pre-trained transformers (devlin2019bert, ) are widely popular
for NLP tasks. BERT like models have shown effectiveness on Q&A datasets such
as SQUAD 101010https://rajpurkar.github.io/SQuAD-explorer/. We fine-tune BERT
base model with two different answer lengths a) 128 _(bert-sl128)_ and b) 256
tokens _(bert-sl256)_ respectively. The intuition is that large scale pre-
trained models are adept at language understanding and can be fine-tuned for
new tasks with small number of samples. We finetune BERT for both datasets
Yahoo! ans and Infobot respectively. It is non-trivial to include external
information in BERT and we leave this for future work.
| Yahoo! Ans | Infobot
---|---|---
Model | P$@$1 | R$@$3 | MRR | P$@$1 | R$@$3 | MRR
$\tau$-att | 0.393 | 0.644 | 0.598 | 0.673 | 0.868 | 0.802
$\lambda$-sim | 0.3743 | 0.633 | 0.578 | 0.551 | 0.817 | 0.7207
bert-sl256 | 0.406 | 0.657 | 0.615 | 0.581 | 0.803 | 0.744
bert-sl128 | 0.363 | 0.604 | 0.589 | 0.557 | 0.799 | 0.731
att | 0.377 | 0.645 | 0.589 | 0.567 | 0.821 | 0.739
qasim | 0.318 | 0.608 | 0.546 | 0.551 | 0.817 | 0.720
random | 0.21 | - | - | 0.239 | - | -
Table 3. Evaluation with WHO external data
### 4.3. Evaluation Metrics
We evaluate the performance of our model using three popular ranking metrics,
mainly Precision (P$@$1), Mean Reciprocal Rank (MRR), and Recall (R$@$3). Each
metric is described below:
* •
Precision (P$@$k): Precision at position $k$ evaluates the fraction of
relevant answers retrieved until position k. For, both datasets Yahoo! ans and
Infobot (emnlp_data2020, ), we evaluate whether the top answer i.e. $(k=1)$ in
the ranked list is indeed correct. It is defined as follows:
(8)
$Prec@k=\frac{1}{|Q|}\sum_{i=1}^{|Q|}\frac{\sum_{j=1}^{k}\mathbb{I}\\{rel_{ij}=1\\}}{k}$
where $\mathbb{I}\\{rel_{ij}=1\\}$ indicates whether the answer at position
$j$ is relevant to the $i^{th}$ question.
* •
Recall (R$@$k): Recall at position $k$ evaluates the fraction of relevant
answers retrieved from all the answers marked relevant for a question. We
report recall averaged for all the queries in test set. For recall, we take a
cutoff as $(k=3)$, which evaluates whether the model is able to retrieve the
correct answers in top 3 positions. It is defined as follows:
(9)
$Recall@k=\frac{1}{|Q|}\sum_{i=1}^{|Q|}\frac{\sum_{j=1}^{k}\mathbb{I}\\{rel_{ij}=1\\}}{|rel_{i}|}$
where $|rel_{i}|$ is the number of relevant answers for the $i$th question.
* •
MRR (MRR): evaluates the average of the reciprocal ranks corresponding to the
most relevant answer for the questions in test set, which is given by:
(10) $MRR=\frac{1}{|Q|}\sum_{i=1}^{|Q|}\frac{1}{rank_{i}}$
where $|Q|$ indicates the number of queries in the test set and $rank_{i}$ is
the rank of the _first_ relevant answer for the $i^{th}$ query.
### 4.4. Parameter Settings
Both primary datasets, Yahoo! ans and Infobot, were divied into three parts:
train ($\sim$60%), validation and test (20%) respectively. The baseline models
$\lambda$-sim and $att$ are initialized with glove embeddings
111111https://nlp.stanford.edu/projects/glove/ of 100 dimensions. We performed
a parameter sweep over $\lambda$ and $\tau$ for $\lambda$-sim and $\tau$-att
models with step size of 0.1 between $\\{0,1.0\\}$ respectively. We used base
uncased model for $bert$ implementation. We fine-tuned the model between 1-10
epochs and found that 3 epochs gave the best result on validation set. We used
LSTM with 64 hidden units to represent question, answer and all the
information in external datasets. We experimented with higher embedding size
and hidden units, but the performance degraded significantly as the model
tends to overfit on training data. Lastly we used batch size of 64 and trained
the model for 30 epochs with early stopping.
| Yahoo! Ans | Infobot
---|---|---
Model | P$@$1 | R$@$3 | MRR | P$@$1 | R$@$3 | MRR
$\tau$-att | 0.532 | 0.778 | 0.715 | 0.606 | 0.842 | 0.766
$\lambda$-sim | 0.326 | 0.616 | 0.555 | 0.556 | 0.813 | 0.722
bert-sl256 | 0.406 | 0.657 | 0.615 | 0.581 | 0.803 | 0.744
bert-sl128 | 0.363 | 0.604 | 0.589 | 0.557 | 0.799 | 0.731
att | 0.291 | 0.495 | 0.494 | 0.601 | 0.833 | 0.762
qasim | 0.318 | 0.608 | 0.546 | 0.551 | 0.817 | 0.720
random | 0.21 | - | - | 0.239 | - | -
Table 4. Evaluation with TREC-COVID external data
## 5\. Results
In this work, our focus is to evaluate the utility of external information in
improving answer ranking for cQ&A task. Thus, we performed experiments to
answer three main research questions listed below.
RQ1: Does external information improve answer ranking?
RQ2: How does temperature ($\tau$) compare with $\lambda$ parameter?
RQ3: What kind of queries/questions does the model attend to when ranking
relevant/non-relevant answers?
##### RQ1: Does external information improve answer ranking?
We evaluated different models for ranking answers in Yahoo! ans and Infobot
dataset in presence of TREC and WHO datasets respectively. We found that
temperature regulated attention models that incorporate external sources
indeed outperform the baselines as shown in Table 4 and Table 3 respectively.
Category | $\tau$-att | $\lambda$-sim | qasim
---|---|---|---
Entertainment (47) | 0.829 | 0.702 | 0.59
Health (62) | 0.693 | 0.69 | 0.645
Politics (143) | 0.727 | 0.629 | 0.587
Society (38) | 0.578 | 0.473 | 0.42
Family (20) | 0.85 | 0.750 | 0.65
Table 5. Recall$@$3 of models across categories
(_$\tau$ -att_) model beats _bert_ models by $\sim$30% in precision, $\sim$18%
in recall and $\sim$16% in MRR respectively on TREC data. However, (_$\tau$
-att_) does only marginally better than _att_ model in precision and MRR on
Infobot data. We suspect that is due to the large set of query-document pairs
in TREC-COVID data compared to fewer number of question-answer pairs in
Infobot dataset. Our results also clearly suggest that embedding based
matching of question-answer pair (_qasim_) would not yield a good ranker,
though it is better than choosing an answer at random (_random_). When WHO is
used as an external dataset, we find that (_$\tau$ -att_) model is slightly
worse than _bert_. This suggests that not all sources would equally benefit
cQ&A task. Since attention is dependent on the input query and key embedding
lengths, it would be interesting to scale the computation in our model to
incorporate several open external datasets to overcome this limitation in the
future.
Yahoo! ans questions are also assigned categories by users. Category based
breakdown of performance on test set is given in Table 6 and Table 5
respectively, where categories with largest number of questions in test set
are listed. In all the categories, our model outperforms best $\lambda$-sim
and _qasim_ model respectively. The largest improvement happens for questions
in Family category where our model achieves an improvement of 71% over the
$\lambda$-sim model. It seems that ranking answers for questions from society
and politics are harder than other categories. All the models, however, are
able to rank the top answer in first three positions effectively as Recall$@$3
is high for all the categories.
Category | $\tau$-att | $\lambda$-sim | qasim
---|---|---|---
Entertainment (47) | 0.446 | 0.382 | 0.297
Health (62) | 0.483 | 0.419 | 0.354
Politics (143) | 0.45 | 0.300 | 0.272
Society (38) | 0.28 | 0.157 | 0.236
Family (20) | 0.6 | 0.350 | 0.40
Table 6. Precision$@$1 of models across categories | Temperature ($\tau$) ¿ 1.0
---|---
| 10 | 100 | 1000 | 10 | 100 | 1000
Src+ Ext | Prec$@$1 | Recall$@$3
Yahoo! + TREC | 0.46 | 0.38 | 0.38 | 0.73 | 0.644 | 0.64
Yahoo! + WHO | 0.37 | 0.38 | 0.36 | 0.64 | 0.65 | 0.64
Infobot + TREC | 0.44 | 0.59 | 0.39 | 0.72 | 0.81 | 0.75
Infobot + WHO | 0.65 | 0.41 | 0.44 | 0.85 | 0.76 | 0.79
Table 7. Variation in P$@$1 and R$@$3 across different temperature values.
(a) Yahoo!+TREC
(b) Yahoo!+WHO
(c) Infobot + TREC
(d) Infobot + WHO
Figure 4. Temperature and $\lambda$ variation impact on Prec$@$1
##### RQ2: How does temperature ($\tau$) compare with $\lambda$ parameter?
We argued that linearly combining similarities between question-answer in
primary dataset and between question-external source may not be sufficient to
boost performance. We observe that in our results too i.e. $\lambda$-sim
models do not perform better than (_$\tau$ -att_) models. This clearly
indicates that more sophisticated models can learn to combine this information
directly from training data. However, our experiments indicate that optimal
value of (_$\tau$_) varies across primary datasets and external sources of
information. For instance, (_$\tau$ -att_) model performed best when
_$\tau=0.4$_ and _$\tau=0.9$_ for Yahoo! ans and Infobot dataset respectively
when TREC was used as external source. It performed best when _$\tau=0.1$_ and
_$\tau=0.5$_ for Yahoo! ans and Infobot dataset respectively when WHO was used
as external source. We also tried to vary _$\tau$_ beyond 1.0 to determine
whether it yielded a trend as shown in Table 7. Higher values of temperature
seem to degrade model performance. We found that optimal temperature range is
between $[0.1-1]$. Existing research in model distillation
(hinton2015distilling, ) has also empirically found that lower values of
temperature yield better performance.
We also compared model performance in terms of precision when $\lambda$ and
$\tau$ are varied for $\lambda$-sim models and temperature based models
respectively as shown in Figure 4. Temperature based models peak at one value
but do not have a clear trend indicating that one needs to explore different
$\tau$ values at the time of training for better performance. On the other
hand, we observe that adding external information also helps the $\lambda$-sim
models until a certain threshold. Overall, both sets of models show that free-
text external information can be incorporated to improve answer ranking
performance.
##### RQ3: What kind of queries/questions does the model attend to when
ranking relevant/non-relevant answers?
Attention based models have a very unique feature: they can aid explaining the
internal workings of neural network models. We inspect what kind of
queries/questions in external datasets does our model pay attention to while
ranking relevant or non-relevant answers. Figure 5 shows one such example of
Yahoo! question and incorporation of TREC data. At the time of scoring
relevant answer, the model gives higher weight to some queries compared to
others. In the example, for instance, it assigns more weight to queries
associated with masks or COVID virus response to weather changes. We observe
higher attention weights for questions when relevant answers are ranked than
when non-relevant answers are scored. An example question, a relevant and non-
relevant answer along with model attention weights on TREC queries are shown
from the Infobot data in Figure 6 respectively. It shows a similar trend where
attention weights are high for external queries that are closely associated
with the question answer text.
Overall, our experiments show that curated external information is useful for
improving community question answering task. Our experiments also indicate
that this external knowledge need not always be structured text. However, it
is worth noting that curated and reliable external sources may not always be
available for all domains. We addressed a very niche task in this work, and
further research is required to extend it to incorporate multiple external
sources. We posit that with scalable attention mechanisms, this work can be
easily made tractable for large external sources containing thousands or
millions of entries in the future.
Figure 5. Y! ques, its rel and non-rel ans and questions with $\tau$-att
model’s attention values for TREC queries. Figure 6. Infobot ques, its rel and
non-rel ans and questions with $\tau$-att model’s attention values for TREC
queries.
## 6\. Conclusion
Question answering platforms provide users with effective and easy access to
information. These platforms also provide content on rapidly evolving
_sensitive_ topics such as disease outbreaks (such as COVID-19) where it is
also useful to use external _vetted_ information for ranking answers. Existing
work only exploits knowledge bases which have some limitations that makes it
difficult to use them for community Q&A for rapidly evolving topics such as
wild-fires or earthquakes. In this work, we tried to evaluate the
effectiveness of external (free text or semi-structured) information in
improving answer ranking models. We argue that simple question-answer text
matching may be insufficient and in presence of external knowledge, but
temperature regulated attention models can distill information better which in
turn yields higher performance. Our proposed model with temperature regulated
attention, when evaluated on two public datasets showed significant
improvements by augmenting information from two _external_ curated sources of
information. In future, we aim to expand these experiments to other categories
such as disaster relief and scale the attention mechanism to include multiple
external sources in one model.
## References
* [1] A. Cohan, S. Feldman, I. Beltagy, D. Downey, and D. S. Weld. Specter: Document-level representation learning using citation-informed transformers. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pages 2270–2282, 2020.
* [2] D. Cohen and W. Croft. End to end long short term memory networks for non-factoid question answering. pages 143–146, 09 2016.
* [3] D. H. Dalip, M. A. Gonçalves, M. Cristo, and P. Calado. Exploiting user feedback to learn to rank answers in q&a forums: A case study with stack overflow. In Proceedings of the 36th International ACM SIGIR Conference on Research and Development in Information Retrieval, SIGIR ’13, page 543–552, New York, NY, USA, 2013. Association for Computing Machinery.
* [4] J. Devlin, M.-W. Chang, K. Lee, and K. Toutanova. Bert: Pre-training of deep bidirectional transformers for language understanding, 2019.
* [5] G. Hinton, O. Vinyals, and J. Dean. Distilling the knowledge in a neural network, 2015.
* [6] L. Hong and B. D. Davison. A classification-based approach to question answering in discussion boards. In Proceedings of the 32nd international ACM SIGIR conference on Research and development in information retrieval, pages 171–178, 2009.
* [7] J. Hu, S. Qian, Q. Fang, and C. Xu. Attentive interactive convolutional matching for community question answering in social multimedia. In Proceedings of the 26th ACM International Conference on Multimedia, MM ’18, page 456–464, New York, NY, USA, 2018. Association for Computing Machinery.
* [8] J. Hu, S. Qian, Q. Fang, and C. Xu. Hierarchical graph semantic pooling network for multi-modal community question answer matching. In Proceedings of the 27th ACM International Conference on Multimedia, MM ’19, page 1157–1165, New York, NY, USA, 2019. Association for Computing Machinery.
* [9] H. Huang, X. Wei, L. Nie, X. Mao, and X.-S. Xu. From question to text: Question-oriented feature attention for answer selection. ACM Transactions on Information Systems, 37:1–33, 10 2018.
* [10] B. Kratzwald, A. Eigenmann, and S. Feuerriegel. Rankqa: Neural question answering with answer re-ranking. CoRR, abs/1906.03008, 2019.
* [11] P. Nakov, D. Hoogeveen, L. Màrquez, A. Moschitti, H. Mubarak, T. Baldwin, and K. Verspoor. Semeval-2017 task 3: Community question answering. arXiv preprint arXiv:1912.00730, 2019.
* [12] L. Nie, X. Wei, D. Zhang, X. Wang, Z. Gao, and Y. Yang. Data-driven answer selection in community qa systems. IEEE transactions on knowledge and data engineering, 29(6):1186–1198, 2017.
* [13] A. Poliak, M. Fleming, C. Costello, K. W. Murray, M. Yarmohammadi, S. Pandya, D. Irani, M. Agarwal, U. Sharma, S. Sun, et al. Collecting verified covid-19 question answer pairs. 2020\.
* [14] A. Rücklé and I. Gurevych. Representation learning for answer selection with LSTM-based importance weighting. In IWCS 2017 — 12th International Conference on Computational Semantics — Short papers, 2017.
* [15] A. Severyn and A. Moschitti. Structural relationships for large-scale learning of answer re-ranking. In Proceedings of the 35th international ACM SIGIR conference on Research and development in information retrieval, pages 741–750, 2012.
* [16] Y. Shen, Y. Deng, M. Yang, Y. Li, N. Du, W. Fan, and K. Lei. Knowledge-aware attentive neural network for ranking question answer pairs. In The 41st International ACM SIGIR Conference on Research & Development in Information Retrieval, SIGIR ’18, page 901–904, New York, NY, USA, 2018. Association for Computing Machinery.
* [17] Y. Shen, W. Rong, Z. Sun, Y. Ouyang, and Z. Xiong. Question/answer matching for cqa system via combining lexical and sequential information. In Proceedings of the Twenty-Ninth AAAI Conference on Artificial Intelligence, AAAI’15, page 275–281. AAAI Press, 2015.
* [18] D. Su, Y. Xu, T. Yu, F. B. Siddique, E. J. Barezi, and P. Fung. CAiRE-COVID: A question answering and multi-document summarization system for covid-19 research. arXiv 2005.03975, 2020.
* [19] M. Surdeanu, M. Ciaramita, and H. Zaragoza. Learning to rank answers on large online qa collections. In ACL, 2008.
* [20] E. Voorhees, T. Alam, S. Bedrick, D. Demner-Fushman, W. R. Hersh, K. Lo, K. Roberts, I. Soboroff, and L. L. Wang. Trec-covid: Constructing a pandemic information retrieval test collection, 2020.
* [21] B. Wang, X. Wang, C.-J. Sun, B. Liu, and L. Sun. Modeling semantic relevance for question-answer pairs in web social communities. In Proceedings of the 48th Annual Meeting of the Association for Computational Linguistics, pages 1230–1238, 2010.
* [22] L. Yang, Q. Ai, D. Spina, R.-C. Chen, L. Pang, W. B. Croft, J. Guo, and F. Scholer. Beyond factoid qa: Effective methods for non-factoid answer sentence retrieval. In European Conference on Information Retrieval, pages 115–128. Springer, 2016.
* [23] X. Yang, M. Khabsa, M. Wang, W. Wang, A. H. Awadallah, D. Kifer, and C. L. Giles. Adversarial training for community question answer selection based on multi-scale matching. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 33, pages 395–402, 2019.
* [24] X. Zhang, S. Li, L. Sha, and H. Wang. Attentive interactive neural networks for answer selection in community question answering. In Proceedings of the Thirty-First AAAI Conference on Artificial Intelligence, AAAI’17, page 3525–3531. AAAI Press, 2017.
* [25] Y. Zhang, S. Qian, Q. Fang, and C. Xu. Multi-modal knowledge-aware hierarchical attention network for explainable medical question answering. In Proceedings of the 27th ACM International Conference on Multimedia, MM ’19, page 1089–1097, New York, NY, USA, 2019. Association for Computing Machinery.
* [26] X. Zhou, B. Hu, Q. Chen, B. Tang, and X. Wang. Answer sequence learning with neural networks for answer selection in community question answering. arXiv preprint arXiv:1506.06490, 2015.
|
# Gain distance matrices for complex unit gain graphs
Aniruddha Samanta Department of Mathematics, Indian Institute of Technology
Kharagpur, Kharagpur 721302, India. Email<EMAIL_ADDRESS>M. Rajesh
Kannan Department of Mathematics, Indian Institute of Technology Kharagpur,
Kharagpur 721302, India. Email<EMAIL_ADDRESS><EMAIL_ADDRESS>
###### Abstract
A complex unit gain graph ($\mathbb{T}$-gain graph), $\Phi=(G,\varphi)$ is a
graph where the function $\varphi$ assigns a unit complex number to each
orientation of an edge of $G$, and its inverse is assigned to the opposite
orientation. In this article, we propose gain distance matrices for
$\mathbb{T}$-gain graphs. These notions generalize the corresponding known
concepts of distance matrices and signed distance matrices. Shahul K. Hameed
et al. introduced signed distance matrices and developed their properties.
Motivated by their work, we establish several spectral properties, including
some equivalences between balanced $\mathbb{T}$-gain graphs and gain distance
matrices. Furthermore, we introduce the notion of positively weighted
$\mathbb{T}$-gain graphs and study some of their properties. Using these
properties, Acharya’s and Stanić’s spectral criteria for balance are deduced.
Moreover, the notions of order independence and distance compatibility are
studied. Besides, we obtain some characterizations for distance compatibility.
Mathematics Subject Classification(2010): 05C22(primary); 05C50,
05C35(secondary).
Keywords. Complex unit gain graph, Signed distance matrix, Distance matrix,
Adjacency matrix, Hadamard product of matrices.
## 1 Introduction
Let $\Phi=(G,\varphi)$ be a connected complex unit gain graph
($\mathbb{T}$-gain graph) on a simple graph $G$ with $n$ vertices. Let
$V(G)=\\{v_{1},v_{2},\dots,v_{n}\\}$ and $E(G)$ be the vertex set and the edge
set of $G$, respectively. If two vertices $v_{i}$ and $v_{j}$ are connected by
an edge, then we write $v_{i}\sim v_{j}$. If $v_{i}\sim v_{j}$, then the edge
between them is denoted by $e_{i,j}$. The _adjacency matrix_ $A(G)$ of a graph
$G$ is a symmetric matrix whose $(i,j)th$ entry is $1$ if $v_{i}\sim v_{j}$
and zero otherwise. A _path_ $P$ in $G$ between the vertices $s$ and $t$ is
denoted by $sPt$. The _distance_ between two vertices $s$ and $t$ in $G$ is
the length of the shortest path between $s$ and $t$, and is denoted by
$d_{G}(s,t)$ (or simply $d(s,t)$). The _distance matrix_ of an undirected
graph $G$, denoted by $D(G)$, is the symmetric $n\times n$ matrix whose
$(i,j)$th entry is $d(v_{i},v_{j})$. The distance matrix of an undirected
graph has been widely studied in the literature, see [2, 3, 4, 5] and the
references therein.
Let $G$ be a simple undirected graph. An oriented edge from the vertex $v_{s}$
to the vertex $v_{t}$ is denoted by $\overrightarrow{e}_{s,t}$. For each
undirected edge $e_{s,t}\in E(G)$, there is a pair of oriented edges
$\overrightarrow{e}_{s,t}$ and $\overrightarrow{e}_{t,s}$. The collection
$\overrightarrow{E}(G):=\\{\overrightarrow{e}_{s,t},\overrightarrow{e}_{t,s}:e_{s,t}\in
E(G)\\}$ is the _oriented edge set associated with $G$_. Let
$\mathbb{T}=\\{z\in\mathbb{C}:|z|=1\\}$. A _complex unit gain graph (or
$\mathbb{T}$-gain graph)_ on a simple graph $G$ is an ordered pair
$(G,\varphi)$, where the gain function
$\varphi:\overrightarrow{E}(G)\rightarrow\mathbb{T}$ is a mapping such that
$\varphi(\overrightarrow{e}_{s,t})=\varphi(\overrightarrow{e}_{t,s})^{-1}$,
for every $e_{s,t}\in E(G)$. A $\mathbb{T}$-gain graph $(G,\varphi)$ is
denoted by $\Phi$. The _adjacency matrix_ of a $\mathbb{T}$-gain graph
$\Phi=(G,\varphi)$ is a Hermitian matrix, denoted by $A(\Phi)$ and its
$(s,t)th$ entry is defined as follows:
$a_{st}=\begin{cases}\varphi(\overrightarrow{e}_{s,t})&\text{if
}\mbox{$v_{s}\sim v_{t}$},\\\ 0&\text{otherwise.}\end{cases}$
The spectrum and the spectral radius of $\Phi$ are the spectrum and the
spectral radius of $A(\Phi)$ and denoted by $\operatorname{spec}(\Phi)$ and
$\rho(\Phi)$, respectively. A _signed graph_ is a graph $G$ together with a
signature function $\psi:E(G)\rightarrow\\{\pm 1\\}$, and is denoted by
$\Psi=(G,\psi)$. The adjacency matrix of $\Psi$, denoted by $A(\Psi)$, is an
$n\times n$ matrix whose $(i,j)$th entry is $\psi(e_{i,j})$. Therefore, a
signed graph can be considered as a $\mathbb{T}$-gain graph $\Psi=(G,\psi)$,
where $\psi$ is a signature function. The notion of adjacency matrix of
$\mathbb{T}$-gain graphs generalize the notion of adjacency matrix of
undirected graphs, adjacency matrix of signed graphs and the Hermitian
adjacency matrix of a mixed graph. For more information about the properties
of gain graphs and $\mathbb{T}$-gain graphs, we refer to [11, 12, 16, 17].
Let $\Psi=(G,\psi)$ be a signed graph. The sign of a path in $\Psi$ is the
product of sign of all edges of the path [15]. Recently, in [6] the authors
introduced the notion of signed distance matrices $D^{\max}(\Psi)$ and
$D^{\min}(\Psi)$ for a signed graph $\Psi$.
###### Definition 1.1 ([6, Definition 1.1]).
Let $\Psi=(G,\psi)$ be a signed graph with vertex set
$V(G)=\\{v_{1},v_{2},\dots,v_{n}\\}$. Then two auxiliary signs are defined as
follows:
* (a)
$\psi_{\max}(v_{i},v_{j})=-1$ if all shortest $v_{i}v_{j}$-paths are negative,
$+1$ otherwise,
* (b)
$\psi_{\min}(v_{i},v_{j})=+1$ if all shortest $v_{i}v_{j}$-paths are positive,
$-1$ otherwise.
The two signed distance matrices are defined as follows:
* (a)
$D^{\max}(\Psi)=(d_{\max}(v_{i},v_{j}))_{n\times n}$,
* (b)
$D^{\min}(\Psi)=(d_{\min}(v_{i},v_{j}))_{n\times n}$,
where $d_{\max}(v_{i},v_{j})=\psi_{\max}(v_{i},v_{j})d(v_{i},v_{j})$ and
$d_{\min}(v_{i},v_{j})=\psi_{\min}(v_{i},v_{j})d(v_{i},v_{j})$.
A signed graph $\Psi$ is distance compatible if and only if
$D^{\max}(\Psi)=D^{\min}(\Psi)$. A characterization of balanced signed graph
in terms of signed distance matrices is obtained in [6]. For more about signed
distance matrices, see [6, 13].
In this article, we introduce the notion of gain distance matrices
$D^{\max}_{<}(\Phi)$ and $D^{\min}_{<}(\Phi)$ for a $\mathbb{T}$-gain graph
$\Phi=(G,\varphi)$ associated with an ordered vertex set $(V(G),<)$. These
concepts generalize the notions of signed distance matrices of signed graphs
and distance matrices of undirected graphs. We define positively weighted
$\mathbb{T}$-gain graphs and establish two new characterizations for balance
of gain graphs. Acharya’s Spectral criterion and Stanić’s spectral criterion
are particular cases of these characterizations. Besides, we introduce two
properties of a $\mathbb{T}$-gain graph, ordered-independence, and distance
compatibility to gain distance matrices. Thereupon we establish two
characterizations for the balance of $\mathbb{T}$-gain graphs in terms of gain
distance matrices and distance compatibility properties. Subsequently, we
present some results on the characterization of distance compatibility.
This paper is organized as follows: In section 2, we collect needed known
definitions and results. In section 3, we define the notion of gain distance
matrices, order-independent and distance compatibility, and discuss their
properties. In section 4, we discuss the positively weighted $\mathbb{T}$-gain
graphs and establish two spectral characterizations for the balance(Theorem
4.2, Theorem 4.3). In section 5, we derive two characterizations for balance
$\mathbb{T}$-gain graph in terms of the gain distance matrices (Theorem 5.2,
Theorem 5.3). In section 6, we obtain a couple of characterizations for
distance compatible $\mathbb{T}$-gain graphs (Theorem 6.1, Theorem 6.2,
Theorem 6.3).
## 2 Definitions, notation and preliminary results
Let $G=(V(G),E(G))$ be a connected undirected graph with no loops and multiple
edges, where $V(G)=\\{v_{1},v_{2},\dots,v_{n}\\}$ is the vertex set and $E(G)$
is the edge set of $G$. A graph $G$ is _geodetic_ , if there exists a unique
shortest path between any two vertices of $G$. Let $\Phi=(G,\varphi)$ be a
$\mathbb{T}$-gain graph on $G$. For $s,t\in V(G)$, $sPt$ denotes a path starts
at $s$ and ends at $t$ in $G$. In case of gain graph, $sPt$ denotes the
oriented path from the vertex $s$ to the vertex $t$. The gain of the path
$sPt$ is $\varphi(sPt)=\prod\limits_{j=1}^{k}\varphi(\overrightarrow{e_{j}})$,
where
$\overrightarrow{e_{1}},\overrightarrow{e_{2}},\dots,\overrightarrow{e_{k}}$
are the consecutive oriented edges in $sPt$. Therefore,
$\varphi(tPs)=\overline{\varphi(sPt)}$. The gain of an oriented cycle
$\overrightarrow{C_{n}}$ with edges
$\overrightarrow{e_{1}},\overrightarrow{e_{2}},\dots,\overrightarrow{e_{n}}$
is
$\varphi(\overrightarrow{C_{n}})=\prod\limits_{j=1}^{n}\varphi(\overrightarrow{e_{j}})$.
A cycle $C$ is _neutral_ in $\Phi$ if $\varphi(\overrightarrow{C})=1$. A gain
graph $\Phi$ is _balanced_ , if all cycles in $\Phi$ are neutral. A
$\mathbb{T}$-gain graph $\Phi$ is _anti-balanced_ if $-\Phi$ is balanced. Let
$\operatorname{Re}(x)$ and $\operatorname{Im}(x)$ denote the real and
imaginary part of a complex number $x$, respectively.
A function $\zeta:V(G)\rightarrow\mathbb{T}$ is a _switching function_. Let
$\Phi_{1}=(G,\varphi_{1})$ and $\Phi_{2}=(G,\varphi_{2})$ be two
$\mathbb{T}$-gain graphs. Then $\Phi_{1}$ and $\Phi_{2}$ are _switching
equivalent_ , denoted by $\Phi_{1}\sim\Phi_{2}$, if there exists a switching
function $\zeta$ such that
$\varphi_{1}(\overrightarrow{e}_{i,j})=\zeta(v_{i})^{-1}\varphi_{2}(\overrightarrow{e}_{i,j})\zeta(v_{j})$,
for all $e_{i,j}\in E(G)$. If $\Phi_{1}\sim\Phi_{2}$, then $A(\Phi_{1})$ and
$A(\Phi_{2})$ are diagonally similar and hence have the same spectra.
###### Lemma 2.1 ([12, Corollary 3.2]).
Let $\Phi_{1}$ and $\Phi_{2}$ be two $\mathbb{T}$-gain graphs on a connected
graph $G$ with a normal spanning tree $T$. Then $\Phi_{1}\sim\Phi_{2}$ if and
only if
$\varphi_{1}(\overrightarrow{C_{j}})=\varphi_{2}(\overrightarrow{C_{j}})$, for
all fundamental cycles $C_{j}$ with respect to $T$.
A signed graph is a $\mathbb{T}$-gain graph $\Psi=(G,\psi)$, where
$\psi(\overrightarrow{e}_{i,j})=1$ or $-1$ for $e_{i,j}\in E(G)$. The _sign_
of a path in $\Psi$ is the product of the signs (the gains) of the edges in
the path.
###### Theorem 2.1 (Harary’s path criterion [7] ).
Let $\Psi$ be a signed graph on an underlying graph $G$. Then $\Psi$ is
balanced if and only if any pair of vertices $s,t$, every $st$-path have the
same signature.
Let $\mathbb{C}^{m\times n}$ denote the set of all $m\times n$ matrices with
complex entries. For $A=(a_{ij})\in\mathbb{C}^{n\times n}$, define
$|A|=(|a_{ij}|)$. For two matrices $A=(a_{ij})$ and $B=(b_{ij})$, we write
$A\leq B$ if $a_{ij}\leq b_{ij}$ for all $i,j$. A matrix is _non-negative_ ,
if all entries of a matrix are non-negative. The spectral radius of a matrix
$A$ is denoted by $\rho(A)$.
###### Theorem 2.2 ([9, Theorem 8.4.5]).
Let $A,B\in\mathbb{C}^{n\times n}$. Suppose $A$ is irreducible and non-
negative and $A\geq|B|$. Let $\mu=e^{i\theta}\rho(B)$ be a given maximum
modulus eigenvalue of $B$. If $\rho(A)=\rho(B)$, then there is a unitary
diagonal matrix $D$ such that $B=e^{i\theta}DAD^{-1}$.
Let $A=(a_{ij}),B=(b_{ij})\in\mathbb{C}^{m\times n}$. The _Hadamard product_
of $A$ and $B$, denoted by $A\circ B$, is defined as $A\circ
B=(a_{ij}b_{ij})_{m\times n}$. For any three matrices $A,B,C$ of same order,
$(A\circ B)\circ C=A\circ(B\circ C)$. Let us recall the following property of
Hadamard product of matrices.
###### Proposition 2.1 ([8, Lemma 5.1.2]).
Let $A,B,C$ be three $n\times n$ matrices and $D,E$ be two $n\times n$
diagonal matrices. Then
$D(A\circ B)E=(DAE)\circ B=(DA)\circ(BE)=(AE)\circ(DB)=A\circ(DBE).$
## 3 Gain distance matrices
This section introduces the notion of gain distance matrices of
$\mathbb{T}$-gain graphs, which generalize the notion of distance matrices of
undirected graphs and signed distance matrices of signed graphs.
Let $\Phi=(G,\varphi)$ be a connected $\mathbb{T}$-gain graph on $G$. For
$s,t\in V(G)$, $sPt$ denotes the oriented path from the vertex $s$ to the
vertex $t$. Define three sets of paths
$\mathcal{P}(s,t),\mathcal{P}^{\max}(s,t)$ and $\mathcal{P}^{\min}(s,t)$ as
follows:
$\mathcal{P}(s,t)=\left\\{sPt:sPt\text{ is a shortest path}\right\\},$
$\mathcal{P}^{\max}(s,t)=\left\\{sPt\in\mathcal{P}(s,t):\operatorname{Re}(\varphi(sPt))=\max\limits_{s\tilde{P}t\in\mathcal{P}(s,t)}\operatorname{Re}(\varphi(s\tilde{P}t))\right\\}$
and
$\mathcal{P}^{\min}(s,t)=\left\\{sPt\in\mathcal{P}(s,t):\operatorname{Re}(\varphi(sPt))=\min\limits_{s\tilde{P}t\in\mathcal{P}(s,t)}\operatorname{Re}(\varphi(s\tilde{P}t))\right\\}.$
Note that $\mathcal{P}^{\max}(s,t)=\mathcal{P}^{\max}(t,s)$ and
$\mathcal{P}^{\min}(s,t)=\mathcal{P}^{\min}(t,s)$.
Let $G$ be a simple graph with vertex set
$V(G)=\\{v_{1},v_{2},\dots,v_{n}\\}$. We denote $\left(V(G),<\right)$ as an
ordered vertex set, where $`<`$ is a total ordering of the vertices of $G$. An
ordering $`<_{r}`$ is the _reverse ordering_ of $`<`$ if $v_{i}<_{r}v_{j}$ if
and only if $v_{j}<v_{i}$, for any $i,j$. An ordering $`<`$ is the _standard
vertex ordering_ if $v_{1}<v_{2}<\dots<v_{n}$.
###### Definition 3.1 (Auxiliary gains).
Let $\Phi=(G,\varphi)$ be a $\mathbb{T}$-gain graph with an ordered vertex set
$(V(G),<)$. We define two auxiliary gains with respect to $<$ as follows.
1. (1)
The function $\varphi^{<}_{\max}:V(G)\times V(G)\rightarrow\mathbb{T}$ is the
maximum auxiliary gain with respect to the vertex order $<$ such that
$\varphi^{<}_{\max}(s,t)=\overline{\varphi^{<}_{\max}(t,s)}$ for each
$(s,t)\in V(G)\times V(G)$ and $\varphi^{<}_{\max}$ is defined by
$\varphi^{<}_{\max}(s,t)=\varphi(sPt)$
where $s<t$, and $sPt\in\mathcal{P}^{\max}(s,t)$ and
$\operatorname{Im}(\varphi(sPt))=\max\limits_{s\tilde{P}t\in\mathcal{P}^{\max}(s,t)}\operatorname{Im}(\varphi(s\tilde{P}t)).$
2. (2)
The function $\varphi^{<}_{\min}:V(G)\times V(G)\rightarrow\mathbb{T}$ is the
minimum auxiliary gain with respect to the vertex order $<$ such that
$\varphi^{<}_{\min}(s,t)=\overline{\varphi^{<}_{\min}(t,s)}$ for each
$(s,t)\in V(G)\times V(G)$ and $\varphi^{<}_{\min}$ is defined by
$\varphi^{<}_{\min}(s,t)=\varphi(sPt)$
where $s<t$, and $sPt\in\mathcal{P}^{\min}(s,t)$ and
$\operatorname{Im}(\varphi(sPt))=\min\limits_{s\tilde{P}t\in\mathcal{P}^{\min}(s,t)}\operatorname{Im}(\varphi(s\tilde{P}t)).$
Note that, for $s<t$,
$\varphi^{<}_{\max}(s,t)(\mbox{~{}resp.,~{}}\varphi^{<}_{\min}(s,t))$ is the
maximum (resp., minimum) gain, with respect to the lexicographic order, over
all the shortest paths between the vertices $s$ and $t$.
###### Definition 3.2 (Gain distances).
Let $\Phi=(G,\varphi)$ be a $\mathbb{T}$-gain graph with an ordered vertex set
$(V(G),<)$. For any two vertices $s,t\in V(G)$, there are two gain distances
from the vertex $s$ to the vertex $t$ which are defined as follows:
1. (1)
$d^{<}_{\max}(s,t)=\varphi^{<}_{\max}(s,t)d(s,t),$
2. (2)
$d^{<}_{\min}(s,t)=\varphi^{<}_{\min}(s,t)d(s,t).$
Next we define the gain distance matrices for $\mathbb{T}$-gain graphs.
###### Definition 3.3 (Gain distance matrices).
Let $\Phi=(G,\varphi)$ be a $\mathbb{T}$-gain graph with an order $<$ on the
vertex set $V(G)$, where $V(G)=\\{v_{1},v_{2},\cdots,v_{n}\\}$. The gain
distance matrices $D_{<}^{\max}(\Phi)$ and $D_{<}^{\min}(\Phi)$ associated
with $<$ are defined as follows:
1. (1)
$D_{<}^{\max}(\Phi)=\left(d^{<}_{\max}(v_{i},v_{j})\right)$,
2. (2)
$D_{<}^{\min}(\Phi)=\left(d^{<}_{\min}(v_{i},v_{j})\right)$.
Here $d^{<}_{\max}(v_{i},v_{j})$ is the $(i,j)$th entry of
$D_{<}^{\max}(\Phi).$
The gain distance matrices are the generalization of the distance matrix of an
undirected graph and signed distance matrices of a signed graph. It is easy to
see that the gain distance matrices $D^{\max}_{<}(\Phi)$ and
$D^{\min}_{<}(\Phi)$ are Hermitian, and hence have real eigenvalues. For any
pair of vertices $(v_{s},v_{t})$, there are two different maximum gain
distances $d^{<}_{\max}(v_{s},v_{t})$ and $d^{<}_{\max}(v_{t},v_{s})$ which
have same absolute value but they are the complex conjugate to each other. The
distance $d^{<}_{\max}(v_{s},v_{t})$ is the maximum gain distance from the
vertex $v_{s}$ to the vertex $v_{t}$ in $\Phi$ with respect to the vertex
ordering $(V(G),<)$. Likewise, the minimum gain distance is defined.
Now we illustrate the definitions with the following example.
Figure 1: $\mathbb{T}$-gain graphs $\Phi=(G,\varphi)$
###### Example 3.1.
Let $\Phi=(G,\varphi)$ be a $\mathbb{T}$-gain graph shown in Figure 1. Let us
consider the standard order $<$ on vertex set $V(G)$, where
$V(G)=\\{v_{1},v_{2},v_{3},v_{4},v_{5}\\}$. Then
$D^{\max}_{<}(\Phi)=\left[\begin{array}[]{ccccc}0&1&2e^{\frac{i\pi}{6}}&1&e^{\frac{i\pi}{3}}\\\
1&0&e^{\frac{i\pi}{6}}&1&2e^{\frac{i\pi}{3}}\\\
2e^{-\frac{i\pi}{6}}&e^{-\frac{i\pi}{6}}&0&e^{\frac{i\pi}{6}}&3e^{\frac{i\pi}{6}}\\\
1&1&e^{-\frac{i\pi}{6}}&0&2e^{\frac{i\pi}{3}}\\\
e^{-\frac{i\pi}{3}}&2e^{-\frac{i\pi}{3}}&3e^{-\frac{i\pi}{6}}&2e^{-\frac{i\pi}{3}}&0\end{array}\right].$
Now, consider the reverse ordering $`<_{r}`$ of the standard order $`<`$. Then
$D^{\max}_{<_{r}}(\Phi)=\left[\begin{array}[]{ccccc}0&1&2e^{-\frac{i\pi}{6}}&1&e^{\frac{i\pi}{3}}\\\
1&0&e^{\frac{i\pi}{6}}&1&2e^{\frac{i\pi}{3}}\\\
2e^{\frac{i\pi}{6}}&e^{-\frac{i\pi}{6}}&0&e^{\frac{i\pi}{6}}&3e^{-\frac{i\pi}{6}}\\\
1&1&e^{-\frac{i\pi}{6}}&0&2e^{\frac{i\pi}{3}}\\\
e^{-\frac{i\pi}{3}}&2e^{-\frac{i\pi}{3}}&3e^{\frac{i\pi}{6}}&2e^{-\frac{i\pi}{3}}&0\end{array}\right].$
Here $D^{\max}_{<}(\Phi)\neq D^{\max}_{<_{r}}(\Phi)$. In fact,
$\operatorname{spec}(D^{\max}_{<}(\Phi))\neq\operatorname{spec}(D^{\max}_{<_{r}}(\Phi))$.
Similarly,$D^{\min}_{<}(\Phi)\neq D^{\min}_{<_{r}}(\Phi)$.
###### Definition 3.4 (Vertex order independent).
A $\mathbb{T}$-gain graph $\Phi=(G,\varphi)$ is vertex order-independent
(simply, order independent), if $D^{\max}_{<}(\Phi)=D^{\max}_{<_{r}}(\Phi)$
and $D^{\min}_{<}(\Phi)=D^{\min}_{<_{r}}(\Phi)$, where $<$ is the standard
vertex order on $V(G)$. In this case, we
define$D^{\max}(\Phi)=D^{\max}_{<}(\Phi)=D^{\max}_{<_{r}}(\Phi)$ and
$D^{\min}(\Phi)=D^{\min}_{<}(\Phi)=D^{\min}_{<_{r}}(\Phi)$.
Now we present a characterization for order independent $\mathbb{T}$-gain
graph.
###### Theorem 3.1.
Let $\Phi=(G,\varphi)$ be a $\mathbb{T}$-gain graph. Then $\Phi$ is not order
independent if and only if at least one of the following holds.
1. (i)
There exists $v_{s},v_{t}\in V(G)$ with at least two shortest paths from
$v_{s}$ to $v_{t}$ in $\mathcal{P}^{\max}(v_{s},v_{t})$ have different gains.
2. (ii)
There exists $v_{s},v_{t}\in V(G)$ with at least two shortest paths from
$v_{s}$ to $v_{t}$ in $\mathcal{P}^{\min}(v_{s},v_{t})$ have different gains.
###### Proof.
Let $\Phi=(G,\varphi)$ be a $\mathbb{T}$-gain graph with vertex set
$V(G)=\\{v_{1},v_{2},\cdots,v_{n}\\}$. Let $<$ be the standard vertex order on
$V(G)$. Suppose $\Phi$ is not order independent, then either
$D^{\max}_{<}(\Phi)\neq D^{\max}_{<_{r}}(\Phi)$ or $D^{\min}_{<}(\Phi)\neq
D^{\min}_{<_{r}}(\Phi)$ hold. Suppose $D^{\max}_{<}(\Phi)\neq
D^{\max}_{<_{r}}(\Phi)$. Then there exists $v_{s},v_{t}\in V(G)$ such that
$d_{\max}^{<}(v_{s},v_{t})\neq d_{\max}^{<_{r}}(v_{s},v_{t})$. Then
$\varphi_{\max}^{<}(v_{s},v_{t})\neq\varphi_{\max}^{<_{r}}(v_{s},v_{t})$. Let
$v_{s}<v_{t}$ and
$\varphi_{\max}^{<}(v_{s},v_{t})=\varphi(v_{s}P_{1}v_{t})=x+iy\in\mathbb{T}$,
for some $v_{s}P_{1}v_{t}\in\mathcal{P}^{\max}(v_{s},v_{t})$. It is clear that
$y\neq 0$. Now $v_{t}<_{r}v_{s}$ and
$\varphi_{\max}^{<_{r}}(v_{t},v_{s})=\varphi(v_{t}P_{2}v_{s})$, for some
$v_{t}P_{2}v_{s}\in\mathcal{P}^{\max}(v_{s},v_{t})$. Then
$\varphi_{\max}^{<_{r}}(v_{s},v_{t})=\overline{\varphi_{\max}^{<_{r}}(v_{t},v_{s})}=\overline{\varphi(v_{t}P_{2}v_{s})}=\varphi(v_{s}P_{2}v_{t})$.
Since $v_{t}P_{2}v_{s}\in\mathcal{P}^{\max}(v_{s},v_{t})$, so
$\varphi(v_{s}P_{2}v_{t})=x-iy_{1}\in\mathbb{T}$, where either $y_{1}=y$ or
$y_{1}=-y$. Also,
$\varphi_{\max}^{<}(v_{s},v_{t})\neq\varphi_{\max}^{<_{r}}(v_{s},v_{t})$, so
$\varphi(v_{s}P_{1}v_{t})\neq\varphi(v_{s}P_{2}v_{t})$. Thus
$\varphi(v_{s}P_{2}v_{t})=x-iy$ and $y>0$. Hence $(i)$ holds. Similarly,
$D^{\min}_{<}(\Phi)\neq D^{\min}_{<_{r}}(\Phi)$ implies $(ii)$.
Conversely, suppose statement $(i)$ holds. Then there exist two shortest
$\overrightarrow{v_{s}v_{t}}$-paths $v_{s}P_{1}v_{t}$ and $v_{s}P_{2}v_{t}$ in
$\mathcal{P}^{\max}(v_{s},v_{t})$ with different gains. If
$\varphi(v_{s}P_{1}v_{t})=x+iy\in\mathbb{T}$, then
$\varphi(v_{s}P_{2}v_{t})=x-iy$, $y\neq 0$. Without loss of generality, assume
that $y>0$. If $v_{s}<v_{t}$, then $\varphi_{\max}^{<}(v_{s},v_{t})=x+iy$ and
$\varphi_{\max}^{<_{r}}(v_{s},v_{t})=\overline{\varphi_{\max}^{<_{r}}(v_{t},v_{s})}=\overline{x+iy}=x-iy$.
Thus $\varphi_{\max}^{<}(v_{s},v_{t})\neq\varphi_{\max}^{<_{r}}(v_{s},v_{t})$.
Therefore $D^{\max}_{<}(\Phi)\neq D^{\max}_{<_{r}}(\Phi)$ and hence $\Phi$ is
not order independent. Similarly if the statement $(ii)$ holds, then $\Phi$ is
not order independent. ∎
###### Proposition 3.1.
Let $\Phi=(G,\varphi)$ be a $\mathbb{T}$-gain graph and $`<`$ be the standard
vertex order. Then $D_{<}^{\max}(\Phi)=D_{<}^{\min}(\Phi)$ if and only if
$D^{\max}(\Phi)$ and $D^{\min}(\Phi)$ are well defined and
$D^{\max}(\Phi)=D^{\min}(\Phi)$.
###### Proof.
Suppose $D_{<}^{\max}(\Phi)=D_{<}^{\min}(\Phi)$. Let $s,t\in V(G)$ such that
$s<t$. Then $d^{<}_{\max}(s,t)=d^{<}_{\min}(s,t)$. Therefore
$\varphi^{\max}_{<}(s,t)=\varphi^{\min}_{<}(s,t)$. Thus all the shortest paths
from $s$ to $t$ have the same gain. Therefore,
$\varphi^{\max}_{<}(s,t)=\varphi^{\max}_{<_{r}}(s,t)$ and
$\varphi^{\min}_{<}(s,t)=\varphi^{\min}_{<_{r}}(s,t)$. Thus $D^{\max}(\Phi)$
and $D^{\min}(\Phi)$ are well defined. Since
$d^{<}_{\max}(s,t)=d^{<}_{\min}(s,t)$, so $D^{\max}(\Phi)=D^{\min}(\Phi)$. The
converse is easy to verify. ∎
###### Theorem 3.2.
Let $\Phi=(G,\varphi)$ be a $\mathbb{T}$-gain graph and $`<`$ be the standard
vertex order. Let $<_{a}$ be any vertex order on $V(G)$. Then
$D^{\max}_{<}(\Phi)=D^{\min}_{<}(\Phi)$ if and only if
$D^{\max}_{<_{a}}(\Phi)=D^{\min}_{<_{a}}(\Phi)$.
###### Proof.
Let $<$ be the standard vertex order, and
$D^{\max}_{<}(\Phi)=D^{\min}_{<}(\Phi)$. Let $v_{i},v_{j}\in V(G)$. Then
$d^{<}_{\max}(v_{i},v_{j})=d^{<}_{\min}(v_{i},v_{j})$ and
$\varphi^{<}_{\max}(v_{i},v_{j})=\varphi^{<}_{\min}(v_{i},v_{j})$. Thus all
the shortest paths from $v_{i}$ to $v_{j}$ have the same gain. Therefore, for
any arbitrary vertex ordering $<_{a}$, we have
$\varphi^{<_{a}}_{\max}(v_{i},v_{j})=\varphi^{<_{a}}_{\min}(v_{i},v_{j})$.
Hence $d^{<_{a}}_{\max}(v_{i},v_{j})=d^{<_{a}}_{\min}(v_{i},v_{j})$. Since
$v_{i}$ and $v_{j}$ are arbitrary, so
$D^{\max}_{<_{a}}(\Phi)=D^{\min}_{<_{a}}(\Phi)$.
Proof of the converse is similar to that of the previous part. ∎
###### Definition 3.5 (Distance compatible).
A $\mathbb{T}$-gain graph $\Phi=(G,\varphi)$ is called _gain distance
compatible (simply, distance compatible)_ if
$D^{\max}_{<}(\Phi)=D^{\min}_{<}(\Phi)$, where $<$ is the standard order. In
this case, we define $D(\Phi)=D^{\max}_{<}(\Phi)=D^{\min}_{<}(\Phi)$.
The proof of the following theorem is easy to verify.
###### Theorem 3.3.
For a $\mathbb{T}$-gain graph $\Phi=(G,\varphi)$, the following are
equivalent:
1. (1)
$\Phi$ is distance compatible.
2. (2)
$D(\Phi)=D^{\max}_{<}(\Phi)=D^{\min}_{<}(\Phi)=D^{\max}(\Phi)=D^{\min}(\Phi)$.
3. (3)
$D(\Phi)$ is well defined.
###### Proof.
$(1)\implies(2):$ Let $\Phi$ be distance compatible. Then, by the definition,
$D^{\max}_{<}(\Phi)=D^{\min}_{<}(\Phi)$ for standard vertex order $<$. Also
$D(\Phi)=D^{\max}_{<}(\Phi)=D^{\min}_{<}(\Phi)$. Now by Proposition 3.1,
$D^{\max}(\Phi)=D^{\min}(\Phi)$. Hence $D^{\max}(\Phi)=D^{\min}_{<}(\Phi)$ and
$D^{\min}(\Phi)=D^{\min}_{<}(\Phi)$.
$(2)\implies(3):$ By the definition of distance-compatible $\mathbb{T}$-gain
graph, $D(\Phi)$ exists.
$(3)\implies(1):$ If $D(\Phi)$ is well defined, then
$D(\Phi)=D^{\max}(\Phi)=D^{\min}(\Phi)$. Hence $\Phi$ is distance-compatible.
∎
###### Proposition 3.2.
Let $\Phi=(G,\varphi)$ be any distance compatible $\mathbb{T}$-gain graph. If
$\Phi\sim\Psi$, then $\Psi$ is distance compatible and
$\operatorname{spec}(D(\Phi))=\operatorname{spec}(D(\Psi))$.
###### Proof.
Let $\Phi=(G,\varphi)$ be a $\mathbb{T}$-gain graph with the standard vertex
order $<$. Let $s,t\in V(G)$. Since $\Phi$ is distance compatible, all
oriented shortest paths $sPt$ from $s$ to $t$ have the same gain. As
$\Phi\sim\Psi$, so there exists a switching function $\zeta$ such that
$\psi(sPt)=\zeta(s)^{-1}\varphi(sPt)\zeta(t)$, for any shortest path $sPt$.
For any shortest path $sPt$, $\varphi(sPt)$ is unique, so $\psi(sPt)$ is
unique. Thus $\psi_{\max}^{<}(s,t)=\psi_{\min}^{<}(s,t)$ and hence
$D^{\max}_{<}(\Psi)=D^{\min}_{<}(\Psi)$. That is, $\Psi$ is distance
compatible and $D(\Psi)$ is well defined. Let $d_{\psi}(s,t)$ and
$d_{\varphi}(s,t)$ be the unique gain distance from $s$ to $t$ in $\Psi$ and
$\Phi$, respectively. Then
$d_{\varphi}(s,t)=\zeta(s)^{-1}d_{\psi}(s,t)\zeta(t)$. Thus $D(\Phi)$ and
$D(\Psi)$ are similar and hence
$\operatorname{spec}(D(\Phi))=\operatorname{spec}(D(\Psi))$. ∎
Converse of the above statement holds for balanced $\mathbb{T}$ gain graph,
see Corollary 5.1.
###### Proposition 3.3.
Let $\Phi=(G,\varphi)$ be a $\mathbb{T}$-gain graph. Then
1. (1)
If $\Phi$ is a signed graph, then $\Phi$ is order-independent.
2. (2)
If $\Phi$ is balanced or anti-balanced, then $\Phi$ is order-independent.
3. (3)
If $\Phi$ is distance compatible, then $\Phi$ is order-independent.
4. (4)
If $\Phi$ is geodetic, then $\Phi$ is order-independent.
However, converse of the above statements need not be true in general, see
Example 3.2. The following result is an extension of the Harary’s path
criterion for $\mathbb{T}$-gain graphs.
###### Lemma 3.1.
Let $\Phi=(G,\varphi)$ be any $\mathbb{T}$-gain graph. Then $\Phi$ is balanced
if and only if every directed $(s,t)$-path have the same gain in $\Phi$, for
any two vertices $s,t$.
###### Proof.
Let $\Phi=(G,\varphi)$ be balanced. Suppose that the two oriented paths
$sP_{1}t$ and $sP_{2}t$ have different gains. That is
$\varphi(sP_{1}t)\neq\varphi(sP_{2}t)$. Then
$\varphi(sP_{1}t)\varphi(tP_{2}s)=\sum\limits_{j=1}^{k}\varphi(\overrightarrow{C_{j}})\neq
1$, where $C_{1},C_{2},\dots,C_{k}$ are the cycles formed by these two paths.
Therefore, there exist at least one cycle, say $C_{j}$ such that
$\varphi(\overrightarrow{C_{j}})\neq 1$ . Thus $\Phi$ is not balanced, a
contradiction. Converse is easy to verify. ∎
Let $\Phi=(G,\varphi)$ be any $\mathbb{T}$-gain graph. Then $\Phi$ is either
order-independent or order-dependent. If $\Phi$ is ordered-independent, then
$\Phi$ may or may not be balanced, anti-balanced, geodetic. If $\Phi$ is
ordered-dependent, then, by Proposition 3.3, $\Phi$ is neither balanced nor
anti-balanced nor geodetic. Therefore, any $\mathbb{T}$-gain graph
$\Phi=(G,\varphi)$ belongs to one of the following classes:
1. (A)
$\Phi$ is balanced or anti-balanced or geodetic and
$D^{\max}(\Phi)=D^{\min}(\Phi)$.
2. (B)
$\Phi$ is neither balanced nor anti-balanced nor geodetic and
$D^{\max}(\Phi)=D^{\min}(\Phi)$.
3. (C)
$\Phi$ is neither balanced nor anti-balanced nor geodetic and
$D^{\max}(\Phi)\neq D^{\min}(\Phi)$.
4. (D)
$\Phi$ is neither balanced nor anti-balanced nor geodetic and at least one of
$D^{\max}(\Phi)$ and $D^{\min}(\Phi)$ are not well defined.
Next we give some examples. Examples of Type $(A)$ can be constructed easily.
Example 3.1 is of Type $(D)$. Examples of Type $(B)$ and Type $(C)$ are given
below.
Figure 2: $\mathbb{T}$-gain graphs $\Phi_{1}$ and $\Phi_{2}$
###### Example 3.2.
Let us consider the $\mathbb{T}$-gain graph $\Phi_{1}=(G,\varphi_{1})$ (in
Figure 2) with the standard vertex order. The graph $\Phi_{1}$ is neither
balanced nor anti-balanced nor geodetic. Also $\Phi$ is order-independent. Now
$D^{\max}(\Phi_{1})=\left[\begin{array}[]{ccccc}0&1&1&2e^{\frac{i\pi}{6}}&e^{\frac{i\pi}{6}}\\\
1&0&1&2&1\\\ 1&1&0&e^{\frac{i\pi}{3}}&1\\\
2e^{-\frac{i\pi}{6}}&2&e^{-\frac{i\pi}{3}}&0&1\\\
e^{-\frac{i\pi}{6}}&1&1&1&0\end{array}\right],D^{\min}(\Phi_{1})=\left[\begin{array}[]{ccccc}0&1&1&2e^{\frac{i\pi}{3}}&e^{\frac{i\pi}{6}}\\\
1&0&1&2e^{\frac{i\pi}{3}}&1\\\ 1&1&0&e^{\frac{i\pi}{3}}&1\\\
2e^{-\frac{i\pi}{3}}&2e^{-\frac{i\pi}{3}}&e^{-\frac{i\pi}{3}}&0&1\\\
e^{-\frac{i\pi}{6}}&1&1&1&0\end{array}\right].$
Thus $D^{\max}(\Phi_{1})\neq D^{\min}(\Phi_{1})$. Therefore, by Theorem 3.3,
$\Phi_{1}$ is distance incompatible.
###### Example 3.3.
The $\mathbb{T}$-gain graph $\Phi_{2}$ (in Figure 2) with the standard vertex
ordering is neither balanced nor anti-balanced nor geodetic. Here $\Phi_{2}$
is order-independent. However, it is distance compatible and
$D^{\max}(\Phi_{2})=D^{\min}(\Phi_{2})=\left[\begin{array}[]{ccccc}0&1&e^{\frac{i\pi}{4}}&2e^{\frac{i\pi}{2}}&e^{\frac{i\pi}{4}}\\\
1&0&1&2e^{\frac{i\pi}{4}}&1\\\ e^{-\frac{i\pi}{4}}&1&0&e^{\frac{i\pi}{4}}&1\\\
2e^{-\frac{i\pi}{2}}&2e^{-\frac{i\pi}{4}}&e^{-\frac{i\pi}{4}}&0&e^{-\frac{i\pi}{4}}\\\
e^{-\frac{i\pi}{4}}&1&1&e^{\frac{i\pi}{4}}&0\end{array}\right].$
## 4 Positively weighted $\mathbb{T}$-gain graph
In this section, we introduce the notion of a positively weighted
$\mathbb{T}$-gain graph. The adjacency matrices of positively weighted
$\mathbb{T}$-gain graphs generalize the following notions: $\mathbb{T}$-gain
adjacency matrices, Hermitian adjacency matrices of mixed graphs, adjacency
matrices of signed graphs, adjacency matrices of undirected graphs.
###### Definition 4.1 (Positively weighted $\mathbb{T}$-gain graph).
Let $\Phi=(G,\varphi)$ be any $\mathbb{T}$-gain graph with $E(G)$ be the
undirected edge set of $G$. Let $w:E(G)\rightarrow\mathbb{R}^{+}$ be a weight
function on the edges of $G$. The positively weighted $\mathbb{T}$-gain graph
associated with $\Phi$ and $w$ is the graph $G$ together with the weighted
gain function $\varphi_{w}$ defined as follows:
$\varphi_{w}(\overrightarrow{e}_{i,j})=\varphi(\overrightarrow{e}_{i,j})w(e_{i,j}).$
A positively weighted $\mathbb{T}$-gain graph on $(G,\varphi)$ is denoted by
$(G,\varphi,w)$(or simply $\Phi_{w}$).
The adjacency matrix associated with $\Phi_{w}$, denoted by $A(\Phi_{w})$, is
an $n\times n$ Hermitian matrix whose $(i,j)th$ entry is
$\varphi_{w}(\overrightarrow{e}_{i,j})$ if $e_{i,j}\in E(G)$, and zero
otherwise. Since $A(\Phi_{w})$ is Hermitian, so all its eigenvalues are real.
The spectrum of $A(\Phi_{w})$ is the spectrum of $\Phi_{w}$.
If $\varphi=1$, then the corresponding positively weighted $\mathbb{T}$-gain
graph is the weighted graph $(G,w)$, and is denoted by $G_{w}$. The adjacency
matrix of $G_{w}$, denoted by $A(G_{w})$, is an $n\times n$ symmetric matrix
with the $(i,j)$th entry $w(e_{i,j})$. Then $A(\Phi_{w})=A(\Phi)\circ
A(G_{w})$, where ${}^{\prime}\circ^{\prime}$ is the Hadamard product.
We establish an expression for the characteristic polynomial of $A(\Phi_{w})$,
which is a generalization of the weighted Sachs formula. Let
$\Phi_{w}=(G,\varphi,w)$ be a positively weighted $\mathbb{T}$-gain graph. The
weight of a cycle $C$ in $G_{w}$ is defined as $w(C)=\prod\limits_{e\in
E(C)}w(e)$. Now $\overrightarrow{C}$ is an oriented cycle. Then the weighted
$\mathbb{T}$-gain of $\overrightarrow{C}$ is
$\varphi_{w}(\overrightarrow{C})=\prod\limits_{\overrightarrow{e}\in\overrightarrow{E(C)}}\varphi_{w}(\overrightarrow{e})=\prod\limits_{\overrightarrow{e}\in\overrightarrow{E(C)}}\varphi(\overrightarrow{e})w(e)=w(C)\varphi(\overrightarrow{C})$.
An elementary subgraph $H$ of $G$ is a subgraph of $G$ such that each
component of $H$ is either a cycle or an edge of $G$. For an elementary
subgraph $H$, $H_{e}$ denotes the set of isolated edges in $H$. The collection
of all elementary subgraphs with $k$ vertices is denoted by $\mathcal{H}_{k}$.
Next, we state the weighted gain Sachs formula. As the proof is similar to
that of the weighted case, so we skip it.
###### Theorem 4.1 (Weighted gain Sachs formula).
Let $\Phi_{w}=(G,\varphi,w)$ be a positively weighted $\mathbb{T}$-gain graph
with characteristic polynomial
$\chi(\Phi_{w};x)=x^{n}+a_{1}x^{n-1}+\dots+a_{n}$. Then
$a_{i}=\sum\limits_{H\in\mathcal{H}_{i}}(-1)^{p(H)}2^{c(H)}w(H_{e})w(H)\prod\limits_{C\in
H}\operatorname{Re}(\varphi(C)),$ (1)
where $c(H),p(H)$ and $C$ denote the number of cycles, the number of
components and cycle in $H$ , respectively.
If we choose $\mathbb{T}$-gain graph $\Phi$ to be the underlying graph $G$,
then above formula become the known Weighted Sachs formula.
###### Corollary 4.1 (Weighted Sachs formula, [6]).
Let $(G,w)$ be a weighted graph with characteristic polynomial
$\chi(G,w;x)=x^{n}+a_{1}x^{n-1}+\dots+a_{n}$. Then the coefficients can be
expressed as
$a_{i}=\sum\limits_{H\in\mathcal{H}_{i}}(-1)^{p(H)}2^{c(H)}w(H_{e})w(H),$ (2)
where $c(H),p(H)$ denote the number of cycles and the number of components in
$H$, respectively.
Now we are ready to state two interesting results which generalize the
corresponding known result for $\mathbb{T}$-gain graph and signed graph.
###### Theorem 4.2.
Let $\Phi_{w}=(G,\varphi,w)$ be a positively weighted $\mathbb{T}$-gain graph.
Then $\Phi_{w}$ and $G_{w}$ are cospectral if and only if $\Phi$ is balanced.
###### Proof.
If $\Phi=(G,\varphi)$ is balanced, then there exists a diagonal unitary matrix
$U$ such that $A(\Phi)=UA(G)U^{*}$. Now, by Proposition 2.1,
$\displaystyle A(\Phi_{w})=A(\Phi)\circ A(G_{w})$
$\displaystyle=UA(G)U^{*}\circ A(G_{w})=U(A(G)\circ
A(G_{w}))U^{*}=UA(G_{w})U^{*}.$
Thus $\Phi_{w}$ and $G_{w}$ are cospectral
Conversely, suppose $\Phi_{w}$ and $G_{w}$ are cospectral. Let
$\chi(\Phi_{w};x)=\sum\limits_{i=0}^{n}a_{i}x^{n-i}$ and
$\chi(G,w;x)=\sum\limits_{i=0}^{n}b_{i}x^{n-i}$ be the characteristic
polynomials of $\Phi_{w}$ and $G_{w}$, respectively. Suppose that $\Phi$ is
not balanced. Then there exists a cycle of smallest length, say $k$, which is
not balanced. Let $\mathcal{C}_{k}$ be the collection of all unbalanced $k$
cycles. Then, by Theorem 4.1,
$b_{k}-a_{k}=2\sum\limits_{C\in\mathcal{C}_{k}}w(C).\\{1-\operatorname{Re}(\varphi(C))\\}>0,$
a contradiction. Thus $\Phi$ is balanced. ∎
The well known Acharya’s spectral criterion for the balance of signed graphs
follows from Theorem 4.2.
###### Corollary 4.2 ([1, Corollary 1.1]).
Let $\Psi=(G,\psi)$ be a signed graph. Then spectra of $\Psi$ and $G$ coincide
if and only if $\Psi$ is balanced.
###### Proof.
By taking $\varphi=\pm 1$ and $w=1$ in Theorem 4.2, we get the result. ∎
Another consequence is the following recent result about the signed graph.
###### Corollary 4.3 ([6, Theorem 2.4]).
Let $\Psi=(G,\psi)$ be a signed graph and $w$ be a positively weighted
function, where $\psi=\pm 1$. Then $\Psi_{w}$ and $G_{w}$ are cospectral if
and only if $\Psi$ is balanced.
Next, we prove one of the main results of this article.
###### Theorem 4.3.
Let $\Phi_{w}=(G,\varphi,w)$ be a connected positively weighted
$\mathbb{T}$-gain graph. Then the spectral radius of $\Phi_{w}$ and $G_{w}$
are equal if and only if either $\Phi$ or $-\Phi$ is balanced.
###### Proof.
Suppose either $\Phi$ or $-\Phi$ is balanced. Then, it is easy to see that,
the spectral radius $\Phi_{w}$ and $G_{w}$ are equal. Conversely, suppose
$\rho(\Phi_{w})=\rho(G_{w})$. Let $\mu_{n}\leq\mu_{n-1}\leq\dots\leq\mu_{1}$
be the eigenvalues of $\Phi_{w}$. Then either $\rho(\Phi_{w})=\mu_{1}$ or
$\rho(\Phi_{w})=-\mu_{n}$.
Case 1: If $\rho(\Phi_{w})=\mu_{1}$, then, by Theorem 2.2, there exists a
diagonal unitary matrix $D$ such that $A(\Phi_{w})=DA(G_{w})D^{*}$. Now
$A(\Phi)\circ A(G_{w})=D(A(G)\circ A(G_{w}))D^{*}$. Then, by Proposition 2.1,
$A(\Phi)\circ A(G_{w})=(DA(G)D^{*})\circ A(G_{w})$. Define $B=(b_{ij})$ as
follows: $b_{ij}$ is the inverse of the nonzero $(i,j)th$-entry of $A(G_{w})$,
otherwise zero. Then $(A(\Phi)\circ A(G_{w}))\circ B=((DA(G)D^{*})\circ
A(G_{w}))\circ B$. Thus, by Proposition 2.1, we have $A(\Phi)=DA(G)D^{*}$.
Thus $\Phi$ is balanced.
Case 2: If $\rho(\Phi_{w})=-\mu_{n}$, then $\mu_{n}=e^{i\pi}\rho(\Phi_{w})$.
By Theorem 2.2, there exists a diagonal unitary matrix $D$, such that
$A(\Phi_{w})=e^{i\pi}DA(G_{w})D^{*}$. That is, $-A(\Phi_{w})=DA(G_{w})D^{*}$.
Then $A(-\Phi)\circ A(G_{w})=D(A(G)\circ A(G_{w}))D^{*}$. By Proposition 2.1,
we have $A(-\Phi)=DA(G)D^{*}$. Thus $-\Phi$ is balanced. ∎
Now we present the following consequences of the above results.
###### Corollary 4.4.
Let $\Phi_{w}=(G,\varphi,w)$ be a connected positively weighted
$\mathbb{T}$-gain graph. Then the largest eigenvalue of $\Phi_{w}$ and $G_{w}$
are equal if and only if $\Phi$ is balanced.
###### Corollary 4.5.
Let $\Phi=(G,\varphi)$ be a connected $\mathbb{T}$-gain graph. Then the
largest eigenvalue of $\Phi$ and $G$ are equal if and only if $\Phi$ is
balanced.
###### Proof.
The proof follows from Corollary 4.4 by assuming $w=1$. ∎
Also Theorem 4.3 unifies the following recent results.
###### Corollary 4.6 ([6, Corollary 2.7]).
Let $\Psi_{w}=(G,\psi,w)$ be a connected positively weighted signed graph.
Then $\Psi$ is balanced if and only if the largest eigenvalue of $\Psi_{w}$
and $G_{w}$ coincide.
###### Proof.
By taking $\varphi=\pm 1$, the result follows from Corollary 4.4. ∎
###### Corollary 4.7 ([10, Theorem 4.4]).
Let $\Phi=(G,\varphi)$ be a connected $\mathbb{T}$-gain graph. Then spectral
radius of $\Phi$ and $G$ coincide if and only if either $\Phi$ is balanced or
$-\Phi$ is balanced.
###### Proof.
Take $w=1$ in Theorem 4.3. ∎
###### Corollary 4.8 ((Stanić’s spectral criterion [14, Lemma 2.1])).
Let $\Psi=(G,\psi)$ be a connected signed graph. Then the largest eigenvalue
of $\Psi$ and $G$ coincide if and only if $\Psi$ is balanced.
###### Proof.
Result follows from Corollary 4.4 by choosing $\varphi=\pm 1$ and $w=1$. ∎
## 5 Characterizations of balanced $\mathbb{T}$-gain graphs in terms of gain
distance matrices
In this section, we establish two characterizations for balanced
$\mathbb{T}$-gain graphs using the gain distance matrices. Let us define two
complete $\mathbb{T}$-gain graphs which are obtained from gain distance
matrices $D^{\max}_{<}(\Phi)$ and $D^{\min}_{<}(\Phi)$.
###### Definition 5.1.
Let $\Phi=(G,\varphi)$ be a $\mathbb{T}$-gain graph and $`<`$ be an order on
$V(G)$. The complete $\mathbb{T}$-gain graph with respect to
$D^{\max}_{<}(\Phi)$, denoted by $K^{D^{\max}_{<}}(\Phi)$, is defined as
follows: keep the edges of $\Phi$ unchanged, and join non adjacent vertices
$v_{i}$ and $v_{j}$ with gain
$\varphi(\overrightarrow{e}_{i,j})=\varphi^{<}_{\max}(v_{i},v_{j})$ for all
$v_{i},v_{j}$. Similarly $K^{D^{\min}_{<}}(\Phi)$ is defined using
$D^{\min}_{<}(\Phi)$.
For a $\mathbb{T}$-gain graph $\Phi=(G,\varphi)$ with order $<$, if
$D^{\max}_{<}(\Phi)=D^{\min}_{<}(\Phi)$, then the associated complete
$\mathbb{T}$-gain graphs $K^{D^{\max}}(\Phi)$ and $K^{D^{\max}}(\Phi)$ are the
same, and it is denoted by $K^{D}(\Phi)$. Then the proof of the following
theorem is easy to verify.
###### Theorem 5.1.
For a $\mathbb{T}$-gain graph $\Phi=(G,\varphi)$, the following are
equivalent:
1. (1)
$K^{D}(\Phi)$ is well defined.
2. (2)
$K^{D^{\max}}(\Phi)=K^{D^{\max}}(\Phi)=K^{D^{\max}_{<}}(\Phi)=K^{D^{\min}_{<}}(\Phi)=K^{D}(\Phi)$.
3. (3)
$D^{\max}_{<}(\Phi)=D^{\min}_{<}(\Phi)$, for some ordering $<$.
###### Theorem 5.2.
Let $\Phi=(G,\varphi)$ be a $\mathbb{T}$-gain graph with vertex order $<$.
Then the following statements are equivalent.
1. (i)
$\Phi$ is balanced.
2. (ii)
$K^{D^{\max}}(\Phi)$ is balanced.
3. (iii)
$K^{D^{\min}}(\Phi)$ is balanced.
4. (iv)
$D^{\max}(\Phi)=D^{\min}(\Phi)$ and associated complete $\mathbb{T}$-gain
graph $K^{D}(\Phi)$ is balanced.
###### Proof.
$(i)\implies(iv)$ Let $V(G)=\\{v_{1},v_{2},\dots,v_{n}\\}$. Suppose $\Phi$ is
balanced. Let $v_{i},v_{j}\in V(G)$. Then, by Lemma 3.1, all shortest oriented
paths $v_{i}Pv_{j}$ have the same gain. Thus
$\varphi_{\max}^{<}(v_{i},v_{j})=\varphi_{\min}^{<}(v_{i},v_{j})$. Therefore,
$D^{\max}_{<}(\Phi)=D^{\min}_{<}(\Phi)$. By Proposition 3.1 and Corollary 3.2,
$D^{\max}(\Phi)=D^{\min}(\Phi)$. Hence $K^{D}(\Phi)$ is well defined.
Claim: $K^{D}(\Phi)$ is balanced.
Let $v_{i}\nsim v_{j}$ in $G$ and $e_{i,j}$ be the edge joining $v_{i}$ and
$v_{j}$ in $K^{D}(\Phi)$. For every oriented path $v_{i}Pv_{j}$ in $\Phi$ have
the same gain. So every cycle passing through the edge $e_{i,j}$ has gain $1$.
Let $T$ be a normal spanning tree of $G$. Suppose $v_{i}\nsim v_{j}$ in $G$.
In $T$, joining the edge $e_{i,j}$ creates a fundamental cycle of
$K^{D}(\Phi)$, say $C_{T}$. Now by previous observation, $\varphi(C_{T})=1$.
Thus all the fundamental cycles in $K^{D}(\Phi)$ are neutral, and hence, by
Lemma 2.1, $K^{D}(\Phi)$ is balanced.
If $K^{D}(\Phi)$ is balanced, then, as $\Phi$ is a subgraph of $K^{D}(\Phi)$,
so $\Phi$ is balanced. Therefore, $(iv)\implies(i),(iii)\implies(i)$ and
$(ii)\implies(i)$ follow.
The proofs $(iv)\implies(iii)$ and $(iv)\implies(ii)$ are easy to see. ∎
###### Theorem 5.3.
Let $\Phi=(G,\varphi)$ be a $\mathbb{T}$-gain graph with vertex order $<$.Then
the following statements are equivalent:
1. (i)
$\Phi$ is balanced.
2. (ii)
$D^{\max}(\Phi)$ is cospectral with $D(G)$.
3. (iii)
$D^{\min}(\Phi)$ is cospectral with $D(G)$.
4. (iv)
The largest eigenvalue of $D^{\max}(\Phi)$ and $D(G)$ are equal.
5. (v)
The largest eigenvalue of $D^{\min}(\Phi)$ and $D(G)$ are equal.
###### Proof.
Let $V(G)=\\{v_{1},v_{2},\dots,v_{n}\\}$, and $\Phi$ be balanced. Then by
Theorem 5.2, $K^{D^{\max}}(\Phi)$ is balanced. Note that
$K^{D^{\max}}(\Phi)=(K_{n},\psi)$ with
$\psi(\overrightarrow{e}_{i,j})=\varphi_{\max}(v_{i},v_{j})=\varphi(v_{i}Pv_{j})$,
where $v_{i}Pv_{j}$ is a shortest path in $\Phi$. Consider ${D^{\max}}(\Phi)$
as the adjacency matrix of a positively weighted $\mathbb{T}$-gain graph
$(K_{n},\psi,w)$ with weight function $w:E(K_{n})\rightarrow\mathbb{R}^{+}$ is
defined as $w(e_{i,j})=d(v_{i},v_{j})$, where $d(v_{i},v_{j})$ is the distance
between $v_{i}$ and $v_{j}$ in $G$. Then the adjacency matrix of $(K_{n},w)$
is same as $D(G)$. By Theorem 4.2 and Theorem 5.2, $\Phi$ is balanced if and
only if $D^{\max}(\Phi)$ is cospectral with $D(G)$. Thus
$(i)\Leftrightarrow(ii)$.
Now, by Corollary 4.4, $\Phi$ is balanced if and only if the largest
eigenvalue of $D^{\max}(\Phi)$ and $D(G)$ coincide. This proves
$(i)\Leftrightarrow(iv)$.
The proofs of $(i)\Leftrightarrow(iii)$ and $(i)\Leftrightarrow(v)$ are
similar . ∎
Both of the above characterizations extend the corresponding known
characterizations [6, Theorem 3.1] and [6, Theorem 3.5] for signed graph.
###### Corollary 5.1.
Let $\Phi=(G,\varphi)$ be a $\mathbb{T}$-gain graph. Then $\Phi$ is balanced
if and only if $D(\Phi)$ exist and it is cospectral with $D(G)$.
## 6 Distance compatible gain graphs
In this final section, we establish a couple of characterizations for distance
compatible $\mathbb{T}$-gain graphs. These results extend the corresponding
known results for signed graph [6].
###### Theorem 6.1.
Let $\Phi=(G,\varphi)$ be any bipartite $\mathbb{T}$-gain graph. Then $\Phi$
is distance compatible if and only if $\Phi$ is balanced.
###### Proof.
If $\Phi$ is balanced, by Theorem 5.3, $\Phi$ is distance compatible.
Conversely, suppose $\Phi$ is distance compatible. Then, by Proposition 3.3,
$\Phi$ is order-independent and $D^{\max}(\Phi)=D^{\min}(\Phi)$. Suppose that
$\Phi$ is unbalanced. Since $\Phi$ is bipartite, there exists an unbalanced
even cycle $C$. Let $v_{i}$ and $v_{j}$ be two diametrical vertices of $C$.
Then $C$ contains two disjoint paths $v_{i}P_{1}v_{j}$ and $v_{i}P_{2}v_{j}$
of same length. Since $\varphi(C)\neq 1$ and
$\varphi(\overrightarrow{C})=\varphi(v_{i}P_{1}v_{j})\varphi(v_{j}P_{2}v_{i})\neq
1$, so $\varphi(v_{i}P_{1}v_{j})\neq\varphi(v_{i}P_{2}v_{j})$.
Claim: $v_{i}P_{1}v_{j}$ and $v_{i}P_{2}v_{j}$ are shortest paths between
$v_{i}$ and $v_{j}$.
Suppose $v_{i}P_{1}v_{j}$ and $v_{i}P_{2}v_{j}$ are not shortest paths. Let
$v_{i}Pv_{j}$ be a shortest path. Then at least one of the even cycle formed
by $v_{i}P_{1}v_{j}$, $v_{i}Pv_{j}$ and $v_{i}P_{2}v_{j}$, $v_{i}Pv_{j}$ is
unbalanced, and has length strictly smaller than that of $C$, which is a
contradiction. Since $\varphi(v_{i}P_{1}v_{j})\neq\varphi(v_{i}P_{2}v_{j})$,
for any ordered vertex set $(V(G),<)$,
$\varphi_{\max}^{<}(v_{i},v_{j})\neq\varphi_{\min}^{<}(v_{i},v_{j})$, a
contradiction. Thus $\Phi$ is balanced. ∎
A _cut vertex_ in a graph $G$ is a vertex whose removal creates more
components than the number of components of $G$. A _block_ of a graph $G$ is a
maximum connected subgraph of $G$ that has no cut vertex.
###### Theorem 6.2.
Let $\Phi=(G,\varphi)$ be a $\mathbb{T}$-gain graph. Then, $\Phi$ is distance
compatible if and only if every block of $\Phi$ is distance compatible.
###### Proof.
Let $B_{1},B_{2},\dots,B_{k}$ be the blocks of $\Phi$. Suppose every block is
distance compatible. Let $s,t\in V(G)$. If $s$ and $t$ are in the same block
then they are distance compatible. Suppose $s$ and $t$ are in different
blocks. Without loss of generality, suppose $s$ is in $B_{1}$ and $t$ is in
$B_{2}$. Then any path $sPt$ in $G$ must passes through the cut vertices
$v_{i}$ and $v_{j}$ where $v_{i}$ and $v_{j}$ are in $B_{1}$ and $B_{2}$,
respectively ($v_{i}$ may be same as $v_{j}$). Any shortest path $sPt$ can be
decompose into $sPv_{i}\cup v_{i}Pv_{j}\cup v_{j}Pt$. Since $B_{1}$ is
distance compatible, so any shortest path $sPv_{i}$ has unique gain. As the
vertices $v_{i}$ and $v_{j}$ are connected by a unique path, so
$\varphi(v_{i}Pv_{j})$ is unique. Proofs of the other cases are similar.
Therefore, any shortest path from $s$ to $t$ has same gain. Thus $s$ and $t$
are distance compatible. Hence $\Phi$ is distance compatible. Converse is easy
to verify. ∎
Let $\Phi=(G,\varphi)$ be $\mathbb{T}$-gain graph with the standard order $<$
on the vertex set. Let $s,t\in V(G)$. If
$\varphi_{\max}^{<}(s,t)=\varphi_{\min}^{<}(s,t)$, then $s$ and $t$ are called
distance compatible. Note that
$\varphi_{\max}^{<}(s,t)=\varphi_{\min}^{<}(s,t)$ if and only if
$\varphi_{\max}^{<_{a}}(s,t)=\varphi_{\min}^{<_{a}}(s,t)$, for any other
vertex order $<_{a}$. Therefore, the vertices $s$ and $t$ are called
_distance-incompatible_ if for some order $<_{a}$,
$\varphi_{\max}^{<_{a}}(s,t)\neq\varphi_{\min}^{<_{a}}(s,t)$ holds.
###### Lemma 6.1.
Let $\Phi=(G,\varphi)$ be a $2$-connected non-geodetic $\mathbb{T}$-gain
graph. If $s$ and $t$ are two incompatible vertices of least distance in $G$
then there exists at least two internally disjoint shortest paths between $s$
and $t$ which have different gains.
###### Proof.
Since $s,t$ are distance-incompatible and $G$ is non-geodetic, so there exist
at least two shortest paths say $sP_{1}t$ and $sP_{2}t$ such that
$\varphi(sP_{1}t)\neq\varphi(sP_{2}t)$. If $sP_{1}t$ and $sP_{2}t$ are
internally disjoint, then we are done. Suppose $sP_{1}t$ and $sP_{2}t$ are not
internally disjoint. Let $v_{1},v_{2},\dots,v_{p}$ be the common internal
vertices of the paths $sP_{1}t$ and $sP_{2}t$ . Let $C_{1},C_{2},\dots,C_{r}$
be the only cycles formed by $sP_{1}t$ and $sP_{2}t$. Thus
$\varphi(sP_{1}t)\varphi(tP_{2}s)=\sum\limits_{i=1}^{r}\varphi(\overrightarrow{C_{i}})\neq
1$. Then there exist a cycle $C_{j}$ which is not balanced. Let $C_{j}$ be
formed by $v_{j}P_{1}v_{j+1}$ and $v_{j}P_{2}v_{j+1}$. Since $sP_{1}t$ and
$sP_{2}t$ are shortest paths, so $v_{j}P_{1}v_{j+1}$ and $v_{j}P_{2}v_{j+1}$
must be shortest paths in between $v_{j}$ and $v_{j+1}$ and of same lengths.
Also
$\varphi(\overrightarrow{C_{j}})=\varphi(v_{j}P_{1}v_{j+1})\varphi(v_{j+1}P_{2}v_{j})\neq
1$. Thus $\varphi(v_{j}P_{1}v_{j+1})\neq\varphi(v_{j}P_{2}v_{j+1})$. Hence
$v_{j}$ and $v_{j+1}$ are distance-incompatible, and distance between them in
$G$ is smaller than the distance between the vertices $s$ and $t$ in $G$, a
contradiction. ∎
###### Theorem 6.3.
Let $\Phi=(G,\varphi)$ be any $2$-connected non-geodetic $\mathbb{T}$-gain
graph. Then $\Phi$ is distance-incompatible if and only if there is an
unbalanced even cycle such that there exist two diametrically opposite
vertices $s$ and $t$ which have no other smaller length path.
###### Proof.
If $\Phi$ is distance-incompatible, then there exist vertices $s,t$ which are
distance-incompatible and of least distance. Then by Lemma 6.1, there exists a
pair of shortest disjoint paths in between $s$ and $t$ such that they have
different gains. Let $C_{2l}$ be the cycle formed by the two disjoin paths.
Therefore, $\varphi(\overrightarrow{C_{2l}})\neq 1$ and $s,t$ do not have any
other shorter length path.
The converse is easy to verify. ∎
## Acknowledgments
The authors thank Prof Thomas Zaslavsky, Binghamton University, for his
comments and suggestions, which improved the paper’s presentation. Aniruddha
Samanta thanks University Grants Commission(UGC) for the financial support in
the form of the Senior Research Fellowship (Ref.No: 19/06/2016(i)EU-V; Roll
No. 423206). M. Rajesh Kannan would like to thank the SERB, Department of
Science and Technology, India, for financial support through the projects
MATRICS (MTR/2018/000986) and Early Career Research Award (ECR/2017/000643).
## References
* [1] B. Devadas Acharya, _Spectral criterion for cycle balance in networks_ , J. Graph Theory 4 (1980), no. 1, 1–11. MR 558448
* [2] R. Bapat, S. J. Kirkland, and M. Neumann, _On distance matrices and Laplacians_ , Linear Algebra Appl. 401 (2005), 193–209. MR 2133282
* [3] R. L. Graham, A. J. Hoffman, and H. Hosoya, _On the distance matrix of a directed graph_ , J. Graph Theory 1 (1977), no. 1, 85–88. MR 505769
* [4] R. L. Graham and L. Lovász, _Distance matrix polynomials of trees_ , Adv. in Math. 29 (1978), no. 1, 60–88. MR 480119
* [5] R. L. Graham and H. O. Pollak, _On the addressing problem for loop switching_ , Bell System Tech. J. 50 (1971), 2495–2519. MR 289210
* [6] Shahul K. Hameed, T. V. Shijin, P. Soorya, K. A. Germina, and Thomas Zaslavsky, _Signed distance in signed graphs_ , Linear Algebra Appl. 608 (2021), 236–247. MR 4143538
* [7] Frank Harary, _On the notion of balance of a signed graph_ , Michigan Math. J. 2 (1953/54), 143–146 (1955). MR 67468
* [8] Roger A. Horn and Charles R. Johnson, _Topics in matrix analysis_ , Cambridge University Press, Cambridge, 1991. MR 1091716
* [9] , _Matrix analysis_ , second ed., Cambridge University Press, Cambridge, 2013. MR 2978290
* [10] Ranjit Mehatari, M. Rajesh Kannan, and Aniruddha Samanta, _On the adjacency matrix of a complex unit gain graph_ , Linear and Multilinear Algebra, 10.1080/03081087.2020.1776672.
* [11] Nathan Reff, _Spectral properties of complex unit gain graphs_ , Linear Algebra Appl. 436 (2012), no. 9, 3165–3176. MR 2900705
* [12] Aniruddha Samanta and M. Rajesh Kannan, _On the spectrum of complex unit gain graph_ , arXiv preprint arXiv:1908.10668 (2019).
* [13] T. V. Shijin, P. Soorya, K. Shahul Hameed, and K. A. Germina, _On signed distance in product of signed graphs_ , arXiv preprint arXiv:2009.08707.
* [14] Zoran Stanić, _Integral regular net-balanced signed graphs with vertex degree at most four_ , Ars Math. Contemp. 17 (2019), no. 1, 103–114. MR 3998150
* [15] Thomas Zaslavsky, _Signed graphs_ , Discrete Appl. Math. 4 (1982), no. 1, 47–74. MR 676405
* [16] , _Vertices of localized imbalance in a biased graph_ , Proc. Amer. Math. Soc. 101 (1987), no. 1, 199–204. MR 897095
* [17] Thomas Zaslavsky, _Biased graphs. I. Bias, balance, and gains_ , J. Combin. Theory Ser. B 47 (1989), no. 1, 32–52. MR 1007712
|
# A Neighborhood-preserving Graph Summarization
Abd Errahmane KIOUCHE Univ Lyon, Université Lyon 1, LIRIS UMR CNRS 5205,
F-69621, Lyon, France.
E-mail<EMAIL_ADDRESS>LCSI, Ecole nationale Supérieure d’Informatique (ESI),Algeria. Julien BASTE
Univ. Lille, CNRS, Centrale Lille, UMR 9189 - CRIStAL - Centre de Recherche en
Informatique Signal et Automatique de Lille, F-59000 Lille, France. E-mail:
<EMAIL_ADDRESS>Mohammed HADDAD Univ Lyon, Université Lyon 1,
LIRIS UMR CNRS 5205, F-69621, Lyon, France.
E-mail<EMAIL_ADDRESS>Hamida SEBA Univ Lyon, Université Lyon 1, LIRIS UMR CNRS 5205, F-69621, Lyon,
France.
E-mail<EMAIL_ADDRESS>
(date)
###### Abstract
We introduce in this paper a new summarization method for large graphs. Our
summarization approach retains only a user-specified proportion of the
neighbors of each node in the graph. Our main aim is to simplify large graphs
so that they can be analyzed and processed effectively while preserving as
many of the node neighborhood properties as possible. Since many graph
algorithms are based on the neighborhood information available for each node,
the idea is to produce a smaller graph which can be used to allow these
algorithms to handle large graphs and run faster while providing good
approximations. Moreover, our compression allows users to control the size of
the compressed graph by adjusting the amount of information loss that can be
tolerated. The experiments conducted on various real and synthetic graphs show
that our compression reduces considerably the size of the graphs. Moreover, we
conducted several experiments on the obtained summaries using various graph
algorithms and applications, such as node embedding, graph classification and
shortest path approximations. The obtained results show interesting trade-offs
between the algorithms runtime speed-up and the precision loss.
_K_ eywords Graph compression $\cdot$ Graph summarization $\cdot$ Algorithm
speed-up $\cdot$ Node embedding $\cdot$ Graph embedding
## 1 Introduction
Graphs are widely used in data modeling because of their ability to represent,
in a simple and intuitive way, complex relations and interactions between
objects: social interactions, protein-protein interactions, chemical molecule
bonds, transport networks, etc. We recall that a graph $G=(V,E)$ is a data
modeling tool consisting of a set $V$ of vertices, also called nodes, and a
set $E$ of edges that connect vertices. Vertices represent objects, while
edges represent relationships between them. Edges can be directed, and both
vertices and edges can have labels. As we are witnessing an explosion in the
number of data generated and processed by our applications, it becomes
important to deal efficiently with large graphs the processing of which
remains a challenging issue. In fact, the amount of generated data is
continuously increasing. Our basic and simple daily activities such as sending
emails, surfing websites, purchasing online, and interacting via social
networks, generate, on their own, a huge amount of data each day. For example,
in 2019 Facebook social network had more than 2.4 Billion monthly active users
with an average of 155 friendship links for each
user111https://www.omnicoreagency.com/facebook-statistics/ visited July.2020.
This large volume of data makes graph querying and analysis a very challenging
task. However, a viable solution seems to emerge from the possibilities
offered by graph summarizing.
Graph summarization, also known by graph compression or simplification, is a
solution that tackles scalability and performance issues when dealing with
massive graph data. Beyond the reduction of the volume of data which is the
main aim of compression, graph summarization looks for significant summaries
that can be used, in graph analysis, without decompression.
In fact, using graph summaries helps to speed-up graph algorithms so that they
can efficiently run on large graphs. Compression algorithms produce smaller
graphs or simpler graph representations, which can be maintained in main
memory and queried and analyzed in reasonable time.
Many graph algorithms, such as node embedding, node classification,
recommendations, shortest path approximation and graph comparison, are based
on the neighborhood information available for each node. Finding this
information may be difficult in practice as dealing with all the neighbors for
each node requires all the edges (links) of the graph to be processed, which
is time and space consuming. This motivated us to introduce a graph
compression that controls the size of the preserved neighborhood of vertices
in the computed summary. So, in this paper, we propose a new graph compression
which retains only a user-specified proportion of the neighbors of each node
to reduce the size of the graph while preserving neighborhood queries.
The main idea is to sparsify the graph by removing edges, while ensuring that
a predefined proportion of the neighbors of each node is included in the set
of $t$-hops neighbors of the node in the compressed graph ($t\geq 1$).
The main advantages of our neighborhood-preserving compression are:
* •
Reduction in storage space of the graph: our compression can decrease
drastically the number of edges in the graph, thus allowing the compressed
graph to be loaded into main memory. The size of the compressed graph can be
controlled by adjusting the proportion of the preserved neighborhood’s
information.
* •
Fast Approximation of graph algorithms: Many graph algorithms, such as
community detection, shortest path lengths and graph comparison, are mainly
based on the neighborhood information of the graph nodes. These algorithms
cannot efficiently run on large graphs since they require all the edges of the
original graph to be loaded in main memory. Since our compression produces a
smaller graph that maintains the principal neighborhood’s information, it can
be used to allow these algorithms to handle large graphs and run faster while
providing good approximations of the original results.
* •
User-controlled trade-off between compression ratio and information loss: with
our compression, the user can control the size of the compressed graph by
adjusting the amount of information loss that can be tolerated. This is a very
useful property, since the amount of tolerated information loss differs
significantly from one application to another, and the desired size of the
compressed graph depends mainly on the available memory.
The remainder of this paper is organized as follows: Section 2 reviews related
works on graph compression methods and their applications. Section 3 formally
defines the problem of neighborhood-preserving graph compression and studies
its complexity. Then, Section 4 provides a description of the algorithms used
to implement this compression. Section 5 presents the results obtained through
the extensive experiments we undertook to evaluate the compression approach,
as well as the usefulness of the obtained summaries. Finally, Section 6
concludes the paper and points out some research perspectives.
## 2 Related Work
Graph compression is attracting increasing interest in various domains and
applications [1]. The aim of graph compression, considered here, is to compute
a graph summary that retains all or part of the original graph properties,
thus allowing use of the summary instead of the original graph in certain
applications. The obtained summary can be either a graph that is simpler or
smaller than the original graph, or any other data structure that is more
compact or is simpler to use than the original graph. Compression algorithms
can be classified in three main categories according to how they simplify the
input graph: (1) sampling, (2) sparsification, and (3) regularity encoding.
Sampling and sparsification based methods generate lossy graph compression and
their results are generally a graph. Regularity encoding based methods allow
having lossy, as well as loss-less, graph summaries, either as graphs or other
data structures:
1. 1.
Graph sampling [2, 3] consists in using a fraction of the graph to make
inferences about the whole dataset. It is generally used for dynamic graphs
for which a sample at time $t$ is a likely representation of the graph. It is
also used with very large graphs, such as protein to protein interactions,
where dealing with the whole graph is too slow. Several graph sampling methods
are proposed in the literature. They generally start with a set of initial
vertices (and/or edges) which can be empty and expand the sample based on a
specific algorithm such as graph exploration and traversal algorithms. As
examples, Breadth-First sampling is used for social network analysis [4] and
graph mining [5]. In [6], the authors apply a traversal based sampling that
utilizes only the local information of nodes, combined with estimated values
of a set of properties, to guide the sampling process and extract tiny samples
that preserve the properties of the graph and closely approximate their
distributions in the original graph. Random walks based methods are also
largely applied in large-scale graph analysis [7, 8, 9]. Frontier sampling, an
edge sampling method using multidimensional random walkers, is used to
estimate the degree distributions and the global clustering coefficient in
[10]. In [11], the authors approximate betweenness centrality based on a
sampled set of shortest paths.
2. 2.
Graph sparsification stands for the methods that compute a sparse subgraph of
the input graph, which preserves some of its properties such as cuts or
shortest paths [12]. Graph sparsification methods can also rely on sampling as
a tool to achieve sparsification. Given a social graph and a log of past
propagations, the authors of [13] prune the network to a prefixed extent,
while maximizing the likelihood of generating the propagation traces in the
log. A similar work is described in [14]. It tackles the problem of
simplifying a graph, while maintaining the connectivity recorded in a given
set of observed activity traces represented by a set of DAGs (or trees) with
specified roots. The problem consists in selecting a subset of arcs in the
graph so as to maximize the number of nodes reachable in all DAGs by the
corresponding DAG roots. This is a cover-maximization problem that the authors
bring to a problem of minimizing a submodular function under size constraints
and using an algorithm introduced in [15] to solve it.
3. 3.
Regularity encoding based methods search for regularities within the graph
structure, i.e., particular patterns or just repetitive patterns, and then
encode these regularities so as to obtain a compact representation of the
graph. Several approaches are proposed in the literature and differ by both
the kind of considered regularities and how these regularities are encoded
within the computed summary. Some methods of this class consist in merging or
combining similar nodes, or subgraphs into super-nodes, and similar edges into
super-edges. Others work directly on the adjacency matrix of the graph using
for example $k2$-trees [16]. In [17], the authors propose a summarizing
approach that iteratively aggregates similar nodes, i.e., those that have the
greatest number of common neighbors. This aggregation is controlled by an
objective function that represents the cost of the compressed output graph and
is defined according to the principle of Minimum Description Length (MDL)
[18]. The graph is encoded with a summary and a set of correcting edges. These
corrections, applied to the summary, enable the initial graph to be
reconstructed. Identifying vertices with a similar neighborhood is a well-
studied topic known as modular decomposition of graphs [19, 20], which aims to
highlight groups of vertices that have the same neighbors outside the group.
These subsets of vertices are called modules. Modular decomposition is used in
[21] to compress a graph and compute its exact list of triangles using solely
the computed summary. The compression consists in considering each module as a
super-node. Several works such as [22] and [23] take advantage of the
regularities of the web graph structure, such as locality and similarity
properties, to compress its adjacency lists and reduce the number of bits
needed to encode a link. In [24, 25], the authors compress graphs using MDL on
a predefined vocabulary of substructures. In [26], graphs are compressed by
recursively detecting repeated substructures and representing them through
grammar rules. In [27], the authors use a clustering algorithm to partition
the original set of vertices into a number of clusters, which will be super-
nodes connected by super-edges to form a complete weighted graph. The super-
edge weights are the edge densities between vertices in the corresponding
super-nodes. The goal is to produce a summary that minimizes the
reconstruction error of the original graph. In [28], the authors merge into
super-nodes graph vertices that have common neighbors so that the obtained
compression ensures that the efficiency of a given task does not drop below a
user-specified threshold. In [29], the authors accelerate node grouping using
a divide and conquer approach that allows parallel node merging. In [30], the
authors use tensor decomposition to group nodes of an evolving graph according
to their connectivity patterns into super-nodes. In [31], the authors address
the problem of preserving node attributes while summarizing diffusion
networks. They propose a sub-quadratic parallelizable algorithm that finds the
best set of candidate nodes and merges them to construct a smaller network of
super-nodes that ensures similar diffusion properties to the original graph.
In [32], the authors use MDL to measure motif relevance based on motif
capacity to compress a graph. In [33], the authors incrementally compute a
summary of an evolving graph using frequent patterns and MDL principle,
combined with a set of operations (merge, split, etc.) on the patterns, in
order to provide changes that have occurred in the data since the previous
state.
In [34], the authors hybrid regularity encoding with sparsification by using
both node grouping and edge sparsification. This allows optimizing the size of
the obtained summary and the graph reconstruction error.
It is interesting to note that few works explore the usefulness of the
computed summaries beyond simple neighborhood or reachability queries.
Summaries obtained by graph sampling are used to estimate graph parameters and
are rarely used as input for graph applications. Most regularity-encoding
based methods do not investigate this issue at all. Graph sparsification
methods are generally designed for specific applications because it is
difficult to have lossy summaries that can be used in several kind of graph
applications. By targeting neighborhood information in our compression and
allowing to control the amount of information loss in the computed summary, we
aim to be able to use our summaries in a variety of graph applications. In
fact, several graph applications, such as node embedding, node classification,
recommendations, etc. are based on the availability of node neighborhood
information. In the remainder of the paper, we show that controlling the
amount of this information in the computed summary allows to reach good trade-
off between algorithm speed-up and precision loss when using the summary as
input instead of the original graph.
## 3 A Neighborhood-preserving graph compression
In this section, we explore a new graph sparsification method that aims to
control the amount of neighborhood information available for each node in the
graph. Our goal is to compute a graph summary that can be used instead of the
original graph in several applications.
### 3.1 Problem Statement
Let $t\geq 1$ be a positive integer. The main idea of neighborhood-preserving
graph compression is to sparsify the input graph by removing edges, while
ensuring that, for all $1\leq i\leq t$, a proportion $p(i)$ of the neighbors
of each node $v$ is included in the set of the $i$-hops neighbors of $v$ in
the compressed graph. We denote such compression by $(p,t)$-compression where:
* •
$p:\mathbb{N}^{*}\to[0,1]$ is a monotonically increasing function, which
represents the proportion of each node’s original neighbors that must be
retrieved in its $i$-hops neighborhood in the compressed graph.
* •
$t$ : is the minimum value for which $p$ reaches its maximal value i.e.,
$p(x)=p(t),\forall x\geq t$.
More formally, given an undirected graph $G=(V,E)$, where $V$ is the set of
vertices and $E$ is the set of edges, a $(p,t)$-compression of $G$ is defined
as follows:
###### Definition 1
Let $t$ be an integer and $p:\mathbb{N}\to[0,1]$ be a monotonically increasing
function satisfying $p(x)=p(t)$, $\forall x>t$. A _$(p,t)$ -compression_ of a
graph $G=(V,E)$ consists in finding a subgraph $G_{c}=(V_{c},E_{c})$ of $G$
such that $V_{c}=V$, $E_{c}\subseteq E$, and, for each $0<x\leq t$ and each
$v\in V$, $\left|N_{G}^{1}(v)\cap
N_{G_{c}}^{x}(v)\right|\geq\left|N_{G}^{1}(v)\right|p(x)$, where
$N_{G}^{x}(v)$ is the set of all $x$-hop neighbors of $v$ in $G$.
In other words, the compressed graph $G_{c}$ contains less edges than $G$ and
preserves a given amount (equal to $p(t)$) of the original neighbors for each
node. In fact, it is required that, for each $0<x\leq t$, a proportion $p(x)$
of the original neighbors must be accessible within a maximum of $x$ hops in
$G_{c}$ using a simple BFS traversal with depth $x$.
Figure 1 illustrates an example of our compression in which $50\%$ of the
original neighbors of each vertex are preserved in the compressed graph and
are reachable within maximum $2$ hops. The resulting compressed graph is
$30\%$ smaller than the original one.
(a) (a) Original graph
(b) (b) Compressed graph
Figure 1: $(p,2)$-compression of Zachary’s karate club network [35] where
$p(1)=0.5$ and $p(2)=1$ for each node.
With $(p,t)$-compression, function $p$ aims to control the loss of
neighborhood information at each neighborhood depth. It is obvious that the
smaller the preserved proportions, the better the compression ratio, and vice-
versa. As regards parameter t, the higher it is, the bigger the stretch factor
of the compressed graph, and vice-versa. The following corollary gives a lower
bound of the size of the compressed graph.
###### property 1
For any $(p,t)$-compression, the number of edges $|E_{c}|$ of the compressed
graph satisfies the following inequality: $|E|p(1)\leq|E_{c}|$
###### proof 1
According to the handshaking theorem, we have $\sum_{v\in V}deg(v)=2|E|$.
Since at least a proportion equal to $p(1)$ of the original neighbors of each
node must be kept in the compressed graph, we have $\sum_{v^{{}^{\prime}}\in
V_{c}}deg(v^{{}^{\prime}})\geq\sum_{v\in V}deg(v)p(1)=2|E|p(1)$, thus
$|E_{c}|\geq p(1)|E|$.
It is interesting to note that spanners [36] are special cases of
$(p,t)$-compression. Given a graph $G$, possibly edge-weighted, a _graph
spanner_ (or _spanner_ for short) is a subgraph $G^{\prime}$ which preserves
lengths of shortest paths in $G$ up to a multiplicative and/or additive error.
A $t$-spanner is a subgraph $G^{\prime}$ such that the distance between two
vertices in $G^{\prime}$ is at most $t$ times the distance between the same
two vertices in $G$. Thus, a $t$-spanner is a particular $(p,t)$-compression
that could be defined such that $p(i)=0$ for every positive integer $i<t$ and
$p(i)=1$ for $i\geq t$. In other words, a $t$-spanner is a $(p,t)$-compression
whose proportion function $p(x)$ is the Shifted Unit Step Function $u(x-t)$.
From a structural point of view, Figure 2 gives an illustration of a 2-spanner
of the Diamond graph (Figure 2.(b)) a 2-spanner of the Diamond graph which is
also a $(p,2)$-compression where $p(1)=\frac{1}{2}$ and $p(2)=1$ (Figure
2.(c)) and a $(p,2)$-compression where $p(1)=\frac{1}{3}$ and
$p(2)=\frac{2}{3}$ which is not a 2-spanner (Figure 2.(d)).
Figure 2: a Diamond graph and some of its compressions.
We can see in this figure that with a $2$-spanner, the left-most vertex is
only connected to $\frac{1}{3}$ of its original neighbors in the Diamond graph
(see Figure 2 (b)), while all the vertices keep at least half of their
original neighbors with a $(p,t)$-compression (see Figure 2 (c)).
### 3.2 NP-hardness and hardness of approximation
Peleg and Schäffer [36] showed that, given an unweighted graph $G$, and
integers $t\geq 2$, $m\geq 1$, determining if $G$ has a $t$-spanner containing
$m$ or fewer edges is NP-complete, even when $t$ is fixed to be $2$. The
reduction is from the edge dominating set problem on bipartite graphs. Since a
$t$-spanner is a particular $(p,t)$-compression of the graph $G$, we can
deduce the following result:
###### theorem 1
Finding the optimal (smallest) compressed graph satisfying the
$(p,t)$-compression constraints for $t\geq 2$ is an NP-Hard problem.
Another work giving us an alternative proof is Cai’s [37]. He showed that for
any fixed $t\geq 2$, the minimum $t$-spanner problem is NP-hard, and for
$t\geq 3$, the problem is NP-hard even when restricted to bipartite graphs.
The reduction is from the 3-SAT problem.
Dinitz et al. [38] show that for $t\leq 3$, and for all $\epsilon>0$, the
$t$-spanner problem cannot be approximated with a ratio better than
$2^{(log^{1-\epsilon}n)/k}$ unless $NP\subseteq BPTIME(2^{polylog(n)})$. This
implies the same inapproximability result for $(p,t)$-compression.
Concerning the best known approximation ratio, Elkin and Peleg [39] propose
approximation algorithms with a sublinear approximation ratio, and study
certain classes of graphs for which logarithmic approximation is feasible. It
is also shown in [40] that for $t=2$, the $t$-spanner problem admits an
$O(log(n))$ approximation. All these results strongly indicate that finding
better or even equivalent approximations for the $(p,t)$-compression problem
will be a hard task. In particular, finding a better result should begin with
finding better approximation than $O(log(n))$ for the $t$-spanner problem.
### 3.3 Integer linear programming formulation
In the following, we provide an integer linear programming formulation of our
problem.
Given an input graph $G=(V,E)$, we denote by $\mathcal{W}$ the set of all
paths in $G$. Given $e=\\{u,v\\}\in E$, we denote by $\mathcal{W}_{uv}$ the
set of all paths in $\mathcal{W}$ from $u$ to $v$. Note that the graph is
undirected and so $\mathcal{W}_{uv}$ also corresponds to the paths from $v$ to
$u$. Moreover, given $i\in\mathbb{N}$ we denote by $\mathcal{W}_{uv}^{i}$ the
set of all paths of $\mathcal{W}_{uv}$ of size at most $i$.
We then define the used variables:
* •
$x_{e}$, for each $e\in E$, denotes whether $e\in E$ is selected ($x_{e}=1$)
or deleted ($x_{e}=0$).
* •
$f_{w}$, for each $w\in\mathcal{W}$, is such that $f_{w}=0$ if at least one
edge $e$ of $w$ is such that $x_{e}=0$. Note that we can have a path
$w\in\mathcal{W}$ such that every edge $e$ of the path is such that $x_{e}=1$
but still have $f_{w}=0$.
We can now write the integer linear programming equation:
$\displaystyle\min\quad$ $\displaystyle\sum_{e\in E}x_{e}$ (1) s.t.
$\displaystyle f_{w}\leq x_{e}$ $\displaystyle\forall w\in\mathcal{W},e\in
E:~{}e\in w$ (2) $\displaystyle\sum_{w\in\mathcal{W}_{uv}}f_{w}\leq 1$
$\displaystyle\forall uv\in E,i\in\mathbb{N}$ (3) $\displaystyle\sum_{v\in
N(u)}\sum_{w\in\mathcal{W}_{uv}^{i}}f_{w}\geq p(i)|N(u)|$
$\displaystyle\forall u\in V,i\in\mathbb{N}$ (4) $\displaystyle
x_{e}\in\\{0,1\\}$ $\displaystyle e\in E$ (5) $\displaystyle
f_{w}\in\\{0,1\\}$ $\displaystyle w\in\mathcal{W}$ (6)
The variable $f_{w}$ can be seen as a flow from the source to the sink. (2)
ensures that if a path $w\in\mathcal{W}$ uses a removed edge $e$ (i.e., $e$ is
such that $x_{e}=0$), then the flow $f_{w}=0$. Using (3), we know that for
each $uv\in E$ there exists at most one path $w\in\mathcal{W}_{uv}$ such that
$f_{w}=1$. By intuition, we assume that we took the path
$w\in\mathcal{W}_{uv}$ of shortest length that is still available in the
remaining graph after removing the edges $e\in E$ such that $x_{e}=0$. Then
condition (4) ensures that the number of neighbors of a vertex $u$, which are
now at distance at most $i$ in the new graph, is at least the proportion given
by $p(i)$.
## 4 Algorithms and approximations
In this section, we present four algorithms for finding the
$(p,t)$-compression of an input graph $G$. Since finding the optimal
$(p,t)$-compression is NP-Hard and cannot be resolved in polynomial time, we
propose polynomial time approximations (sub-optimal algorithms). Algorithm 1
gives the basic implementation of our compression. It has the advantage of
simplicity and speed. Algorithm 1 takes as input a simple graph to compress
$G=(V,E)$, the compression parameters $p$ and $t$ and an order $E_{o}$ for
processing the edges of the input graph. $E_{o}$ is by default a random
ordering of the vertices. The idea is to process the edges of the initial
graph in the order $E_{o}$. The algorithm processes the edges of $G$
incrementally as follows: If an edge $e$ can be removed from $G$ without
violating the neighborhood preservation constraints, the algorithm does not
keep this edge in the summary. Otherwise, the algorithm keeps the edge in the
summary. Assume that the average branching factor ( Out/In degree) of the
graph is equal to $b$, then average time complexity of Algorithm 1 is
$O(|E|b^{t})$.
Data: $G=(V,E)$ a simple Graph,
$t$ an integer,
$p$ a monotonically increasing function $p:\mathbb{N}\to[0,1]$,
$E_{o}$ a possible order of the graph edges
Result: $G_{c}=(V_{c},E_{c})$ a compressed graph
1 $//$ Initialization Step ;
2 $G_{c}=(V_{c},E_{c})\leftarrow(V,\emptyset)$;
3 $G^{\prime}=(V^{\prime},E^{\prime})\leftarrow(V,\emptyset)$;
4 for _$e=(u,v)\in E_{o}$_ do
5 $E^{\prime}\leftarrow E^{\prime}\cup\\{(u,v)\\}$;
6 $N_{G^{\prime}}^{1}(u)\leftarrow$ direct neighbors of node $u$ in
$G^{\prime}$;
7 $N_{G^{\prime}}^{1}(v)\leftarrow$ direct neighbors of node $v$ in
$G^{\prime}$;
8 insert $\leftarrow False$;
9 for _$i=1$ to $t$_ do
10 $N_{Gc}^{i}(u)\leftarrow$ neighbors of node $u$ in $G_{c}$ within at most
$i$-hops;
11 $N_{Gc}^{i}(v)\leftarrow$ neighbors of node $v$ in graph $G_{c}$ within at
most $i$-hops;
12 if _$|N_{Gc}^{i}(u)\cap N_{G^{\prime}}{1}(u)|
<p(i)|N_{G^{\prime}}^{1}(u)|$ or $|N_{Gc}^{i}(v)\cap
N_{G^{\prime}}^{1}(v)|<p(i)|N_{G^{\prime}}^{1}(v)|$_ then
13 insert $\leftarrow True$;
14 Break;
15
16 end if
17
18 end for
19 if _insert_ then
20 $E_{c}\leftarrow E_{c}\cup\\{(u,v)\\}$;
21
22 end if
23
24 end for
Algorithm 1 Basic Algorithm
We note that different edge orderings lead to different compression
performances. Therefore, in order to improve the compression performance of
our algorithm, we propose in the three following subsections, 3 sub-optimal
algorithms which are based on the basic algorithm and try to find a near
optimal edge processing order.
### 4.1 Linear programming
We provide in Section 3.3, an optimal integer linear programming formulation
of $(p,t)$-compression. However, such a resolution is NP-hard to solve, so we
use the standard tricks consisting in relaxing the problem. For this, we keep
the same formulation but allow the values $x_{e}$, $e\in E$, and $f_{w}$,
$w\in\mathcal{W}$, to be any real values between $0$ and $1$. As we only have
to consider paths of a length at most $t$, we have a polynomial number of
variables (the degree of which depends on the fixed value $t$). The average
number of variables is of the order $O(|E|+|V|b^{t})$, where $b$ is the
average branching factor of the graph. This resolution provides a value for
each $x_{e}$, $e\in E$. The interpretation we give to this resolution is that
the higher the value of $x_{e}$, the more likely we want to keep $e$ in our
solution. In reverse, the lower the value of $x_{e}$, the more likely we want
to remove the edge $e$. Thus, we can use the values of $x_{e}$, $e\in E$, in
order to obtain an ordering for the edges and give this ordering to the basic
compression algorithm (see Algorithm 2). Since this linear problem is solvable
in polynomial time, the time complexity of Algorithm 2 is
$O(poly(|E|+|V|b^{t}))$.
Data: $G=(V,E)$ a simple Graph,
$t$ an integer,
$p$ a monotonically increasing function $p:\mathbb{N}\to[0,1]$
Result: $G_{c}=(V_{c},E_{c})$ a compressed graph
1 $//$ Computing the greedy edge order $E_{go}$ ;
2
3Solve the LP Relaxed problem to compute the edge scores $x_{e}$;
4
5$E_{go}\leftarrow$ sort edges E in descending order according to their score
$x_{e}$;
6
7$G_{c}\leftarrow$ Basic Algorithm ($G$,$t$,$p$,$E_{go}$);
Algorithm 2 LP Algorithm
### 4.2 Greedy order based on edge connectivity
Computing the LP-based order is time-consuming for large graphs according to
its time complexity. So, we propose in this subsection another edge ordering
that can be computed much faster than the LP order. The idea is to first
process the edges with a high centrality value. The centrality we consider
here is a relaxation of local edge betweenness defined in [41]. An edge with a
high edge betweenness centrality represents a bridge-like connector between
two parts of a network, the removal of which may affect the shortest paths
between them. The local edge betweenness of an edge $e$ is the number of
shortest paths running along $e$, the length of which is less than or equal to
some constant $t$. In our relaxation, we consider all simple paths of a length
at most $t$, i.e., not necessarily shortest paths. Thus, we compute for every
edge $e$ a centrality score $s(e)$ according to Equation 7.
In Equation 7, $\sigma_{t}(u,v|e)$ is the number of simple paths from $u$ to
$v$ of length $\leq t$ that pass through the edge $e$. Once all scores are
computed, we sort the edges in descending order according to their score
$s(e)$ and pass the obtained order as input to the basic algorithm (See
Algorithm 3). The average time complexity of Algorithm 3 is
$O((|E|+|V|b^{t})log(|E|+|V|b^{t}))$.
$s(e)=\sum_{(u,v)\in E}\sigma_{t}(u,v|e)\ \forall(u,v)\in E$ (7)
Data: $G=(V,E)$ a simple Graph,
$t$ an integer,
$p$ a monotonically increasing function $p:\mathbb{N}\to[0,1]$
1
Result: $G_{c}=(V_{c},E_{c})$ a compressed graph
2 $//$ Computing the greedy edge order $E_{go}$ ;
3
4for _$e\in E$_ do
5 compute the score s(e) using Equation 7;
6
7 end for
8
9$E_{go}\leftarrow$ sort the edges of $G$ in descending order according to
their score $s(e)$;
10
11$G_{c}\leftarrow$ Basic Algorithm ($G$,$t$,$p$,$E_{go}$);
Algorithm 3 Greedy Algorithm based on edge connectivity (EC)
### 4.3 Sub-optimal order based on Simulated Annealing
In the previous two subsections, we have proposed two greedy edge orderings to
improve compression performance. However, the drawback of these two solutions
is that they are more time-consumingthan Algorithm 1 with a random edge
ordering, as we will reveal in the next section with the experimental
evaluation. Moreover, the computation time cannot be controlled by the user
since the computation of both orderings, i.e., the LP ordering and the greedy
ordering based on edge connectivity, cannot be suspended, and we need to go to
the end of the calculation. To overcome this problem, we propose a third
algorithm based on Simulated Annealing (SA) [42]. The advantage of this
solution is that the computation time can be controlled by the user by
adjusting the number of SA iterations. SA is an optimization scheme that
allows efficient search space exploration by accepting, with a given
probability, worst solutions to avoid a premature convergence [42]. SA for
$(p,t)$-compression is illustrated in Algorithm 4. The initial state of the
algorithm is a random order of edges. Then, in each iteration, the algorithm
makes a slight modification to edge order by performing two permutations of
two elements in the vector representing the order and recomputes the cost of
the new solution. If the new order is better, the algorithm keeps the order.
Otherwise, the algorithm keeps it with a probability which increases over the
iterations (see line 19 of Algorithm 4).
Data: $G=(V,E)$ a simple Graph,
$t$ an integer,
$p$ a monotonically increasing function $p:\mathbb{N}\to[0,1]$,
$N$ an integer (Number of iterations),
$T_{0}$ a double ( Initial temperature),
$\alpha$ a double ( decreasing factor)
Result: $G_{c}=(V_{c},E_{c})$ a compressed graph
1
2$S\leftarrow$ Random order of $E$;
3 $T\leftarrow T_{0}$;
4 $G_{t}(V_{t},E_{t})\leftarrow$ Basic Algorithm( $G$,$t$,$p$,$S$);
5 $C_{best}\leftarrow|E_{t}|$;
6 $C_{S}\leftarrow|E_{t}|$;
7 for _$i=1$ to $N$_ do
8 $S_{2}\leftarrow$ Perturbing $S$ by swapping the order of two random edges;
9 $G_{t}(V_{t},E_{t})\leftarrow$ Basic Algorithm($G$,$t$,$p$,$S_{2}$);
10
11 if _$|E_{t}| <C_{best}$_ then
12 $E_{best}\leftarrow S$;
13 $C_{best}\leftarrow|E_{t}|$;
14
15 end if
16 if _$|E_{t}| <COST_{S}$_ then
17 $S\leftarrow S_{2}$;
18 $C_{S}\leftarrow|E_{t}|$;
19
20 end if
21 else
22 $r\leftarrow$ random number between $0$ and $1$;
23 if _$\exp(\frac{C_{S}-|E_{t}|}{T}) >r$_ then
24 $S\leftarrow S_{2}$;
25 $COST_{S}\leftarrow|E_{t}|$;
26
27 end if
28
29 end if
30 $T\leftarrow\alpha*T$;
31 end for
32$G_{c}\leftarrow$ Basic Algorithm ($G$,$t$,$p$,$E_{best}$);
Algorithm 4 $(p,t)$ Compression based on simulated annealing
## 5 Experimental Analysis
In this section, we present an experimental analysis of our compression.
First, we evaluate the approximation algorithms provided to compute the
compression. Then, we provide an analysis of the sensitivity of the
compression to parameters $p$ and $t$. Finally, we evaluate its effectiveness
on several tasks such as graph properties estimation, node embedding and whole
graph embedding. All the experiments are carried out on an Intel core $i7$
processor with $64$ Gigabytes of memory.
### 5.1 Evaluation of approximation algorithms
In this subsection, we present a comparative experimental study of the four
proposed approximations of $(p,t)$-compression. For this, we launched the 4
approximations, i.e., the basic algorithm with random edge ordering (Algorithm
1), the LP approximation (Algorithm 2), the EC approximation (Algorithm 3),
and the SA approximation (Algorithm 4), on 3 families of synthetic graphs the
properties of which are given in Table 1.
Table 1: Characteristics of the synthetic graph families Name | number of graphs | $|V|$ | $|E|$
---|---|---|---
SYNTHETIC 1 | $30$ | $20$ | $60$
SYNTHETIC 2 | $30$ | $50$ | $350$
SYNTHETIC 3 | $30$ | $100$ | $1.4K$
We use the following compression parameters $t=2$ , $p(1)=0.0$ and $p(2)=0.5$.
For a reliable and accurate comparison, we carried out around thirty tests on
each family of graphs for each algorithm. The results of the comparison are
depicted in Table 2. Note that the user configuration of the SA is $T_{0}=10$,
$N=1000$ and $\alpha=0.99$. We notice that the two greedy algorithms LP and
EC, and the SA algorithm outperform the basic algorithm with a random ordering
of edges in terms of compression performance. The results clearly show that
the greedy (EC) and the SA algorithms are the best algorithms. The greedy (EC)
algorithm seems really interesting and offers the best trade-off between
compression performance and runtime. However, all the approximations are still
much slower than the basic algorithm with a random order of edges. Therefore,
we recommend using the basic algorithm for large graphs.
Table 2: Evaluation of the approximation algorithms Dataset | Basic | Greedy ( LP) | Greedy ( EC) | SA
---|---|---|---|---
Avg $|E_{c}|$ | time | Avg $|E_{c}|$ | time | Avg $|E_{c}|$ | time | Avg $|E_{c}|$ | time
SYNTHETIC 1 | 28 | 0.001 | 25.24 | 0.02 | 23.55 | 0.01 | 21.56 | 0.5
SYNTHETIC 2 | 121.63 | 0.008 | 113.26 | 2.5 | 105.66 | 0.02 | 105.9 | 5.2
SYNTHETIC 3 | 367.03 | 0.05 | 354.36 | 212 | 323.23 | 0.09 | 340.4 | 40
### 5.2 Impact of the compression parameters
In this series of experiments, we study the effect of parameters $p$ and $t$
on compression performance. To this end, we evaluate our compression using two
metrics: the compression runtime measured in seconds and the compression ratio
that represents the ratio of the number of deleted edges over the total number
of edges (see Equation 8).
$compression\;ratio=\frac{|E|-|E^{{}^{\prime}}|}{|E|}$ (8)
Note that higher is the compression ratio better is the storage space gain
ensured by the compression. For these experiments and all the following ones
we use real graph datasets. Table 3 gives the main characteristics of these
datasets.
Table 3: Characteristics of the real datasets used in our experiments Name | number of graphs | $|V|$ | $|E|$
---|---|---|---
BLOG-CATALOG | $1$ | $10.31K$ | $333.98K$
CA-ASTROPH | $1$ | $18.77K$ | $198.11K$
CA-HEPTH | $1$ | $9.8K$ | $25.9K$
COLLAB | $5000$ | $372.5K$ | $49.1M$
ENZYMES | $600$ | $19.5K$ | $74.6K$
FLICKR | $1$ | $80.51K$ | $5.89M$
PROTEINS | $1113$ | $43.5K$ | $162.1K$
Table 4 gives the compression ratio obtained by our compression on the CA-
AstroPh dataset, while varying the neighborhood preservation proportion $p$.
As expected, the compression ratio decreases as the preserved proportion of
neighborhood increases and vice-versa. Most of the values of the compression
ratio obtained with the various combinations of parameters are satisfactory.
In addition, we remark that the compression ratio range is wide (from $7\%$ to
$75\%$) which confirms the possibility of controlling effectively the trade-
off information loss/compression ratio using parameters $p$ and $t$.
Furthermore, we set $p(t)=1$ in all experiments, which means that the whole
initial neighborhood of each node can be retrieved in a neighborhood of radius
$r=t$ at maximum. This ensures that reachability queries are fully preserved
for all vertices. The choice of the best combination of parameters depends
essentially on the nature of the graph to be compressed and the user needs.
Particularly, for this example, the combinations $(0.5,1)$ and $(0.7,1)$ seem
really interesting.
Table 4: Compression ratio of the Ca-AstroPh dataset with different combinations of parameters $p$ and $t$ $t$ | $p(1)$ | $p(2)$ | $p(3)$ | compression ratio
---|---|---|---|---
$2$ | $0.2$ | $1.0$ | - | $58.13\%$
$0.5$ | $1.0$ | - | $45.82\%$
$0.7$ | $1.0$ | - | $26.39\%$
$0.9$ | $1.0$ | - | $7.43\%$
$3$ | $0.0$ | $0.2$ | $1.0$ | $75.00\%$
$0.2$ | $0.5$ | $1.0$ | $71.50\%$
$0.5$ | $0.7$ | $1.0$ | $46.73\%$
$0.7$ | $0.9$ | $1.0$ | $26.43\%$
The curves depicted in Figure 3 show the runtime and the compression ratio as
a function of the value of $t$ where $p(0<x<t)=0$ and $p(t)=1$. Note that this
combination of parameters gives a particular type of subgraphs called
$t$-spanners [36]. We notice that the compression ratio grows slowly and
starts to level off from $t=5$. However, the execution time increases
exponentially and rapidly. This is due to the complexity of the compression,
which is of the order $O(|E|b^{t})$ in the average case. Although spanners
give good compression ratios on this dataset ranging from $58\%$ up to $85\%$,
they do not allow good control of the trade-off between information loss and
compression ratio. Indeed, for this dataset, spanners give a control margin of
$(85\%-58\%=27\%)$ for a maximal stretch factor $t=6$, which represents a
significant loss of neighborhood information. However, with
$(p,t)$-compression we get a larger control margin of $(75\%-8\%=67\%)$ with a
maximal stretch factor $t=3$. This confirms once again the efficiency and
usefulness of our compression and its parameter $p$ when compared to spanners.
Figure 3: Compression performance of t-spanners on the Ca-AstroPh dataset
### 5.3 Applications of $(p,t)$compression
Several graph algorithms are based on the availability of neighborhood
information of nodes. Our first motivation is to be able to use these kinds of
algorithms directly on the compressed graphs. So, the purpose of these
experiments is to show the effectiveness of our compression in terms of
speeding-up for such graph algorithms, while handling large graphs and
providing good approximations of the original results. For this, and for all
the following experiments, we use the datasets presented in Table 3 and
compute two new metrics in addition to the compression ratio namely:
* •
Speed-up factor: the ratio between the algorithm run-time on the original
graph and its run-time on the compressed graph. he higher the speed-up factor,
the faster the graph algorithm on the compressed graph.
* •
Performance loss: the difference between the performance metric value on the
original graph and the performance metric value on the compressed graph. For
example, for a classification task the performance loss is the difference
between the accuracy on the original graph and the accuracy on the compressed
graph.The smaller the performance loss, the better the approximation of the
graph properties on the compressed graph.
#### 5.3.1 Shortest paths approximation
The most suitable application for our compression is the approximation of the
shortest paths between all nodes, since every $(p,t)$-compression where
$p(t)=1.0$ preserves all the connectivity properties between the nodes of the
graph, by stretching all the connecting paths by a factor equal to $t$ in the
worst case. In this experimental phase, we compressed three unweighted
undirected graphs with the following combination of parameters $t=2$,
$p(1)=0.5$, and $p(2)=1.0$. Then, we applied the BFS (Breadth First Search)
algorithm to compute all shortest paths between all nodes. Table 5 summarizes
the obtained compression ratio and the shortest path speed-up obtained on
three datasets: CA-ASTROPH, CA-HEPTH and BLOG-CATALOG. We notice that our
compression saves a considerable amount of storage space (compression ratio
$>31\%$) while approximating faster (speed-up ranges from $1.06$ to $1.49$)
the shortest path lengths for the 3 chosen datasets. Indeed, reducing the
number of edges of the graphs reduces the run-time of the shortest paths
computed by the BFS algorithm, which is of complexity $O(|V|(|E|+|V|))$. This
speed-up is more noticeable for denser graphs.
Table 5: Speeding-up all shortest paths computation Dataset | Space gain | Speed-up
---|---|---
CA-ASTROPH | $45.82\%$ | $1.496$
BLOG-CATALOG | $46.52\%$ | $1.323$
CA-HEPTH | $31.08\%$ | $1.065$
Figure 4 shows the distribution of the shortest path lengths in the original
and compressed graphs for the three datasets. We note that the two curves have
almost the same pace. This shows that our compression preserves the
distribution of the lengths of the shortest paths in the 3 datasets. However,
the curves of the compressed graphs are slightly stretched and shifted from
the original curves. This is due to the stretching of the paths as a result of
compression. This stretch is not really considerable because of the
preservation of $50\%$ of the direct neighbors of each vertex in the graph. In
addition, unlike $t$-spanners, $(p,t)$-compression compression controls the
shift between the two curves by adjusting the value of parameter $p$.
(a) (a) Ca-AstroPh dataset
(b) (b) Blog-Catalog dataset
(c) (c) Ca-HepTh dataset
Figure 4: Distribution of shortest path lengths of the original and compressed
graphs.
#### 5.3.2 Whole graph embedding speed-up
Many whole graph embedding methods are based on the local neighborhood
information of the nodes. These methods learn graph representations by
exploring the node neighborhood and extracting some features such as walks,
shortest paths, and local substructures. Since our compression preserves the
local neighborhood within radius $t$, it is thus worth running these
algorithms on compressed graphs to see if our compression speeds up these
algorithms and to evaluate the performance loss. For this, we compressed three
different datasets COLLAB, ENZYMES, and PROTEINS where $t=3$, $p(x<t)=0.0$ and
$p(t)=1.0$ and we run three graph embedding algorithms on the compressed
graphs: shortest path delta kernel [43], graphlet kernel [44] and Graph2vec
[45]. We evaluated the performance of these algorithms on both the original
and the compressed graphs in graph classification tasks as follows: we train
an SVM classifier with $90\%$ of the graphs chosen randomly and then compute
classification accuracy on the test set composed of the remaining $10\%$
graphs. For Graph2vec, we used the best configuration of parameters given in
the original paper [45].
Table 6: Graph kernel performance on the compressed graphs
cr: compression ratio. pl: performance loss
Dataset | Kernel | cr | Speed-up | Original accuracy | Accuracy | pl
---|---|---|---|---|---|---
COLLAB | SP delta | $79\%$ | $4.52$ | $65.78\%$ | $64.78\%$ | $1.00\%$
3-Graphlet | $9.39$ | $64.62\%$ | $53.80\%$ | $10.82\%$
4-Graphlet | $\textgreater 10$ | (out of time) | $53.68\%$ | -
PROTEINS | SP delta | $40\%$ | $1.123$ | $71.91\%$ | $72.18\%$ | $0.0\%$
3-Graphlet | $1.49$ | $71.60\%$ | $71.18\%$ | $0.42\%$
4-Graphlet | $2.02$ | $71.58\%$ | $71.35\%$ | $0.23\%$
ENZYMES | SP delta | $39\%$ | $1.06$ | $29.31\%$ | $24.71\%$ | $4.60\%$
3-Graphlet | $1.6$ | $24.58\%$ | $19.58\%$ | $5.00\%$
4-Graphlet | $1.87$ | $30.03\%$ | $19.70\%$ | $10.33\%$
Table 6 shows the performance of the graph kernels on the compressed graphs.
We notice that the three kernels run faster on compressed graphs in all
experiments. This Kernel computation speed-up is more noticeable on denser
datasets, especially for the graphlet kernels. Indeed, we notice that the
4-Graphlet kernel exceeded the time limit (10 hours) on COLLAB’s original
graphs, while it takes less than one hour on the compressed dataset. Regarding
performance loss, we notice a small loss in values for the shortest path
kernels ranging from $0.0$ to $4.6\%$. However, the loss is more noticeable
for the graphlet kernel but it remains acceptable ($<11.0\%$ in all
experiments). This is due to the fact that our $(p,t)$-compression does not
preserve all graphlets, for example only $6$ graphlets are preserved among the
$11$ graphlets of size=4. Despite this, kernel computation speed-up is very
satisfactory.
Table 7: Graph2vec performance on compressed datasets Dataset | Compression ratio | Speed-up | Performance Loss
---|---|---|---
COLLAB | $79\%$ | $2.54$ | $0.00\%$
PROTEINS | $40\%$ | $1.004$ | $2.93\%$
ENZYMES | $39\%$ | $1.27$ | $5.99\%$
Table 7 shows the performance of Graph2vec on the compressed datasets.
Globally, the algorithm runs faster on the compressed graphs, and the speed-up
factor is bigger on denser datasets. Performance loss is acceptable on the
ENZYMES dataset ($<6\%$) and very satisfactory on the first two datasets
($<3.0\%$). This is due to the fact that Graph2vec considers graphs as sets of
Weisfeiler-Lehman relabeled subgraphs [46] that encompass higher order
neighborhoods of graph nodes. These subgraphs are highly preserved by our
compression. Figure 5 depicts the distribution of the classification accuracy
obtained using Graph2vec on the 3 datasets (compressed and original) by
running 10 experiments on each dataset and for each type of graph (compressed
or original). We notice that the performances on compressed graphs and
original graphs are nearly equivalent. The performance on original graphs is
slightly better for the ENZYMES and PROTEINS datasets. Moreover, the loss of
performance due to the compression is not great.
Figure 5: Graph2vec performance boxplot
#### 5.3.3 Node embedding speed-up
In this series of experiments, we use two main algorithms, i.e., Node2vec [47]
and DeepWalk [48], to learn node representations for both compressed and
original graphs. We use BLOG-CATALOG and FLICKR datsets. The compressed graphs
are obtained using $(p,t)$-compression where $t=2$, $p(1)=0.5$ and $p(2)=1.0$.
The compression ratios are $\geq 45\%$ for the two datasets. Node2vec and
DeepWalk are run using the best parameter combinations given in the original
papers. To evaluate their performance on both compressed and original graphs,
we run multiple multilabel classification tasks on the obtained
representations. To this end, a sample $P_{tr}=\\{0.1,...,0.9\\}$ of the
labeled nodes is used as training data. The rest of the nodes are used for
testing. This process is repeated $10$ times, and we use Macro-F1 and Micro-F1
as performance metrics.
Table 8 shows the performance of Node2vec and Deepwalk on the compressed
graphs where $P_{tr}=0.5$. We notice that Deepwalk runs at the same speed on
both the original and compressed graphs. This is because the time complexity
of Deepwalk depends only on the number of nodes in the graph, which remains
the same after our compression. However, Node2vec runs much faster on the
compressed graphs than on the original graphs. This is justified by the fact
that Node2vec’s time complexity depends on the square of the branching factor
$b$ of the graph [49], which is implicitly related to the number of edges in
the graph. The Micro-F1 and Macro-F1 scores obtained on the compressed graphs
are nearly equivalent to the original scores. The performance loss rates are
low ($\leq 4\%$) for both methods, and insignificant on the FLICKR dataset
($\leq 1.2\%$).
For more fine-grained results, we also compared the performance of the two
algorithms on compressed and original graphs while varying the size of the
training sample $P_{tr}$ from $0.1$ to $0.9$. We summarize the results
graphically for the Micro-F1 and Macro-F1 scores for both methods, i.e.,
Node2vec and Deepwalk, in Figures 6 and 7 respectively. Here we make the same
observations: the performances of the two algorithms on compressed graphs are
almost similar to their performances on the original graphs for all training
rates. With the BLOG-CATALOG dataset, the performance drops by $4\%$ on
compressed graphs in the worst case. However the performance curves on
original and compressed graphs are almost identical in the FLICKR dataset.
Table 8: Performance of node embedding algorithms on compressed datasets Dataset | Method | Space gain | Speed-up | Loss ( Micro F1) | Loss ( Macro F1)
---|---|---|---|---|---
BlogCatalog | DeepWalk | $46.52\%$ | $0.99$ | $3.6\%$ | $3.4\%$
Node2vec | $2.87$ | $3.4\%$ | $2.5\%$
Flickr | DeepWalk | $45.59\%$ | $0.99$ | $0.9\%$ | $1.2\%$
Node2vec | $5.32$ | $0.3\%$ | $0.0\%$
Figure 6: Performance of Node2vec on compressed graphs Figure 7: Performance
of DeepWalk on compressed graphs
## 6 Conclusion and future work
In this paper, we presented a graph summarization approach designed to control
the amount of neighborhood information preserved in the computed summary. This
approach relies on two parameters: a function $p$ that gives the proportion of
each node’s original neighbors to be preserved in its $i$-hops neighborhood in
the compressed graph, and a threshold $t$ for which $p$ reaches its maximal
value.
We presented algorithms to compute this compression with the minimum cost, and
showed their effectiveness in compressing input graphs through experimental
evaluation on multiple real life as well as synthetic graph datasets.
We also showed that the summaries computed by the proposed approach can be
used without any decompression as input to multiple graph applications, such
as node embedding, graph classification, and shortest path approximations. The
results show interesting trade-offs between algorithm runtime speed-up and
precision loss.
As for future work, we consider a more thorough analysis of
$(p,t)$-compression impact on walks based graph learning algorithms such as
Node2vec and DeepWalk. In fact, we observed some situations where learning
accuracy increased when the graph was compressed. This was a quite unexpected
observation. While we guess that walks are biased in the right direction by
removing edges, characterizing such edges remains an open question. Another
important open question is to find an efficient method to order graph edges.
This would allow us to significantly improve the time complexity of the
approach. In addition, we aim to design an incremental version of our
compression to deal with dynamic graphs or graph streams.
We note also that our approach can be used on both directed and undirected
graph. However, our compression do not consider the labels of the edges. To
compress edge-labelled graphs, a new model need to be defined so as to take
into account these labels for example when defining the edge ordering.
##### Acknowledgement:
This work is funded by ANR under grant ANR-20-CE23-0002.
## References
* [1] Y. Liu, T. Safavi, A. Dighe, and D. Koutra. Graph summarization methods and applications: A survey. ACM Comput. Surv., 51(3):62:1–62:34, June 2018.
* [2] A.R. Bloemena. Sampling from a Graph. Mathematical Centre tracts. Mathematisch Centrum, 1976.
* [3] L.-C. Zhang and M. Patone. Graph sampling. METRON, 75(3):277–299, Dec 2017.
* [4] Yong-Yeol Ahn, Seungyeop Han, Haewoon Kwak, Sue Moon, and Hawoong Jeong. Analysis of topological characteristics of huge online social networking services. In Proceedings of the 16th International Conference on World Wide Web, WWW ’07, page 835–844, New York, NY, USA, 2007. Association for Computing Machinery.
* [5] Irma Ravkic, Martin Žnidaršič, Jan Ramon, and Jesse Davis. Graph sampling with applications to estimating the number of pattern embeddings and the parameters of a statistical relational model. Data Mining and Knowledge Discovery, 32(4):913–948, July 2018.
* [6] Muhammad Irfan Yousuf and Suhyun Kim. Guided sampling for large graphs. Data Mining and Knowledge Discovery, 34(4):905–948, July 2020.
* [7] Azade Nazi, Zhuojie Zhou, Saravanan Thirumuruganathan, Nan Zhang, and Gautam Das. Walk, not wait: Faster sampling over online social networks. Proc. VLDB Endow., 8(6):678–689, February 2015.
* [8] Y. Li, Z. Wu, S. Lin, H. Xie, M. Lv, Y. Xu, and J. C. S. Lui. Walking with perception: Efficient random walk sampling via common neighbor awareness. In 2019 IEEE 35th International Conference on Data Engineering (ICDE), pages 962–973, 2019.
* [9] Junzhou Zhao, Pinghui Wang, John C. S. Lui, Don Towsley, and Xiaohong Guan. Sampling online social networks by random walk with indirect jumps. Data Mining and Knowledge Discovery, 33(1):24–57, January 2019\.
* [10] Bruno Ribeiro and Don Towsley. Estimating and sampling graphs with multidimensional random walks. In Proceedings of the 10th ACM SIGCOMM Conference on Internet Measurement, IMC ’10, page 390–403, New York, NY, USA, 2010. Association for Computing Machinery.
* [11] Matteo Riondato and Evgenios M. Kornaropoulos. Fast approximation of betweenness centrality through sampling. Data Mining and Knowledge Discovery, 30(2):438–475, March 2016\.
* [12] L. Paul Chew. There are planar graphs almost as good as the complete graph. Journal of Computer and System Sciences, 39(2):205 – 219, 1989\.
* [13] Michael Mathioudakis, Francesco Bonchi, Carlos Castillo, Aristides Gionis, and Antti Ukkonen. Sparsification of influence networks. In Proceedings of the 17th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, KDD ’11, page 529–537, New York, NY, USA, 2011. Association for Computing Machinery.
* [14] Francesco Bonchi, Gianmarco De Francisci Morales, Aristides Gionis, and Antti Ukkonen. Activity preserving graph simplification. Data Mining and Knowledge Discovery, 27(3):321–343, November 2013\.
* [15] Kiyohito Nagano, Yoshinobu Kawahara, and Kazuyuki Aihara. Size-constrained submodular minimization through minimum norm base. In Proceedings of the 28th International Conference on Machine Learning, ICML 2011, Proceedings of the 28th International Conference on Machine Learning, ICML 2011, pages 977–984, 2011. 28th International Conference on Machine Learning, ICML 2011 ; Conference date: 28-06-2011 Through 02-07-2011.
* [16] Nieves R. Brisaboa, Susana Ladra, and Gonzalo Navarro. k2-trees for compact web graph representation. In Jussi Karlgren, Jorma Tarhio, and Heikki Hyyrö, editors, String Processing and Information Retrieval, pages 18–30, Berlin, Heidelberg, 2009. Springer Berlin Heidelberg.
* [17] Saket Navlakha, Rajeev Rastogi, and Nisheeth Shrivastava. Graph summarization with bounded error. In Proceedings of the 2008 ACM SIGMOD International Conference on Management of Data, SIGMOD ’08, page 419–432, New York, NY, USA, 2008. Association for Computing Machinery.
* [18] J. Rissanen. Modeling by shortest data description. Automatica, 14(5):465 – 471, 1978.
* [19] T. Gallai. Transitiv orientierbare graphen. Acta Mathematica Hungarica, 18:25–66, 1967.
* [20] Michel Habib and Christophe Paul. A survey of the algorithmic aspects of modular decomposition. Computer Science Review, 4(1):41 – 59, 2010.
* [21] Sofiane Lagraa and Hamida Seba. An efficient exact algorithm for triangle listing in large graphs. Data Mining and Knowledge Discovery, 30(5):1350–1369, September 2016.
* [22] T. Suel and Jun Yuan. Compressing the graph structure of the web. In Proceedings DCC 2001. Data Compression Conference, pages 213–222, 2001.
* [23] P. Boldi and S. Vigna. The webgraph framework i: Compression techniques. In Proceedings of the 13th International Conference on World Wide Web, WWW ’04, page 595–602, New York, NY, USA, 2004. Association for Computing Machinery.
* [24] Danai Koutra, U Kang, Jilles Vreeken, and Christos Faloutsos. Summarizing and understanding large graphs. Stat. Anal. Data Min., 8(3):183–202, June 2015.
* [25] Neil Shah, Danai Koutra, Tianmin Zou, Brian Gallagher, and Christos Faloutsos. Timecrunch: Interpretable dynamic graph summarization. KDD ’15, page 1055–1064, New York, NY, USA, 2015. Association for Computing Machinery.
* [26] Sebastian Maneth and Fabian Peternek. Grammar-based graph compression. Information Systems, 76:19 – 45, 2018.
* [27] Matteo Riondato, David García-Soriano, and Francesco Bonchi. Graph summarization with quality guarantees. Data Mining and Knowledge Discovery, 31(2):314–349, March 2017\.
* [28] K. Ashwin Kumar and Petros Efstathopoulos. Utility-driven graph summarization. Proc. VLDB Endow., 12(4):335–347, December 2018.
* [29] Kijung Shin, Amol Ghoting, Myunghwan Kim, and Hema Raghavan. Sweg: Lossless and lossy summarization of web-scale graphs. In The World Wide Web Conference, WWW ’19, page 1679–1690, New York, NY, USA, 2019. Association for Computing Machinery.
* [30] Sofia Fernandes, Hadi Fanaee-T, and João Gama. Dynamic graph summarization: a tensor decomposition approach. Data Mining and Knowledge Discovery, 32(5):1397–1420, September 2018.
* [31] Sorour E. Amiri, Liangzhe Chen, and B. Aditya Prakash. Efficiently summarizing attributed diffusion networks. Data Mining and Knowledge Discovery, 32(5):1251–1274, September 2018.
* [32] Peter Bloem and Steven de Rooij. Large-scale network motif analysis using compression. Data Mining and Knowledge Discovery, 34(5):1421–1453, September 2020.
* [33] Sarang Kapoor, Dhish Kumar Saxena, and Matthijs van Leeuwen. Online summarization of dynamic graphs using subjective interestingness for sequential data. Data Mining and Knowledge Discovery, September 2020.
* [34] Kyuhan Lee, Hyeonsoo Jo, Jihoon Ko, Sungsu Lim, and Kijung Shin. Ssumm: Sparse summarization of massive graphs. In Proceedings of the 26th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, KDD ’20, page 144–154, New York, NY, USA, 2020. Association for Computing Machinery.
* [35] Wayne W Zachary. An information flow model for conflict and fission in small groups. Journal of anthropological research, 33(4):452–473, 1977.
* [36] David Peleg and Alejandro A. Schäffer. Graph spanners. Journal of Graph Theory, 13(1):99–116, 1989.
* [37] Leizhen Cai. Np-completeness of minimum spanner problems. Discrete Applied Mathematics, 48(2):187 – 194, 1994.
* [38] Michael Dinitz, Guy Kortsarz, and Ran Raz. Label cover instances with large girth and the hardness of approximating basic k-spanner. ACM Trans. Algorithms, 12(2), December 2016.
* [39] Michael Elkin and David Peleg. Approximating k-spanner problems for k>2\. Theoretical Computer Science, 337(1):249 – 277, 2005.
* [40] G. Kortsarz and D. Peleg. Generating sparse 2-spanners. Journal of Algorithms, 17(2):222 – 236, 1994.
* [41] Steve Gregory. A fast algorithm to find overlapping communities in networks. In Walter Daelemans, Bart Goethals, and Katharina Morik, editors, Machine Learning and Knowledge Discovery in Databases, pages 408–423, Berlin, Heidelberg, 2008. Springer Berlin Heidelberg.
* [42] Peter JM Van Laarhoven and Emile HL Aarts. Simulated annealing. In Simulated annealing: Theory and applications, pages 7–15. Springer, 1987.
* [43] Karsten M Borgwardt and Hans-Peter Kriegel. Shortest-path kernels on graphs. In Fifth IEEE international conference on data mining (ICDM’05), pages 8–pp. IEEE, 2005.
* [44] Nino Shervashidze, SVN Vishwanathan, Tobias Petri, Kurt Mehlhorn, and Karsten Borgwardt. Efficient graphlet kernels for large graph comparison. In Artificial Intelligence and Statistics, pages 488–495, 2009\.
* [45] Annamalai Narayanan, Mahinthan Chandramohan, Rajasekar Venkatesan, Lihui Chen, Yang Liu, and Shantanu Jaiswal. graph2vec: Learning distributed representations of graphs. arXiv preprint arXiv:1707.05005, 2017.
* [46] Nino Shervashidze, Pascal Schweitzer, Erik Jan Van Leeuwen, Kurt Mehlhorn, and Karsten M Borgwardt. Weisfeiler-lehman graph kernels. Journal of Machine Learning Research, 12(9), 2011.
* [47] Aditya Grover and Jure Leskovec. node2vec: Scalable feature learning for networks. In Proceedings of the 22nd ACM SIGKDD international conference on Knowledge discovery and data mining, pages 855–864, 2016.
* [48] Bryan Perozzi, Rami Al-Rfou, and Steven Skiena. Deepwalk: Online learning of social representations. In Proceedings of the 20th ACM SIGKDD international conference on Knowledge discovery and data mining, pages 701–710, 2014.
* [49] Tiago Pimentel, Adriano Veloso, and Nivio Ziviani. Fast node embeddings: Learning ego-centric representations. 2018\.
|
# Detecting Deepfake Videos Using Euler Video Magnification
Rashmiranjan Das 1, Gaurav Negi 1 and Alan F. Smeaton 1,2
1School of Computing and 2Insight Centre for Data Analytics
Dublin City University, Glasnevin, Dublin, 9, Ireland.
Email<EMAIL_ADDRESS>
###### Abstract
Recent advances in artificial intelligence make it progressively hard to
distinguish between genuine and counterfeit media, especially images and
videos. One recent development is the rise of deepfake videos, based on
manipulating videos using advanced machine learning techniques. This involves
replacing the face of an individual from a source video with the face of a
second person, in the destination video. This idea is becoming progressively
refined as deepfakes are getting progressively seamless and simpler to
compute. Combined with the outreach and speed of social media, deepfakes could
easily fool individuals when depicting someone saying things that never
happened and thus could persuade people in believing fictional scenarios,
creating distress, and spreading fake news. In this paper, we examine a
technique for possible identification of deepfake videos. We use Euler video
magnification which applies spatial decomposition and temporal filtering on
video data to highlight and magnify hidden features like skin pulsation and
subtle motions. Our approach uses features extracted from the Euler technique
to train three models to classify counterfeit and unaltered videos and compare
the results with existing techniques.
## Introduction
Deepfakes describes a technique for artificially manipulating video, initially
applied to swap celebrities faces into video recordings which were shared on
sites like Reddit [13]. They operate by replacing the face of one person in an
original video, with a second person, inserted so that head movement, facial
expressions, lighting and lip syncing when talking, are all exactly as in the
original video. While many thousands of images of the second person to be
superimposed into the deepfake, are usually required, recent work has shown
that good deepfakes can be generated with a reduced number [21]. This means
fake videos can be realistic and can be generated with small amounts of
training data.
Deepfakes can be recognised as both an opportunity and a threat because they
permit users with relatively little computing experience in machine learning
or computer programming to generate almost seamless fake videos. The
availability of state of the art deep learning libraries such as TensorFlow
[1] and Keras [7], with enough accessible training data of facial images,
allows generation of fake video recordings whose quality is so good they can
be very persuading [11].
The initial implementations of deepfakes relied on convolutional autoencoders
[22]. Images of both subjects are reduced to lower dimensions using an encoder
and reconstructed using a decoder. This training is performed for both source
and destination facial expressions. In order to perform a face swap, a trained
encoder of the source is mapped with a decoder trained on the target subject’s
face. An upgrade to this technique is by adding a generative adversarial
network (GAN) in the decoder [10, 2]. GANs consist of two modules, a generator
and a discriminator. The task of the generator is to develop images resembling
the source while the discriminator determines if the image is counterfeited.
It is an iterative process, which makes deepfakes realistic as they are
constantly learning.
The availability of such sophisticated techniques for deepfake generation in
the hands of ordinary researchers and their possible exploitation by other
persons have escalated concerns about their possible misuse. Applications such
as Deepfacelab [20], FakeApp and OpenFaceSwap are GUI based tools made
accessible to relatively untrained researchers to create deepfake videos. With
these tools, it becomes progressively possible for video evidence to be
altered for political tension, false video evidence and fake news. Hence, this
poses a challenge for society as well as an opportunity for creating novel
entertainment, but it demands an effective technique for the detection of such
counterfeit video.
## Related Work on Deepfake Detection
One approach to deepfake detection focuses on psychological signals in the
video [16], proposing a detection method by observing eye blinking in videos,
a psychological signal not well presented in synthesised videos. This is based
on a novel deep learning model combining a convolutional neural network (CNN)
with a recursive neural network (RNN) that captures phenomenological and
temporal regularities in the eye blinking process. Since the training images
used to generate deep fakes do not usually include images of the subject with
eyes closed, this is a clever approach, though it can be circumvented by
intentionally integrating images into the training data with eyes closed.
The work in [15] exploits colour disparity between GAN-generated images and
real images in the non-RGB colour space to classify them. Again in the work
reported in [18] analysed the colour difference between GAN images and real
images. However, it is not clear if this approach can be applied to inspecting
local areas in the image, as would be needed in the deepFake case.
The paper which is most similar to our work is by Fernandes et al. [9], where
they investigate the heart rates of people in deepfake videos and real videos
using Neural ODE. The objective of this paper is to generate a heart rate from
deepfake videos, which are assumed to have no heartbeat. However, this
approach may not perform very well in case of detecting deepfakes as deepfakes
are usually not stable and are not under perfect lighting conditions, so it
can be difficult to obtain a stable heartbeat.
## Euler Video Magnification
Eulerian Video Magnification (EVM) [25] is a technique which can uncover
fleeting and hidden details in videos that will, in general, be hard to see
with the human eye. EVM magnifies and visualises temporal variations in
spatial and/or colour aspects in videos. The method emphasises subtle changes
which occur naturally and are encoded within the video but not seen when
viewed. For instance, one can enhance the slight colour changes in videos
which include exposed human skin such as around the face where the capillaries
in the skin show pulsation due to the bloodstream and blood flow changes
caused, in turn from heart rate. This is similar to Differential Imaging
Forensics introduced in [4] which also reveal latent visual features in videos
that are not perceivable by human observers.
The basis for EVM is that these difficult-to-see changes happen at specific
frequencies that we can expand using a static window in the frequency space.
For instance, in a plucked guitar, each guitar string resonates at a different
frequency, so to amplify a string’s vibration EVM looks at pixel variations in
the respective frequencies of the note being played. Similarly to amplifying
human pulse visualisation in exposed skin around the face, one can consider
pixel changes in frequencies somewhere in the range 1.0 to 3Hz and process
this as it resonates to between 60 and 180 beats per minute.
Figure 1: Euler magnification architecture taken from [25]
The Eulerian amplification process consists of steps shown in Figure 1. A
video is decomposed into images broken down using a Laplacian pyramid into
various frequency ranges. The temporal changes concerning pixels in all
frequency ranges of the Laplacian pyramid are bandpass filtered to select
important and relevant frequency bands which are amplified by a magnification
factor and this outcome is added to the respective signal. Amplified signals
which belong to different frequency bands in the Laplacian pyramid are
flattened to generate the last yield. The key attribute is the temporal
frequency band which can be specified by adjusting the high and low cutoff
frequencies for the filter. One way to consider this is as continuously
stacking small variants of the picture on top of each other. This procedure
makes a pyramid shape with the base as the first picture and gradually
compressing as the pyramid rises.
When we extract the frequency band of interest we amplify the signal and add
it back to the source data. Adjusting the amplification factor of the bandpass
signal results in a larger boost to the temporal bandpass. Changing these
parameters can make variations in the scene more apparent but large
amplification can add artefacts to the result.
## Video Data For Deepfake Detection
During 2020, AWS, Facebook, Microsoft, and others joined together to build and
run a Deepfake Detection Challenge, offering a prize fund of US$1M to
researchers taking part. The Deepfake Detection Challenge Dataset (DFDC) is
described in [8]. The challenge was hosted on the Kaggle website and 2,265
teams took part in the activity, making more than 3,000 submissions.
The full DFDC dataset consists of 124,000 videos, some of which are real, some
are deepfakes. Subjects in the videos are from varied ethnicities and have
different skin tones, genders, lighting conditions and head poses, and
activities. DFDC deepfakes were generated using the whole range of
manipulation techniques, such tampering with the intent of representing the
real adversarial space of facial manipulation, though no further details of
methods used were provided to participants in the challenge.
All video clips in the DFDC training set were left at their original
resolution and quality, so deriving appropriate augmentations of the training
set was left as an exercise to researchers.
When the DFDC was complete and results processed, the best system achieved
82.56% accuracy, based on an ensemble of techniques. This is important for two
reasons. First, because there are many ways to generate deepfakes, these need
to be counteracted by using many ways to detect deepfakes. Secondly, because
an individual generation technique may require different detection techniques,
depending on the video that is generated, thus an ensemble of detection
techniques, is appropriate.
Along with the DFDC data set, we generated our own set of deepfake videos,
created using the Deepfacelab application. Participants’ consent was given and
participants signed an agreement approved by the DCU School of Computing
Research Ethics Committee. A group of 30 participants each submitted a video
of 10 seconds recorded in a controlled environment with suitable lighting at
1080p resolution and the participants remained quite still unlike some of the
DFDC videos. Several face swaps were performed using the H64, H128 and SAE
techniques [17].
These are autoencoder techniques that reduce the data to smaller dimensions,
for example, the H64 model compresses the data into 64x64 pixels. Each face
swap video was trained for 30,000 epochs with a mean loss of 0.0630. All these
videos are of the same dimension as in DFDC dataset with 30 frames per second.
The duration of the videos is 10 secs. A labelled data set was then created by
merging the above data sets and this was also used in our experiments. All
processing was done on an IdeaPad L340 with an NVidia graphic card GTX 1650
and 8 cores, Cuda enabled version 10.1.
## Methodology
This paper explores the use of Euler Video Magnification as a way to pre-
process video and use the result as an indicator of whether a video is a
deepfake or not. We perform both EVM-based colour and movement amplification
on videos to explore if the resulting differences can distinguish between
original and deepfake videos.
A related technique to our work on colour-based EVM is photoplethysmography
(PPG) [19]. This is a process to identify fluctuations in blood volume by
shining light of a given wavelength onto the skin and measuring changes in
light assimilation. The pumping of heart drives blood to the skin surface in
an oscillation cycle and it is the differences in the colour of oxygenated
blood that causes changes in light assimilation which is thus a measure of
heart rate. Photoplethysmography can also be used to recognise human
activities [6] but in this paper we are interested in using colour-based EVM
on exposed skin areas such as the face, to see if a pulse can be detected.
Such a process is robust to different skin tones and small motion of the
subject and our interest is to see if such EVM-based pulses are present in
deepfakes as well as in original videos.
Along with measuring minute colour changes, EVM can also magnify tiny motions
by a subject in a video. One of the characteristics of people is natural
tremor. This is a naturally occurring oscillatory motion which is frequent but
not observable to the naked eye due to its very small amplitude. Their
recurrence is within the scope of 8 to 12 Hz. These periodic motions have been
observed to stay with age and are caused by compressions that are caused in
muscles of the limbs.
To illustrate this, Figure 2 taken from [24] displays vertical and horizontal
displacements that occur naturally to a participant while taking a series of
burst shot images with a smartphone camera, those displacements being caused
by natural human tremors. The graph was based on 86 burst shots and the circle
formed with red dots marks one standard deviation of movement in any
direction. The observation we make from the graph is that human tremors are
symmetrically distributed across all directions. Such small motions due to
natural tremors could be magnified by Euler magnification so that they may be
detectable as a differentiator between real and deepfake videos.
Figure 2: Horizontal and vertical displacement of camera shots taken in burst
mode, caused by natural tremors, taken from [24].
We processed videos with EVM, both deepfakes and originals, and then subjected
them to three different techniques to extract features which we used for video
classification.
### Technique 1: SSIM
The first of the techniques we use is SSIM which is a comparison technique
used to compare two frames, evaluate their likeness and calculate a similarity
index for the video based on visual structures. The Structural Similarity
Index (SSIM) is a perceptual metric that quantifies image quality degradation
caused by processing such as data compression or by losses in data
transmission, and for deepfake videos, a quality degradation in frames are due
to a less well-trained neural net. It is a full reference metric that requires
two images from similar image capture.
SSIM is calculated based on luminance, contrast and structure. Comparing the
time series of SSIM of an original video and its EVM equivalent in Figure 3 it
can be seen that Euler magnification enhances inconsistencies among adjacent
frames of deepfake videos. Such inconsistencies in frames are due to pixelated
faces and such irregularities will have been intensified and magnified as a
result of Euler magnification where the spatial amplification factor of EVM
magnifies the irregularity and hence there are drops in the similarity index.
SSIM is defined in Figure 4 [23].
Figure 3: Inter-frame dissimilarity for original, deepfake, Euler magnified
original and Euler magnified deepfake videos. Figure 4: Structural similarity
index of two windows x and y of common size NxN. $\mu$ is the average value of
(x and y). $\sigma^{2}$ is the variance. $\sigma$ is the covariance of x and
y. c1 and c2 two variables to stabilise the division with weak denominator.
### Technique 2: LSTM
The idea of using a long short term memory (LSTM) network within a neural
network architecture is to help the model learn long term dependencies across
the data series. LSTM networks were first presented by Hochreiter &
Schmidhuber in 1997 [12] and their original idea has been upgraded numerous
times. The LSTM model’s primary objective is to recollect information over an
extensive stretch of time. Unlike an RNN with its single tanh layer, LSTMs
have four strategically arranged modules. On this basis, LSTM has been used by
us for our classification task.
The second technique we use builds on the amount of success achieved by CNN
models in video analysis, by adding a Long Short Term Memory (LSTM) network
into a neural network architecture. This can be used to learn any long term
dependencies in a data sequence. The LSTM is coupled with the inception module
to learn discriminative features from video frames [14].
Inception V3 includes an Inception module where there are changes in the
spatial convolutions to depth-wise separable convolutions. Our model is build
using CNN network layers for feature extraction followed by an LSTM layer for
temporal analysis of Euler magnified videos. Our network has fully connected
layers and a dropout layer to make sure that there is no over-fitting. The
total number of trainable parameters used is 5,500,898 and these are used as
input for the LSTM network and 2 node network working as a detector for
deepfake videos from original videos.
To obtain ground truth, the neural network was trained on videos which were
not modified by the Euler magnified method. The hidden layer had a ‘relu’
activation function, while the last layer had ‘softmax’ as the activation
function. We calculated the loss and accuracy of the technique on both the
training and test sets. We performed the same sequential steps on the same
sets of the video but on the Euler magnified form of the video for comparison
on how the technique compares to the standard classification techniques. We
used an LSTM network with 512 widths and dropout of 0.5 to randomly set values
of outgoing edges of hidden layers to zero. The last layer is constructed
using a softmax activation layer to predict video class.
### Technique 3: Heart Rate Estimation
When computing Euler magnification on a deepfake video, we observed that
deepfake videos also exhibit pulsation as seen in Figure 5. Our third
technique is to estimate the heart rates of subjects appearing in videos by
focusing on an area of facial skin, and analysing if any differences in heart
rate calculations due to skin pulsation could be observed between them.
Figure 5: EVM on deepfake video
The main hypothesis behind the work in this paper is to Euler magnify both
source and counterfeit videos in order to extract features which could
highlight skin pulsation or minuscule movement by a subject in the videos.
Through pulsation, we can visualise and thus be able to extract the heart rate
of the subject in the video by calculating the number of colour change peaks
and counting each one as a heartbeat.
EVM can amplify spatial as well as the temporal aspects of a video. Spatial
magnifies the motion while temporal magnifies colour changes on skin tone. We
use the temporal aspect to visualise the pulse on exposed facial skin. The
features of this that are customisable are filter type, magnification factor
and range of frequency. Videos in our test set were subjected to a range of
EVMs with a frequency range of between 1Hz (60 BPM) and 1.33 Hz (80 BPM). The
amplification factor was set to 50. To fetch a heartbeat, a Fast Fourier
transform algorithm was used. The temporal signal was transformed into a
frequency domain to fetch the signal measured in hertz.
Our data set consists of 19.25% of REAL videos which have not been altered,
With the deepfakes accounting for 80.75% of the samples. There is a huge
imbalance between the categories thus models might be biased towards
categorise videos as deepfake hence the data needs to be up-scaled to balance
it.
We used OpenCV [5] to detect the locations of faces in videos using the face
recognition package. We observe that in some cases, when the subject is not
looking frontally at the camera or when the luminosity is low, the algorithm
for face detection does not detect the face or eyes correctly.
## Results
The results of our experiments were produced using Python 3 on a computer with
8 GB RAM and a 4 core Ryzen 5 AMD processor. We used 400 sets of videos from
the DFDC Kaggle dataset [8] and 30 assembled directly by us. We generated 5
video datasets by changing parameters of Euler magnification, notably
amplification factors. The complete sequence can be divided into the following
steps
1. 1.
Create metadata of the videos extracted from multiple sources;
2. 2.
Detect faces and crop video to leave only the face in the video;
3. 3.
Euler magnify the video with a specific frequency range and amplification
factor;
4. 4.
Train and evaluate a model on Euler magnified video.
### Results for Technique 1: SSIM
As observed anecdotally from multiple graphs of SSIM scores for videos, there
were more similarity score drops in counterfeit deepfake videos when compared
to their real video counterparts when magnified by the Euler magnification
process. After magnifying the data using EVM, we calculated SSIM scores for
all videos in the dataset. Deepfake video detection is particularly difficult
to train as the manipulation can be observed only on a few frames and it is
restricted to certain areas of the face. When there is much movement in the
video, there can be inconsistencies and important areas in the frame appear
only briefly.
Below are the results we obtained when we used a range of standard machine
learning models to classify videos:
Technique | Logistic | Decision | NNet | NNet +
---|---|---|---|---
Regression | Tree | LSTM
Original Videos | 68.7% | 65% | 77% | 77%
EVM Videos | 53.7% | 62% | 70% | 62%
These results are below the 82.56% accuracy achieved by the best-performing
submission to the DFDC but that was based on an ensemble while our results are
one single technique. The results show consistently that SSIM-derived features
are more discriminative of real vs. deepfake videos, before Euler Video
Magnification was applied, suggesting that our default settings for EVM could
be tweaked to improve performance.
In the table below, we processed our test videos with Euler magnification with
multiple amplification factors (15, 20, 30, 40 and 50). The frequency band was
restricted to between 0.8hz and 1.0 Hz. As we can see from the results the
lower amplification factor performed better than the higher amplification
factor. In these videos, a higher amplification factor led to additional noise
which blocked some of the features in the videos. This indicates that Euler
magnification is introducing noise as the amplification factor increases.
Amplification | Accuracy | Loss
---|---|---
Factor
10 | 70.24% | 0.6036
20 | 68.36% | 0.6043
30 | 65.77% | 0.6051
40 | 63.54% | 0.6243
50 | 60.49% | 0.6189
### Results for Technique 2: LSTM
We ran the Inception v3 inspired LSTM model for 100 epochs to give the
following results on our test video set. For these experiments we found that
classification on the original unprocessed videos was more accurate and
outperformed classification accuracy when EVM had been applied to the videos
on an amplification factor of 30.
Video set | Accuracy | Loss
---|---|---
Original Videos | 77.24% | 0.88
EVM Videos | 61.79% | 2.52
This result is another apparent setback to the idea of using EVM for deepfake
detection. The difference in accuracy between original and EVM videos is even
more pronounced than in Model 1.
### Results for Technique 3: Heart Rate Estimation
A comparison of heart rate based on pulsation observed in the Euler magnified
video of an original and a deepfake video revealed that beats per minute (BPM)
for both the videos were very similar, with slight changes only in the first
decimal position.
As an example in Figure 6, the heart rate of both the original and deepfake
videos came to 65.78 beats per minute (1.096 Hz). Thus the temporal variance
calculated through Euler video magnification is insufficient to differentiate
deepfake from original videos [3].
Subjects appearing in both deepfake videos and original videos have an
estimated heartbeat which is indistinguishable. This reveals that the GANs
used for creating the deepfakes does not simply superimpose a new image of the
subject on top of the real image which could have concealed facial pulsation
colour changes, but the GAN manages to faithfully model the true data
distribution of the real data at the pixel level and in this way it keeps the
colour-temporal changes from the genuine video, intact.
Figure 6: Heartbeat calculation using Fast Fourier transform on EVM of
Original and Deepfake video
## Conclusions
In this paper, we tested the effect and impact of Euler Video Magnification as
a technique for video pre-processing leading to possible detection of deepfake
videos. Both the colour and the spatial aspects of EVM were tested as
possibilities for a number of classification models we built for
discriminating between real and deepfake. We used accuracy as a metric, even
though accuracy is known not to be a great metric for evaluation when using
imbalanced datasets like ours, which is why we include accuracy figures for
both original and EVM processed videos, so we can compare.
What the accuracy performance figures do not show is that using EVM vs. not
using EVM as a video pre-process actually detects different videos, so it is
not the case that EVM pre-processing simply eliminates some videos from being
accurately classified as real or deepfake. Thus we should include a range of
metrics including Precision, Recall, F1 and others so as to more fully
understand what EVM is doing, but as this paper is preliminary work, that
forms part of future work.
The results of the best performing systems at the Deepfake Detection Challenge
[8] achieved an accuracy of 82.56% across a much larger test dataset than we
have used here, so we cannot compare our results directly against the data in
the full DFDC results. However because these best results were based on
ensembles of techniques, it follows that the more variety among the systems
used in the ensemble, the better will be will be the final, overall result.
We believe that our results help us to understand how deepfake video
generation incorporate pulsation information and subject tremor motion into
their generated videos. In future work, we would like to explore the most
difficult facial objects to alter like lips and eyes, for fake detection as
well as a deeper exploration into the results by using different evaluation
metrics. We would also investigate different feature sets and limit the
parameters of EVM to focus on its impact. We would also like to explore how
compression artefacts in video storage interact with the EVM process as it is
possible it is masking the contribution of EVM.
### Acknowledgements
This work was part-funded by Science Foundation Ireland through the Insight
Centre for Data Analytics (SFI/12/RC/2289_P2), co-funded by the European
Regional Development Fund.
## Author Biography
Rashmiranjan Das works in the field of Artificial Intelligence at Deciphex
Limited. He received his Masters in Data Analytics from Dublin City University
in 2020 and holds a Bachelors in Computer Engineering from Mumbai University.
After his undergraduate education, he joined Vertisystem limited as a software
developer and then as an Analyst at Ernst & Young.
Gaurav Negi completed his Masters in Data Analytics from Dublin City
University in 2020 and is now working at Xcelerator Machine Translations Ltd.
Gaurav holds a bachelor’s degree in Computer Science from Amity University,
India (2016) and worked for 3 years as a Data Analyst for Genpact India.
Alan Smeaton is Professor of Computing at Dublin City University. He received
his PhD from University College Dublin (1987). He is an elected member of the
Royal Irish Academy and a winner of the Academy’s Gold Medal in Engineering
Sciences, an award given to individuals who have made a demonstrable and
internationally recognised outstanding scholarly contribution in their fields.
Alan is Chair of ACM SIGMM.
## References
* [1] Martín Abadi, Paul Barham, Jianmin Chen, Zhifeng Chen, Andy Davis, Jeffrey Dean, Matthieu Devin, Sanjay Ghemawat, Geoffrey Irving, Michael Isard, et al. Tensorflow: A system for large-scale machine learning. In 12th USENIX Symposium on Operating Systems Design and Implementation, pages 265–283, 2016.
* [2] Martin Arjovsky, Soumith Chintala, and Léon Bottou. Wasserstein gan. arXiv preprint arXiv:1701.07875, 2017.
* [3] Stephanie L Bennett, Rafik Goubran, and Frank Knoefel. Adaptive eulerian video magnification methods to extract heart rate from thermal video. In 2016 IEEE International Symposium on Medical Measurements and Applications (MeMeA), pages 1–5. IEEE, 2016.
* [4] Aurélien Bourquard and Jeff Yan. Differential imaging forensics. arXiv preprint arXiv:1906.05268, 2019.
* [5] Gary Bradski and Adrian Kaehler. Learning OpenCV: Computer vision with the OpenCV library. O’Reilly Media, Inc., 2008.
* [6] Eoin Brophy, José Juan Dominguez Veiga, Zhengwei Wang, Alan F. Smeaton, and Tomás E. Ward. An interpretable machine vision approach to human activity recognition using photoplethysmograph sensor data. In Proceedings of the 26th Irish Conference on Artificial Intelligence & Cognitive Science http://ceur-ws.org/Vol-2259/, pages 244–255, December 2018.
* [7] François Chollet et al. Keras. https://keras.io, 2015. Last Accessed: 21 December 2020.
* [8] Brian Dolhansky, Russ Howes, Ben Pflaum, Nicole Baram, and Cristian Canton Ferrer. The deepfake detection challenge (DFDC) preview dataset. arXiv preprint arXiv:1910.08854, 2019.
* [9] Steven Fernandes, Sunny Raj, Eddy Ortiz, Iustina Vintila, Margaret Salter, Gordana Urosevic, and Sumit Jha. Predicting heart rate variations of deepfake videos using neural ode. In Proceedings of the IEEE International Conference on Computer Vision Workshops, pages 0–0, 2019.
* [10] Ian Goodfellow. NIPS 2016 tutorial: Generative adversarial networks. arXiv preprint arXiv:1701.00160, 2016.
* [11] David Güera and Edward J Delp. Deepfake video detection using recurrent neural networks. In 2018 15th IEEE International Conference on Advanced Video and Signal Based Surveillance (AVSS), pages 1–6. IEEE, 2018.
* [12] Sepp Hochreiter and Jürgen Schmidhuber. LSTM can solve hard long time lag problems. In Advances in neural information processing systems, pages 473–479, 1997.
* [13] Marissa Koopman, Andrea Macarulla Rodriguez, and Zeno Geradts. Detection of deepfake video manipulation. In The 20th Irish Machine Vision and Image Processing Conference, pages 133–136, 2018.
* [14] Akash Kumar, Arnav Bhavsar, and Rajesh Verma. Detecting deepfakes with metric learning. In 8th International Workshop on Biometrics and Forensics, pages 1–6. IEEE, 2020.
* [15] Haodong Li, Bin Li, Shunquan Tan, and Jiwu Huang. Detection of deep network generated images using disparities in color components. arXiv preprint arXiv:1808.07276, 2018.
* [16] Yuezun Li, Ming-Ching Chang, and Siwei Lyu. In ictu oculi: Exposing AI generated fake face videos by detecting eye blinking. arXiv preprint arXiv:1806.02877, 2018.
* [17] Artem A Maksutov, Viacheslav O Morozov, Aleksander A Lavrenov, and Alexander S Smirnov. Methods of deepfake detection based on machine learning. In IEEE Conference of Russian Young Researchers in Electrical and Electronic Engineering (EIConRus), pages 408–411. IEEE, 2020.
* [18] Scott McCloskey and Michael Albright. Detecting GAN-generated imagery using color cues. arXiv preprint arXiv:1812.08247, 2018.
* [19] Seyedeh Somayyeh Mousavi, Mohammad Firouzmand, Mostafa Charmi, Mohammad Hemmati, Maryam Moghadam, and Yadollah Ghorbani. Blood pressure estimation from appropriate and inappropriate PPG signals using a whole-based method. Biomedical Signal Processing and Control, 47:196–206, 2019.
* [20] Ivan Petrov, Daiheng Gao, Nikolay Chervoniy, Kunlin Liu, Sugasa Marangonda, Chris Umé, Jian Jiang, Luis RP, Sheng Zhang, Pingyu Wu, et al. Deepfacelab: A simple, flexible and extensible face swapping framework. arXiv preprint arXiv:2005.05535, 2020.
* [21] Simranjeet Singh, Rajneesh Sharma, and Alan F. Smeaton. Using GANs to synthesise minimum training data for deepfake generation. In Proceedings of the 28th Irish Conference on Artificial Intelligence & Cognitive Science Dublin, Ireland http://ceur-ws.org/Vol-2771/, pages 193–204, December 2020.
* [22] Justus Thies, Michael Zollhofer, Marc Stamminger, Christian Theobalt, and Matthias Nießner. Face2face: Real-time face capture and reenactment of RGB videos. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 2387–2395, 2016.
* [23] Zhou Wang, Eero P Simoncelli, and Alan C Bovik. Multiscale structural similarity for image quality assessment. In The 27th Asilomar Conference on Signals, Systems & Computers, volume 2, pages 1398–1402. IEEE, 2003.
* [24] Bartlomiej Wronski, Ignacio Garcia-Dorado, Manfred Ernst, Damien Kelly, Michael Krainin, Chia-Kai Liang, Marc Levoy, and Peyman Milanfar. Handheld multi-frame super-resolution. ACM Transactions on Graphics, 38(4):1–18, 2019.
* [25] Hao-Yu Wu, Michael Rubinstein, Eugene Shih, John Guttag, Frédo Durand, and William Freeman. Eulerian video magnification for revealing subtle changes in the world. ACM Transactions on Graphics (TOG), 31(4):1–8, 2012.
|
Further author information: Send correspondence to C. P.; E-mail:
<EMAIL_ADDRESS>
# Performance of the Dark Energy Spectroscopic Instrument (DESI) Fiber System
Claire Poppett Space Sciences Laboratory, University of California, Berkeley,
Berkeley, CA Patrick Jelinsky Space Sciences Laboratory, University of
California, Berkeley, Berkeley, CA Julien Guy Lawrence Berkeley National
Laboratory, Berkeley, CA Jerry Edelstein Space Sciences Laboratory,
University of California, Berkeley, Berkeley, CA Sharon Jelinsky Space
Sciences Laboratory, University of California, Berkeley, Berkeley, CA Jessica
Aguilar Lawrence Berkeley National Laboratory, Berkeley, CA Ray Sharples
Centre for Advanced Instrumentation, Durham University, Durham, UK Jurgen
Schmoll Centre for Advanced Instrumentation, Durham University, Durham, UK
David Bramall Centre for Advanced Instrumentation, Durham University, Durham,
UK Luke Tyas Centre for Advanced Instrumentation, Durham University, Durham,
UK Paul Martini The Ohio State University Kevin Fanning The Ohio State
University Michael Levi Lawrence Berkeley National Laboratory, Berkeley, CA
David Brooks University College London Peter Doel University College London
Duan Yutong Boston University Gregory Tarle University of Michigan Erique
Gazta$\tilde{\text{n}}$aga University of Barcelona Francisco Prada
Instituto de Astrofisica de Andalucia CSIC the DESI Collaboration
###### Abstract
The recently commissioned Dark Energy Spectroscopic Instrument (DESI) will
measure the expansion history of the universe using the Baryon Acoustic
Oscillation technique. The spectra of 35 million galaxies and quasars over
14000 sq deg will be measured during the life of the experiment. A new prime
focus corrector for the KPNO Mayall telescope delivers light to 5000 fiber
optic positioners. The fibers in turn feed ten broad-band spectrographs. We
describe key aspects and lessons learned from the development, delivery and
installation of the fiber system at the Mayall telescope.
###### keywords:
DESI, fiber system, focal plane, robotic positioners
## 1 INTRODUCTION
The Dark Energy Spectroscopic Instrument (DESI) is a fiber-fed spectroscopic
instrument installed on the 4-meter Mayall telescope at Kitt Peak National
Observatory (KPNO). During its 5 year survey, DESI will measure the
spectroscopic redshift of up to 35 million galaxies and quasars, enabling the
completion of the largest 3D map of the universe out to a redshift of
$\sim$3.5. This unprecedented dataset will be used to measure Baryon Acoustic
Oscillations and Redshift Space Distortions leading to a more accurate
measurement of the expansion history of the universe and ultimately furthering
our understanding of dark energy [1].
The instrument delivers light from the 4-m telescope through a new corrector
[2], which provides a 3 degree field of view to the focal plane [3]. The focal
plane contains 5,000 optical fibers that are reconfigurable using robotic
positioners [4]. These fibers carry the light from the focal plane to ten
broad-band spectrographs, sensitive from 360-980nm with a resolution of
2,000-5,000 [5].
The DESI fiber system consists of 5020 custom built positioner fiber
assemblies (PFAs) that are installed into 5020 robotic fiber positioners. 5000
of these fibers are used for science and 20 are routed to a sky camera that is
used to estimate the exposure times. The 107$\mu$m core fibers are precision
cleaved and then bonded into fused silica ferrules that allow them to be
bonded into the positioner. A polyimide tube is bonded into the back of the
ferrule to provide strain relief from the positioner as it moves between
targets. Finally, an anti-reflection (AR) coating is applied to the front face
[6]. The PFA is installed into a positioner and the ferrule is bonded into the
ferrule arm after being aligned in focus.
Ten 47.5 m cables run from the focal plane enclosure to spectrograph slits.
Each slit consists of 500 fibers in 20 blocks of 25 fibers, where the blocks
follow the slit curvature. The cable and slit assembles are connected to the
focal plane via fusion splicing. The fusion splice not only facilitates an
effective fabrication, integration and testing flow but also allows the system
to retain maximum throughput and minimize the focal ratio degradation (FRD). A
schematic overview of the DESI fiber system is shown in figure 1.
---
Figure 1: Schematic of the DESI fiber system. The 5000 fibers are divided into
ten identical bundles of 500 fibers, each of which is integrated with one
focal plane petal.
An international collaboration of institutes worked together with industry to
design, produce and test the fiber system. The fiber system was installed at
the 4m-Mayall telescope with $\geq 99\%$ of fibers intact, $\geq 90\%$ fibers
with collimated FRD $\leq 1.8^{\circ}$, and $\geq 90\%$ throughput delivered
from the prime focus corrector to the spectrograph. This high performance is
due to many innovations such as precision cleaved front end fibers [6], and
fusion splices between the focal plane and spectrograph slit assemblies [7].
In this paper we describe key aspects and lessons learned from the successful
development, delivery and installation of the fiber system at the Mayall
telescope. This paper will focus on several new topics. In section 2 we
discuss the requirements and specifications that drove the testing performed
prior to shipping and the performance that was achieved. Section 3 focuses on
the installation of the fiber system at the Mayall Telescope as the cables
were routed from the focal plane to the spectrograph shack. The final section
(4), shows the preliminary performance of the fiber system using on-sky data
taken during early commissioning.
## 2 Fiber system Requirements, Specifications, and Throughput Measurements
As light propagates through an optical fiber, it suffers from fabrication- and
stress-induced increases in entropy that manifest as beam attenuation and
diffusion of the input entrance angles (modal diffusion). Mechanical fiber
stresses can be induced by the fiber termination method, optomechanical
bonding method, and other mechanical forces. Broadly speaking, optimizing
fiber performance is therefore motivated by controlling the entropy gain in
the system by, e.g., minimizing mechanical stresses and/or intentionally
scrambling the light path.
The first-order concern when optimizing fiber performance is throughput loss.
For an end-to-end system, throughput losses due to the fiber cable can be
broken into (1) unavoidable attenuation as the light propagates through the
cable (e.g., cladding-mode losses) and (2) losses that occur when attempting
to capture the fiber output beam. Fiber manufacturers continue to improve the
former with new materials and fabrication techniques. A critical aspect of the
latter is the increase in angles from the input to the output beam. This non
conservation of étendue is referred to as focal-ratio degradation (FRD) [8].
There are two standard experiments used to characterize fiber FRD: (1) Cone
(or solid angle) tests measure the angular diffusion of a fiber by
illuminating it with a filled cone of light with constant surface brightness
over a specified angle.[9] To better simulate the telescope illumination
pattern, scaled obscurations, such as the secondary mirror and support
structure, are added to mimic the telescope. Cone tests therefore directly
measure how much energy will be enclosed within some $f$-ratio in a fiber,
which is the ratio of the focal length to the diameter of the entrance pupil
($f/\\#$), for a known input illumination geometry. (2) Ring (or collimated)
tests measure FRD by injecting a collimated beam (e.g., from a low-power
laser) into the fiber.[10] The fiber azimuthally scrambles the beam to form a
ring illumination pattern, recorded by an imaging detector. The diameter of
the ring gives a direct measurement of the incidence angle and FRD is measured
by the thickness of the ring in the radial direction. Although less direct
than the cone test, the ring test is very simple to perform and interpret
since there are very few sources of measurement error.
The requirement on the focal ratio degradation (FRD) for the fiber system
states that the FRD-induced throughput of the fiber system should be $\geq
90~{}\%$ enclosed energy (EE) averaged over all fibers. This requirement must
be met within an f/3.57 output beam, illuminated by a uniform f/3.9 (7.3
degree half angle) input beam at a wavelength of 625 nm that includes a scaled
Mayall-secondary obstruction (3.46 degree half angle) with a chief ray that is
less than 26 arc minutes from the fiber’s optical axis. This requirement is
verified through full cone FRD testing as described above. However, since it
was impractical to perform a full cone FRD test for every science fiber, it
was necessary to find a relationship between full cone FRD and collimated FRD.
It was found through testing that at f/in=3.9, a $0.2^{\circ}$ increase in the
ring width of a collimated FRD test would result in a 2$\%$ loss in throughput
to the spectrograph as described in Poppett (2018) [11].
Through multiple tests of splicing fibers with differed FRD performances
together we established through that a model that fits the spliced FRD fairly
well is given in the following equation:
$\text{FRD}_{SP}=\sqrt{\text{FRD}^{2}_{PFA}+\text{FRD}^{2}_{slit}}$ (1)
where $\text{FRD}_{SP}$ is the collimated FRD of the resultant splice,
$\text{FRD}_{PFA}$ is the collimated FRD of the positioner fiber assembly and
$\text{FRD}_{slit}$ is the FRD of the slit and cable.
### 2.1 FRD Budget
Collimated FRD testing of first 1000 PFAs is shown in figure 2 and allowed us
to establish a basline for PFA performance. Any PFA with an FRD greater than
1.5∘ should be rejected. As is shown by this figure, the best fit normal
distribution has an average of 0.88∘ and a sigma of 0.18∘, however, the
distribution is not normal and has an extended wing with increasing FRD.
In order to establish the FRD requirements for the fiber cable (fiber in cable
with terminated slit) we assumed that the FRD distribution would be a
Gaussian. We then ran a Monte Carlo simulation of this Gaussian distribution
convolved with the PFA FRD distribution using equation 1 and throughput
measurements, obtained from full cone testing during R&D, in order to estimate
the distribution of throughputs for the fibers. The FRD mean was varied from
1.2∘ to 2.2∘ and the $\sigma$ from 0.2∘ to 0.8∘. The results of a few
throughput distributions are plotted in Figure 2.
From this analysis, and since the distribution is not likely to be Gaussian
and will have a tail for larger FRDs, it was determined that the fiber
cable+slit FRD should have a mean $\leq 1.8^{\circ}$ with a $\sigma$ $\leq
0.6^{\circ}$ in order to meet the FRD budget for the final spliced fiber.
---
Figure 2: Left: Collimated FRD budget of the first 1000 PFAs manufactured and
Right: the modelled throughput distribution for various Cable FRD Gaussian
distributions. The mean and the $\sigma$ of the distribution was varied. The
blue curve has a $\sigma$ of 0.6, and the red curve has a $\sigma$ of 0.8.
### 2.2 As Delivered Throughput
In order to measure the absolute throughput of the full fiber system we
combined collimated FRD results with the measured throughput of a calibrated
fiber. By comparing the total counts in the FRD image of the spliced fiber
with the calibrated fiber it was possible to determine the absolute throughput
of the end-to end system. Figure 3 shows the results from the measurements of
one petal (1 wedge of the focal plane containing 500 fibers) as an example.
All petals met the throughput requirements and the fiber system was delivered
with $\geq 99\%$ of fibers intact and $\geq 90\%$ throughput from the prime
focus corrector to the spectrograph.
---
Figure 3: Collimated FRD performance, Absolute throughput, and throughput
normalised for FRD performance for 1 section of the focal plane prior to
shipment to the telescope.
## 3 Fiber system Installation
The installation of the fiber system at the Mayall telescope was performed in
2019 in three main stages [12, 13]. The first stage was to install the focal
plane system onto the back of the prime focus corrector assembly. In June and
July, 2019, the telescope was parked in the southeast annex location in the
Mayall telescope dome and the petals were installed using a custom
installation system, referred to as the sled, from a platform that enabled the
petals to be guided into position. This process is shown in the first two
panels of figure 4. The final panel in this figure shows the cable management
system, which was designed to route the cables without violating their minimum
bend radius of 200 mm whilst retaining the ability to remove any petal without
others being affected.
---
Figure 4: Ten optical fiber cables from the focal plane petals are strain
relieved to a frame that will later be covered within the FPE. The routing of
each cable was carefully engineered to ensure the bend radius of the cables
was everywhere greater than 200 mm. The panel on the left shows the first
petal loaded onto the sled prior to insertion. The middle panel shows the back
of the focal plane when 8/10 were loaded. The panel on the right shows the
focal plane cable management system.
During August and September, 2019, the thermally insulated focal plane
enclosure (FPE, from Berkeley Lab) was installed and the fiber cables were
routed down the telescope to where the spectrographs are located. The cables
and slits were stored on their shipping carts during the focal plane
installation and handled by an expert team of NOIRlab engineers (see figure
5).
---
Figure 5: The fiber cables and slits were stored on their shipping carts
during the focal plane installation and handled by an expert team of NOIRlab
engineers during cable routing.
The fiber cables emerge from the FPE and cross to the upper ring, along both
the southeast and northeast spider vanes. They are stacked parallel to the
optical axis when they cross the spider vanes in order to to minimize
obscuration. From the upper ring, the two groups of five cables are
constrained by custom hardware designed, built, and installed by NOIRLab. The
two sets of cables converge into a single group of ten cables about halfway
down the serrurier truss.
The fiber cables then follow an arc-shaped path through the declination and
hour axes. They were installed in articulated cable carriers with reversible
bend directions to ensure that they do not twist or otherwise violate their
minimum bend radius. The bundle of ten cables coming down the telescope truss
were loaded into the declination wrap carrier on a bench near the east
declination bearing, which was then craned into its custom guides in the gap
between the telescope center section and the hour angle “horseshoe” [13].
The bundle of ten cables was strain relieved to the oval tube of the
horseshoe, and then the bundle passed into the hour angle cable wrap in one
layer, ten cables wide. The fixed end of the hour angle cable wrap was
attached to the fixed telescope mount, and from there the cables passed
straight through penetrations in the wall of the large coudé room on the main
floor to the east of the telescope mount. The cables continue from that wall
through penetrations in the DESI spectrograph clean room shack. The
penetrations through the walls of the large coudé room and the shack were
sized to allow the passage of the slitheads while they were in their
protective shipping boxes.
---
Figure 6: Routing of the cables through the Dec and Polar Axis before they
penetrate the wall of the spectrograph shack.
## 4 On Sky Fiber system Performance
On sky fiber system performance was evaluated by measuring the stability of
both the bulk throughput and the PSF stability. If the PSF changes from the
calibration to the science exposures, this will result in an extraction bias.
It was a requirement that the spectroscopic PSF should be characterized for
all fibers in each science exposure over the full wavelength range such that
the PSF bias did not exceed 1%.
### 4.1 Fiber Throughput Stability
Throughput variations in the fiber system were measured under a variety of
conditions. The first test measured throughput variations as the robotic fiber
positioners were moved. The measurements were made by moving the telescope
from Zenith and pointing instead to the calibration screen in the dome and
measuring PSF flux variations in the spectrographs as the positioners moved
around their patrol zone. Figure 7 shows that most fibers were found to have a
rms flux variation below 1%. Only 3% of fibers have a flux variation that
exceeds 1%.
---
Figure 7: Example of the aperture flux variations around isolated arc lines
when moving the positioners. Most fibers were found to have a rms flux
variation below 1 %
The second test measured the PSF stability under different telescope
pointings. This was necessary in order to prove that different bends in the
cable did not result in different fiber output distributions. Again, the
telescope was pointed to the white spot and then both the dome and telescope
were moved through 4 different dome azimuth angles (0, 107, 180, 253 deg).
Figure 8 show that the variation in throughput is within $\pm$1%, and there is
no evidence for an effect of the dome azimuth / telescope pointing on the PSF
stability.
---
Figure 8: Example of fiber throughput stability as measured in Spectrograph-0
NIR camera. The variation in throughput is within $\pm$1%, and there is no
evidence for an effect of the dome azimuth / telescope pointing on the PSF
stability. The fibers in the slit are organised into 20 blocks of 25 fibers
and this is clearly seen in the data.
Finally, an analysis of sky residuals was performed on multiple night time
exposures taken during early commissioning. When reducing science data, the
sky spectrum recorded during each integration is used as the normalization,
when considering the residual sky-subtraction error in any given data set
[14]. If the PSF is not stable, this error will increase. The results show
that the sky residuals RMS on emission lines is $\leq$ CCD noise + 1 % of sky
lines. These results are shown in figure 9
---
Figure 9: Sky residuals RMS on emission lines is $\leq$ noise + 1 % of sky.
Results shown are from 1 sky tile in the red and blue cameras observed during
instrument commissioning
## 5 Conclusions
The Dark Energy Spectroscopic Instrument (DESI) is a fiber-fed spectroscopic
instrument installed on the 4-meter Mayall telescope at Kitt Peak National
Observatory (KPNO). DESI fiber system starts with 5000 custom built positioner
fiber assemblies (PFAs) that are installed into 5000 robotic fiber
positioners. Ten 47.5 m cables run from the focal plane enclosure to
spectrograph slits.
The DESI fiber system was delivered to the 4m-Mayall telescope with $\geq
99\%$ of fibers intact and more than $\geq 90\%$ throughput delivered from the
prime focus corrector to the spectrograph. It was installed on the Mayall
telescope at KPNO in 2019. The instrument was commissioned over a 6 month
period in 2019/2020 and finished earlier than planned due to the global
pandemic of COVID-19. Preliminary analysis from this commissioning data has
shown that the PSF stability is $\leq$1 % for most fibers. This requirement is
met when moving or not the positioners, and as a function of dome azimuth.
Finally it has been proven that the PSF stability is sufficient to obtain a
sky subtraction precision better than 1%.
###### Acknowledgements.
This research is supported by the Director, Office of Science, Office of High
Energy Physics of the U.S. Department of Energy under Contract No.
DE–AC02–05CH1123, and by the National Energy Research Scientific Computing
Center, a DOE Office of Science User Facility under the same contract;
additional support for DESI is provided by the U.S. National Science
Foundation, Division of Astronomical Sciences under Contract No. AST-0950945
to the NSF’s National Optical-Infrared Astronomy Research Laboratory; the
Science and Technologies Facilities Council of the United Kingdom; the Gordon
and Betty Moore Foundation; the Heising-Simons Foundation; the French
Alternative Energies and Atomic Energy Commission (CEA); the National Council
of Science and Technology of Mexico; the Ministry of Economy of Spain, and by
the DESI Member Institutions. The authors are honored to be permitted to
conduct astronomical research on Iolkam Du’ag (Kitt Peak), a mountain with
particular significance to the Tohono O’odham Nation.
## References
* [1] DESI Collaboration, “The DESI Experiment Part I: Science,Targeting, and Survey Design,” arXiv e-prints , arXiv:1611.00036 (Oct. 2016).
* [2] Doel, P., Besuner, R., Brooks, D., Flaugher, B., Gallo, G., Gutierrez, G., Kent, S., Lampton, M., Levi, M., Liang, M., Miller, T., and Sprayberry, D., “The prime focus corrector for dark energy spectroscopic instrument,” in [Ground-based and Airborne Instrumentation for Astronomy VI ], Evans, C. J., Simard, L., and Takami, H., eds., Society of Photo-Optical Instrumentation Engineers (SPIE) Conference Series 9908, 99088D (Aug. 2016).
* [3] Lambert, A. R., Besuner, R. W., Claybaugh, T. M., and Silber, J. H., “DESI focal plate mechanical integration and cooling,” in [Ground-based and Airborne Instrumentation for Astronomy VI ], Evans, C. J., Simard, L., and Takami, H., eds., 9908, 2568 – 2579, International Society for Optics and Photonics, SPIE (2016).
* [4] Schubnell, M., Ameel, J., Besuner, R. W., Gershkovich, I., Hoerler, P., Kneib, J.-P., Heetderks, H. D., Silber, J. H., Tarlé, G., and Weaverdyck, C., “The DESI fiber positioner system,” in [Ground-based and Airborne Instrumentation for Astronomy VI ], Evans, C. J., Simard, L., and Takami, H., eds., 9908, 2715 – 2720, International Society for Optics and Photonics, SPIE (2016).
* [5] Edelstein, J., Jelinsky, P., Levi, M., Tarle, G., and Brooks, D., “The DESI spectrograph system and production,” in [Ground-based and Airborne Instrumentation for Astronomy VII ], Evans, C. J., Simard, L., and Takami, H., eds., 10702, 2255 – 2272, International Society for Optics and Photonics, SPIE (2018).
* [6] Poppett, C., Edelstein, J., Aguilar, J., Jelinsky, S., and McCauley, J., “Design, production, and performance of the DESI front end fiber system,” in [Ground-based and Airborne Instrumentation for Astronomy VII ], Evans, C. J., Simard, L., and Takami, H., eds., Society of Photo-Optical Instrumentation Engineers (SPIE) Conference Series 10702, 107027P (July 2018).
* [7] Fagrelius, P., Poppett, C., and Edelstein, J., “Fusion splicing: a novel approach to fiber connections for the Dark Energy Spectroscopic Instrument,” in [Ground-based and Airborne Instrumentation for Astronomy VI ], Evans, C. J., Simard, L., and Takami, H., eds., 9908, 2268 – 2276, International Society for Optics and Photonics, SPIE (2016).
* [8] Ramsey, L. W., “Focal ratio degradation in optical fibers of astronomical interest.,” in [Fiber Optics in Astronomy ], Barden, S. C., ed., Astronomical Society of the Pacific Conference Series 3, 26–39 (Jan. 1988).
* [9] Belland, B., Gunn, J., Reiley, D., Cohen, J., Kirby, E., Cesar de Oliveira, A., Souza de Oliveira, L., Roberts, M., and Seiffert, M., “Focal Ratio Degradation for Fiber Positioner Operation in Astronomical Spectrographs,” Journal of Astronomical Instrumentation 8, 1950007 (Jan. 2019).
* [10] Finstad, D., Wishnow, E., Poppett, C., Sirk, M., Edelstein, J., Gibson, S., Marcy, G., and Howard, A., “Collimated focal ratio degradation testing for highly multiplexed fiber systems - an improvement to a standard test,” the Optical Society 55, 6829 (Sept. 2016).
* [11] Poppett, C., Sharples, R., Edelstein, J., Schmoll, J., and Bramall, D., “The DESI fiber system,” in [Ground-based and Airborne Instrumentation for Astronomy VII ], Evans, C. J., Simard, L., and Takami, H., eds., 10702, 2345 – 2352, International Society for Optics and Photonics, SPIE (2018).
* [12] Allen, L., Sprayberry, D., Evatt, M., Marshall, R., Abareshi, B., Lavoie, T., Joyce, R., Probst, R., Dey, A., Blum, R., Besuner, R., Jelinsky, P., Silber, J., Claybaugh, T., Poppett, C., Doel, P., Brooks, D., Gutierrez, G., Flaugher, B., Gallo, G., Friend, D., Sharples, R., Bramall, D., Honscheid, K., Martini, P., Derwent, M., and O’Brien, T., “DESI installation: preparations and progress (Conference Presentation),” in [Ground-based and Airborne Telescopes VII ], Marshall, H. K. and Spyromilio, J., eds., 10700, International Society for Optics and Photonics, SPIE (2018).
* [13] Sprayberry, D., Goble, W., Allen, L., Elias, J., Probst, R., Joyce, R., Dey, A., Marshall, R., Evatt, M., Blum, R., Abbott, T. M. C., Walker, A., Muñoz, F., Besuner, R., Sholl, M., Jelinsky, P., Silber, J., Lafever, R., Bebek, C., and Flaugher, B., “Planning the installation of the dark energy spectroscopic instrument on the Mayall Telescope,” in [Ground-based and Airborne Telescopes V ], Stepp, L. M., Gilmozzi, R., and Hall, H. J., eds., 9145, 1421 – 1432, International Society for Optics and Photonics, SPIE (2014).
* [14] Sharp, R. and Parkinson, H., “Sky subtraction at the Poisson limit with fibre-optic multiobject spectroscopy,” Monthly Notices of the Royal Astronomical Society 408, 2495–2510 (10 2010).
|
∎
11institutetext: The authors are with the Department of Informatics,
Bioengineering, Robotics, and Systems Engineering, University of Genoa, Via
All’Opera Pia 13, 16145 Genoa, Italy.
11email<EMAIL_ADDRESS>
# An Integrated Localisation, Motion Planning and Obstacle Avoidance Algorithm
in Belief Space
Antony Thomas Fulvio Mastrogiovanni Marco Baglietto
(Received: date / Accepted: date)
###### Abstract
As robots are being increasingly used in close proximity to humans and
objects, it is imperative that robots operate safely and efficiently under
real-world conditions. Yet, the environment is seldom known perfectly. Noisy
sensors and actuation errors compound to the errors introduced while
estimating features of the environment. We present a novel approach (1) to
incorporate these uncertainties for robot state estimation and (2) to compute
the probability of collision pertaining to the estimated robot configurations.
The expression for collision probability is obtained as an infinite series and
we prove its convergence. An upper bound for the truncation error is also
derived and the number of terms required is demonstrated by analyzing the
convergence for different robot and obstacle configurations. We evaluate our
approach using two simulation domains which use a roadmap-based strategy to
synthesize trajectories that satisfy collision probability bounds.
###### Keywords:
Motion Planning Belief Space Planning Collision Probability
## 1 Introduction
Planning and decision making under uncertainty are fundamental requirements
for autonomous robots. Uncertainties often arise due to insufficient knowledge
about the environment, imperfect sensing and inexact robot motion. In these
conditions, the robot poses or other variables of interest can only be dealt
with in terms of probabilities. Planning is therefore performed in the belief
space, which corresponds to the set of all probability distributions over
possible robot states platt2010RSS . At a given time instant, we consider the
belief or the belief state of the robot which corresponds to a probability
distribution of the robot state (or other variables of interest) given the
measurements and controls thus far van_den_berg2012IJRR . Consequently, for
efficient planning and decision making, it is required to reason about future
belief distributions due to candidate actions and the corresponding expected
observations. Such a problem falls under the category of Partially Observable
Markov Decision Processes (POMDPs) kaelbling1998AI .
Robots are becoming ubiquitous in our day-to-day lives and are being
increasingly used in close proximity to humans and other objects in service-
oriented scenarios such as factories, living spaces, or elderly care
facilities. It is therefore of vital importance that robots operate
efficiently and safely in real-world conditions. Localization is a key aspect
for safe and efficient robot motion as it is a precursor to solving the
problems “where to move to” and “how to reach there”. A robot perceives the
environment through its sensors and distinct objects known as landmarks aid
the robot in localizing. However, most approaches assume that these landmarks
are known with high certainty. For example, given the map of the environment,
while planning for future actions the standard Markov localization111The
application of Bayes filter to the localization problem is called Markov
localization thrun2005book . does not take into account the map uncertainty,
that is, the landmark location uncertainties are ignored and the locations are
assumed to be perfectly known. This means that given the map and the sensor
range222Note that the concepts discussed here are applicable to any sensor
used for robot localization. In particular, in this work (Section 5) we use a
laser range finder and beacons that give signal measurements in terms of the
distance to the beacons., for any landmark, there exists a set of viewpoints
from which an observation may be obtained. Let us consider for example a robot
equipped with a laser range finder and observing a landmark. Whenever the
robot location is such that the landmark falls within the sensing range, a
measurement is obtained. Thus, there exists a set of robot locations or
viewpoints from which a measurement of the landmark may be obtained.
Therefore, when the landmark locations are assumed to be perfect, this set of
viewpoints can be easily determined since it depends on the environment map
and the sensing capabilities of the sensor employed. Yet, this might not be
true in practice. For example, consider the map of an environment obtained
from a Simultaneous Localization and Mapping (SLAM) session. Due to the
dynamic nature of the environment, the objects of interests could be occluded
when viewed from the set of viewpoints which would have otherwise produced a
full observation. Moreover, an erroneous localization, for example due to
wrong data association, could lead to wrongly estimated object poses. Thus, in
such cases it is more fitting to consider the uncertainty in the landmark
locations. This landmark uncertainty directly translates to the fact that the
viewpoints whence the object can be observed is uncertain. This is visualized
in Fig. 1. As seen on the left hand side of the figure, when the object
location is known perfectly, there exists a region (green) from which the
object can be observed. Note that as discussed before this region is
determined from the environment map and the sensor capabilities. Some of the
viewpoints inside this region are shown in black. On the right hand side of
the figure, we consider the uncertainty in landmark location the red shaded
region denotes the uncertainty in landmark location. Since the object location
is not known precisely (object can be anywhere within the uncertainty region),
given a viewpoint, it cannot be said with certainty that the object will be
observed. This is so because given a viewpoint, a landmark is observed if it
falls within the sensing range. However, since the landmark location is not
fixed and is uncertain, the landmark may or may not be within the sensing
range. For example, if the landmark location is Gaussian distributed, then the
landmark, in practice can be anywhere within the (say) 3-$\sigma$ uncertainty
region. Thus we cannot define a precise region from which the landmark can be
observed. Therefore, one can only reason in terms of the probability of
observing the object from the considered viewpoint. This results in a
probability distribution function for the viewpoints. Consequently, not
considering this uncertainty can wrongly localize the robot, leading to
inefficient plans causing catastrophes. From now on, we will use the term
object uncertainty to refer to the notion of uncertainty in landmark location.
Figure 1: The red blob denotes an object in the environment. The green region
corresponds to the set of viewpoints from which the object can be observed;
some of these viewpoints are shown as black dots. On the right hand side, the
red shaded region denotes the uncertainty in object location, with the red
blob denoting its mean position. The corresponding viewpoint region is
visualized as the intersection between different viewpoint regions that
correspond to the object being at different locations (left hand side shows
one instance of this).
In order to ensure safe robots’ motion, it is also essential to consider
collision avoidance strategies. As robots are being increasingly used in
service-oriented scenarios with both static and dynamic obstacles,
deterministic approaches do not fare well. Moreover, in the case of dynamic
obstacles, their future states have to be predicted. Yet this is an added
difficulty due to the lack of perfect knowledge of their motions. As a result,
providing safety guarantees is difficult.
### 1.1 Notations and Problem Definition
Throughout this paper vectors will be assumed to be column vectors and will be
denoted by lower case letters, that is x. The transpose of x will be denoted
by $\textbf{x}^{T}$ and its Euclidean norm by
$\mathinner{\\!\left\lVert\textbf{x}\right\rVert}=\sqrt{\textbf{x}^{T}\textbf{x}}$.
A multivariate Gaussian distribution of x with mean $\bm{\mu}$ and covariance
$\Sigma$ will be denoted using the notation
$\textbf{x}\sim\mathcal{N}(\bm{\mu},\Sigma)$. Matrices will be denoted by
capital letters. The trace of a square matrix $M$ will be denoted by $tr(M)$.
The identity matrix will be denoted by $I$ or $I_{n}$ when the dimension needs
to be stressed. A diagonal matrix with diagonal elements
$\lambda_{1},\ldots,\lambda_{n}$ will be denoted by
$diag(\lambda_{1},\ldots,\lambda_{n})$. Sets will be denoted using
calligraphic capital letters like $\mathcal{S}$ or $\mathcal{R}$. Unless
otherwise mentioned, subscripts on vectors/matrices will be used to denote
time indexes and (whenever necessary) superscripts will be used to indicate
the robot or the object that it refers to. For example, $\textbf{x}_{k}^{i}$
represents the state of robot $i$ at time instant $k$. The notation $P(\cdot)$
will be used to denote the probability of an event and the probability density
function (pdf) will be denoted by $p(\cdot)$. While deriving the Belief Space
Planning (BSP) framework to incorporate object uncertainties we will mainly
follow the notations and formalisms in thrun2005book .
We now formally define the problem that we tackle in this paper. Consider a
robot operating in a partially-observable environment. The map of the
environment is either known a priori or is built using a standard SLAM
algorithm. At any time $k$, we denote the robot pose (or configuration) by
$\textbf{x}_{k}\doteq(x_{k},y_{k},\theta_{k})$, the acquired measurement from
objects is denoted by $\textbf{z}_{k}$ and the applied control action is
denoted as $\textbf{u}_{k}$. Note that by objects we refer to both the
landmarks and the obstacles in the environment. We consider a standard motion
model with Gaussian noise
$\textbf{x}_{k+1}=f(\textbf{x}_{k},\textbf{u}_{k})+\textbf{w}_{k}\ ,\
\textbf{w}_{k}\sim\mathcal{N}(0,R_{k})$ (1)
where $\textbf{w}_{k}$ is the random unobservable noise, modeled as a zero
mean Gaussian. We note that modeling the random unobservable noise variables
as Gaussians with zero mean is a common practice in robotics thrun2005book .
The objects are detected through the robot’s sensors and, assuming data
association is known, the observation model can be written as
$\textbf{z}_{k}=h(\textbf{x}_{k},O_{k}^{i})+\textbf{v}_{k}\ ,\
\textbf{v}_{k}\sim\mathcal{N}(0,Q_{k})$ (2)
where $O_{k}^{i}$ is the detected $i$-th object and $\textbf{v}_{k}$ is the
zero mean Gaussian noise. The function $h(\mathbf{x}_{k},O_{k}^{i})$ denotes
the fact that at time $k$, the measurement $\mathbf{z}_{k}$ is obtained by
observing the $i-th$ object $O_{k}^{i}$ from viewpoint (robot location)
$\textbf{x}_{k}$. In the case of a laser-range finder the function $h$ could
be defined as the distance between $\mathbf{x}_{k}$ and the location of the
object (or any particular point on the object) $O_{k}^{i}$. If we consider the
case of a camera, $h$ may be defined as a pinhole projection operator,
projecting the object $O_{k}^{i}$ onto the image plane. Given the models in
(1) and (2), in this paper we focus on two aspects. First, we consider the
object uncertainties while localizing the robot. Second, we compute the exact
probability of collision under obstacle uncertainty, which is modeled as a
Gaussian distribution. Finally, we evaluate our approach in two simulation
domains: a 2D mobile robot domain and a 2D manipulator domain. It is to be
noted that for the manipulator domain we will be concerned with the collision
avoidance of the manipulator’s end-effector.
### 1.2 Related Work
BSP has been researched extensively in the past with applications spanning a
variety of areas including autonomous navigation, multi-modal planning, and
active SLAM van_den_berg2012IJRR ; Kurniawati2016ISRR ; agha_mohammadi2014IJRR
; prentice2009IJRR ; thomas2019ISRR ; thomas2020STAIRS ; kaelbling2013IJRR ;
pathak2018IJRR .kaelbling2013IJRR consider object uncertainty since they are
planning in an unknown environment and require several measurements to obtain
confidence estimates of object locations. Thus they perform active perception,
that is, to look for robot actions that enhances information to reduce the
object uncertainty. This context is different from ours since we consider a
known environment with object uncertainty and focus on active localization
incorporating these uncertainties. In pathak2018IJRR , the concept of object
uncertainty is commented upon (they call it scene uncertainty); however they
do not show how it affects the state estimation. Dynamic environments are
considered in Kurniawati2016ISRR ; agha_mohammadi2014IJRR however the
landmark/beacon locations are assumed to be known perfectly; thomas2019ISRR ;
thomas2020STAIRS also consider perfect landmark locations in the context of
task and motion planning. Thus most active and passive localization-based
approaches focus on robot state uncertainty and assume perfect knowledge about
the location of the objects in the environment. However, in practice, the
environment is seldom known with high certainty and hence providing formal
guarantees for safe navigation is imperative.
Patil et al. patil2012ICRA estimate the probability of collision under robot
state uncertainty by truncating the state distributions. In bry2011ICRA ,
future state distributions are predicted and the uncertainties are used to
compute bounded collision probabilities. Lee et al. lee2013IROS use sigma
hulls333 Sigma hulls are convex hulls of the geometry of individual robot
links transformed according to the sigma points in joint space lee2013IROS .
to formulate collision avoidance constraints in terms of the signed distance
to the obstacles. Du Toit and Burdick dutoit2011IEEE , Park et al.
park2018IEEE compute the collision probability by marginalizing the joint
distribution between the robot and obstacle location. The distributions are
assumed to be Gaussian and the marginalization is computed with an indicator
function that is true under the collision condition. However, since there is
no closed-form solution to this formulation, an approximation is assumed.
Furthermore, Park et al. compute an upper bound for the collision probability.
An approximation is computed using Monte Carlo Integration in
lambert2008ICCARV , albeit computationally intensive. Another impressive work
that uses Monte Carlo approach is Monte Carlo Motion Planning (MCMP)
janson2018ISRR . This approach first solves a deterministic motion planning
problem with inflated obstacles and then adjusts the inflation to compute a
path that is exactly as safe as desired.
Linear chance constraints are used to compute bounded collision-free
trajectories with dynamic obstacles in zhu2019RAL . Axelrod et al.
axelrod2018IJRR focus exclusively on obstacle uncertainty. They formalize a
notion of “shadows”, which are the geometric equivalent of confidence
intervals for uncertain obstacles. The shadows fundamentally give rise to
loose bounds but the computational complexity of bounding the collision
probability is greatly reduced. Uncertain obstacles are modelled as polytopes
with Gaussian-distributed faces in shimanuki2018WAFR . Planning a collision-
free path in the presence of “risk zones” is considered in salzman2017ICAPS
by penalizing the time spent in these risk zones. Risk contour maps which give
the risk information (uncertainties in location, size and geometry of
obstacles) in uncertain environments are used in jasour2019RSS to obtain safe
paths with bounded risks. A related approach for randomly moving obstacles is
presented in hakobyan2019RAL . Formal verification methods have also been used
to construct safe plans ding2013ICRA ; sadigh2016RSS .
Most approaches discussed above compute the collision probability along a path
by summing or multiplying the probabilities along different waypoints in the
path. Boole’s inequality is used to decouple the total probability in terms of
individual waypoint probabilities. Such approaches tend to be overly
conservative and rather than computing bounded collision probabilities along a
path, the bound should be checked for each configuration along a path.
Moreover, in most approaches, the collision probability computed along each
waypoint is an approximation of the true value. On the one hand, such
approximations can overly penalize paths and could gauge all plans to be
infeasible. On the other hand some approximations can be lower444For example,
the approach in dutoit2011IEEE computes a value lower than the actual when
the robot state covariance is small. than the true collision probability
values and can lead to synthesizing unsafe plans.
### 1.3 Contributions
In this paper two main theoretical contributions are presented. First, we
incorporate object uncertainties in the BSP planning framework and derive the
resulting Bayes filter in terms of the prediction and measurement updates of
the Extended Kalman Filter (EKF). The second is the computation of the
probability of collision under environment uncertainty. We formulate the
collision avoidance constraint as a quadratic form in random variables. This
provides an exact expression for the collision probability in terms of a
converging infinite series. A notion of safety is also formalized to compute
configurations that satisfy the required collision probability bounds.
We make the following assumptions: (1) the uncertainties are modelled using
Gaussian distributions; (2) while formulating the collision constraint, we
assume that the robot and obstacles have circular geometries. However, this is
by no means a limitation and the approach can be extended to objects with
different geometries by considering the configuration spaces.
## 2 Object Uncertainty
In this Section, we focus on a BSP formulation that incorporates object
uncertainties, that is, the viewpoints whence the objects can be observed are
not precisely known. We define the object space
$\mathcal{O}=\\{O^{i}|\text{$O^{i}$ is an object, and}\ 1\leq
i\leq|\mathcal{O}|\\}$ to be the set of all objects in the environment. The
motion (1) and observation (2) models can be written in a probabilistic
framework as $p(\textbf{x}_{k+1}|\textbf{x}_{k},\textbf{u}_{k})$ and
$p(\textbf{z}_{k}|\textbf{x}_{k},O_{k}^{i})$, respectively. Let us consider
that at time $k$ the robot received a measurement $\textbf{z}_{k}$ which was
originated by observing object $O^{i}_{k}$. Given an initial distribution
$p(\textbf{x}_{0})$, and the motion and observation models
$p(\textbf{x}_{k+1}|\textbf{x}_{k},\textbf{u}_{k})$ and
$p(\textbf{z}_{k}|\textbf{x}_{k},O_{k}^{i})$, the posterior probability
distribution at time $k$ is the belief $b[\textbf{x}_{k}]$ and can be written
as
$p(\textbf{x}_{k}|\textbf{z}_{k},O_{k}^{i},\textbf{z}_{0\mathrel{\mathop{\mathchar
58\relax}}k-1},\textbf{u}_{0\mathrel{\mathop{\mathchar 58\relax}}k-1})$, where
$O^{i}_{k}$ is the object observed at time $k$,
$\textbf{z}_{0\mathrel{\mathop{\mathchar
58\relax}}k-1}\doteq\\{\textbf{z}_{0},...,\textbf{z}_{k-1}\\}$ is the sequence
of measurements up to $k-1$ and $\textbf{u}_{0\mathrel{\mathop{\mathchar
58\relax}}k-1}\doteq\\{\textbf{u}_{0},...,\textbf{u}_{k-1}\\}$ is the sequence
of controls up to $k-1$. Using Bayes rule and theorem of total probability,
$b[\textbf{x}_{k}]$ can be expanded as
$p(\textbf{x}_{k}|\textbf{z}_{k},O_{k}^{i},\textbf{z}_{0\mathrel{\mathop{\mathchar
58\relax}}k-1},\textbf{u}_{0\mathrel{\mathop{\mathchar 58\relax}}k-1})\\\
=\eta_{k}p(\textbf{z}_{k}|\textbf{x}_{k},O_{k}^{i})p(O_{k}^{i}|\textbf{x}_{k})\int_{\textbf{x}_{k-1}}p(\textbf{x}_{k}|\textbf{x}_{k-1},\textbf{u}_{k-1})b[\textbf{x}_{k-1}]$
(3)
where $\eta_{k}=1/p(\textbf{z}_{k}|\textbf{z}_{0\mathrel{\mathop{\mathchar
58\relax}}k-1},\textbf{u}_{0\mathrel{\mathop{\mathchar 58\relax}}k-1})$ is the
normalization constant and
$b[\textbf{x}_{k-1}]\sim\mathcal{N}(\bm{\mu_{k-1}},\Sigma_{k-1})$ is the
belief at time $k-1$. The term $p(O_{k}^{i}|\textbf{x}_{k})$ denotes the
probability of observing the object $O_{k}^{i}$ from the pose $\textbf{x}_{k}$
and models the object uncertainty. Similarly, given an action
$\textbf{u}_{k}$, the propagated belief can be written as
$b[\bar{\textbf{x}_{k+1}}]=\int_{\textbf{x}_{k}}p(\textbf{x}_{k+1}|\textbf{x}_{k},\textbf{u}_{k})b[\textbf{x}_{k}]$
(4)
Given the current belief $b[\textbf{x}_{k}]$ and the control $\textbf{u}_{k}$,
the propagated belief parameters, that is, mean and covariance, can be
computed using the standard EKF prediction as
$\begin{split}\bar{\bm{\mu}}_{k+1}&=f(\bm{\mu}_{k},\bm{u}_{k})\\\
\bar{\Sigma}_{k+1}&=F_{k}\Sigma_{k}F_{k}^{T}+R_{k}\end{split}$ (5)
where $F_{k}$ is the Jacobian of $f(\cdot)$ with respect to $\textbf{x}_{k}$.
To compute the posterior belief using EKF update equations, we first need to
model the term $p(O_{k}^{i}|\textbf{x}_{k})$. In this work we model the object
distribution as a Gaussian distribution given by
$p(O_{k}^{i}|x_{k})\sim\mathcal{N}(\bm{\mu}_{O_{k}^{i}},\Sigma_{O_{k}^{i}})$
(6)
where $\bm{\mu}_{O_{k}^{i}}$ is the mean viewpoint/pose that corresponds to
the maximum probability of observing $O_{k}^{i}$ and $\Sigma_{O_{k}^{i}}$ is
the associated covariance.
For convenience we state the probability density function (pdf) of
multivariate Gaussian distributions. For
$\textbf{x}\sim\mathcal{N}(\bm{\mu},\Sigma)$ the pdf is of the form
$p(\textbf{x})=det\left(2\pi\Sigma\right)^{-\frac{1}{2}}\textrm{exp}\left(-\frac{1}{2}(\textbf{x}-\bm{\mu})^{T}\Sigma^{-1}(\textbf{x}-\bm{\mu})\right)$
(7)
where $det(\cdot)$ denotes the determinant. Expanding the right hand side of
(3), we have
$b[\textbf{x}_{k+1}]=\eta^{\prime}_{k}\int\text{exp}(-\mathcal{J}_{k+1})$,
where $\eta^{\prime}_{k}$ contains the non-exponential terms and
$\mathcal{J}_{k+1}$ is given by
$\mathcal{J}_{k+1}=\frac{1}{2}\left(\textbf{z}_{k+1}-h\left(\bar{\bm{\mu}}_{k+1}\right)-H_{k+1}\left(\textbf{x}_{k+1}-\bar{\bm{\mu}}_{k+1}\right)\right)^{T}\\\
Q_{k+1}^{-1}\left(\textbf{z}_{k+1}-h\left(\bar{\bm{\mu}}_{k+1}\right)-H_{k+1}\left(\textbf{x}_{k+1}-\bar{\bm{\mu}}_{k+1}\right)\right)\\\
+\frac{1}{2}(\bm{x}_{k+1}-\bm{\mu}_{O_{k+1}^{i}})^{T}\Sigma_{O_{k+1}^{i}}^{-1}(\textbf{x}_{k+1}-\bm{\mu}_{O_{k+1}^{i}})\\\
+\frac{1}{2}(\textbf{x}_{k+1}-\bar{\bm{\mu}}_{k+1})^{T}\bar{\Sigma}_{k+1}^{-1}(\textbf{x}_{k+1}-\bar{\bm{\mu}}_{k+1})$
(8)
where $H_{k+1}$ is the Jacobian of $h(\cdot)$ with respect to
$\textbf{x}_{k+1}$. We note that when object uncertainty is not considered,
the second term in (8) disappears and the results that we derive below reduce
to that of the standard EKF update case. The parameters of this Gaussian can
be obtained by taking the first and second derivatives of $\mathcal{J}_{k+1}$
with respect to $\textbf{x}_{k+1}$,
$\frac{\partial\mathcal{J}_{k+1}}{\partial\textbf{x}_{k+1}}=-H_{k+1}^{T}Q_{k+1}^{-1}\left(\textbf{z}_{k+1}-h(\bar{\bm{\mu}}_{k+1})-\right.\\\
\left.H_{k+1}(\textbf{x}_{k+1}-\bar{\bm{\mu}}_{k+1})\right)+\Sigma_{O_{k+1}^{i}}^{-1}\left(\textbf{x}_{k+1}-\bm{\mu}_{O_{k+1}^{i}}\right)+\\\
\bar{\Sigma}_{k+1}^{-1}\left(\textbf{x}_{k+1}-\bar{\bm{\mu}}_{k+1}\right)$ (9)
$\frac{\partial^{2}\mathcal{J}_{k+1}}{\partial\textbf{x}_{k+1}^{2}}=H_{k+1}^{T}Q_{k+1}^{-1}H_{k+1}+\Sigma_{O_{k+1}^{i}}^{-1}+\bar{\Sigma}_{k+1}^{-1}$
(10)
The term (10) is the inverse of the covariance of $b[\textbf{x}_{k+1}]$
thrun2005book , that is,
$\Sigma_{k+1}=\left(H_{k+1}^{T}Q_{k+1}^{-1}H_{k+1}+\Sigma_{O_{k+1}^{i}}^{-1}+\bar{\Sigma}_{k+1}^{-1}\right)^{-1}$
(11)
Since the mean of $b[\textbf{x}_{k+1}]$ is the value that minimizes
$\mathcal{J}_{k+1}$, it is obtained by equating (9) to zero
$H_{k+1}^{T}Q_{k+1}^{-1}\left(\textbf{z}_{k+1}-h\left(\bar{\bm{\mu}}_{k+1}\right)-H_{k+1}\left(\textbf{x}_{k+1}-\bar{\mu}_{k+1}\right)\right)\\\
=\Sigma_{k+1}^{-1}\left(\bm{\mu}_{k+1}-\bar{\bm{\mu}}_{k+1}\right)-\Sigma_{O_{k+1}^{i}}^{-1}\left(\bm{\mu}_{O_{k+1}^{i}}-\bar{\bm{\mu}}_{k+1}\right)\\\
\implies\bm{\mu}_{k+1}=\bar{\bm{\mu}}_{k+1}+K_{k+1}\left(\textbf{z}_{k+1}-h\left(\bar{\bm{\mu}}_{k+1}\right)\right)\\\
+\Sigma_{k+1}\Sigma_{O_{k+1}^{i}}^{-1}\left(\bm{\mu}_{O_{k+1}^{i}}-\bar{\bm{\mu}}_{k+1}\right)$
(12)
where $K_{k+1}=\Sigma_{k+1}H_{k+1}^{T}Q_{k+1}^{-1}$ is the Kalman gain.
As in the case of standard EKF, the gain $K_{k+1}$ can be transformed to an
expression that does not depend on $\Sigma_{k+1}$, by post-multiplying with an
identity matrix $I=AA^{-1}$, where
$A=\\\
\left(H_{k+1}\bar{\Sigma}_{k+1}\left(\bar{\Sigma}_{k+1}+\Sigma_{O_{k+1}^{i}}\right)^{-1}\Sigma_{O_{k+1}^{i}}H_{k+1}^{T}+Q_{k+1}\right)$
(13)
This gives
$K_{k+1}=\Sigma_{k+1}\left(H_{k+1}^{T}Q_{k+1}^{-1}H_{k+1}\bar{\Sigma}_{k+1}\left(\bar{\Sigma}_{k+1}+\Sigma_{O_{k+1}^{i}}\right)^{-1}\right.\\\
\left.\Sigma_{O_{k+1}^{i}}H_{k+1}^{T}+H_{k+1}^{T}\right)A^{-1}$ (14)
In order to simplify the above expression for $K_{k+1}$, we first compute the
inverse of the term
$\bar{\Sigma}_{k+1}\left(\bar{\Sigma}_{k+1}+\Sigma_{O_{k+1}^{i}}\right)^{-1}\Sigma_{O_{k+1}^{i}}$
(15)
The inverse is computed as
$\left(\bar{\Sigma}_{k+1}\left(\bar{\Sigma}_{k+1}+\Sigma_{O_{k+1}^{i}}\right)^{-1}\Sigma_{O_{k+1}^{i}}\right)^{-1}\\\
=\Sigma_{O_{k+1}^{i}}^{-1}\left(\bar{\Sigma}_{k+1}+\Sigma_{O_{k+1}^{i}}\right)\bar{\Sigma}_{k+1}^{-1}\\\
=\Sigma_{O_{k+1}^{i}}^{-1}\bar{\Sigma}_{k+1}\bar{\Sigma}_{k+1}^{-1}+\Sigma_{O_{k+1}^{i}}^{-1}\Sigma_{O_{k+1}^{i}}\bar{\Sigma}_{k+1}^{-1}\\\
=\Sigma_{O_{k+1}^{i}}^{-1}+\bar{\Sigma}_{k+1}^{-1}$ (16)
Using (16) and (11), the expression in (14) simplifies to
$K_{k+1}=\Sigma_{k+1}\left(H_{k+1}^{T}Q_{k+1}^{-1}H_{k+1}+\Sigma_{O_{k+1}^{i}}^{-1}+\bar{\Sigma}_{k+1}^{-1}\right)\bar{\Sigma}_{k+1}\\\
\left(\bar{\Sigma}_{k+1}+\Sigma_{O_{k+1}^{i}}\right)^{-1}\Sigma_{O_{k+1}^{i}}H_{k+1}^{T}\\\
\left(H_{k+1}\bar{\Sigma}_{k+1}\left(\bar{\Sigma}_{k+1}+\Sigma_{O_{k+1}^{i}}\right)^{-1}\Sigma_{O_{k+1}^{i}}H_{k+1}^{T}+Q_{k+1}\right)^{-1}\\\
=\bar{\Sigma}_{k+1}\left(\bar{\Sigma}_{k+1}+\Sigma_{O_{k+1}^{i}}\right)^{-1}\Sigma_{O_{k+1}^{i}}H_{k+1}^{T}\\\
\left(H_{k+1}\bar{\Sigma}_{k+1}\left(\bar{\Sigma}_{k+1}+\Sigma_{O_{k+1}^{i}}\right)^{-1}\Sigma_{O_{k+1}^{i}}H_{k+1}^{T}+Q_{k+1}\right)^{-1}\\\
\vspace{-0.2cm}$ (17)
By treating the sum $\Sigma_{O_{k+1}^{i}}^{-1}+\bar{\Sigma}_{k+1}^{-1}$ in
(11) as a single term and applying the matrix inversion lemma on the right
hand side of (11) and further simplifying using the expression for the inverse
computed in (16), it can be shown that
$\Sigma_{k+1}=\left(I-K_{k+1}H_{k+1}\right)\bar{\Sigma}_{k+1}\left(\bar{\Sigma}_{k+1}+\Sigma_{O_{k+1}^{i}}\right)^{-1}\Sigma_{O_{k+1}^{i}}$
(18)
We note that when no object uncertainty is considered the update step of the
standard EKF gives
$\bm{\mu}_{k+1}=\bar{\bm{\mu}}_{k+1}+K_{k+1}\left(\textbf{z}_{k+1}-h\left(\bar{\bm{\mu}}_{k+1}\right)\right)$
and $\Sigma_{k+1}=\left(I-K_{k+1}H_{k+1}\right)\bar{\Sigma}_{k+1}$. The
additional term in (12) rightly adjusts the mean $\bm{\mu}_{k+1}$ accounting
for the fact that the object location is uncertain. Similarly, the extra terms
in (18) account for the object uncertainty and scale the posterior covariance
accordingly.
## 3 Collision Probability
Let $\mathcal{R}$ represent the set of all points occupied by a rigid-body
robot at any given time. Thus, $\mathcal{R}$ represents the collection of
points that form the rigid-body robot. Similarly, let $\mathcal{S}$ represent
the set of all points occupied by a rigid-body obstacle. A collision occurs if
$\mathcal{R}\cap\mathcal{S}\neq\\{\phi\\}$ and we denote the probability of
collision as $P\left(\mathcal{R}\cap\mathcal{S}\neq\\{\phi\\}\right)$. In this
work we assume circular geometries for $\mathcal{R}$ and $\mathcal{S}$ with
radii $r_{1}$ and $s_{1}$, receptively and we denote the center of mass of the
robot and the obstacle by $\textbf{x}_{k}$ and s, receptively. By abuse of
notation we will use $\textbf{x}_{k}$ and s equivalently to $\mathcal{R}$ and
$\mathcal{S}$. The collision condition will be written in terms of the center
of mass as $\mathcal{C}_{\textbf{x}_{k},\textbf{s}}\mathrel{\mathop{\mathchar
58\relax}}\mathcal{R}\cap\mathcal{S}\neq\\{\phi\\}$. It is noteworthy that
both $\textbf{x}_{k}$ and s are not known precisely but can only be estimated
probabilistically, as seen in the previous section. At this point we would
like to stress the fact that the concepts and the derivations herein are valid
for any 2D rigid-body robot. A mobile robot may be represented by a minimum
area enclosing circle. In the case of a 2D manipulator robot each link can be
approximated by bounding circles that tightly enclose the link. For such
robots, the collision with an obstacle has to be checked for each bounding
circle. For example, consider a manipulator robot with $l$ bounding circles.
Then the collision condition for the $i-$th circle ($1\leq i\leq l$) is given
by $\mathcal{C}_{\textbf{x}_{k}^{i},\textbf{s}}$, where $\textbf{x}_{k}^{i}$
is the center of the $i-$th circle.
Let us now consider an obstacle at any given time instant, distributed
according to the Gaussian
$\textbf{s}\sim\mathcal{N}\left(\bar{\textbf{s}},\Sigma_{s}\right)$, where
$\bar{s}$ represents the mean and $\Sigma_{s}$ the uncertainty in the
estimation of the object. Given the belief at time $k$, that is,
$b[\textbf{x}_{k}]$, the probability of collision is given by
$P\left(\mathcal{C}_{\textbf{x}_{k},\textbf{s}}\right)=\int_{\textbf{x}_{k}}\int_{\textbf{s}}I_{c}(\textbf{x}_{k},\textbf{s})p(\textbf{x}_{k},\textbf{s})$
(19)
where $\mathcal{C}_{\textbf{x}_{k},\textbf{s}}$ as defined above represents
the fact that robot configuration $\textbf{x}_{k}$ and its collision with
obstacle at location s is considered, and $I_{c}$ is an indicator function
defined as
$I_{c}(\textbf{x}_{k},\textbf{s})=\begin{cases}1\ &\text{if}\
\mathcal{R}\cap\mathcal{S}\neq\\{\phi\\}\\\ 0\ &\text{otherwise}.\end{cases}$
(20)
Du Toit and Burdick dutoit2011IEEE , Park et al. park2018IEEE approximate the
integral in (19) as $Vp(\textbf{x}_{k},\textbf{s})$, where $V$ is the 2D
footprint (area) occupied by the robot. For this approximation, in
dutoit2011IEEE it is assumed that the robot radius $\varepsilon$ is
negligible and a point obstacle is considered for this derivation.
To do away with this approximation, we formulate the above problem by
considering an alternative approach. Since the robot and obstacle are assumed
to be spherical objects, the collision constraint can be written as
$\mathinner{\\!\left\lVert\textbf{x}_{k}-\textbf{s}\right\rVert}^{2}\leq(r_{1}+s_{1})^{2}$
(21)
where $\textbf{x}_{k}$ and s are the random vectors that denote the robot and
obstacle pose respectively. Here, $\textbf{x}_{k}$ and s corresponds to the
body-fixed frames in the global frame. As noted before, the two random vectors
in (19) are distributed according to
$\textbf{s}\sim\mathcal{N}\left(\mathbf{\bar{\textbf{s}}},\Sigma_{s}\right)$
and $\textbf{x}_{k}\sim\mathcal{N}\left(\bm{\mu}_{k},\Sigma_{k}\right)$. Let
us denote by $\textbf{w}=\textbf{x}_{k}-\textbf{s}$, the difference between
the two random variables. Then we know that w is also a Gaussian, distributed
as
$\textbf{w}\sim\mathcal{N}\left(\bm{\mu}_{k}-\bm{\bar{s}},\Sigma_{k}+\Sigma_{s}\right)$.
The collision constraint can now be written as
$\textbf{v}=\mathinner{\\!\left\lVert\textbf{w}\right\rVert}^{2}=\textbf{w}^{T}\textbf{w}\leq(r_{1}+s_{1})^{2}$
(22)
where v is a random vector distributed according to the squared $L_{2}$-norm
of w. Now, given the probability density function (pdf) of v, the collision
constraint in (21)reduces to solving the integral
$P\left(\mathcal{C}_{\textbf{x}_{k},\textbf{s}}\right)=\int_{0}^{(r_{1}+s_{1})^{2}}p(v)$
(23)
where $p(v)=P_{\textbf{v}}(\textbf{v}=v)$ is the pdf of v. It is noteworthy
that the above expression is the cumulative distribution function (cdf) of v,
which is defined as
$F_{\textbf{v}}\left((r_{1}+s_{1})^{2}\right)=P\left(\textbf{v}\leq(r_{1}+s_{1})^{2}\right)$.
### 3.1 Quadratic Form in Random Variables
A quadratic form in random variables is defined as provost1992book ,
###### Definition 1
Let $\textbf{x}=\left(x_{1},\ldots,x_{n}\right)^{T}$ denote a random vector
with mean $\bm{\mu}=\left(\mu_{1},\ldots,\mu_{n}\right)^{T}$ and covariance
matrix $\Sigma$. Then the quadratic form in the random variables
$x_{1},\ldots,x_{n}$ associated with an $n\times n$ symmetric matrix
$A=(a_{ij})$ is
$Q(\textbf{x})=Q(x_{1},\ldots,x_{n})=\textbf{x}^{T}A\textbf{x}=\sum_{i=1}^{n}\sum_{j=1}^{n}a_{ij}X_{i}X_{j}$
(24)
Let us define $\textbf{y}=\Sigma^{-\frac{1}{2}}\textbf{x}$ and define a random
vector $\textbf{z}=\left(\textbf{y}-\Sigma^{-\frac{1}{2}}\bm{\mu}\right)$. The
resulting distribution of z is thus zero mean with covariance being the
identity matrix. Thus the quadratic form becomes
$Q(\textbf{x})=\left(\textbf{z}+\Sigma^{-\frac{1}{2}}\bm{\mu}\right)^{T}\Sigma^{\frac{1}{2}}A\Sigma^{\frac{1}{2}}\left(\textbf{z}+\Sigma^{-\frac{1}{2}}\bm{\mu}\right)$
(25)
Suppose there exists an orthogonal matrix $P$, that is, $PP^{T}=I$ which
diagonalizes $\Sigma^{\frac{1}{2}}A\Sigma^{\frac{1}{2}}$, then
$P^{T}\Sigma^{\frac{1}{2}}A\Sigma^{\frac{1}{2}}P=\textrm{diag}\left(\lambda_{1},\ldots,\lambda_{n}\right)$,
where $\lambda_{1},\ldots,\lambda_{n}$ are the eigenvalues of
$\Sigma^{\frac{1}{2}}A\Sigma^{\frac{1}{2}}$. The quadratic form can now be
written as
$\begin{split}Q(\textbf{x})&=\left(\textbf{z}+\Sigma^{-\frac{1}{2}}\bm{\mu}\right)^{T}\Sigma^{\frac{1}{2}}A\Sigma^{\frac{1}{2}}\left(\textbf{Z}+\Sigma^{-\frac{1}{2}}\bm{\mu}\right)\\\
&=\left(\textbf{u}+\textbf{b}\right)^{T}\textrm{diag}\left(\lambda_{1},\ldots,\lambda_{n}\right)\left(\textbf{u}+\textbf{b}\right)\end{split}$
(26)
where $\textbf{u}=P^{T}\textbf{z}=(u_{1},\ldots,u_{n})^{T}$ and
$\textbf{b}=P^{T}\Sigma^{-\frac{1}{2}}\bm{\mu}=(b_{1},\ldots,b_{n})^{T}$. The
expression in (26) can be written concisely,
$Q(\textbf{x})=\textbf{x}^{T}A\textbf{x}=\sum_{i=1}^{n}\lambda_{i}(u_{i}+b_{i})^{2}$
(27)
###### Theorem 3.1
The cdf of
$Q(\textnormal{{x}})=\textnormal{{y}}=\textnormal{{x}}^{T}A\textnormal{{x}}$
with $A=A^{T}>0,\textnormal{{x}}\sim\mathcal{N}(\bm{\mu},\Sigma),\Sigma>0$ is
$F_{\textnormal{{y}}}(y)=P(\textnormal{{y}}\leq
y)=\sum_{k=0}^{\infty}(-1)^{k}c_{k}\frac{y^{\frac{n}{2}+k}}{\Gamma\left(\frac{n}{2}+k+1\right)}$
(28)
and its pdf is given by
$p_{\textnormal{{y}}}(y)=P(\textnormal{{y}}=y)=\sum_{k=0}^{\infty}(-1)^{k}c_{k}\frac{y^{\frac{n}{2}+k-1}}{\Gamma\left(\frac{n}{2}+k\right)}$
(29)
where $\Gamma$ denotes the gamma function and
$\begin{split}&c_{0}=\textrm{exp}(-\frac{1}{2}\sum\limits_{i=1}^{n}b_{i}^{2})\prod_{i=1}^{n}\left(2\lambda_{i}\right)^{-\frac{1}{2}}\\\
&c_{k}=\frac{1}{k}\sum\limits_{i=0}^{k-1}d_{k-i}c_{i}\\\
&d_{k}=\frac{1}{2}\sum\limits_{i=1}^{n}\left(1-kb_{i}^{2}\right)\left(2\lambda_{i}\right)^{-k}\end{split}$
The proof of the above theorem is beyond the scope of this paper and we refer
the interested readers to provost1992book . It is easily seen that the left
hand side of (22), is in the quadratic form $Q(\textbf{y})$ with $A=I$, the
identity matrix. Thus the collision probability can be computed from (28) as
$P\left(\mathcal{C}_{\textbf{x}_{k},\textbf{s}}\right)=F_{\textbf{y}}\left((r_{1}+s_{1})^{2}\right)$
(30)
### 3.2 Convergence and Truncation Error
In this section we will prove the convergence the infinite series in (28) and
(29). Note that the series expansion of the pdf in Theorem 29 is of the form
$p_{\textbf{y}}(y)=\sum\limits_{k=0}^{\infty}c_{k}h_{k}(y)$ (31)
From kotz1967AMS we have the following lemma.
###### Lemma 1
Let $\\{h_{k}\\}_{0}^{\infty}$ be a sequence of measurable complex valued
functions on $[0,\infty]$ and $\\{c_{k}\\}_{0}^{\infty}$ be a sequence of
complex numbers such that
$\sum_{k=0}^{\infty}|c_{k}||h_{k}(y)|\leq\alpha e^{(\beta y)}\ \textrm{for}\
y\in[0,\infty]$ (32)
where $\alpha$, $\beta$ are real constants. Then $L\left(h_{k}(y)\right)$ and
$L(p_{\textbf{y}}(y))$ exist for $Re(s)>\beta$, and
$L(p_{\textbf{y}}(y))=\sum_{k=0}^{\infty}c_{k}L(h_{k}(s))$ (33)
where $L(\cdot)$ denotes the Laplace transform. Let us now define the term
$M(\theta)$ such that
$M(\theta)=\sum_{k=0}^{\infty}c_{k}\theta^{k}$ (34)
where the infinite series is a uniformly convergent series for $\theta$ in
some region with $M(\theta)>0$. Let the Laplace transform of $h_{k}(y)$ be the
form $L(h_{k}(y))=\xi(s)\eta^{k}(s)$, where, for $Re(s)>\beta$ with $\beta$
being a real constant, $\xi(s)$ is a non-vanishing analytic function and
$\eta(s)$ is an analytic function with an inverse function
$\eta(\zeta(\theta))=\theta$. For $h_{k}(y)$ in (29), we have,
$\xi(s)=(2s)^{-n/2}$, $\eta(s)=-(2s)^{-1}$ and
$\zeta(\theta)=-(2\theta)^{-1}$. Now let us define,
$\begin{split}M(\theta)&=\left(L(p_{\textbf{y}})\circ\zeta/\xi\circ\zeta\right)(\theta)=\sum_{k=0}^{\infty}c_{k}\theta^{k}\end{split}$
(35)
where $\circ$ denotes function composition. Using Cauchy’s inequality, we get
$|c_{k}|\leq\frac{m(\rho)}{\rho^{k}},\ \
m(\rho)=\textrm{max}_{|\theta|=\rho}|M(\theta)|$ (36)
Since $h_{k}(y)$ is bounded and, using (36), the condition (32) in Lemma 33 is
satisfied and the series $p_{\textbf{y}}(y)$ converges uniformly in every
bounded interval of $y>0$. As a result, integrating $p_{\textbf{y}}(y)$ term-
by-term, the obtained series $F_{\textbf{y}}(y)$ is uniformly convergent in
every bounded interval of $y>0$.
If the series in (29) is truncated after $N$ terms, the truncation error is
$e(N)=\sum_{k=N+1}^{\infty}|c_{k}h_{k}(y)|=\left|\sum_{k=N+1}^{\infty}c_{k}\frac{y^{\frac{n}{2}+k-1}}{\Gamma\left(\frac{n}{2}+k\right)}\right|$
(37)
Using (36), an upper bound for the truncation error can hence be obtained as
$e(N)\leq\frac{m(\rho)}{\rho^{k}}\left|\sum_{k=N+1}^{\infty}\frac{y^{\frac{n}{2}+k-1}}{\Gamma\left(\frac{n}{2}+k\right)}\right|$
(38)
where the summation term can be further simplified using the gamma function
identity, $\forall\varsigma>0,\
\Gamma(\varsigma+1)=\varsigma\Gamma(\varsigma)$, giving
$e(N)\leq
m(\rho)\left(\Gamma\left(\frac{n}{2}\right)N!\right)^{-1}(\frac{y}{2})^{\frac{n}{2}-1}(\frac{y}{2\rho})^{N+1}\text{exp}(\frac{y}{2\rho})$
(39)
The truncation error for (28) is obtained in a similar manner,
$\displaystyle E(N)\leq
m(\rho)\left(\Gamma\left(\frac{n}{2}\right)(N+1)!\right)^{-1}(\frac{y}{2})^{\frac{n}{2}}(\frac{y}{2\rho})^{N+1}\text{exp}(\frac{y}{2\rho})$
(40)
The expression for $m(\rho)$ is obtained from kotz1967AMS2 ,
$m(\rho)=\prod_{j=1}^{n}\lambda_{j}^{-\frac{1}{2}}\text{exp}\left(-\frac{1}{2}\sum_{j=1}^{n}\frac{b_{j}^{2}\lambda_{j}}{\lambda_{j}+\rho}\right)\prod_{j=1}^{n}(1-\frac{\rho}{\lambda_{j}})^{-\frac{1}{2}}$
(41)
The expression in (41) is valid only if $\rho<\lambda_{j}$ kotz1967AMS and
hence $\rho<\textrm{min}\ \lambda_{j}$. Thus we have $m(\rho)\rightarrow 0$
with $\sum_{j=1}^{n}b_{j}^{2}\rightarrow\infty$. The larger the distance from
the obstacles and the higher the certainty in the robot and obstacle
positions, the greater is the $b_{j}$ (see 26) value. In such scenarios,
convergence is often attained within the first few terms of the series. For a
given robot configuration and obstacle parameters, we see that the only
varying term in (40) is $(y/2\rho)^{N+1}/(N+1)!$ which depends on
$\lambda_{j}$’s, that is the eigenvalues of $\Sigma_{k}+\Sigma_{s}$. Clearly,
at time instant $k$, the parameter that influences the convergence is the
degree of uncertainty in both the robot and obstacle location, that is,
$\Sigma_{k}+\Sigma_{s}$.
The convergence is visualized for different configurations in Fig. 2. The blue
and green circles represent a robot and an obstacle, respectively. The red
ellipses corresponds to the 3$\sigma$ uncertainties for different covariances
$diag(0.04,0.04),\ diag(0.08,0.08),\ \ldots,\ diag(0.74,0.74)$. In Fig. 2(a)
the robot and the obstacle are touching each other. For each of these
covariances, the number of terms for convergence is shown in Fig. 2(b). The
worst case corresponds to the covariance of $diag(0.04,0.04)$, requiring 16
terms for convergence (dashed blue line with spikes in Fig. 2(b)). In Fig.
2(c) the distance between the robot and the obstacle is increased by 0.2$m$
and the covariance $diag(0.04,0.04)$ needed 12 terms for convergence. The
distances are further increased by 0.4$m$ and 0.8$m$ in Fig. 2(e), (g) and
their worst case convergences are 9 and 5 respectively as seen in Fig.2(f),
(h). The number of terms for worst case convergence that corresponds to
covariance $diag(0.04,0.04)$ and the respective time for collision probability
computation are shown in Table 1.
Configuration | Terms for convergence | Computation time (s)
---|---|---
A | 16 | 0.0412 $\pm$ 0.0086
B | 12 | 0.0044 $\pm$ 0.0041
C | 9 | 0.0008 $\pm$ 0.0003
D | 5 | 0.0004 $\pm$ 0.0002
Table 1: The maximum number of terms required for convergence and the
corresponding collision probability computation time. The values correspond to
the covariance $diag(0.04,0.04)$ for each of the configurations.
(a) Configuration A
(b) Collision probability evolution
(c) Configuration B
(d) Collision probability evolution
(e) Configuration C
(f) Collision probability evolution
(g) Configuration D
(h) Collision probability evolution
Figure 2: Different configurations for a robot of radius 0.3$m$ and obstacle
of radius 0.5$m$. For each configuration the evolution of probability of
collision is plotted for different covariances. In each of the 4
configurations, maximum terms for convergence is for the minimum covariance of
$diag(0.04,0.04)$.
### 3.3 Safe Configuration
In the presence of perception and motion uncertainty, providing safety
guarantees for robot motion is imperative. Let us assume that the obstacle
position is known with high certainty as a result of perfect sensing. However,
since the true state of the robot is not known and only a distribution of
these states can be estimated, collision checking has to be performed for this
distribution of states. Moreover, in practice, the observations are noisy and
this renders the estimated obstacle location (and shape) uncertain. Hence,
this uncertainty should be taken into account while considering collision
avoidance.
Given a robot configuration $\textbf{x}_{k}$, we define the following notion
of $\epsilon-$safe configuration.
###### Definition 2
A robot configuration $\textbf{x}_{k}$ is an $\epsilon-$safe configuration
with respect to an obstacle location s, if the probability of collision is
such that $P\left(\mathcal{C}_{\textbf{x}_{k},\textbf{s}}\right)\leq
1-\epsilon$.
For example, a $0.99-$safe configuration implies that the probability of this
configuration colliding with the obstacle is at most $0.01$. We use the
sampling based Probabilistic Roadmap (PRM) kavraki1996IEEE to compute motion
plans. As a result we can only guarantee probabilistic completeness for
returning $\epsilon-$safe configurations since the PRM motion planner is
probabilistically complete karaman2011IJRR , that is the probability of
failure decays to zero exponentially with the number of samples used in the
construction of the roadmap. The failure to find an $\epsilon-$safe
configuration might be because such a configuration indeed does not exist or
simply because there were not enough samples.
### 3.4 Complexity Analysis
It is known that for $m$ nodes, the computational complexity of PRM is
$O(m\log m)$ karaman2011IJRR . First let us consider the case of belief space
planning over the PRM graph, without computing the collision probabilities.
Finding a trajectory to the goal requires performing Bayesian (EKF) update
operations. This basically involves performing matrix operations— matrix
multiplication and inversion of matrices. For a state of dimension $n$, the
covariance matrix is of dimension $O(n^{2})$. Therefore, each step of the
Bayesian update has a complexity of $O(n^{3})$. If $T$ denotes the number of
time steps in the trajectory, then the overall computational complexity is
$O(n^{3}T)$. Let us now analyze the complexity of collision probability
computation. From (40) we see that for each iteration, the truncation error
varies with $(y/2\rho)$. Therefore, to achieve $E(N)\leq\delta$, for an
$\epsilon-$safe configuration,
$k=O\left(\log\frac{\delta\rho}{y(1-\epsilon)}\right)$ iterations are
required. We note that for each obstacle, the runtime is increased by this
factor.
## 4 Cost Function
At each time instant the robot is required to minimize its control usage and
proceed towards the goal $\textbf{x}^{g}$, while minimizing its state
uncertainty. We quantify the state uncertainty by computing the trace of the
marginal covariance of the robot position. As a result, we have the following
cost function
$c\doteq\mathinner{\\!\left\lVert\xi(\textbf{u}_{k})\right\rVert}^{2}_{M_{u}}+\mathinner{\\!\left\lVert\textbf{x}_{k}-\textbf{x}^{g}\right\rVert}^{2}_{M_{g}}+tr\left(\mathinner{\\!\left\lVert
M_{\Sigma}\right\rVert}^{2}_{\Sigma_{k}}\right)+M_{C}P(\mathcal{C})$ (42)
where $\mathinner{\\!\left\lVert x\right\rVert}_{S}=\sqrt{x^{T}Sx}$ is the
Mahalanobis norm, $M_{u},M_{g},M_{C}$ are weight matrices and
$\xi(\textbf{u}_{k})$ is a function that quantifies control usage. The choice
of weight matrices and the control function vary with application. The term
$tr\left(\mathinner{\\!\left\lVert
M_{\Sigma}\right\rVert}^{2}_{\Sigma_{k}}\right)=tr\left(M_{\Sigma}^{T}\Sigma_{k}M_{\Sigma}\right)$,
returns the marginal covariance of the robot location. Therefore,
$M_{\Sigma}=\tau\bar{M}_{\Sigma}$, where $\tau$ is a positive scalar and
$\bar{M}_{\Sigma}$ is a matrix filled with zero or identity entries.
$P(\mathcal{C})$ represents the probability of collision and $M_{C}$ penalizes
the belief states with higher collision probabilities.
The failure to find an $\epsilon-$safe configuration might be because such a
configuration indeed does not exist or simply because there was not enough
samples in the roadmap. In such scenario the roadmap has to be extended.
Different strategies could be implemented to efficiently extend the roadmap
but is not the main focus of the current paper. Therefore we follow a
straightforward approach to add more samples when an $\epsilon-$safe
configuration cannot be found. Given a node from which no $\epsilon-$safe
configuration can be found, a circle of certain radius (half the maximum
distance allowed between two edges) is drawn. Samples are then added to the
roadmap and the PRM graph is updated until an $\epsilon-$safe configuration is
found or until time-out.
## 5 Simulation Results
In this section we first provide a comparison of our approach with
park2018IEEE and dutoit2011IEEE . We then explore the capabilities of our
approach in two simulation domains. Performance are evaluated on an Intel®
Core i7-6500U<EMAIL_ADDRESS>with 8GB RAM under Ubuntu 16.04 LTS.
### 5.1 Comparison to Other Approaches
(a)
(b)
(c)
Figure 3: Comparison of our approach to other methods. (a) The robot state is known perfectly, however the obstacle location is uncertain. (b) Robot state uncertainty is considered (contours in blue). The collision probability value computed with park2018IEEE gave a much higher value. (c) A point-like robot and obstacle are considered. The values computed with park2018IEEE ; dutoit2011IEEE are much lower than expected. Case | Algorithm | Collision probability | Computation time (s) | Feasible
---|---|---|---|---
(a) | Numerical integral | $4.62\%$ | 0.8896 $\pm$ 0.0356 | Yes
Du Toit and Burdick dutoit2011IEEE | $5.84\%$ | 0.0026 $\pm$ 0.0003 | Yes
Park et al. park2018IEEE | $33.26\%$ | 0.2367 $\pm$ 0.2081 | No
Our approach | $4.61\%$ | 0.0232 $\pm$ 0.0024 | Yes
(b) | Numerical integral | $8.25\%$ | 1.2309 $\pm$ 0.0298 | Yes
Du Toit and Burdick dutoit2011IEEE | $14.20\%$ | 0.0021$\pm$ 0.0001 | No
Park et al. park2018IEEE | $36.31\%$ | 0.2108 $\pm$ 0.3067 | No
Our approach | $8.22\%$ | 0.0208 $\pm$ 0.0021 | Yes
(c) | Numerical integral | $14.82\%$ | 1.2450 $\pm$ 0.0301 | No
Du Toit and Burdick dutoit2011IEEE | $0.46\%$ | 0.0019 $\pm$ 0.0004 | Yes
Park et al. park2018IEEE | $0.61\%$ | 0.3145 $\pm$ 0.4610 | Yes
Our approach | $14.83\%$ | 0.0271 $\pm$ 0.0087 | No
Table 2: Comparison of collision probability methods.
Park et al. park2018IEEE approximate the integral in (19) as
$Vp(\textbf{x}_{k},\textbf{s})$, where $V$ is the 2D footprint or area
occupied by the robot. For computing $p(\textbf{x}_{k},\textbf{s})$, they
first assume a distribution centered around the obstacle with the covariance
being the sum of the robot and obstacle location uncertainties. The collision
probability is then computed by finding the $\textbf{x}_{k}$ that maximizes
$p(\textbf{x}_{k},\textbf{s})$ and formulate the problem as an optimization
problem with a Lagrange multiplier. In dutoit2011IEEE the density of the
center of the robot is used. For comparing with these approaches, we formulate
the problem as given in each of these works555For the comparison, the
approaches in dutoit2011IEEE ; park2018IEEE have been reproduced to the best
our understanding and the reproduced codes (including numerical integration
and our approach) can be found here—
https://bitbucket.org/1729antony/comparison_cp_methods/src/master/. In order
to validate the values computed using our approach, we perform numerical
integration of the expression in (19), which gives the exact collision
probability value.
Three different cases are considered as shown in Fig. 3. The solid green
circle denotes an obstacle of radius 0.5m and its corresponding uncertainty
contours are shown as green circles. The solid blue circle denotes a robot of
radius 0.3m with the blue circles showing the Gaussian contours. We define a
collision probability threshold of $0.1$, that is, a $0.9-$safe configuration.
The collision probability values and the computation times are provided in
Table 2. In Fig. 3(a), the robot position known with high certainty and our
approach computes collision probability as $4.61\%$ and hence the given
configuration is a $0.9-$safe configuration. The numerical integral provides
the actual value and as seen in Table 2, it is computed to be $4.62\%$, thus
proving the exactness of our method. However, the collision probability
computed as given in park2018IEEE is $33.26\%$ (almost seven times our
value), predicting the configuration to be unsafe. The approach in
dutoit2011IEEE gave the value of $5.84\%$, a much tighter upper bound. In
Fig. 3(b), there is robot uncertainty along the horizontal axis and the
collision probability computed using our approach is $8.22\%$. The actual
value is computed to be $8.25\%$. As compared to the previous case, the
probability has almost doubled. This is quite intuitive as seen from the robot
uncertainty spread and hence there is greater chance for intersection between
the robot and the obstacle. The value computed using the approach in
park2018IEEE is $36.31\%$ ($4.5$ times our value). The approach in
dutoit2011IEEE also gave a higher value of $14.20\%$. Unlike the approaches
in dutoit2011IEEE ; park2018IEEE our approach rightly predicts the
configuration to be a $0.9-$safe configuration. The higher values obtained
using dutoit2011IEEE ; park2018IEEE are due to the overly conservative nature
of the estimates.
The approach of Park et al. park2018IEEE and dutoit2011IEEE assumes that the
robot radius is very small. We also compute the collision probabilities for a
robot and an obstacle with radius $0.05$m each, where the robot and the
obstacle are touching each other (Fig. 3(c)). The obstacle location is also
much more certain, with the uncertainty reduced by $97\%$ as compared to cases
in Fig. 3(a),(b). Actual value obtained using numerical integral is $14.82\%$.
The probability of collision computed using our approach is $14.83\%$,
whereas, using the approach in park2018IEEE the computed value is $0.61\%$
and the approach in dutoit2011IEEE computes it to be $0.46\%$. Thus our
approach predicts the configuration to be unsafe. To get a sense of the actual
value, we compute the area of the covariance matrix, which is $6.28\times
10^{-4}m^{2}$. This clearly indicates that $0.61\%$ is too small a value and
the configuration is not $0.9-$safe configuration. Using the approaches in
dutoit2011IEEE ; park2018IEEE would lead to collision as it predicts the
configuration to be safe. Our approach computes the exact probability of
collision and outperforms the approaches in dutoit2011IEEE ; park2018IEEE .
### 5.2 2D Environment Domain
(a)
(b)
Figure 4: Simulation environment. (a) Scaled-down ($\times\frac{1}{4}$) top
view of the environment with the sampled roadmap and start and goal locations
of the robot. (b) Pioneer robot at the starting node of the roadmap.
(a)
(b)
(c)
(d)
(e)
(f)
Figure 5: Trajectory and the covariance evolution for single planning instantiations are shown. Different cases with obstacle uncertainty for a point robot and a robot of radius 0.3$m$ are shown in (a), (b), (c) and (d). (e) The planned trajectory when there is uncertainty in beacon locations. (f) True beacon locations are shown in yellow. Approach | Robot radius | Obstacle uncertainty | Beacon (object) uncertainty | Planned trajectory
---|---|---|---|---
Our | Point | No | No | Fig. 5(a)
Our | Point | Yes | No | Fig. 5(b)
Our | 0.3 m | No | No | Fig. 5(a)
dutoit2011IEEE | 0.3 m | No | No | Fig. 5(a)
park2018IEEE | 0.3 m | No | No | Fig. 5(d)
Our | 0.3 m | Yes | No | Fig. 5(c)
Our | 0.3 m | No | Yes | Fig. 5(e)
Our | 0.3 m | No | No (true beacon location) | Fig. 5(f)
Our | 0.3 m | No | No (mean beacon location) | Fig. 5(a)
Table 3: Different configurations used for the 2D environment domain.
We consider the case of a environment where a mobile robot moving in an
environment of $30m\times 20m$. A scaled-down top view is seen in Fig. 4(a).
The underlying PRM graph, the start (S in the figure) and goal (G in the
figure) locations can also be seen. The gray circles denote the obstacles in
the environment. Fig. 4(b) shows a Pioneer P3DX robot at the start location.
For the robot motion model, we consider the following non-linear dynamics
thrun2005book
$\begin{split}x_{k+1}&=x_{k}+\delta_{trans}\cos(\theta_{k}+\delta_{rot1})\\\
y_{k+1}&=y_{k}+\delta_{trans}\sin(\theta_{k}+\delta_{rot1})\\\
\theta_{k+1}&=\theta_{k}+\delta_{rot1}+\delta_{rot2}\end{split}$ (43)
where $\textbf{x}_{k}\doteq(x,y,\theta)$ is the robot pose at time $k$ and
$\textbf{u}_{k}\doteq(\delta_{rot1},\delta_{trans},\delta_{rot2})$ is the
applied control. The model assumes that the robot ideally implements the
following commands in order: rotation by an angle of $\delta_{rot1}$,
translation of $\delta_{trans}$ and a final rotation of $\delta_{rot2}$
orienting the robot in the required direction. The robot accrue translational
and rotational errors while executing $\textbf{u}_{k}$ and localizes itself by
estimating its position using signal measurements from beacons
$\bar{b}_{1},\ldots,\bar{b}_{7}$, which are located at
$(x_{\bar{b}_{1}},y_{\bar{b}_{1}}),\ldots,(x_{\bar{b}_{7}},y_{\bar{b}_{7}})$.
The signal strength decays quadratically with the distance to the beacon,
giving the following observation model with sensor noise $v_{k}$,
$\textbf{z}_{k}=\begin{bmatrix}1/\left((x_{k}-x_{\bar{b}_{1}})^{2}+(y_{k}-y_{\bar{b}_{1}})^{2}+1\right)\\\
\vdots\\\
1/\left((x_{k}-x_{\bar{b}_{7}})^{2}+(y_{k}-y_{\bar{b}_{7}})^{2}+1\right)\end{bmatrix}+v_{k}$
(44)
We validate our approach in the above discussed environment by varying
different parameters, a summary of which is provided in Table 3. Below we
detail each of cases considered in Table 3. We first consider the motion
planning approach for a point-like robot. The cost function is of the form in
(42) with $M_{u}=0.3$, $M_{g}=diag(0.8,0.8)$, $M_{\Sigma}=diag(1,1)$ and
$M_{C}=10$. The underlying PRM graph with 65 nodes is shown in Fig. 5, with
the green dots denoting the sampled nodes. The robot, starting from its
initial belief state (mean pose denoted by S in the figure) has to reach the
node $\textbf{x}_{g}$ (G in the figure), while reducing its uncertainty. The
blue triangles denote the beacons that aid in localization. The solid black
circles with radius 0.5m, represent obstacles in the environment and the red
ellipses denote the 3$\sigma$ covariances (only the ($x$,$y$) portion is
shown). Unless otherwise mentioned, in all the experiments, $0.99-$safe
configurations are solicited and the total planning time is the average time
for 25 different runs.
We first consider a case with a point robot and no uncertainty in obstacle
location. The planned trajectory in this case is seen in cyan in Fig. 5(a)
with total planning time of $0.0051s(\pm 0.0008s)$. Please note that the total
planning time also includes the collision probability computation time. Next,
we consider uncertainty in one of the obstacle location, whose covariance
ellipse is shown in gray. The planned trajectory is seen in cyan in Fig. 5(b)
and the planning was completed under $0.0279s(\pm 0.0043s)$. Due to the
uncertainty in the obstacle location, the robot takes a longer route to avoid
collision. A robot of radius 0.3$m$ and certain (negligible uncertainty)
obstacles gave the same trajectory as in Fig. 5(a) with a planning time of
$0.0055s(\pm 0.0009s)$. However, when the obstacle location is uncertain the
resulting trajectory is as shown in Fig. 5(c). A change in the trajectory is
observed, as compared to the case of a point robot in Fig. 5(b). The planning
time in this case is $0.0294s(\pm 0.0047s)$. It is also worth mentioning that
in Fig. 5(b) and (c), the roadmap was updated by adding a node since a
$0.99-$safe configuration could not be found. The added node is seen in brown,
with its coordinates being approximately $(9,11)$. We also run the case with
no obstacle uncertainty and a robot of radius 0.3$m$ using the approach of
Park et al. park2018IEEE . In this case the planned trajectory is as given in
Fig. 5(d). Note that using our approach, the same scenario gives a shorter
trajectory (Fig. 5(a)). The longer trajectory computed using the approach in
park2018IEEE is due to the fact that a loose upper bound is computed for the
collision probability. As a result a longer trajectory is obtained. Contrary
to this, we compute the exact collision probability and hence a shorter
trajectory is synthesized. The same scenario is also run with the approach in
dutoit2011IEEE and produced a trajectory similar to ours. However, since the
uncertainties are significantly lower, the approximate collision probability
values computed using dutoit2011IEEE are much smaller than the actual values.
Next, we consider the case with uncertainty in the location of the beacons.
The considered robot radius is 0.3$m$ with the bottom obstacle being uncertain
with covariance $diag(0.49,0.49)$. Taking object uncertainty into account, the
planned trajectory with covariance evolution is as shown in Fig .5(e). Fig.
5(f), shows the trajectory planned with true beacon locations. The beacons are
shown in yellow to denote the true location. Considering only the mean
position of the beacons and neglecting the position uncertainty, the planned
trajectory is as shown in Fig. 5(a). Actual execution of this would lead to
collision with the bottom obstacle. However, executing the planned trajectory
obtained by considering the uncertainty in beacon locations does not violate
the $\epsilon-$safety criterion and all the configurations are $0.99-$safe.
It is noteworthy that though we have discussed a 2D environment, the approach
directly extends to a mobile robot navigating in a 3D environment. In such
domains, the mobile robot may be represented by a minimum volume enclosing
sphere. Similarly, the obstacles can also be approximated by their
corresponding minimum volume enclosing spheres. Hence the collision condition
is the same as given in (21) and therefore the approach discussed in this
paper remains valid.
### 5.3 Laser-grasp Domain
(a)
(b)
Figure 6: Trajectory of the end-effector; green dots denote its mean and the
red ellipses denote the covariance matrix. The puck is shown in black and the
end-effector is shown to its right. (a) Trajectory and covariance evolution
when object uncertainty is not considered and (b) when object uncertainty is
considered.
(a)
(b)
Figure 7: Green dots denote the mean of the state trajectory and the red
ellipses denote the covariance matrix. Mean position of the obstacle at each
time instant is visualized in blue. (a) State trajectory and covariance
evolution during offline collision avoidance planning. (b) More information is
acquired during online planning, reducing the uncertainty of the obstacle and
thereby leading to a change in the planned trajectory.
We consider two modified versions of the laser-grasp domain as suggested in
platt2010RSS . In this domain, a planar robot manipulator must locate and
proceed towards a round puck. The state space is the position of the
manipulator’s end-effector relative to a grasping point defined directly in
front of the puck. Though the end-effector position is assumed to be known
completely, the state is not directly observed since the puck position is
unknown. Its position can be determined using the laser range finder that
points out as a horizontal line from the end-effector. The underlying system
dynamics is
$f(\textbf{x}_{t},\textbf{u}_{t})=\textbf{x}_{t}\ +\ \textbf{u}_{t}$ (45)
where $\textbf{x}\in\mathbb{R}^{2}$ denotes the state space and
$\textbf{u}\in\mathbb{R}^{2}$ is the end-effector velocity. The cost function
is of the form in (42) with $M_{u}=diag(10,10)$, $M_{g}=diag(100,100)$,
$M_{\Sigma}=diag(10000,10000)$ and $M_{C}=10$.
First, we consider a scenario wherein an additional object is placed that aids
in localization. In this scenario, the state is the end-effector position
which is not known precisely due to actuation errors. The goal is to place the
end-effector directly in front of the puck so as to be able to grasp it. Both
the object and the puck can be detected by the horizontal laser. However, the
object location is not known exactly and the $3\sigma$ uncertainty ellipse is
shown in light blue in Fig. 6(a) and (b). The mean position is visualized by
the blue blob and the yellow blob denotes the actual object location. The red
ellipses represent the state covariance at different points along the
trajectory. Fig. 6(a) shows the case in which object uncertainty is not
considered and the object is assumed to be its mean position. The manipulator
moves towards the object first, localizing the end-effector position and then
proceeds further to place the end-effector at the grasping point. However, as
seen in Fig. 6(a), while executing this plan produced offline, not considering
the object uncertainty leads to the collision of the end-effector with the
true object (in yellow). When the object uncertainty is considered, the
execution of the plan do not lead to collision, as it can be seen in Fig.
6(b). This illustrates the fact that not considering object uncertainty can
wrongly localize the robot, leading to catastrophes.
Next, we consider a scenario wherein the state space is the position of the
manipulator’s end-effector relative to a grasping point defined directly in
front of the puck. The state is not directly observed since the puck position
is unknown. However, as soon as the manipulator starts to move, a ball starts
to roll in between the manipulator and the puck. The ball follows a Gaussian
velocity distribution, and therefore at each time instant, the mean position
of the ball and the corresponding uncertainty can be estimated. The mean
position of the ball at each time instant is shown in blue in Fig. 7(a) and
(b). The green dots denote the mean of the state trajectory. As seen in Fig.
7(a), the manipulator initially moves downwards. However, as the ball comes
closer, the manipulator retraces its path and move upwards towards its
starting position to avoid collision. This is so because the safety constraint
for $\epsilon=0.99$ is violated. As the ball keeps moving upwards, after a
while, it is seen that the manipulator takes a downward action just before
reaching its starting position since the configuration is a $0.99-$safe
configuration.
The scenario in Fig. 7(b) is similar to that of Fig. 7(a). However, it is seen
that once the manipulator retraces its path backward towards the starting
position, it takes a downward action much earlier. This is because more
information is acquired during online planning and the uncertainty bound on
the obstacle changes with time.
The 2D manipulator domain studied here directly extends to 3D manipulator
scenarios for both static and mobile manipulators. In the case of static
manipulators, the end-effector is approximated as a sphere. Each link is
approximated as a set of spheres kept side by side. However, in heavily
cluttered environments such an approximation can be computationally intensive
since each sphere has to be checked for collision with obstacles. An
alternative and effective approach is to consider the minimum-volume enclosing
ellipsoid for each link rimon1997JINT . It is known that for every convex
polyhedron, there exits a unique ellipsoid of minimal volume that contains the
polyhedron and is called the Löwner-John ellipsoid of the polyhedron
grotschel1988geometric . Thus each link can be represented by their
corresponding Löwner-John ellipsoids. The distance between two ellipsoids is
used to modify the collision condition in (21). For mobile manipulators, the
collision condition should also checked for the base as discussed in the 2D
robot section.
## 6 Conclusion
In this paper, we have addressed a novel approach to compute the probability
of collision under robot and obstacle pose uncertainties. The collision
probability is computed as an infinite series whose convergence is proved. An
upper bound for the truncation error is also derived. As shown in Fig. 2,
convergence analysis is performed for different configurations and it is seen
that our approach is of the order of milliseconds and therefore can be used in
online planning. We also provide a comparison with the approaches in
park2018IEEE ; dutoit2011IEEE . In addition, we incorporate landmark
uncertainties in belief space planning and derive the resulting Bayes filter
in terms of the prediction and measurement updates of the EKF. Finally,
experimental evaluation for a mobile robot scenario and a 2D manipulator is
performed to illustrate our approach. We have considered static obstacles in
this paper and the immediate future work is to realize the approach in
simulated and real-world environments with dynamic obstacles.
## References
* [1] Ali-Akbar Agha-Mohammadi, Suman Chakravorty, and Nancy M Amato. FIRM: Sampling-based feedback motion-planning under motion uncertainty and imperfect measurements. The International Journal of Robotics Research, 33(2):268–304, 2014\.
* [2] Brian Axelrod, Leslie Pack Kaelbling, and Tomás Lozano-Pérez. Provably safe robot navigation with obstacle uncertainty. The International Journal of Robotics Research, 37(13-14):1760–1774, 2018.
* [3] Adam Bry and Nicholas Roy. Rapidly-exploring random belief trees for motion planning under uncertainty. In IEEE International Conference on Robotics and Automation, pages 723–730, 2011.
* [4] Xu Chu Ding, Alessandro Pinto, and Amit Surana. Strategic planning under uncertainties via constrained markov decision processes. In IEEE International Conference on Robotics and Automation, pages 4568–4575, 2013.
* [5] Noel E Du Toit and Joel W Burdick. Probabilistic collision checking with chance constraints. IEEE Transactions on Robotics, 27(4):809–815, 2011.
* [6] Martin Grötschel, László Lovász, and Alexander Schrijver. Geometric Algorithms and Combinatorial Optimization. Springer-Verlag, New York, 1988.
* [7] Astghik Hakobyan, Gyeong Chan Kim, and Insoon Yang. Risk-Aware Motion Planning and Control Using CVaR-Constrained Optimization. IEEE Robotics and Automation Letters, 4(4):3924–3931, 2019.
* [8] Lucas Janson, Edward Schmerling, and Marco Pavone. Monte Carlo motion planning for robot trajectory optimization under uncertainty. In Robotics Research, pages 343–361. Springer, 2018.
* [9] Ashkan M Jasour and Brian C Williams. Risk contours map for risk bounded motion planning under perception uncertainties. Robotics: Science and Systems, 2019.
* [10] Leslie Pack Kaelbling, Michael L Littman, and Anthony R Cassandra. Planning and acting in partially observable stochastic domains. Artificial Intelligence, 101(1-2):99–134, 1998.
* [11] Leslie Pack Kaelbling and Tomás Lozano-Pérez. Integrated task and motion planning in belief space. The International Journal of Robotics Research, 32(9-10):1194–1227, 2013.
* [12] Sertac Karaman and Emilio Frazzoli. Sampling-based algorithms for optimal motion planning. The International Journal of Robotics Research, 30(7):846–894, 2011\.
* [13] Lydia E Kavraki, Petr Svestka, J-C Latombe, and Mark H Overmars. Probabilistic roadmaps for path planning in high-dimensional configuration spaces. IEEE Transactions on Robotics and Automation, 12(4):566–580, 1996\.
* [14] Samuel Kotz, Norman L Johnson, and DW Boyd. Series representations of distributions of quadratic forms in normal variables. I. Central case. The Annals of Mathematical Statistics, 38(3):823–837, 1967.
* [15] Samuel Kotz, Norman L Johnson, and DW Boyd. Series representations of distributions of quadratic forms in normal variables ii. non-central case. The Annals of Mathematical Statistics, 38(3):838–848, 1967.
* [16] Hanna Kurniawati and Vinay Yadav. An Online POMDP Solver for Uncertainty Planning in Dynamic Environment. In Robotics Research: The 16th International Symposium ISRR, pages 611–629, 2016.
* [17] Alain Lambert, Dominique Gruyer, and Guillaume Saint Pierre. A fast Monte Carlo algorithm for collision probability estimation. In 10th IEEE International Conference on Control, Automation, Robotics and Vision, pages 406–411, 2008.
* [18] Alex Lee, Yan Duan, Sachin Patil, John Schulman, Zoe McCarthy, Jur Van Den Berg, Ken Goldberg, and Pieter Abbeel. Sigma hulls for gaussian belief space planning for imprecise articulated robots amid obstacles. In IEEE/RSJ International Conference on Intelligent Robots and Systems, pages 5660–5667, 2013.
* [19] Chonhyon Park, Jae S Park, and Dinesh Manocha. Fast and bounded probabilistic collision detection for high-DOF trajectory planning in dynamic environments. IEEE Transactions on Automation Science and Engineering, 15(3):980–991, 2018.
* [20] Shashank Pathak, Antony Thomas, and Vadim Indelman. A unified framework for data association aware robust belief space planning and perception. The International Journal of Robotics Research, 37(2-3):287–315, 2018.
* [21] Sachin Patil, Jur Van Den Berg, and Ron Alterovitz. Estimating probability of collision for safe motion planning under Gaussian motion and sensing uncertainty. In IEEE International Conference on Robotics and Automation, pages 3238–3244, 2012.
* [22] Robert Platt Jr, Russ Tedrake, Leslie Kaelbling, and Tomas Lozano-Perez. Belief space planning assuming maximum likelihood observations. In Proceedings of Robotics: Science and Systems, Zaragoza, Spain, 2010.
* [23] Samuel Prentice and Nicholas Roy. The belief roadmap: Efficient planning in belief space by factoring the covariance. The International Journal of Robotics Research, 28(11-12):1448–1465, 2009.
* [24] Serge B Provost and AM Mathai. Quadratic forms in random variables: theory and applications. M. Dekker, 1992.
* [25] Elon Rimon and Stephen P Boyd. Obstacle collision detection using best ellipsoid fit. Journal of Intelligent and Robotic Systems, 18(2):105–126, 1997\.
* [26] Dorsa Sadigh and Ashish Kapoor. Safe control under uncertainty with probabilistic signal temporal logic. Robotics: Science and Systems, 2016.
* [27] Oren Salzman, Brian Hou, and Siddhartha Srinivasa. Efficient motion planning for problems lacking optimal substructure. In Twenty-Seventh International Conference on Automated Planning and Scheduling, 2017.
* [28] Luke Shimanuki and Brian Axelrod. Hardness of 3D Motion Planning Under Obstacle Uncertainty. Workshop on Algorithmic Foundations of Robotics, 2018.
* [29] Antony Thomas, Fulvio Mastrogiovanni, and Marco Baglietto. Task-Motion Planning for Navigation in Belief Space. In The International Symposium on Robotics Research, 2019.
* [30] Antony Thomas, Fulvio Mastrogiovanni, and Marco Baglietto. Towards Multi-Robot Task-Motion Planning for Navigation in Belief Space. In European Starting AI Researchers’ Symposium. CEUR, 2020.
* [31] Sebastian Thrun, Wolfram Burgard, and Dieter Fox. Probabilistic Robotics. MIT press, 2005.
* [32] Jur Van Den Berg, Sachin Patil, and Ron Alterovitz. Motion planning under uncertainty using iterative local optimization in belief space. The International Journal of Robotics Research, 31(11):1263–1278, 2012.
* [33] Hai Zhu and Javier Alonso-Mora. Chance-constrained collision avoidance for mavs in dynamic environments. IEEE Robotics and Automation Letters, 4(2):776–783, 2019.
|
# Closed-form Quadrangulation of $n$-Sided Patches
Marco Tarini Università degli Studi di Milano<EMAIL_ADDRESS>
###### Abstract
We analyze the problem of quadrangulating a $n$-sided patch, each side at its
boundary subdivided into a given number of edges, using a single irregular
vertex (or none, when $n=4$) that breaks the otherwise fully regular lattice.
We derive, in an analytical closed-form, (1) the necessary and sufficient
conditions that a patch must meet to admit this quadrangulation, and (2) a
full description of the resulting tessellation(s).
## 1 Introduction
Consider a polygonal-shaped, planar region patch $P$, delimited by $n>1$
sides, each subdivided in a number of edges. Let $e_{i}\in\mathbb{N}$ be the
number of edges found on side $i$, $i\in[0..n-1]$.
We are interested in determining whether or not $P$ can be quad-tessellated
using only one irregular vertex, of valency $n$, somewhere in the interior
(even this vertex is regular when $n=4$).
This tessellation, when it exists, can also be described as the one obtained
by applying one step of Catmull-Clark (CC) subdivision [3] to the polygon $P$,
which creates $n$ quadrilateral regions, followed by a conforming, fully
regular tessellation of each of these regions, each at some appropriate grid
resolution. For this reason, we refer to this property of $P$ as it being “CC-
able”.
For example, the construction in Fig. 1 shows that a pentagonal-shaped patch
delimited by $e_{i}=(6,4,3,5,4)$ edges happens to be CC-able. We show a
closed-form formulation to determine the existence (and the uniqueness) of
this quadrangulation, for an arbitrary $P$, and to construct it, when it
exists.
Figure 1: An example: a pentagonal patch with $e_{0..4}=(6,4,3,5,4)$ is CC-
able, as graphically shown here.
### 1.1 Motivations and context
An intensively studied recent problem in Geometry Processing is Coarse layout
decomposition (see [2] for a survey), where a given surface must be
partitioned into a layout of patches homologous to a disk. This is a
precondition for numerous tasks (such as surface parametrization, shape
segmentation, shape matching, application of machine learning on 3D shapes),
each coming with its own set of useful final applications (ranging from
texture mapping, digital fabrication, shape recognition, shape modelling,
etc). Clearly, each such scenario dictates different requirements on the
layout. Among others, one central application is surface semi-regular quad-
remeshing (see [1] for a survey, and specifically Section 3.2 there). For
other potential contexts, see the Conclusions.
The layout is typically required to be comprised of rectangular regions, as
they admit a “natural” tessellation consisting of a fully-regular lattice –
but only if opposite sides of the region are subdivided in an equal number of
edges.
The present work can be understood as a way to _extend the concept of a
“natural tessellation” to non-rectangular patches_.
Figure 2: Schema for parameters $e_{i}$ and unknowns $s_{i}$ for patches with
$n$ = 2, 3, 4, 5 and 6 sides. The same schema generalises to larger $n$.
## 2 Related work
The literature addresses a problem similar to ours, but where multiple
internal irregular vertices are allowed. In this case, solutions are now known
to always exists [10] (as long as the “parity” condition holds, see below);
note that layouts presenting more internal irregular vertices are typically
considered less desirable in many contexts.
This version of the problem was solved in [10] for $n\leqslant 6$, and an
algorithm is offered that always constructs one valid solution. The problem
was previously tackled for $n=3$ or $4$, with similar results [9]. These
algorithms are combinatorial in nature; this is in contrast with our closed-
form characterization of the instances admitting a solution with a single-
irregular vertex (in addition to solving also for $n>6$).
The algorithm in [10] returns the one solution considered the “best” among
valid alternatives, according to some definition of desirability (although not
all possibilities are always considered). In a similar spirit, Machine-
Learning [6] and Data-Driven [5] heuristics have been leveraged to pick one
“best” solution among the ones admitted by a given instance (again, according
to some targeted criteria). In older proposals, users, such as digital
modellers, are allowed to interactively navigate inside the space of
admissible solutions, in search of the “preferred” one for a given mesh [8].
In most cases (although not necessarily all), the “best” or “preferred”
solution is exactly the one featuring a single irregular vertex inside the
patch, when such a solution exists; to our knowledge, our is the first
criterion to determine _a priori_ whether that is the case or not.
## 3 Solution
One well-known precondition for $P$ to be tasselable with only quads is that
$\left(\sum_{i}{e_{i}}\right)\text{mod}\,2=0.$ (1)
We refer to this as the _parity condition_ , and we assume it always holds.
We are looking for a tessellation with a single, internal irregular vertex. In
the only possible construction for a solution is depicted in Figure 2, where
each side around $P$ is split into two sub-sides. Due to the constraint on
regularity, a pair of sub-sides at the left and the right of another edge $i$
must share the same number of edges $s_{i}$.
Thus, we have that
$\forall i<n,\;\;e_{i}=s_{i-1}+s_{i+1}$ (2)
(all indices, in the above and in all following equations, are considered
modulo $n$).
Because each side of the polygon must be split in two sub-sides, we also need
to assume $e_{i}>1$.
For $P$ to be CC-able, equations 2 must be fulfilled for some choice of
unknown positive integer values $s_{i}$. This general property translates into
different sets of conditions for each value of $n$, as we analyze in the
following sections.
### 3.1 Two-sided shapes ($n=2$)
For $n=2$, the problem is trivial. Equation (2) becomes simply
$e_{0}=2\;s_{1}\;\;\;\;\text{ , }\;\;\;\;e_{1}=2\;s_{0}$ (3)
(as $i-1$ and $i+1$ denote the same index modulo 2).
Note that it is not necessarily the case that valency-2 vertices (sometimes
called _doublets_ , [4]) are considered valid configurations, but they can be
[11]. See also Fig. 3.
Figure 3: A CC-able 2-sided patch, shown with its the corresponding internal
quadrangulation.
In conclusion, _a two-sided patch is CC-able (in only one way) iff both
$e_{0}$ and $e_{1}$ are even numbers_ (a stronger condition than the Parity
condition).
### 3.2 Triangular shapes ($n=3$)
For $n=3$, we rewrite Equation (2) in matrix form:
$\left(\begin{matrix}0&1&1\\\ 1&0&1\\\
1&1&0\end{matrix}\right)\left(\begin{matrix}s_{0}\\\ s_{1}\\\
s_{2}\end{matrix}\right)=\left(\begin{matrix}e_{0}\\\ e_{1}\\\
e_{2}\end{matrix}\right)$ (4)
which implies (by matrix inversion):
$\frac{1}{2}\left(\begin{matrix}-1&+1&+1\\\ +1&-1&+1\\\
+1&+1&-1\end{matrix}\right)\left(\begin{matrix}e_{0}\\\ e_{1}\\\
e_{2}\end{matrix}\right)=\left(\begin{matrix}s_{0}\\\ s_{1}\\\
s_{2}\end{matrix}\right)$ (5)
In other terms,
$\forall i,\;\;s_{i}=\frac{1}{2}\Big{(}e_{i-1}-e_{i}+e_{i+1}\Big{)}.$ (6)
For $s_{i}$ to be integer, the expression in parenthesis must be even, which
is already guaranteed by Eq. (1).
For $s_{i}$ to be positive, we also need the above expression to be positive,
therefore:
$\forall i,\;\;e_{i-1}+e_{i+1}>e_{i}.$ (7)
In conclusion, _a triangular patch is CC-able iff each side has fewer edges
than the other two sides combined._ This condition is a discrete version of
the familiar triangle inequality.
### 3.3 Quadrilateral shapes ($n=4$)
For a quadrilateral shape, to be CC-able amounts to be regularly griddable.
The conditions for this to be the case are well-known and obvious, and are re-
derived here (consistently with the other cases) only for completeness.
With $n=4$, Eq. (2) can be written as
$\left(\begin{matrix}0&1&0&1\\\ 1&0&1&0\\\ 0&1&0&1\\\
1&0&1&0\end{matrix}\right)\left(\begin{matrix}s_{0}\\\ s_{1}\\\ s_{2}\\\
s_{3}\end{matrix}\right)=\left(\begin{matrix}e_{0}\\\ e_{1}\\\ e_{2}\\\
e_{3}\end{matrix}\right)$ (8)
The matrix is non-singular, as the first two rows match the second two;
therefore, the system has either multiple solutions for $s_{i}$, when
$e_{0}=e_{2}$ and $e_{1}=e_{3}$, or no solution otherwise. This condition also
implies the parity condition, as $\sum e_{i}=2(e_{0}+e_{1})$.
Because $n=4$, the internal “irregular” vertex is, actually, regular like the
others, and every valid choice of $s_{i}$ produces the same tessellation.
In conclusion, _a quadrilateral patch is CC-able (in only one way) when the
opposite sides have a matching number of edges._
### 3.4 Pentagonal shapes ($n=5$)
For $n=5$, we rewrite Eq. (2) as
$\left(\begin{matrix}0&1&0&0&1\\\ 1&0&1&0&0\\\ 0&1&0&1&0\\\ 0&0&1&0&1\\\
1&0&0&1&0\end{matrix}\right)\left(\begin{matrix}s_{0}\\\ s_{1}\\\ s_{2}\\\
s_{3}\\\ s_{4}\end{matrix}\right)=\left(\begin{matrix}e_{0}\\\ e_{1}\\\
e_{2}\\\ e_{3}\\\ e_{4}\end{matrix}\right)$ (9)
which implies (by matrix inversion)
$\frac{1}{2}\left(\begin{matrix}+1&+1&-1&-1&+1\\\ +1&+1&+1&-1&-1\\\
-1&+1&+1&+1&-1\\\ -1&-1&+1&+1&+1\\\
+1&-1&-1&+1&+1\end{matrix}\right)\left(\begin{matrix}e_{0}\\\ e_{1}\\\
e_{2}\\\ e_{3}\\\ e_{4}\end{matrix}\right)=\left(\begin{matrix}s_{0}\\\
s_{1}\\\ s_{2}\\\ s_{3}\\\ s_{4}\end{matrix}\right)$ (10)
In other terms,
$\forall
i,\;\;s_{i}=\frac{1}{2}\Big{(}\left(e_{i-1}+e_{i}+e_{i+1}\right)-\left(e_{i-2}+e_{i+2}\right)\Big{)}.$
(11)
For $s_{i}$ to be integer, the expressions above must be even before halving,
which is already guaranteed by Eq. (1).
The sign constraint ($s_{i}>0$) produces
$\forall i,\;\;e_{i-1}+e_{i}+e_{i+1}>e_{i-2}+e_{i+2}$ (12)
In conclusion, _a pentagonal patch is CC-able (in only one way) iff the total
number of edges in any three consecutive edges is larger than the number of
edges in the other two._
### 3.5 Hexagonal shapes ($n=6$)
For $n=6$, we rewrite Eq. (2) as
$\left(\begin{matrix}0&1&0&0&0&1\\\ 1&0&1&0&0&0\\\ 0&1&0&1&0&0\\\
0&0&1&0&1&0\\\ 0&0&0&1&0&1\\\
1&0&0&0&1&0\end{matrix}\right)\left(\begin{matrix}s_{0}\\\ s_{1}\\\ s_{2}\\\
s_{3}\\\ s_{4}\\\ s_{5}\end{matrix}\right)=\left(\begin{matrix}e_{0}\\\
e_{1}\\\ e_{2}\\\ e_{3}\\\ e_{4}\\\ e_{5}\end{matrix}\right)$ (13)
which implies (by matrix inversion)
$\frac{1}{2}\left(\begin{matrix}[r]0&+1&0&-1&0&+1\\\ +1&0&+1&0&-1&0\\\
0&+1&0&+1&0&-1\\\ -1&0&+1&0&+1&0\\\ 0&-1&0&+1&0&+1\\\ +1&0&-1&0&+1&0\\\
\end{matrix}\right)\left(\begin{matrix}e_{0}\\\ e_{1}\\\ e_{2}\\\ e_{3}\\\
e_{4}\\\ e_{5}\end{matrix}\right)=\left(\begin{matrix}s_{0}\\\ s_{1}\\\
s_{2}\\\ s_{3}\\\ s_{4}\\\ s_{5}\end{matrix}\right)$ (14)
In other terms,
$\forall i,\;\;s_{i}=\frac{1}{2}\Big{(}e_{i-1}+e_{i+1}-e_{i+3}\Big{)}.$ (15)
The integrity constraint ($s_{i}\in\mathbb{Z}$) translates in the requirement
for the sum of $e_{0,2,4}$, and the sum of $e_{1,3,5}$, to be even numbers.
The sign constraint ($s_{i}>0$) gives, by $j=(i+3)$ modulo 6:
$\forall j,\;\;e_{j-2}+e_{j+2}>e_{j}.$ (16)
In conclusion, _a hexagonal patch is CC-able (in only one way) iff (1) each
even side has fewer edges than the other two even sides combined (and
likewise, for odd sides), and (2), both the even sides, and the odd sides,
have an even total number of edges._
### 3.6 Heptagonal shapes ($n=7$)
For $n=7$, we rewrite Eq. (2) as
$\left(\begin{matrix}0&1&0&0&0&0&1\\\ 1&0&1&0&0&0&0\\\ 0&1&0&1&0&0&0\\\
0&0&1&0&1&0&0\\\ 0&0&0&1&0&1&0\\\ 0&0&0&0&1&0&1\\\
1&0&0&0&0&1&0\end{matrix}\right)\left(\begin{matrix}s_{0}\\\ s_{1}\\\ s_{2}\\\
s_{3}\\\ s_{4}\\\ s_{5}\\\
s_{6}\end{matrix}\right)=\left(\begin{matrix}e_{0}\\\ e_{1}\\\ e_{2}\\\
e_{3}\\\ e_{4}\\\ e_{5}\\\ e_{6}\end{matrix}\right)$ (17)
which implies (by matrix inversion)
$\frac{1}{2}\left(\begin{matrix}-1\\!&\\!+1\\!&\\!+1\\!&\\!-1\\!&\\!-1\\!&\\!+1\\!&\\!+1\\\
+1\\!&\\!-1\\!&\\!+1\\!&\\!+1\\!&\\!-1\\!&\\!-1\\!&\\!+1\\\
+1\\!&\\!+1\\!&\\!-1\\!&\\!+1\\!&\\!+1\\!&\\!-1\\!&\\!-1\\\
-1\\!&\\!+1\\!&\\!+1\\!&\\!-1\\!&\\!+1\\!&\\!+1\\!&\\!-1\\\
-1\\!&\\!-1\\!&\\!+1\\!&\\!+1\\!&\\!-1\\!&\\!+1\\!&\\!+1\\\
+1\\!&\\!-1\\!&\\!-1\\!&\\!+1\\!&\\!+1\\!&\\!-1\\!&\\!+1\\\
+1\\!&\\!+1\\!&\\!-1\\!&\\!-1\\!&\\!+1\\!&\\!+1\\!&\\!-1\end{matrix}\right)\left(\begin{matrix}e_{0}\\\
e_{1}\\\ e_{2}\\\ e_{3}\\\ e_{4}\\\ e_{5}\\\
e_{6}\end{matrix}\right)=\left(\begin{matrix}s_{0}\\\ s_{1}\\\ s_{2}\\\
s_{3}\\\ s_{4}\\\ s_{5}\\\ s_{6}\end{matrix}\right)$ (18)
In other terms,
$\displaystyle\begin{split}\forall
i,\;\;s_{i}=\frac{1}{2}\Big{(}e_{i-2}+e_{i-1}+e_{i+1}+e_{i+2}&\\\
-\left(e_{i}+e_{i+3}+e_{i-3}\right)\;\;\;\;&\Big{)}.\end{split}$ (19)
For $s_{i}$ to be integer, the expression in parenthesis must be even, which
is already guaranteed by Eq. (1).
The sign constraint ($s_{i}>0$) becomes
$\forall i,\;\;e_{i-2}+e_{i-1}+e_{i+1}+e_{i+2}>e_{i}+e_{i+3}+e_{i-3}.$ (20)
In conclusion, _a heptagonal patch is CC-able (in only one way) iff each side
plus its two opposite sides have fewer edges than the remaining four sides
combined._
### 3.7 Octagonal shapes ($n=8$)
For $n=8$, we can rewrite Eq. (2) as two separate linear systems:
$\begin{split}\left(\begin{matrix}1&0&0&1\\\ 1&1&0&0\\\ 0&1&1&0\\\ 0&0&1&1\\\
\end{matrix}\right)\left(\begin{matrix}s_{1}\\\ s_{3}\\\ s_{5}\\\
s_{7}\end{matrix}\right)&=\left(\begin{matrix}e_{0}\\\ e_{2}\\\ e_{4}\\\
e_{6}\end{matrix}\right)\\\ \left(\begin{matrix}1&1&0&0\\\ 0&1&1&0\\\
0&0&1&1\\\ 1&0&0&1\\\ \end{matrix}\right)\left(\begin{matrix}s_{0}\\\ s_{2}\\\
s_{4}\\\ s_{6}\end{matrix}\right)&=\left(\begin{matrix}e_{1}\\\ e_{3}\\\
e_{5}\\\ e_{7}\end{matrix}\right)\end{split}$ (21)
Neither matrix is invertible, being deficient by one rank: specifically, their
two even rows and two odd rows sum up to the same row-vector. Therefore, the
system can have solutions only when
$\begin{split}e_{0}+e_{4}&=e_{2}+e_{6},\\\
e_{1}+e_{5}&=e_{3}+e_{7}.\end{split}$ (22)
In the following, we show that this condition is also sufficient.
The parity condition is already implied by Eq. (22), because
$\sum{e_{i}}=2(e_{0}+e_{4}+e_{1}+e_{5})$.
The only remaining condition is that $s_{i}>0$. Let $k_{0},k_{1}$ be the
choices for $s_{0},s_{1}$ (consented by the rank deficits). Using Eq. (2), we
get the values of all other $s_{i}$:
$\displaystyle s_{0}$ $\displaystyle=k_{0},$ $\displaystyle s_{1}$
$\displaystyle=k_{1},$ $\displaystyle s_{2}$ $\displaystyle=e_{1}-s_{0}$
$\displaystyle s_{3}$ $\displaystyle=e_{2}-s_{1}$ $\displaystyle=e_{1}-k_{0},$
$\displaystyle=e_{2}-k_{1},$ $\displaystyle s_{4}$ $\displaystyle=e_{3}-s_{2}$
$\displaystyle s_{5}$ $\displaystyle=e_{4}-s_{3}$ (23)
$\displaystyle=e_{3}-e_{1}+k_{0},$ $\displaystyle=e_{4}-e_{2}+k_{1},$
$\displaystyle s_{6}$ $\displaystyle=e_{5}-s_{4}$ $\displaystyle s_{7}$
$\displaystyle=e_{6}-s_{5}$ $\displaystyle=e_{5}-e_{3}+e_{1}-k_{0}$
$\displaystyle=e_{6}-e_{4}+e_{2}-k_{1}$ $\displaystyle=e_{7}-k_{0},$
$\displaystyle=e_{0}-k_{1}.$
(the bottom line uses Eq. (22)). Therefore, to have $s_{i}>0$, the choices
$k_{0}$ and $k_{1}$ must be, each, subject to two upper-bounds and two lower-
bounds:
$\displaystyle 0$ $\displaystyle<k_{0}<e_{1},$ $\displaystyle 0$
$\displaystyle<k_{1}<e_{2},\;\;\;\;\;\;$ $\displaystyle e_{1}-e_{3}$
$\displaystyle<k_{0}<e_{7},$ $\displaystyle e_{2}-e_{4}$
$\displaystyle<k_{1}<e_{0}.\;\;\;\;\;\;$
The above set of constraints is always feasible. The existence of a valid
integer solution for $k_{0}$ (and likewise, for $k_{1}$) is guaranteed because
each of the two lower bounds is smaller, by at least two units, than each of
the two upper bounds. _Proof:_ it is immediate to verify this for $0<e_{1,7}$
and $e_{1}-e_{3}<e_{1}$ (as $e_{1,3,7}>1$); finally, it also holds for
$e_{1}-e_{3}<e_{7}$ because $e_{1}<e_{7}+e_{3}=e_{1}+e_{5}$ (and $e_{5}>1$;
the last equality is Eq. 22). $\square\;$
In conclusion, _an octagonal patch is CC-able (in general, in multiple ways)
iff one pair of even opposite sides have the same combined number of edges as
the other pair of even opposite sides, and same for odd sides._
Figure 4: Examples of (strict or non-strict) CC-ability being met or not met
for triangular (left) and pentagonal (right) patches. The solitary singular
vertex is the red dot; the inequalities (respectively, equations 7,12) which
are not met, or met in a non-strict sense, are annotated in red. See Section
4.
### 3.8 Generalization to $n>8$
Our CC-ability analysis for each $n\leqslant 8$ covers (we think) the cases
used by most practical scenarios.
The case for any other $n>8$ can be constructed similarly, presenting only
more instances of the situations encountered for $n\leqslant 8$. We only
briefly surmise a generalization here.
For odd values of $n$, Eq. (2) can always be expressed with an invertible
linear system, which can be inverted to extract the values of $s_{i}$ as a
linear expression of the form
$s_{i}=\frac{1}{2}\sum_{j}\pm e_{j}>0$ (25)
(as exemplified for $n=3,5,7$). Because $\sum\pm e_{j}$ has always the same
parity as $\sum e_{j}$, the condition that $s_{i}\in\mathbb{Z}$ is already
guaranteed by the Parity condition, Eq. (1). The condition for CC-ability is
thus entirely determined by the signs of $s_{i}$, which translates into a set
of $n$ linear inequalities, each involving all $e_{i}$. For a given patch,
there is either one or no solution.
For $n=2h$ ($h\in\mathbb{N}$), Eq. (2) can be expressed as two separate linear
systems, partitioning the patch sides in two alternating subsets of $h$
elements each. When $h$ is odd (i.e., $n$ is not a multiple of 4), both
systems are invertible, and CC-ability is then determined by a total of $n$
inequalities, each involving $h$ sides, plus the condition that the total
number of edges on each subset must be even; the solution is always unique, if
it exists. When $h$ is even (i.e., $n$ is a multiple of 4) neither system is
invertible, each being deficient by one rank; the CC-ability is determined by
the equality conditions on $e_{i}$ required by either system to admit
solutions. Solutions can be multiple (but equivalent when $n=4$, as noted),
and their space is spanned by two degrees of freedom.
## 4 Non-strict CC-ability
The definition of CC-ability can be relaxed by accepting that the one
singularity is found on the boundary of the patch, rather than in its interior
(see Figure 4 for examples). We term the relaxed definition “non-strict CC-
ability”.
For a patch to be non-strictly CC-able, we need $s_{i}$ to be just non-
negative, rather than strictly positive, thus allowing for all the
inequalities in the conditions for CC-ability to be fulfilled in the non-
strict sense, in Equations (7,12,16,20,LABEL:eq:octaKi) for the cases $n=$
3,5,6,7,8, respectively. The assumption that $e_{i}>1$ can also be dropped,
allowing for $e_{i}=1$ in the input patch.
When _one_ inequality is fulfilled as equality, the irregular vertex will be
on the boundary, i.e. as a boundary vertex with edge-valency $\neq 3$. When
_two_ inequalities are fulfilled as equalities, the irregular vertex will be
found on a corner, i.e. as a corner vertex with edge-valency $\neq 2$ (as
shown in bottom-right of Figure 4). It is never possible for more than two
inequalities to be fulfilled as equalities (without infringing at least
another inequality).
Figure 5: An octagonal patch (top) fulfills the conditions in Eq. (22) and is
therefore CC-able. After Eq. (LABEL:eq:octaKi), we have two choices
$k_{0},k_{1}$, ranging in the intervals $0<k_{0}<3$ and $0<k_{1}<4$; all
resulting quadrangulations are shown.
## 5 Resulting quadrangulation
In addition to stating the conditions that must be met for a patch to be CC-
able (either strictly or non-strictly), our construction also provides a
closed-form description of the resulting quadrangulation.
The set of values $s_{i}$ can be understood as a compact and complete way to
describe the internal tessellation of the patch. Specifically, they describe
the grid-sizes of the regular rectangular areas constructed on the corners of
the original patch; or, equivalently, the number of edges of the sub-sides
splitting each side, and, thus, the topological position of the singularity
inside $P$ (see Figure 2).
The Equations (3,6,11,15,19,23) define the values of $s_{i}$ as a closed
function of $e_{i}$ for the cases of $n=2,3,5,6,7,8$ respectively. In the last
case, the equations use two arbitrary values $k_{i}$, which must be chosen
inside the intervals defined in closed-form by Eq. (LABEL:eq:octaKi), spanning
all and only the possible solutions.
## 6 Discussion
In this work, we identified a problem statement for a self-contained simple
task, useful in the context of Geometry Processing, and derived a complete
answer. Specifically, we expressed in closed form the conditions for CC-
ability, the description of the resulting quadrangulation, and the set of
available choices (if multiple solutions are possible)
#### About the uniqueness of the solution
As we have shown, the sought single-irregular-vertex quadrangulation can be
non-unique for $n=8$. This contradicts the commonly held notion that, in a
quadrangulation, no irregular vertex can be “moved alone” (e.g. [8]), meaning
that its topological position inside a region cannot be modified without also
affecting either the tessellation of the region boundary or the topological
position of another irregular vertex in the same region. As it turns out,
valency-8 vertices are exceptional in that they _can_ be “moved alone”, as
exemplified in Figure 5. As we have shown, this is the lowest valency for
which this happens (but it also occurs for valencies 12, 16, 20, and so on).
#### Usability
The property of being CC-able can be easily embedded in optimization systems,
because it is expressed in closed form as a set of linear inequality,
equality, or parity constraints, which can be enforced for example in convex
numerical solvers.
#### Potential applications
While the main motivation for our analysis stems from Coarse-layout
construction (see discussion in Section 1.1), another potential context is
that of shape modelling, where digital modellers often define 3D shape high-
level description, leaving an automatic system in the background to deal with
the minutiae of the surface tessellation (for example [10]), such as in the
software suits 3D Coat (PilgWay), Z-Brush (Pixologic) and others. Due to the
mentioned direct relationship with Catmull-Clark subdivision, our formulation
can be a useful tool in the context of “reverse subdivision” (for example,
[7]), where the objective is to seek a subdivision surface approximating a
given shape.
More broadly speaking, the generality of the analyzed problem statement leads
us to believe that our closed-form formulation can find numerous applications.
## References
* [1] Bommes, D., Lévy, B., Pietroni, N., Puppo, E., Silva, C. T., Tarini, M., and Zorin, D. Quad-mesh generation and processing: A survey. Comput. Graph. Forum 32, 6 (2013), 51–76.
* [2] Campen, M. Partitioning surfaces into quadrilateral patches: A survey. Comput. Graph. Forum 36, 8 (2017), 567–588.
* [3] Catmull, E., and Clark, J. Recursively generated b-spline surfaces on arbitrary topological meshes. Computer-Aided Design 10, 6 (1978), 350 – 355.
* [4] Daniels, J., Silva, C. T., Shepherd, J., and Cohen, E. Quadrilateral mesh simplification. ACM Trans. Graph. 27, 5 (Dec. 2008).
* [5] Marcias, G., Takayama, K., Pietroni, N., Panozzo, D., Sorkine-Hornung, O., Puppo, E., and Cignoni, P. Data-driven interactive quadrangulation. ACM Trans. Graph. 34, 4 (2015), 65:1–65:10.
* [6] Matveev, A., Artemov, A., Rakhimov, R., Bobrovskikh, G., Panozzo, D., Zorin, D., and Burnaev, E. Def: Deep estimation of sharp geometric features in 3d shapes, 2020.
* [7] Panozzo, D., Puppo, E., Tarini, M., Pietroni, N., and Cignoni, P. Automatic construction of quad-based subdivision surfaces using fitmaps. IEEE Transactions on Visualization and Computer Graphics 17, 10 (2011), 1510–1520.
* [8] Peng, C.-H., Zhang, E., Kobayashi, Y., and Wonka, P. Connectivity editing for quadrilateral meshes. ACM Trans. Graph. 30, 6 (Dec. 2011), 1–12.
* [9] Takayama, K., Panozzo, D., Sorkine-Hornung, A., and Sorkine-Hornung, O. Robust and controllable quadrangulation of triangular and rectangular regions. Technical Report/ETH Zurich, Department of Computer Science 784 (2013).
* [10] Takayama, K., Panozzo, D., and Sorkine-Hornung, O. Pattern-based quadrangulation for _N_ -sided patches. Comput. Graph. Forum 33, 5 (2014), 177–184.
* [11] Tarini, M., Pietroni, N., Cignoni, P., Panozzo, D., and Puppo, E. Practical quad mesh simplification. Computer Graphics Forum (2010).
|
# Iterated and mixed discriminants
Alicia Dickenstein, Sandra di Rocco, Ralph Morrison Department of
Mathematics, FCEN, University of Buenos Aires and IMAS (UBA-CONICET), Ciudad
Universitaria, Pab. I, C1428EGA Buenos Aires, Argentina<EMAIL_ADDRESS>KTH, Royal Institute of Technology, 10044, Stockholm, Sweden<EMAIL_ADDRESS>Department of Mathematics, Williams College, Williamstown, MA 01267, USA
<EMAIL_ADDRESS>
###### Abstract.
We consider systems of Laurent polynomials with support on a fixed point
configuration. In the non-defective case, the closure of the locus of
coefficients giving a non-degenerate multiple root of the system is defined by
a polynomial called the mixed discriminant. We define a related polynomial
called the multivariate iterated discriminant, generalizing the classical
Schäfli method for hyperdeterminants. This iterated discriminant is easier to
compute and we prove that it is always divisible by the mixed discriminant. We
show that tangent intersections can be computed via iteration if and only if
the singular locus of a corresponding dual variety has sufficiently high
codimension. We also study when point configurations corresponding to Segre-
Veronese varieties and to the lattice points of planar smooth polygons, have
their iterated discriminant equal to their mixed discriminant.
The first author acknowledges the support of ANPCyT PICT 2016-0398, UBACYT
20020170100048BA, CONICET PIP 11220150100473 and SOURCES, Argentina. The
second author acknowledges supported by KTH and Williams College. The third
author acknowledges support by ICERM and VR grants NT:2014-4763,
NT:2018-03688. All three authors acknowledge support by the Knut and Alice
Wallenberg foundation.
## 1\. Introduction
Let $K$ be an algebraically closed field of characteristic zero and
$A\subset\mathbb{Z}^{n}$ a finite lattice subset. A (Laurent) polynomial with
support on the point configuration $A$, $p=\sum_{a\in A}c_{a}x^{a}\in
K[x_{1},\ldots,x_{n}],$ is called an _$A$ -polynomial_.
Consider a system of $(r+1)$ polynomials equations supported on
$A_{0},\ldots,A_{r}\subset\mathbb{Z}^{n}$ respectively:
(1.1) $p_{0}=p_{1}=\ldots=p_{r}=0,\quad p_{i}=\sum_{a\in A_{i}}c_{i,a}x^{a}.$
When $r+1=n$, the Bernstein-Kouchnirenko Theorem says that for a generic
choice of the coefficients $c_{i,a}$ there is a finite number of non-
degenerate solutions to (1.1) in $(K^{*})^{n},$ equal to the mixed volume
$MV(A_{0},\ldots,A_{n-1})$ of the support subsets. These intersections are
transverse except for solutions on which the Jacobian associated to the system
vanishes. When $r+1<n,$ vanishing points of the Jacobian correspond to
singularities of the variety cut out by the hypersurfaces $p_{i}=0$.
Classical work by Salmon [Sal82] and Bromwich [Bro71] classified singular
intersections of two quadric surfaces, corresponding to the case $r=1$ and $A$
equal to the lattice points in the dilated simplex $2\Delta_{3}$ in
$\mathbb{R}^{3}$. The basic idea of these results was already pursued by
Cayley in connection with tangent intersections of conics in $\mathbb{C}^{2}.$
More recently, the problem has been revisited with similar tools in [FNO89],
in the context of geometric modeling with focus on the real case; and in
[ZJT+19], where these techniques are used to classify singular Darboux
cyclides. These are surfaces in 3-space that are the projection of the
intersection of two quadrics in dimension four. A generalization to the case
of two higher dimensional quadric hypersurfaces is given in [Ott13].
Consider two space quadrics, given in matrix form by
(1.2)
$p_{i}=\begin{bmatrix}1&x_{1}&x_{2}&x_{3}\end{bmatrix}\,M_{i}\,\begin{bmatrix}1\\\
x_{1}\\\ x_{2}\\\ x_{3}\end{bmatrix},\quad i=0,1.$
For generic matrices, the intersection $(p_{1}=p_{2}=0)$ describes a non-
sigular curve of degree $4$. The non-generic intersections are described in
[Sch53, GKZ94] in the following way. Consider the pencil of quadrics given by
$p_{0}+tp_{1}$. Using the Schäfli decomposition method, the existence of a
tangential intersection can be studied by considering the zero locus of the
the following polynomial in the entries of $M_{0},M_{1}$:
(1.3) $D_{4\Delta_{1}}(\det(M_{0}+tM_{1})),$
where $D_{4,\Delta_{1}}$ is the univariate discriminant of the degree $4$
polynomial $\det(M_{0}+tM_{1})$, considered as a polynomial in $t.$ For
generic matrices this is a polynomial of degree $6$ in its entries (that is,
in the coefficients of $p_{0},p_{1}$), and it vanishes whenever
$\det(M_{0}+tM_{1})$ does not have four simple roots. To classify the
different singular intersections, they then studied the Segre characteristics
arising from the Jordan normal form of $M_{0}+tM_{1}$.
Figure 1. Transverse (left) and non-transverse (right) hyperbolas
In this paper, we propose and study a generalization of this approach for any
support $A$.
We consider Equation (1.3) to be an iterated process, as we are computing the
discriminant of a discriminant. Factorizations of iterated discriminants and
resultants for polynomials of three variables where studied in [BM09]. Our aim
is to define and study an iterated discriminant generalizing Schäfli’s method
for hyperdeterminants, and to show when tangent intersections can be computed
via iteration.
The theory of $A$-discriminants was introduced in [GKZ94] and has been
extensively studied both from a geometric and a computational viewpoint
[DFS07, DRRS07, Est10, GHRS16]. Denote by $X_{A}\subset{\mathbb{P}^{|A|-1}}$
the projective variety defined as the closed image of the monomial embedding
given by the $A$-monomials. The dual variety
$X_{A}^{\nu}\subset{\mathbb{P}^{|A|-1}}^{\nu}$ is the closure of the
coefficient vectors of the $A$-polynomials $p$ whose zero-locus $(p=0)$ has a
singular point $x\in(K^{*})^{n}$ with nonzero coordinates. Equivalently, the
dual variety is the closure of the hyperplane sections of $X_{A}$ which are
singular at a point with nonzero coordinates. The expected codimension of
$X_{A}^{\nu}$ is one and when this is the case we say that $A$ is _non-
defective_. When $A$ is non-defective the irreducible polynomial
$D_{A}\in{\mathbb{Z}}[(c_{a})_{a\in A}]$ defining (up to sign) the dual
variety: $X_{A}^{\nu}=(D_{A}=0),$ is called the _$A$ -discriminant_ [GKZ94].
We will use the notation $D_{A}((c_{a})_{a\in A})=D_{A}(p)$.
If $n=1$ and $A=\\{0,1,2\\}$, then $p=c_{2}x^{2}+c_{1}x+c_{0}$ and
$D_{A}(p)=c_{1}^{2}-4c_{0}c_{2}$ is the classical discriminant of a degree two
polynomial. More generally, $D_{0,1,\ldots,\delta}$ coincides with the
classical discriminant of univariate polynomials of degree $\delta$. This
$D_{0,1,\ldots,\delta}$ is a polynomial of degree $2(\delta-1)$ in the
coefficients, which we denote by $D_{\delta\Delta_{1}}$. The case of multi-
linear polynomials (i.e. tensors) corresponds to the case in which the convex
hull of $A$ equals the product
$\Delta_{n_{1}}\times\ldots\times\Delta_{n_{l}}$ where $\Delta_{s}$ denotes
the unit simplex of dimension $s.$ This multivariate $A$-discriminant is also
referred to as the hyperdeterminant of size
$(n_{1}+1)\times\ldots\times(n_{l}+1)$ [GKZ94, Chapter 14]. This is a
classical object defined originally by Cayley [Cay45].
Note that for a quadratic polynomial $p$ with associated matrix $M$ as in
(1.2), that is for $A$ consisting of the lattice points in $2\Delta_{3}$, the
existence of a singular point in $(p=0)$ implies that the linear forms given
by its partial derivatives vanish and so $\det(M)=0$. Indeed,
$D_{A}(p)=\det(M)$. This suggests that an iterated discriminant should be
connected to the notion of discriminant for a system of polynomials. This
notion is called the _mixed discriminant_ [GKZ94, CCD+13, DEK14], which is a
natural generalization of the classical $A$-discriminant.
Given $r+1$ finite configurations $A_{0},\ldots,A_{r}\subset\mathbb{Z}^{n}$,
we call an isolated solution $x\in(K^{*})^{n}$ a _non-degenerate multiple
root_ for the system (1.1) if the $r+1$ gradient vectors
$\nabla_{x}p_{i}(x),i=0,\ldots,r$ are linearly dependent but any subset of $r$
of them is linearly independent. The associated _mixed discriminantal variety_
is the closure of the locus of coefficients for which the system has a non-
degenerate multiple root. If this variety is a hypersurface, it is defined by
a single irreducible polynomial which we call the mixed discriminant, denoted
$MD_{A_{0},\ldots A_{r}}$. If it is not a hypersurface, we call the system
_defective_ and set $MD_{A_{0},\ldots A_{r}}=1$.
Observe that when $r=0$, $MD_{0,A_{0}}=D_{A_{0}}$ equals the $A_{0}$
discriminant. In fact, in the non-defective case, mixed discriminants are
special cases of discriminants of a single polynomial. This was settled in
[GKZ94], but without the hypothesis of non-degeneracy of the common multiple
root and in [CCD+13] for the case $r+1=n$. Given $A_{0},\ldots A_{r}$, the
associated Cayley configuration
$C=C(A_{0},\dots,A_{r})\subset\mathbb{Z}^{n+r}$ is the union of the lifted
configurations $e_{i}\times A_{i}\in\mathbb{Z}^{n+r}$ for $i=0,\ldots,r$,
where $e_{0}=0$ and $e_{i}$ is the standard $i^{th}$ basis vector in
$\mathbb{Z}^{r}$ for $i\geq 1.$ As sparse discriminants are affine invariants
of lattice configurations [GKZ94], we could equivalently consider
$C\subset\mathbb{Z}^{n+r+1}$, where now $e_{0},\dots,e_{r}$ denote the
canonical basis in $\mathbb{Z}^{r+1}$. In Proposition 3.3 we prove that:
###### Proposition.
If $C$ is non-defective, then $MD(A_{0},\ldots,A_{r})$ equals $D_{C}$.
This characterization leads to the following definition of multivariate
iterated discriminant of order $r.$ We introduce $(r+1)$ new variables
$\lambda_{0},\ldots,\lambda_{r}$ and encode the initial system by one
auxiliary $C$-polynomial:
$P_{\lambda}=\lambda_{0}p_{0}+\ldots+\lambda_{r}p_{r}\in
K[\lambda_{0},\dots,\lambda_{r},x_{1},\dots,x_{n}].$
We will denote both this polynomial and its tuple of coefficients by
$P_{\lambda}$ where $\lambda=(\lambda_{0},\ldots,\lambda_{r}).$ In the present
paper we consider the case when $A_{0}=\ldots=A_{r}=A$ and use the notation
$MD_{r,A}:=MD_{A,\ldots,A}.$ Notice that $D_{A}(P_{\lambda})$ is a homogeneous
polynomial of degree $\deg(D_{A})$ in $\lambda_{0},\ldots,\lambda_{r}.$
###### Definition 1.1.
Given $A\subset\mathbb{Z}^{n}$ non-defective, denote by $d$ the codimension of
the singular locus of the dual variety $X_{A}^{\nu}$. Given $r\geq 0$, the
_multivariate iterated discriminant of order $r$_ is the polynomial $ID_{r,A}$
on the coefficients of $(r+1)$ $A$-polynomials $p_{0},\dots,p_{r}$ defined by
$\begin{cases}ID_{r,A}(p_{0},\dots,p_{r}):=D_{\delta_{A}\Delta_{r}}(D_{A}(P_{\lambda})),\,\text{
if }d\geq r+1,\\\ ID_{r,A}(p_{0},\dots,p_{r}):=0,\,\text{ otherwise.
}\end{cases}$
It is worth noting that in the classical case of $r=0,$ all these polynomials
coincide by definition:
$MD_{0,A}=ID_{0,A}=D_{A},\;\;\text{ and }D_{\delta\Delta_{0}}(D_{A}(\lambda
p_{A}))=D_{A}.$
The latter equality is a consequence of the fact that the discriminant (in the
variable $\lambda$) of the monomial $D_{A}\lambda^{\delta}$ is the coefficient
$D_{A}$ [Jou91]. Moreover, when $A$ consists of the vertices of a simplex,
$ID_{r,A}$ coincides with the hyperdeterminant Schäfli decomposition [GKZ94,
Ch. 14].
Our main results give a precise relation between $MD_{r,A}$ and $ID_{r,A}$.
The advantage of relating $MD_{r,A}$ with $ID_{r,A}$ is that the latter
polynomial is much easier to compute. We show that in the non-defective case
$MD_{r,A}$ is always an irreducible factor of $ID_{r,A}$, as a consequence of
biduality (see Section 4). Therefore, if $ID_{r,A}(p_{0},\dots,p_{r})\neq 0$,
we get a certificate that the intersection $(p_{0}=\dots=p_{r})$ is smooth.
When $A$ is non-defective, we denote by $\textrm{sing}(X_{A}^{\nu})$ the
subscheme of $X_{A}^{\nu}$ defined by the ideal generated by the partial
derivatives of $D_{A}.$ We show that $ID_{r,A}$ can have other irreducible
factors given by the Chow forms $Ch_{Y_{k}}$ of the higher dimensional
irreducible components of the schematic singular locus of the dual variety
$X_{A}^{\nu}$. We recall the notion of Chow forms at the beginning of Section
4. Theorem 4.4 and Proposition 4.6 imply the following Theorem.
###### Theorem.
Assume $A\subset\mathbb{Z}^{n}$ is non-defective and let $r\in\mathbb{Z}$ with
$0\leq r\leq\dim(X_{A}).$ Then, the mixed discriminant $MD_{r,A}$ always
divides the iterated discriminant $ID_{r,A}$. Moreover:
1. (1)
If ${\rm codim}_{X_{A}^{\nu}}(sing(X_{A}^{\nu}))>r$, then $ID_{r,A}=MD_{r,A}$.
2. (2)
If ${\rm codim}_{X_{A}^{\nu}}(sing(X_{A}^{\nu}))=r$, then
$ID_{r,A}=MD_{r,A}\prod_{i=k}^{\ell}Ch_{Y_{k}}^{\mu_{k}},$ where
$Y_{1},\dots,Y_{\ell}$ are the irreducible components of $sing(X_{A}^{\nu})$
of maximal dimension $r$, with respective multiplicities $\mu_{k}\geq 1$.
3. (3)
If ${\rm codim}_{X_{A}^{\nu}}(sing(X_{A}^{\nu}))<r$, then $ID_{r,A}=0.$
As a corollary of our results, we show in Proposition 5.2 that an iterated
method to characterize singular complete intersections is unfortunately only
possible for $r+1$ hypersurfaces of the same degree $d$ in $\mathbb{P}^{n}$
and only if $r=1$ and $d=2$ (the case already found in [Ott13, Theorem 8.2]):
###### Proposition.
Let $1<d$ and $1\leq r\leq n$. Then $MD_{r,d\Delta_{n}}=ID_{r,d\Delta_{n}}$ if
and only if $r=1$ and $d=2$.
The paper is organized as follows. In Section 2 we present some examples that
motivate the theory of iterated discriminants. In Section 3 we present
material on mixed discriminants and Cayley configurations, and in Section 4 we
develop the theory of iterated discriminants and prove our main results.
In Section 5 we ask more broadly when mixed and iterated discriminants are
equal, for products of scaled simplices, that is, when $X_{A}$ is a Segre-
Veronese variety. The case of Segre varieties was solved in [WZ96], via a
careful study of the singularities of hyperdeterminant varieties. Our
Conjecture 5.3 is the following, notation as in Section 5:
###### Conjecture.
The equality $\deg(ID_{r,A_{\ell,d,k}})=\deg(MD_{r,A_{\ell,d,k}})$ holds if
and only if
$\mathbb{P}^{r}(1)\times\mathbb{P}^{k_{1}}(d_{1})\times\cdots\times\mathbb{P}^{k_{\ell}}(d_{\ell})$
is of one of the following cases:
1. (1)
$\mathbb{P}^{r}\times\mathbb{P}^{m}\times\mathbb{P}^{m},\,m\geq 1,\,r=1,2$,
2. (2)
$(\mathbb{P}^{1})^{4}$,
3. (3)
$\mathbb{P}^{1}\times\mathbb{P}^{n}(2)$.
A partial answer is given in Theorem 5.6 and Proposition 5.2.
Finally, in Section 6 we analyze the case of plane curves. Theorem 6.3 shows
that for planar configurations $A$ consisting of the lattice points of a
smooth polygon, the only case where $MD_{1,A}$ equals $ID_{1,A}$ are the known
cases in which the polygon is the unit square (the bilinear case) or
$2\Delta_{2}$, the standard triangle of size $2$. This implies that in all
other cases, the singularities of the discriminant locus have codimension one;
that is, there are “many” different types of singular hypersurfaces defined by
$A$-polynomials. A factorization of the iterated discriminants gives all
components of the singular locus of codimension one.
### Acknowledgements
We thank Carlos D’Andrea, Frédéric Bihan, Laurent Busé, Bernard Mourrain, and
Giorgio Ottaviani for helpful discussions and references to previous work in
this direction.
## 2\. Motivating examples
In this section we present some motivating examples that we abstract in the
paper. The first two correspond to two classical cases in which the iterated
discriminant actually computes the mixed discriminant. The last two are the
simplest cases which already show the occurrence of other factors of the
iterated discriminant.
###### Example 2.1.
Let $A=\\{(0,0),(1,0),(0,1),(1,1)\\}$ be the vertices of the unit cube and let
$f=c_{00}+c_{10}x_{1}+c_{01}x_{2}+c_{11}x_{1}x_{2}$ be an $A$-polynomial. In
this case, $D_{A}(f)=c_{00}c_{11}-c_{10}c_{01}$ is a polynomial of degree $2$,
which equals the determinant of the matrix
$\left(\begin{array}[]{cc}c_{00}&c_{01}\\\ c_{10}&c_{11}\end{array}\right).$
In case $r+1=2$, the mixed discriminant associated with two $A$-polynomials
$p_{0}=c^{1}_{00}+c^{1}_{10}x_{1}+c^{1}_{01}x_{2}+c^{1}_{11}x_{1}x_{2}$ and
$p_{1}=c^{2}_{00}+c^{2}_{10}x_{1}+c^{2}_{01}x_{2}+c^{2}_{11}x_{1}x_{2}$, is
the following degree four irreducible polynomial, which is the
hyperdeterminant of format $2\times 2\times 2$ (see [GKZ94], pp. 475–479):
$MD_{1,A}(p_{0},p_{1})={c^{2}_{00}}^{2}{c^{1}_{11}}^{2}-2c^{2}_{00}c^{2}_{01}c^{1}_{10}c^{1}_{11}-2c^{2}_{00}c^{2}_{10}c^{1}_{01}c^{1}_{11}-2c^{2}_{00}c^{2}_{11}c^{1}_{00}c^{1}_{11}+$
$4c^{2}_{00}c^{2}_{11}c^{1}_{01}c^{1}_{10}+2c^{2}_{00}c^{1}_{00}{c^{1}_{11}}^{2}-4c^{2}_{00}c^{1}_{01}c^{1}_{10}c^{1}_{11}+{c^{2}_{01}}^{2}{c^{1}_{10}}^{2}+4c^{2}_{01}c^{2}_{10}c^{1}_{00}c^{1}_{11}-$
$2c^{2}_{01}c^{2}_{10}c^{1}_{01}c^{1}_{10}-2c^{2}_{01}c^{2}_{11}c^{1}_{00}c^{1}_{10}+2c^{2}_{01}c^{1}_{00}c^{1}_{10}c_{11}+{c^{2}_{10}}^{2}{c^{1}_{01}}^{2}-$
$2c^{2}_{10}c^{2}_{11}c^{1}_{00}c^{1}_{01}+2c^{2}_{10}c^{1}_{00}c^{1}_{01}c^{1}_{11}+{c^{2}_{11}}^{2}{c^{1}_{00}}^{2}-2c^{2}_{11}{c^{1}_{00}}^{2}c^{1}_{11}+{c^{1}_{00}}^{2}{c^{1}_{11}}^{2},$
It vanishes at $(p_{0},p_{1})$ with respective coefficient vectors
$(1,1,-2,-1)$ and $(1,1,-3,-2)$, corresponding to the tangent hyperbolas in
Figure 1.
One form of computing $D_{A}$ is as the iterated discriminant $ID_{1,A}$.
Write
$\det\left(\begin{array}[]{cc}c^{1}_{00}+\lambda c^{2}_{00}&c^{1}_{01}+\lambda
c^{2}_{01}\\\ c^{1}_{10}+\lambda c^{2}_{10}&c^{1}_{11}+\lambda
c^{2}_{11}\end{array}\right)\,=\,\Delta_{0}+\Delta_{1}\lambda+\Delta_{2}\lambda^{2},$
and then compute
$MD_{1,A}(c^{1},c^{2})=\Delta_{1}^{2}-4\Delta_{0}\Delta_{2},$
as the univariate resultant of the degree $2$ polynomial
$\Delta_{0}+\Delta_{1}\lambda+\Delta_{2}\lambda^{2}$ in $\lambda$ with
coefficients in $\mathbb{Z}[c^{1},c^{2}]$. This compact formula is the
simplest case of Schäfli’s formula to compute the mixed discriminant
$MD_{1,A}$.
###### Example 2.2.
Let us consider again the case discussed in the Introduction corresponding to
the singular intersections of two quadric surfaces $p_{0},p_{1}$ in three-
space. We display their common support $A$ as the columns of the following
$3\times 10$ matrix:
$\begin{bmatrix}0&1&0&0&2&1&1&0&0&0\\\ 0&0&1&0&0&1&0&2&1&0\\\
0&0&0&1&0&0&1&0&1&2\end{bmatrix}.$
We also display the corresponding Cayley configuration $C=\Delta_{1}\times A$
as the columns of the following $5\times 20$-matrix:
$\small\setcounter{MaxMatrixCols}{20}\begin{bmatrix}0&1&0&0&2&1&1&0&0&0&0&1&0&0&2&1&1&0&0&0\\\
0&0&1&0&0&1&0&2&1&0&0&0&1&0&0&1&0&2&1&0\\\
0&0&0&1&0&0&1&0&1&2&0&0&0&1&0&0&1&0&1&2\\\
1&1&1&1&1&1&1&1&1&1&0&0&0&0&0&0&0&0&0&0\\\
0&0&0&0&0&0&0&0&0&0&1&1&1&1&1&1&1&1&1&1\end{bmatrix}.$
In this case, we know that $(X_{C})^{\nu}$ is a hypersurface by [DR14]. Thus
we have that the polynomial $MD_{1,A}(p_{0},p_{1})$ cuts out the closure of
the locus of coefficients for which the two quadrics lie tangent to one
another at a point and it can be computed via the discriminant $D_{C}$ by
Proposition 3.3. It can be also computed as the iterated discriminant in
(1.3). This polynomial can be studied through tropical discriminants as in
[DFS07].
Moreover, one can compute the univariate discriminant $D_{4\Delta_{1}}$ of a
degree $4$ polynomial as the discriminant of its cubic resolvent from Galois
theory. Let $\Delta_{i}$ denote the coefficient $\lambda^{i}$ in
$\det(M_{0}+tM_{1})$. Then
$MD_{1,A}(p_{0},p_{1})=\frac{4p^{3}-q^{2}}{27},$
where $p=12\Delta_{4}\Delta_{0}-3\Delta_{3}\Delta_{1}+\Delta_{2}^{2}$, and
$q=72\Delta_{4}\Delta_{2}\Delta_{0}+9\Delta_{3}\Delta_{2}\Delta_{1}-27\Delta_{4}\Delta_{1}^{2}-27\Delta_{0}\Delta_{3}^{2}-2\Delta_{2}^{3}$.
This gives a compact and feasible way of computing the mixed discriminant
$MD_{1,A}$ we are interested in. In fact, expanding this expression in terms
of the coefficients of $p_{0},p_{1}$ is beyond the capabilities of the
excellent Computer Algebra System Macaulay2 [GS] in a standard computer,
because it is a polynomial of degree $24$ which has degree $12$ in both the
coefficients of $p_{0}$ and $p_{1}$. Note that a general polynomial of
bidegree $(12,12)$ in two groups of $10$ variables has more than $4\cdot
10^{11}$ monomials!
The general case is hinted in the following simple examples.
###### Example 2.3.
Consider the two dimensional configuration
$A=\\{(0,0),(1,0),(2,0),(0,1),(1,1)\\}$ corresponding to the first Hirzebruch
surface ${\mathbb{F}}_{1}$.
Given a generic polynomial $f$ with support $A$:
$f(x,y)=a_{0}+a_{1}x+a_{2}x^{2}+y(b_{0}+b_{1}x),$
the $A$-discriminant coincides with the resultant of the two univariate
polynomials $a_{0}+a_{1}x+a_{2}x^{2}$ and $b_{0}+b_{1}x$ and thus is equal to
the degree $3$ polynomial
$D_{A}(f)\,=a_{0}b_{1}^{2}-a_{1}b_{0}b_{1}+a_{2}b_{0}^{2}.$
The mixed discriminant $MD_{1,A}$ has degree $8$, while the iterated
discriminant $ID_{2,A}$ has degree $2\cdot 3\cdot(3-1)^{1}=12$ by (4.9) .
There is another irreducible factor that we explain in Theorem 4.4 and compute
in Example 4.7.
###### Example 2.4.
We now consider the case of a univariate polynomial of degree $3$ with
$A=\\{0,1,2,3\\}$ and $r=1$. Given two cubic polynomials $p_{0},p_{1}$
depending on a variable $x$, their mixed discriminant equals the discriminant
of the Cayley configuration
$C=\\{(0,0),(0,1),(0,2),(0,3),(1,0),(1,1),(1,2),(1,3)\\}$
at the polynomial $p_{0}+tp_{1}$ in one more variable $t$. In fact,
$D_{C}(p_{0}+tp_{1})$ equals the univariate resultant ${\rm
Res}_{3,3}(p_{0},p_{1})$. This resultant can be computed as the determinant of
the associated Sylvester matrix and therefore has degree $6$ in the vectors of
coefficients of $p_{0},p_{1}$. Since the discriminant $D_{A}$ of a cubic
univariate polynomial has degree $4$, the iterated discriminant
$ID_{1,A}=D_{4\Delta 1}(D_{A}(p_{0}+tp_{1}))$ instead has degree $2\cdot
4\cdot 3=24$ according to (4.9). It has another irreducible factor of degree
$6$ raised to the third power, which corresponds to the Chow form of the
singular locus of $D_{A}=0$ corresponding to degree $3$ polynomials with a
triple root (a degenerate multiple root), predicted by Theorem 4.4.
## 3\. The Mixed Discriminant and the discriminant of the Cayley
configuration
In this section we show in Proposition 3.3 that in the non-defective case the
mixed discriminant $MD(A_{0},\ldots,A_{r})$ coincides with the discriminant of
the associated Cayley configuration $D_{C}$, thus generalizing Theorem 2.1 in
[CCD+13]. Note that when $C$ is defective these varieties need not coincide,
as shown in Example 2.2 in [CCD+13]. We also characterize, in Proposition 3.4,
non-defectivity of $C$ when all $A_{i}$ are equal. The latter result relies on
a classical criterion by Katz, stated as Lemma 3.1 below, and is a simple
consequence of [WZ94, Th. 0.1].
Recall that for a projective variety $X$, the _dual defect_ of $X$ is defined
to be
(3.1) $\text{def}(X):=\text{codim}(X^{\nu})-1,$
where $X^{\nu}$ is the dual variety consisting of singular hyperplane sections
to $X$. In particular, if the dual variety is a hypersurface as expected, then
the dual defect is equal to $0$ and $X$ is said to be non-defective. When
$X=X_{A}$ for some finite lattice configuration $A$, we also say that $A$ is
non-defective. In this context, we have the following lemma due to Katz.
###### Lemma 3.1.
[Kat73] Let $A\subset\mathbb{Z}^{n}$ be a lattice configuration with
$|A|=N+1.$ Let $H_{p}(f)$ denote the Hessian matrix of an $A$-polynomial $f$.
Then
${\rm codim}{X_{A}^{\nu}}=1+{\rm min}_{f}\left({{\rm
corank}(H_{u}(f))}\right)$
where $u$ is a general point and $f$ varies among the polynomials with support
in $A$ vanishing at $u.$
In particular, ${\rm codim}{X_{A}^{\nu}}=1$ implies that polynomials vanishing
at a general point $u$ together with their partial derivatives have Hessian of
maximal rank.
Observe that Lemma 3.1 is equivalent to saying that in the non-defective case,
the closure of the singular $A$-polynomials coincides with the closure of the
nodal $A$-polynomials, that is, polynomials only admitting non-degenerate
multiple roots (see Introduction).
###### Corollary 3.2.
If $A$ is a non-defective finite lattice configuration, then
$X_{A}^{\nu}=\overline{\left\\{f\in{\mathbb{P}^{N}}^{\nu}\,:\,f(u)=0,\frac{\partial
f}{\partial x_{i}}(u)=0\text{ and }\det(H(f))(u)\neq 0\text{ for some
}u\in(K^{*})^{n}\right\\}}.$
###### Proof.
The inclusion “$\subseteq$” follows by definition and the inclusion
“$\supseteq$” follows by Lemma 3.1. ∎
Let us now consider the Cayley configuration $C$ associated to $r+1$ finite
lattice configurations $A_{0},\ldots,A_{r}\in\mathbb{Z}^{n}.$ We remark that
the $r$ in [GKZ94] corresponds to our $r-1.$ We use the following notation:
$(y,x)=(y_{0},\ldots,y_{r},x_{1},\dots,x_{n})$. A polynomial $f$ with support
on $C$ has the form
$f\,=\,\sum_{0}^{r}y_{i}p_{i},$
where $p_{i}$ are $A_{i}$-polynomials in the variables $x$. Consider the
Jacobian matrix of $p_{0},\dots,p_{r}$ at $u$
$\begin{bmatrix}\nabla_{x}(p_{0})(u)&\ldots&\nabla_{x}(p_{r})(u)\end{bmatrix}.$
Notice that $f\in X_{C}^{\nu}$ if there exists
$(\lambda,u)\in(\mathbb{C}^{\nu})^{r+1}\times(\mathbb{C}^{\nu})^{n}\text{ s.t.
}p_{0}(u)=\dots=p_{r}(u)=0\text{ and
}\lambda\in\ker\begin{bmatrix}\nabla_{x}(p_{0})(u)&\ldots&\nabla_{x}(p_{r})(u)\end{bmatrix}^{T};$
or equivalently, if $\sum_{i=0}^{r}\lambda_{i}\nabla_{x}(p_{i})(u)=0$ and thus
the gradients are linearly dependent. In particular,
${\rm
rank}(\begin{bmatrix}\nabla_{x}(p_{0})(u)&\ldots&\nabla_{x}(p_{r})(u)\end{bmatrix})\leq
r.$
We will now prove that the locus where the rank is exactly $r$ characterizes
the dual variety $X_{C}^{\nu},$ assuming it is a hypersurface.
###### Proposition 3.3.
Let $A_{0},\dots,A_{r}$ and $C$ as above and assume that $C$ is non-defective.
Then,
$MD(A_{0},\ldots,A_{r})(p_{0},\dots,p_{r})=D_{C}(\sum_{i=0}^{r}y_{i}p_{i}),$
where $p_{i}$ are $A_{i}$-polynomials for $i=0,\dots,r$ and
$(y_{0},\dots,y_{r})$ are variables.
###### Proof.
Let $\phi_{f}$ be the tuple of coefficients of $f$. Corollary 3.2 implies that
$X_{C}^{\nu}=\overline{\\{\phi_{f}\;:\;f(\lambda,u)=0,p_{A_{i}}(u)=0,i=0,\dots,r,\,\sum_{0}^{r}\lambda_{j}\frac{\partial
p_{A_{j}}}{\partial x_{i}}(u)=0\text{ and }\det(H(f))(\lambda,u)\neq 0\\}},$
where $(\lambda,u)\in(\mathbb{C}^{\nu})^{r+1}\times(\mathbb{C}^{\nu})^{n}.$
Here, $H(f)(\lambda,u)$ means the following: as $f$ is homogeneous in the
variables $y$ and $\lambda\in(\mathbb{C}^{\nu})^{r+1}$, we assume that
$\lambda_{0}=1$ and that $(\lambda_{1},\dots,\lambda_{r})$ are its affine
coordinates. Thus, $H(f)(\lambda,u)$ is the Hessian of $f$ with respect to the
variables $(\lambda_{1},\dots,\lambda_{r},x_{1},\dots,x_{n})$ .
This Hessian matrix is of the form
$H_{(u,\lambda)}(\phi)=\begin{bmatrix}\lambda_{0}H(p_{A_{0}})(u)+\sum_{1}^{r}\lambda_{i}H(p_{A_{i}})(u)&\begin{bmatrix}\nabla(p_{A_{1}})(u)\\\
\ldots\\\ \nabla(p_{A_{r}})(u)\end{bmatrix}^{T}\\\
\begin{bmatrix}\nabla(p_{A_{1}})(u)\\\ \ldots\\\
\nabla(p_{A_{r}})(u)\end{bmatrix}&0\end{bmatrix}.$
It follows that if $\phi_{f}\in X_{C}^{\nu}$ and $\det H(f)(\lambda,u)\neq 0,$
which happens for generic points in $X_{C}^{\nu}$ by Corollary 3.2, then
$\textrm{rank}\begin{bmatrix}\nabla(p_{A_{1}})(u)\\\ \ldots\\\
\nabla(p_{A_{r}})(u)\end{bmatrix}=r.$ As $\lambda_{j}\neq 0$ for any
$j=0,\dots,r$, we can choose instead the chart setting $\lambda_{j}=1$. We
would then get a Hessian matrix whose determinant will vanish if and only if
the one we are computing vanishes, and we would deduce that the gradients of
the polynomials $p_{A_{i}}(u)$ for $i\neq j$ form a matrix of rank $r$, that
is, they are linearly independent. Moreover, this is exactly the condition
implying that $\phi_{f}$ belongs to the mixed-discriminantal variety
$MD(A_{0},\ldots,A_{r})=0$ which we denote by $X_{MD}$. It follows that
$X_{C}^{\nu}\subseteq X_{MD}$ and that $X_{MD}$ is also a hypersurface, i.e.
$MD(A_{0},\ldots,A_{r})\neq 1.$
The reverse inclusion follows essentially from the definition. In fact if
$\phi_{f}\in X_{MD}$ is generic, then there is a common zero
$u\in(C^{\nu})^{n}$ of $p_{A_{0}},\dots,p_{A_{r}}$ and a linear dependency
$\sum\lambda_{i}\nabla(p_{A_{i}})(u)=0$ with all $\lambda_{i}\neq 0,$ because
all the maximal minors in the matrix $\begin{bmatrix}\nabla(p_{A_{0}})(u)\\\
\ldots\\\ \nabla(p_{A_{r}})(u)\end{bmatrix}$ are assumed to be nonzero. It
follows that $\phi_{f}\in X_{C}^{\nu}.$ ∎
Notice that if $A_{0}=A_{1}=\ldots=A_{r}=A$ then
$C=\\{e_{0},\dots,e_{r}\\}\times A$, which is usually written as
$C=\Delta_{r}\times A$. Following [WZ94], we define the following quantity
associated to a projective variety $X$:
(3.2) $\mu(X)\,=\,\dim(X)+{\rm def}(X),$
where the defect of $X$ has been defined in (3.1). We end this section with
the following result about non-defectivity.
###### Proposition 3.4.
Let $A$ be a non-defective finite lattice configuration. Then, the associated
Cayley configuration $C=\Delta_{r}\times A$ is non-defective if and only if
$r\leq\dim(X_{A})$.
###### Proof.
Note that $X_{C}=\mathbb{P}^{r}\times X_{A}$. We can then use Theorem 0.1 in
[WZ96], which says that
$\mu(\mathbb{P}^{r}\times X_{A})={\rm max}(r+\dim(X_{A}),r+{\rm
def}(\mathbb{P}^{r}),\dim(X_{A})+{\rm def}(X_{A})).$
According to (3.2), we have that $\mu(X_{C})=r+\dim(X_{A})+{\rm def}(X_{C})$.
Since ${\rm def}(\mathbb{P}^{r})=r$, and by hypothesis ${\rm def}(X_{A})=0$,
we get that
$\mu(X_{C})={\rm max}(r+\dim(X_{A}),2r,\dim(X_{A})).$
When $r\leq\dim(X_{A})$, we get that $\mu(X_{C})=r+\dim(X_{A})$ which implies
that ${\rm def}(X_{C})=0$. On the other side, when $r>\dim(X_{A})$, we have
that $\mu(X_{C})=2r$ and so ${\rm def}(X_{C})=r-\dim(X_{A})>0$. ∎
## 4\. The multivariate iterated discriminant
In the remainder of the paper we will consider the case $A_{i}=A$ for
$i=0,\ldots,r.$ In order to establish an iterated process for the mixed
discriminant it is convenient to consider the geometric iterated discriminant
$JD_{r,A}$ introduced in Definition 4.2 below. In Proposition 4.6 we prove
that this polynomial coincides with the iterated discriminant $ID_{r,A}$ from
Definition 1.1. It implies that Theorem 4.4, which can be considered the main
result of this paper, also holds for $ID_{r,A}$, as stated in the
Introduction.
Recall that given an irreducible and reduced projective variety
$Y\subset\mathbb{P}^{N}$ of codimension $s$, its Chow form $Ch_{Y}$ is defined
as follows. Consider linear subspaces of dimension $\ell$ in $\mathbb{P}^{N}$,
$L\in Gr(\ell+1,N+1).$ If $s\geq\ell+1$, any generic $L$ will not intersect
$Y.$ The irreducible subvariety
$\\{L\in Gr(\ell+1,N+1)\,:\,L\cap Y\neq\emptyset\\}$
parametrizing the exceptional intersection locus, has codimension $(s-\ell)$
in $Gr(\ell+1,N+1).$ In case $\ell=s-1$ the defining polynomial is denoted by
$Ch_{Y}$ and it is called the Chow form of $Y$ [GKZ94, page 99].
We also need to recall two classical facts which will be used in the proof of
our main Theorem 4.4.
###### Remark 4.1.
Given a finite lattice configuration $A$ and a generic singular hyperplane
section of $X_{A}$, we can recover the intersection point by means of the
gradient of the discriminant $D_{A}$. Precisely,
1. (1)
As we are assuming that ${\rm char}(K)=0$, if a regular point $H$ in the dual
variety $X_{A}^{\nu}$ is tangent to $X_{A}$ at a regular point $y_{H}$, then
this projective point is unique and $y_{H}=\nabla D_{A}(H)$ [GKZ94, Th.1.1,
Ch. 1]. This is referred to as biduality.
2. (2)
When $X_{A}^{\nu}\subset(\mathbb{P}^{N})^{\nu}$ is a hypersurface, biduality
implies that the Gauss map $\gamma:X_{A}^{\nu}\dashrightarrow\mathbb{P}^{n}$
is defined by $H\mapsto\nabla D_{A}(H)=y_{H}$ and the closure of its image
equals $X_{A}$.
Let $A=\\{a_{0},\ldots,a_{N}\\}\subset\mathbb{Z}^{n}$ be a lattice
configuration. We will assume henceforth that $A$ is non-defective and that it
is a homogeneous polynomial of degree $\delta>0.$
Given $A$-polynomials
$p_{i}=\sum_{j=0}^{N}c_{ij}x^{a_{j}},\quad\quad i=0,\ldots r,$
we also denote by $(p_{0},\ldots,p_{r})\in(\mathbb{P}^{(r+1)(N+1)-1})^{\nu}$
the vector of their coefficients. For any
$\lambda=(\lambda_{0},\ldots,\lambda_{r})\in\mathbb{P}^{r}$ we write
$P_{\lambda}:=\lambda_{0}p_{0}+\ldots+\lambda_{r}p_{r}\in(\mathbb{P}^{N})^{\nu}.$
###### Definition 4.2.
Consider the incidence variety
(4.1)
$\Sigma=\left\\{((p_{0},p_{1},\ldots,p_{r}),\lambda)\in(\mathbb{P}^{(r+1)(N+1)-1})^{\nu}\times\mathbb{P}^{r}\,:\,\sum_{j}c_{ij}\frac{\partial
D_{A}}{\partial c_{j}}(P_{\lambda})=0,\,i=0,\ldots,r\right\\}.$
Let $\pi:\Sigma\to(\mathbb{P}^{(r+1)(N+1)-1})^{\nu}$ be the linear projection
onto the first factor. The $r$-multivariate iterated dual scheme $\pi(\Sigma)$
is defined by the projective elimination ideal
$\pi I=(I:m^{\infty})\cap\mathbb{C}[c],$
where $\mathbb{C}[c]$ is the ring of polynomials in the variables $c_{ij}$ and
$I$ is the ideal
$I=\left\langle\sum_{j}c_{ij}\frac{\partial D_{A}}{\partial
c_{j}}(P_{\lambda}),i=0,\ldots,r\right\rangle.$
When $\pi(\Sigma)$ has codimension one, we denote by
$JD_{r,A}\in\mathbb{Z}[c]$ a generator (unique up to multiplication by a
nonzero constant) of the union of the codimension one components of $\pi I$
and we call it the _geometric iterated discriminant_.
Notice that the projection is in general not irreducible; see for instance
Example 2.3. We will see in Proposition 4.6 below that the geometric iterated
discriminant $JD_{r,A}$ coincides with the more naive definition of the
iterated discriminant $ID_{r,A}$ from Definition 1.1.
Let
$(p,\lambda)=((p_{0},\ldots,p_{r}),(\lambda_{0},\ldots,\lambda_{r}))\in\Sigma$.
In order to understand the projection $\pi$ we consider two auxiliary maps,
$\phi:\Sigma\to X_{A}^{\nu}$ and
$T:(\mathbb{P}^{(r+1)(N+1)-1})^{\nu}\dashedrightarrow Gr(r+1,N+1):$
${\Sigma}$${X_{A}^{\nu}}$${\pi(\Sigma)}$${Gr(r+1,N+1)}$$\scriptstyle{\phi}$$\scriptstyle{T}$
defined by $\phi(p,\lambda)=P_{\lambda}$ and $T(p_{0},\ldots,p_{r})=T_{p},$
where we denote by $T_{p}$ the projective linear span of $p_{0},\ldots,p_{r}.$
###### Lemma 4.3.
Let $p=(p_{0},\ldots,p_{r})\in(\mathbb{P}^{(r+1)(N+1)-1})^{\nu},$ then
$JD_{r,A}(p)=0$ if and only if $T_{p}$ is tangent to $X_{A}^{\nu}$ at some
point $\xi.$
###### Proof.
If $JD_{r,A}(p_{0},\ldots,p_{r})=0$ then there exists $\lambda$ such that
$(p,\lambda)\in\Sigma$; let $\xi=P_{\lambda}.$ Consider the equalities
(4.2) $\sum_{j}c_{ij}\frac{\partial D_{A}}{\partial
c_{j}}(P_{\lambda})=0,\quad i=0,\ldots,r.$
The Euler relation implies $\deg(D_{A})\cdot D_{A}(P_{\lambda})=0,$ so
$D_{A}(P_{\lambda})=0$ and thus $P_{\lambda}\in X_{A}^{\nu}.$ Moreover (4.2)
implies that each $p_{i}$ lies in $T_{X^{\nu},\xi}$, which is equivalent to
$T_{p}\subset T_{X^{\nu},\xi}.$ The converse follows the same line of proof. ∎
Recall that we denote by $\textrm{sing}(X_{A}^{\nu})$ the subscheme of
$X_{A}^{\nu}$ defined by the ideal generated by the partial derivatives of
$D_{A}.$
###### Theorem 4.4.
Assume $A\subset\mathbb{Z}^{n}$ is non-defective and let $r\in\mathbb{Z}$ with
$0\leq r\leq\dim(X_{A}).$ Then, the mixed discriminant $MD_{r,A}$ always
divides the geometric iterated discriminant $JD_{r,A}$. Moreover:
1. (1)
If ${\rm codim}_{X_{A}^{\nu}}(sing(X_{A}^{\nu}))>r$, then $JD_{r,A}=MD_{r,A}$.
2. (2)
If ${\rm codim}_{X_{A}^{\nu}}(sing(X_{A}^{\nu}))=r$, then
$JD_{r,A}=MD_{r,A}\prod_{i=k}^{\ell}Ch_{Y_{k}}^{\mu_{k}},$ where
$Y_{1},\dots,Y_{\ell}$ are the irreducible components of $sing(X_{A}^{\nu})$
of maximal dimension $r$, with respective multiplicities $\mu_{k}\geq 1$.
3. (3)
If ${\rm codim}_{X_{A}^{\nu}}(sing(X_{A}^{\nu}))<r$, then
$\pi(\Sigma)=(\mathbb{P}^{(r+1)(N+1)-1})^{\nu},$ and $JD_{r,A}=0.$
###### Proof.
As already observed, in the classical case of $r=0$ we have
$ID_{0,A}=MD_{0,A}=D_{A}.$ Note also that by Propositions 3.4 and 3.3,
$\deg(MD_{r,A})>0$ and it is irreducible.
Observe that the map $\phi$ is surjective since for any $F\in X_{A}^{\nu},$
$F=\phi(F,\ldots,F,\frac{1}{(r+1)},\ldots,\frac{1}{(r+1)})$ and we have
$(F,\ldots,F,\frac{1}{(r+1)},\ldots,\frac{1}{(r+1)})\in\Sigma.$ The rational
map $T$ is defined over the open dense subset
$U_{T}=\\{p\;:\,T_{p}\simeq\mathbb{P}^{r}\\}$ of all $p$ with linear span of
projective dimension $r$. Notice also that $T$ is surjective and that for each
$H\in Gr(r+1,N+1)$, the fiber $T^{-1}(H)$ has dimension $(r+1)^{2}-1.$ Let
$\Sigma^{\circ}=\phi^{-1}((X_{A}^{\nu})_{reg})$ and
$\Sigma^{\prime}=\phi^{-1}(sing(X_{A}^{\nu}))$; that is, let
$\Sigma^{\circ}=\\{(p,\lambda)\in\Sigma\,:\,P_{\lambda}\in(X_{A}^{\nu})_{reg}\\},\quad\Sigma^{\prime}=\\{(p,\lambda)\in\Sigma\,:\,P_{\lambda}\in
sing(X_{A}^{\nu})\\}.$
It follows that $\pi(\Sigma)=\pi(\Sigma^{\circ})\cup\pi(\Sigma^{\prime}).$
We claim that $\overline{\pi(\Sigma^{\circ})}\subseteq V(MD_{r,A}).$ In fact,
take a generic point $(p,\lambda)\in\pi(\Sigma^{\circ})$. We can then assume
that not only $P_{\lambda}\in(X_{A}^{\nu})_{reg}$, but also there is a unique
regular point $y=(x^{m_{0}}:\ldots:x^{m_{N}})\in X_{A}$ with $x\in(K^{*})^{n}$
such that $P_{\lambda}(x)=0$ and $\frac{\partial P_{\lambda}}{\partial
x_{i}}(x)=0,\,\,i=1,\ldots,n.$ By Remark 4.1, $y=(\frac{\partial
D_{A}}{\partial c_{0}}(P_{\lambda}):\ldots:\frac{\partial D_{A}}{\partial
c_{n}}(P_{\lambda})).$ The equations $\frac{\partial P_{\lambda}}{\partial
x_{i}}(x)=0$ for $i=1,\ldots,n$ mean that $\sum\lambda_{i}\nabla(p_{i})(x)=0.$
Moreover $p_{i}(x)=0$ for all $i$ because
(4.3) $p_{i}(x)=\sum_{j}c_{ij}y_{j}=k\sum_{j}c_{ij}\frac{\partial
D_{A}}{\partial c_{j}}(P_{\lambda})=0,\quad i=0,\ldots r$
for some $k\in K^{*}$ such that $y=k\,\nabla D_{A}(P_{\lambda}).$ This implies
that $MD_{r,A}(p_{0},\ldots,p_{r})=0.$
We now show that $V(MD_{r,A})\subseteq\pi(\Sigma).$ Let $(p_{0},\ldots,p_{r})$
be a generic element in the zero locus of $MD_{r,A}.$ Then there exists
$(u,\lambda)\in(K^{*})^{n+r+1}$ such that
$p_{0}(u)=\cdots=p_{r}(u)=0\text{ and }\begin{bmatrix}\nabla
p_{0}(u)&\cdots&\nabla p_{r}(u)\end{bmatrix}\begin{bmatrix}\lambda_{0}\\\
\vdots\\\ \lambda_{r}\end{bmatrix}=0.$
We claim that $(p,\lambda)\in\Sigma.$ If $P_{\lambda}\in sing(X_{A}^{\nu})$ is
generic then $\frac{\partial D_{A}}{\partial c_{j}}(P_{\lambda})=0$ and thus
$(p_{0},\ldots,p_{r})\in\pi(\Sigma).$ If instead
$P_{\lambda}\in(X_{A}^{\nu})_{reg}$ is generic, then biduality gives $\nabla
D_{A}(P_{\lambda})=y$ with $y=(u^{m_{0}}:\ldots:u^{m_{N}})$ and thus
$\sum_{j}c_{ij}\frac{\partial D_{A}}{\partial c_{j}}(P_{\lambda})=p_{i}(u)=0$
as in (4.3), implying again that $(p_{0},\ldots,p_{r})\in\pi(\Sigma).$ We have
then proved that
(4.4) $\overline{\pi(\Sigma^{\circ})}\subseteq
V(MD_{r,A})\subseteq\pi(\Sigma).$
Consider the non-embedded primary components of the ideal
$\langle\frac{\partial D_{A}}{\partial c_{j}},j=0,\ldots,N\rangle$ defining
the singular locus of $X_{A}^{\nu}.$ Correspondingly, we consider the
decomposition into irreducible components $sing(X_{A}^{\nu})=\bigcup Y_{k}.$
Define
(4.5) $V_{k}=\\{H\in Gr(r+1,N+1)\,:H\cap Y_{k}\neq\emptyset\\}\text{ and
}\Sigma_{k}=\phi^{-1}(Y_{k}).$
Recall that $\text{\rm codim}_{Gr(r+1,N+1)}(V_{k})=\max\\{0,\text{\rm
codim}_{\mathbb{P}^{N}}(Y_{k})-r\\}.$
Assume that $\text{\rm codim}_{X_{A}^{\nu}}(sing(X_{A}^{\nu}))>r$. Then
$\text{\rm codim}_{Gr(r+1,N+1)}(V_{k})\geq 2$ for all $k.$ It follows that
$\text{\rm
codim}_{{\mathbb{P}^{(r+1)(N+1)-1}}^{\nu}}(\pi(\Sigma_{i}))=\text{\rm
codim}_{{\mathbb{P}^{(r+1)(N+1)-1}}^{\nu}}(\overline{\pi(\Sigma_{i})\cap
U_{T}})\geq\text{\rm
codim}_{{\mathbb{P}^{(r+1)(N+1)-1}}^{\nu}}(\overline{T^{-1}(V_{i})})\geq 2$
for all $i$. The containment in Equation (4.4) then implies that $\pi(\Sigma)$
is of codimension one and set-theoretically coincides with $V(MD_{r,A})$.
As the mixed discriminant $MD_{r,A}$ is irreducible, it remains to show that
the multiplicity of $MD_{r,A}$ in $ID_{r,A}$ is one. For that, it is enough to
show that there exists $(p_{0}^{\nu},\ldots,p_{r}^{\nu})\in V(MD_{r,A})$ and
$\lambda^{\nu}$ such that $(p^{\nu},\lambda^{\nu})\in\Sigma$ and
$d\pi((p^{\nu},\lambda^{\nu}))$ has maximal rank.
We start by choosing a point $\xi\in reg(X_{A}^{\nu})$ such that
$\textrm{rank}(H)=n,$ where $H=Hess(D_{A})(\xi).$ Notice that $H={\rm
Jac}(\gamma)(\xi),$ where $\gamma:X_{A}^{\nu}\dasharrow X_{A}$ is the Gauss
map defined as $\gamma(y)=\nabla D_{A}(y)$ which has generic rank equal to
$n=\dim(X_{A}).$ Up to a change of coordinates, $H$ can be assumed to be of
the form
(4.6) $H=\begin{bmatrix}I_{n}&0\\\ 0&0\end{bmatrix}.$
Consider $Z=\\{M\in Gr(r+1,N+1)\,:\,\xi\in M\subset T_{X_{A}^{\nu},\xi}\\}$
and recall that we assume that $r\leq n.$
Note that for every $p\in T^{-1}(Z),$ we have $\xi\in T_{p}\subset
T_{X_{A}^{\nu},\xi}$ and $\dim(T_{p})=r$ as $p\in U_{T}$. It follows that
there exists a unique $\lambda$ such that $(p,\lambda)\in\Sigma$ and
$\phi(p,\lambda)=\xi.$ Consider the $(r+1)\times(N+1)$ matrix $M_{p}$ whose
$i$-th row corresponds to the coefficients of $p_{i}$. Without loss of
generality, the matrix $M_{p}$ can be assumed to be of the form $[I_{r+1},C].$
Let
$\chi(p)=\det\begin{bmatrix}I_{r+1}&C\\\ -C&I_{r+1}\end{bmatrix}.$
Notice that if all the entries of $C$ are $0$ then $\chi(p)=1$, and thus
$\chi$ is not identically zero. Choose now $p\in T^{-1}(Z)$ such that
$\chi(p)\neq 0$ and let $\lambda^{\nu}$ be the unique solution to
$\xi=\sum\lambda^{\nu}_{i}p_{i}.$ Note that
$d\pi(p,\lambda^{\nu})=M_{p}HM_{p}^{t}.$ If this matrix is not of maximal rank
then there is $\alpha\in\mathbb{P}^{r}$ such that $M_{p}HM_{p}^{t}\alpha=0.$
Since $H$ is as in (4.6) and $M_{p}^{t}\alpha\neq 0$ then
$HM_{p}^{t}\alpha\in\ker(M_{p}),$ which is a contradiction to $\chi(p)\neq 0.$
Assume that $\text{\rm codim}_{X_{A}^{\nu}}(sing(X_{A}^{\nu}))<r$. Then
$\text{\rm codim}_{\mathbb{P}^{N}}(Y_{k})<r+1$ for all $k.$ The assumption
also implies that any element of the Grassmannian belongs to $V_{k}$ for all
$k$ (defined in (4.5)) and that $\pi(\Sigma^{\prime})\cap
U_{T}=T^{-1}(Gr(r+1,N+1))=U_{T}.$ It follows that
$\mathbb{P}^{(r+1)(N+1)-1}=\overline{U_{T}}=\overline{\pi(\Sigma_{i})\cap
U_{T}}=\pi(\Sigma_{i})\subset\pi(\Sigma)$ and thus
$\pi(\Sigma)=(\mathbb{P}^{(r+1)(N+1)-1})^{\nu}$. Assume now that $\text{\rm
codim}_{X_{A}^{\nu}}(sing(X_{A}^{\nu}))=r.$ Then $\text{\rm
codim}_{Gr(r+1,N+1)}(V_{k})=1$ and thus by definition
$\pi(\Sigma_{k})=V(Ch_{Y_{k}}^{\mu_{k}})$. ∎
###### Remark 4.5.
= We have not identified the exponents $\mu_{k}$ occurring in the
factorization of the iterated discriminant. Based on the evidence we have
collected through examples (see for instance Examples 2.3 and Example 4.7) and
the results in [BM09], we see some evidence of the following. We conjecture
that $\mu_{k}=2$ if $Y_{k}$ is a component corresponding to the closure of the
locus of those $p$ for which there are two different non-degenerate multiple
roots (the double point locus), while $\mu_{k}=3$ when $Y_{k}$ is a component
corresponding to the locus of those $p$ for which there is a degenerate
multiple root (the cusp locus). Weyman and Zelevinsky showed in [WZ96] that
for some hyperdeterminants there can be more than one irreducible component of
each type.
The following Proposition 4.6 explains the name geometric iterated
discriminant: we show that under the hypotheses of Theorem 4.4, $JD_{r,A}$
equals $ID_{r,A}$ and thus when it is nonzero it can be computed as a
discriminant of a discriminant.
Recall that, given a natural number $d$, we denote by $d\Delta_{r}$ in
$\mathbb{R}^{r+1}$ the lattice configuration given by the integer points in
the dilated unit simplex $d$ times, and by $D_{d\Delta_{r}}$ the associated
discriminant. For any homogeneous polynomial
$H=H(\lambda_{0},\dots,\lambda_{r})$ of degree $d$, the discriminant of $H$
equals, up to constant, the resultant of its partial derivatives:
(4.7) $D_{d\Delta_{r}}(H)\,=\,{\rm Res}_{d-1}\left(\frac{\partial
H}{\partial\lambda_{0}},\dots,\frac{\partial H}{\partial\lambda_{r}}\right),$
where ${\rm Res}_{d-1}$ denotes the homogeneous resultant associated to $r+1$
homogeneous polynomials of degree $d-1$ [GKZ94, Jou91]. Moreover, the
following universal property is proved in [Jou91]. Let
$G_{0},\dots,G_{r}\in\mathbb{Z}[u][\lambda_{0},\dots,\lambda_{r}]$ have degree
$d-1$ with generic coefficients $u$:
$G_{i}(u,\lambda)=\sum_{|\alpha|=d_{i}}\,u_{i,\alpha}\,\lambda^{\alpha}.$
Denote by $\langle
G_{0},\dots,G_{r}\rangle\subset\mathbb{Z}[u][\lambda_{0},\dots,\lambda_{r}]$,
the ideal generated by $G_{0},\dots,G_{r}$ in this ring. Then for any variable
$\lambda_{i}$ and any $N>\sum_{i}d_{i}-n$, it holds that
(4.8) $\lambda_{i}^{N}\,{\rm Res}_{d-1}(G_{0},\dots,G_{r})\in\langle
G_{0},\dots,G_{r}\rangle.$
Thus, such an equality holds for any specialization of the coefficients $u$ in
a ring.
Let $A$ be a non-defective configuration with $\text{\rm
codim}(sing(X_{A}^{\nu}))\geq r$. Call $\delta=\deg(D_{A})$ and for a choice
of $A$-polynomials $p_{0},\dots,p_{r}$ consider the evaluation
$D_{A}(P_{\lambda})=D_{A}(\sum_{i=0}^{r}\lambda_{i}p_{i})$, which is either
zero or a homogeneous polynomial in $\lambda=(\lambda_{0},\dots,\lambda_{r})$
of degree $\delta$.
###### Proposition 4.6.
Under the hypotheses of Theorem 4.4, the following equality holds:
$JD_{r,A}=ID_{r,A}.$
Moreover, when ${\rm codim}_{X_{A}^{\nu}}(sing(X_{A}^{\nu}))\geq r$, the
degree of the iterated discriminant equals
(4.9) $\deg(ID_{r,A})=(r+1)\delta(\delta-1)^{r}.$
###### Proof.
By Theorem 4.4 and Definition 1.1, we can assume that ${\rm
codim}_{X_{A}^{\nu}}(sing(X_{A}^{\nu}))\geq r$.
Let $(p_{0}^{0},\dots,p_{r}^{0},\lambda^{0})$ be a point in the incidence
variety $\Sigma$ defined in (4.1). Note that for any $i=0,\dots,r$, we have
that
$0\,=\,\sum_{j}c_{ij}\frac{\partial D_{A}}{\partial
c_{j}}(\sum_{i=0}^{r}\lambda_{i}^{0}p_{i}^{0})\,=\frac{\partial}{\partial\lambda_{i}}D_{A}(\sum_{i=0}^{r}\lambda_{i}p_{i}^{0})(\lambda^{0}).$
Then, $D_{\delta\Delta_{r}}(D_{A}(\sum_{i=0}^{r}\lambda_{i}p_{i}^{0}))=0.$
Conversely, as we pointed out in (4.7), this homogeneous discriminant equals
the resultant
${\rm
Res}_{\delta-1}\left(\frac{\partial}{\partial\lambda_{0}}D_{A}(\sum_{i=0}^{r}\lambda_{i}p_{i}^{0}),\dots,\frac{\partial}{\partial\lambda_{r}}D_{A}(\sum_{i=0}^{r}\lambda_{i}p_{i}^{0})\right).$
It then follows from [Jou91] that if
$D_{\delta\Delta_{r}}(D_{A}(\sum_{i=0}^{r}\lambda_{i}p_{i}^{0}))=0$, then
there exists $\lambda^{0}\in\mathbb{P}^{r}$ which is a common zero of all
these partial derivatives. Moreover, we deduce from Equation (4.8) that for
any ring $R$ containing the coefficients of $D_{A}(P_{\lambda})$ and for any
$i=0,\dots,r$, the iterated discriminant $D_{\delta\Delta_{r}}(P_{\lambda})$
lies in the ideal generated by the partial derivatives
$\frac{\partial}{\partial\lambda_{j}}D_{A}(\sum_{i=0}^{r}\lambda_{i}p_{i}^{0}),j=0,\dots,r,$
in the localization $R[\lambda_{0},\dots,\lambda_{r}]_{\lambda_{i}}$. It
follows that
$D_{\delta\Delta_{r}}(D_{A}(\sum_{i=0}^{r}\lambda_{i}p_{i}^{0}))=0$ defines
the schematic image of $\Sigma$ and thus it coincides with
$ID_{r,A}(p_{0},\dots,p_{r})$.
To see that Equation (4.9) holds, recall that $\deg(D_{A})=\delta$ and so the
degree of $D_{A}(\sum_{i=0}^{r}\lambda_{i}p_{i})$ in the coefficients of
$p_{0},\dots,p_{r}$ as well as in the $\lambda$ variables is equal to
$\delta$. On the other side, the degree of $D_{\delta\Delta_{r}}$ is equal to
$(r+1)\delta^{r}$. ∎
We conclude this section with two examples that illustrate Theorem 4.4 with
$r+1=2.$ In the first one, the singular locus has codimension $r+1=2$, which
implies a factor (with multiplicity $2$) of the iterated discriminant. In the
second one, the singular locus has codimension bigger than $2$, which implies
equality between $MD_{1,A}$ and $ID_{2,A}.$
###### Example 4.7.
[Example 2.3, continued.] Consider again the two dimensional configuration
corresponding to the first Hirzebruch surface ${\mathbb{F}}_{1}$:
$A=\\{(0,0),(1,0),(2,0),(0,1),(1,1)\\}.$
Given a generic $A$-polynomial
$f(x,y)=a_{0}+a_{1}x+a_{2}x^{2}+y(b_{0}+b_{1}x)$, we saw that
$D_{A}(f)\,=a_{0}b_{1}^{2}-a_{1}b_{0}b_{1}+a_{2}b_{0}^{2}$. The ideal defining
the singular locus $S$ of $X_{A}^{\nu}$ is generated by
$b_{0}^{2},b_{0}b_{1},b_{1}^{2},-a_{1}b_{1}+2a_{2}b_{0},2a_{0}b_{1}-a_{1},b_{0}.$
This ideal has multiplicity $2$ and its radical is generated by $b_{0},b_{1}$.
In this case, $ID_{2,A}$ has another irreducible factor $Ch_{S}$ of degree $2$
coming from the Chow form of $S$, to the second power:
$ID_{2,A}=MD_{1,A}\,\cdot\,Ch_{S}^{2},$
where
$Ch_{S}((a_{0},a_{1},a_{2},b_{0},b_{1}),(A_{0},A_{1},A_{2},B_{0},B_{1}))=B_{0}b_{1}-B_{1}b_{0}$.
###### Example 4.8.
Let $X_{A}$ be the Segre embedding of
$\mathbb{P}^{1}\times\mathbb{P}^{1}\times\mathbb{P}^{1}$, so $D_{A}$ is the
hyperdeterminant of format $(2,2,2)$ of degree $4$, whose singular locus has
codimension greater than $2$ by [WZ96]. Take $r=2$, so that $MD_{1,A}$ equals
the discriminant of the hyperdeterminant of format $(2,2,2,2)$ (corresponding
to the Segre embedding of
$\mathbb{P}^{1}\times\mathbb{P}^{1}\times\mathbb{P}^{1}\times\mathbb{P}^{1}$.)
In this case, $MD_{1,A}$ equals the iterated discriminant $ID_{2,A}$ and thus
has degree $2\cdot 4\cdot(4-1)^{1}=24$.
This is the only known case of polynomials of degree bigger than $2$ for which
the iterated and the mixed discriminants coincide.
## 5\. Comparing mixed and iterated discriminants
In this section we consider the case when $A$ equals the lattice points in a
cartesian product of dilates of standard simplices:
$d_{1}\Delta_{k_{1}}\times\cdots\times d_{\ell}\Delta_{k_{l}}$, for some
$\ell\geq 1$. In other words we investigate Segre-Veronese varieties
$X_{A}=\mathbb{P}^{k_{1}}(d_{1})\times\cdots\times\mathbb{P}^{k_{\ell}}(d_{\ell}).$
The symbol $\mathbb{P}^{k}(d)$ denotes the Veronese embedding of degree $d$ in
dimension $k$, i.e. the variety $\mathbb{P}^{k}$ embedded in
$\mathbb{P}^{{k+d\choose d}-1}$ by the global sections of the line bundle
${\mathcal{O}}_{{\mathbb{P}}^{k}}(d).$ We occasionally denote
$\mathbb{P}^{k}(1)$ by $\mathbb{P}^{k}.$
The symbol
$\mathbb{P}^{k_{1}}(d_{1})\times\cdots\times\mathbb{P}^{k_{\ell}}(d_{\ell})$
denotes the Segre embedding of the above defined Veronese embeddings, more
precisely the variety
$\mathbb{P}^{k_{1}}\times\cdots\times\mathbb{P}^{k_{\ell}}$ embedded via the
global sections of the line bundle
$\pi^{*}_{1}{\mathcal{O}}_{{\mathbb{P}}^{k_{1}}}(d_{1})\otimes\ldots\otimes\pi_{\ell}^{*}{\mathcal{O}}_{{\mathbb{P}}^{k_{\ell}}}(d_{\ell}),$
where $\pi_{i}$ denotes the $i^{th}$ projection
$\pi_{i}:\mathbb{P}^{k_{1}}\times\cdots\times\mathbb{P}^{k_{\ell}}\to\mathbb{P}^{k_{i}}.$
These are toric embeddings corresponding to the configurations of lattice
points of the polytopes $d_{1}\Delta_{k_{1}}\times\ldots\times
d_{l}\Delta_{k_{\ell}}.$
When $d_{i}=1$ we recover the case of hyperdeterminants, which has been
completely solved in [WZ96]. In Proposition 5.2 we show that that when
$\ell=1$ there is equality if and only if $r=1$ and $d_{1}=2$. We then
conjecture that these are all the possible cases (see Conjecture 5.3), that
is, in all other cases the singularities of the discriminantal locus have
codimension one in the dual variety. We conclude with Theorem 5.6, which
covers the case in which all $d_{i}>1$.
To determine when the iterated and mixed discriminants of Segre-Veronese
varieties are equal, we start with the following lemma, which allows us to
compute the degree of $MD_{r,d\Delta_{n}}$, that is, the case in which we
consider $(r+1)$ polynomials of degree $d$ in $n$ variables. Recall that when
$r\leq n$, we know by Proposition 3.3 that the mixed discriminant equals the
discriminant of the Cayley configuration given by the lattice points in the
product of simplices $\Delta_{r}\times d\Delta_{n}$.
###### Lemma 5.1.
If $r\leq n$ then
$\deg(MD_{r,d\Delta_{n}})=(n+1)\binom{n}{r}d^{r}(d-1)^{n-r}$.
###### Proof.
We will use [GKZ94, Ch. 13, Theorem 2.4], which tells us that this degree is
equal to the coefficient of the monomial $x^{r}y^{n}$ in the expansion of
$S(x,y)=\frac{1}{\left((1+x)(1+y)-x(1+y)-dy(1+x)\right)^{2}}=\frac{1}{(1-(d-1)y-dxy)^{2}}.$
We may write
$S(x,y)=\left(\frac{1}{1-q}\right)^{2}=\sum_{n\geq 0}(n+1)q^{n},$
where $q=(d-1)y+dxy=y((d-1)+dx)$.
Since
$q^{n}=\left(\sum_{j=0}^{n}\binom{n}{j}(d-1)^{n-j}d^{j}x^{j}\right)y^{n},$
we have
(5.1) $S(x,y)\,=\,\sum_{n\geq
0}\sum_{j=0}^{n}(n+1)\binom{n}{j}(d-1)^{n-j}d^{j}x^{j}y^{n}.$
From this expansion, we see that the coefficient of $x^{r}y^{n}$ is equal to
$(n+1)\binom{n}{r}(d-1)^{n-r}d^{r}$
when $r\leq n$, and is equal to $0$ if $r>n$. This completes the proof.
∎
###### Proposition 5.2.
Let $1<d$ and $1\leq r\leq n$. Then $MD_{r,d\Delta_{n}}=ID_{r,d\Delta_{n}}$ if
and only if $r=1$ and $d=2$.
The fact that this equality holds in the case of $r=1$ and $d=2$ was shown in
[Ott13, Theorem 8.2]. Although we include this in our proof for completeness,
the main contribution of this result is that equality does not hold in any
other case.
###### Proof.
For any $d>1$, the configuration of lattice points in $d\Delta_{n}$ is non-
defective [BJ14] and as $r\leq n$, it is enough to check that
$\deg(MD_{r,d\Delta_{n}})=\deg(ID_{r,d\Delta_{n}})$ by Propositions 3.3 and
3.4.
From Lemma 5.1 we know that
$\deg(MD_{r,d\Delta_{n})})=(n+1)\binom{n}{r}d^{r}(d-1)^{n-r}.$
By Proposition 4.6 we also know that
$\deg(ID_{r,d\Delta_{n}})=(n+1)(d-1)^{n}(r+1)((n+1)(d-1)^{n}-1)^{r}.$
To determine when these are equal, we will consider the ratio of the two
degrees, both of which are nonzero for $d>1$. We have
$\frac{\deg(ID_{r,d\Delta_{n}})}{\deg(MD_{r,d\Delta_{n}})}=\frac{(n+1)(d-1)^{n}(r+1)((n+1)(d-1)^{n}-1)^{r}}{(n+1)\binom{n}{r}d^{r}(d-1)^{n-r}}=\frac{(d-1)^{r}(r+1)((n+1)(d-1)^{n}-1)^{r})}{\binom{n}{r}d^{r}}.$
For any $d>1$ we have
$(n+1)(d-1)^{n}-1=n(d-1)^{n}+(d-1)^{n}-1\geq n(d-1)^{n},$
with equality if and only if $d=2$. Thus the numerator satisfies
$(d-1)^{r}(r+1)((n+1)(d-1)^{n}-1)^{r}\geq(d-1)^{r}(r+1)(n(d-1)^{n})^{r}=(d-1)^{r(n+1)}(r+1)n^{r}.$
Since $\binom{n}{r}\leq\frac{n^{r}}{r!}$, we have
$\frac{\deg(ID_{r,d\Delta_{n}})}{\deg(MD_{r,d\Delta_{n}})}\geq\frac{(d-1)^{r(n+1)}(r+1)n^{r}}{\frac{n^{r}}{r!}d^{r}}=\frac{(d-1)^{r(n+1)}(r+1)!}{d^{r}}=\left(\frac{(d-1)^{n+1}}{d}\right)^{r}\cdot(r+1)!$
If $d=2$, then this ratio is $\frac{(r+1)!}{d^{r}}=\frac{(r+1)!}{2^{r}}\geq
1$, with equality if and only if $r=1$. If $d>2$, then
$(d-1)^{n+1}\geq(d-1)^{2}\geq 2(d-1)=2d-2>d$. Thus $\frac{(d-1)^{n+1}}{d}>1$,
and so
$\frac{\deg(ID_{r,d\Delta_{n}})}{\deg(MD_{r,d\Delta_{n}})}>(r+1)!.$
Thus except possibly in the case of $d=2$ and $r=1$, we have
$\deg(ID_{r,d\Delta_{n}})>\deg(MD_{r,d\Delta_{n}})$.
To see that $d=2$ and $r=1$ gives
$\deg(MD_{r,d\Delta_{n})})=\deg(ID_{r,d\Delta_{n}})$, note that in this case
the ratio of the degrees is
$\frac{(2-1)^{1}(1+1)((n+1)(2-1)^{n}-1)^{1})}{\binom{n}{1}2^{1}}=\frac{2(n+1-1)}{2n}=1.$
∎
Geometrically, Proposition 5.2 shows that for
$\mathbb{P}^{r}(1)\times\mathbb{P}^{n}(d)$ the associated mixed discriminant
is equal to the iterated discriminant only when $r=1$ and $d=2$. Note that we
don’t consider the case $d=1$ because this case is defective.
It is natural to consider the same question for any product-of-simplices :
$\mathbb{P}^{r}(1)\times\mathbb{P}^{k_{1}}(d_{1})\times\cdots\times\mathbb{P}^{k_{\ell}}(d_{\ell}).$
Setting $d=(d_{1},\ldots,d_{\ell})$ and $k=(k_{1},\ldots,k_{\ell})$, let
$A_{\ell,d,k}$ denote the configuration corresponding to
$\mathbb{P}^{k_{1}}(d_{1})\times\cdots\times\mathbb{P}^{k_{\ell}}(d_{\ell})$.
We conjecture the following:
###### Conjecture 5.3.
We have $ID_{r,A_{\ell,d,k}}=MD_{r,A_{\ell,d,k}}$ if and only if
$\mathbb{P}^{r}(1)\times\mathbb{P}^{k_{1}}(d_{1})\times\cdots\times\mathbb{P}^{k_{\ell}}(d_{\ell})$
is of one of the following forms:
1. (1)
$\mathbb{P}^{r}\times\mathbb{P}^{m}\times\mathbb{P}^{m},m\geq 1,r=1,2$,
2. (2)
$(\mathbb{P}^{1})^{4}$,
3. (3)
$\mathbb{P}^{1}\times\mathbb{P}^{n}(2)$.
This conjecture was inspired by the question posed in [GKZ94, Chapter 14, pg
479], which coincides with the above conjecture when $d_{i}=1$ for all $i$.
Their conjecture (and thus our conjecture in this special case) was proved in
[WZ96]. Note that Proposition 5.2 implies that Conjecture 5.3 is true when
$\ell=1$, which puts us into case (3).
To study our conjecture in general, the following theorem giving the degree of
the mixed discriminant $MD_{r,A_{\ell,d,k}}$ will be useful. Let $B=B_{\ell}$
be the set of all non-empty subsets $\Omega\subset\\{0,1,\ldots,\ell\\}$. For
each $\Omega\in B$, let
$d_{\Omega}=\sum_{j\in\Omega}d_{j}.$
Let $\delta(\Omega)\in\mathbb{Z}^{\ell+1}_{+}$ be the characteristic vector of
$\Omega$. For every $\kappa=(r,k_{1},\ldots,k_{\ell})\in\mathbb{Z}_{+}^{r+1}$,
let $\mathcal{P}(\kappa)$ denote the set of all partitions of $\kappa$ into a
sum of vectors $\delta(\Omega)$; in other words, $\mathcal{P}(\kappa)$ is the
set of all non-negative integral vectors $(m_{\Omega})_{\Omega\in B}$ such
that $\sum_{\Omega\in B}m_{\Omega}\delta(\Omega)=\kappa$.
###### Theorem 5.4 (Theorem 13.2.5, [GKZ94]).
The degree of $MD_{r,A_{\ell,d,k}}$ is given by
$\sum_{(m_{\Omega})\in\mathcal{P}(\kappa)}\left(1+\sum_{\Omega\in
B}m_{\Omega}\right)!\prod_{\Omega\in
B}\frac{(d_{\Omega}-1)^{m_{\Omega}}}{m_{\Omega}!}.$
Note that any partition using the vector $\delta(\\{0\\})=(1,0,\ldots,0)$ will
not contribute to this sum, since $d_{\\{0\\}}-1=0$. Letting $C$ be the set of
all nonempty subsets of $\\{1,\ldots,\ell\\}$ and letting
$k=(k_{1},\ldots,k_{\ell})$ as before, we have that
$\deg\left(ID_{r,A_{\ell,d,k}}\right)=(r+1)\delta(\delta-1)^{r},$ where
$\delta=\sum_{(m_{\Omega})\in\mathcal{P}(k)}\left(1+\sum_{\Omega\in
C}m_{\Omega}\right)!\prod_{\Omega\in
C}\frac{(d_{\Omega}-1)^{m_{\Omega}}}{m_{\Omega}!}.$
When it is clear from context, we will abbreviate $\deg(MD_{r,A_{\ell,d,k}})$
as $\deg(MD)$ and $\deg(ID_{r,A_{\ell,d,k}})$ as $\deg(ID)$.
###### Example 5.5.
Let us compare the degrees of the mixed and the iterated discriminant when
$r=1$, $\ell=2$, and $k_{1}=k_{2}=1$. To compute the degree of the mixed
discriminant, we consider all partitions of $\kappa=(1,1,1)$. We may discount
any partition with the vector $(1,0,0)$, as this partition would contribute a
term of $0$ to $\deg(MD)$. Thus, the only relevant partitions are
* •
$(1,1,1)$,
* •
$(1,1,0)+(0,0,1)$, and
* •
$(1,0,1)+(0,1,0)$.
The contributions from these terms to $\deg(MD)$ are
* •
$2!\cdot\frac{(d_{1}+d_{2})^{1}}{1!}=2(d_{1}+d_{2})$ ,
* •
$3!\cdot\frac{d_{1}^{1}\cdot(d_{2}-1)^{1}}{1!\cdot 1!}=6d_{1}(d_{2}-1)$, and
* •
$3!\cdot\frac{d_{2}^{1}\cdot(d_{1}-1)^{1}}{1!\cdot 1!}=6d_{2}(d_{1}-1)$,
respectively. (Note that some of these contributions will be zero if one or
both of $d_{1}$ and $d_{2}$ are equal to $1$.) Adding these gives
$\deg(MD)=2(d_{1}+d_{2})+6d_{1}(d_{2}-1)+6d_{2}(d_{1}-1)=12d_{1}d_{2}-4d_{1}-4d_{2}=4(3d_{1}d_{2}-d_{1}-d_{2}).$
To compute $\deg(ID)$, we must consider the partitions of $k=(1,1)$. There are
only two: $(1,1)$ and $(1,0)+(0,1)$. The contributions of these to $\delta$
are
* •
$2!\cdot\frac{(d_{1}+d_{2}-1)^{1}}{1!}=2(d_{1}+d_{2}-1)$ and
* •
$3!\cdot\frac{(d_{1}-1)^{1}(d_{2}-1)^{1}}{1!\cdot 1!}=6(d_{1}-1)(d_{2}-1),$
respectively. Thus
$\delta=2(d_{1}+d_{2}-1)+6(d_{1}-1)(d_{2}-1)=6d_{1}d_{2}-4d_{1}-4d_{2}+4.$ It
follows that
$\deg(ID)=2\delta(\delta-1)=2(6d_{1}d_{2}-4d_{1}-4d_{2}+4)(6d_{1}d_{2}-4d_{1}-4d_{2}+3).$
We will now argue that $\deg(ID)>\deg(MD)$, unless $d_{1}=d_{2}=1$. First we
perform the change of variables $d_{1}=d_{1}^{\prime}+1$ and
$d_{2}=d_{2}^{\prime}+1$, to remove some of the negatives. This gives
$\deg(MD)=4(3d_{1}^{\prime}d_{2}^{\prime}+2d_{1}^{\prime}+2d_{2}^{\prime}+1)$
and
$\deg(ID)=2(6d_{1}^{\prime}d_{2}^{\prime}+2d_{1}^{\prime}+2d_{2}^{\prime}+2)(6d_{1}^{\prime}d_{2}^{\prime}+2d_{1}^{\prime}+2d_{2}^{\prime}+1).$
Now, if either $d_{1}$ or $d_{2}$ is greater than $1$, then
$(6d_{1}^{\prime}d_{2}^{\prime}+2d_{1}^{\prime}+2d_{2}^{\prime}+1)$ is at
least $3$, meaning that
$\deg(ID)\geq
6(6d_{1}^{\prime}d_{2}^{\prime}+2d_{1}^{\prime}+2d_{2}^{\prime}+2)=36d_{1}^{\prime}d_{2}^{\prime}+12d_{1}^{\prime}+12d_{2}^{\prime}+12.$
This is certainly greater than
$\deg(MD)=12d_{1}^{\prime}d_{2}^{\prime}+8d_{1}^{\prime}+8d_{2}^{\prime}+4,$
since $d_{1}^{\prime}$ and $d_{2}^{\prime}$ are nonnegative. So, in this case
$\deg(ID)>\deg(MD)$. If we do have $d_{1}=d_{2}=1$, then
$\deg(MD)=4=\deg(ID)$. This equality was predicted by case (1) of Conjecture
5.3.
We will now prove that Conjecture 5.3 holds in the case that $r=1$ and
$d_{i}>1$ for all $i$.
###### Theorem 5.6.
Suppose $d_{i}>1$ for all $i$. Then the only case where
$MD_{1,A_{\ell,d,k}}=ID_{1,A_{\ell,d,k}}$ is when $\ell=1$ and $d_{1}=2$.
###### Proof.
This proposition holds when $\ell=1$ by Proposition 5.2, and when $\ell=2$ and
$k_{1}=k_{2}=1$ by Example 5.5. Thus it suffices to prove that for $\ell=2$
with $(k_{1},k_{2})\neq(1,1)$, and for $\ell\geq 3$, we have
$\deg\left(MD_{1,A_{\ell,d,k}}\right)<\deg\left(ID_{1,A_{\ell,d,k}}\right)$.
First we consider how partitions of $\kappa=(1,k_{1},\ldots,k_{\ell})$ relate
to partitions of $k=(k_{1},\ldots,k_{\ell})$. Each partition of $\kappa$ gives
rise to a partition of $k$ simply by deleting the first coordinate and
grouping together vectors that are now identical. Note that no partition
$(m_{\Omega})$ contributing to $\deg(MD)$ uses the vector $(1,0,\ldots,0)$.
Also, exactly one vector in each partition of $\kappa$ is of the form
$(1,*,\ldots,*)$. Call the support of this vector $\Psi((m_{\Omega}))$, or
simply $\Psi$ when the context is clear. Note that $m_{\Psi}=1$. Let $\Xi$
denote $\Psi\setminus\\{0\\}$. Isolating $\Psi$ and $\Xi$, we may write
$\displaystyle\deg\left(MD\right)=$
$\displaystyle\sum_{(m_{\Omega})\in\mathcal{P}(\kappa)}\left(1+\sum_{\Omega\in
B}m_{\Omega}\right)!\prod_{\Omega\in
B}\frac{(d_{\Omega}-1)^{m_{\Omega}}}{m_{\Omega}!}$ $\displaystyle=$
$\displaystyle\sum_{(m_{\Omega})\in\mathcal{P}(\kappa)}\left(1+\sum_{\Omega\in
B}m_{\Omega}\right)!\cdot\frac{(d_{\Psi}-1)^{m_{\Psi}}}{m_{\Psi}!}\cdot\frac{(d_{\Xi}-1)^{m_{\Xi}}}{m_{\Xi}!}\prod_{\Omega\in
B,\Omega\neq\Psi,\Xi}\frac{(d_{\Omega}-1)^{m_{\Omega}}}{m_{\Omega}!}$
$\displaystyle=$
$\displaystyle\sum_{(m_{\Omega})\in\mathcal{P}(\kappa)}\left(1+\sum_{\Omega\in
B}m_{\Omega}\right)!\cdot(d_{\Psi}-1)\cdot\frac{(d_{\Xi}-1)^{m_{\Xi}}}{m_{\Xi}!}\prod_{\Omega\in
B,\Omega\neq\Psi,\Xi}\frac{(d_{\Omega}-1)^{m_{\Omega}}}{m_{\Omega}!}.$
Given $(m_{\Omega})$ a partition of $\kappa$, let $(n_{\Omega})$ be the
corresponding partition of $k$. So, if the term in $\deg(MD)$ coming from
$(m_{\Omega})$ is
$\left(1+\sum_{\Omega\in
B}m_{\Omega}\right)!\cdot(d_{\Psi}-1)\cdot\frac{(d_{\Xi}-1)^{m_{\Xi}}}{m_{\Xi}!}\prod_{\Omega\in
B,\Omega\neq\Psi,\Xi}\frac{(d_{\Omega}-1)^{m_{\Omega}}}{m_{\Omega}!},$
then the term in $\delta$ coming from $(n_{\Omega})$ is
$\left(1+\sum_{\Omega\in
B}m_{\Omega}\right)!\cdot\frac{(d_{\Xi}-1)^{m_{\Xi}+1}}{(m_{\Xi}+1)!}\prod_{\Omega\in
B,\Omega\neq\Psi,\Xi}\frac{(d_{\Omega}-1)^{m_{\Omega}}}{m_{\Omega}!}.$
Note that $d_{\Xi}=d_{\Psi}-1$. We know that $d_{\Xi}\neq 1$ by our assumption
that $d_{i}>1$ for all $i$, so the change in factor between these two
contributions is
$\frac{d_{\Xi}-1}{d_{\Xi}(m_{\Xi}+1)}.$
Since $d_{\Xi}>1$, this is at least $\frac{1}{2(m_{\Xi}+1)}$. In general,
$m_{\Omega}\leq\max\\{k_{i}\\}$ for any $\Omega$; since the vector
$\delta(\Psi)$ also appears in the partition of $\kappa$, we in fact have
$m_{\Xi}\leq\max\\{k_{i}\\}-1$. So, $m_{\Xi}+1\leq\max\\{k_{i}\\}$. It follows
that $\frac{1}{2(m_{\Xi}+1)}$ is greater than or equal to
$\frac{1}{2\max\\{k_{i}\\}}$. So, passing from a partition of $\kappa$ to a
partition of $k$, the corresponding term in $\delta$ is at least
$\frac{1}{2\max\\{k_{i}\\}}$ times the corresponding term in $\deg(MD)$.
Now we consider how many partitions of $\kappa$ give rise to the same
partition of $k$. Given a partition $(n_{\Omega})$ of $k$, all relevant
partitions of $\kappa$ that map to it can be constructed by choosing a single
vector used in $(n_{\Omega})$, and appending a $1$ to the $0^{th}$ coordinate.
Thus, the number of partitions of $\kappa$ mapping to $(n_{\Omega})$ is equal
to the number of distinct vectors used in $(n_{\Omega})$. The number of
distinct vectors in this partition can be bounded by $k_{1}+\ldots+k_{\ell}$,
since this is the total sum of all the entries of all the vectors used. Thus,
we have that
$\delta\geq\frac{1}{2\max\\{k_{i}\\}(k_{1}+\cdots+k_{\ell})}\deg(MD).$
It follows that
$\displaystyle\deg(ID)=2\delta(\delta-1)\geq$ $\displaystyle
2\frac{1}{2\max\\{k_{i}\\}(k_{1}+\cdots+k_{\ell})}\deg(MD)\cdot(\delta-1)$
$\displaystyle=$
$\displaystyle\frac{\delta-1}{\max\\{k_{i}\\}(k_{1}+\cdots+k_{\ell})}\cdot\deg(MD).$
To show that $\deg(ID)>\deg(MD)$, it remains to show that
$\delta-1>\max\\{k_{i}\\}(k_{1}+\cdots+k_{\ell})$.
First, rewrite
$\displaystyle\delta=$
$\displaystyle\sum_{(m_{\Omega})\in\mathcal{P}(k)}\left(1+\sum_{\Omega\in
C}m_{\Omega}\right)!\prod_{\Omega\in
C}\frac{(d_{\Omega}-1)^{m_{\Omega}}}{m_{\Omega}!}$ $\displaystyle=$
$\displaystyle\sum_{(m_{\Omega})\in\mathcal{P}(k)}\left(1+\sum_{\Omega\in
C}m_{\Omega}\right)\cdot\frac{\left(\sum_{\Omega\in
C}m_{\Omega}\right)!}{\prod_{\Omega\in C}m_{\Omega}!}\prod_{\Omega\in
C}(d_{\Omega}-1)^{m_{\Omega}}$ $\displaystyle=$
$\displaystyle\sum_{(m_{\Omega})\in\mathcal{P}(k)}\left(1+\sum_{\Omega\in
C}m_{\Omega}\right)\cdot{\sum_{\Omega\in C}m_{\Omega}\choose
m_{\Omega_{1}},\ldots,m_{\Omega_{t}}}\prod_{\Omega\in
C}(d_{\Omega}-1)^{m_{\Omega}}.$
For any partition of $k$, we have that $\sum_{\Omega\in
C}m_{\Omega}\geq\max\\{k_{i}\\}$, since $k_{i}$ vectors (counted with
multiplicity) must have nonzero $i^{th}$ coordinate. Moreover, a multinomial
coefficient ${a_{1}+a_{2}+\cdots+a_{s}\choose a_{1},a_{2},\cdots,a_{s}}$ can
be rewritten as the product ${a_{1}\choose a_{1}}{a_{1}+a_{2}\choose
a_{2}}\cdots{a_{1}+a_{2}+\cdots+a_{s}\choose a_{s}}$, so it is at least as
large as ${a_{1}+a_{2}+\cdots+a_{s}\choose a_{s}}$. Of course, we may reorder
the $a_{i}$’s in any way we desire. So, as long as some $a_{i}$ satisfies
$0<a_{i}<a_{1}+\cdots+a_{s}$, we have ${a_{1}+a_{2}+\cdots+a_{s}\choose
a_{1},a_{2},\cdots,a_{s}}\geq{a_{1}+a_{2}+\cdots+a_{s}\choose
a_{i}}\geq{a_{1}+a_{2}+\cdots+a_{s}\choose 1}=(a_{1}+\cdots+a_{s})$. This
means that if $(m_{\Omega})$ is a partition of $k$ that uses at least two
different vectors, we have ${\sum_{\Omega\in C}m_{\Omega}\choose
m_{\Omega_{1}},\ldots,m_{\Omega_{t}}}\geq\sum_{\Omega\in
C}m_{\Omega}\geq\max\\{k_{i}\\}$. Finally, the product $\prod_{\Omega\in
C}(d_{\Omega}-1)^{m_{\Omega}}$ is greater than or equal to $1$. Thus, every
partition of $k$ that uses at least distinct two vectors contributes at least
$(1+\max\\{k_{i}\\})\cdot\max\\{k_{i}\\}$ to $\delta$. We will now argue that
there are at least $\ell$ such partitions of $k$.
To do this, we split into two cases: where $\ell=2$, and where $\ell\geq 3$.
If $\ell=2$ and $(k_{1},k_{2})\neq(1,1)$, then there are indeed at least two
such partitions of $k=(k_{1},k_{2})$. For instance, we could use
$(1,1)+(k_{1}-1)(1,0)+(k_{2}-1)(0,1)$ and $k_{1}(1,0)+k_{2}(0,1)$. Both do
indeed use at least two distinct vectors since at least one of $k_{1}-1$ and
$k_{2}-1$ is nonzero.
Assume now $\ell\geq 3$. We can construct a partition of $k$ that uses at
least two vectors by choosing any $0-1$ vector with support size at least $2$
and at most $\ell-1$, and then completing the partition by using standard
basis vectors. The condition on the support size guarantees that at least one
other vector will be used, and that this new standard basis vector has not
already been used. There are $2^{\ell}-2-\ell$ such initial vectors, which is
greater than or equal to $\ell$ since $\ell\geq 3$. Thus, at least $\ell$
partitions of $k$ contribute at least
$(1+\max\\{k_{i}\\})\cdot\max\\{k_{i}\\}$ to $\delta$.
Note that $\ell\max\\{k_{1}\\}\geq k_{1}+\cdots+k_{\ell}$. It follows that
$\displaystyle\delta\geq$
$\displaystyle\ell(1+\max\\{k_{i}\\})\cdot\max\\{k_{i}\\}$ $\displaystyle\geq$
$\displaystyle\ell\max\\{k_{i}\\}\cdot\max\\{k_{i}\\}+\ell$ $\displaystyle>$
$\displaystyle\ell\max\\{k_{i}\\}\cdot\max\\{k_{i}\\}+1$ $\displaystyle\geq$
$\displaystyle\max\\{k_{i}\\}(k_{1}+\cdots+k_{\ell})+1$
Equivalently, $\delta-1>\max\\{k_{i}\\}(k_{1}+\cdots+k_{\ell})$. This implies
that $\deg(ID)>\deg(MD)$, as desired. ∎
## 6\. Curves in the plane
In this section we will determine when the mixed and iterated discriminants
associated to a planar configuration are equal. Let $A=P\cap\mathbb{Z}^{2}$,
where $P$ is a smooth lattice polygon of dimension $2$. Let $v_{A}$, $p_{A}$,
and $V_{A}$ denote the normalized area ${\rm area}_{\mathbb{Z}}(P)$ (that is,
twice its Euclidean area), the lattice perimeter (that is, the number of
points in $A$ on the edges of $P$), and the number of vertices of $P$,
respectively. It is well known [GKZ94] that in this smooth case the degree
$\delta_{A}$ of $D_{A}$ equals
$\delta_{A}=3v_{A}-2p_{A}+V_{A}.$
The degree of the mixed discriminant can be computed from Corollary 3.15 in
[CCD+13] as
$\deg(MD(A,A))=2({\rm area}_{\mathbb{Z}}(2P)-{\rm
area}_{\mathbb{Z}}(P)-p_{A})=2(4v_{A}-v_{A}-p_{A})=6v_{A}-2p_{A}.$
We can reformulate these equations in terms of the number of interior lattice
points of $P$. Let $i_{A}$ denote the number of interior lattice points of
$P$. Then we know by Pick’s Theorem that
$v_{A}=2i_{A}+p_{A}-2,$
which can be rewritten as $v_{A}-p_{A}=2i_{A}-2$. This allows us to write
$\deg(MD(A,A))=6v_{A}-2p_{A}=4v_{A}+2(v_{A}-p_{A})=4v_{A}+4i_{A}-4=4(v_{A}+i_{A}-1).$
and
$\delta_{A}=3v_{A}-2p_{A}+V_{A}=v_{A}+2(v_{A}-p_{A})+V_{A}=v_{A}+4(i_{A}-1)+V_{A}$
###### Example 6.1.
Let $A=((0,0),(2,0),(0,2))$. Let us verify that
$\deg(MD(A,A))=\deg(ID_{1,A})$, as implied by Proposition 5.2. We have
$v_{A}=4$, $i_{A}=0$, and $V_{A}=3$. This gives us
$\deg(MD(A,A))=4(v_{A}+i_{A}-1)=4(4-0-1)=12$
and
$\delta_{A}=v_{A}+4(i_{A}-1)+V_{A}=4+4(0-1)+3=3.$
This means that $\deg(ID_{1,A})=2\delta_{A}(\delta_{A}-1)=2\cdot 3\cdot
2=12=\deg(MD(A,A))$.
###### Example 6.2.
Assume $A=\text{conv}((0,0),(1,0),(0,1),(1,1))$. Let us verify that
$\deg(MD(A,A))=\deg(ID_{1,A})$, as implied by Example 2.1. We have $v_{A}=2$,
$i_{A}=0$, and $V_{A}=4$. This gives us
$\deg(MD(A,A))=4(v_{A}+i_{A}-1)=4(2-0-1)=4$
and
$\delta_{A}=v_{A}+4(i_{A}-1)+V_{A}=2+4(0-1)+4=2.$
This means that $\deg(ID_{1,A})=2\delta_{A}(\delta_{A}-1)=2\cdot 2\cdot
1=4=\deg(MD(A,A))$.
It turns out that these two examples are the only smooth polygons $P$ where
the iterated and the mixed discriminants associated to the configuration of
lattice points in $P$ coincide.
###### Theorem 6.3.
The only smooth polygons $P$ with an associated discriminant without
singularities in codimension bigger than $1$ are the known cases of the
triangle $2\Delta_{2}$ and the unit square.
###### Proof.
Assume $P$ is such a polygon, $A=P\cap\mathbb{Z}^{2}$ and
$\delta_{A}=\deg(D_{A})$. Using our formulas for $MD(A,A)$ and $\delta_{A}$,
we have that
$4(v_{A}+i_{A}-1)=2(v_{A}+4(i_{A}-1)+V_{A})(v_{A}+4(i_{A}-1)+V_{A}-1),$
which is equivalent to
$2(v_{A}+i_{A}-1)=(v_{A}+4(i_{A}-1)+V_{A})(v_{A}+4(i_{A}-1)+V_{A}-1).$
Suppose for the sake of contradiction that $i_{A}>0$. Then $v_{A}+i_{A}-1\leq
v_{A}+4(i_{A}-1)<v_{A}+4(i_{A}-1)+V_{A}$. Now, if $a,b,c,d$ are positive real
numbers with $ab=cd$, then $b<c$ implies $a>d$. This means that
$2>v_{A}+4(i_{A}-1)+V_{A}-1\geq v_{A}+V_{A}-1\geq v_{A}+2$. In other words,
$v_{A}<0$, a contradiction. Thus we know that $i_{A}=0$.
Setting $i_{A}=0$ reduces our equation to
$2(v_{A}-1)=(v_{A}+V_{A}-4)(v_{A}+V_{A}-5).$
By a classification result due to [Koe91] and presented again in [Cas12], all
convex lattice polygons with no interior lattice points are equivalent to
either the triangle $2\Delta_{2}=\text{conv}((0,0),(2,0),(0,2))$, or to a
polygon of the form $\text{conv}((0,0),(0,1),(a,0)),(b,1)$, where $a\geq b\geq
0$ and $a\geq 1$. These polygons are illustrated in Figure 2. All of these
polygons have either three or four vertices. So, we must have $V_{A}=3$ or
$V_{A}=4$.
Figure 2. All lattice polygons with no interior lattice points
If $V_{A}=3$, our equation becomes
$2(v_{A}-1)=(v_{A}-1)(v_{A}-2).$
This means that either $v_{A}=1$, or $2=v_{A}-2$; that is, $v_{A}=1$ or
$v_{A}=4$. If $v_{A}=1$, the only possibility for $P$ is the primitive lattice
triangle of normalized $1$; but this gives a degenerate system, and so is
removed from our consideration. If $v_{A}=4$, the only possibilities of $P$
are $\text{conv}((0,0),(2,0),(0,2))$ and $\text{conv}((0,0),(4,0),(0,1))$. The
second polygon is not smooth, so the only possible triangle is
$\text{conv}((0,0),(2,0),(0,2))$.
If $V_{A}=4$, our equation becomes
$2(v_{A}-1)=v_{A}(v_{A}-1).$
This means that either $v_{A}=1$ (which is impossible impossible with
$V_{A}=4$), or that $v_{A}=2$. The only polygon with $4$ vertices and area $2$
is the square $\text{conv}((0,0),(1,0),(0,1),(1,1))$
Thus we have shown that $\text{conv}((0,0),(2,0),(0,2))$ and
$\text{conv}((0,0),(1,0),(0,1),(1,1))$ are the only possibilities for $P$.
Having already verified that $\deg(MD(A,A))=2\delta_{A}(\delta_{A}-1)$ for
both these polygons, this completes the proof. ∎
## References
* [BJ14] Laurent Busé and Jean-Pierre Jouanolou. On the discriminant scheme of homogeneous polynomials. Math. Comput. Sci., 8(2):175–234, 2014.
* [BM09] Laurent Busé and Bernard Mourrain. Explicit factors of some iterated resultants and discriminants. Math. Comp., 78(265):345–386, 2009.
* [Bro71] T. J. I’A. Bromwich. Quadratic forms and their classification by means of invariant-factors. Hafner Publishing Co., New York, 1971. Reprint of the 1906 edition, Cambridge Tracts in Mathematics and Mathematical Physics, No. 3.
* [Cas12] Wouter Castryck. Moving out the edges of a lattice polygon. Discrete Comput. Geom., 47(3):496–518, 2012.
* [Cay45] A. Cayley. On the theory of linear transformations. Cambridge. Mat. J., 4:1–16, 1845.
* [CCD+13] Eduardo Cattani, María Angélica Cueto, Alicia Dickenstein, Sandra Di Rocco, and Bernd Sturmfels. Mixed discriminants. Math. Z., 274(3-4):761–778, 2013.
* [DEK14] Alicia Dickenstein, Ioannis Z. Emiris, and Anna Karasoulou. Plane mixed discriminants and toric Jacobians. In SAGA—Advances in ShApes, Geometry, and Algebra, volume 10 of Geom. Comput., pages 105–121. Springer, Cham, 2014\.
* [DFS07] Alicia Dickenstein, Eva Maria Feichtner, and Bernd Sturmfels. Tropical discriminants. J. Amer. Math. Soc., 20(4):1111–1133, 2007.
* [DR14] Sandra Di Rocco. Linear toric fibrations. In Combinatorial algebraic geometry, volume 2108 of Lecture Notes in Math., pages 119–147. Springer, Cham, 2014.
* [DRRS07] Alicia Dickenstein, J. Maurice Rojas, Korben Rusek, and Justin Shih. Extremal real algebraic geometry and $A$-discriminants. Mosc. Math. J., 7(3):425–452, 574, 2007.
* [Est10] A. Esterov. Newton polyhedra of discriminants of projections. Discrete Comput. Geom., 44(1):96–148, 2010.
* [FNO89] R. T. Farouki, C. Neff, and M. A. O’Conner. Automatic parsing of degenerate quadric-surface intersections. ACM Trans. Graph., 8(3):174–203, July 1989.
* [GHRS16] Elizabeth Gross, Heather A. Harrington, Zvi Rosen, and Bernd Sturmfels. Algebraic systems biology: a case study for the Wnt pathway. Bull. Math. Biol., 78(1):21–51, 2016.
* [GKZ94] I. M. Gelfand, M. M. Kapranov, and A. V. Zelevinsky. Discriminants, resultants, and multidimensional determinants. Mathematics: Theory & Applications. Birkhäuser Boston, Inc., Boston, MA, 1994.
* [GS] Daniel R. Grayson and Michael E. Stillman. Macaulay2, a software system for research in algebraic geometry. Available at http://www.math.uiuc.edu/Macaulay2/.
* [Jou91] J.-P. Jouanolou. Le formalisme du résultant. Adv. Math., 90(2):117–263, 1991.
* [Kat73] N. M. Katz. Pinceaux de Lefschetz: Theoreme d’existence. In Groupes de Monodromie en Géométrie Algébrique, pages 212–253, Berlin, Heidelberg, 1973. Springer Berlin Heidelberg.
* [Koe91] R. Koelman. The number of moduli of families of curves on a toric surface. PhD thesis, Katholieke Universiteit de Nijmegen, 1991.
* [Ott13] Giorgio Ottaviani. Introduction to the hyperdeterminant and to the rank of multidimensional matrices. In Commutative algebra, pages 609–638. Springer, New York, 2013\.
* [Sal82] George Salmon. A treatise on the analytic geometry of three dimensions. W. Metcalfe and Son Printers, Cambridge, 1882.
* [Sch53] Ludwig Schläfli. Gesammelte mathematische Abhandlungen. Band II. Verlag Birkhäuser, Basel, 1953.
* [WZ94] Jerzy Weyman and Andrei Zelevinsky. Multiplicative properties of projectively dual varieties. Manuscripta Math., 82(2):139–148, 1994.
* [WZ96] J. Weyman and A. Zelevinsky. Singularities of hyperdeterminants. Ann. Inst. Fourier (Grenoble), 46(3):591–644, 1996.
* [ZJT+19] Mingyang Zhao, Xiaohong Jia, Changhe Tu, Bernard Mourrain, and Wenping Wang. Enumerating the morphologies of non-degenerate Darboux cyclides. Comput. Aided Geom. Design, 75:101776, 15, 2019.
|
11institutetext: Auckland University of Technology, Auckland, New Zealand,
22institutetext: University of Tasmania, Tasmania, Australia
33institutetext: Nagoya Institute of Technology, Japan
33email<EMAIL_ADDRESS>weihua.li<EMAIL_ADDRESS><EMAIL_ADDRESS><EMAIL_ADDRESS>
# Privacy Information Classification: A Hybrid Approach
Jiaqi Wu 11 Weihua Li 11 Quan Bai 22 Takayuki Ito 33 Ahmed Moustafa 33
###### Abstract
A large amount of information has been published to online social networks
every day. Individual privacy-related information is also possibly disclosed
unconsciously by the end users. Identifying privacy-related data and
protecting the online social network users from privacy leakage turn out to be
significant. Under such a motivation, this study aims to propose and develop a
hybrid privacy classification approach to detect and classify privacy
information from OSNs. The proposed hybrid approach employs both deep learning
models and ontology-based models for privacy-related information extraction.
Extensive experiments are conducted to validate the proposed hybrid approach,
and the empirical results demonstrate its superiority in assisting online
social network users against privacy leakage.
###### Keywords:
Privacy detection Online Social Networks Deep Learning Privacy Information.
## 1 Introduction
With the proliferation and popularisation of the World Wide Web, Online Social
Networks (OSNs) become of one of the essential channels for social
interactions and communications [1, 2]. OSNs provide great convenience to the
users, but these online social platforms also raise potential risks, such as
privacy leakage. A vast amount of private information can be accessed publicly
through OSNs, such as preferences, email address, marital status, hometown,
activities attended, etc., which may lead to severe security issues.
Therefore, it is significant to explore a useful and practical approach to
protect OSN users from privacy information disclosure. Users should be
reminded before posting any privacy-related messages to the public. In the
context of OSNs, “user privacy” refers to a sequence of words, stating or
implying any individual’ s personal information, preferences, events that he
or she involved; privacy leakage describes a situation when an individual
shares stories including private information with their contacts or even those
they are not familiar with. Thus, it is necessary to develop a tool to detect
and identify all the possible privacy-related information contained in any
posting messages [20, 7]. More importantly, the justifications of privacy-
related information classification would be helpful for OSN users to get rid
of posting similar messages again.
In our previous research work [11], we conducted preliminary studies and
developed a generic framework of privacy leakage detection for OSN users based
on deep learning models. The proposed framework is capable of capturing
privacy-related entities after giving sufficient training. However, two
significant limitations are to be covered. Firstly, it can only remind users
regarding the possibility of privacy leakage. As a result, detailed leaking
information in terms of what kind of leakage is missing. Secondly, the privacy
model is not conceptually modelled or presented.
Ontology models conceptually reflect the domain-specific knowledge in the form
of terms and demonstrate two apparent advantages, i.e., shareability and
reusability [22]. Therefore, ontology-based privacy models can be easily
extended and applied to various OSNs [16]. Moreover, as ontology organises the
concepts in the form of taxonomy or hierarchy based on a pre-defined natural
relationship, it is suitable to introduce ontology into the research of
privacy-related information extraction and classification.
As an extension of our previous work [11], in this paper, we leverage a hybrid
privacy classification approach, incorporating both deep learning and ontology
models, for individual users of OSNs. The extended framework is capable of
addressing the privacy leakage problem for individuals by effectively
identifying privacy information and classifying into a detailed category. More
specifically, the proposed hybrid approach is composed of two major
components, i.e., a deep learning based approach to detect the privacy leakage
on online social data and an ontology privacy model classifying the privacy-
related information into fine-grained privacy categories. Deep learning models
are utilised to conduct the Name Entity Recognition (NER) and detect the pre-
defined privacy-related entities. An ontology model is developed based on the
analysis of massive data collected from real-world OSNs. Given the predictive
results carried out by the deep learning model, the privacy ontology model
further classifies the recognised privacy-related entities into sub-classes.
The rest of the paper is organised as follows. Section 2 reviews the existing
literature of privacy information classification approaches on OSNs. Section 3
introduces the entire privacy information detection and fine-grained
classification approach. In Section 4, experiments are conducted to evaluate
the ontology-based approach on the dataset crawled from twitter. Section 5
concludes the findings of this paper and points out the limitations and future
direction.
## 2 Related Work
Most contemporary privacy information classification approaches aim to detect
some specific categories of privacy information on OSNs or perform a binary
classification, i.e., sensitive or non-sensitive, rather than identifying and
classifying privacy information for the end users. [8] indicates that it is
easy for users to leak the matters or activities that one involves anytime
anywhere unconsciously. For example, a tweet saying that user’s family will go
out for a holiday implies no one stays at home, which may cause robbery
accidents. Therefore, the privacy information of OSNs users should be prompted
before sharing with the public.
[6] proposes a machine learning classification approach by adopting Named
Entity Recognition technology, which can classify privacy information on OSNs
into different categories, e.g., electronic devices and brands. Similarly,
[13] present a privacy classifier for three kinds of sensitive tweets, i.e.,
drunk, holiday and disease, and classifies tweets into binary results, i.e.,
sensitive tweets or nonsensitive tweets.
From the research above, we can see even there are several papers about
classifying privacy revealing information on OSNs, most of them pay attention
to specific categories by machine learning approaches. Very few studies
classify them according to a domain privacy ontology and protect individual
OSNs users from online privacy leaks. Consequently, an ontology-based
classification approach is comparatively a new research field of privacy
information classification on OSNs. Ontology-based classification approaches
show outstanding performance in many areas, e.g., online job offers [3] and
trust requirements on semantic web services [4]. Whereas, it is necessary to
build an ontology into the OSNs [10].
[5] survey the privacy ontology and point out that although some security
ontologies for fulfilling security requirements have been presented, these
studies focus on security rather than privacy [19]. They present a novel
privacy ontology to identify the key concepts and relations to satisfy the
privacy requirement. However, it aims to deal with the privacy requirements
for software engineers. Actually, there is little research about privacy
information ontology which can apply for OSNs.
Therefore, in this paper, we aim to build an ontology about privacy
information on online social media. Because almost all the privacy-related
entities are words or phrases, the semantic similarity degree calculating
plays a significant role in the proposed ontology-based privacy approach.
## 3 Automated Hybrid Privacy Detection Framework
The proposed automated hybrid privacy information detection and classification
approach are demonstrated in Fig 1. There are two key parts in the proposed
framework, i.e., the privacy-related entities recognition and further
classification based on ontology models. In the former, a deep learning model
is trained to detect four types of entities that potentially cause privacy
leakage. While, in the latter, a privacy ontology model is developed based on
the analysis of messages posted by the OSN users.
Figure 1: Automated Hybrid Privacy Detection Framework
The rationale of utilising a hybrid model of both deep learning and ontology-
based classification is clarified as follows: ontology-only approach
classifies privacy information merely based on the domain-specific vocabulary
of terms or concepts, which have to be systematically defined [17]. Deep
learning based NER does not necessarily require a lexicon or domain-specific
words, but it is difficult to control the classification of specific terms as
deep learning models turn out to be a “black-box”. By considering the factors
mentioned above, a combination of deep learning and ontology model is adopted.
The details of these two key modules are introduced in the following
subsections.
### 3.1 Deep-learning Based Privacy Information Detection
Users’ data collection turns out to be the preliminary of deep-learning based
privacy information detection. OSN users’ public data can be collected from
OSNs through web crawlers or available APIs. The collected raw data are
supposed to be filtered and pre-processed, e.g., removing messages which are
advertisements or spams, removing meaningless words and characters and parsing
word sequences to tokens.
Next, the processed data are enriched by running through the pre-annotation
process if a privacy-detection NER model is already available. Given the pre-
annotated dataset, privacy-related entities are required to be annotated
manually, and the data size of the training and testing data for the deep
learning model are nearly 20k. As mentioned previously, in the context of
OSNs, “privacy” is associated with an individual’ s personal information,
preferences, events that he or she involved. Thus, messages containing four
types of entities, i.e., “PERSON”, “TRAIT”, “PREF”, and “EVENT”, potentially
cause privacy leakage. After annotation, the annotated dataset is then fed
into the deep learning model for training.
### 3.2 Ontology-based Privacy Information Classification
As the second part of our hybrid privacy detection approach, an ontology-based
classification approach can remind individual users regarding what is to be
disclosed instead of simply giving general information.
#### 3.2.1 The Domain and Scope of The Privacy Domain Ontology
The defining domain and scope of “privacy” is the first step to build a
private-related information ontology model [17]. Naturally, the privacy
information concepts describing different subclasses (class corresponds to the
entity in this thesis) of the four privacy-related entities will be fed into
our ontology. Specifically, the privacy domain ontology on OSNs includes:
1. 1.
A hierarchical classification of privacy concepts from general classes to
specific subclasses.
2. 2.
A set of relations between privacy classes to link concepts in a more
complicated way that implied by an underlying hierarchy.
#### 3.2.2 Privacy-related Keywords Extraction
Initially, it is significant to obtain a comprehensive list of privacy-related
terms and concepts in order to form the hierarchy of privacy ontology. To
construct an ontology, we extracted all the values of privacy-related entities
recognized by the deep learning based model. Next, based on the word
frequencies, representative keywords are selected as the major indicators of
the subcategories of the entities. For example, we want to find some keywords
regarding private events under the “Event” main class. Some verbs
representative of private events, e.g., eating, shopping, etc., as well as
some nouns, e.g., concert, meeting, journey, etc., are frequently mentioned in
event-related entities. Additionally, some words are significant indicators of
privacy-related entities, which can imply the user is leaking his/her ’TRAIT’
sensitive information.
Because the creation process of an ontology is an interactive process [12], we
searched the selected keywords in the dataset to find out the occurrence of
these words and how important of them according to the term frequency [17].
Through the interactive procedure, the terms and concepts of the privacy
information ontology are finally determined. For example, because the keyword
of “interview” frequently appears in the extracted entities, it turns out to
be a keyword, representing the subclass of “Corporate Event”.
Table 1 shows the representative keywords extracted from the collected data,
which are used for building the privacy domain ontology.
Table 1: Corresponding Keywords with Classes and Subclasses Class | Subclass | Keywords
---|---|---
Person | Individual | I
| Third Party | you, we, they, he, she, classmate, uncle
Preference | Item | book, chocolate, keyboard, tea
| Hobby | cosplay, paint, fishing, dancing, reading
| Specific Person | girlfriend, teacher
Event | Private Event | eat, shopping, concert, movie, exercise, spa
| Corporate Event | wedding, interview, meeting, conference, festival, party, parade, salon
| Journey | fly, holiday, travel, island, hotel, airport
Trait | Individual Identity | years-old, Auckland
| Linked Information | lawyer, female, gay, Christian, married, white, disable
#### 3.2.3 Privacy Ontology
Among the possible approaches in developing a class hierarchy, a top-down
process has been selected by considering the relationships among the privacy
concepts in this paper [17]. The ontology hierarchy presents a tree structure,
having most general classes on top and specific associative classes connected
with the general ones. For example, given “PERSON” as a superclass,
“Individual” and “Third Party” can be the subclasses based on the recognised
entity-value pairs. Three other superclasses, i.e., “TRAIT”, “EVENT”, and
“PREFERENCE”, are also included.
TRAIT describes personally identifiable information(PII) or sensitive personal
information, which is defined and classified into two types from the usage of
the PII in United Stated legal fields, i.e., distinguish identity and relating
information [14]. Similarly, TRAITS can be classified into two subclasses as
below:
1. 1.
Any information can be used to distinguish an individuals identity, such as
birth date and hometown.
2. 2.
Any information can be linked to an individual, such as medical, educational,
marital status and employment information.
Through this kind of classification approach, two subclasses of Trait are
identified: Individual identity and Linked information. For example, the date
of birth can be recognised as an individuals identity. Whereas, race, gender,
sexual orientation, marital status, religion, belief, and education background
are categorized as linked information. Similarly, EVENT can be classified as a
private event, corporate event, and journey in terms of the event is social or
non-social. Among the subclasses of “Event”, tweets about the journey is
individually classified because we think users who reveal their journey plans
will make them very vulnerable to theft crimes. Preference can be classified
as a specific person, item and hobby according to the characteristic of the
leaking hobby information. Therefore, the privacy domain ontology for OSNs is
presented in Fig. 2:
Figure 2: Privacy Ontology
#### 3.2.4 Semantic Phrase Similarity Degree Based On GloVe
To classify the privacy-related entities into fine-grained subclasses, it is
essential to find an approach in finding the subclass where the entities have
the highest belonging degree. Moreover, the highest belonging degree is
decided by the highest similarity degree between the extracted entities and
the representative terms. Comparing the similarity of word embedding of each
word can decide their similarity degree [9]. Word embedding is defined as
semantic vector space models that use vectors to represent each word.
Consequently, in this subsection, we will propose a semantic phrase similarity
degree approach based on Glove. It is divided into two steps. Firstly, a word
semantics vector space model is decided. Secondly, the construction of the
classification model will be described by taking “Event” entity as an example.
$\bullet$ Word Semantics Vector Space Model
In the first step, the main step is to choose a word semantics vector space
model. A pre-trained statistical model (called “en_core_web_lg” in spacy) is
used in this paper, which is trained on blogs, news, and comments with GloVe.
Global Vectors for Word Representation (GloVe) is a state-of-art tool using
word embedding techniques. Most word embedding approaches like Word2Vec exist
a disadvantage, which is the lack of co-occurrence between words. Luckily, the
GloVe approach trains global word-word co-occurrence counts which fills this
gap, outperforming than other current word embedding approaches in common word
similarity tasks[18] [15]. That is the reason we choose this word semantics
vector space model for us to use in order to compare the similarity degree
between words.
$\bullet$ The Construction of The Classification Model
To explain the construction of the classification model, we make use of
“Event” entities as an example. In Fig. 2, the entity consists of three
subclasses, “Private Event”, “Corporate Event” and “Journey”. Moreover, in
Table 1, there are 20 representative terms representing the subclasses
associated with them. Each representative term can be represented as
$term_{i}$, where $i$ belongs to {0,1,2,…,19}. Among them, ’Private
Event’:{0,1,2,…,5}, ’Corporate Event’:{6,7…,13}, and ’Journey’:{14,15…,19}.
Similarly, each word in the extracted entities can be represented as
$entity_{j}$.
Consequently, $S_{ij}$ can represent the semantic similarity degree between
each word in the extracted entities and each representative term:
$S_{ij}=similarity(entity_{j},term_{i})$ (1)
So the similarity of event-related entities and the event representative terms
in the ontology can be shown as follows:
$S_{i(sum)}=\sum_{j=1}^{n}S_{ij}$ (2)
After the calculation of all the similarity degree, then the subclass of the
extracted event entities can be decided according to the maximum degree of
$S_{i(sum)}$, which means the subclass which obtains the maximum $S_{i(sum)}$
is the corresponding subclass of the event-related entity:
$\lambda=\max\left(S_{0}{(sum)},S_{1}{(sum)},...,S_{19}{(sum)}\right)$ (3)
Then the $i$ which obtains the maximum degree of $S_{i(sum)}$ can be decided
and the corresponding subclass can be decided as below:
$subclass=\left\\{\begin{array}[]{ll}PrivateEvent&\textrm{$i\in{0,1,...,5}$}\\\
CorporateEvent&\textrm{$i\in{6,7,...,13}$}\\\
Journey&\textrm{$i\in{14,15,...,19}$}\end{array}\right.$
Hence, by using the semantic information which the word embedding technique
capture [21], we can classify privacy-related information into the
corresponding subclass in the privacy ontology like the “Event” procedure we
list.
## 4 Experiments
Experiments have been conducted to evaluate the proposed hybrid privacy
detection approach. The experiment uses a real-world testing dataset that aims
to classify fine-grained privacy-related entities. Moreover, we also focus on
one interesting problem around private tweets: What type of personal private
information is leaked most on OSNs?
### 4.1 Data Description
Twitter 111https://twitter.com/ is one of the largest Online Social Media
platforms as a micro-blogging service, where a large amount of information is
broadcast publicly by individual users. In Twitter, the information posted by
end users is named as tweets. Twitter provides APIs, allowing developers to
search and store tweets based on certain criteria. Therefore, we search and
collect 18k tweets from Twitter through API. Most tweets are selected by
searching for keywords related to sensitive activities and plans.
Consequently, we use Twitter API to search for some terms which contain
sensitive keywords, e.g., sensitive activities and plans, which may result in
privacy information leakage and cause negative consequences. Then we collect
around 18k tweets as a testing dataset in this experiment. As we demonstrated
before, the tweets will be conducted with the deep learning based NER approach
and the ontology-based classification approach in this chapter, then the
corresponding subclasses will be decided.
### 4.2 Evaluation
To evaluate the performance of the hybrid privacy detection approach, we
manually annotate the subclasses of privacy leaking information. Moreover, the
”ground truth” is prepared by allowing the users themselves to provide
opinions on whether they leak the privacy and what types of private
information they are leaking on tweets. For example, a tweet “I watch a
movie.”, has “I” annotated as “Individual” and “watch a movie” annotated as
“Private Event”.
Four traditional measures to utilised to evaluate the performance of the
proposed hybrid approach:
1. 1.
Accuracy: the fraction of correct classification in the testing data set;
2. 2.
Precision: the fraction of correct classification among all results are
classified in this subclass in the testing data set;
3. 3.
Recall: the fraction of correct classification among all actual results belong
to this subclass in the testing data set;
4. 4.
F1-value: the harmonic value of precision and recall, which is a balance
measurement.
### 4.3 Experimental Results
Our hybrid privacy classification approach utilizes a deep learning based NER
approach and an ontology-based approach to perform classification of specific
privacy information. After the NER, we use the ontology vocabulary for
performing semantic phrase similarity degree calculation with the extracted
privacy-related entities.
We evaluate our hybrid approach on the testing dataset and get a considerable
performance, as shown in Table 2, with high accuracy in each type. We believe
this accuracy is high enough to demonstrate the effectiveness of our automated
detection and classification approach. However, we observe the accuracy of
categories under “TRAIT” entity is much lower than other types of entities.
The extraction result of “TRAIT” entity is lower than other types of entity
and a trait-related privacy entity will be classified to categories under
other types of entities. That is why the accuracy value of classification of
“TRAIT” entity is lower than other entities.
Table 2: Performance of Hybrid Privacy Information Classification Class | Subclass | Accuracy
---|---|---
Person | Individual | 0.94
| Third Party | 0.85
Preference | Item | 0.82
| Hobby | 0.81
| Specific Person | 0.76
Event | Private Event | 0.74
| Corporate Event | 0.76
| Journey | 0.77
Trait | Individual Identity | 0.62
| Linked Information | 0.64
### 4.4 Privacy Leaking Information Categorization
After the whole automated hybrid privacy information detection approach, all
the types of private information leaks on the 18k testing dataset can be
extracted. Then we plot the distribution of the results of the privacy
information types of the testing dataset in Fig. 3, which includes all the
subclasses under the “TRAIT”, “PREF”, and “EVENT” entities. According to the
privacy rule in this detection approach, each tweet with privacy-related
entities which is identified as the private tweet contains the “Person”
entity, so the distribution of the “Person” entity is not necessary to be
analyzed its categorization.
Figure 3: Distribution of Different Types of Privacy Information Leaking
In this section, we also explore the question of what type of sensitive
information users leak most. We counted the percentage of the eight types of
privacy information in the 18k testing dataset. The results are shown in Fig.
3. From Fig. 3, we can see the most leaking information is the “Event” entity,
where “Private Event” counts the most in it. Additionally, “Journey” under the
“Event” and “Specific person” under the “Preference” are also leaked a lot in
tweets, which means the information about all subclasses under the entity of
“Event” both is leaked a lot on OSNs. From the results, we suggest that OSNs
users should exercise a little more restraint about posting relevant tweets
about these types. On the other hand, people are more conservative about
“Individual Identity” and “Hobby” information because the privacy leaks (as
the percentage) of the two types is much smaller than other types.
## 5 Conclusions and Future Works
In this paper, we propose a hybrid approach for classifying private
information generated by OSN users. Through characterising the nature of
privacy information leaking on OSNs, a deep learning model has been employed
for privacy-related entities recognition, and an ontology-based classification
approach is conducted to automatically classify fine-grained privacy
information, i.e., nine subtypes of private leaking. The ontology-based
approach calculated the semantic similarity between entities extracted from
the deep learning model and the representative terms. We evaluated the result
with the accuracy value, which demonstrates it gains a considerable
performance. Moreover, what specific types of personal private information
users are leaking on OSNs can be understood.
This research can be extended by investigating the following directions.
Firstly, we can recognise privacy-related entities on tweets other than on all
tweets by one user. In the future, all privacy information revealing by a user
can be collected by the model and protect the user. Secondly, different tweets
are associated with different degrees of privacy leakage. In this paper, we
demonstrate what types of privacy-related entities OSNs users reveal. However,
we do not analyse the detailed privacy leakage degree of users. In the future,
we plan to evaluate the privacy leakage degree and explore the insights based
on predictive results.
## References
* [1] Batra, A., Sidhu, K., Sharma, S.: Characteristics of women whatsapp users and use pattern. Journal of Education, Society and Behavioural Science pp. 1–7 (2018)
* [2] Cormode, G., Krishnamurthy, B.: Key differences between web 1.0 and web 2.0. First Monday 13(6) (2008)
* [3] ul haq Dar, E., Dorn, J.: Ontology based classification system for online job offers. In: 2018 International Conference on Computing, Mathematics and Engineering Technologies (iCoMET). pp. 1–8. IEEE (2018)
* [4] Galizia, S.: Wsto: a classification-based ontology for managing trust in semantic web services. In: European Semantic Web Conference. pp. 697–711. Springer (2006)
* [5] Gharib, M., Giorgini, P., Mylopoulos, J.: Towards an ontology for privacy requirements via a systematic literature review. In: International Conference on Conceptual Modeling. pp. 193–208. Springer (2017)
* [6] Gomez-Hidalgo, J.M., Martin-Abreu, J.M., Nieves, J., Santos, I., Brezo, F., Bringas, P.G.: Data leak prevention through named entity recognition. In: Social Computing (SocialCom), 2010 IEEE Second International Conference on. pp. 1129–1134. IEEE (2010)
* [7] Hasan, O., Habegger, B., Brunie, L., Bennani, N., Damiani, E.: A discussion of privacy challenges in user profiling with big data techniques: The eexcess use case. In: Big Data (BigData Congress), 2013 IEEE International Congress on. pp. 25–30. IEEE (2013)
* [8] Humphreys, L., Gill, P., Krishnamurthy, B.: How much is too much? privacy issues on twitter. In: Conference of International Communication Association, Singapore. Citeseer (2010)
* [9] Jiang, J.J., Conrath, D.W.: Semantic similarity based on corpus statistics and lexical taxonomy. arXiv preprint cmp-lg/9709008 (1997)
* [10] Kim, A., Hoffman, L.J., Martin, C.D.: Building privacy into the semantic web: An ontology needed now. In: Proc. of Semantic Web Workshop, Hawaii, USA (2002)
* [11] Li, W., Wu, J., Bai, Q.: An automated privacy information detection approach for protecting individual online social network users. In: Proceedings of the 33th Annual Conference of Japanese Society for Artificial Intelligence (JSAI), Japanese Society for Artificial Intelligence (JSAI) (2019)
* [12] Madsen, M., et al.: A health information privacy ontology: toward decision support for compliance assessment. HIC 2006 and HINZ 2006: Proceedings p. 184 (2006)
* [13] Mao, H., Shuai, X., Kapadia, A.: Loose tweets: an analysis of privacy leaks on twitter. In: Proceedings of the 10th annual ACM workshop on Privacy in the electronic society. pp. 1–12. ACM (2011)
* [14] McCallister, E., Grance, T., Scarfone, K.A.: Sp 800-122. guide to protecting the confidentiality of personally identifiable information (pii) (2010)
* [15] Mihalcea, R., Corley, C., Strapparava, C., et al.: Corpus-based and knowledge-based measures of text semantic similarity. In: AAAI. vol. 6, pp. 775–780 (2006)
* [16] Mitra, P., Liu, P., Pan, C.C.: Privacy-preserving ontology matching. In: AAAI Workshop on Context and Ontologies (2005)
* [17] Noy, N.F., McGuinness, D.L., et al.: Ontology development 101: A guide to creating your first ontology (2001)
* [18] Pennington, J., Socher, R., Manning, C.: Glove: Global vectors for word representation. In: Proceedings of the 2014 conference on empirical methods in natural language processing (EMNLP). pp. 1532–1543 (2014)
* [19] Souag, A., Salinesi, C., Comyn-Wattiau, I.: Ontologies for security requirements: A literature survey and classification. In: International Conference on Advanced Information Systems Engineering. pp. 61–69. Springer (2012)
* [20] Wang, Y., Norcie, G., Komanduri, S., Acquisti, A., Leon, P.G., Cranor, L.F.: I regretted the minute i pressed share: A qualitative study of regrets on facebook. In: Proceedings of the seventh symposium on usable privacy and security. p. 10. ACM (2011)
* [21] Yaguinuma, C.A., Ferraz, V.R.T., Santos, M.T.P., Camargo, H.A., Nogueira, T.M.: A model for representing vague linguistic terms and fuzzy rules for classification in ontologies. In: ICEIS (2). pp. 438–442 (2010)
* [22] Zhao, Y., Dong, J., Peng, T.: Ontology classification for semantic-web-based software engineering. IEEE Transactions on Services Computing 2(4), 303–317 (2009)
|
# Mining Large-Scale Low-Resource Pronunciation Data From Wikipedia
Tania Chakraborty, Manasa Prasad, Theresa Breiner, Sandy Ritchie, Daan van
Esch
Google Research
{taniarini, pbmanasa, tbreiner, sandyritchie<EMAIL_ADDRESS>
(June 2020)
###### Abstract
Pronunciation modeling is a key task for building speech technology in new
languages, and while solid grapheme-to-phoneme (G2P) mapping systems exist,
language coverage can stand to be improved. The information needed to build
G2P models for many more languages can easily be found on Wikipedia, but
unfortunately, it is stored in disparate formats. We report on a system we
built to mine a pronunciation data set in 819 languages from loosely
structured tables within Wikipedia. The data includes phoneme inventories, and
for 63 low-resource languages, also includes the grapheme-to-phoneme (G2P)
mapping. 54 of these languages do not have easily findable G2P mappings online
otherwise. We turned the information from Wikipedia into a structured,
machine-readable TSV format, and make the resulting data set publicly
available so it can be improved further and used in a variety of applications
involving low-resource languages.
## 1 Introduction
There are thousands of languages spoken around the world, and many efforts to
learn about them and document them. However, information about low-resource
languages is not always easy to find, or leveraged to its full potential.
Wikipedia is a useful resource for language-specific information, with a
community of native speakers and linguistic experts who continue to improve
coverage and quality111See the LingWiki effort:
https://en.wikipedia.org/wiki/Wikipedia:GLAM/SOAS/Lingwiki. While Wikipedia is
often used to build monolingual or parallel text corpora [Prasad et al., 2018,
Rahma et al., 2018], its articles on specific languages contain useful
information that is not frequently extracted.
In this paper we describe how we extracted pronunciation information from
these Wikipedia pages for hundreds of languages, some of which are not
represented in other sources, which can be used in technologies for low-
resource languages. The data was mined from tables containing information such
as phonemes, associated graphemes, and sample words in each language. While
the tables are understandable to the human eye, it is nontrivial to
automatically parse and generate standardized TSV files from them. We publish
our mined data and encourage others to leverage and improve this source.
## 2 Uses for Pronunciation Data
While large-scale pronunciation data can be useful in many linguistic
pursuits, grapheme-to-phoneme (G2P) mappings are also a building block for
automatic language processing systems. Tools to generate pronunciations from
input words [Mortensen et al., 2018, Novak et al., 2016] typically support a
specific set of languages but can often be extended using G2P rules for new
languages. G2P data can also be used in cross-lingual transfer learning, e.g.
in named entity recognition [Bharadwaj et al., 2016].
G2P mappings are also required in typical automatic speech recognition (ASR)
and text-to-speech (TTS) systems. While these systems usually train on paired
audio and transcription data, several recent areas of research leverage G2P
mappings to overcome this requirement. If there is a G2P mapping and some text
available in a low-resource language, an ASR system can be built by
repurposing an acoustic model from a similar higher-resource language [Prasad
et al., 2019]. G2P mappings can even be bootstrapped for new languages using
only the language’s phoneme inventory combined with higher-resource language
text written in the same script [Bleyan et al., 2019], making even simple
phoneme data a more useful tool for scaling technologies to new languages.
## 3 Similar Resources
One similar resource in this space is PHOIBLE [Moran et al., 2014], a database
listing the phonemes of over 2,100 languages, their possible allophones, and
how common each phoneme is in the language, but not any G2P information.
Wikimedia’s Wiktionary and
Incubator222https://meta.wikimedia.org/wiki/Wiktionary#List_of_Wiktionaries
are open-content dictionaries written in about 500 languages, and can be used
to automatically extract phoneme and pronunciation information [Schlippe et
al., 2010, Lee et al., 2020]. Deri and Knight [Deri and Knight, 2016] as well
as Peters et al. [Peters et al., 2017] have been able to adapt this data even
to build G2P models for new languages that do not appear in the mined data.
Deri and Knight also published the G2P mappings that they mined from Wikipedia
IPA Help
tables333https://en.wikipedia.org/wiki/Category:International_Phonetic_Alphabet_help
for 98 languages. Our data set, which includes phoneme inventories for over
800 languages, offers G2P data in 63 low-resource languages, 54 of which are
not covered by Deri and Knight. There is also more potential G2P information
stored for an additional 224 languages, as we detail in section 4.3. We hope
that this additional data set will be a useful supplement to the existing
resources.
## 4 Mining Pronunciation Data from Wikipedia
The English Wikipedia contains hundreds of articles about specific
languages444https://en.wikipedia.org/wiki/Index_of_language_articles, often
including details on the phonology and writing systems, and covers languages
that are not represented in other data sources. We wanted to aggregate this
data into an easily comparable and processable data set to make it more widely
usable for the research that we described in section 2.
### 4.1 Extracting the Wikipedia Pages
Wikipedia’s category structure enabled us to target relevant pages inter-
linked from the category pages for Languages by
Country555http://en.wikipedia.org/wiki/Category:Languages_by_country and
Language
Phonologies666http://en.wikipedia.org/wiki/Category:Language_phonologies. We
also checked for other potential language pages by leveraging the Wikipedia
search query for ISO639 codes777https://iso639-3.sil.org/; for example, the
page on Amarasi (ISO 639 code ‘aaz’) can be found at
https://en.wikipedia.org/wiki/ISO_639:aaz. Using tooling similar to
wikitable2csv888https://github.com/gambolputty/wikitable2csv, we extracted the
table data embedded in each targeted page.
### 4.2 Parsing the Tables
From the tables on each language’s page, we wanted to parse: a) the phoneme
inventory of the language b) the pronunciation features of the phonemes,
including voicing, place/manner of articulation, and other phonetic features,
useful for clarity and in case of nonstandard transcriptions c) grapheme to
phoneme mappings d) words in the native script and e) the phonemic
transcriptions of the words.
While some or all of this data is present in most of the language pages,
different pages format their tables in different ways, which makes data
extraction and processing tricky. We aim to cover as many edge cases as
possible and successfully parse a majority of the table formats. Some
especially odd cases and ones that we left unaddressed can be found in section
4.3.
We mainly saw two different ways that data in a table are presented, which we
will call Type A and Type B. Type A has one set of headers, where the contents
are related to each other horizontally (or vertically if headers are on the
side). For example, Figure 1(a)999Screenshots of pages accessed on June 24,
2020: https://en.wikipedia.org/wiki/Amarasi_language and
https://en.wikipedia.org/wiki/Alekano_language is a Type A table where the
data in the column titled “Amarasi Alphabet” are graphemes, and the data in
the column titled IPA are the corresponding phonemes. Type B has two sets of
headers, in which case the data between columns is not related, but rather
classified in a similar way. For example, in Figure 1(b), the phonemes in the
columns are not related to each other; the data in each cell is classified by
two or more pronunciation features, namely place of articulation (column
headers) and manner of articulation/voicing (row headers).
(a) Amarasi (iso:aaz) grapheme-phoneme table
(b) Alekano (iso:gah) consonant phoneme table
Figure 1: Example Wikipedia tables with relatively simple formats.
Since there is no obvious indicator of a table’s type from its representation,
we have to automatically determine how to parse the data by observing the
structure of the table; i.e., how many headers the table has, if the table has
both column and row headers, etc. We determine that a given row or column is a
header if it has consecutive ‘header’ type cells, which are annotated with a
‘!’ symbol101010According to the Wikipedia Table Help page:
https://en.wikipedia.org/wiki/Help:Table. We treat tables with headers only on
the top or the side as Type A, while tables with both are Type B.
Category | Targeted Keywords
---|---
Grapheme | letter, grapheme, alphabet, written
Phoneme | IPA, pronunciation
Pronunciation Feature | description, vowel/consonant [found in table caption rather than header]
Example word | example, word
Transcription | transcription
Unclassified | [any unmatchable data]
Table 1: Keywords in table headers that helped us determine the category of
data in the table.
The data extracted from the tables is classified into one of the categories
listed in Table 1. The correct category for each data item is based on the
type of the table and the text in the header corresponding to the item’s
column. For Type A tables, a regular expression is used to determine the
category, based on keywords as you can see in Table 1. If the header text
contains any variation of the keywords we are searching for, we can classify
that data into the matching category.
For Type B tables, the data is classified as a pronunciation feature, with the
cell value representing the associated phoneme and the actual pronunciation
features extracted from the row and column headers.
### 4.3 Challenges
As we described in Section 4, the primary challenge in mining the data was the
variety of the table structures. We tried to account for the most common types
of tables, including the simpler Type A and Type B tables described above, as
well as tables with repeated headers, as seen in Figure 2111111Screenshot of
page accessed on June 24, 2020: https://en.wikipedia.org/wiki/Wakhi_language.
If repetition is not accounted for, the contents of an entire column would be
incorrectly associated with each other. Instead, we need to associate only two
adjacent rows with each other as we parse the table. The same principle
applies to tables that have a repetitions every N rows or every N columns for
any value of N. Some more table examples can be found in Appendix A.
Figure 2: Wakhi (iso:wbl) data table with repeated headers across rows
Another challenge is that the keywords in Table 1 may be insufficient to
classify data correctly for all cases. For example, if a table uses the name
of the language as the header for the graphemes in a G2P table instead of a
more generic keyword, we will not parse the data correctly. As we developed
the parser, we added in new keywords to cover more cases, but do not cover all
edge cases. When we find a table whose header is not a keyword, we store the
data item and its associated header together in the Unclassified category
rather than discarding it. Some examples can be found in Appendix A.
If a keyword is used in a way we do not expect, or if there is a very
intricate table format, there can be other problems. Some tables list multiple
data within a cell, and it is not clear what the relationship is in order to
split the data or classify it correctly. Some examples of these tables can be
found in Appendix B. Our dataset may include some of this noise, although we
hope that in most cases, tables that confused our system would store the data
in the Unclassified category to at least be human readable.
## 5 Our Data Set
We found phoneme data for 819 languages, and G2P mappings for 63 languages. 54
of these do not appear in Deri and Knight’s G2P data for 98 languages mined
from the Wikipedia IPA Help Tables
121212https://drive.google.com/drive/u/1/folders/0B7R_gATfZJ2aWkpSWHpXUklWUmM
[Deri and Knight, 2016]. An additional 224 languages had some data we could
only parse as Unclassified, but further analysis hints that 72 of them may
contain G2P information. Since the Wikipedia pages for higher-resource
languages do not often contain G2P tables (but tend to focus on grammar or
history), our data set does not include G2P for these languages, which are
more likely covered by Deri and Knight, and instead represents more low-
resource languages.
We publish our data 131313https://github.com/google/language-
resources/tree/master/mined-wiki-phoneme-tables in TSV format, where the
columns correspond to: grapheme, phoneme, pronunciation features, example
word, IPA transcription of the example word, and Unclassified data. Phonemes
can appear in more than one row if, for example, they are mapped to more than
one grapheme or example word. If a field is empty, it is filled in with
‘(n/a)’. See Figure 3 for an example excerpt, with more examples available in
the appendix. We also include a txt with all Wikipedia links that were used.
Figure 3: Excerpt of Amarasi (iso:aaz) TSV data file, containing g2p mappings
## 6 Conclusion
We mined a data set covering 819 languages’ phonemes, including 63 languages’
grapheme-to-phoneme mappings, 54 of which are low-resource languages that do
not have easily findable G2P mappings in other available sources. This
information is available on Wikipedia, but inconsistencies with how it is
formatted make it quite challenging to use in extending language technology.
While our extraction system might be improved to handle more edge cases in
future work, it may in fact be even more fruitful to collaborate with relevant
Wikipedia editors to standardize their contributions to make the data more
machine-readable. At any rate, we hope this work helps the research community
continue to scale research and technologies to more and more languages.
## Appendix A Additional Examples of Table Formats We Handle
This appendix gives some further examples of table formats that are nontrivial
to parse in order to associate and classify the data correctly.
Figure 4: Ocaina (iso:oca) data table with repeated headers across columns
In Figure 4 141414Screenshot of page accessed on June 24, 2020:
https://en.wikipedia.org/wiki/Ocaina_language, we see that tables may have not
only repeated row headers (as seen in Figure 2) but may have repeated column
headers. We handle this case.
Figure 5: Albanian (iso:sq) data table with multiple layers of headers
In Figure 5 151515Screenshot of page accessed on June 24, 2020:
https://en.wikipedia.org/wiki/Albanian_language every element must be
connected with multiple columns and rows of headers to ensure that we do not
miss any information. We handle this case.
(a) Adyghe (iso:ady) data table with multiple orthographies
(b) Adyghe (iso:ady) extracted data example
Figure 6: Adyghe Language extraction example
Figure 6(a) 161616Screenshot of page accessed on June 24, 2020:
https://en.wikipedia.org/wiki/Adyghe_language combines the G2P mappings for
two scripts used by the language, with headers that are too specific compared
to our keywords for the correct category (grapheme). We handle this case by
storing the data and its associated header in the Unclassified category, which
is labeled as “Best Guess” in Figure 6(b), and allows the user to see some
more information for the language that wasn’t straightforward to parse
automatically.
## Appendix B Additional Examples of Table Formats We Don’t Handle Well
(a) E (iso:eee) data table where keyword is
used unexpectedly
(b) Madurese (iso:mad) data table with multiple items per cell
In Figure 7(a) 171717Screenshot of page accessed on June 25, 2020:
https://en.wikipedia.org/wiki/E_language, a table has a header containing one
of the keywords, “letter”, but not in the way that we generally expect, which
would be in a table that shows G2P. We do not handle this case and this data
is parsed into the data set mistakenly.
In Figure 7(b) 181818Screenshot of page accessed on June 24, 2020:
https://en.wikipedia.org/wiki/Madurese_language, there are multiple data items
within each cell, and there is no clear way to understand automatically how to
split up or classify these items. We do not handle this case.
## Appendix C Additional Example Excerpts from TSV Data Files
Figure 8 is an snippet from the oca.tsv file, and shows how G2P mappings could
be captured in the Unclassified column. Figure 9 is another example of the
type of data that could be captured by the Unclassified column. This language
did not have corresponding phonemes in the table, instead it had words from
the language and their English meaning.
Figure 8: Ocaina (iso:oca) tsv format Figure 9: Krio (iso:kri) tsv format
## References
* [Bharadwaj et al., 2016] Akash Bharadwaj, David Mortensen, Chris Dyer, and Jaime Carbonell. 2016\. Phonologically aware neural model for named entity recognition in low resource transfer settings. In Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing, pages 1462–1472, Austin, Texas, November. Association for Computational Linguistics.
* [Bleyan et al., 2019] Harry Bleyan, Sandy Ritchie, Jonas Fromseier Mortensen, and Daan van Esch. 2019\. Developing pronunciation models in new languages faster by exploiting common grapheme-to-phoneme correspondences across languages. In Proceedings of Interspeech 2019.
* [Deri and Knight, 2016] Aliya Deri and Kevin Knight. 2016\. Grapheme-to-phoneme models for (almost) any language. In Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 399–408, Berlin, Germany, August. Association for Computational Linguistics.
* [Lee et al., 2020] Jackson L. Lee, Lucas F.E. Ashby, M. Elizabeth Garza, Yeonju Lee-Sikka, Sean Miller, Alan Wong, Arya D. McCarthy, and Kyle Gorman. 2020\. Massively multilingual pronunciation modeling with WikiPron. In Proceedings of The 12th Language Resources and Evaluation Conference, pages 4223–4228, Marseille, France, 05. European Language Resources Association.
* [Moran et al., 2014] Steven Moran, Daniel McCloy, and Richard Wright, editors. 2014\. PHOIBLE Online. Max Planck Institute for Evolutionary Anthropology, Leipzig.
* [Mortensen et al., 2018] David R. Mortensen, Siddharth Dalmia, and Patrick Littell. 2018\. Epitran: Precision G2P for many languages. In Proceedings of the Eleventh International Conference on Language Resources and Evaluation (LREC 2018), Miyazaki, Japan, May. European Language Resources Association (ELRA).
* [Novak et al., 2016] Josef Robert Novak, Nobuaki Minematsu, and Keikichi Hirose. 2016\. Phonetisaurus: Exploring grapheme-to-phoneme conversion with joint n-gram models in the wfst framework. Natural Language Engineering, 22(6):907–938.
* [Peters et al., 2017] Ben Peters, Jon Dehdari, and Josef van Genabith. 2017\. Massively multilingual neural grapheme-to-phoneme conversion. In Proceedings of the First Workshop on Building Linguistically Generalizable NLP Systems, pages 19–26, Copenhagen, Denmark, 09. Association for Computational Linguistics.
* [Prasad et al., 2018] Manasa Prasad, Theresa Breiner, and Daan van Esch. 2018\. Mining training data for language modeling across the world’s languages. In Proceedings of the 6th International Workshop on Spoken Language Technologies for Under-resourced Languages (SLTU 2018).
* [Prasad et al., 2019] Manasa Prasad, Daan van Esch, Sandy Ritchie, and Jonas Fromseier Mortensen. 2019\. Building large-vocabulary asr systems for languages without any audio training data. In Proceedings of Interspeech 2019.
* [Rahma et al., 2018] Sellami Rahma, Fatiha Sadat, and Lamia Belguith, 2018. Building and Exploiting Domain-Specific Comparable Corpora for Statistical Machine Translation, pages 659–676. 01\.
* [Schlippe et al., 2010] Tim Schlippe, Sebastian Ochs, and Tanja Schultz. 2010\. Wiktionary as a source for automatic pronunciation extraction. In Proceedings of Interspeech 2010.
|
capbtabboxtable[][]
Modeling of Nonlinear Interference Power for Dual-Polarization 4D Formats
Gabriele Liga1, Bin Chen1,2, Astrid Barreiro1, and Alex Alvarado1
1Department of Electrical Engineering, Eindhoven University of Technology,
Eindhoven, The Netherlands
2School of Computer Science and Information Engineering, Hefei University of
Technology, Hefei, China
<EMAIL_ADDRESS>
###### Abstract
We assess the accuracy of a recently introduced nonlinear interference model
for general dual-polarization 4D formats. Unlike previous models for
polarization-multiplexed 2D formats, an average gap from split-step Fourier
simulations within 0.1 dB is demonstrated.
## 1 Introduction
Nonlinear interference (NLI) modeling in optical fiber transmission is a key
tool to analyze and optimize the performance of optical communication systems.
In recent years, remarkable progress has been made in this direction, and as
new model’s applications emerge, further work is required to meet new accuracy
and computational requirements. Constellation shaping is today a popular area
of application for NLI models for two reasons: i) the performance of a given
constellation significantly depends on the amount of NLI power induced during
propagation; ii) NLI models can provide easy-to-compute as well as accurate
cost functions for the performance optimization. However, as research focus
moves towards multidimensional constellation shaping, where the different
dimensions are mapped onto the degrees of freedom of the fiber channel
(quadratures, polarization, wavelengths, etc), extensions of the established
NLI models for two-dimensional (2D) formats are needed.
NLI power models such as the enhanced Gaussian noise (EGN) model [1]
introduced analitical expressions explicitly linking the properties of the
transmitted constellation to the resulting NLI power after propagation.
Nonetheless, such expressions have been developed only for so-called
polarization-multiplexed 2D (PM-2D) formats, i.e. when a single 2D
constellation is used to independently map information over 2 orthogonal
polarization modes of the optical field. The resulting dual polarization four-
dimensional (DP-4D) constellation is, thus, given by the Cartesian product of
the component 2D constellation by itself. However, the class of PM-2D formats
represents only a limited subset of all possible DP-4D constellations. A vast
literature on assessing the performance of DP-4D formats that do not fall
within the PM-2D class is available, (see, e.g., [2]), and, lately, DP-4D
constellations have shown improved shaping gains compared to other
conventional PM-2D formats [3, 4, 5]. However, no NLI model has been so far
available to support the design of nonlinearity-tolerant DP-4D formats.
Recently, in [6], we extended the model in [1] to account for the entire DP-4D
class of modulation formats. In this contribution, we present a first
numerical validation of our 4D model. Moreover, we show that heuristic
extensions of the EGN model to non PM-2D formats may lead to inaccuracies in
the prediction of the NLI power coefficient beyond 1 dB, even for fairly
regular 4D modulation formats. Our 4D model [6] is instead proven to be
accurate on average within 0.1 dB for all 4D modulation formats studied in
this work.
## 2 Analytical Formulation of the NLI coefficient for DP-4D formats
The model we aim to validate in this work consists of an analytical formula
for the computation of the vector $(\sigma^{2}_{x},\sigma^{2}_{y})$, where
$\sigma^{2}_{x}$ and $\sigma^{2}_{y}$ represent the NLI power within the
received signal bandwidth over the $x$\- and $y$ polarizations, respectively.
The corresponding NLI power coefficient vector
$(\eta_{x},\eta_{y})\triangleq(\sigma^{2}_{x},\sigma^{2}_{y})/P^{3}$, where
$P$ denotes the transmitted signal power. The model’s expression was derived
using a first-order perturbational approach under the hypothesis of single-
channel transmission with quasi-rectangular pulse spectrum. A comprehensive
discussion on the assumptions of the model as well as its mathematical
derivation can be found in [6].
Based on [6, eqs. (42)-(43)], the NLI power coefficient for the $x$
polarization can be found as
$\displaystyle\begin{split}\eta_{x}&=\left(\frac{8}{9}\right)^{2}\frac{\gamma^{2}}{P^{3}}\left[R_{s}^{3}\left(\Phi_{1}\overline{{\mathchoice{\raisebox{0.0pt}{$\displaystyle\chi$}}{\raisebox{0.0pt}{$\textstyle\chi$}}{\raisebox{0.0pt}{$\scriptstyle\chi$}}{\raisebox{0.0pt}{$\scriptscriptstyle\chi$}}}}_{1}+\Phi_{2}\overline{{\mathchoice{\raisebox{0.0pt}{$\displaystyle\chi$}}{\raisebox{0.0pt}{$\textstyle\chi$}}{\raisebox{0.0pt}{$\scriptstyle\chi$}}{\raisebox{0.0pt}{$\scriptscriptstyle\chi$}}}}_{2}+\Phi_{3}\overline{{\mathchoice{\raisebox{0.0pt}{$\displaystyle\chi$}}{\raisebox{0.0pt}{$\textstyle\chi$}}{\raisebox{0.0pt}{$\scriptstyle\chi$}}{\raisebox{0.0pt}{$\scriptscriptstyle\chi$}}}}_{3}\right)+R_{s}^{2}\left(\Psi_{1}\overline{{\mathchoice{\raisebox{0.0pt}{$\displaystyle\chi$}}{\raisebox{0.0pt}{$\textstyle\chi$}}{\raisebox{0.0pt}{$\scriptstyle\chi$}}{\raisebox{0.0pt}{$\scriptscriptstyle\chi$}}}}_{4}+2\operatorname{Re}\\{\Psi_{2}\overline{{\mathchoice{\raisebox{0.0pt}{$\displaystyle\chi$}}{\raisebox{0.0pt}{$\textstyle\chi$}}{\raisebox{0.0pt}{$\scriptstyle\chi$}}{\raisebox{0.0pt}{$\scriptscriptstyle\chi$}}}}_{5}+\Psi_{3}\overline{{\mathchoice{\raisebox{0.0pt}{$\displaystyle\chi$}}{\raisebox{0.0pt}{$\textstyle\chi$}}{\raisebox{0.0pt}{$\scriptstyle\chi$}}{\raisebox{0.0pt}{$\scriptscriptstyle\chi$}}}}_{5}^{*}\\}+\Psi_{4}\overline{{\mathchoice{\raisebox{0.0pt}{$\displaystyle\chi$}}{\raisebox{0.0pt}{$\textstyle\chi$}}{\raisebox{0.0pt}{$\scriptstyle\chi$}}{\raisebox{0.0pt}{$\scriptscriptstyle\chi$}}}}_{6}\right.\right.\\\
&\left.\left.+2\operatorname{Re}\\{\Lambda_{1}\overline{{\mathchoice{\raisebox{0.0pt}{$\displaystyle\chi$}}{\raisebox{0.0pt}{$\textstyle\chi$}}{\raisebox{0.0pt}{$\scriptstyle\chi$}}{\raisebox{0.0pt}{$\scriptscriptstyle\chi$}}}}_{7}+\Lambda_{2}\overline{{\mathchoice{\raisebox{0.0pt}{$\displaystyle\chi$}}{\raisebox{0.0pt}{$\textstyle\chi$}}{\raisebox{0.0pt}{$\scriptstyle\chi$}}{\raisebox{0.0pt}{$\scriptscriptstyle\chi$}}}}_{7}^{*}(f)\\}+\Lambda_{3}\overline{{\mathchoice{\raisebox{0.0pt}{$\displaystyle\chi$}}{\raisebox{0.0pt}{$\textstyle\chi$}}{\raisebox{0.0pt}{$\scriptstyle\chi$}}{\raisebox{0.0pt}{$\scriptscriptstyle\chi$}}}}_{8}+2\operatorname{Re}\\{\Lambda_{4}\overline{{\mathchoice{\raisebox{0.0pt}{$\displaystyle\chi$}}{\raisebox{0.0pt}{$\textstyle\chi$}}{\raisebox{0.0pt}{$\scriptstyle\chi$}}{\raisebox{0.0pt}{$\scriptscriptstyle\chi$}}}}_{9}+\Lambda_{5}\overline{{\mathchoice{\raisebox{0.0pt}{$\displaystyle\chi$}}{\raisebox{0.0pt}{$\textstyle\chi$}}{\raisebox{0.0pt}{$\scriptstyle\chi$}}{\raisebox{0.0pt}{$\scriptscriptstyle\chi$}}}}_{9}^{*}\\}+\Lambda_{6}\overline{{\mathchoice{\raisebox{0.0pt}{$\displaystyle\chi$}}{\raisebox{0.0pt}{$\textstyle\chi$}}{\raisebox{0.0pt}{$\scriptstyle\chi$}}{\raisebox{0.0pt}{$\scriptscriptstyle\chi$}}}}_{10}\right)+R_{s}\Xi_{1}\overline{{\mathchoice{\raisebox{0.0pt}{$\displaystyle\chi$}}{\raisebox{0.0pt}{$\textstyle\chi$}}{\raisebox{0.0pt}{$\scriptstyle\chi$}}{\raisebox{0.0pt}{$\scriptscriptstyle\chi$}}}}_{11}\right],\end{split}$
(1)
where $\gamma$ is the fiber nonlinearity coefficient, $R_{s}$ is the symbol
rate and the coefficients $\Phi_{1}$, $\Phi_{2}$, $\Phi_{3}$, $\Psi_{1}$,
$\Psi_{2}$, …, $\Psi_{4}$, $\Lambda_{1}$, $\Lambda_{2}$, …, $\Lambda_{6}$,
$\Xi_{1}$ are functions of several different intra- and cross- polarization
moments of the DP-4D transmitted modulation format, and are given in [6, Table
8]. The
$\overline{{\mathchoice{\raisebox{0.0pt}{$\displaystyle\chi$}}{\raisebox{0.0pt}{$\textstyle\chi$}}{\raisebox{0.0pt}{$\scriptstyle\chi$}}{\raisebox{0.0pt}{$\scriptscriptstyle\chi$}}}}_{i}$
coefficients are obtained integrating over the channel bandwidth (see [6, eq.
(42)-(43)]) the frequency-dependent integrals
${\mathchoice{\raisebox{0.0pt}{$\displaystyle\chi$}}{\raisebox{0.0pt}{$\textstyle\chi$}}{\raisebox{0.0pt}{$\scriptstyle\chi$}}{\raisebox{0.0pt}{$\scriptscriptstyle\chi$}}}_{i}(f)$,
$i=1,2,\dots,11$ in [6, Table 8]. The expression for $\eta_{y}$ can be found
from (1) by simply applying the transformation $x\rightarrow y$ and
$y\rightarrow x$. As discussed in [6, Sec. 8], (1) reduces to the EGN formula
for conventional PM-2D formats.
16641281284096$30$$30.5$$31$$31.5$$32$$32.5$c4_164D-64PRSl4_1284D_128OSPM-64QAM0.8
dB$M$$\eta_{x}$ [dB 1/W2]4D model (1)SSFM4D-EGN model (a)
16641281284096$30$$30.5$$31$$31.5$$32$$32.5$c4_164D-64PRSl4_1284D_128OSPM-64QAM0.9
dB$M$$\eta_{y}$ [dB 1/W2]4D model (1)SSFM4D-EGN model (b)
Fig. 1: NLI power coefficients $\eta_{x}$ (a), and $\eta_{y}$ (b), for DP-4D
constellations with different cardinality $M$.
## 3 Methodology
Parameter | Value
---|---
TX parameters
Symbol rate ($R_{s}$) | 32 Gbaud
No. of channels | 1
RRC rolloff | 0.01 %
TX power ($P$) | $-20$ dBm
Fiber parameters
Attenuation coeff. ($\alpha$) | 0.2 (dB/km)
Dispersion par. ($D$) | 17 ps/nm/km
Nonlinear coeff. ($\gamma$) | 1.3 (W$\cdot\text{km})^{-1}$)
Link parameters
Span length | 100 km
No. of spans | 10
SSFM parameters
Step distribution | Adaptive step
$\phi_{NLmax}$ | $10^{-3}$ rad
Sim. bandwidth | 96 GHz
Table 1: System parameters.
The numerical validation of the model in this work is performed via the
estimation of the signal-to-noise ratio in split-step Fourier method (SSFM)
simulations where optical nonlinearity is kept as the sole source of noise.
Indeed, in such a scenario, $\mathbf{E}\approx(P_{x}(\text{SNR}_{x}\cdot
P^{3})^{-1},P_{y}(\text{SNR}_{y}\cdot P^{3})^{-1})$, where $P_{x}$, SNRx,
$P_{y}$, and SNRy are the transmitted powers and signal-to-noise ratios over
$x$ and $y$ polarization, resp. The previous formula becomes increasingly
accurate as higher-order NLI terms vanish compared to the first-order one,
i.e. for small values of $P$. The simulated single-channel, multi-span optical
system is described in Table 1. At the transmitter, 4D symbols are jointly
modulated using a root-raised cosine (RRC) pulse shape. At the receiver,
chromatic dispersion compensation and matched filtering followed by sampling
are performed. $\text{SNR}_{x}$ and $\text{SNR}_{y}$ are then computed via a
data-aided approach. Finally, an adaptive step size SSFM with a maximum
nonlinear phase rotation $\phi_{NLmax}$ per step is used to simulate the fiber
propagation.
The EGN model is also used as a reference to show the increased accuracy of
(1). However, in its standard formulation, the EGN model does not provide a
suitable expression to account for general DP-4D formats. An intuitive way to
extend the EGN model to general DP-4D formats is to consider the EGN
expressions as separately applicable to the two 2D constellations obtained by
the projection of the transmitted DP-4D format over the $x$ and $y$
polarization plane. In the following, we adopt this approach, labelled 4D-EGN
model, to compute EGN-based estimates of E.
## 4 Results
The numerical results in this section are based on a set of DP-4D formats
which combines the full list of 4D constellations in [7] (4D sphere packings),
and formats used in optical communications such as the 4D 64-ary polarization
ring switched (4D-64PRS) format [4], the 4D orthant-symmetric 128 (4D-OS128)
format [5], and the family of 4D 2-amplitudes 8 phase-shift keying (4D-2A8PSK)
formats [3].
In Fig. 1, $\eta_{x}$ (Fig. 1(a)) and $\eta_{y}$ (Fig. 1(b)) are shown for
some of the above mentioned constellations with different constellation
cardinalities $M$, using: i) (1) (blue bars); ii) the 4D-EGN model (yellow
bars); iii) the SSFM (red bars). For all constellations shown, our 4D model is
within 0.1 dB from the SSFM estimates for both $\eta_{x}$ and $\eta_{y}$. The
4D-EGN model leads to inaccuracies of up to 0.9 dB, for $\eta_{y}$ in
“l4_128”, or 0.8 dB for $\eta_{x}$ in “c4_16”. Neither “l4_128” nor “c4_16”
are symmetric constellations with respect to the $y=x$ plane, and this is
reflected by uneven values of $\eta_{x}$ and $\eta_{y}$. However, even for
symmetric constellations such as 4D-64PRS and 4D-128OS, the 4D-EGN model
deviates from SSFM estimates by approx. 0.6 dB in both $\eta_{x}$ and
$\eta_{y}$. As expected, for a PM-2D quadrature amplitude modulation (QAM)
format such as PM-64QAM, all three estimation methods are in perfect
agreement.
To further validate (1), the average deviation
$\overline{\Delta\eta}\triangleq(\Delta\eta_{x}+\Delta\eta_{y})/2$ is
computed, where $\Delta\eta_{x}$ and $\Delta\eta_{y}$ are the deviations (in
dB and in absolute value) for $\eta_{x}$ and $\eta_{y}$, respectively, between
a given model and the SSFM estimates. In Fig. 3, $\overline{\Delta\eta}$ is
illustrated for (1) and the 4D-EGN model as a function of the constellation
cardinality $M$. The solid lines show the average $\overline{\Delta\eta}$
across all considered constellations, whereas dashed lines indicate the
maximum and minimum deviation. The results show an average
$\overline{\Delta\eta}$ within 0.1 dB for (1) across all cardinalities, with
maximum $\overline{\Delta\eta}$ of 0.25 dB (for $M=16$). On the contrary, the
4D-EGN model shows deviations in excess of 1 dB for low cardinality formats
($M\leq$64), with worst-case scenario $\overline{\Delta\eta}$ of 1.75 dB for
$M=16$. At higher cardinalities ($M\geq$64), the 4D-EGN model accuracy
improves as the average $\overline{\Delta{\eta}}$ lies between 0.3 dB and 0.5
dB.
Const. label | $M$ | $(\eta_{x},\eta_{y})$ [dB 1/W2]
---|---|---
dicyclic4_16 [7] | 16 | (30.2, 30.2)
4D-2A-8PSK5b [3] | 32 | (30.3, 30.3)
4D-2A-8PSK6b [3] | 64 | (30.3, 30.3)
4D-2A-8PSK7b [3] | 128 | (30.3, 30.3)
w4_256 [7] | 256 | (30.7, 30.7)
sphere4_512 [7] | 512 | (30.7, 30.7)
120cell4_600 [7] | 600 | (30.3, 30.3)
a4_2048 [7] | 2048 | (30.7, 30.8)
a4_4096 [7] | 4096 | (30.8, 30.7)
Table 2: $\overline{\eta}$-optimal formats in Fig. 3.
Finally, in Fig. 3, the minimum values of the average NLI power coefficient
$\overline{\eta}\triangleq(\eta_{x}+\eta_{y})/2$ are shown for each $M$ within
the set of constellations analyzed in this work. The corresponding
$\overline{\eta}$-optimal constellations are listed in Table 2 for some values
of $M$. The 4D-EGN model (yellow markers) almost always overestimates
$\overline{\eta}$, with deviations up to 0.75 dB ($M=81$). Conversely, (1)
(blue markers) is consistently within 0.1 dB from the SSFM $\overline{\eta}$
(red markers). Among these optimal constellations, we highlight: i) “w4_256”
and “a4_4096”, already studied in the context of coded modulation (see refs.
in [7]), which outperform PM-16QAM and PM-64QAM, respectively; ii)
“120cell600” (see $x$ and $y$ projections in the inset in Fig. 3), which is a
constant-modulus constellation in 4D, and has the lowest $\overline{\eta}$
among all formats with $M\geq$128.
[] 163264128256512102420484096$0.1$$0.5$$1$$1.5$$1.8$$2$Max.
$\overline{\Delta\eta}$Min.
$\overline{\Delta\eta}$$M$$\overline{\Delta\eta}\text{ [dB]}$4D-EGN model4D
model (1)
Fig. 2: Average NLI power coeff. gap across the two polarizations
($\overline{\Delta\eta}$) as a function of the cardinality $M$.
[] 16326412825660020484096$30.5$$31$$31.5$$32$$32.5$$\approx$ 0.75
dB“120cell4_600”$x$ pol.$y$ pol.$M$$\overline{\eta}$ [dB 1/W2]4D-EGN model4D
model (1)SSFM
Fig. 3: Minimum values of $\overline{\eta}$ vs. cardinality $M$ for
the 4D constellations investigated in this work.
## 5 Conclusions
We tested a novel model predicting the NLI power for general DP-4D
constellations. Based on this preliminary study, the assessed model shows a
superior accuracy compared to a heuristic 4D extension of the EGN model. In
particular, a 0.1 dB average deviation from SSFM simulations is demonstrated
for the NLI power coefficient over a wide variety of regular 4D formats.
Although further numerical validation is required, we foresee the model in
this work as a powerful analytical tool for optimizing constellations in the
DP-4D space of the optical field.
Acknowledgements: The work of G. Liga is funded by the EuroTechPostdoc
programme under the European Union’s Horizon 2020 research and innovation
programme (Marie Skłodowska-Curie grant agreement No. 754462). This work has
received funding from the European Research Council (ERC) under the European
Union’s Horizon 2020 research and innovation programme (grant agreement No.
757791).
## References
* [1] A. Carena, G. Bosco, V. Curri, Y. Jiang, P. Poggiolini, and F. Forghieri, “EGN model of non-linear fiber propagation,” Opt.Express 22, 16335–16362 (2014).
* [2] E. Agrell and M. Karlsson, “Power-efficient modulation formats in coherent transmission systems,” JLT 27, 5115–5126 (2009).
* [3] K. Kojima, T. Yoshida, T. Koike-Akino, D. S. Millar, K. Parsons, M. Pajovic, and V. Arlunno, “Nonlinearity-tolerant four-dimensional 2A8PSK family for 5–7 bits/symbol spectral efficiency,” JLT 35, 1383–1391 (2017).
* [4] B. Chen, C. Okonkwo, H. Hafermann, and A. Alvarado, “Polarization-ring-switching for nonlinearity-tolerant geometrically shaped four-dimensional formats maximizing generalized mutual information,” JLT 37, 3579–3591 (2019).
* [5] B. Chen, A. Alvarado, S. van der Heide, M. v. d. Hout, H. Hafermann, and C. Okonkwo, “Analysis and experimental demonstration of orthant-symmetric four-dimensional 7 bit/4D-sym modulation for optical fiber communication,” arXiv:2003.12712 (2020).
* [6] G. Liga, A. Barreiro, H. Rabbani, and A. Alvarado, “Extending fibre nonlinear interference power modelling to account for general dual-polarisation 4D modulation formats,” Entropy 22, 1324 (2020).
* [7] “Sphere packings of dimension 4,” https://codes.se/packings/4.htm.
|
# A unified approach to Local Quantum Uncertainty and Interferometric Power by
Metric Adjusted Skew Information
Paolo Gibilisco Department of Economics and Finance, University of Rome “Tor
Vergata”, Via Columbia 2, Rome 00133, Italy<EMAIL_ADDRESS>Davide Girolami<EMAIL_ADDRESS>Politecnico di Torino, Corso Duca
degli Abruzzi 24, Torino 129, Italy Frank Hansen Department of Mathematical
Sciences, University of Copenaghen, Universitetsparken 5 DK-2100 Copenhagen,
Denmark<EMAIL_ADDRESS>
###### Abstract
Local quantum uncertainty and interferometric power have been introduced by
Girolami et al. in GTA:2013 ; GSGTFSSOA:2014 as geometric quantifiers of
quantum correlations. The aim of the present paper is to discuss their
properties in a unified manner by means of the the metric adjusted skew
information defined by Hansen in Hansen:2006b .
## I Introduction
One of the key traits of many-body quantum systems is that the full knowledge
of their global configurations does not imply full knowledge of their
constituents. The impossibility to reconstruct the local wave functions
$|\psi_{1}\rangle,\,|\psi_{2}\rangle$ (pure states) of two interacting quantum
particles from the wave function of the whole system,
$|\psi_{12}\rangle\neq|\psi_{1}\rangle\otimes|\psi_{2}\rangle$, is due to the
existence of entanglement ent . Investigating open quantum systems, whose
(mixed) states are described by density matrices
$\rho_{12}=\sum_{i}p_{i}|\psi_{i}\rangle_{12}\langle\psi_{i}|$, revealed that
the boundary between the classical and quantum worlds is more blurred than we
thought. There exists a genuinely quantum kind of correlation, quantum
discord, which manifests even in absence of entanglement, i.e. in separable
density matrices $\rho_{12}=\sum_{i}p_{i}\rho_{1,i}\otimes\rho_{2,i}$ oz ; hv
. The discovery triggered theoretical and experimental studies to understand
the physical meaning of quantum discord, and the potential use of it as a
resource for quantum technologies rev . Relying on the known interplay between
geometrical and physical properties of mixed states Uhlmann:1992 ; book , a
stream of works employed information geometry techniques to construct
quantifiers of quantum discord AyGibiliscoMatus: 2018 ; BogaertGirolami:2017 ;
CFTA:2018 ; GII:2009 ; FPA:2017 ; GibiliscoIsola:2011 . In particular, two of
the most popular ones are the Local Quantum Uncertainty (LQU) and the
Interferometric Power (IP) GTA:2013 ; GSGTFSSOA:2014 . A merit of these two
measures is that they admit an analytical form for $N$ qubit states across the
$1\,vs\,N-1$ qubit partition. Also, they have a clear-cut physical
interpretation. The lack of certainty about quantum measurement outcomes is
due to the fact that density matrices are changed by quantum operations. The
LQU evaluates the minimum uncertainty about the outcome of a local quantum
measurement, when performed on a bipartite system. It is proven that two-
particle density matrices display quantum discord if and only if they are not
“classical-quantum” states. That is, they are not (mixture of) eigenvalues of
local observables, $\rho_{12}\neq\sum_{i}p_{i}|i\rangle_{1}\langle
i|\otimes\rho_{2,i}$, or
$\rho_{12}\neq\sum_{i}p_{i}\rho_{1,i}\otimes|i\rangle_{2}\langle i|$, in which
$\\{|i\rangle\\}$ is an orthonormal basis. Indeed, this is the only case in
which one can identify a local measurement that does not change a bipartite
quantum state, whose spectral decomposition reads
$A_{1}=\sum_{i}\lambda_{i}|i\rangle_{1}\langle i|$, or
$A_{2}=\sum_{i}\lambda_{i}|i\rangle_{2}\langle i|$. The LQU was built as the
minimum of the Wigner-Yanase skew information, a well-known information
geometry measure WignerYanase:1963 , between a density matrix and a finite-
dimensional observable (Hermitian operator). It quantifies how much a density
matrix $\rho_{12}$ is different from being a zero-discord state. The IP was
concocted by following a similar line of thinking. Quantum discord implies a
non-classical sensitivity to local perturbations. This feature of quantum
particles, while apparently a limitation, translates into an advantage in the
context of quantum metrology metro . It was theoretically proven and
experimentally demonstrated that quantum systems sharing quantum discord are
more sensitive probes for interferometric phase estimation. The figure of
merit of such measurement protocols is the quantum Fisher information of the
state under scrutiny with respect to a local Hamiltonian (in Information
Geometry the QFI is known as the SLD or Bures-Uhlmann metric). The latter
generates a unitary evolution that imprints information about a physical
parameter on the quantum probe. The IP is the minimum quantum Fisher
information over all the possible local Hamiltonians, being zero if and only
if the probe states are classically correlated.
Here, we polish and extend the mathematical formalization of information-
geometric quantum correlation measures. We build a class of parent quantities
of the LQU (and consequently of the IP) in terms of the the metric adjusted
skew informations Hansen:2006b . In Sections 2,3, we review definition and
main properties of operator means. In Sections 4-6, we discuss information-
geometric quantities that capture complementarity between quantum states and
observables. In particular, we focus on the quantum $f$-covariances and the
quantum Fisher information. They quantify the inherent uncertainty about
quantum measurement outcomes. After having recalled the definition of metric
adjusted skew information (Section 7), we build a new quantum discord measure,
the metric adjusted local quantum uncertainty ($f$-LQU), in Section 8. Finally
we are able to show that LQU and IP are just two particular members of this
family allowing a unified treatment of their fundamental properties.
## II Means for positive numbers
We use the notation ${\mathbb{R}}_{+}=(0,+\infty)$.
###### Definition 1.
A bivariate mean PetzTemesi:2005 is a function
$m\colon{\mathbb{R}}_{+}\times{\mathbb{R}}_{+}\to{\mathbb{R}}_{+}$ such that
1. 1.
$m(x,x)=x.$
2. 2.
$m(x,y)=m(y,x).$
3. 3.
$x<y$ $\,\Rightarrow\,$ $x<m(x,y)<y.$
4. 4.
$x<x^{\prime}$ and $y<y^{\prime}$ $\,\Rightarrow\,$
$m(x,y)<m(x^{\prime},y^{\prime}).$
5. 5.
$m$ is continuous.
6. 6.
$m$ is positively homogeneous; that is $m(tx,ty)=t\cdot m(x,y)$ for $t>0.$
We use the notation $\mathcal{M}_{num}$ for the set of bivariate means
described above.
###### Definition 2.
Let $\mathcal{F}_{num}$ denote the class of functions
$f\colon\mathbb{R}_{+}\to\mathbb{R}_{+}$ such that
1. 1.
$f$ is continuous.
2. 2.
$f$ is monotone increasing.
3. 3.
$f(1)=1.$
4. 4.
$tf(t^{-1})=f(t)$.
The following result is straightforward.
###### Proposition 1.
There is a bijection $f\mapsto m_{f}$ betwen ${\mathcal{F}}_{nu}$ and
${\mathcal{M}}_{nu}$ given by
$m_{f}(x,y)=yf(y^{-1}x)\qquad\text{and in reverse}\qquad f(t)=m(1,t)$
for positive numbers $x,y$ and $t.$
In Table 1 we have some examples of means.
Table 1: $\begin{array}[]{|c|c|c|}\hline\cr{\rm Name}&{f}&{m_{f}}\\\
\hline\cr{\rm
arithmetic}&\displaystyle\frac{1+x}{2}&\displaystyle\frac{x+y}{2}\\\
\hline\cr{\rm
WYD},\beta\in(0,1)&\displaystyle\frac{x^{\beta}+x^{1-\beta}}{2}&\displaystyle\frac{x^{\beta}y^{1-\beta}+x^{1-\beta}y^{\beta}}{2}\\\
\hline\cr{\rm geometric}&\sqrt{x}&\sqrt{xy}\\\ \hline\cr{\rm
harmonic}&\displaystyle\frac{2x}{x+1}&\displaystyle\frac{2}{x^{-1}+y^{-1}}\\\
\hline\cr{\rm logarithmic}&\displaystyle\frac{x-1}{\log
x}&\displaystyle\frac{x-y}{\log x-\log y}\\\ \hline\cr\end{array}$
## III Means for positive operators in the sense of Kubo-Ando
The celebrated Kubo-Ando theory of operator means KuboAndo79/80 ;
PetzTemesi:2005 ; GibiliscoHansenIsola:2009 may be viewed as the operator
version of the results of Section II.
###### Definition 3.
A bivariate mean $m$ for pairs of positive operators is a function
$(A,B)\to m(A,B),$
defined in and with values in positive definite operators on a Hilbert space,
that satisfies, mutatis mutandis, conditions $(1)$ to $(5)$ in Definition 1.
In addition, the transformer inequality
$Cm(A,B)C^{*}\leq m(CAC^{*},CBC^{*}),$
should also hold for positive definite $A,B$ and arbitrary $C.$
Note that the transformer inequality replaces condition $(6)$ in Definition 1.
We denote by $\displaystyle{\mathcal{M}}_{op}$ the set of matrix means.
###### Example 1.
The arithmetic, geometric and harmonic operator means are defined,
respectively, by setting
$\begin{array}[]{rcl}A\nabla B&=&\frac{1}{2}(A+B)\\\\[6.45831pt]
A\\#B&=&A^{1/2}\bigl{(}A^{-1/2}BA^{-1/2}\bigr{)}^{1/2}A^{1/2}\\\\[8.61108pt]
A{\rm!}B&=&2(A^{-1}+B^{-1})^{-1}.\end{array}$
We recall that a function $f\colon(0,\infty)\to\mathbb{R}$ is said to be
operator monotone (increasing) if
$A\leq B\quad\Rightarrow\quad f(A)\leq f(B)$
for positive definite matrices of arbitrary order. It then follows that the
inequality also holds for positive operators on an arbitrary Hilbert space. An
operator monotone function $f$ is said to be symmetric if $f(t)=tf(t^{-1})$
for $t>0$ and normalized if $f(1)=1.$
###### Definition 4.
${\mathcal{F}}_{op}$ is the class of functions
$f:{\mathbb{R}}_{+}\to{\mathbb{R}}_{+}$ such that
1. 1.
$f$ is operator monotone increasing,
2. 2.
$tf(t^{-1})=f(t)\qquad t>0,$
3. 3.
$f(1)=1.$
The fundamental result, due to Kubo and Ando, is the following.
###### Theorem.
There is a bijection $f\mapsto m_{f}$ between ${\mathcal{M}}_{op}$ and
${\mathcal{F}}_{op}$ given by the formula
$m_{f}(A,B)=A^{1/2}f(A^{-1/2}BA^{-1/2})A^{1/2}.$
###### Remark 1.
The function in ${\mathcal{F}}_{op}$ are (operator) concave which makes the
operator case quite different from the numerical (commutative) case. For
example, there exist convex functions in $\mathcal{F}_{num},$ see GH:2017 .
If $\rho$ is a density matrix (a quantum state) and $A$ is a self-adjoint
matrix (a quantum observable), then the expectation of $A$ in the state $\rho$
is defined by setting
${\rm E}_{\rho}(A)={\rm Tr\hskip-1.99997pt}~{}(\rho A).$
## IV The correspondence between Fisher information and metric adjusted skew
information
We introduce now a technical tool which is useful to establish some
fundamental relations between quantum covariance, quantum Fisher information
and the metric adjusted skew information.
###### Definition 5.
For $f\in{\mathcal{F}}_{op}$ we define $f(0)=\lim_{x\to 0}f(x).$ We say that a
function $f\in{\mathcal{F}}_{op}$ is regular if $f(0)\not=0,$ and non-regular
if $f(0)=0,$ cf. PetzSudar:1996 ; Hansen:2006b .
###### Definition 6.
A quantum Fisher information is extendable if its radial limit exists and it
is a Riemannian metric on the real projective space generated by the pure
states.
For the definition of the radial limit see PetzSudar:1996 where the following
fundamental result is proved.
###### Theorem.
An operator monotone function $f\in{\mathcal{F}}_{op}$ is regular, if and only
if $\langle\cdot,\cdot\rangle_{\rho,f}$ is extendable.
###### Remark 2.
The reader should be aware that there is no negative connotation associated
with the qualification “non-regular”. For example, a very important quantum
Fisher information in quantum physics (see FickSauermann:1990 ), namely the
Kubo-Mori metric related to the function $f(x)=(x-1)/\log x,$ is non-regular.
We introduce the sets of regular and non-regular functions
${\mathcal{F}}_{op}^{\,r}:=\\{f\in{\mathcal{F}}_{op}\mid
f(0)\not=0\\},\quad{\mathcal{F}}_{op}^{\,n}:=\\{f\in{\mathcal{F}}_{op}\mid
f(0)=0\\}$
and notice that trivially
${\mathcal{F}}_{op}={\mathcal{F}}_{op}^{\,r}\dot{\cup}{\mathcal{F}}_{op}^{\,n}$
.
###### Definition 7.
We introduce to $f\in{\mathcal{F}}_{op}^{\,r}$ the transform $\tilde{f}$ given
by
$\tilde{f}(x)=\frac{1}{2}\left[(x+1)-(x-1)^{2}\frac{f(0)}{f(x)}\right]$
for $x>0.$ We may also write ${\tilde{f}}={\mathcal{G}}(f),$ cf.
GibiliscoImparatoIsola:2007 ; GibiliscoHansenIsola:2009 .
The following result is taken from (GibiliscoHansenIsola:2009, , Theorem 5.1).
###### Theorem.
The correspondence $f\to\tilde{f}$ is a bijection between
${\mathcal{F}}_{op}^{\,r}$ and ${\mathcal{F}}_{op}^{\,n}\,.$
In Table 2 we have some examples (where $0<\beta<1$).
Table 2: $\displaystyle\begin{array}[]{|c|c|}\hline\cr f&\tilde{f}\\\
\hline\cr\displaystyle\frac{1+x}{2}&\displaystyle\frac{2x}{x+1}\\\
\hline\cr\displaystyle\frac{(\sqrt{x}+1)^{2}}{4}&\sqrt{x}\\\
\hline\cr\displaystyle\beta(1-\beta)\frac{(x-1)^{2}}{(x^{\beta}-1)(x^{1-\beta}-1)}&\displaystyle\frac{x^{\beta}+x^{1-\beta}}{2}\\\
\hline\cr\end{array}$
## V Quantum f-Covariance
The notion of quantum $f$-covariance has been introduced by Petz, see
Petz:2003 ; GibiliscoHiaiPetz:2009 . Any Kubo-Ando function
$m_{f}(x,y)=yf(y^{-1}x)$ for $x,y>0$ has a continuous extension to
$[0,+\infty)\times[0,+\infty)$ given by
$m_{f}(0,y)=f(0)y,\quad m_{f}(x,0)=f(0)x,\quad m_{f}(0,0)=0,\quad x,y>0.$
The operator $m_{f}(L_{\rho},R_{\rho})$ is well-defined by the spectral
theorem for any state, see (GibiliscoImparatoIsola:2007, , Proposition 11.1
page 11). To self-adjoint $A$ we set $A_{0}=A-({\rm Tr\hskip-1.99997pt}~{}\rho
A)I,$ where $I$ is the identity operator. Note that
${\rm Tr\hskip-1.99997pt}~{}\rho A_{0}={\rm Tr\hskip-1.99997pt}~{}\rho A-({\rm
Tr\hskip-1.99997pt}~{}\rho A){\rm Tr\hskip-1.99997pt}~{}\rho=0,$
if $\rho$ is a state.
###### Definition 8.
Given a state $\rho,$ a function $f\in\mathcal{F}_{op}$ and self-adjoint $A,B$
we define the quantum $f$-covariance by setting
${\rm Cov}_{\rho}^{f}(A,B)={\rm
Tr\hskip-1.99997pt}~{}B_{0}\,m_{f}(L_{\rho},R_{\rho})A_{0}$
and the corresponding quantum $f$-variance by ${\rm Var}_{\rho}^{f}(A)={\rm
Cov}_{\rho}^{f}(A,A).$
The $f$-variance is a positive semi-definite sesquilinear form and
$f\leq g\quad\Rightarrow\quad{\rm Var}_{\rho}^{f}(A)\leq{\rm
Var}_{\rho}^{g}(A).$ (1)
Note that for the standard covariance we have ${\rm Cov}_{\rho}(A,B)={\rm
Cov}_{\rho}^{SLD}(A,B),$ where the SLD or Bures-Uhlmann metric is the one
associated with the function $(1+x)/2$.
###### Proposition 2.
If $\rho$ is a pure state then ${\rm Var}_{\rho}^{f}(A)=2\,m_{f}(1,0)\cdot{\rm
Var}_{\rho}(A),$ cf. TothPetz:2013 .
###### Corollary.
If $\rho$ is a pure state and $f$ is non-regular, then ${\rm
Var}_{\rho}^{f}(A)=0.$
###### Proof.
If $f$ is non regular $m_{f}(1,0)=0$ QED
## VI Quantum Fisher Information
The theory of quantum Fisher information is due to Petz and we recall here the
basic results. If ${\mathcal{N}}$ is a differentiable manifold we denote by
$T_{\rho}\mathcal{N}$ the tangent space to $\mathcal{N}$ at the point
$\rho\in{\mathcal{N}}$. Recall that there exists a natural identification of
$T_{\rho}{\mathcal{D}}^{1}_{n}$ with the space of self-adjoint traceless
matrices; namely, for any $\rho\in{\mathcal{D}}^{1}_{n}$
$T_{\rho}{\mathcal{D}}^{1}_{n}=\\{A\in M_{n}\mid
A=A^{*}\,,\,\hbox{Tr}\,A=0\\}.$
A stochastic map is a completely positive and trace preserving operator
$T:M_{n}\to M_{m}$. A monotone metric is a family of Riemannian metrics
$g=\\{g^{n}\\}$ on $\\{{\mathcal{D}}^{1}_{n}\\}$, $n\in\mathbb{N}$, such that
$g^{m}_{T(\rho)}(TX,TX)\leq g^{n}_{\rho}(X,X)$
holds for every stochastic map $T:M_{n}\to M_{m}$, every faithful state
$\rho\in{\mathcal{D}}^{1}_{n},$ and every $X\in
T_{\rho}{\mathcal{D}}^{1}_{n}$. Usually monotone metrics are normalized in
such a way that $[A,\rho]=0$ implies $g_{\rho}(A,A)={\rm
Tr}({\rho}^{-1}A^{2})$. A monotone metric is also called (an example of)
quantum Fisher information (QFI). This notation is inspired by Chentsov’s
uniqueness theorem for commutative monotone metrics Chentsov:1982 .
Define $L_{\rho}(A)=\rho A$ and $R_{\rho}(A)=A\rho$, and observe that
$L_{\rho}$ and $R_{\rho}$ are commuting positive superoperators on $M_{n}.$
For any $f\in{\mathcal{F}}_{op}$ one may also define the positive (non-linear)
superoperator $m_{f}(L_{\rho},R_{\rho})$. The fundamental theorem of monotone
metrics may be stated in the following way:
###### Theorem.
(See Petz:1996 ). There exists a bijective correspondence between monotone
metrics (quantum Fisher information(s)) on ${\mathcal{D}}^{1}_{n}$ and
functions $f\in{\mathcal{F}}_{op}$. The correspondence is given by the formula
$\langle A,B\rangle_{\rho,f}={\rm Tr}(A\cdot
m_{f}(L_{\rho},R_{\rho})^{-1}(B))$
for positive matrices $A$ and $B.$
## VII Metric adjusted skew information
By using the general form of the quantum Fisher information it is possible to
greatly generalize the Wigner-Yanase information measure. To
$f\in\mathcal{F}_{op}$ the so-called Morosova function $c_{f}(x,y)$ is defined
by setting
$c_{f}(x,y)=\frac{1}{yf(xy^{-1})}=m_{f}(x,y)^{-1}\qquad x,y>0.$ (2)
The corresponding monotone symmetric metric $K_{\rho}$ is given by
$K_{\rho}^{f}(A,B)={\rm
Tr\hskip-1.99997pt}~{}A^{*}c_{f}\bigl{(}L_{\rho},R_{\rho}\bigr{)}B,$ (3)
where $L_{\rho}$ and $R_{\rho}$ denote left and right multiplication with
$\rho.$ Note that $K^{f}_{\rho}(A)$ is increasing in $c_{f}$ and thus
decreasing in $f.$ If furthermore $f$ is regular, the notion of metric
adjusted skew information (Hansen:2006b, , Definition 1.2) is defined by
setting
$I^{f}_{\rho}(A)=I^{f}(\rho,A)=\frac{f(0)}{2}K^{f}_{\rho}\bigl{(}i[\rho,A^{*}],i[\rho,A]\bigr{)},$
(4)
where $\rho>0.$ We use the second notation, $I^{f}(\rho,A),$ when the
expression of the state takes up too much space. We also tacitly extended the
metric adjusted skew information to arbitrary (non-self-adjoint) operators
$A.$ It is convex (Hansen:2006b, , Theorem 3.7) in the state variable $\rho$
and
$0\leq I^{f}_{\rho}(A)\leq{\rm Var}_{\rho}(A)$ (5)
with equality if $\rho$ is pure (Hansen:2006b, , Theorem 3.8), see also the
summery with interpretations in (CaiHansen:2010, , Theorem 1.2). Furthermore,
the notion of unbounded metric adjusted skew information for non-regular
functions in $\mathcal{F}_{op}$ is introduced in (CaiHansen:2010, , Theorem
5.1). For regular $f\in\mathcal{F}_{op}$ the metric adjusted skew information
may be written as
$I_{\rho}^{f}(A)={\rm Tr\hskip-1.99997pt}~{}\rho A^{2}-{\rm
Tr\hskip-1.99997pt}~{}A\,m_{\tilde{f}}(L_{\rho},R_{\rho})A,$
se (AudenaertCaiHansen:2008, , equation (7)). We thus obtain that the metric
adjusted skew information is decreasing in the transform $\tilde{f}$ for
arbitrary self-adjoint $A,$ that is
$\tilde{f}\leq\tilde{g}\quad\Rightarrow\quad I_{\rho}^{f}(A)\geq
I_{\rho}^{g}(A)\qquad\text{for}\quad f,g\in{\mathcal{F}}_{op}^{\,r}\,.$ (6)
We may also write
$\check{f}=\frac{f(0)}{f(t)}\qquad\text{and}\qquad\check{c}(x,y)=y^{-1}\check{f}(xy^{-1})$
and obtain
$I_{\rho}^{f}(A)=\frac{1}{2}{\rm
Tr\hskip-1.99997pt}~{}i[\rho,A^{*}]\check{c}\bigl{(}L_{\rho},R_{\rho}\bigr{)}i[\rho,A],$
cf. (AudenaertCaiHansen:2008, , equation (10)). It follows that the metric
adjusted skew information is increasing in $\check{f}$ for arbitrary $A.$ It
may be derived from (GibiliscoImparatoIsola:2007, , Proposition 6.3, page 11),
that the metric adjusted skew information can be expressed as the difference
$I^{f}_{\rho}(A)={\rm Var}_{\rho}(A)-{\rm Var}^{\tilde{f}}_{\rho}(A)$
with extension to the sesquilinear form
$I^{f}_{\rho}(A,B)={\rm Cov}_{\rho}(A,B)-{\rm Cov}^{\tilde{f}}_{\rho}(A,B).$
### VII.1 Information inequalities
A function $f\colon\mathbb{R_{+}}\to\mathbb{R_{+}}$ is in $\mathcal{F}_{op}$
if and only if it allows a representation of the form
$f(t)=\frac{1+t}{2}\exp\int_{0}^{1}\frac{(\lambda^{2}-1)(1-t)^{2}}{(\lambda+t)(1+\lambda
t)(1+\lambda)^{2}}\,h_{f}(\lambda)\,d\lambda,$ (7)
where the weight function $h_{f}\colon[0,1]\to[0,1]$ is measurable. The
equivalence class containing $h_{f}$ is uniquely determined by $f,$ cf.
(AudenaertCaiHansen:2008, , Theorem 2.1). This representation gives rise to an
order relation in $\mathcal{F}_{op}.$
###### Definition 9.
Let $f,g\in\mathcal{F}_{op}.$ We say that $f$ is majorized by $g$ and write
$f\preceq g,$ if the function
$\varphi(t)=\frac{t+1}{2}\,\frac{f(t)}{g(t)}\qquad t>0$
is in $\mathcal{F}_{op}\,.$
The partial order relation $\preceq$ is stronger that the usual order relation
$\leq,$ and it renders $(\mathcal{F}_{op}\,,\preceq)$ into a lattice with
$f_{\text{min}}(t)=\frac{2t}{t+1}\qquad\text{and}\qquad
f_{\text{max}}(t)=\frac{t+1}{2}\qquad$ (8)
as respectively minimal element and maximal element. Furthermore,
$f\preceq g\quad\text{if and only if}\quad h_{f}\geq h_{g}\qquad\text{almost
everywhere},$ (9)
cf. (AudenaertCaiHansen:2008, , Theorem 2.4). The restriction of $\preceq$ to
the regular part of $\mathcal{F}_{op}$ induces a partial order relation
$\preceq$ on the set of metric adjusted skew informations.
###### Proposition 3.
The restriction of the order relation $\preceq$ renders the regular part of
$\mathcal{F}_{op}$ into a lattice. In addition, if one of two functions
$f,g\in\mathcal{F}_{op}$ is non-regular, then the minorant $f\wedge g$ is also
non-regular.
###### Proof.
Take $f\in\mathcal{F}_{op}$ with representative function $h_{f}$ as given in
(7). Then it follows that $f$ is regular if and only if the integral
$\int_{0}^{1}\frac{h_{f}(\lambda)}{\lambda}\,\hskip-1.99997pt\operatorname{\mathit{d}}\hskip-2.29996pt{}\lambda<\infty.$
(10)
Take now regular functions $f,g\in\mathcal{F}_{op\,.}$ We know that
$\bigl{(}\mathcal{F}_{op}\,,\preceq\bigr{)}$ is a lattice
(AudenaertCaiHansen:2008, , bottom of page 141), and that the representative
function in (7) for the minorant $f\wedge g$ is given by
$h_{f\wedge g}=\max\\{h_{f},h_{g}\\}\leq h_{f}+h_{g}$
showing that also $h_{f\wedge g}$ satisfies the integrability condition (10)
implying that $f\wedge g$ is regular. Since
$h_{f\vee g}=\min\\{h_{f},h_{g}\\}\leq h_{f}$
it also follows that the majorant is regular.
We now take functions $f,g\in\mathcal{F}_{op}$ with representative functions
$h_{f}$ and $h_{g}$ and assume that $f$ is non-regular. Since
$h_{f\wedge g}=\max\\{h_{f},h_{g}\\}\qquad\text{and thus}\qquad h_{f}\leq
h_{f\wedge g}$
we obtain that also the minorant $f\wedge g$ is non-regular. QED
### VII.2 The Wigner-Yanase-Dyson skew informations
The Wigner-Yanase-Dyson skew information (with parameter $p)$ is defined by
setting
$I_{p}(\rho,A)=-\frac{1}{2}{\rm
Tr\hskip-1.99997pt}~{}[\rho^{p},A[[\rho^{1-p},A],\qquad 0<p<1.$
It is an example of a metric adjusted skew information and reduces to the
Wigner-Yanase skew information for $p=1/2\,.$ The representing function
$f_{p}$ of $I_{p}(\rho,A)$ is given by
$f_{p}(t)=p(1-p)\cdot\frac{(t-1)^{2}}{(t^{p}-1)(t^{1-p}-1)}\qquad 0<p<1,$
that is $I_{p}(\rho,A)=I^{f_{p}}_{\rho}(A).$ The weight-functions
$h_{p}(\lambda)$ in equation (7) corresponding to the representing functions
$f_{p}$ are given by
$h_{p}(\lambda)=\frac{1}{\pi}\arctan\frac{(\lambda^{p}+\lambda^{1-p})\sin
p\pi}{1-\lambda-(\lambda^{p}-\lambda^{1-p})\cos p\pi}\qquad 0<\lambda<1.$
It is a non-trivial result that the Wigner-Yanase-Dyson skew informations
$I_{p}(\rho,A)$ are increasing in the parameter $p$ for $0<p\leq 1/2$ and
decreasing in $p$ for $1/2\leq p<1$ with respect to the order relation
$\preceq,$ cf. (AudenaertCaiHansen:2008, , Theorem 2.8). The Wigner-Yanase
skew information is thus the maximal element among the Wigner-Yanase-Dyson
skew informations with respect to the order relation $\preceq.$
### VII.3 The monotonous bridge
The family of metrics with representing functions
$f_{\alpha}(t)=t^{\alpha}\left(\frac{1+t}{2}\right)^{1-2\alpha}\qquad t>0,$
decrease monotonously (with respect to $\preceq)$ from the largest monotone
symmetric metric down to the Bures metric for $\alpha$ increasing from $0$ to
$1.$ They correspond the the constant weight functions
$h_{\alpha}(\lambda)=\alpha$ in equation (7). However, the only regular metric
in this bridge is the Bures metric $(\alpha=1).$ It is however possible to
construct a variant bridge by choosing the weight functions
$h_{p}(\lambda)=\left\\{\begin{array}[]{lrl}0,&\lambda&<1-p\\\\[4.30554pt]
p,&\lambda&\geq 1-p\end{array}\right.\qquad 0\leq p\leq 1$
in equation (7) instead of the constant weight functions. It is non-trivial
that these weight functions provide a monotonously decreasing bridge (with
respect to $\preceq)$ of monotone symmetric metrics between the smallest and
the largest (monotone symmetric) metric. The benefit of this variant bridge is
that all the constituent metrics are regular except for $p=1.$
## VIII Metric adjusted local quantum uncertainty
We consider a bipartite system
$\mathcal{H}=\mathcal{H}_{1}\otimes\mathcal{H}_{2}$ of two finite dimensional
Hilbert spaces.
###### Definition 10.
Let $f\in\mathcal{F}_{\text{op}}$ be regular and take a vector
$\Lambda\in\mathbf{R}^{d}.$ We define the Metric Adjusted Local Quantum
Uncertainty (or $f$-LQU) by setting
$\mathcal{U}_{1}^{\Lambda,f}(\rho)=\inf\\{I^{f}_{\rho}(K_{1}\otimes 1_{2})\mid
K_{1}\text{ has spectrum $\Lambda$}\\},$ (11)
where $\rho_{12}$ is a bipartite state, and $K_{1}$ is the partial trace of an
observable $K$ on $\mathcal{H}.$
The minimum in the above definition is thus taken over local observables
$K_{1}\otimes 1_{2}\in B(\mathcal{H}_{1}\otimes H_{2})$ such that $K_{1}$ is
unitarily equivalent with the diagonal matrix $\text{diag}(\Lambda).$
###### Remark 3.
The metric adjusted LQU has been studied in the literature for specific
choices of $f.$
* •
If $f(x)=f_{WY}(x)=\Bigl{(}\frac{1+\sqrt{x}}{2}\Bigr{)}^{2}$ then
${\mathcal{U}}_{1}^{\Lambda,f}$ coincide with the LQU introduced in (GTA:2013,
, equation 2).
* •
If $f(x)=f_{SLD}(x)=\frac{1+x}{2}$ then $\mathcal{U}_{1}^{\Lambda,f}$
coincides with the Interferometric Power introduced in GSGTFSSOA:2014 .
###### Proposition 4.
For $f,g\in\mathcal{F}_{op}^{\,r}$ with $\tilde{g}\leq\tilde{f}$ we have the
inequality
${\mathcal{U}}^{\Lambda,f}_{1}(\rho_{12})\leq{\mathcal{U}}^{\Lambda,g}_{1}(\rho_{12}).$
In particular the LQU is smaller than the IP.
###### Proof.
Let $\tilde{K}_{1}$ be the local observable with spectrum $\Lambda$ minimizing
the metric adjusted skew information. Then
${\mathcal{U}}^{\Lambda,f}_{1}(\rho_{12})=I^{f}_{\rho_{12}}\bigl{(}\tilde{K}_{1}\otimes
1_{2}\bigr{)}\geq I^{g}_{\rho_{12}}\bigl{(}\tilde{K}_{1}\otimes
1_{2}\bigr{)}\geq{\mathcal{U}}^{\Lambda,g}_{1}(\rho_{12}),$
where we used the inequality in (6). QED
###### Corollary.
Let $g_{1}$ and $g_{2}$ be regular functions in $\mathcal{F}_{op}$ and set
$f=\tilde{g}_{1}\wedge\tilde{g}_{2}$ with respect to the lattice structure in
$\mathcal{F}_{op}\,.$ Then there is a regular function $g$ in
$\mathcal{F}_{op}$ such that $\tilde{g}=f=\tilde{g}_{1}\wedge\tilde{g}_{2}$
and
$\max\bigl{\\{}{\mathcal{U}}^{\Lambda,g_{1}}_{1}(\rho_{12}),\,{\mathcal{U}}^{\Lambda,g_{2}}_{1}(\rho_{12})\bigr{\\}}\leq{\mathcal{U}}^{\Lambda,g}_{1}(\rho_{12})$
for arbitrary $\rho_{12}\,.$
###### Proof.
The functions $\tilde{g}_{1}$ and $\tilde{g}_{2}$ are non-regular by Theorem
Theorem. By Proposition 3 we thus obtain that also the minorant $f$ is non-
regular. Therefore there exists, by the correspondence in Theorem Theorem, a
(unique) regular function $g$ in $\mathcal{F}_{op}$ such that $\tilde{g}=f.$
The assertion then follows by Proposition 4. QED
Following BogaertGirolami:2017 we prove that the metric adjusted LQU is a
measure of non-classical correlations, i.e. it meets the criteria which
identify discord-like quantifiers, see rev .
###### Theorem.
If the state $\rho$ is classical-quantum in the sense of Piani , then the
metric adjusted LQU vanishes, that is $\mathcal{U}_{1}^{\Lambda,f}(\rho)=0.$
Conversely, if the coordinates of $\Lambda$ are mutually different (thus
rendering the operator $K_{1}$ non-degenerate) and
$\mathcal{U}_{1}^{\Lambda}(\rho)=0,$ then $\rho$ is classical-quantum.
###### Proof.
We note that the metric adjusted skew information $I_{\rho}^{f}(A)$ for a
faithful state $\rho$ is vanishing if and only if $\rho$ and $A$ commute. If
$\rho$ is classical-quantum, then
$P_{1}(\rho)=\sum_{i}(P_{1,i}\otimes 1_{2})\rho(P_{1,i}\otimes 1_{2})=\rho$
for some von Neumann measurement $P$ given by a resolution $(P_{i})$ of the
identity $1_{1}$ in terms of one-dimensional projections. We may choose
$K_{1}$ diagonal with respect to this resolution, so $K_{1}\otimes 1_{2}$ and
$\rho$ commute and thus $\mathcal{U}_{1}^{\Lambda,f}(\rho)=0.$
If on the other hand the $f$-LQU $\mathcal{U}_{1}^{\Lambda,f}(\rho)=0,$ then
there exist a local observable $K_{1}\otimes 1_{2}$ such that
$[\rho,K_{1}\otimes 1_{2}]=0.$ By the spectral theorem we write
$K_{1}=\sum_{i}\lambda_{i}P_{1,i}=\sum_{i}\lambda_{i}|i\rangle_{1}\langle i|$
and since
$\rho(K_{1}\otimes 1_{2})=(K_{1}\otimes 1_{2})\rho$
we obtain by multiplying with $P_{1,i}\otimes 1_{2}$ from the left and
$P_{1,j}\otimes 1_{2}$ from the right the identity
$\lambda_{j}(P_{1,i}\otimes 1_{2})\rho(P_{1,j}\otimes
1_{2})=\lambda_{i}(P_{1,i}\otimes 1_{2})\rho(P_{1,j}\otimes 1_{2}).$
If $K_{1}$ is non-degenerate, it thus follows that
$(P_{1,i}\otimes 1_{2})\rho(P_{1,j}\otimes 1_{2})=0\qquad\text{for}\quad i\neq
j.$
By summing over all $j$ different from $i$ we obtain
$(P_{1,i}\otimes 1_{2})\rho((1_{1}-P_{1,i})\otimes 1_{2})=0,$
thus
$(P_{1,i}\otimes 1_{2})\rho=(P_{1,i}\otimes 1_{2})\rho(P_{1,i}\otimes 1_{2}),$
so $P_{1,i}\otimes 1_{2}$ and $\rho$ commute. It follows that
$P_{1}(\rho)=\sum_{i}(P_{1,i}\otimes 1_{2})\rho(P_{1,i}\otimes 1_{2})=\rho,$
so $\rho$ is left invariant under the von Neumann measurement $P$ given by
$(P_{i}).$ Therefore, $\rho$ is classical-quantum. QED
Recall that Luo and Zhang LuoZang:2008 proved that a state $\rho$ is
classical-quantum if and only if there exists a resolution $(P_{i})$ of the
identity $1_{1}$ such that
$\rho=\sum_{i}p_{i}P_{1,i}\otimes\rho_{i},$
where each $\rho_{i}$ is a state on $\mathcal{H}_{2}$ and $p_{i}\geq 0,$ and
the sum $\sum_{i}p_{i}=1.$ By (CaiHansen:2010, , Lemma 3.1) the inequality
$I_{\rho}^{f}(K_{1}\otimes 1_{2})\geq I_{\rho_{1}}^{f}(K_{1})$
is valid, where $\rho_{1}={\rm Tr\hskip-1.99997pt}~{}_{2}\,\rho_{12}\,.$
Consequently, we obtain that
$\mathcal{U}_{1}^{\Lambda,f}(\rho)\geq\inf_{K_{1}}\,I^{f}_{\rho_{1}}(K_{1})=\inf_{\sigma_{1}}\,I^{f}_{\sigma_{1}}\bigl{(}K_{1}\bigr{)},$
(12)
where the minimum is taken over states $\sigma_{1}$ on $\mathcal{H}_{1}$
unitarily equivalent with $\rho_{1}.$
###### Theorem.
The metric adjusted LQU is invariant under local unitary transformations.
###### Proof.
For the metric adjusted skew information and local unitary transformations we
have
$\begin{array}[]{l}\mathcal{U}_{1}^{\Lambda,f}\bigl{(}(U_{1}\otimes
U_{2})\rho_{12}(U_{1}\otimes U_{2})^{\dagger}\bigr{)}\\\\[6.45831pt]
=\min_{K_{1}}I^{f}\bigl{(}(U_{1}\otimes U_{2})\rho_{12}(U_{1}\otimes
U_{2})^{\dagger},K_{1}\otimes 1_{2}\bigr{)}\\\\[6.45831pt]
=\min_{K_{1}}I^{f}\bigl{(}\rho_{12},(U_{1}\otimes
U_{2})^{\dagger}(K_{1}\otimes 1_{2})(U_{1}\otimes
U_{2})\bigr{)}\\\\[6.45831pt]
=\min_{K_{1}}I^{f}\bigl{(}\rho_{12},(U_{1}^{\dagger}K_{1}U_{1}\otimes
1_{2}\bigr{)}=\mathcal{U}^{\Delta,f}_{1}(\rho_{12}),\end{array}$
where we used the definition in (11). QED
###### Theorem.
The metric adjusted LQU is contractive under completely positive trace-
preserving maps on the non-measured subsystem.
###### Proof.
Let $\tilde{K}_{1}$ be the local observable minimizing the metric adjusted
skew information. A completely positive trace preserving map $\Phi_{2}$ on
system 2 is obtained as an amplification followed by a partial trace
(Stinespring dilation): ${\rm
Tr\hskip-1.99997pt}~{}_{3}(U_{23}\rho_{23}U_{23}^{\dagger})=\Phi_{2}\,\rho_{2}$.
The metric adjusted LQU is invariant under local unitaries. Also, the metric
adjusted skew information is contractive under partial trace. Calling $d_{3}$
the dimension of the Hilbert space of the ancillary system 3, one has
$\displaystyle\mathcal{U}_{1}^{\Lambda,f}(\rho_{12})=$
$\displaystyle\,I^{f}\left(\rho_{12},\tilde{K}_{1}\otimes
1_{2}\right)=I^{f}\left(\rho_{12}\otimes\frac{1}{d_{3}}1_{3},\tilde{K}_{1}\otimes
1_{23}\right)$ $\displaystyle=$ $\displaystyle\,\ I^{f}\left((1_{1}\otimes
U_{23})\left(\rho_{12}\otimes\frac{1}{d_{3}}1_{3}\right)(1_{1}\otimes
U_{23}^{\dagger}),\tilde{K}_{1}\otimes 1_{23}\right)$ $\displaystyle\geq$
$\displaystyle\,I^{f}\left({\rm
Tr\hskip-1.99997pt}~{}_{3}\left\\{(1_{1}\otimes
U_{23})\left(\rho_{12}\otimes\frac{1}{d_{3}}1_{3}\right)(1_{1}\otimes
U_{23}^{\dagger})\right\\},\tilde{K}_{1}\otimes 1_{2}\right)$ $\displaystyle=$
$\displaystyle\,I^{f}\left((1_{1}\otimes\Phi_{2})\rho_{12},\tilde{K}_{1}\otimes
1_{2}\right)$ $\displaystyle\geq$
$\displaystyle\,\mathcal{U}_{1}^{\Lambda,f}\bigl{(}(1_{1}\otimes\Phi_{2})\rho_{12}\bigr{)},$
as desired. QED
###### Theorem.
The metric adjusted LQU reduces to an entanglement monotone for pure states.
###### Proof.
The metric adjusted $f$-LQU coincides with the standard variance on pure
states, that is
$I_{\rho}^{f}(A)={\rm Var}_{\rho}(A)={\rm Tr\hskip-1.99997pt}~{}\rho
A^{2}-({\rm Tr\hskip-1.99997pt}~{}\rho A)^{2}$
whenever $\rho$ is pure (Hansen:2006b, , Theorem 3.8). But in GTA:2013 it has
been proved that the minimum local variance is an entanglement monotone for
pure states. QED
## IX Conclusion
In this work, we have built a unifying information-geometric framework to
quantify quantum correlations in terms of metric adjusted skew informations.
We extended the physically meaningful definition of LQU to a more general
class of information measures. Crucially, metric adjusted quantum correlation
quantifiers enjoy, by construction, a set of desirable properties which make
them robust information measures.
An important open question is whether information geometry methods may help
characterize many-body quantum correlations. In general, the very concept of
multipartite statistical dependence is not fully grasped in the quantum
scenario. In particular, we do not have axiomatically consistent and
operationally meaningful measures of genuine multipartite quantum discord.
Unfortunately, the LQU and IP cannot be straightforwardly generalized to
capture joint properties of more than two quantum particles. A promising
starting point could be to translate into the information-geometry language
the entropic multipartite correlation measures developed in weaving . We plan
to investigate the issue in future studies.
###### Acknowledgements.
This research is supported by a Rita Levi Montalcini Fellowship of the Italian
Ministry of Research and Education (MIUR), grant number 54$\\_$AI20GD01.
## References
* (1) D. Girolami, T. Tufarelli and G. Adesso. Characterizing Nonclassical Correlations via Local Quantum Uncertainty. Phys. Rev. Lett. 110, 240402 (2013).
* (2) D. Girolami, A. M. Souza, V. Giovannetti, T. Tufarelli, J. G. Filgueiras, R. S. Sarthour, D. O. Soares-Pinto, I. S. Oliveira and G. Adesso. Quantum discord determines the interferometric power of quantum states Phys. Rev. Lett. 112, 210401 (2014).
* (3) F. Hansen. Metric adjusted skew information. Proc. Nat. Acad. Sci. USA 105(29), pp 9909–9916, 2008.
* (4) R. Horodecki, P. Horodecki, M. Horodecki and K. Horodecki. Quantum Entanglement. Rev. Mod. Phys. 81, pp 865–942, 2009.
* (5) H. Ollivier and W. H. Zurek. Quantum Discord: A Measure of the Quantumness of Correlations. Phys. Rev. Lett. 88, 017901, 2001.
* (6) L. Henderson and V. Vedral. Classical, quantum and total correlations. J. Phys. A 34, 6899, 2001.
* (7) K. Modi, A. Brodutch, H. Cable, T. Paterek and V. Vedral. The classical-quantum boundary for correlations: discord and related measures. Rev. Mod. Phys. 84 (4), 1655, 2012.
* (8) A. Uhlmann. The metric of Bures and the geometric phase. In R. Gielerak et al. (eds.), Groups and related topics, Kluwer Academic Publishers, 1992.
* (9) I. Bengtsson and K. Zyczkowski. Geometry of Quantum States, Cambridge University Press, Cambridge, 2007.
* (10) N. Ay, P. Gibilisco, F. Matus (eds.) Information Geometry and its Applications, Springer Proceedings in Mathematics & Statistics, vol. 252, 2018.
* (11) M. Cianciaruso, I. Frérot, T. Tufarelli and G. Adesso. Characterising two-sided quantum correlations beyond entanglement via metric-adjusted $f$-correlations, pp 411–430 in AyGibiliscoMatus: 2018 .
* (12) F. F. Fanchini, D. O. Soares Pinto, G. Adesso (eds.). Lectures on General Quantum Correlations and their Applications, Springer, 2017.
* (13) P. Bogaert and D. Girolami. Metrological measures of non-classical correlations, pp 159–179 in FPA:2017 .
* (14) P. Gibilisco, D. Imparato and T. Isola. Inequalities for quantum Fisher information. Proc. Am. Math. Soc. 137(1), pp 317–327, 2009.
* (15) P. Gibilisco and T. Isola. How to distinguish quantum covariances using uncertainty relations. J. Math. Anal. App. 384, pp 670–676, 2011.
* (16) E. P. Wigner and M. M. Yanase. Information content of distributions. Proc. Nat. Acad. Sci. USA 49, pp 910–918, 1963.
* (17) V. Giovannetti, S. Lloyd and L. Maccone Advances in Quantum Metrology. Nature Phot. 5, 222, 2011.
* (18) D. Petz and R. Temesi. Means of positive numbers. SIAM J. Matrix Anal. Appl., 27, pp 712–720, 2005.
* (19) P. Gibilisco, F. Hansen and T. Isola. On a correspondence between regular and non-regular operator monotone functions. Lin. Alg. Appl. 430(8-9), pp 2225–2232, 2009.
* (20) F. Kubo and T.Ando. Means of positive linear operators. Math. Ann. 246(3), 205D224, 1979/80.
* (21) P. Gibilisco and F. Hansen An inequality for expectation of means of positive random variable. Annals of Functional Analysis 8(1), pp 142–151, 2017.
* (22) D. Petz and C. Sudár. Geometry of quantum states. J. Math Phys. 37, pp 2662–2673, 1996.
* (23) E. Fick and G. Sauermann. The quantum statistics of dynamic processes. Springer Series in Solid-State Sciences, 86. Berlin, 1990.
* (24) P. Gibilisco, D. Imparato and T. Isola. Uncertainty principle and quantum Fisher information II. J. Math. Phys 48, 072109, 2007.
* (25) P. Gibilisco, F. Hiai and D. Petz. Quantum covariance, quantum Fisher information and the uncertainty relations. IEEE Trans. Inform. Theory 55, 2009.
* (26) D. Petz. Covariance and Fisher information in quantum mechanics. J. Phys. A 35, 79D91, 2003.
* (27) G. Toth and D. Petz. Extremal properties of the variance and the quantum Fisher information. Phys. Rev. A 87, 032324, 2013.
* (28) N. N. Čencov Statistical decision rules and optimal inference. American Mathematical Society, Providence, R.I., 1982.
* (29) D. Petz. Monotone metrics on matrix spaces. Linear Algebra Appl. 244, pp 81–96, 1996.
* (30) L. Cai and F. Hansen. Metric adjusted skew information: Convexity and restricted forms of superadditivity. Lett. Math. Phys. 93, pp 1–13, 2010.
* (31) K. Audenaert, L. Cai and F. Hansen. Inequality for quantum skew information. Lett. Math. Phys. 85, pp 135–146, 2008.
* (32) M. Piani, P. Horodecki and R. Horodecki. No-local-broadcasting theorem for multipartite quantum correlations. Phys. Rev. Lett. 100(9), 90502, 2008.
* (33) S. Luo and Q. Zhang. Superadditivity of Wigner-Yanase-Dyson information revisited. J. Stat. Phys. 131, pp 1169–1177, 2008.
* (34) D. Girolami, T Tufarelli, and C. E. Susa Quantifying genuine multipartite correlations and their pattern complexity. Phys. Rev. Lett. 119, 140505, 2017.
|
# Computational strategies and estimation performance with Bayesian
semiparametric Item Response Theory models
Sally Paganin<EMAIL_ADDRESS>University of California, Berkeley
Christopher J. Paciorek University of California, Berkeley Claudia Wehrhahn
University of California, Santa Cruz Abel Rodríguez University of
Washington, Seattle Sophia Rabe-Hesketh University of California, Berkeley
Perry de Valpine University of California, Berkeley
###### Abstract
Item response theory (IRT) models typically rely on a normality assumption for
subject-specific latent traits, which is often unrealistic in practice.
Semiparametric extensions based on Dirichlet process mixtures offer a more
flexible representation of the unknown distribution of the latent trait.
However, the use of such models in the IRT literature has been extremely
limited, in good part because of the lack of comprehensive studies and
accessible software tools. This paper provides guidance for practitioners on
semiparametric IRT models and their implementation. In particular, we rely on
NIMBLE, a flexible software system for hierarchical models that enables the
use of Dirichlet process mixtures. We highlight efficient sampling strategies
for model estimation and compare inferential results under parametric and
semiparametric models.
Keywords: binary IRT models, Dirichlet process mixture, MCMC strategies,
NIMBLE.
## 1 Introduction
Traditional approaches in item response theory (IRT) modeling rely on the
assumption that subject-specific latent traits follow a normal distribution.
This assumption is often considered for computational convenience, but there
are many situations in which it may be unrealistic (Samejima, 1997). For
example, Micceri (1989) gives a comprehensive review of many psychometric
datasets where the distribution of latent individual trait does not respect
the normality assumption and presents instead asymmetries, heavy-tails or
multimodality. In addition, estimation of IRT parameters in the presence of
non-normal latent traits has been shown to produce biased estimates of the
parameters (see, for example Seong, 1990; Kirisci ., 2001; Schmitt ., 2006;
Finch Edwards, 2016).
Different proposals have been made in the general IRT literature for relaxing
this normality assumption, using either Markov chain Monte Carlo (MCMC) or
Marginal Maximum Likelihood (MML) estimation methods. One option is to rely on
more general parametric assumptions. For example, Azevedo . (2011) considered
a skew-normal distribution (Azzalini, 1985), while others have suggested
finite mixtures of normal distributions (Bolt ., 2001; Bambirra Gonçalves .,
2018). Alternatively, one can refrain from making distributional assumption on
the latent abilities by using nonparametric maximum likelihood estimation
(Laird, 1978; Mislevy, 1984), B-splines (Woods Thissen, 2006; Johnson, 2007)
or empirical histograms (Woods, 2007).
This paper considers a Bayesian nonparametric approach that uses a Dirichlet
process mixture (Ferguson, 1973; Lo, 1984; Escobar West, 1995) as a
nonparametric prior on the distribution of the subject-specific latent trait.
Dirichlet process mixtures are often used as flexible models to describe the
unknown distribution of an heterogeneous population of interest. These models
are sometimes interpreted as mixture models with an infinite number of
components. In practice these model treat the number of groups as an unknown
parameter and estimate it from the data, so that the model can easily account
for multi-modality, asymmetries or outliers in the latent trait distribution.
We focus in particular on semiparametric extensions of logistic IRT models for
binary responses. Such models are semiparametric because they retain other,
parametric, assumptions of binomial mixed models, such as the functional form
of the link function.
Even though some Bayesian nonparametric extensions of binary IRT models have
been presented in the literature, they have been given only limited
consideration. Within this approach, the semiparametric 1PL model has been the
focus of more effort as well as software (Jara ., 2011, DPpackage, no longer
actively maintained). San Martín . (2011) investigated a semiparametric
generalization of the 1PL model from a theoretical perspective, while Finch
Edwards (2016) provided results from simulation studies. An example using the
semiparametric 2PL model is given in Duncan MacEachern (2008). However, such
semiparametric models have not received much attention in applied IRT
modeling, in good part because of the lack of comprehensive studies and
accessible software tools.
The goal of this paper is to provide a practical guide to semiparametric IRT
models for both (i) applied researchers interested in using Dirichlet process
mixtures, and (ii) those familiar with Bayesian nonparametrics concepts who
are interested in IRT models. To achieve these goals, we fill three major gaps
that hinder widespread application of semiparametric Bayesian IRT models.
First, we implement the semiparametric 1PL, 2PL and 3PL models in NIMBLE (de
Valpine ., 2017) (R package nimble, de Valpine ., 2020), a flexible R-based
system for hierarchical modeling. In particular, NIMBLE provides functionality
for fitting hierarchical models that involve Dirichlet process priors either
via a Chinese Restaurant Process (CRP) (Aldous, 1985; Pitman, 1996; Blackwell
MacQueen, 1973) or a truncated stick-breaking (SB) (Sethuraman, 1994)
representation of the prior. Hence, NIMBLE supports a much wider class of
models than those that are implemented in standard software packages. Code is
provided for all examples in a publicly accessible GitHub repository
(https://github.com/salleuska/IRT_nimble_code).
Second, focusing on the 2PL model, we study the efficiency of several MCMC
sampling strategies in both simulated and real-data scenarios. We define
sampling strategies as the combination of model parameterization,
identifiability constraints and sampling algorithms, focusing on general MCMC
algorithms available in easy-to-access software tools for Bayesian
hierarchical models. We find that some choices of parameterization and
identifiability constraints can yield order-of-magnitude differences in
sampling efficiency compared to others. This approach also allows us to
compare various random walk Metropolis-Hastings MCMC strategies to the
Hamiltonian Monte Carlo (HMC) strategy implemented in the widely used Stan
package (Stan Development Team, 2018). Although there is Stan support for many
parametric IRT models (Bürkner, 2021; Furr, 2017), HMC algorithms are not
readily available for Dirichlet process prior models, since HMC cannot sample
discrete parameters (the component indicators), which cannot be easily
integrated out in infinite mixture models.
Finally, we present a comparison of inferential results for item and subject
parameters under parametric and semiparametric specifications. To make these
comparisons fair, we carefully elicit prior distributions for the models by
matching the prior predictive distribution of the data to a common
distribution (Berger Pericchi, 1996). We also illustrate how to estimate the
entire distribution of latent traits and its functionals under the two
specifications. As expected, we find that the semiparametric model improves
recovery of item and individual latent trait parameters in the case of non-
normal latent traits. More suprisingly, there seems to be little inferential
penalty in using a semiparametric model when a parametric model would be
correct, supporting the benefit of greater robustness to mis-specification.
These conclusions are based on analyses carried out on simulated data as well
as two real datasets related to education and medical assessments: the 2007
Trends in International Mathematics and Science Study (TIMSS) and the 1996
Health Survey for England. For both the real data examples, the semiparametric
model performs better than the parametric counterpart, with the semiparametric
model identifying distinct modes in the distribution of the latent trait
missed by the parametric model.
We note that other authors have considered nonparametric IRT models that rely
on a general monotonic function in place of the logistic/probit link function.
These models are sometimes referred as NIRT models. Some work using the
Dirichlet process falls in this class of models (Qin, 1998; Miyazaki Hoshino,
2009; Karabatsos, 2017). While we do not pursue this direction in this paper,
focusing instead on nonparametric modeling of the latent trait distribution,
such an extension is relatively straightforward.
The remainder of the paper is organized as follows. In Section 2 we present
the standard IRT model and the Bayesian semiparametric extension along with
considerations for identifiability. We then present different potential
sampling strategies (Section 3) and discuss the goals of our experiments. To
fairly compare the different strategies, we give guidance on selecting prior
distributions in Section 4. We introduce simulated and real-world data in
Section 5. Comparison of the results in terms of MCMC efficiency and
statistical inference is presented in Sections 6 and 7. In Section 8, we
conclude that having access to semiparametric models can be broadly useful, as
it allows inference on the entire underlying latent trait distribution and its
functionals, with NIMBLE being a flexible framework for estimation of such
models.
## 2 IRT models and background
IRT models are widely used in various social science disciplines to scale
binary responses into continuous constructs. For conciseness, in this section
we introduce model notation in the context of educational assessment, where
typically data are answers to exam questions from a set of individuals and the
latent trait is interpreted as an individual’s ability. In particular, let
$y_{ij}$ denote the answer of individual $j$ to item $i$ for $j=1,\ldots,N$
and $i=1,\ldots,I$, with $y_{ij}=1$ when the answer is correct and $0$
otherwise. Responses from different individuals are assumed to be independent,
while responses from the same individual are assumed independent conditional
on the latent trait (this is sometimes called the _local independence
assumption_ in the psychometric literature).
### 2.1 Binary logistic IRT models
Let $\pi_{ij}$ denote the probability that individual $j$ answers item $i$
correctly, given the model parameters $\eta_{j},\lambda_{i},\beta_{i}$; i.e.,
$\pi_{ij}=\Pr(y_{ij}=1\mid\eta_{j},\lambda_{i},\beta_{i})$ for $i=1,\ldots,I$
and $j=1,\ldots,N$. The parameter $\eta_{j}$ represents the latent ability of
the $j$-th individual, while $\beta_{i}$ and $\lambda_{i}$ encode the item
characteristics for the $i$-th item. In the two-parameter logistic (2PL)
model, the probability $\pi_{ij}$ is determined using the logistic function as
$\text{logit}(\pi_{ij})=\lambda_{i}(\eta_{j}-\beta_{i}),\quad
i=1,\ldots,I,\quad j=1,\ldots,N.$ (1)
A further assumption for the latent abilities is that they are independently
and identically distributed according to some distribution $G$,
$\eta_{j}\stackrel{{\scriptstyle iid}}{{\sim}}G,\quad j=1,\ldots,N,$ (2)
with $G$ traditionally a standard normal distribution. The parameter
$\lambda_{i}>0$ is often referred to as _discrimination_ , since items with a
large $\lambda_{i}$ are better at discriminating between subjects with similar
abilities, while $\beta_{i}$ is called _difficulty_ because for any fixed
$\eta_{j}$ the probability of a correct response to item $i$ is decreasing in
$\beta_{i}$.
Often, the log-odds in (1) are reparameterized as
$\lambda_{i}\eta_{j}+\gamma_{i},$
with $\gamma_{i}=-\lambda_{i}\beta_{i}$. The two parameterizations are
sometimes referred to as _IRT parameterization_ and _slope-intercept (SI)
parameterization_ , respectively. While the slope-intercept parameterization
is often considered for computational convenience, the IRT parameterization is
the most traditional in terms of interpretation. In exploring different
strategies for Bayesian estimation, we will consider both alternatives and
investigate potential differences in terms of computational performance.
Alternative models can be obtained by considering a different number of item
parameters. When $\lambda_{i}=1$ for all $i=1,\ldots,I$, the model in (1)
reduces to the one-parameter logistic (1PL) model, also known as Rasch model
(Rasch, 1990). In some settings one may wish to account for the probability of
answering correctly by chance, by introducing a third set of item parameters,
$\upsilon_{i},i=1,\ldots,I$, referred to as _guessing_ parameters, so that
$\Pr(y_{ij}=1\mid\eta_{j},\lambda_{i},\beta_{i},\upsilon_{i})=\upsilon_{i}+(1-\upsilon_{i})\text{expit}\\{\lambda_{i}(\eta_{j}-\beta_{i})\\},$
where $\text{expit}\\{\cdot\\}$ denotes the inverse of the logistic function.
This model is typically referred to as the three-parameter logistic model
(3PL), and is often relevant in educational assessments.
### 2.2 Semiparametric IRT models
The classical formulation of IRT models assumes that the latent abilities in
(2) follow a normal distribution. This assumption can be relaxed, modeling the
distribution of ability as a mixture of normal distributions, where the number
of mixture components does not need to be specified in advance but rather is
learned from the data. This can be achieved using a Dirichlet process mixture
(DPM) model for the distribution of ability. In particular, the distribution
of ability $G$ in (2) can be specified as a convolution involving a Dirichlet
process (DP) prior, i.e.
$\displaystyle G=\int\mathcal{K}(\eta_{j}\mid\theta)F(d\theta),\quad
F\sim\mbox{DP}(\alpha,G_{0}),$ (3)
where $\mathcal{K}(\cdot\mid\theta)$ is a suitable probability kernel indexed
by the parameter $\theta$, while $\alpha$ and $G_{0}$ are, respectively, the
concentration parameter and the base distribution of the Dirichlet process.
In the context of binary IRT models, it seems natural to choose a normal
kernel for $\mathcal{K}(\cdot\mid\theta)$, indexed by parameters
$\theta=\\{\mu,\sigma^{2}\\}$. This means the distribution of ability is a
mixture of normal distributions, where the number of mixture components and
their means and variances are unknown. Furthermore, under this choice, taking
$\alpha\to 0$ leads to the original parametric model discussed in Section 2.1,
in this case a single normal. Parameters characterizing each mixture component
are drawn from the base distribution, $G_{0}$. For computational convenience
the base distribution is typically the product of conjugate distributions,
e.g., a normal distribution for $\mu$ and an inverse-gamma distribution for
$\sigma^{2}$.
We proceed now to discuss the Dirichlet process prior in more detail. There
are two main representations of the Dirichlet process, each leading to a
different MCMC posterior sampling strategy, namely the stick-breaking
representation (SB) (Sethuraman, 1994) and the Chinese Restaurant Process
(CRP) (Aldous, 1985; Pitman, 1996; Blackwell MacQueen, 1973). In this work we
use the CRP representation. The CRP representation is derived from (3)
integrating out the random measure $F$. More specifically, let
$\theta_{1},\ldots,\theta_{N}$ be an independent sample from $F$, with some
values possibly repeated. Integrating over $F$ one can obtain the joint prior
distribution on $(\theta_{1},\ldots,\theta_{N})$, which can be written as the
product of a sequence of conditional distributions, where
$(\theta_{j}\mid\theta_{j-1},\ldots,\theta_{1})\sim\frac{\alpha}{\alpha+j-1}G_{0}+\sum_{l=1}^{j-1}\frac{1}{\alpha+j-1}\delta_{\theta_{l}},$
(4)
for $j=1,\ldots,N$, where $\delta_{a}$ is the Dirac probability measure
concentrated at $a$. The second term in (4) represents the probability that a
new observation is equal to one of the previous ones, while the first term
captures the possibility that we observe a new value, which would be drawn
from the base measure $G_{0}$.
The CRP name comes from an analogy often used to describe the process in (4).
Consider a Chinese restaurant with an infinite number of tables, each table
serving one dish shared by all customers sitting at that table. In this
metaphor, each table represents a possible mixture component, while each dish
represents the parameter indexing the distribution associated with the mixture
component. Customers entering the restaurant can seat themselves at a
previously occupied table and share the same dish (with probability
proportional to the number of customers already sitting at the table), or go
to a new table and order another dish (with probability proportional to
$\alpha$). The dishes are selected according to the centering distribution
$G_{0}$.
One way to make the Chinese restaurant analogy clearer is by reparameterizing
the model. Denote by $\theta_{k}^{*}$ the dish served in table $k$ (which is a
draw from $G_{0}$) and let $z_{j}$ be the variable denoting the table chosen
by the $j$th customer. Then
$\displaystyle p(z_{j}=k\mid
z_{j-1},\ldots,z_{2},z_{1},\alpha)=\begin{cases}\frac{n_{k}^{j-1}}{\alpha+j-1},\quad
k=1,\ldots,K^{j-1},\\\ \frac{\alpha}{\alpha+j-1},\quad
k=K^{j-1}+1,\end{cases}$ (5)
where $K^{j-1}$ is the total number of occupied tables by the first $j-1$
customers, and $n^{j-1}_{k}$ is the number of customers at table $k$ among the
first $j-1$. This new parameterization can be related to the old one by noting
that $\theta_{i}=\theta_{z_{i}}^{*}$. The concentration parameter $\alpha$
controls the distribution of the number of tables (components), with larger
values favoring more tables. Using the indicators
$\mathbf{z}=\\{z_{j},j=1,\ldots,N\\}$, we can denote by
$\mathbf{z}\mid\alpha\sim\mbox{CRP}(\alpha)$ the joint distribution induced by
(5), and rewrite the DPM model for the distribution of ability in (3) using
$\displaystyle\eta_{j}\mid z_{j},\theta^{*}_{1},\theta^{*}_{2},\ldots$
$\displaystyle\stackrel{{\scriptstyle
ind}}{{\sim}}\mathcal{K}(\cdot\mid\theta^{*}_{z_{j}}),\quad j=1,\ldots,N,$
$\displaystyle\mathbf{z}\mid\alpha$ $\displaystyle\sim\mbox{CRP}(\alpha),$
$\displaystyle\theta^{*}_{k}$ $\displaystyle\stackrel{{\scriptstyle
iid}}{{\sim}}G_{0},\quad k=1,2,\ldots.$
Together, the probability kernel, base measure, and CRP form a Dirichlet
process mixture. Alternatively, the distribution $F$ can be written using the
stick-breaking representation:
$F(\cdot)=\sum_{k=1}^{\infty}w_{k}\delta_{\tilde{\theta}_{k}},$
where $\tilde{\theta}_{1},\tilde{\theta}_{2},\ldots$ is a sequence of
independent draws from $G_{0}$ and the weights are constructed by letting
$w_{k}=v_{k}\prod_{l=1}^{k-1}(1-v_{l})$, with $v_{1},v_{2},\ldots$ being a
sequence of independent draws from a $\mbox{Beta}(1,\alpha)$ distribution.
This construction makes it clear that, as long as the kernel
$\mathcal{K}(\cdot\mid\theta)$ is continuous, the distribution of ability $G$
is also continuous, but $F$ is almost surely discrete, naturally inducing
clustering via repeats in the parameter indexing the distribution of ability.
Note that an alternative to the formulation described above is to model the
distribution of the ability $G$ directly using a DP, e.g., centered around a
normal distribution. Such a model also comprises the standard parametric model
as a limiting case (now, when $\alpha\to\infty$) and leads to slightly simpler
computational algorithms. However, we believe that such an approach has some
serious drawbacks in the context of most IRT applications. By definition,
realizations from a Dirichlet process are almost surely discrete. This
property has made the Dirichlet process a useful tool in clustering
applications. However, in our context, it implies that we believe that two (or
more) individuals potentially have exactly the same ability. Not only is this
assumption not realistic, but it potentially prevents us from distinguishing
individuals based on their abilities, which is one common goal in IRT
modeling. The use of a Dirichlet process mixture with a continuous kernel
(Gaussian, in this case) sidesteps this issue.
### 2.3 Identifiability and constraints
Without additional constraints, the parameters of the models presented in
Section 2 are not identifiable (e.g., see Geweke Singleton, 1981; Bafumi .,
2005, as well as Section A in the Supplementary Materials). For example in the
2PL and 3PL models, increasing all $\eta_{j}$ and $\beta_{i}$ values by the
same amount yields the same probabilities in (1) for all $i$ and $j$. More
generally, the ability parameters are known up to a linear transformation, and
constraints are needed to identify them. To address this problem, traditional
work on parametric IRT models assumes that latent abilities in (2) come from a
standard normal distribution, i.e., $G\equiv\mathcal{N}(0,1)$, and constrains
the discrimination parameters $\lambda_{i}$ for $i=1,\ldots,I$ to be positive.
Alternative constraints can also establish identifiability and could yield
different computational performance for MCMC sampling. A common alternative
considers sum-to-zero constraints for the item parameters (Fox, 2010)
$\sum_{i=1}^{I}\beta_{i}=0,\quad\left(\text{or
}\sum_{i=1}^{I}\gamma_{i}=0\right),\quad\textbf{}\sum_{i=1}^{I}\log(\lambda_{i})=0.$
(6)
Centering the difficulty parameters addresses the invariance to translations,
while centering the log of the discrimination parameters (setting their
product to one) accounts for the invariance to rescalings of the latent space.
Another potential set of constraints, popular in political science
applications, involves fixing the value of the latent traits for two
individuals (e.g., see Clinton ., 2004). Whatever the set of constraints, it
is worthwhile to note that they can be either directly incorporated in the
model as part of the prior (and, therefore in the structure of the sampling
algorithms), or they can be applied as a postprocessing step (after running an
unconstrained MCMC). This last approach is typical of parameter-expanded
algorithms, which embed the target model in a larger specification. Parameter
expansion has been proposed in the literature to accelerate EM (C. Liu .,
1998) and Gibbs sampler (JS. Liu Wu, 1999) convergence, as well as to induce
new classes of priors (Gelman, 2004). Although targeting the same posterior,
constrained priors and parameter expansion can lead to very different results
in terms of convergence and mixing of the MCMC algorithms.
Similar arguments apply for the semiparametric extensions using the Dirichlet
process mixture. In that setting, one identifiability strategy may be to
constrain the base distribution $G_{0}$, e.g., by letting
$G_{0}\sim\mathcal{N}(0,1)$ (for example, see Duncan MacEachern, 2008).
However, even if the prior expectation and variance of $G_{0}$ are zero and
one, the corresponding posterior quantities can deviate substantially from
these values, leading to biased inference (Yang Dunson, 2010). More general
centering approaches have been proposed in the literature when a DP
distribution is used to model random effects or latent variables in a
hierarchical model (Yang ., 2010; Yang Dunson, 2010; Li ., 2011). These
approaches rely on parameter expansion by sampling from the unconstrained DP
model and then applying a post-processing procedure to the posterior samples.
This post-processing procedure requires the analytical evaluation of the
posterior mean and variance of the DP random measure, with Li . (2011)
providing results under the CRP representation and Yang . (2010) under the
stick-breaking one. Although such strategies are useful for general
hierarchical models to avoid identifiability issues, for the semiparametric
1PL, 2PL, and 3PL models it is simpler to use the sum-to-zero constraints on
the item parameters in (6), and that is the approach we adopt in this work. As
in the parametric case, we can either include these constraints in the prior
or use the parameter expansion approach for sampling and then center and
rescale the posterior samples as appropriate.
## 3 Sampling strategies for logistic IRT models
In this work we explore different _sampling strategies_ for Bayesian
estimation of the logistic parametric and semiparametric IRT models. We define
a _sampling strategy_ to include the combination of model parameterization,
identifiability constraints and sampling algorithms. We focus on the case of
the 2PL model, as it contains the 1PL as a special case and presents the same
identifiability challenges as the 3PL model. The strategies considered are
summarized in Table 1.
Identifiability constraints Parametric Semi-parametric Slope-intercept IRT
Slope-intercept IRT Constrained abilities MH/conjugate MH/conjugate Centered
HMC (Stan) Constrained item parameters MH/conjugate MH/conjugate∗ MH/conjugate
MH/conjugate∗ Unconstrained MH/conjugate MH/conjugate MH/conjugate
MH/conjugate Centered Centered
Table 1: Summary of the 14 sampling strategies considered for the parametric
and semiparametric 2PL model. Each of the 14 entries is a different strategy
with “MH/conjugate”, “Centered”, and “HMC (Stan)” referring to three different
sampling algorithms discussed below. The asterisk symbol denotes the sampling
strategies that lead directly to samples parameterized as model (3). Others
need post-processing to correspond to model (3).
We explore both parameterizations of the 2PL model mentioned in Section 2.1:
the IRT and the slope-intercept parameterization. To compare estimates
obtained from different parameterizations on a common scale, we post-process
posterior samples (using transformations described in the Supplementary
Materials, Section A) to respect the following base parameterization
$\displaystyle\text{logit}(\pi_{ij})$
$\displaystyle=\lambda_{i}(\eta_{j}-\beta_{i}),\quad i=1,\ldots,I,$
$\displaystyle\sum_{i=1}^{I}\log(\lambda_{i})$
$\displaystyle=0\quad\sum_{i=1}^{I}\beta_{i}=0,$ $\displaystyle\eta_{j}$
$\displaystyle\sim G,\quad j=1,\ldots,N,$ (7)
where $G$ denotes a general distribution for the latent abilities, either
parametric or nonparametric. The model in (3) follows the IRT parameterization
with sum-to-zero identifiability constraints, which is typically the target
one for inference for interpretability reasons.
As discussed in Section 2.3, our target inferential model in (3) can be
estimated directly, accounting for identifiability constraints. This can be
achieved by introducing in the model formulation a set of auxiliary item
parameters, $\\{\lambda_{i}^{\prime},\beta^{\prime}_{i}\\}$ for each
$i=1,\ldots,I$ and defining $\\{\lambda_{i},\beta_{i}\\}$ as
$\displaystyle\log(\lambda_{i})$
$\displaystyle=\log(\lambda^{\prime}_{i})-\frac{1}{I}\sum_{i=1}^{I}\log(\lambda^{\prime}_{i})\quad\beta_{i}=\beta^{\prime}_{i}-\frac{1}{I}\sum_{i=1}^{I}\beta^{\prime}_{i},\quad
i=1,\ldots,I.$ (8)
In Table 1 we label this model as the _constrained item parameters model_.
Unconstrained priors are then placed on the auxiliary parameters
$\\{\lambda_{i}^{*},\beta^{*}_{i}\\}$. The same formulation applies under the
slope-intercept parameterization, where the sum-to-zero constraints are placed
on the pairs $\\{\log(\lambda_{i}),\gamma_{i}\\}$ for $i=1,\ldots,I$.
Alternatively, we can consider _unconstrained_ versions of the 2PL model,
treating the unconstrained model as a parameter-expanded version of the target
inferential model in (3), where the redundant parameters are the means of the
difficulty and log discrimination parameters. Hence we conduct MCMC sampling
with known model unidentifiability, and before using the results for
inference, we transform samples to follow the model in (3), as described in
the Supplementary Materials, Section A.
Finally, for the parametric case only, we consider the traditional version of
the 2PL model that assumes the ability parameters $\eta_{j}$ for
$j=1,\ldots,N$ follow a standard normal distribution (_constrained abilities
model_).
### 3.1 Sampling algorithms
We focus on general MCMC sampling algorithms available in easy-to-access
software tools for Bayesian hierarchical models, such as NIMBLE and Stan, that
can flexibly accommodate different choices of prior distributions and link
functions. In this section we give an overall description of the algorithms
considered for the different sampling strategies. We refer to the
Supplementary Materials, Section B, for a detailed summary of the samplers
used for each parameter.
For both the parametric and semiparametric models we consider NIMBLE’s default
sampling configuration (_MH/conjugate algorithm_). NIMBLE’s MCMC uses an
overall one-at-a-time sampling strategy, cycling over individual parameters,
or parameter blocks for parameters with a multivariate prior. By default,
specific sampler types are assigned to the parameters or parameter blocks, but
the user can choose to change sampler types, control blocking strategies, and
modify details of sampling algorithm behavior. NIMBLE’s default MCMC
configuration assigns a conjugate (sometimes called “Gibbs”) sampler where
possible, sampling from the corresponding full conditional posterior
distribution. For non-conjugate continuous-valued parameters, NIMBLE’s default
sampler assignment is an adaptive random walk Metropolis-Hastings. For the
parametric versions of the 2PL model, the strategies using the default NIMBLE
assignments (_MH/conjugate algorithm_) correspond to these conjugate and
adaptive random walk Metropolis-Hastings samplers, with the latter also used
for most parametric components of the semiparametric 2PL. Specialized samplers
are assigned when Bayesian nonparametric priors are considered in the
semiparametric 2PL.
In the case of the slope-intercept parameterization, we take advantage of
NIMBLE’s flexibility to include user-programmed custom samplers (_centered
sampler_). The proposed centered sampler uses an adaptive random walk
Metropolis-Hastings sampler with a joint proposal for each pair of item
parameters $\\{\lambda_{i},\gamma_{i}\\}$ for $i=1,\ldots,I$, thereby
accounting for their posterior correlation. The proposal is made under a
reparameterization of the model that centers the abilities to have mean zero.
Implementation details are provided in the Supplementary Materials, Section B.
Finally, in the parametric setting only, we consider a Hamiltonian Monte-Carlo
(HMC) algorithm, as implemented in the Stan software (Carpenter ., 2017). Stan
implements an adaptive HMC sampler (Betancourt ., 2017) based on the No-U-Turn
sampler (NUTS) of Hoffman Gelman (2014). HMC algorithms are known to produce
samples that are much less autocorrelated than those of other samplers but at
more computational cost given the need to calculate the gradient of the log-
posterior. In this work, we limit the comparison to the IRT parameterization
with constraints on the abilities distribution, as that is the model provided
in the edstan R package (Furr, 2017).
### 3.2 Aims
In the remainder of the paper, we study the efficiency of the MCMC sampling
strategies in Table 1 to fit binary logistic IRT models, and we compare
inferential results under parametric and semiparametric specifications. Using
both simulated and real-world data, we aim to answer the following questions:
1. Q.1
For the parametric binary logistic IRT model, which of the Metropolis-
Hastings-based MCMC sampling strategies in Table 1 are most efficient? Do
different strategies work better in different scenarios for the distribution
of ability?
2. Q.2
For the parametric binary logistic IRT model, how does efficiency of random
walk Metropolis-Hastings sampling compare to Hamiltonian Monte Carlo (HMC), as
implemented in the popular Stan package? This question is of interest because
HMC is not readily available for semiparametric models using Dirichlet process
priors.
3. Q.3
How does MCMC efficiency of a semiparametric model compare to that of a
parametric one? Does this comparison differ when the parametric model is
correctly vs. incorrectly specified?
4. Q.4
To what degree does the use of a parametric model when its assumptions are
violated yield bad inference? Does use of a semiparametric model change
inference even when a parametric model would be valid?
5. Q.5
How much do results differ between the semiparametric and parametric models
for the real data examples?
In Section 6 we discuss the choice of efficiency metrics to address Q.1-Q.3
and present the results obtained using simulated and real-world data. In
Section 7, we discuss differences in the inferential results to investigate
Q.4-Q.5.
## 4 Choice of prior distributions
Past research on Bayesian IRT models has warned about the use of either vague
priors or highly informative priors when there is little information about the
parameters (Sheng, 2010; Natesan ., 2016). In particular Natesan . (2016)
investigated the use of different prior choices in 1PL and 2PL models using
MCMC and variational Bayes algorithms and found that the use of vague priors
tends to produce biased inference or convergence issues. Similarly, it is well
known that highly informative prior distributions on parameters can strongly
affect model comparison procedures.
To ensure a fair comparison between results from different strategies, we
chose the parameters of the priors in such a way that the induced prior
predictive distribution of the data is similar across all the different model
parameterizations. This “predictive matching approach” has been widely used to
guide prior elicitation in model comparison settings (Berger Pericchi, 1996;
Bedrick ., 1996; Ibrahim, 1997).
In the context of binary logistic IRT models, we aim to match the prior
marginal predictive distribution of a response $y_{ij}$, which in turn can be
achieved by matching the induced prior distribution on the marginal prior
probability of a correct response,
$\pi_{ij}=\mbox{expit}\\{\lambda_{i}(\eta_{j}-\beta_{i})\\}$. Note that all
the priors discussed in this paper are separately exchangeable, which means
that this prior marginal will be the same for any values of $i$ and $j$. In
particular, we attempt to match a $\mbox{Beta}(0.5,0.5)$ distribution, which
is both the reference and the Jeffreys prior for the Bernoulli likelihood in
the fully exchangeable case (Bernardo, 1979; Berger ., 2009). A similar
approach to prior elicitation in the context of latent space models for
networks can be found in Guhaniyogi Rodriguez (2020) and Sosa Rodrìguez
(2021). Because there are no analytical expressions available for the prior
distribution of $\pi_{ij}$, we use simulations to estimate the shape of the
prior distribution and obtain an approximate match. This is facilitated by our
implementation in NIMBLE. Indeed, one of the advantages of the NIMBLE system
is that it provides a seamless way to simulate from the model of interest.
Histograms of samples from the resulting induced priors can be seen in Figure
1 for a set of parametric and semiparametric models. Further details are
presented in the following subsections.
Figure 1: Histogram of samples from the induced prior on $\pi_{ij}$ under each
of the considered models. Dashed line indicates the density function of a
$\mbox{Beta}(0.5,0.5)$ distribution. Samples for the semiparametric models use
a prior distribution $\mbox{Gamma}(2,4)$ for the DP concentration parameter
$\alpha$, but similar results are obtained under the other settings presented
Section 4.
### 4.1 Priors for the item parameters
In Bayesian IRT modeling, normal distributions are typically chosen as priors
for the item parameters. This is true under both parameterizations. In
addition, the discrimination parameters, $\\{\lambda_{i}\\}_{i=1}^{I}$, are
typically assumed positive, so we consider a normal distribution on the log-
scale. To summarize, priors on the item parameters are:
$\log{\lambda_{i}}\sim\mathcal{N}(\mu_{\lambda},\sigma^{2}_{\lambda}),\quad\beta_{i}\sim\mathcal{N}(0,\sigma^{2}_{\beta}),\quad\gamma_{i}\sim\mathcal{N}(0,\sigma^{2}_{\gamma})\quad
i=1,\ldots,I.$
By default, we center on the difficulty parameters $\beta_{i}$ (or the
reparameterized version $\gamma_{i}$) on 0 for $i=1,\ldots,I$, while we set
$\sigma^{2}_{\beta}=\sigma^{2}_{\gamma}=3$. For the discrimination parameters,
we set $\mu_{\lambda}=\sigma^{2}_{\lambda}=0.5$ such that the prior
probability mass on the original scale is mostly in the range $(0.5,2.5)$.
### 4.2 Priors for the distribution of ability
In choosing priors for the abilities, we distinguish between the parametric
and semiparametric cases. In the parametric case, excluding the strategies in
which the distribution is a standard normal, we assume
$G\equiv\mathcal{N}(\mu_{\eta},\sigma^{2}_{\eta})$. We specify hyperpriors for
the unknown mean and variance, using a normal distribution for the mean
$\mu_{\eta}\sim\mathcal{N}(0,3)$, and an inverse-gamma distribution for the
variance, $\sigma^{2}_{\eta}\sim\mbox{InvGamma}(2.01,1.01)$ as in Paulon .
(2018), with hyperparameter values implying an a priori marginal expected
value of $1$ and an a priori variance equal to $100$.
In the semiparametric case, we need to specify the base distribution $G_{0}$
of the DP mixture prior along with the hyperparameters. We choose
$G_{0}\equiv\mathcal{N}(0,\sigma^{2}_{0})\times\mbox{InvGamma}(\nu_{1},\nu_{2})$
where $\mbox{InvGamma}(\nu_{1},\nu_{2})$ denotes an inverse-gamma distribution
with shape parameter $\nu_{1}$ and mean $\nu_{2}/(\nu_{1}-1)$. In choosing
values for the hyperparameters $\\{\sigma^{2}_{0},\nu_{1},\nu_{2}\\}$, we
first considered the concentration parameter $\alpha$ as fixed and evaluated
the induced prior distribution on $\boldsymbol{\pi}$ for values of
$\alpha\in\\{0.01,0.05,0.5,1,1.5,2\\}$. Recall that $\alpha$ controls the
prior expectation and variance of the number of clusters induced by the DP,
which are both of the order $\alpha\log(N)$. We discuss prior choice for the
$\alpha$ in Section 4.3. As in the parametric case, we center the normal
distribution for the mixture component means on $0$ with $\sigma_{0}^{2}=3$
and set $\nu_{1}=2.01$ and $\nu_{2}=1.01$ for the inverse-gamma distribution.
Given these settings, we found that choosing
$\alpha\in\\{0.01,0.05,0.5,1,1.5,2\\}$ does not have much effect on the
marginal prior distribution of the $\pi_{ij}$s.
### 4.3 Prior on the DP concentration parameter
One may be interested in placing a prior distribution on the concentration
parameter $\alpha$ of the Dirichlet process. A typical choice for the DP
concentration parameter is a $\mbox{Gamma}(a,b)$, with shape $a>0$ and scale
$b>0$, due to its computational convenience (Escobar West, 1995). As
previously stated, the concentration parameter controls the prior distribution
of the number of clusters (Escobar West, 1995; SJ. Liu, 1996). In choosing
values $a$ and $b$, we considered the implied prior mean and variance of the
number of clusters.
Let $K_{N}$ denote the number of clusters for a sample of size $N$. Results
from Antoniak (1974) and SJ. Liu (1996) show that the expected value and
variance of $K_{N}$ given $\alpha$ is
$\mathbb{E}(K_{N}\mid\alpha)=\sum_{i=1}^{N}\frac{\alpha}{\alpha+N-i},\quad\mathbb{V}ar(K_{N}\mid\alpha)=\sum_{i=1}^{N}\frac{\alpha(i-1)}{(\alpha+N-i)^{2}}.$
(9)
We exploit these results to choose values $a$ and $b$ that lead to reasonable
a priori values for the moments of the number of clusters for each of our
applications. For a given N and for different values of $a$ and $b$, we
evaluated the marginal expectation and variance of the quantities in (9) via
Monte Carlo approximation. We sample $\alpha_{r}$ for $r=1,\ldots,R$ from its
prior and compute
$\widehat{\mathbb{E}}(K_{N})=\frac{1}{R}\sum_{r=1}^{R}\mathbb{E}[K_{N}\mid\alpha_{r}],\quad\widehat{\mathbb{V}ar}(K_{N})=\frac{1}{R}\sum_{r=1}^{R}\mathbb{V}ar(K_{N}\mid\alpha_{r})+\widehat{\mathbb{V}ar}\left(\mathbb{E}[K_{N}\mid\alpha]\right),$
where
$\widehat{\mathbb{V}ar}\left(\mathbb{E}[K_{N}\mid\alpha]\right)=R^{-1}\sum_{r=1}^{R}\left[\mathbb{E}[K_{N}\mid\alpha_{r}]-\widehat{\mathbb{E}}[K_{N}]\right]^{2}$.
We explored a few prior choices and tabulate approximated moments in Table 2,
for the values of $N$ in our datasets. We consider the popular choice of
$a=2$, $b=4$ for the hyperparameters as in Escobar West (1995) along with
values favoring a small number of clusters ($a=1,b=3$) and values leading to a
more vague prior ($a=1,b=1$). For our applications we decided to favor a
relatively small number of clusters, choosing $a=2,b=4$ as hyperparameters for
the simulated data, and $a=1,b=3$ for the real-world data.
$\alpha\sim Gamma(a,b)$ | $\mathbb{E}[\alpha]$ | $\mathbb{V}ar(\alpha)$ | $\widehat{\mathbb{E}}(K_{2,000})$ | $\widehat{\mathbb{V}ar}(K_{2,000})$ | $\widehat{\mathbb{E}}(K_{14,525})$ | $\widehat{\mathbb{V}ar}(K_{14,525})$ | $\widehat{\mathbb{E}}(K_{7,377})$ | $\widehat{\mathbb{V}ar}(K_{7,377})$
---|---|---|---|---|---|---|---|---
$a=2,b=4$ | $0.5$ | $0.12$ | $4.7$ | $9.3$ | $5.6$ | $14.12$ | $5.3$ | $12.2$
$a=1,b=3$ | $0.3$ | $0.11$ | $3.5$ | $7.6$ | $4.2$ | $11.76$ | $3.9$ | $10.2$
$a=1,b=1$ | $1$ | $1$ | $7.8$ | $43.7$ | $9.8$ | $73.35$ | $9.2$ | $64.5$
Table 2: Approximate expectation and variance of the a priori number of
clusters, $K_{N}$, under different choices of the concentration parameter
distribution, for $N=\mbox{2,000, 14,525, 7,377}$ as in our data example
presented in Section 5.
## 5 Data examples
### 5.1 Synthetic data
We specify three different simulation scenarios for the distribution of
ability: _unimodal_ , _bimodal_ and _multimodal_ distributions. For all the
scenarios, we simulate responses from $N=2,000$ individuals on $I=15$ binary
items. Values for the discrimination parameters $\\{\lambda_{i}\\}_{i=1}^{15}$
are sampled from a $\mbox{Uniform}(0.5,1.5)$ distribution, while values for
difficulty parameters $\\{\beta_{i}\\}_{i=1}^{15}$ are taken to be equally
spaced in $(-3,3)$. We center the log of the discrimination parameters on
zero, while the difficulty parameters are already centered based on how they
are generated. We consider three different underlying distributions for the
latent abilities, $\eta_{j}$ for $j=1,\ldots,2,000$. In the _unimodal_
scenario, latent abilities are generated from a normal distribution with mean
$0$ and variance $(1.25)^{2}$. In the _bimodal_ scenario, we use a equal-
weights mixture of two normal distributions with means $\\{-2,2\\}$ and common
variance $(1.25)^{2}$. Finally, for the _multimodal_ scenario, latent
abilities are generated from the following mixture
$\frac{1}{5}\mathcal{N}(-2,1)+\frac{2}{5}\mathcal{N}(0,(0.5)^{2})+\frac{2}{5}\mathcal{SN}(3,1,-3),$
where $\mathcal{SN}(\xi,\omega,\zeta)$ indicates a skew-normal distribution
(Azzalini, 1985) with location parameter $\xi$, scale parameter $\omega>0$ and
parameter $\zeta$ that controls the asymmetry of the distribution.
As a sensitivity analysis, we considered other values of $N$ and $I$,
simulating data for each scenario following a factorial design with
$I\in\\{10,30\\}$ and $N\in\\{1,000,5,000\\}$. We discuss efficiency results
of the different sampling strategies in Section 6 and report results in the
Supplementary Materials, Section D.
### 5.2 Real world data
The first example is a subset of data from the 1996 Health Survey for England
(Joint Health Surveys Unit of Social and Community Planning Research and
University College London, 2017), a survey conducted yearly to collect
information concerning health and behavior of households in England. In
particular, we have data for 10 items measuring Physical Functioning (PF-10),
which is a sub-scale of the SF-36 Health Survey (Ware, 2003) administered to
people aged $16$ and above. In this case the latent trait quantifies the
physical status of a given individual (McHorney ., 1997; Hays ., 2000).
Participants in the survey were asked whether they perceived limitations in a
variety of physical activities (e.g., running, walking, lifting heavy objects)
and if so the degree of limitation. We list the original questions in the the
Supplementary Materials, Section C. Answers to items comprised three possible
responses (“yes, a lot”, “yes, limited a little”, “no, not limited at all”);
however, in our analysis we consider the dichotomous indicator for not being
limited at all. The left panel of Figure 2 shows the distribution of raw
scores, i.e. the total of correct answers. We consider the 2PL model for this
data, as it reasonable to assume that some of the questions are more
informative in defining individuals with high physical impairment (see
Supplementary Materials, Section C). For simplicity, we analyzed complete case
data from $14,525$ individuals out of $15,592$ respondents, although the model
can easily accommodate missing data.
The second example uses data from the TIMSS (Trends in International
Mathematics and Science Study) survey, which is an international comparative
educational survey dedicated to improving teaching and learning in mathematics
and science for students around the world
(http://timssandpirls.bc.edu/TIMSS2007/about.html). We used data from the 2007
eighth-grade mathematics assessment for the United States (N = $7,377$),
publicly available at https://timssandpirls.bc.edu/TIMSS2007/idb_ug.html. The
dataset comprises $214$ items, with $192$ of them dichotomous, while the
remaining $22$ have three category responses (“incorrect”, “partially
correct”, “correct”). We dichotomized these latter questions, considering
partially correct answers as incorrect ones. Like other large-scale
assessments, participants in TIMSS only received a subset of the items
according to a booklet design, resulting in $28$-$32$ item responses per
student. Distribution of the raw scores for the data are shown in the right
panel of Figure 2. As for the previous example, it is reasonable to assume
that some items discriminate differently between students with high and low
ability. However, in the context of educational testing, the 3PL model is
often considered because it accounts for the probability of answering
correctly by chance. Hence we consider both the 2PL and the 3PL models when
evaluating the different strategies.
Figure 2: Distribution of the raw scores (total of correct answers) for the
real data examples: health data (left panel) and TIMSS data (right panel).
Finally, we note that participants in the survey were sampled according to a
complex two-stage clustered sampling design that we did not consider in our
application. In other contexts the design is typically taken into account
using sampling weights for model estimation, as discussed for example by
Rutkowski . (2010).
## 6 Comparing results in terms of efficiency
MCMC performance is often evaluated in terms of mixing, often by calculating
the effective sample size (ESS), which is the equivalent number of independent
samples that would contain the same statistical information as the actual non-
independent samples. However, comparison between different MCMC algorithms
based solely on mixing can be misleading, as different samplers can vary
greatly in terms of computational cost (Nguyen ., 2020). Hence, it is
appropriate to consider ESS per computation time (in seconds), the rate at
which effectively independent samples are generated. A second issue is how to
combine ESS results for multiple parameters. For this purpose, we used the
_multivariate ESS_ (mESS) recently introduced by Vats . (2019), which accounts
for cross-correlations among parameters.
Computation time is typically measured for the actual MCMC run, not accounting
for steps to prepare for a run, thereby focusing on the algorithms of interest
rather than unrelated aspects of the software. Comparison between HMC and MCMC
algorithms raises the question of how to fairly account for computation times,
given that these two classes of algorithms use different tuning phases. Since
there is not an established way to compare these two algorithms in the
literature, we decided to consider different timings when using the two
algorithms: (i) _sampling time_ , which accounts only for the time to draw the
posterior samples, hence discarding the time needed for the burn-in and warm-
up phases of the two algorithms and (ii) _total time_ comprising also the
burn-in and warm-up phases. Although one can use alternative metrics for the
comparison, this choice can provide interesting and useful insights. When
computing efficiency based on the sampling time, we can assess pure efficiency
of sampling from the posterior. Using total time accounts for potentially
different times needed for warm-up/burn-in by the different algorithms but
introduces the difficulty of determining the optimal burn-in/warm-up time,
which we avoided here in favor of using basic defaults.
Estimate of the mESS is based on the multivariate batch means estimator as
described in Vats . (2019) and implemented in the mcmcse (Flegal ., 2021)
package. Since we used different specifications for the distribution of
ability across the different sampling strategies, we calculate the mESS
considering only the common parameters (i.e., the item parameters and sampled
abilities) after transforming samples to the parameterization of our target
inferential model (3). The mESS provides a single scalar measure of joint
mixing for all the parameters of interest in a model, but it does not
necessarily reflect univariate ESS values of each parameter. For example, mESS
can be larger than all the univariate ESS values (see Supplementary Materials,
Section D, Figures 16-18 for some insights). Given this, it can be useful as a
simple overall performance metric but does not replace ESS for specific
parameters of interest.
Given the large number of examples and sampling strategies, we performed a
preliminary experiment to choose the number of iterations and number of burn-
in or warm-up samples. In particular, for a portion of the simulations we used
multiple runs to determine the number of iterations and samples needed to
obtain reliable estimates of the mESS. For all MCMCs using NIMBLE, we decided
to use a total of $50,000$ iterations, with a $10\%$ burn-in of $5,000$ for
all examples. When running the HMC algorithm via Stan, we used a total of
$15,000$ iterations, with the first $5,000$ iterations as warm-up steps.
However, we observed highly variable values of the mESS estimates when using
the HMC algorithm. We also found that values of mESS are correlated with those
of the HMC tuning parameters (see Supplementary Materials, Section D). Given
this variability in mixing performance, we decided to limit comparisons with
HMC to the simulation scenarios, reporting results from the run with the
median ESS across multiple replications. All the models were estimated using a
Linux cluster with 4 nodes having 24 cores and 128 GB RAM per node (Intel(R)
Xeon(R) CPU E5-2643 v2 @ 3.50GHz). Across simulations, the running times
ranged between 15-94 minutes for parametric models and between 37-124 minutes
for semiparametric ones. For the data applications, running times ranged
between 85-384 minutes for the parametric models, and 410-610 minutes for
semiparametric ones.
### 6.1 Efficiency results for simulated data
For the three simulation scenarios we estimated the 2PL parametric model using
the different sampling strategies summarized in Table 1. Figure 3 compares
efficiency for all these strategies using the multivariate ESS per second,
computed with respect to both the total and sampling time.
Figure 3: Multivariate ESS per second for various sampling strategies used to
estimate the 2PL parametric model for the unimodal (left column), bimodal
(middle column), and multimodal (right column) scenarios. Results are computed
using total time (top row) and sampling time (bottom row). Note that SI stands
for “slope-intercept”.
Using different time baselines when computing efficiency changes the ranking
of the MCMC algorithms for the unimodal and multimodal simulation,
highlighting the trade-off for the HMC algorithm between sampling efficiency
and computational cost. While the HMC is highly efficient in producing samples
with low correlation, warm-up steps are computationally expensive. Recall that
efficiency values presented for the HMC strategy are relative to a median
performance across multiple runs. Amongst the non-HMC strategies,
unconstrained scenarios generally mix well, as do scenarios with constraints
on the abilities. However, imposing constraints on the item parameters
directly in the sampling performs poorly because obtaining each sample is
time-consuming. This is because the constraints in (8) require calculation of
all the likelihood terms for each parameter update, whereas for other
strategies only the likelihood terms for individuals’ responses on the item
under consideration need to be calculated. While incorporating constraints on
the item parameters is time-consuming, such a strategy could be useful in more
complicated hierarchical models, in particular when it is unclear how to
rescale posterior samples. The centered strategy for the slope-intercept
parameterization seems to have little impact across the scenarios.
Moving to the semiparametric models, recall that we did not consider
identifiability constraints on the abilities. We either included
identifiability constraints on the item parameters in the sampling or sampled
from the unconstrained model and rescaled the posterior samples.
Figure 4: Multivariate ESS per second (computed using the total time) for
semiparametric 2PL models (bottom row) in comparison with their parametric
version (top row) under the two simulation scenarios. Note that SI stands for
“slope-intercept”.
As expected when using a more complicated model, we observed some reduction in
efficiency in the semiparametric model compared to the parametric model, but
not a drastic one (Figure 4). Results for the semiparametric case are similar
in relative terms, but not in absolute magnitudes, when comparing the
different parameterizations and constraints.
For the non-HMC strategies, we also looked at how different combinations of
numbers of items and individuals affect efficiency of the different sampling
strategies (as discussed in Section 5.1), with results shown in the
Supplementary Materials, Section D. We found that the ranking across
strategies is generally stable across the different scenarios; strategies
using constraints on items are the worst overall, and the benefit of using
other strategies is most evident when the number of individuals $N$ is low.
### 6.2 Efficiency results for real-world data
We did similar comparisons using the real-data examples (Figure 5), noting
that we excluded the constrained items strategy, given its poor performance on
the simulated datasets, and the HMC strategy, because of the high variability
in mixing performance.
Figure 5: Multivariate ESS per second (computed using the total time) for the
health data (left column) parametric and semiparametric 2PL models, and TIMSS
data for the 2PL (middle column) and 3PL (right column) models. Note the
scales are different for the different rows and that SI stands for “slope-
intercept”.
The efficiency is lower than for the simulated datasets because the real data
have many more individuals or items, and therefore more parameters. The same
consideration applies when comparing efficiency of the 2PL and 3PL models for
the TIMSS data. We also noted that for the health data, when considering the
posteriors for the abilities in the semiparametric model, we saw evidence for
multimodality and some difficulty moving between modes for individuals with
high raw scores. The multimodality is likely related to it being difficult for
the semiparametric model to identify the exact magnitude of the ability for
such individuals. The use of more informative priors, with careful elicitation
of the prior distribution, may be important in such cases.
## 7 Comparing results in terms of statistical inference
In this section we compare results for the parametric and semiparametric
models in terms of statistical inference, regardless of the sampling strategy
used to obtain posterior samples. All posterior samples follow the
parameterization in (3), and we use samples from the most efficient sampling
strategy for each dataset. We use posterior means as point estimates for the
item and ability parameters. For the simulated datasets, we measure how well
the models recover the (known) true value of the parameters using absolute
error, e.g., $|\hat{\beta_{i}}-\beta_{i}|$, and squared error, e.g.,
$(\hat{\beta_{i}}-\beta_{i})^{2}$.
A crucial point of this paper is to make inference on the distribution of
latent abilities. An estimate of this distribution is sometimes based on the
posterior means of the abilities (Duncan MacEachern, 2008; Bambirra Gonçalves
., 2018), and histograms or kernel density plots are reported. Such an
estimate ignores uncertainty in the estimates of individual abilities.
Instead, one should directly obtain the point estimate of the posterior
distribution of the latent abilities $p(\eta\mid\mathbf{Y})$ (for any value of
$\eta$) using the posterior samples. In the parametric case, this reduces to:
$\widehat{p(\eta\mid\mathbf{Y})}=\frac{1}{T}\sum_{t=1}^{T}\mathcal{N}(\eta;\mu^{(t)},\sigma^{2(t)}),$
(10)
with $\mathcal{N}(\cdot;\mu,\sigma^{2})$ indicating the probability density
function of a normal distribution with mean $\mu$ and variance $\sigma^{2}$,
and $t=1,\ldots,T$ denoting an MCMC iteration. In the semiparametric case, a
point estimate of $p(\eta\mid\mathbf{Y})$ is the posterior mean of the mixing
measure $G$ of the Dirichlet process. This can be obtained using posterior
samples, averaging over the DP conditional distribution in (4) computed for
each iteration $t=1,\ldots,T$,
$p\widehat{(\eta\mid\mathbf{Y})}=\frac{1}{T}\sum_{t=1}^{T}\left\\{\left[\sum_{k=1}^{K^{(t)}}\frac{n_{k}^{(t)}}{\alpha^{(t)}+N}\mathcal{N}(\eta;\mu_{k}^{(t)},\sigma_{k}^{2^{(t)}})\right]+\frac{\alpha^{(t)}}{\alpha^{(t)}+N}\mathcal{N}(\eta;\mu_{K^{(t)}+1},\sigma_{K^{(t)}+1}^{2})\right\\},$
(11)
with $n_{k}^{(t)}$ the number of observations in cluster $k$ at iteration $t$,
$K^{(t)}$ the total number of clusters at iteration $t$, and $\mu_{K^{(t)}+1}$
and $\sigma_{K^{(t)}+1}^{2}$ sampled from $G_{0}$ (conditional on the data).
We graphically compare estimates for the distribution of ability resulting
from (10)–(11) with the estimates obtained using the posterior means.
It is possible to make full inference on $p(\eta\mid\mathbf{Y})$ in the
semiparametric setting; this requires sampling from the posterior of the
mixing distribution $F$. A computational approach to obtain the entire
posterior distribution has been presented in Gelfand Kottas (2002), a version
of whose algorithm is implemented in NIMBLE in the function
getSamplesDPMeasure. This function provides samples of a truncated version of
the infinite mixture to a level $L$. The value of $L$ varies at each iteration
of the MCMC’s output when $\alpha$ is random, while it is the same at each
iteration when $\alpha$ is fixed. In our case, for every MCMC iteration, we
can obtain samples of the vector of mixture weights
$\\{w_{1}^{(t)},\ldots,w_{L^{(t)}}^{(t)}\\}$ and parameters of the mixture
components. We can use these samples to make inference on functionals of the
distribution, such as the percentile for an individual, $100\times p_{j}$,
where $p_{j}=\int_{-\infty}^{\eta_{j}}p(\eta\mid\mathbf{Y})d\eta$, typically
paired with test scores when giving results for educational assessments. For
an individual $j$ for $j=1,\ldots,N$ we estimate $p_{j}$ at each MCMC
iteration as
$p_{j}^{(t)}=\sum_{l=1}^{L^{(t)}}w_{l}^{(t)}F_{\mathcal{N}}(\eta_{j}^{(t)};\mu_{l}^{(t)},\sigma_{l}^{2(t)}),$
(12)
where $F_{\mathcal{N}}$ denotes the distribution function of the normal
distribution. For comparison, we define the parametric counterpart as
$p_{j}^{(t)}=F_{\mathcal{N}}(\eta_{j}^{(t)};\mu^{(t)},\sigma^{2(t)})$.
| Unimodal Simulation | Bimodal Simulation | Multimodal Simulation
---|---|---|---
| Parametric | Semi-parametric | Parametric | Semi-parametric | Parametric | Semi-parametric
| MAE | MSE | MAE | MSE | MAE | MSE | MAE | MSE | MAE | MSE | MAE | MSE
Difficulty parameters | 0.0996 | 0.0185 | 0.0988 | 0.0183 | 0.0895 | 0.0137 | 0.0673 | 0.0083 | 0.0647 | 0.0080 | 0.0737 | 0.0103
Discrimination parameters | 0.0734 | 0.0069 | 0.0731 | 0.0070 | 0.0832 | 0.0105 | 0.0397 | 0.0020 | 0.0721 | 0.0082 | 0.0677 | 0.0062
Ability parameters | 0.4836 | 0.3719 | 0.4836 | 0.3720 | 0.5944 | 0.5571 | 0.5501 | 0.4775 | 0.5477 | 0.4753 | 0.5212 | 0.4444
Table 3: MAE and MSE for the item and ability parameters estimates, under the
three simulation scenarios, using samples from most efficient MCMC-based
sampling strategies.
### 7.1 Inferential results for the simulated datasets
We report results using posterior samples from the unconstrained sampling
strategy under the IRT parameterization. Using the absolute error and the
squared error for each parameter, we report in Table 3 the mean absolute error
(MAE) and the mean squared error (MSE) across item and ability parameters.
Figure 6: Unimodal simulated data. Comparison of the posterior mean estimates
(with $95\%$ credible interval) of item parameters (difficulties
$\boldsymbol{\beta}$ and discriminations $\boldsymbol{\lambda}$) for
parametric and semiparametric 2PL models and true simulated values. Note that
some estimates from the parametric model overlap almost exactly with
semiparametric ones.
In the unimodal scenario, we observe similar performance for the parametric
and semiparametric 2PL in estimating item parameters (Figure 6). As expected,
we observe some differences when considering the bimodal and multimodal
scenario (Figure 7). The use of a model with a more flexible distribution
improves recovery of both item and ability parameters in the bimodal scenarios
(Table 3). This is especially evident when comparing estimates of the
discrimination parameters, in particular for larger values. Instead, in the
multimodal case the inference for the item parameters seems relatively
insensitive to the specification.
Figure 7: Bimodal Simulation (top row) and multimodal simulation (bottom row).
Comparison of the posterior mean estimates (with $95\%$ credible interval) of
item parameters (difficulties $\boldsymbol{\beta}$ and discriminations
$\boldsymbol{\lambda}$) for parametric and semiparametric 2PL models and true
simulated values.
Results for the ability parameters are similar when estimating abilities using
the posterior means of the individual abilities (Figure 8). However, results
are very different when one looks at estimates of the distribution of ability
(Figure 9).
Figure 8: Histogram and density estimate of individual posterior mean
abilities, under unimodal (left column), bimodal (middle column), and
multimodal (right column) scenarios compared with the true density (dotted
line). Figure 9: Distribution of ability estimated under the unimodal (left
column), bimodal (middle column), and multimodal (right column) scenarios,
compared with the true density (dotted line). Dashed lines indicate $95\%$
credible intervals for the estimated distributions.
The normality assumption of the parametric model leads to unimodal density
estimates, inconsistent with the true distribution, whereas the semiparametric
model can recover it. The posterior means of individual abilities in Figure 8
are a compromise between the inferred distribution of ability and the
information in the data, so with sufficient observations, one can obtain
estimates of the distribution that are reasonable even with severe model mis-
specification. In other words, for mis-specified parametric models, the in-
sample predictions for observed individuals can be reasonable, while the out-
of-sample predictions based on (10) for new individuals are poor. Note that
when using the parametric model, inspection of the posterior means of
individual abilities can be used to assess model mis-specification relative to
the assumed parametric distribution.
Mis-specification of the distribution of ability has limited effect when
estimating individual percentiles. In Figure 10 we compare the posterior mean
estimates of individual percentiles with the percentiles calculated using the
true distribution assumed for the simulation, for a subset of $50$
individuals. Overall, the parametric and semiparametric estimates produce
similar results even when the estimated density is largely different. This is
because estimation of percentile is basically ranking the individuals; it
makes sense that ranking is relatively insensitive to the estimation of the
distribution of ability.
Figure 10: Estimates of individual percentiles (with $95\%$ credible interval)
for a subset of $50$ individuals with varying (true) ability levels under the
unimodal (left column), bimodal (middle column), and multimodal (right
column). Black dots correspond to true percentiles.
### 7.2 Inferential results for real-world data
For the real data examples we graphically inspect results from the parametric
and semiparametric models. To compare the overall model fit in the parametric
and semiparametric cases, we computed the Widely Applicable Information
Criterion (WAIC) (Watanabe Opper, 2010). We refer to the NIMBLE user manual
for details about WAIC calculation (de Valpine ., 2022, see Section 7.7).
We found that for the health data the semiparametric model performs better
than the parametric one (WAIC of 68,527 versus 70,414), while for the TIMSS
data the best model is the semiparametric 3PL (WAIC of 228,871; parametric
3PL: 229,123; semiparametric 2PL: 229,529; parametric 2PL: 229,575). Hence in
this Section, we show inferential results using samples from the IRT
unconstrained model for both the health and TIMSS data, using the 3PL model
for TIMMS. We also report inferential results for the 2PL model in the
Supplementary Materials (Section E).
In Figure 11 we compare item parameter estimates from the two models for the
health data application, while Figure 12 shows estimates for the distribution
of abilities. Recall that, in this case, we interpret the latent ability as
physical ability, with high values characterizing healthy individuals. As with
the bimodal simulation, estimates from the parametric model of the
distribution of physical ability are quite different than the distribution of
individual posterior mean abilities. It is clear that the parametric model is
badly mis-specified and would produce bad out-of-sample predictions. In
contrast, the semiparametric model seems to nicely characterize multi-modality
in the latent distribution. We observe in Figure 12 large credible intervals
for high values of this distribution that can be explained by the presence of
many individuals with high raw scores, (i.e., 9 or 10 out of 10, Figure 2) for
whom the model can clearly determine that their physical abilities are high,
but with the exact magnitudes being difficult to identify.
The two modeling assumptions yield different estimates of the item parameters
(Figure 11), with this difference being higher for extreme values. However,
the relative ranking of the items is roughly the same in both cases, with for
example item 1 (_Vigorous activities_) being the most difficult item and the
one with lowest value of the discrimination parameter. According to the
parametric model, discrimination parameters for item $3$ (_Lift/carry_) and
item $10$ (_Bathing/dressing_) should have similar values, while the
semiparametric model separates them.
Figure 11: Health data. Comparison of item parameter estimates from the
parametric and semiparametric models. In each panel items are ordered by
increasing values of the parameter estimate under the semiparametric model.
Figure 12: Health Data. Histogram and density estimate of the posterior means
of the latent abilities (left panel) and estimate of the posterior
distribution for the latent abilities (right panel). Dashed lines indicate
$95\%$ credible intervals for the estimated distributions.
For the TIMSS data we only compare point estimates of the item parameters
(Figure 13), due to the large number of parameters. Estimates for the
difficulty parameters and discrimination parameters have the largest
difference for low/high values of the parameters, while estimates of the
guessing parameters are quite different. Figure 14 shows estimates of the
distribution of ability. In this case the semiparametric model estimate shows
departure from the normal parametric assumption, with multimodality in the
estimated distribution. Differences in the distribution of ability between the
semiparametric and parametric model may explain the large discrepancies
between the guessing parameter estimates; however, further investigation of
this matter is outside the scope of the paper. We found also that the estimate
of the ability distribution is quite different under the semiparametric 3PL
and semiparametric 2PL model with the distribution under the 2PL model being
unimodal and right-skewed instead of multimodal (see Supplementary Materials,
Section F).
Figure 13: TIMSS data. Comparison of posterior estimates of the item
parameters between the parametric and semiparametric 3PL model both using the
SI unconstrained centered sampling strategy. Figure 14: TIMSS Data. Histogram
and density estimate of the posterior means of the latent abilities (left
panel), and estimate of the posterior distribution for the latent abilities
(right panel). Dashed lines indicate $95\%$ credible intervals for the
estimated distributions. Figure 15: Estimates of individual percentiles (with
$95\%$ credible interval) for a subset of $50$ individuals, for the health
data (left panel) and TIMSS data (right panel).
Figure 15 compares estimates of the percentiles for both the health and TIMSS
data for a sample of $50$ individuals sorted according to the point estimates
of the abilities from the semiparametric model. There are moderate differences
in percentile values and individual ordering between the parametric and
semiparametric models; in particular some estimates are associated with larger
intervals than in the semiparametric case.
## 8 Discussion
In this paper, we consider a semiparametric extension for binary logistic IRT
models, using Dirichlet process mixtures as a nonparametric prior to flexibly
characterize the distribution of ability. We provide an overview of these
models and study how different sets of constraints can address identifiability
issue and lead to different MCMC estimation strategies.
Focusing on the 2PL and 3PL models, we compare efficiency and inferential
results under different sampling strategies based on model parametrization,
constraints and sampling algorithms. We find that MCMC performance across
strategies can vary in relation to underlying shape of the latent distribution
and the total number of parameters.
When moving to semiparametric modeling, the computational cost can be high for
large datasets, given that sampling from the Dirichlet process requires
iteration through all individuals. However we find computational costs to be
reasonable in our applications in light of the better inferential results.
In particular under model mis-specification, inference for item parameters
worsens noticeably in the parametric model compared to the semiparametric
model. With sufficient data, inference for the abilities of observed
individuals can be decent even under mis-specification of the distribution of
ability, but inference for the unknown latent distribution (i.e., the
predictive distribution for new individuals) as a whole can be quite bad.
Although parametric IRT models can work well in applications in educational
assessment, having access to semiparametric models can be broadly useful as it
allows inference on the entire underlying distribution of ability and its
functionals. This is particularly relevant in contexts where the distribution
of the individual latent trait is more complicated, for example, when
measuring health (Smits ., 2020) or psychological outcomes (Reise Rodriguez,
2016).
Results of this work potentially can be generalized to versions of binary IRT
models using different prior distributions or link functions (e.g. probit),
since we considered general MCMC sampling algorithms as opposed to algorithms
tailored to specific choices of such model components. As a general
recommendation for IRT models, we found that sampling strategies using
parameter expansion are more efficient than those embedding the constraints.
In this work we extensively use the NIMBLE software for hierarchical modeling,
with code reproducing results in the paper available at
https://github.com/salleuska/IRT_nimble_code. Although there are other
software solutions enabling Bayesian nonparametric modeling, these are often
limited in the type of algorithms or in the class of models available. NIMBLE
offers a high degree of flexibility in that the models considered in this
paper could be easily embedded in more complicated ones. Sampler assignment
can be highly customized by the user, including user-defined sampling
algorithms. This customizability makes NIMBLE a powerful platform for
comparing different sampling strategies. At the same time, NIMBLE allows easy
sharing of the most successful strategies as block-box implementations for end
users.
## Supplementary materials
## A. Identifiability
The 2PL model is not identifiable based on the likelihood. Here we demonstrate
the non-identifiability for the two parameterizations we consider, showing how
different linear transformations lead to the same probabilities. Note that
these transformations are defined for every parameter associated with each
item $i=1,\ldots,I$ and individual $j=1,\ldots,N$.
Under the IRT parameterization:
1. 1.
$\eta_{j}^{\prime}=\eta_{j}/s$ and $\lambda_{i}^{\prime}=s\lambda_{i}$
$\lambda_{i}^{\prime}(\eta_{j}^{\prime}-\beta_{i})=s\lambda_{i}(\eta_{j}/s-\beta_{i})=\lambda_{i}\eta_{j}-\lambda_{i}\beta_{i}=\lambda_{i}(\eta_{j}-\beta_{i}),$
2. 2.
$\eta_{j}^{\prime}=\eta_{j}+c$ and $\beta_{i}^{\prime}=\beta_{i}+c$,
$\lambda_{i}(\eta_{j}^{\prime}-\beta_{i}^{\prime})=\lambda_{i}(\eta_{j}+c-(\beta_{i}+c))=\lambda_{i}(\eta_{j}-\beta_{i}).$
Under the slope-intercept parameterization:
1. 1.
$\eta_{j}^{\prime}=\eta_{j}/s$ and $\lambda_{i}^{\prime}=s\lambda_{i}$,
$\lambda_{i}^{\prime}\eta_{j}^{\prime}+\gamma_{i}=s\lambda_{i}\eta_{j}/s+\gamma_{i}=\lambda_{i}\eta_{j}+\gamma_{i},$
2. 2.
$(\lambda_{i}\eta_{j})^{\prime}=\lambda_{i}\eta_{j}+c$ and
$\gamma_{i}^{\prime}=\gamma_{i}-c$, or $\eta_{j}^{\prime}=\eta_{j}+c$ and
$\gamma_{i}^{\prime}=\gamma_{i}-\lambda_{i}c$
$\lambda_{i}\eta_{j}^{\prime}+\gamma_{i}^{\prime}=\lambda_{i}(\eta_{j}+c)+\gamma_{i}-\lambda_{i}c=\lambda_{i}\eta_{j}+\gamma_{i}.$
### Post-processing to satisfy identifiability constraints
This section reports the transformations we apply to item and ability
parameters in order to satisfy the identifiability constraints in our base
parameterization (3). These transformations are applied to each posterior
sample.
Under the IRT parameterization, the set of transformations for each posterior
sample of $\\{\lambda_{i},\beta_{i},\eta_{j}\\}$ for
$i=1,\ldots,I,j=1,\ldots,N$ takes these forms:
$\displaystyle\lambda^{*}_{i}=s\lambda_{i},\quad\beta^{*}_{i}=\frac{\beta_{i}-b}{s},\quad\eta^{*}_{j}=\frac{\eta_{j}-b}{s},$
subject to $\prod_{i=1}^{I}\lambda^{*}_{i}=1$,
$\sum_{i=1}^{I}\beta^{*}_{i}=0$. By solving the system of equations given by
the transformations and the set of identifiability constraints, we obtain
$\displaystyle s=\exp\left\\{\sum_{i=1}^{I}\log(\lambda_{i})/I\right\\},\quad
b=\frac{\sum_{i=1}^{I}\beta_{i}}{I}.$ (A1)
Under the slope-intercept parameterization, the set of transformations for
each posterior sample of $\\{\lambda_{i},\gamma_{i},\eta_{j}\\}$ for
$i=1,\ldots,I,j=1,\ldots,N$ takes these forms:
$\displaystyle\tilde{\lambda}_{i}=s\lambda_{i},\quad\tilde{\gamma}_{i}=\gamma_{i}-\lambda_{i}c,\quad\tilde{\eta}_{j}=\frac{\eta_{j}+c}{s},$
subject to $\prod_{i=1}^{I}\tilde{\lambda}_{i}=1$,
$\sum_{i=1}^{I}\tilde{\gamma}_{i}=0$. Similarly, by solving the system of
equations given by the transformations and the set of identifiability
constraints, we obtain
$\displaystyle s=\exp\left\\{\sum_{i=1}^{I}\log(\lambda_{i})/I\right\\},\quad
c=\frac{\sum_{i=1}^{I}\gamma_{i}}{\sum_{i=1}^{I}\lambda_{i}}.$ (A2)
Finally, to get from the slope-intercept parameterization to the IRT
parameterization, we define
$\tilde{\beta}_{i}\vcentcolon=-\tilde{\gamma}_{i}/\tilde{\lambda}_{i}$ and
then calculate $\beta_{i}^{*}=\tilde{\beta}_{i}-\sum_{i}\tilde{\beta}_{i}/I$.
### Rescaling the DP density
We can obtained posterior samples from the mixing distribution $F$ via
NIMBLE’s getSamplesDPmeasure function, allowing us to estimate the density for
the latent ability distribution. However, when comparing these estimated
densities between models, for some of the sampling strategies, we need to
transform the estimated density to account for the transformations of the
abilities from the scale on which sampling is done to the scale in (3).
As an example, consider the IRT parameterization without constraints. From the
MCMC output we can obtain $p(\tilde{\eta})$ evaluated for different values of
$\tilde{\eta}$, but we want $p(\tilde{\eta}^{*})$ with
$\tilde{\eta}^{*}=(\tilde{\eta}-b)/s$. To do so we need the Jacobian of the
transformation, which is simply $s$. Then, we obtain $p(\tilde{\eta}^{*})$
$p(\tilde{\eta}^{*})=p_{\tilde{\eta}}(s\tilde{\eta}^{*}+b)\left|\frac{\partial(s\tilde{\eta}^{*}+b)}{\partial\tilde{\eta}^{*}}\right|=p_{\tilde{\eta}}(s\tilde{\eta}^{*}+b)s.$
(A3)
## B. Details about the sampling algorithms
Tables 4-5 summarize the sampling algorithms used for each sampling strategy
in Table 1.
Table 4: Summary of the sampling algorithms used for each parameter under the
sampling strategies considered for the parametric 2PL model.
Model constraints IRT parameterization SI parameterization MH/conjugate HMC
(Stan) MH/conjugate Centered Constrained abilities Adaptive MH
$\\{\log(\lambda_{i}),\beta_{i},\eta_{j}\\}$ HMC
$\left\\{\\{\log(\lambda_{i})\\},\\{\beta_{i}\\},\\{\eta_{j}\\}\right\\}$
Adaptive MH $\\{\log(\lambda_{i}),\gamma_{i},\eta_{j}\\}$ Centered sampler for
pairs $\\{\log(\lambda_{i}),\gamma_{i}\\}$ Adaptive MH $\\{\eta_{j}\\}$
Constrained item Adaptive MH
$\\{\log(\lambda^{*}_{i}),\beta^{*}_{i},\eta_{j}\\}$ - Adaptive MH
$\\{\log(\lambda^{*}_{i}),\gamma^{*}_{i},\eta_{j}\\}$ - Unconstrained Adaptive
MH $\\{\log(\lambda_{i}),\beta_{i},\eta_{j}\\}$ Conjugate
$\\{\mu,\sigma^{2}\\}$ - Adaptive MH
$\\{\log(\lambda_{i}),\gamma_{i},\eta_{j}\\}$ Conjugate $\\{\mu,\sigma^{2}\\}$
Centered sampler for pairs $\\{\log(\lambda_{i}),\gamma_{i}\\}$ Adaptive MH
$\\{\eta_{j}\\}$ Conjugate $\\{\mu,\sigma^{2}\\}$
Table 5: Summary of the sampling algorithms used for each parameter under the
sampling strategies considered for the semiparametric 2PL model.
Model constraints IRT parameterization SI parameterization MH/conjugate
MH/conjugate Centered Constrained item Adaptive MH
$\\{\log(\lambda^{*}_{i}),\beta^{*}_{i},\eta_{j}\\}$ CRP sampler
$\\{\alpha,\\{z_{j}\\}\\}$ Conjugate
$\\{\\{\mu^{*}_{k}\\},\\{\sigma^{2*}_{k}\\}\\}$ Adaptive MH
$\\{\log(\lambda^{*}_{i}),\gamma^{*}_{i},\eta_{j}\\}$ CRP sampler
$\\{\alpha,\\{z_{j}\\}\\}$ Conjugate
$\\{\\{\mu^{*}_{k}\\},\\{\sigma^{2*}_{k}\\}\\}$ - Unconstrained Adaptive MH
$\\{\log(\lambda_{i}),\beta_{i},\eta_{j}\\}$ CRP sampler
$\\{\alpha,\\{z_{j}\\}\\}$ Conjugate
$\\{\\{\mu^{*}_{k}\\},\\{\sigma^{2*}_{k}\\}\\}$ Adaptive MH
$\\{\log(\lambda_{i}),\gamma_{i},\eta_{j}\\}$ CRP sampler
$\\{\alpha,\\{z_{j}\\}\\}$ Conjugate
$\\{\\{\mu^{*}_{k}\\},\\{\sigma^{2*}_{k}\\}\\}$ Centered sampler for pairs
$\\{\log(\lambda_{i}),\gamma_{i}\\}$ Adaptive MH $\\{\eta_{j}\\}$ CRP sampler
$\\{\alpha,\\{z_{j}\\}\\}$ Conjugate
$\\{\\{\mu^{*}_{k}\\},\\{\sigma^{2*}_{k}\\}\\}$
#### Description of the sampling algorithms
* •
Conjugate sampler: for the models used in the paper, we exploit conjugancy
results for the Normal distribution with Normal Inverse-gamma priors for the
mean and variance.
* •
Adaptive MH (Metropolis-Hastings): uses a normal proposal distribution, with
initial proposal variance equal to $1$ and and adaptation interval of $200$
iterations. The adaptation routine is implemented as given in Shaby Wells
(2010).
* •
Centered sampler: this is a custom defined sampler implemented by the authors.
Details are given in the below.
* •
CRP sampler: under the CRP specification, the random measure $G$ is integrated
out from the model and NIMBLE assigns a collapsed sampler.
* –
clustering indicators $\mathbf{z}$ are updated as in described in Neal (2000);
* –
the DP concentration parameter $\alpha$ is sampled as described in Escobar
West (1995) (Section 6) when a Gamma prior is used, as in the models
considered in the paper. If another prior is considered, NIMBLE uses a random
walk Metropolis-Hastings.
### Centered sampler
We consider a custom sampler for the 2PL model under the slope-intercept
parameterization. Intuition for this sampling strategy comes from the
resemblance to a linear model. In order to sample
$\\{\lambda_{i},\gamma_{i}\\}$ efficiently, we propose centering the implied
covariate, $\eta_{j}$, to have mean zero. This is analogous to centering
covariates in a linear model, but in this case the ”covariate” values are not
fixed, so the centering needs to be done in each iteration. For a given item
$i$ for $i=1,\ldots,I$ we can rewrite
$\displaystyle\lambda_{i}\eta_{j}+\gamma_{i}$
$\displaystyle=\lambda_{i}(\eta_{j}-\bar{\eta})+\lambda_{i}\bar{\eta}+\gamma_{i},$
$\displaystyle=\lambda_{i}\eta_{j}^{c}+\gamma_{i}^{c},$
such that the quantity $\eta_{j}^{c}=\eta_{j}-\bar{\eta}$ is centered. The
idea is to propose a new value $\lambda_{i}^{*}$ in this new parameterization
at each MCMC iteration, using a random walk on the log scale. Translating to
the original parameterization, we have:
$\displaystyle\lambda_{i}^{*}\eta_{j}^{c}+\gamma_{i}^{c}$
$\displaystyle=\lambda_{i}^{*}(\eta_{j}-\bar{\eta})+\lambda_{i}\bar{\eta}+\gamma_{i},$
$\displaystyle=\lambda_{i}^{*}\eta_{j}-\lambda_{i}^{*}\bar{\eta}+\lambda_{i}\bar{\eta}+\gamma_{i}.$
This means that we are proposing
$\gamma_{i}^{*}=\gamma_{i}+\bar{\eta}(\lambda_{i}-\lambda_{i}^{*})$. Thus we
have a joint proposal $(\lambda_{i}^{*},\gamma_{i}^{*})$ that accounts for the
usual correlation in a regression between intercept and slope. Apart from
accounting for sampling $\lambda_{i}$ on the log scale, the proposal is
symmetric, so no Hastings correction is needed. The original sampler for
$\gamma_{i}$ can stay the same. This is because in the reparameterization with
$\gamma_{i}^{c}$ above, shifting $\gamma_{i}$ by a certain amount is
equivalent to shifting $\gamma_{i}^{c}$.
## C. Health data questions
The following items are about activities you might do during a typical day.
Does your health now limit you in these activities? If so, how much?
1. 1.
Vigorous activities: Vigorous activities, such as running, lifting heavy
objects, participating in strenuous sports.
2. 2.
Moderate activities: Moderate activities, such as moving a table, pushing a
vacuum cleaner, bowling or playing golf.
3. 3.
Lift/Carry: Lifting or carrying groceries.
4. 4.
Several stairs: Climbing several flights of stairs.
5. 5.
One flight stairs: Climbing one flight of stairs.
6. 6.
Bend/Kneel/Stoop: Bending, kneeling, or stooping.
7. 7.
Walk more mile: Walking more than a mile.
8. 8.
Walk several blocks: Walking several blocks.
9. 9.
Walk one block: Walking one block.
10. 10.
Bathing/Dressing: Bathing or dressing yourself.
## D. A note on efficiency comparisons
### Comments on the multivariate ESS
In this section we compare univariate and multivariate efficiency metrics. In
particular, we compare efficiency values based on the mESS (multivariate
efficiency), with the distribution of efficiencies calculated for each
parameter (univariate efficiency) using the total time. We report these
metrics for each simulation scenario, selecting three representative
strategies under the IRT parameterization for the parametric 2PL model.
Figures 16-18 show the distribution of univariate efficiency for difficulty,
discrimination and ability parameters, along with a table reporting
information for the multivariate efficiency. There are some differences
between the multivariate and univariate efficiency results. This is expected
because the mESS provides a single scalar measure of mixing performance that
accounts for cross-correlation among the parameters and does not necessarily
reflect the distribution of univariate ESSs. In fact, values of the mESS can
be larger than all the univariate ESSs. For example, the IRT HMC strategy has
larger univariate ESS values compared to other sampling strategies, but lower
mESS.
Strategy | mESS | total time (second) | efficiency (mESS/second)
---|---|---|---
IRT HMC (Stan)** | 31802 | 1405 | 23
IRT unconstrained | 27540 | 934 | 29
IRT constrained item | 27904 | 7318 | 4
Figure 16: Unimodal simulation ($N=2000,I=15)$. Distribution of univariate
efficiencies (univariate ESS/seconds) for each group of parameters used to
compute the multivariate efficiency (mESS/second) using the total time. The
symbol ** denotes median results across 11 runs.
Strategy | mESS | total time (second) | efficiency (mESS/second)
---|---|---|---
IRT HMC (Stan)** | 30473 | 2191 | 14
IRT unconstrained | 27167 | 937 | 29
IRT constrained item | 27463 | 4508 | 6
Figure 17: Bimodal simulation ($N=2000,I=15)$. Distribution of univariate
efficiencies (univariate ESS/seconds) for each group of parameters used to
compute the multivariate efficiency (mESS/second) using the total time. The
symbol ** denotes median results across 11 runs.
Strategy | mESS | total time (second) | efficiency (mESS/second)
---|---|---|---
IRT HMC (Stan)** | 22698 | 1176 | 19
IRT unconstrained | 27442 | 1407 | 19
IRT constrained item | 27756 | 5248 | 5
Figure 18: Multimodal simulation ($N=2000,I=15)$. Distribution of univariate
efficiencies (univariate ESS/seconds) for each group of parameters used to
compute the multivariate efficiency (mESS/second) using the total time. The
symbol ** denotes median results across 11 runs.
### Variability of mESS when using HMC from the Stan software
When deciding on the number of posterior , burn-in and warm-up samples, we
tried to obtain a reliable estimate of the multivariate ESS. To ensure the
chains were long enough, we used multiple runs for some of the experiments. We
found that mESS estimates based on the chosen settings of the MCMC algorithm
(i.e., number of iterations, number of burn-in or warm-up samples) have
negligible variability across multiple runs for all strategies with the
exception of using HMC as implemented in Stan. As an example, Figure 19 shows
the distribution of multivariate ESS for the IRT constrained abilities
approach.
Figure 19: Comparison of mESS estimates across multiple runs for the IRT
constrained abilities approach using the unimodal simulation scenario.
We also found that lower and higher values of ESS for the strategies using the
HMC are highly correlated with values of the tuning parameters of the HMC
(i.e., leapfrog and step-size parameters), which are typically estimated
during the warm-up phase (see Figure 20).
Figure 20: Estimates of the mESS versus average values of the HMC tuning
parameters (step-size and leapfrog) across post warm-up iterations.
## E. Results from additional simulations
We investigated how different combinations of numbers of items and individuals
affect efficiency of the different sampling strategies. In particular, we
simulated data under the three scenarios presented in Section 5.1 following a
factorial design with $I\in\\{10,30\\}$ and individuals
$N\in\\{1,000,5,000\\}$. We omit results for the strategy using HMC due to the
high variability in estimating the mESS.
Figure 21: Multivariate ESS per second of the parametric sampling strategies
(excluding HMC) across different combinations of numbers of items and
individuals. Figure 22: Multivariate ESS per second of the semiparametric
sampling strategies (excluding HMC) across different combinations of numbers
of items and individuals.
## F. Inferential results for the TIMSS data using the 2PL model
Figure 23: TIMSS data, 2PL model. Comparison of posterior estimates of the
item parameters between the parametric and semiparametric model both using the
SI unconstrained centered sampling strategy. Figure 24: TIMSS Data, 2PL model.
Histogram and density estimate of the posterior means of the latent abilities
(left panel), and estimate of the posterior distribution for the latent
abilities (right panel). Dashed lines indicate $95\%$ credible intervals for
the estimated distributions. Figure 25: Estimates of individual percentiles
(with $95\%$ credible interval) for a subset of $50$ individuals, for the
health data (left panel) and TIMSS data (right panel) under the 2PL model.
## References
* Aldous (1985) aldous1985exchangeabilityAldous, DJ. 1985\. Exchangeability and related topics Exchangeability and related topics. École d’Été de Probabilités de Saint-Flour XIII—1983 École d’été de probabilités de saint-flour xiii—1983 ( 1–198). Springer.
* Antoniak (1974) antoniak1974Antoniak, CE. 197411\. Mixtures of Dirichlet Processes with Applications to Bayesian Nonparametric Problems Mixtures of Dirichlet Processes with applications to Bayesian nonparametric problems. Annals of Statistics261152–1174. https://doi.org/10.1214/aos/1176342871 10.1214/aos/1176342871
* Azevedo . (2011) azevedo2011bayesianAzevedo, CL., Bolfarine, H. Andrade, DF. 2011\. Bayesian inference for a skew-normal IRT model under the centred parameterization Bayesian inference for a skew-normal IRT model under the centred parameterization. Computational Statistics & Data Analysis551353–365.
* Azzalini (1985) azzalini1985classAzzalini, A. 1985\. A class of distributions which includes the normal ones A class of distributions which includes the normal ones. Scandinavian Journal of Statistics171–178.
* Bafumi . (2005) bafumi2005practicalBafumi, J., Gelman, A., Park, DK. Kaplan, N. 2005\. Practical issues in implementing and understanding Bayesian ideal point estimation Practical issues in implementing and understanding Bayesian ideal point estimation. Political Analysis132171–187.
* Bambirra Gonçalves . (2018) bambirra2018bayesianBambirra Gonçalves, F., da Costa Campos Dias, B. Machado Soares, T. 2018\. Bayesian item response model: a generalized approach for the abilities’ distribution using mixtures Bayesian item response model: a generalized approach for the abilities’ distribution using mixtures. Journal of Statistical Computation and Simulation885967–981.
* Bedrick . (1996) bedrick1996newBedrick, EJ., Christensen, R. Johnson, W. 1996\. A new perspective on priors for generalized linear models A new perspective on priors for generalized linear models. Journal of the American Statistical Association914361450–1460.
* Berger . (2009) berger2009formalBerger, JO., Bernardo, JM. Sun, D. 2009\. The Formal Definition of Reference Priors The formal definition of reference priors. The Annals of Statistics372905–938. http://www.jstor.org/stable/30243652
* Berger Pericchi (1996) berger1996intrinsicBerger, JO. Pericchi, LR. 1996\. The intrinsic Bayes factor for model selection and prediction The intrinsic Bayes factor for model selection and prediction. Journal of the American Statistical Association91433109–122.
* Bernardo (1979) bernardo1979referenceBernardo, JM. 1979\. Reference posterior distributions for Bayesian inference Reference posterior distributions for Bayesian inference. Journal of the Royal Statistical Society: Series B (Methodological)412113–128.
* Betancourt . (2017) betancourt2017geometricBetancourt, M., Byrne, S., Livingstone, S. Girolami, M. 2017\. The geometric foundations of Hamiltonian Monte Carlo The geometric foundations of Hamiltonian Monte Carlo. Bernoulli234A2257–2298.
* Blackwell MacQueen (1973) blackwell1973fergusonBlackwell, D. MacQueen, JB. 1973\. Ferguson distributions via Pólya urn schemes Ferguson distributions via Pólya urn schemes. The Annals of Statistics12353–355.
* Bolt . (2001) bolt2001mixtureBolt, DM., Cohen, AS. Wollack, JA. 2001\. A Mixture Item Response Model for Multiple-Choice Data A mixture item response model for multiple-choice data. Journal of Educational and Behavioral Statistics264381–409. http://www.jstor.org/stable/3648167
* Bürkner (2021) burker2021bayesianBürkner, PC. 2021\. Bayesian Item Response Modeling in R with brms and Stan Bayesian item response modeling in r with brms and stan. Journal of Statistical Software10051–54. https://www.jstatsoft.org/index.php/jss/article/view/v100i05 10.18637/jss.v100.i05
* Carpenter . (2017) stan2017jssCarpenter, B., Gelman, A., Hoffman, M., Lee, D., Goodrich, B., Betancourt, M.Riddell, A. 2017\. Stan: A Probabilistic Programming Language Stan: A probabilistic programming language. Journal of Statistical Software, Articles7611–32. https://www.jstatsoft.org/v076/i01 10.18637/jss.v076.i01
* Clinton . (2004) clinton2004statisticalClinton, J., Jackman, S. Rivers, D. 2004\. The statistical analysis of roll call data The statistical analysis of roll call data. American Political Science Review355–370.
* de Valpine . (2022) nimble-manual:2021de Valpine, P., Paciorek, C., Turek, D., Michaud, N., Anderson-Bergman, C., Obermeyer, F.Hug, J. 2022\. NIMBLE User Manual Nimble user manual []. https://r-nimble.org 10.5281/zenodo.1211190
* de Valpine . (2020) nimble-software2020de Valpine, P., Paciorek, CJ., Turek, D., Michaud, N., Anderson-Bergman, C., Obermeyer, F.Paganin, S. 2020\. NIMBLE: MCMC, Particle Filtering, and Programmable Hierarchical Modeling. NIMBLE: MCMC, particle filtering, and programmable hierarchical modeling. https://cran.r-project.org/package=nimble R package version 0.9.1 10.5281/zenodo.1211190
* de Valpine . (2017) devalpine2017nimblede Valpine, P., Turek, D., Paciorek, CJ., Anderson-Bergman, C., Lang, DT. Bodik, R. 2017\. Programming With Models: Writing Statistical Algorithms for General Model Structures With NIMBLE Programming with models: Writing statistical algorithms for general model structures with nimble. Journal of Computational and Graphical Statistics262403-413. https://doi.org/10.1080/10618600.2016.1172487 10.1080/10618600.2016.1172487
* Duncan MacEachern (2008) duncan2008nonparametricDuncan, KA. MacEachern, SN. 2008\. Nonparametric Bayesian modelling for item response Nonparametric Bayesian modelling for item response. Statistical Modelling8141–66.
* Escobar West (1995) escobar1995bayesianEscobar, MD. West, M. 1995\. Bayesian density estimation and inference using mixtures Bayesian density estimation and inference using mixtures. Journal of the American Statistical Association90430577–588.
* Ferguson (1973) ferguson1973Ferguson, TS. 197303\. A Bayesian Analysis of Some Nonparametric Problems A Bayesian analysis of some nonparametric problems. The Annals of Statistics12209–230. 10.1214/aos/1176342360
* Finch Edwards (2016) finch2016raschFinch, H. Edwards, JM. 2016\. Rasch Model Parameter Estimation in the Presence of a Nonnormal Latent Trait Using a Nonparametric Bayesian Approach Rasch model parameter estimation in the presence of a nonnormal latent trait using a nonparametric Bayesian approach. Educational and Psychological Measurement764662-684. https://doi.org/10.1177/0013164415608418 10.1177/0013164415608418
* Flegal . (2021) mcmcseFlegal, JM., Hughes, J., Vats, D., Dai, N., Gupta, K. Maji, U. 2021\. mcmcse: Monte Carlo Standard Errors for MCMC mcmcse: Monte Carlo standard errors for MCMC []. Riverside, CA, and Kanpur, India. R package version 1.5-0
* Fox (2010) fox2010bayesianFox, JP. 2010\. Bayesian Item Response modeling: Theory and Applications Bayesian item response modeling: Theory and applications. Springer Science & Business Media.
* Furr (2017) edstan2017Furr, DC. 2017\. edstan: Stan models for Item Response Theory edstan: Stan models for Item Response Theory []. https://CRAN.R-project.org/package=edstan R package version 1.0.6
* Gelfand Kottas (2002) gelfand2002Gelfand, AE. Kottas, A. 2002\. A Computational Approach for Full Nonparametric Bayesian Inference Under Dirichlet Process Mixture Models A computational approach for full nonparametric Bayesian inference under Dirichlet Process mixture models. Journal of Computational and Graphical Statistics112289-305. https://doi.org/10.1198/106186002760180518 10.1198/106186002760180518
* Gelman (2004) gelman2004parameterizationGelman, A. 2004\. Parameterization and Bayesian modeling Parameterization and Bayesian modeling. Journal of the American Statistical Association99466537–545.
* Geweke Singleton (1981) geweke1981maximumGeweke, JF. Singleton, KJ. 1981\. Maximum likelihood “confirmatory” factor analysis of economic time series Maximum likelihood “confirmatory” factor analysis of economic time series. International Economic Review37–54.
* Guhaniyogi Rodriguez (2020) guhaniyogi2020jointGuhaniyogi, R. Rodriguez, A. 2020\. Joint modeling of longitudinal relational data and exogenous variables Joint modeling of longitudinal relational data and exogenous variables. Bayesian Analysis152477–503.
* Hays . (2000) hays2000itemHays, RD., Morales, LS. Reise, SP. 2000\. Item response theory and health outcomes measurement in the 21st century Item response theory and health outcomes measurement in the 21st century. Medical Care389 SupplII28.
* Hoffman Gelman (2014) hoffman2014noHoffman, MD. Gelman, A. 2014\. The No-U-Turn sampler: adaptively setting path lengths in Hamiltonian Monte Carlo. The No-U-Turn sampler: adaptively setting path lengths in Hamiltonian Monte Carlo. Journal of Machine Learning Research1511593–1623.
* Ibrahim (1997) ibrahim1997propertiesIbrahim, JG. 1997\. On Properties of Predictive Priors in Linear Models On properties of predictive priors in linear models. The American Statistician514333-337. https://www.tandfonline.com/doi/abs/10.1080/00031305.1997.10474408 10.1080/00031305.1997.10474408
* Jara . (2011) jara2011dppackageJara, A., Hanson, TE., Quintana, FA., Müller, P. Rosner, GL. 2011\. DPpackage: Bayesian semi-and nonparametric modeling in R DPpackage: Bayesian semi-and nonparametric modeling in R. Journal of Statistical Software4051.
* Johnson (2007) johnson2007modelingJohnson, MS. 2007\. Modeling dichotomous item responses with free-knot splines Modeling dichotomous item responses with free-knot splines. Computational Statistics & Data Analysis5194178–4192.
* Joint Health Surveys Unit of Social and Community Planning Research and University College London (2017) healthdata1996Joint Health Surveys Unit of Social and Community Planning Research and University College London. 2017\. Health Survey for England, 1996. [data collection]. Health survey for england, 1996. [data collection]. 5th Edition. UK Data Service. SN: 3886. http://doi.org/10.5255/UKDA-SN-3886-2
* Karabatsos (2017) karabastos2017Karabatsos, G. 2017\. Bayesian Nonparametric Response Models Bayesian nonparametric response models. WJ. Van der Linden (), Handbook of item response theory, Volume One: Models Handbook of item response theory, volume one: Models ( 323-336). CRC Press. https://www.taylorfrancis.com/books/e/9781466514423/chapters/10.1201/9781315374512-32
* Kirisci . (2001) kirisci2001robustnessKirisci, L., chi Hsu, T. Yu, L. 2001\. Robustness of Item Parameter Estimation Programs to Assumptions of Unidimensionality and Normality Robustness of item parameter estimation programs to assumptions of unidimensionality and normality. Applied Psychological Measurement252146-162. https://doi.org/10.1177/01466210122031975 10.1177/01466210122031975
* Laird (1978) laird78Laird, NM. 1978\. Nonparametric maximum likelihood estimation of a mixing distribution Nonparametric maximum likelihood estimation of a mixing distribution. Journal of the American Statistical Association73805-807.
* Li . (2011) li2011centerLi, Y., Müller, P. Lin, X. 2011\. Center-adjusted inference for a nonparametric Bayesian random effect distribution Center-adjusted inference for a nonparametric Bayesian random effect distribution. Statistica Sinica2131201–1223.
* C. Liu . (1998) liu1998parameterLiu, C., Rubin, DB. Wu, YN. 1998\. Parameter expansion to accelerate EM: the PX-EM algorithm Parameter expansion to accelerate EM: the PX-EM algorithm. Biometrika854755–770.
* JS. Liu Wu (1999) liu1999parameterLiu, JS. Wu, YN. 1999\. Parameter expansion for data augmentation Parameter expansion for data augmentation. Journal of the American Statistical Association944481264–1274.
* SJ. Liu (1996) liu1996nonparametricLiu, SJ. 1996\. Nonparametric hierarchical Bayes via sequential imputations Nonparametric hierarchical Bayes via sequential imputations. The Annals of Statistics243911 – 930. https://doi.org/10.1214/aos/1032526949 10.1214/aos/1032526949
* Lo (1984) lo1984classLo, AY. 1984\. On a class of Bayesian nonparametric estimates: I. Density estimates On a class of Bayesian nonparametric estimates: I. density estimates. The Annals of Statistics351–357.
* McHorney . (1997) mchorney1997evaluationMcHorney, CA., Haley, SM. Ware Jr, JE. 1997\. Evaluation of the MOS SF-36 physical functioning scale (PF-40): II. Comparison of relative precision using Likert and Rasch scoring methods Evaluation of the MOS SF-36 physical functioning scale (PF-40): II. comparison of relative precision using Likert and Rasch scoring methods. Journal of Clinical Epidemiology504451–461.
* Micceri (1989) micceri1989unicornMicceri, T. 1989\. The unicorn, the normal curve, and other improbable creatures. The unicorn, the normal curve, and other improbable creatures. Psychological bulletin1051156.
* Mislevy (1984) mislevy1984estimatingMislevy, RJ. 1984\. Estimating latent distributions Estimating latent distributions. Psychometrika493359–381. https://doi.org/10.1007/BF02306026 10.1007/BF02306026
* Miyazaki Hoshino (2009) miyazaki2009bayesianMiyazaki, K. Hoshino, T. 2009\. A Bayesian semiparametric item response model with Dirichlet Process priors A Bayesian semiparametric item response model with Dirichlet Process priors. Psychometrika743375–393.
* Natesan . (2016) natesan2016bayesianNatesan, P., Nandakumar, R., Minka, T. Rubright, JD. 2016\. Bayesian prior choice in IRT estimation using MCMC and variational Bayes Bayesian prior choice in IRT estimation using MCMC and variational Bayes. Frontiers in Psychology71422.
* Neal (2000) neal2000markovNeal, RM. 2000\. Markov chain sampling methods for Dirichlet Process mixture models Markov chain sampling methods for Dirichlet Process mixture models. Journal of Computational and Graphical Statistics92249–265.
* Nguyen . (2020) nguyen2020Nguyen, D., de Valpine, P., Atchade, Y., Turek, D., Michaud, N. Paciorek, CJ. 202012\. Nested Adaptation of MCMC Algorithms Nested adaptation of mcmc algorithms. Bayesian Analysis1541323–1343. https://doi.org/10.1214/19-BA1190 10.1214/19-BA1190
* Paulon . (2018) paulon2018Paulon, G., De Iorio, M., Guglielmi, A. Ieva, F. 201807\. Joint modeling of recurrent events and survival: a Bayesian non-parametric approach Joint modeling of recurrent events and survival: a Bayesian non-parametric approach. Biostatistics2111-14. https://doi.org/10.1093/biostatistics/kxy026 10.1093/biostatistics/kxy026
* Pitman (1996) pitman1996somePitman, J. 1996\. Some developments of the Blackwell-MacQueen urn scheme Some developments of the blackwell-macqueen urn scheme. TS. Ferguson, LS. Shapley JB. MacQueen (), Statistics, probability and game theory Statistics, probability and game theory ( Volume 30, 245–267). Hayward, CAInstitute of Mathematical Statistics. https://doi.org/10.1214/lnms/1215453576 10.1214/lnms/1215453576
* Qin (1998) qin1998nonparametricQin, L. 1998\. Nonparametric Bayesian models for item response data Nonparametric Bayesian models for item response data . The Ohio State University.
* Rasch (1990) rasch1990probabilisticRasch, G. 1990\. Probabilistic Models for Some Intelligence and Attainment Tests. Probabilistic models for some intelligence and attainment tests. Copenhagen: Danish Institute for Educational Research.
* Reise Rodriguez (2016) reise2016itemReise, S. Rodriguez, A. 2016\. Item response theory and the measurement of psychiatric constructs: some empirical and conceptual issues and challenges Item response theory and the measurement of psychiatric constructs: some empirical and conceptual issues and challenges. Psychological Medicine46102025–2039.
* Rutkowski . (2010) rutkowski2010internationalRutkowski, L., Gonzalez, E., Joncas, M. von Davier, M. 2010\. International large-scale assessment data: issues in secondary analysis and reporting International large-scale assessment data: issues in secondary analysis and reporting. Educational Researcher392142–151.
* Samejima (1997) samejima1997departureSamejima, F. 1997\. Departure from normal assumptions: a promise for future psychometrics with substantive mathematical modeling Departure from normal assumptions: a promise for future psychometrics with substantive mathematical modeling. Psychometrika624471–493.
* San Martín . (2011) sanmartin2011San Martín, E., Jara, A., Rolin, JM. Mouchart, M. 2011Jul01. On the Bayesian Nonparametric Generalization of IRT-Type Models On the Bayesian nonparametric generalization of IRT-type models. Psychometrika763385–409. https://doi.org/10.1007/s11336-011-9213-9 10.1007/s11336-011-9213-9
* Schmitt . (2006) schmitt2006semiSchmitt, JE., Mehta, PD., Aggen, SH., Kubarych, TS. Neale, MC. 2006\. Semi-nonparametric methods for detecting latent non-normality: A fusion of latent trait and ordered latent class modeling Semi-nonparametric methods for detecting latent non-normality: A fusion of latent trait and ordered latent class modeling. Multivariate Behavioral Research414427–443.
* Seong (1990) seong1990sensitivitySeong, T. 1990\. Sensitivity of marginal maximum likelihood estimation of item and ability parameters to the characteristics of the prior ability distributions Sensitivity of marginal maximum likelihood estimation of item and ability parameters to the characteristics of the prior ability distributions. Applied psychological measurement143299–311.
* Sethuraman (1994) sethuraman1994Sethuraman, J. 1994\. A constructive definition of Dirichlet priors A constructive definition of Dirichlet priors. Statistica Sinica42639–650.
* Shaby Wells (2010) shaby2010Shaby, B. Wells, M. 2010\. Exploring an adaptive Metropolis algorithm Exploring an adaptive Metropolis algorithm Technical Report 1011-14. Duke University Department of Statistical Science.
* Sheng (2010) sheng2010sensitivitySheng, Y. 2010\. A sensitivity analysis of Gibbs sampling for 3PNO IRT models: Effects of prior specifications on parameter estimates A sensitivity analysis of gibbs sampling for 3PNO IRT models: Effects of prior specifications on parameter estimates. Behaviormetrika37287–110.
* Smits . (2020) smits2020studySmits, N., Öğreden, O., Garnier-Villarreal, M., Terwee, CB. Chalmers, RP. 2020\. A study of alternative approaches to non-normal latent trait distributions in item response theory models used for health outcome measurement A study of alternative approaches to non-normal latent trait distributions in item response theory models used for health outcome measurement. Statistical Methods in Medical Research2941030–1048.
* Sosa Rodrìguez (2021) sosa2021latentSosa, J. Rodrìguez, A. 2021\. A latent space model for cognitive social structures data A latent space model for cognitive social structures data. Social Networks6585 - 97.
* Stan Development Team (2018) stan2018Stan Development Team. 2018\. Stan Modeling Language Users Guide and Reference Manual, Version 2.18.0. Stan modeling language users guide and reference manual, version 2.18.0. http://mc-stan.org
* Vats . (2019) vats2019multivariateVats, D., Flegal, JM. Jones, GL. 2019\. Multivariate output analysis for Markov chain Monte Carlo Multivariate output analysis for Markov chain Monte Carlo. Biometrika1062321–337.
* Ware (2003) ware2003sf36Ware, JE. 2003\. SF-36 health survey: Manual and interpretation guide Sf-36 health survey: Manual and interpretation guide..
* Watanabe Opper (2010) watanabe2010asymptoticWatanabe, S. Opper, M. 2010\. Asymptotic equivalence of Bayes cross validation and widely applicable information criterion in singular learning theory. Asymptotic equivalence of Bayes cross validation and widely applicable information criterion in singular learning theory. Journal of Machine Learning Research1112.
* Woods (2007) woods2007empiricalWoods, CM. 2007\. Empirical histograms in item response theory with ordinal data Empirical histograms in item response theory with ordinal data. Educational and Psychological Measurement67173–87.
* Woods Thissen (2006) woods2006itemWoods, CM. Thissen, D. 2006\. Item response theory with estimation of the latent population distribution using spline-based densities Item response theory with estimation of the latent population distribution using spline-based densities. Psychometrika712281.
* Yang Dunson (2010) yang2010bayesianYang, M. Dunson, DB. 2010\. Bayesian semiparametric structural equation models with latent variables Bayesian semiparametric structural equation models with latent variables. Psychometrika754675–693.
* Yang . (2010) yang2010semiparametricYang, M., Dunson, DB. Baird, D. 2010\. Semiparametric Bayes hierarchical models with mean and variance constraints Semiparametric Bayes hierarchical models with mean and variance constraints. Computational Statistics & Data Analysis5492172–2186.
|
# The number of prime factors of integers with dense divisors
Andreas Weingartner Department of Mathematics, 351 West University Boulevard,
Southern Utah University, Cedar City, Utah 84720, USA<EMAIL_ADDRESS>
(Date: November 11, 2021)
###### Abstract.
We show that for integers $n$, whose ratios of consecutive divisors are
bounded above by an arbitrary constant, the normal order of the number of
prime factors is $C\log\log n$, where $C=(1-e^{-\gamma})^{-1}=2.280...$ and
$\gamma$ is Euler’s constant. We explore several applications and resolve a
conjecture of Margenstern about practical numbers.
###### 2010 Mathematics Subject Classification:
11N25, 11N37
## 1\. Introduction
We say that a positive integer $n$ is $t$-dense if the ratios of consecutive
divisors of $n$ do not exceed $t$. Let $\mathcal{D}(x,t)$ denote the set of
$t$-dense integers $n\leq x$ and write $D(x,t)=|\mathcal{D}(x,t)|$. Let
$\omega(n)$ (resp. $\Omega(n)$) be the number of prime factors of $n$, counted
without (resp. with) multiplicity.
Theorem 1 gives the average and normal order of $\omega(n)$ and $\Omega(n)$
for the $t$-dense integers. We write $\log_{2}x$ for $\log\log x$ and define
$E(x,t)$, the approximate expected value of $\omega(n)$ on $\mathcal{D}(x,t)$,
by
$E(x,t):=C\log_{2}x-(C-1)\log_{2}t,\quad C:=(1-e^{-\gamma})^{-1}=2.280291...$
###### Theorem 1.
Let $\xi(x)\to\infty$. Uniformly for $x\geq t\geq 2$,
$\frac{\sum_{n\in\mathcal{D}(x,t)}\omega(n)}{D(x,t)}=E(x,t)+O\left(1\right)$
(1)
and
$\Bigl{|}\left\\{n\in\mathcal{D}(x,t):|\omega(n)-E(x,t)|>\xi(x)\sqrt{\log_{2}x}\right\\}\Bigr{|}\ll\frac{D(x,t)}{\xi(x)^{2}}.$
(2)
These results also hold with $\Omega$ in place of $\omega$.
Note that $\log_{2}x\leq E(x,t)\leq C\log_{2}x+O(1)$, as $x\geq t\geq 2$. If
$t=x$, then $E(x,t)=\log_{2}x$ and $\mathcal{D}(x,t)=[1,x]\cap\mathbb{N},$ so
Theorem 1 contains the well-known results about the average and normal order
of $\omega(n)$ on $\mathbb{N}$. If $t\geq 2$ is constant, then
$E(x,t)=C\log_{2}x+O(1)$, so that the average and normal order of $\omega(n)$
for $t$-dense integers $n$ is $C\log_{2}n$.
The $t$-dense integers are a special case of a family of integer sequences
that arise as follows. Let $\theta$ be an arithmetic function. Let
$\mathcal{B}=\mathcal{B}_{\theta}$ be the set of positive integers containing
$n=1$ and all those $n\geq 2$ with prime factorization
$n=p_{1}^{\alpha_{1}}\cdots p_{k}^{\alpha_{k}}$, $p_{1}<p_{2}<\ldots<p_{k}$,
which satisfy
$p_{i}\leq\theta\big{(}p_{1}^{\alpha_{1}}\cdots
p_{i-1}^{\alpha_{i-1}}\big{)}\qquad(1\leq i\leq k).$ (3)
We write $\mathcal{B}(x)=\mathcal{B}\cap[1,x]$ and $B(x)=|\mathcal{B}(x)|$.
When $\theta(n)=nt$, then $\mathcal{B}$ is the set of $t$-dense integers [6,
10, 14]. If $\theta(n)=\sigma(n)+1$, where $\sigma(n)$ is the sum of the
positive divisors of $n$, then $\mathcal{B}$ is the set of practical numbers
[2, 6, 7, 8, 9, 10, 14], i.e. integers $n$ such that every $m\leq n$ can be
expressed as a sum of distinct positive divisors of $n$.
We will derive Corollaries 1 through 3 from Theorem 1 in Section 2.
###### Corollary 1.
Assume $\theta$ satisfies $\max(2,n)\leq\theta(n)\leq nf(n)$, where $f$ is
non-decreasing. Let $\xi(x)\to\infty$. We have
$\frac{\sum_{n\in\mathcal{B}(x)}\omega(n)}{B(x)}=C\log_{2}x\left\\{1+O\left(\left(\frac{\log
f(x)}{\log_{2}x}\right)^{1/3}\right)\right\\}$ (4)
and
$\Bigl{|}\bigl{\\{}n\in\mathcal{B}(x):|\omega(n)-C\log_{2}x|>\xi(x)\sqrt{\log_{2}x}\bigr{\\}}\Bigr{|}\ll
B(x)\frac{\log f(x)}{\xi(x)^{2}}.$ (5)
These results also hold with $\Omega$ in place of $\omega$.
Conjecture 5 of Margenstern [2] proposes that, for practical numbers,
$\sum_{n\in\mathcal{B}(x)}\omega(n)\sim\mu x/(\log x)^{\eta}$, for some
constants $\mu>0$ and $1/2<\eta<1$. The estimate (4) disproves this
conjecture, since $B(x)\sim c_{\theta}x/\log x$ by [14, Thm. 1.2], if
$\max(2,n)\leq\theta(n)\ll n(\log 2n)/(\log_{2}3n)^{1+\varepsilon}$ for $n\geq
1$.
Corollary 2 shows that almost all large practical numbers $n$ have about
$C\log_{2}n$ prime factors.
###### Corollary 2.
If $\theta$ satisfies $\max(2,n)\leq\theta(n)\ll n(\log 2n)^{o(1)}$, then the
average and normal order of $\omega(n)$ on $\mathcal{B}$ is $C\log_{2}n$. That
is, as $x\to\infty$,
$\sum_{n\in\mathcal{B}(x)}\omega(n)\sim\sum_{n\in\mathcal{B}(x)}C\log_{2}n$
(6)
and all but $o(B(x))$ integers $n\in\mathcal{B}(x)$ satisfy
$\omega(n)=(1+o(1))C\log_{2}n.$ (7)
These results also hold with $\Omega$ in place of $\omega$.
Let $\tau(n)$ be the number of positive divisors of $n$.
###### Corollary 3.
If $\theta$ satisfies $\max(2,n)\leq\theta(n)\ll n(\log 2n)^{o(1)}$, then
$(\log x)^{C\log 2-o(1)}<\frac{\sum_{n\in\mathcal{B}(x)}\tau(n)}{B(x)}\ll(\log
x)^{e\log 2},$ (8)
as $x\to\infty$, and all but $o(B(x))$ integers $n\in\mathcal{B}(x)$ satisfy
$\tau(n)=(\log n)^{C\log 2+o(1)}=(\log n)^{1.580577...+o(1)}.$ (9)
Conjecture 4 of Margenstern [2] says that, in the case of practical numbers,
$\sum_{n\in\mathcal{B}(x)}\tau(n)\sim\nu x(\log x)^{\delta}$, for constants
$1/2<\nu,\,\delta<1$. The estimate (8) implies $\delta\in[C\log 2-1,e\log
2-1]=[0.580...,0.884...]$, since $B(x)\asymp x/\log x$. In [17], we prove this
conjecture with $\delta=0.713...$ and some constant $\nu>0$.
The next two corollaries are improvements to the lower bounds of Theorems 1
and 3 of [5]. The proofs of both of these theorems rely on the fact that
almost all $n\in\mathcal{B}$ satisfy $\Omega(n)<(e+\varepsilon)\log_{2}n$.
With Corollary 2, this can be improved to
$\Omega(n)<(C+\varepsilon)\log_{2}n$, under the assumption $\theta(n)\ll
n(\log n)^{o(1)}$. In the lower bound of [5, Thm. 1] for the count of
practical numbers that are also shifted primes, which has an exponent of
$(e+1)\log(e+1)-e\log(e)+1+\varepsilon=3.16470...$, we can replace $e$ by $C$
to get $3.01711...$.
###### Corollary 4.
Fix a nonzero integer $h$ and assume $\theta$ satisfies
$\max(2,n)\leq\theta(n)\ll n(\log 2n)^{o(1)},\quad\theta(mn)\ll
m^{O(1)}\theta(n)\quad(n,m\in\mathbb{N}).$ (10)
We have
$\frac{x}{(\log x)^{3.01712}}\ll_{h}\bigl{|}\\{p\leq x:p\mbox{ prime},\
p-h\in\mathcal{B}\\}\bigr{|}\ll_{h}\frac{x}{(\log x)^{2}},$
where $h$ is not divisible by $\prod_{p\leq\theta(1)}p$ in the lower bound.
Similarly, in the lower bound of [5, Thm. 3] for the count of twin practical
numbers, which has an exponent of $2+4e\log 2+\varepsilon=9.53667...$, we can
replace $e$ by $C$ to get $8.32230...$
###### Corollary 5.
Fix a nonzero integer $h$ and assume $\theta$ satisfies (10). We have
$\frac{x}{(\log x)^{8.32231}}\ll_{h}\bigl{|}\\{n\leq
x:n\in\mathcal{B},n+h\in\mathcal{B}\\}\bigr{|}\ll_{h}\frac{x}{(\log x)^{2}}.$
For the lower bound, assume that (i) $n\in\mathcal{B}$ and $m\leq 3n/|h|$
imply $mn\in\mathcal{B}$, and (ii) if $\theta(1)<3$, then $h\in 2\mathbb{Z}$
if $\theta(2)\geq 3$, and $h\in 4\mathbb{Z}$ if $\theta(2)<3$.
Conditions (i) and (ii) in Corollary 5 are satisfied by the practical numbers
and by the $2$-dense integers for any nonzero even integer $h$, and by the
$t$-dense integers for any nonzero integer $h$, provided $t\geq 3$.
The $\varphi$-practical numbers [4, 12] are integers $n$ such that $x^{n}-1$
has a divisor in $\mathbb{Z}[x]$ of every degree up to $n$. Although not an
example of a set $\mathcal{B}_{\theta}$, they are a superset of
$\mathcal{B}_{\theta_{1}}$ with $\theta_{1}(n)=n+1$, and a subset of
$\mathcal{B}_{\theta_{2}}$ with $\theta_{2}(n)=n+2$. Therefore, Corollaries 1
through 5 also apply to the $\varphi$-practical numbers, provided $h$ is odd
in the lower bound of Corollary 4, while $h$ is even in the lower bound of
Corollary 5.
Theorem 1 is a consequence of Theorems 2 and 3. Theorem 2 gives an estimate
for
$D_{q}(x)=D_{q}(x,t):=|\\{n\in\mathcal{D}(x,t):q|n\\}|,$
when $q$ has a bounded number of prime factors. As in the case $q=1$ (see [14,
Thm. 1.3]), the main term contains the function $d(v)$, which is defined by
$d(v)=0$ for $v<0$ and
$d(v)=1-\int_{0}^{\frac{v-1}{2}}\frac{d(u)}{u+1}\
w\left(\frac{v-u}{u+1}\right)du\qquad(v\geq 0),$ (11)
where $w(u)$ denotes Buchstab’s function.
###### Theorem 2.
Let $k\in\mathbb{N}\cup\\{0\\}$ be fixed. Uniformly for $x\geq 1$, $t\geq 2$,
$q\in\mathbb{N}$ with $\Omega(q)=k$, $v=\log x/\log t$, we have
$\displaystyle D_{q}(x,t)$
$\displaystyle=xd(v)\eta_{q,t}\left\\{1+O_{k}\left(\frac{1}{\log
xt}+\frac{\log 2q\log qt}{\log^{2}xt}\right)\right\\}+O(1),$ (12)
$\displaystyle D_{q}(x,t)$
$\displaystyle=xd(v)\eta_{q,t}\left\\{1+O_{k}\left(\frac{\log 2q}{\log
2x}\right)\right\\},$ (13)
where
$q^{-1}\ll\eta_{q,t}\ll_{k}q^{-1}.$ (14)
###### Corollary 6.
Let $k\in\mathbb{N}\cup\\{0\\}$ be fixed. Uniformly for $x\geq 1$, $t\geq 2$,
$q\in\mathbb{N}$ with $\Omega(q)=k$, we have
$D_{q}(x,t)=\frac{c_{q}x}{\log xt}\left\\{1+O_{k}\left(\frac{1}{\log
xt}+\frac{\log^{2}qt}{\log^{2}xt}\right)\right\\}+O(1),$ (15)
where
$c_{q}=c_{q,t}=C\eta_{q,t}\log t,\quad q^{-1}\log t\ll c_{q}\ll_{k}q^{-1}\log
t,$ (16)
and
$D_{q}(x,t)\ll_{k}\frac{x\log t}{q\log xt}.$ (17)
The estimates (15) and (16) follow from (12) and (14), since
$d(v)=C(v+1)^{-1}\left\\{1+O\left((v+1)^{-2}\right)\right\\}$ by [13, Thm. 1].
The upper bound (17) follows from (13), because $D_{q}(x,t)=0$ if $q>x$.
Theorem 3 gives estimates for $c_{q}$ when $q$ is a prime or a product of two
primes. These estimates are needed to derive Theorem 1 from Theorem 2.
###### Theorem 3.
Let $p\leq q$ be primes. The constant factor in (15) satisfies
$c_{\theta}:=c_{1}=C(\log t-\gamma)+O\left(e^{-\sqrt{\log t}}\right),$ (18)
$qc_{q}=Cc_{\theta}\left\\{1+O\left(\frac{1}{\log
q}+\frac{\log^{2}t}{\log^{2}q}\right)\right\\}$ (19) $qc_{q}=c_{\theta}+C\log
q+O\left(\exp\left(-\sqrt{\log t}\right)\right)\quad(q\leq t),$ (20)
$pqc_{pq}=C^{2}c_{\theta}\left\\{1+O\left(\frac{1}{\log
p}+\frac{\log^{2}t}{\log^{2}p}+\frac{\log^{2}p}{\log^{2}q}\right)\right\\},$
(21) $pqc_{pq}=\left(Cc_{\theta}+C^{2}\log
p\right)\left\\{1+O\left(\frac{1}{\log
q}+\frac{\log^{2}t}{\log^{2}q}+e^{-\sqrt{\log t}}\right)\right\\}\quad(p\leq
t),$ (22) $pqc_{pq}=c_{\theta}+C\log pq+O\left(e^{-\sqrt{\log
t}}\right)\quad(p\leq q\leq t).$ (23)
In Section 2 we derive Corollaries 1, 2 and 3 from Theorem 1. Section 3
contains several lemmas, used in the proofs of Theorems 2 and 3, about members
of $\mathcal{B}$ that are multiples of a natural number $q$. The proof of
Theorem 2 is given in Section 4. In Section 5 we establish Theorem 3 with the
help of Corollary 6, which is a consequence of Theorem 2. Finally, in Section
6 we apply Theorems 2 and 3 to prove Theorem 1.
## 2\. Proof of Corollaries 1, 2 and 3
###### Lemma 1.
We have
$D(x,t)\ll\frac{x\log t}{\log xt}\qquad(x>1/t,t\geq 2),$ $D(x,t)\gg\frac{x\log
t}{\log xt}\qquad(x\geq 1,t\geq 2),$ $D(x/q,t)-D(q/t,t)\leq D_{q}(x,t)\leq
D(x/q,qt)\qquad(x\geq 0,t\geq 2,q\geq 1).$
###### Proof.
The first two estimates follow from [6, Thm. 1].
If $m\in\mathcal{D}(x/q,t)$ and $m>q/t$, then $mq\in\mathcal{D}_{q}(x,t)$.
This shows that $D(x/q,t)-D(q/t,t)\leq D_{q}(x,t)$.
If $n\in\mathcal{D}(x,t)$ and $q|n$, we write $n=qm$ and observe that
$m\in\mathcal{D}(x/q,qt)$. Thus, $D_{q}(x,t)\leq D(x/q,qt)$. ∎
###### Proof of Corollary 1.
We first show (5). For $1\leq n\leq x$, we have $\max(2,n)\leq\theta(n)\leq
nf(n)\leq nf(x)$. Thus, $\mathcal{B}(x)\subset\mathcal{D}(x,f(x))$. If
$f(x)\leq x$, (2) yields
$\left|\left\\{n\in\mathcal{B}(x):|\omega(n)-E(x,f(x))|>\frac{\xi(x)}{2}\sqrt{\log_{2}x}\right\\}\right|\ll\frac{D(x,f(x))}{\xi(x)^{2}}.$
The assumption $\theta(n)\geq\max(2,n)$ implies $B(x)\gg x/\log x$, by [14,
Thm. 1.2]. By Lemma 1,
$\frac{D(x,f(x))}{\xi(x)^{2}}\ll\frac{x\log f(x)}{\xi(x)^{2}\log x}\ll
B(x)\frac{\log f(x)}{\xi(x)^{2}}.$
The result being trivial if $\log f(x)>\xi(x)^{2}$, we may assume $\log
f(x)\leq\xi(x)^{2}$, so that
$|E(x,f(x))-C\log_{2}x|=|(C-1)\log_{2}f(x)|\leq\frac{\xi(x)}{2}\sqrt{\log_{2}x},$
for $x\geq x_{0}$. Thus, (5) holds if $f(x)\leq x$. If $f(x)>x$, then
$\xi(x)^{2}\geq\log f(x)>\log x$, so that
$|\omega(n)-C\log_{2}x|>\xi(x)\sqrt{\log_{2}x}$ implies
$\omega(n)>\xi(x)>\sqrt{\log x}$. The result now follows from Nicolas’ Theorem
[3], an asymptotic estimate for the quantity $|\\{n\leq x:\Omega(n)=k\\}|$,
which easily implies
$|\\{n\leq x:\Omega(n)\geq y\log_{2}x\\}|\ll\frac{x}{(\log x)^{y\log 2-1}},$
(24)
uniformly for $x\geq 2$ and $y\geq 2+\delta$, for any fixed $\delta>0$.
Next, we show that (5) implies (4). From (24) we have
$|\\{n\leq x:\omega(n)\geq 6\log_{2}x\\}|\leq|\\{n\leq x:\Omega(n)\geq
6\log_{2}x\\}|\ll\frac{B(x)}{\log^{2}x}.$ (25)
Since $\omega(n)\leq\Omega(n)\ll\log n$, the contribution to
$\sum_{n\in\mathcal{B}(x)}\omega(n)$ from $n$ with $\omega(n)>6\log_{2}x$ is
$\ll B(x)/\log x$, while the contribution from $n$ with $\omega(n)\leq
6\log_{2}x$ and $|\omega(n)-C\log_{2}x|>\xi(x)\sqrt{\log_{2}x}$ is
$\ll(6\log_{2}x)B(x)\frac{\log f(x)}{\xi(x)^{2}},$
by (5). The contribution to $\sum_{n\in\mathcal{B}(x)}\omega(n)$ from $n$ with
$\omega(n)\leq 6\log_{2}x$ and
$|\omega(n)-C\log_{2}x|\leq\xi(x)\sqrt{\log_{2}x}$ is
$B(x)\left(1+O\left(\frac{\log
f(x)}{\xi(x)^{2}}+\frac{1}{\log^{2}x}\right)\right)C\log_{2}x\left(1+O\left(\frac{\xi(x)}{\sqrt{\log_{2}x}}\right)\right).$
If $\log f(x)\leq\log_{2}x$, (4) now follows with $\xi(x)=(\log
f(x))^{1/3}(\log_{2}x)^{1/6}$. If $\log f(x)>\log_{2}x$, (4) follows directly
from (25). The argument works the same with $\Omega(n)$ in place of
$\omega(n)$. ∎
###### Proof of Corollary 2.
Assume $\max(2,n)\leq\theta(n)\ll n(\log 2n)^{o(1)}$. Define $f(x)=\max_{n\leq
x}\theta(n)/n$, so that $f$ is non-decreasing and $f(x)=(\log x)^{o(1)}$ as
$x\to\infty$, that is $\log f(x)=o(\log_{2}x)$. The relation (6) follows from
(4). Choosing $\xi(x)=(\log f(x)\log_{2}x)^{1/4}$ in (5) yields (7). ∎
###### Lemma 2.
Let $\varepsilon>0$. For $2\leq\alpha\leq 4-\varepsilon$ we have
$\sum_{n\leq x\atop\Omega(n)\geq\alpha\log_{2}x}\tau(n)\ll x(\log
x)^{\alpha(\log 2-\log\alpha+1)-1}.$
###### Proof.
This is a variation of Exercise 05 in [1]. Write
$y^{\Omega(n)}=\sum_{d|n}f(d)$, so that $f(n)$ is multiplicative and
$f(p^{k})=y^{k}(1-1/y)$ for $k\geq 1$, by Möbius inversion. For $0\leq y\leq
2-\varepsilon$,
$\begin{split}\sum_{n\leq x}\tau(n)y^{\Omega(n)}&=\sum_{n\leq
x}\tau(n)\sum_{d|n}f(d)\leq\sum_{d\leq x}f(d)\tau(d)\sum_{m\leq x/d}\tau(m)\\\
&\leq x\log x\sum_{d\leq x}f(d)\tau(d)/d\leq x\log x\sum_{P^{+}(d)\leq
x}f(d)\tau(d)/d\\\ &=x\log x\prod_{p\leq x}\sum_{k\geq
0}f(p^{k})\tau(p^{k})/p^{k}\ll x(\log x)^{2y-1}.\end{split}$
If $1\leq y\leq 2-\varepsilon$, we get
$\sum_{n\leq x\atop\Omega(n)\geq\alpha\log_{2}x}\tau(n)y^{\alpha\log_{2}x}\ll
x(\log x)^{2y-1}.$
The result now follows with $y=\alpha/2$. ∎
###### Proof of Corollary 3.
Since $2^{\omega(n)}\leq\tau(n)\leq 2^{\Omega(n)}$ for all $n\geq 1$, the
estimate (9) and the lower bound in (8) follow at once from (7). For the upper
bound in (8), we write
$\sum_{n\in\mathcal{B}(x)\atop\Omega(n)\leq e\log_{2}x}\tau(n)\leq
B(x)2^{e\log_{2}x}=B(x)(\log x)^{e\log 2}$
and
$\sum_{n\in\mathcal{B}(x)\atop\Omega(n)\geq e\log_{2}x}\tau(n)\leq\sum_{n\leq
x\atop\Omega(n)\geq e\log_{2}x}\tau(n)\ll x(\log x)^{e\log 2-1}\ll B(x)(\log
x)^{e\log 2},$
by Lemma 2 with $\alpha=e$. ∎
## 3\. Multiples of $q$ in $\mathcal{B}$
In this section we develop some general identities for sets $\mathcal{B}$,
defined by (3), with
$\theta:\mathbb{N}\to\mathbb{R}\cup\\{\infty\\},\quad\theta(1)\geq
2,\quad\theta(n)\geq P^{+}(n)\quad(n\geq 2),$ (26)
where $P^{+}(n)$ denotes the largest prime factor of $n$. Let
$\Phi(x,y)=1_{x\geq 1}+|\\{2\leq n\leq x:P^{-}(n)>y\\}|,$
where $P^{-}(n)$ denotes the smallest prime factor of $n$. Let
$\psi(n):=\begin{cases}1&\mbox{if }n\in\mathcal{B}\\\
0&\mbox{else.}\end{cases}$
and define
$\lambda_{n}(s):=\frac{\psi(n)}{n^{s}}\prod_{p\leq\theta(n)}\left(1-\frac{1}{p^{s}}\right),\quad\lambda_{n}:=\lambda_{n}(1),$
$\mu_{n}(s):=\sum_{p\leq\theta(n)}\frac{\log p}{p^{s}-1}-\log
n,\quad\mu_{n}:=\mu_{n}(1).$
###### Lemma 3.
Let $\theta$ satisfy (26) and let $q_{1}\leq q_{2}\leq\ldots\leq q_{k}$ be
primes. For $x\geq 0$,
$\sum_{n\geq 1}\psi(n)\Phi\left(\frac{x}{n},\theta(n)\right)=\lfloor x\rfloor$
(27)
and
$\sum_{n\geq 1\atop q_{1}\cdots
q_{k}|n}\psi(n)\Phi\left(\frac{x}{n},\theta(n)\right)=\sum_{\theta(n)\geq
q_{k}\atop q_{1}\cdots
q_{k-1}|n}\psi(n)\Phi\left(\frac{x}{nq_{k}},\theta(n)\right).$ (28)
###### Proof.
The relation (27) is [15, Lemma 3]. We will show (28). Every $m\in q_{1}\cdots
q_{k}\mathbb{N}$ factors uniquely as $m=nr$ where $n\in\mathcal{B}$ and
$P^{-}(r)>\theta(n)$ if $r>1$. If $q_{1}\nmid n$ then $\theta(n)<q_{1}$. If
$q_{1}|n$, let $j$ be the largest index such that $q_{1}\cdots q_{j}|n$, so
that $q_{j+1}>\theta(n)$ if $j<k$. We count all integer multiples of
$q_{1}\cdots q_{k}$ up to $x$ according to $j$ and $n$:
$\left\lfloor\frac{x}{q_{1}\cdots
q_{k}}\right\rfloor=\sum_{\theta(n)<q_{1}}\psi(n)\Phi\left(\frac{x}{nq_{1}\cdots
q_{k}},\theta(n)\right)\\\ +\sum_{j=1}^{k-1}\sum_{\theta(n)<q_{j+1}\atop
q_{1}\cdots q_{j}|n}\psi(n)\Phi\left(\frac{x}{nq_{j+1}\cdots
q_{k}},\theta(n)\right)+\sum_{q_{1}\cdots
q_{k}|n}\psi(n)\Phi\left(\frac{x}{n},\theta(n)\right).$ (29)
We can now establish (28) by induction on $k$. When $k=1$, (29) and (27) yield
(28). For the inductive step, we write the inner sum of (29) as
$\sum_{\theta(n)<q_{j+1}\atop q_{1}\cdots q_{j}|n}=\sum_{q_{1}\cdots
q_{j}|n}-\sum_{\theta(n)\geq q_{j+1}\atop q_{1}\cdots q_{j}|n}$
and use the inductive hypothesis on the sum $\sum_{q_{1}\cdots q_{j}|n}$ to
get
$\sum_{\theta(n)<q_{j+1}\atop q_{1}\cdots
q_{j}|n}\psi(n)\Phi\left(\frac{x}{nq_{j+1}\cdots
q_{k}},\theta(n)\right)=\sum_{\theta(n)\geq q_{j}\atop q_{1}\cdots
q_{j-1}|n}\psi(n)\Phi\left(\frac{x}{nq_{j}\cdots q_{k}},\theta(n)\right)\\\
-\sum_{\theta(n)\geq q_{j+1}\atop q_{1}\cdots
q_{j}|n}\psi(n)\Phi\left(\frac{x}{nq_{j+1}\cdots q_{k}},\theta(n)\right).$
Thus, the sum over $j$ in (29) is telescoping and the result follows from
(27). ∎
###### Lemma 4.
Let $\theta$ satisfy (26) and let $q_{1}\leq q_{2}\leq\ldots\leq q_{k}$ be
primes. For $\operatorname{Re}(s)>1$ we have
$\sum_{n\geq 1}\lambda_{n}(s)=1$ (30)
and
$\sum_{n\geq 1\atop q_{1}\cdots
q_{k}|n}\lambda_{n}(s)=\frac{1}{q_{k}^{s}}\sum_{\theta(n)\geq q_{k}\atop
q_{1}\cdots q_{k-1}|n}\lambda_{n}(s).$ (31)
Both relations hold at $s=1$ if $B(x)=o(x)$.
###### Proof.
The relation (30) is [16, Lemma 1] when $\operatorname{Re}(s)>1$ and [15,
Theorem 1] when $s=1$. The proof of (31) mirrors that of (28). We first assume
$\operatorname{Re}(s)>1$. Every $m\in q_{1}\cdots q_{k}\mathbb{N}$ factors
uniquely as $m=nr$ where $n\in\mathcal{B}$ and $P^{-}(r)>\theta(n)$ if $r>1$.
If $q_{1}\nmid n$ then $\theta(n)<q_{1}$. If $q_{1}|n$, let $j$ be the largest
index such that $q_{1}\cdots q_{j}|n$, so that $q_{j+1}>\theta(n)$ if $j<k$.
We rearrange the terms of the Dirichlet series $\sum_{q_{1}\cdots
q_{k}|m}m^{-s}$ according to $n$ and $j$. After dividing by $\zeta(s)$, this
shows that, for $\operatorname{Re}(s)>1$,
$\frac{1}{(q_{1}\cdots
q_{k})^{s}}=\sum_{\theta(n)<q_{1}}\frac{\lambda_{n}(s)}{(q_{1}\cdots
q_{k})^{s}}+\sum_{j=1}^{k-1}\sum_{\theta(n)<q_{j+1}\atop q_{1}\cdots
q_{j}|n}\frac{\lambda_{n}(s)}{(q_{j+1}\cdots q_{k})^{s}}+\sum_{q_{1}\cdots
q_{k}|n}\lambda_{n}(s).$ (32)
We establish (31) by induction on $k$. When $k=1$, the result follows from
applying (30) to the first sum of (32). For the inductive step, note that the
inner sum in (32) is
$\begin{split}\sum_{\theta(n)<q_{j+1}\atop q_{1}\cdots
q_{j}|n}\frac{\lambda_{n}(s)}{(q_{j+1}\cdots q_{k})^{s}}&=\sum_{q_{1}\cdots
q_{j}|n}\frac{\lambda_{n}(s)}{(q_{j+1}\cdots q_{k})^{s}}-\sum_{\theta(n)\geq
q_{j+1}\atop q_{1}\cdots q_{j}|n}\frac{\lambda_{n}(s)}{(q_{j+1}\cdots
q_{k})^{s}}\\\ &=\sum_{\theta(n)\geq q_{j}\atop q_{1}\cdots
q_{j-1}|n}\frac{\lambda_{n}(s)}{(q_{j}\cdots q_{k})^{s}}-\sum_{\theta(n)\geq
q_{j+1}\atop q_{1}\cdots q_{j}|n}\frac{\lambda_{n}(s)}{(q_{j+1}\cdots
q_{k})^{s}},\\\ \end{split}$
by the inductive hypothesis. Thus, the sum over $j$ in (32) is a telescoping
sum and the result follows from (30).
If $B(x)=o(x)$, the validity of (31) at $s=1$ follows from (28), in much the
same way that the validity of (30) at $s=1$ follows from (27), which is
demonstrated in the proof of [15, Thm. 1]. ∎
###### Lemma 5.
Let $\theta$ satisfy (26) and let $q_{1}\leq q_{2}\leq\ldots\leq q_{k}$ be
primes. For $\operatorname{Re}(s)>1$ we have
$\sum_{n\geq 1}\lambda_{n}(s)\mu_{n}(s)=0$ (33)
and
$\sum_{n\geq 1\atop q_{1}\cdots
q_{k}|n}\lambda_{n}(s)\mu_{n}(s)=\frac{1}{q_{k}^{s}}\sum_{\theta(n)\geq
q_{k}\atop q_{1}\cdots q_{k-1}|n}\lambda_{n}(s)\bigl{(}\mu_{n}(s)-\log
q_{k}\bigr{)}.$ (34)
###### Proof.
Differentiate (30) and (31) with respect to $s$. ∎
## 4\. Proof of Theorem 2
The following estimate for $\Phi(x,y)$, which differs from the one we used in
[14], simplifies the proof of Theorem 2.
###### Lemma 6.
Uniformly, for $x\geq 0$, $y\geq 2$, we have
$\begin{split}\Phi(x,y)&=1_{x\geq 1}+x\prod_{p\leq
y}\left(1-\frac{1}{p}\right)+\frac{x}{\log
y}\\!\left\\{w(u)-e^{-\gamma}\\!-\left.\frac{y}{x}\right|_{x\geq
y}\\!\\!+O\left(\frac{e^{-u/3}}{\log y}\right)\\!\right\\}\\\ &=1_{x\geq
1}+x\prod_{p\leq y}\left(1-\frac{1}{p}\right)+O\left(\frac{xe^{-u/3}}{\log
y}\right),\end{split}$
where $u=\frac{\log\max(1,x)}{\log y}$ and $w(u)$ is Buchstab’s function.
###### Proof.
The second estimate follows from the first, since $w(u)-e^{-\gamma}\ll e^{-u}$
and, if $x\geq y\geq 2$, then $y/x\ll e^{-u/2}$.
When $x\geq y\geq 2$ and $\log y\geq(\log_{2}x)^{2}$, the first estimate
follows from combining equations (49), (52), (59) and (60) of [11, Sec.
III.6], with equation (6) of [11, Sec. III.5], where we estimate the integral
in (52) as
$\int_{0}^{\infty}|w^{\prime}(u-v)|y^{-v}dv\ll\int_{0}^{\infty}e^{-(u-v)/2}y^{-v}dv=\frac{e^{-u/2}}{\log
y-1/2}\asymp\frac{e^{-u/2}}{\log y}.$
When $x\geq y\geq 2$ and $\log y<(\log_{2}x)^{2}$, then $u\gg\sqrt{\log x}$
and the result follows from [11, Thm. III.6.1 and Thm. III.5.1].
When $x<y$, then $\Phi(x,y)=1_{x\geq 1}$ and $w(u)=0$, so that the result
follows from Mertens’ formula [11, Thm. I.1.11]. ∎
###### Lemma 7.
Assume $B(x)=B_{t}(x)$ is the counting function of a set
$\mathcal{B}_{t}\subset\mathbb{N}$ that depends on the parameter $t$. Assume
$B(x)=x\int_{1}^{\infty}\frac{B(y)}{y^{2}\log
yt}\left(e^{-\gamma}-w\left(\frac{\log x/y}{\log
yt}\right)\right)dy+R(x)\quad(x\geq 1),$
such that the integrals
$\alpha_{t}:=e^{-\gamma}\int_{1}^{\infty}\frac{B(y)}{y^{2}\log
yt}dy,\quad\beta_{t}:=\frac{-1}{\log t}\int_{1}^{\infty}R(y)\frac{dy}{y^{2}}$
converge. Then
$B(x)=x\eta_{t}d(v)+O\Bigl{\\{}1+x\beta_{t}(v+1)^{-3.03}+R(x)+I(x)+J(x)\Bigr{\\}},$
where $v=\log x/\log t$, $d(v)$ is given by (11),
$\eta_{t}=\alpha_{t}+\beta_{t}$,
$I(x)=\frac{x}{\log xt}\int_{x}^{\infty}R(y)\frac{dy}{y^{2}},\quad
J(x)=\frac{x}{(\log xt)^{3.03}}\int_{1}^{x}R(y)(\log
yt)^{2.03}\frac{dy}{y^{2}}.$
###### Proof.
We follow the second half of the proof of [14, Thm. 1.3]. The only
modification needed is the use of the improved estimates
$(v+1)d(v)=C+O((v+1)^{-2.03}),\quad(v\geq 0),$ (35)
and
$(v+1)^{2}d^{\prime}(v)=-C+O((v+1)^{-2.03}),\quad(v\geq 0).$ (36)
The estimate (35) is a consequence of [13, Cor. 6], while (36) follows from
inserting (35) in the proof of [13, Cor. 5]. In [14], we used slightly weaker
estimates for simplicity, with an exponent of $-2$ instead of $-2.03$ in the
error terms. In the proof of Theorem 2, the improved exponent will save a
factor of $\log_{2}x$ (when estimating the contribution from $R_{2}(x)$ to
$J(x)$). ∎
For $n\in\mathbb{N}$ with prime factorization $n=p_{1}\cdots p_{k}$, where
$p_{1}\leq\ldots\leq p_{k}$, define
$F(n):=\max_{1\leq j\leq k}p_{j}^{2}p_{j+1}\cdots p_{k}.$
###### Lemma 8.
If $x<\max(m,F(m)/t)$, then $D_{m}(x,t)=0$.
###### Proof.
Note that $n\in\mathcal{D}_{m}(x,t)$ if and only if $n\leq x$, $m|n$ and
$F(n)\leq nt$. Also, $m|n$ implies $F(m)\leq F(n)$. Thus, if $D_{m}(x,t)\neq
0$ and $n\in\mathcal{D}_{m}(x,t)$, then $m\leq n\leq x$ and $F(m)\leq F(n)\leq
nt\leq xt$, so $x\geq\max(m,F(m)/t)$. ∎
###### Lemma 9.
Let $n\geq 2$ with prime factorization $n=p_{1}\cdots p_{k}$,
$p_{1}\leq\ldots\leq p_{k}$. If $F(n)\leq x$, then $p_{k-j+1}\cdots p_{k}\leq
x^{1-2^{-j}}$ for $1\leq j\leq k$.
###### Proof.
We use induction on $j$. When $j=1$, the claim is that $p_{k}\leq x^{1/2}$,
which follows from $p_{k}^{2}\leq F(n)\leq x$, for all $k\geq 1$. Assume now
that the claim is correct for some $j\in\mathbb{N}$ and all $k\geq j$. Let
$k\geq j+1$. We have $p_{k-j}^{2}p_{k-j+1}\cdots p_{k}\leq F(n)\leq x$. Thus,
if $p_{k-j}p_{k-j+1}\cdots p_{k}\geq x^{1-2^{-j-1}}$, then $p_{k-j}\leq
x^{2^{-j-1}}$. By the inductive hypothesis,
$p_{k-j}(p_{k-j+1}\cdots p_{k})\leq x^{2^{-j-1}}x^{1-2^{-j}}=x^{1-2^{-j-1}},$
for all $k\geq j+1$. ∎
###### Lemma 10.
Let $k\geq 0$ be fixed. For $m\geq 1$ with $\Omega(m)=k$, $t\geq 2$ and
$x\geq\max(m,F(m)/t)$, we have
$\frac{x\log t\log 2m}{m\log xt\log 2x}\gg_{k}1.$
###### Proof.
This is obvious if $m=1$ or if $t\geq x^{2^{-k}}$. If $m\geq 2$ and
$t<x^{2^{-k}}$, then $F(m)\leq xt$ and Lemma 9 imply
$m\leq(xt)^{1-2^{-k}}<x^{1-4^{-k}}\ll_{k}x/(\log xt)^{2}$, from which the
claim follows. ∎
###### Proof of Theorem 2.
In the remainder of this paper, we write $D_{q}(x)$ for $D_{q}(x,t)$ and
$D(x)$ for $D(x,t)$. We will show by induction on $k\geq 0$ that, for
$k=\Omega(q)$, the estimates (12), (13) and (14) hold and that, for primes $r$
with $r\geq P^{+}(q)$, we have
$\sum_{n\in\mathcal{D}_{q}\atop n\geq
r/t}\Phi(x/rn,nt)-\frac{x}{r}\sum_{n\in\mathcal{D}_{q}\atop n\geq
r/t}\lambda_{n}\ll 1+\frac{x\log t\log qr\log
qrt}{qr\log^{3}xt}=:R_{2}(x,qr),$ (37)
for $x\geq\max(qr,F(qr)/t)$. Note that (13) and (35) imply
$D_{q}(x)\ll_{k}xd(v)\eta_{q,t}\asymp_{k}\frac{x\log t}{q\log xt},\quad(x\geq
1,q\geq 1,t\geq 2),$ (38)
since $D_{q}(x)=0$ if $q>x$.
When $k=0$, $q=1$, (12) is [14, Eq. (13)], (13) is [14, Thm. 1.3] and (14) is
[14, Eq. (6)]. To show (37) for $k=0$ and $q=1$, assume that $r$ is prime and
$x\geq\max(r,r^{2}/t)$. Equations (27) and (30) show that
$\begin{split}&\sum_{n\in\mathcal{D}\atop n\geq
r/t}\Phi(x/rn,nt)-\frac{x}{r}\sum_{n\in\mathcal{D}\atop n\geq
r/t}\lambda_{n}\\\ &=-\\{x/r\\}-\sum_{n\in\mathcal{D}\atop
n<r/t}\Phi(x/rn,nt)+\frac{x}{r}\sum_{n\in\mathcal{D}\atop n<r/t}\lambda_{n}\\\
&\ll 1+D(r/t)+\sum_{n\in\mathcal{D}\atop n<r/t}\frac{x}{rn\log
nt}\exp\left(-\frac{\log xt/r}{3\log nt}\right),\end{split}$ (39)
by the second estimate in Lemma 6. We have $D(r/t)\ll r\log t/(t\log r)$ by
Lemma 1, so $D(r/t)\ll R_{2}(x,r)$ follows from
$\frac{r}{\log^{2}r\log
rt}\ll\frac{xt/r}{\log^{3}xt}\asymp\frac{xt/r}{\log^{3}(xt/r)},$
since $xt/r\geq\sqrt{xt}$. This holds because $r\leq xt/r$. Finally, the last
sum in (39) is $\ll R_{2}(x,r)$ by Lemma 1. Thus, (37) holds for $k=0$.
For the inductive step, assume that (13) (and hence (38)) and (14) hold for
$q\in\mathbb{N}$ with $\Omega(q)=k$ for some $k\geq 0$. If $k\geq 1$, assume
that (37) holds for $\Omega(q)=k-1$. Let $r$ be a prime with $r\geq P^{+}(q)$
and write $m=qr$. We note that in the remainder of this proof, all implied
constants in the $\ll$ and big-O notation may depend on $k$.
We estimate the first sum in (28) with Lemma 6 and apply (31) to get
$\begin{split}D_{m}(x)&+\sum_{n\in\mathcal{D}_{m}}\frac{x}{n\log
nt}\left\\{w\left(\frac{\log x/n}{\log
nt}\right)-e^{-\gamma}-\left.\frac{n^{2}t}{x}\right|_{n^{2}\leq\frac{x}{t}}\\!+O\left(\frac{e^{-\frac{\log
x/n}{3\log nt}}}{\log nt}\right)\right\\}\\\ &=\sum_{n\in\mathcal{D}_{q}\atop
n\geq r/t}\Phi(x/rn,nt)-\frac{x}{r}\sum_{n\in\mathcal{D}_{q}\atop n\geq
r/t}\lambda_{n}.\end{split}$ (40)
The contribution from the last two terms in the first sum in (40) is
$\ll\tilde{R}_{1}(x):=\frac{x\log mt}{m(\log xt)^{2}},$
by Lemma 1. In the second application of this argument we will be able to
replace $\tilde{R}_{1}(x)$ by the smaller
$R_{1}(x):=\frac{x\log t}{m(\log xt)^{2}}.$
The error from applying Abel summation to the remaining terms of the first sum
in (40) is also $\ll\tilde{R}_{1}(x)$, since $w(u)-e^{-\gamma}\ll e^{-u}$ and
$w^{\prime}(u)\ll e^{-u}$. Thus,
$\begin{split}D_{m}(x)=&x\int_{1}^{\infty}\frac{D_{m}(y)}{y^{2}\log
yt}\left(e^{-\gamma}-w\left(\frac{\log x/y}{\log
yt}\right)\right)dy+O(\tilde{R}_{1}(x))\\\ &+\sum_{n\in\mathcal{D}_{q}\atop
n\geq r/t}\Phi(x/rn,nt)-\frac{x}{r}\sum_{n\in\mathcal{D}_{q}\atop n\geq
r/t}\lambda_{n}.\end{split}$ (41)
If $x/r<\max(q,F(q)/t,r/t)$, that is $x<\max(m,F(m)/t)$, then Lemma 8 shows
that the first sum in (41) vanishes, while the second sum is
$\ll\frac{x}{m}\min\left(1,\frac{\log t}{\log
r}\right)\ll\frac{x}{m}\min\left(1,\frac{\log t}{\log m}\right),$
by the inductive hypothesis (38) and since $\log r\leq\log m\leq\log
r^{k+1}\ll_{k}\log r$. Define
$R_{2}(x):=\begin{cases}\frac{x}{m}\min\left(1,\frac{\log t}{\log
m}\right)&\text{if }x<\max(m,F(m)/t)\\\ 1+\frac{x\log t\log m\log
mt}{m\log^{3}xt}&\text{if }x\geq\max(m,F(m)/t).\end{cases}$
Thus,
$D_{m}(x)=x\int_{1}^{\infty}\frac{D_{m}(y)}{y^{2}\log
yt}\left(e^{-\gamma}-w\left(\frac{\log x/y}{\log yt}\right)\right)dy+R(x)$
(42)
holds with $R(x)\ll\tilde{R}_{1}(x)+R_{2}(x)$ when $x<\max(m,F(m)/t)$.
Assume now that $x\geq\max(m,F(m)/t)$. If $k=0$ and $q=1$, we have already
shown that the last row of (41) is $\ll R_{2}(x)$. If $k\geq 1$, write $q=uv$
where $v=P^{+}(q)$. Equations (28) and (31) show that the last row of (41)
equals
$\begin{split}=&\sum_{n\in\mathcal{D}_{u}\atop n\geq
v/t}\Phi(x/nrv,nt)-\frac{x}{rv}\sum_{n\in\mathcal{D}_{u}\atop n\geq
v/t}\lambda_{n}-\sum_{n\in\mathcal{D}_{q}\atop
n<r/t}\Phi(x/nr,nt)+\frac{x}{r}\sum_{n\in\mathcal{D}_{q}\atop
n<r/t}\lambda_{n}\\\ =&S-T-(U-V),\end{split}$
say. Now $x\geq\max(m,F(m)/t)$ implies $x/r\geq\max(q,F(q)/t,r/t)$. The
inductive hypothesis (37) yields
$S-T\ll 1+\frac{(x/r)\log t\log q\log qt}{q\log^{3}(xt/r)}\ll 1+\frac{x\log
t\log m\log mt}{m\log^{3}xt}=R_{2}(x),$
since $xt/r\geq\sqrt{xt}$. The second estimate in Lemma 6 and (38) show that
$\begin{split}U-V&\ll D_{q}(r/t)+\sum_{n\in\mathcal{D}_{q}\atop
n<r/t}\frac{x}{nr\log nt}\exp\left(-\frac{\log\frac{x}{nr}}{3\log
nt}\right)\\\ &\ll\frac{r^{2}\log t}{mt\log r}+R_{2}(x)\ll
R_{2}(x),\end{split}$
where the last assertion is implied by
$\frac{r}{\log r\log m\log mt}\ll\frac{xt/r}{\log^{3}(xt/r)},$
which holds because $mt>m\geq r$ and $r\leq xt/r$. As the last row of (41) is
$\ll R_{2}(x)$, we have established that (37) holds for $\Omega(q)\leq k$ and
that (42) holds with $R(x)\ll\tilde{R}_{1}(x)+R_{2}(x)$.
We need to estimate $I(x)$ and $J(x)$ from Lemma 7. For this purpose we may
assume that $x\geq\max(m,F(m)/t)$. If not, Lemma 8 shows that $D_{m}(x)=0$, so
that (12) holds since the main term in (12) is absorbed by the error terms. We
find that $I(x)+J(x)\ll\tilde{R}_{1}(x)+R_{2}(x)$ and $\beta_{t}\ll\frac{\log
m}{m\log t}$ so that $x\beta_{t}(v+1)^{-3}\ll R_{2}(x)$.
The conclusion of Lemma 7 is that
$D_{m}(x)=x\eta_{m,t}d(v)+O\bigl{(}\tilde{R}_{1}(x)+R_{2}(x)\bigr{)}.$ (43)
The lower bound in Lemma 1 yields $\eta_{m,t}\gg m^{-1}$. Lemma 11 shows that
$\eta_{m,t}\ll_{k}m^{-1}$. Together with (43), Lemmas 8 and 10, this implies
$D_{m}(x)\ll_{k}\frac{x\log t}{m\log xt},\quad(x\geq 1,t\geq 2,\Omega(m)\leq
k+1).$ (44)
Running through this proof a second time with this upper bound replacing the
one in Lemma 1, we can replace $\tilde{R}_{1}(x)$ by $R_{1}(x)$ to obtain
$D_{m}(x)=x\eta_{m,t}d(v)+O\bigl{(}R_{1}(x)+R_{2}(x)\bigr{)},$
where $1/m\ll\eta_{m,t}\ll_{k}1/m$. This shows that (12) and (14) hold with
$q$ replaced by $m=qr$.
To see that (12) implies (13), note that the term $O(1)$ is acceptable by
Lemma 10, provided $x\geq\max(m,F(m)/t)$. If $x<\max(m,F(m)/t)$, then
$D_{m}(x)=0$ and $x<F(m)\leq m^{2}$, so (13) (with $q$ replaced by $m$)
follows from $\log x\leq 2\log m$. This completes the proof of Theorem 2. ∎
###### Lemma 11.
Assume that (43) holds for $m$ with $\Omega(m)\leq k+1$, and that
$\eta_{q,t}\ll_{k}q^{-1}$ for $\Omega(q)\leq k$. Then
$\eta_{m,t}\ll_{k}m^{-1}$ for $\Omega(m)=k+1$.
###### Proof.
Let $m=qr$ where $r$ is prime and $r\geq P^{+}(q)$. Equation (34) yields
$\sum_{n\geq 1\atop qr|n}\lambda_{n}(s)\mu_{n}(s)=\frac{1}{r^{s}}\sum_{n\geq
r/t\atop q|n}\lambda_{n}(s)\bigl{(}\mu_{n}(s)-\log r\bigr{)}.$
As in [16], we let $s=1+1/\log^{2}N$ and split the sum on the left according
to $n\leq N$ and $n>N$. As $N\to\infty$, the contribution from $n\leq N$
converges to (see [16, Lemma 3]) $\sum_{qr|n}\lambda_{n}\mu_{n}$, while the
contribution from $n>N$ converges to (see [16, Lemma 4])
$-c_{qr}(1-e^{-\gamma}),$ by (43), where $\eta_{qr,t}=Cc_{qr}\log t$. Applying
the same reasoning to the sum on the right-hand side, we get
$\sum_{n\geq 1\atop
qr|n}\lambda_{n}\mu_{n}-c_{qr}(1-e^{-\gamma})=\frac{1}{r}\Bigl{(}\sum_{n\geq
r/t\atop q|n}\lambda_{n}\bigl{(}\mu_{n}-\log
r\bigr{)}-c_{q}(1-e^{-\gamma})\Bigr{)}.$
Since $\mu_{n}\ll\log t$ , we obtain
$rc_{qr}\ll S\log t+S\log r+c_{q},\quad S:=r\sum_{n\geq 1\atop
qr|n}\lambda_{n}=\sum_{n\geq r/t\atop q|n}\lambda_{n},$
by (31). Now $\eta_{q,t}\ll q^{-1}$ and (44) holds with $q$ in place of $m$.
Thus, $c_{q}\ll q^{-1}\log t$ and $S=\sum_{n\geq r/t\atop q|n}\lambda_{n}\ll
q^{-1}\min(1,\log t/\log r).$ This shows that $rc_{qr}\ll q^{-1}\log t$, which
is the desired result. ∎
## 5\. Proof of Theorem 3
###### Proof of (18), (19) and (20).
The estimate (18) is [16, Cor. 3]. Equation (34), with $k=1$ and $q_{1}=q$
prime, is
$\sum_{n\geq 1\atop q|n}\lambda_{n}(s)\mu_{n}(s)=\frac{1}{q^{s}}\sum_{n\geq
q/t}\lambda_{n}(s)\bigl{(}\mu_{n}(s)-\log q\bigr{)}.$ (45)
As in [16], we let $s=1+1/\log^{2}N$ and split the sum on the left according
to $n\leq N$ and $n>N$. As $N\to\infty$, the contribution from $n\leq N$
converges to (see [16, Lemma 3]) $\sum_{n\geq 1\atop q|n}\lambda_{n}\mu_{n}$,
while the contribution from $n>N$ converges to (see [16, Lemma 4])
$-c_{q}(1-e^{-\gamma}),$ by Corollary 6. Applying the same reasoning to the
right-hand side of (45), we get
$\sum_{n\geq 1\atop
q|n}\lambda_{n}\mu_{n}-c_{q}(1-e^{-\gamma})=\frac{1}{q}\Bigl{(}\sum_{n\geq
q/t}\lambda_{n}\bigl{(}\mu_{n}-\log
q\bigr{)}-c_{\theta}(1-e^{-\gamma})\Bigr{)}.$
With the estimate [16, Lemma 13]
$\mu_{n}=\log t-\gamma+O\left(e^{-\sqrt{\log nt}}\right),$ (46)
we obtain
$qc_{q}(1-e^{-\gamma})=S\log
q+c_{\theta}(1-e^{-\gamma})+O\left(Se^{-\sqrt{\log\max(q,t)}}\right),$ (47)
where
$S:=\sum_{n\geq q/t}\lambda_{n}=q\sum_{n\geq 1\atop q|n}\lambda_{n},$
by (31). If $q\leq t$, then $S=1$ by (30), so that (20) follows from (47).
In general, we have
$S=\frac{e^{-\gamma}c_{\theta}}{\log q}\left(1+O\left(\frac{1}{\log
q}+\frac{\log^{2}t}{\log^{2}q}\right)\right),\quad(q\geq 2,\,t\geq 2).$ (48)
If $q\leq t$, this holds because $S=1$ and $c_{\theta}\asymp\log t$. If $q\geq
t$, (48) follows from applying Abel summation to $S=\sum_{n\geq
q/t}\lambda_{n}$, together with Mertens’ formula and Corollary 6 (with $q=1$).
Combining (48) with (47) proves (19). ∎
###### Proof of (21), (22) and (23).
Equation (34), with $k=2$ and $q_{1}=p\leq q_{2}=q$ is
$\sum_{n\geq 1\atop pq|n}\lambda_{n}(s)\mu_{n}(s)=\frac{1}{q^{s}}\sum_{n\geq
q/t\atop p|n}\lambda_{n}(s)\bigl{(}\mu_{n}(s)-\log q\bigr{)}.$ (49)
As in [16], we let $s=1+1/\log^{2}N$ and split the sum on the left according
to $n\leq N$ and $n>N$. As $N\to\infty$, the contribution from $n\leq N$
converges to (see [16, Lemma 3]) $\sum_{pq|n}\lambda_{n}\mu_{n}$, while the
contribution from $n>N$ converges to (see [16, Lemma 4])
$-c_{pq}(1-e^{-\gamma}),$ by Corollary 6. Applying the same reasoning to the
right-hand side of (49), we get
$\sum_{n\geq 1\atop
pq|n}\lambda_{n}\mu_{n}-c_{pq}(1-e^{-\gamma})=\frac{1}{q}\Bigl{(}\sum_{n\geq
q/t\atop p|n}\lambda_{n}\bigl{(}\mu_{n}-\log
q\bigr{)}-c_{p}(1-e^{-\gamma})\Bigr{)}.$
We estimate $\mu_{n}$ with (46) to obtain
$qc_{pq}(1-e^{-\gamma})=T\log
q+c_{p}(1-e^{-\gamma})+O\left(T\exp\left(-\sqrt{\log\max(pt,q)}\right)\right),$
(50)
where
$T:=\sum_{n\geq q/t\atop p|n}\lambda_{n}=q\sum_{n\geq 1\atop
pq|n}\lambda_{n},$
by (31).
If $p\leq q\leq t$, then
$T=\sum_{n\geq 1\atop p|n}\lambda_{n}=\frac{1}{p}\sum_{n\geq
p/t}\lambda_{n}=\frac{1}{p}\sum_{n\geq 1}\lambda_{n}=\frac{1}{p},$
by (31) and (30). The estimate (23) now follows from (50) with $T=1/p$ and
$c_{p}$ estimated by (20).
In general, we have
$T=\frac{e^{-\gamma}c_{p}}{\log q}\left(1+O\left(\frac{1}{\log
q}+\frac{\log^{2}pt}{\log^{2}q}\right)\right),\quad(q\geq p\geq 2,\,t\geq 2).$
(51)
If $q<pt$, this is implied by $T\ll c_{p}/\log q$, which follows from (17)
(with $q$ replaced by $p$). If $q\geq pt$, we estimate $\lambda_{n}$ with
Mertens’ formula and use Abel summation and the estimate (15). The
contribution from the first two error terms in (15) is clearly acceptable,
while the term $O(1)$ contributes
$\ll\int_{q/t}^{\infty}\frac{dy}{y^{2}\log yt}\leq\frac{t}{q\log
q}\asymp\frac{c_{p}}{\log q}\cdot\frac{pt}{q\log t}\ll\frac{c_{p}}{\log
q}\cdot\frac{\log^{2}pt}{\log^{2}q}.$
Now substitute (51) into (50) and estimate $c_{p}$ with (19) to get (21), and
with (20) to get (22). ∎
## 6\. Proof of Theorem 1
###### Proof of (1)..
We have
$\sum_{n\in\mathcal{D}(x)}\omega(n)=\sum_{p\leq x}D_{p}(x)=xd(v)\sum_{p\leq
x}\eta_{p,t}\left(1+O\left(\frac{\log p}{\log x}\right)\right),$
by (13). The contribution from the error term is $\ll xd(v)\asymp D(x)$, since
$\eta_{p,t}\ll 1/p$ by (14). Now $\eta_{p,t}C\log t=c_{p}$ and
$\eta_{1,t}C\log t=c_{\theta}$, by Corollary 6. With (18), (19) and (20), we
find that
$\sum_{p\leq x}\eta_{p,t}=\frac{1}{C\log t}\sum_{p\leq
x}c_{p}=\frac{c_{\theta}}{C\log t}(E(x,t)+O(1))=\eta_{1,t}(E(x,t)+O(1)).$
The result now follows from (13) with $q=1$, that is
$D(x)=xd(v)\eta_{1,t}(1+O(1/\log x))$.
To see that (1) remains valid when $\omega$ is replaced by $\Omega$, note that
$\sum_{n\in\mathcal{D}(x)}(\Omega(n)-\omega(n))=\sum_{n\in\mathcal{D}(x)}\sum_{k\geq
2}\sum_{p^{k}|n}1=\sum_{p\leq x}\sum_{k\geq 2}D_{p^{k}}(x)\ll\frac{x\log
t}{\log xt}\ll D(x),$
by Lemma 1. ∎
###### Proof of (2)..
We have
$\sum_{n\in\mathcal{D}(x)}\omega(n)^{2}=\sum_{p,q\leq
x}D_{pq}(x)+O\Bigl{(}\sum_{p\leq x}D_{p}(x)\Bigr{)},$
where $p,q$ run over primes. The last term is $\ll D(x)\log_{2}x$, by (1).
Thus, (13) yields
$\sum_{n\in\mathcal{D}(x)}\omega(n)^{2}=O(D(x)\log_{2}x)+xd(v)\sum_{p,q\leq
x}\eta_{pq,t}\left(1+O\left(\frac{\log pq}{\log x}\right)\right).$
Since $\eta_{pq,t}\ll 1/pq$, the contribution from the error term is $\ll
xd(v)\log_{2}x\ll D(x)\log_{2}x$. With (21), (22) and (23), we find that
$\begin{split}\sum_{p,q\leq x}\eta_{pq,t}=\frac{1}{C\log t}\sum_{p,q\leq
x}c_{pq}&=\frac{c_{\theta}}{C\log t}\bigl{(}E(x,t)^{2}+O(\log_{2}x)\bigr{)}\\\
&=\eta_{1,t}\bigl{(}E(x,t)^{2}+O(\log_{2}x)\bigr{)}.\end{split}$
Combining this with (1) and (13) (with $q=1$), we get
$\sum_{n\in\mathcal{D}(x)}(\omega(n)-E(x,t))^{2}\ll D(x)\log_{2}x,$
which implies (2). This estimate remains valid if $\omega(n)$ is replaced by
$\Omega(n)$, since
$\sum_{n\in\mathcal{D}(x)}(\Omega(n)-\omega(n))^{2}\leq\sum_{p\neq q\leq
x}\sum_{k,j\geq 2}D_{p^{k}q^{j}}(x)+\sum_{p\leq x}\sum_{k\geq
2}2kD_{p^{k}}(x)\ll D(x),$
by Lemma 1. ∎
## Acknowledgments
The author thanks Eric Saias and the anonymous referee for several very
helpful suggestions.
## References
* [1] R. R. Hall and G. Tenenbaum, Divisors, Cambridge University Press, 1988.
* [2] M. Margenstern, Les nombres pratiques: théorie, observations et conjectures, J. Number Theory 37 (1991), 1–36.
* [3] J.-L. Nicolas, Sur la distribution des nombres entiers ayant une quantité fixée de facteurs premiers, Acta Arith. 44 (1984), 191–200.
* [4] C. Pomerance, L. Thompson, A. Weingartner, On integers $n$ for which $X^{n}-1$ has a divisor of every degree, Acta Arith. 175 (2016), no. 3, 225–243.
* [5] C. Pomerance and A. Weingartner, On primes and practical numbers, arXiv: 2007.11062
* [6] E. Saias, Entiers à diviseurs denses 1, J. Number Theory 62 (1997), 163–191.
* [7] W. Sierpiński, Sur une propriété des nombres naturels, Ann. Mat. Pura Appl. (4) 39 (1955), 69–74.
* [8] A. K. Srinivasan, Practical numbers, Current Sci. 17 (1948), 179–180.
* [9] B. M. Stewart, Sums of distinct divisors, Amer. J. Math. 76 (1954), 779–785.
* [10] G. Tenenbaum, Sur un problème de crible et ses applications, Ann. Sci. École Norm. Sup. (4) 19 (1986), 1–30.
* [11] G. Tenenbaum, Introduction to Analytic and Probabilistic Number Theory, Cambridge Studies in Advanced Mathematics, Vol. 46, Cambridge Univ. Press, Cambridge, 1995.
* [12] L. Thompson, Polynomials with divisors of every degree, J. Number Theory 132 (2012), 1038–1053.
* [13] A. Weingartner, Integers with dense divisors 3, J. Number Theory 142 (2014), 211-222.
* [14] A. Weingartner, Practical numbers and the distribution of divisors, Q. J. Math. 66 (2015), no. 2, 743–758.
* [15] A. Weingartner, A sieve problem and its application, Mathematika 63 (2017), no. 1, 213–229.
* [16] A. Weingartner, On the constant factor in several related asymptotic estimates, Math. Comp. 88 (2019), no. 318, 1883–1902.
* [17] A. Weingartner, The mean number of divisors for rough, dense and practical numbers, preprint arXiv:2104.07137.
|
# QCD EQUATION OF STATE AT FINITE CHEMICAL POTENTIALS FOR RELATIVISTIC NUCLEAR
COLLISIONS
AKIHIKO MONNAI Department of Mathematical and Physical Sciences, Japan
Women’s University
Bunkyo-ku, Tokyo 112-8681, Japan
<EMAIL_ADDRESS>BJÖRN SCHENKE Physics Department, Brookhaven National
Laboratory
Upton, New York 11973, USA
<EMAIL_ADDRESS>CHUN SHEN Department of Physics and Astronomy, Wayne State
University
Detroit, Michigan 48201, USA
RIKEN BNL Research Center, Brookhaven National Laboratory
Upton, New York 11973, USA
<EMAIL_ADDRESS>
(Day Month Year)
###### Abstract
We review the equation of state of QCD matter at finite densities. We discuss
the construction of the equation of state with net baryon number, electric
charge, and strangeness using the results of lattice QCD simulations and
hadron resonance gas models. Its application to the hydrodynamic analyses of
relativistic nuclear collisions suggests that the interplay of multiple
conserved charges is important in the quantitative understanding of the dense
nuclear matter created at lower beam energies. Several different models of the
QCD equation of state are discussed for comparison.
###### keywords:
quantum chromodynamics; equation of state; nuclear collision.
PACS numbers:
## 1 Introduction
The collective properties of quantum chromodynamic (QCD) matter have been a
topic of great interest in nuclear physics. A milestone has been the discovery
of the quark-gluon plasma (QGP) [1, 2, 3, 4], a high-temperature phase of QCD,
at the Relativistic Heavy Ion Collider (RHIC) at Brookhaven National
Laboratory (BNL) in the year 2000 [5, 6, 7, 8]. The QGP is speculated to have
filled the universe about $10^{-5}$-$10^{-4}$ seconds after the Big Bang. The
collider experiments have allowed the quantitative study of QCD matter through
comparison of theoretical calculations and experimental data and consequently
provided a glimpse of the early universe. The nearly-perfect fluidity and
rapid thermalization of the QGP are major discoveries and have opened up a
world of possibilities to study thermodynamics of strongly-interacting
elementary particles in collider experiments.
The high-energy frontier has been explored by the Large Hadron Collider (LHC)
at the European Organization for Nuclear Research (CERN), which has been in
operation since 2009 [9, 10, 11]. It has extended our experimental knowledge
of the QCD phase diagram (Fig. 1) in the direction of temperature, getting
closer to the beginning of the universe. The fluidity has been shown to
persist at higher temperatures, though the fluid may become less perfect [12]
as the system would be less strongly-coupled.
Figure 1: A schematic illustration of the QCD phase diagram and the beam
energy scan experiments. The phase structure in the dense regions are
conjectured based on model estimations.
The next frontier on the phase diagram is the high-density regime [13, 14],
where first principles calculations are known to suffer from the fermion sign
problem [15]. Estimations based on the chiral model indicate that the quark-
hadron transition turns from a crossover to a first-order phase transition at
a finite baryon chemical potential, suggesting the existence of a critical
point [16]. Further theoretical model analyses indicate that the QCD phase
structure can be quite nontrivial; possible scenarios include the color
superconducting (CSC) phase at low temperature and high baryon density, where
quarks form a condensate of Cooper pairs [17, 18, 19], the second critical
point at the high-density end of the quark-hadron phase boundary implied by
the QCD axial anomaly [20], the chiral and color superconducting phase
transitions enhanced with vector interaction [21], and the quarkyonic phase,
suggested by studies in the large $N_{c}$ limit [22]. Exploration of the dense
quark matter is of particular importance since the experimental detection of
gravitational waves, emerging from e.g. neutron star mergers, now give more
stringent constraints on the properties of the compact stars themselves,
including the equation of state [23, 24]. See 25, 26 for recent reviews.
The collider experiments are a powerful tool to obtain bottom-up insight into
the QCD matter at finite baryon chemical potential with high precision in
controlled environments (Fig. 2). The Beam Energy Scan programs, being
preformed at BNL RHIC and planned at various facilities including the GSI
Facility for Antiproton and Ion Research (FAIR), JINR Nuclotron-based Ion
Collider fAility (NICA), and JAEA/KEK Japan Proton Accelerator Research
Complex (J-PARC). The heavy-ion programs at BNL Alternating Gradient
Synchrotron (AGS), CERN Super Proton Synchrotron (SPS) and GSI
Schwerionensynchrotron 18 (SIS 18), provide complimentary data for
understanding the properties of dense quark matter.
Figure 2: A schematic illustration of the nuclear collisions at high energies
where the system has larger temperature and smaller baryon density (left) and
at intermediate to low energies where the system has lower temperature and
larger baryon density (right).
One of the most successful models for the description of the dynamical
evolution of QGP is the relativistic hydrodynamic model [27, 28]. The observed
spectra of hadronic particles up to moderate transverse momenta
($p_{T}\lesssim 3\,{\rm GeV}$) are known to be in quantitative agreement with
hydrodynamic model calculations. Azimuthal momentum anisotropies,
characterized with flow harmonics $v_{n}$ [29, 30, 31, 32], are considered to
be one of the most prominent pieces of evidence for the nearly-perfect
fluidity of the produced QCD medium, because they are found to clearly reflect
the geometrical anisotropy of the overlap region of colliding nuclei, implying
that the system is strongly coupled. The physics of QCD enters the model
through the equation of state – and the transport coefficients in off-
equilibrium cases – along with details of initial conditions. Thus, once a
realistic initial geometry is given, it is in principle possible to extract
information on the QCD equation of state by comparing numerical results based
on trial input with experimental data [33, 34, 35, 36, 37, 38].
The equation of state is a fundamental relation among thermodynamic variables.
Earliest studies of the QCD equation of state date back to the MIT bag model
[39, 40], where confinement is introduced phenomenologically. Hadrons are
treated as quarks in bags within the QCD vacuum. Consequently, the model has a
first-order phase transition between the hadron and QGP phases. Since then,
our understanding of QCD thermodynamics has been deepened with the advent of
model approaches such as the potential model [41] and the Nambu-Jona-Lasinio
model [42, 43]. A breakthrough was brought when first principle calculations
became possible with the advancement of the computational method of lattice
QCD (at zero chemical potentials). SU(3) pure glue studies predict a first-
order QCD phase transition while more realistic (2+1)-flavor calculations
suggest a crossover transition, implying the importance of quark contributions
in the phenomenon [44, 45, 46]. State-of-the-art lattice QCD simulations with
a physical pion mass provide high-precision results of the QCD equation of
state over a wide-range of temperatures [47, 48, 49]. The lattice QCD equation
of state, when embedded in a hydrodynamic model, is known to reproduce the
experimental data of nuclear collisions at top RHIC and LHC energies well.
It was considered in the earlier days of QGP phenomenology that the fluidity
appears only at and above energies around $\sqrt{s_{NN}}=\mathcal{O}(10^{2})$
GeV. The net baryon density in most cases was not considered important because
it would be small and have negligible effects for such systems, except at
forward rapidities. As the hydrodynamic model became more sophisticated, on
the other hand, it was rediscovered that the hydrodynamic description can be
valid for hadronic yields at lower energies, down to
$\sqrt{s_{NN}}=\mathcal{O}(10)$ GeV, which is the typical energy scale covered
by the beam energy scan programs[50, 51]. This may be partially owing to the
fact that one has a better understanding of non-equilibrium processes –
initial dynamics of local equilibration, viscosity and diffusion, and hadronic
transport, which occur before, during, and after the hydrodynamic evolution,
respectively – and can now show that hydrodynamics, which is based on the idea
of local equilibrium, is compatible with experimental results.
For a quantitative description of nuclear collisions in the beam energy scan
programs, one needs the equation of state at finite densities [52, 53, 54, 55,
56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69] as input for
hydrodynamic simulations. First principle calculations are known to be
challenging at finite densities, owing to the aforementioned sign problem.
Several intriguing methods to circumvent the sign problem have been proposed,
such as the Taylor expansion method [70, 71], the reweighting method [72, 71,
73], the imaginary chemical potential method [74, 75, 76], the complex
Langevin method [77, 78, 79, 80], the Lefschetz thimble method [81, 82], and
the path optimization method [83, 84], but so far no complete description is
available at small temperatures and large chemical potentials.
In this review, we will discuss the phenomenological construction of the QCD
equation of state at finite chemical potentials for relativistic nuclear
collisions. Oftentimes, only net baryon number is taken into account as the
conserved charge in the equation of state, especially when hydrodynamic
modeling is concerned. We review the neos model [85, 86] based on the lattice
QCD equation of state and susceptibilities at vanishing densities from Refs.
48, 87, 88, 89, 90 (see also Refs. 91, 92, 93, 49, 94), and the hadron
resonance gas equation of state, that include three conserved charges relevant
in nuclear collisions: net baryon (B), electric charge (Q), and strangeness
(S) as a successor to the version including only net baryon chemical
potentials [95, 96, 97, 98, 99, 100, 101]. The selection of conserved charges
is based on the assumption that only light quarks ($u$, $d$, and $s$) would
thermalize in nuclear collisions. It has been employed in recent hydrodynamic
model analyses [102, 103, 104]. A similar approach also has recently been
proposed in Ref. 105, and the importance of multiple conserved charges has
been discussed in various situations before our model realization [106, 107,
108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119]. We demonstrate by
explicit calculations within a hydrodynamic model that the description of
experimental data is improved by our comprehensive treatment of the conserved
charges in the construction of the equation of state. Finally, we compare
different models for the QCD equation of state, both at zero and finite
densities, and present conclusions and summary.
The natural units $c=\hbar=k_{B}=1$ and the mostly-minus Minkowski metric
$g^{\mu\nu}=\mathrm{diag}(+,-,-,-)$ are used.
## 2 Status
We review the status of the study of the QCD equation of state. QCD
thermodynamics is a topic of interest to a broad range of studies from nuclear
physics to particle physics to astrophysics. Here we focus on the
phenomenological equations of state intended for use in hydrodynamic studies
of relativistic nuclear collisions.
### 2.1 From bag model to lattice QCD
Early hydrodynamic models often employed the equation of state inspired by the
MIT bag model, such as EOS Q, which has a first-order phase transition at zero
densities. The hadronic phase is described using a resonance gas and the QGP
phase using a parton gas with a bag constant [120, 121, 122, 27]. A finite net
baryon density was relatively easy to implement in such models, though it was
neglected in many cases, because it would have small effects around mid-
rapidity at top RHIC energies. These equations of state are, despite involving
a first-order phase transition, able to reproduce the experimental data of
hadronic spectra and elliptic flow reasonably well with an appropriate choice
of initial conditions in inviscid models.
Crossover-like equations of state have also been discussed in the literature.
Early studies include a functional parametrization of pioneering lattice QCD
results [123, 124, 125] by matching a parton gas with a pion gas equation of
state and encoding the details of the transition using the choice of the
connection width $\Delta T$ (defined later as in Eq. (8)). A more
sophisticated equation of state was developed by connecting the results of the
hadron resonance gas and an effective theory for finite temperature SU(3)
gauge theory, [126, 127] and was used for viscous hydrodynamic analyses [128].
With the advent of first principle calculations, the connection of the results
of lattice QCD simulations and hadron resonance gas became a topic of interest
[129, 130, 131, 132, 133, 134, 135]. A variation of such approach includes the
quasi-particle model fit to the lattice QCD data [53, 136]. Sometimes the
lattice QCD equation of state is used directly at vanishing densities down
into the hadronic phase, though caution is needed because the energy-momentum
conservation at particlization is no longer automatically guaranteed if the
hadron resonance gas description is not used [137], and the inconsistency may
be hidden by the normalization of initial conditions. Additionally, the
lattice QCD data in the continuum limit typically have uncertainty bands of a
few percent, though they have been improved considerably in recent
simulations.
As mentioned earlier, the lattice-based QCD equation of state is considered to
give an accurate description of the hot matter created and observed in the
collider experiments at RHIC and LHC, where the conserved charges can be
neglected [33, 34, 35, 36, 37, 38]. It is important to next elucidate the high
density regions of the QCD phase diagram for fully utilizing the data from the
ongoing and upcoming beam energy scan programs and for understanding
microscopic properties of the QCD matter near equilibrium.
### 2.2 Equation of state at finite densities
The finite-density version of a hybrid equation of state s95p-v1 [57] is one
of the pioneering studies to use the coefficients of the Taylor expansion
method for construction. A temperature shift was introduced to the
susceptibilities estimated in lattice QCD calculations with larger than
physical pion mass, which tend to produce a higher $T_{c}$ than those with
physical pion mass, for smooth matching to the resonance gas results. As the
lattice QCD calculations improved, one has begun to use the bare result of the
baryon susceptibility in hydrodynamic simulations [138, 139]. The connection
of the lattice QCD results using the physical pion mass to the hadron
resonance gas results has been discussed at finite density of net baryons [96]
and of net baryons, electric charge and strangeness [85, 105].
The implementation of a critical point in the equation of state is also a
topic of importance [16, 140, 141]. The 3-dimensional Ising model is often
used for this purpose because it is considered to be in the same universality
class as QCD [52, 64]. The rescaled magnetic field and the reduced temperature
in the latter model [142] are mapped onto the reduced temperature and (baryon)
chemical potential in QCD, respectively. Experimental elucidation of the
critical point is a long-standing goal to which no complete answer is
available yet [143, 144].
Perturbative QCD calculations have also been improved to include higher order
contributions [145, 146, 147, 148, 149, 150, 151], though the convergence of
the weak-coupling expansion for the pressure is slow, even at the $g^{6}$
order, and the dependence on the renormalization scale is large [152]. In
light of this situation, improved versions of the perturbation theory have
been proposed, such as the two-particle irreducible (2PI) formalism [153, 154,
155] and hard thermal loop (HTL) perturbation theory [156, 157, 158]. There
are quantitative studies to match perturbative results [159, 160, 161, 162,
163] to the hadron resonance gas ones with the help of lattice QCD data to
approach the finite density regime [59].
The effective model approaches to the finite-density phase structure include
the Nambu-Jona-Lasinio model with the Polyakov loop [164, 106, 165, 56],
polyakov loop enhanced quark-meson models [66], and a more phenomenological
quasi-particle model, [54] where the result of Ref. 56 has been employed in
one of the first modern hydrodynamic simulations of the beam energy scan
experiments [115].
An alternative approach to describe strongly-coupled matter is via holographic
gauge-string duality [166, 167, 168]. The original anti-de Sitter/conformal
field theory (AdS/CFT) correspondence is conjectured for the $\mathcal{N}=4$
super Yang-Mills theory, which is scale invariant and thus has no phase
transition. The primary role of the conjecture in the phenomenology of nuclear
collisions is perhaps the prediction of transport coefficients [169, 170, 171,
172], for which first principle calculations are difficult. Extensions of this
method to non-conformal theories have been proposed in order to preserve
consistency with the thermodynamic properties of QCD. Such examples include
the Einstein-Maxwell-Dilation model [173, 174, 175] which can reproduce the
known lattice QCD data.
The QCD equation of state is also a topic of importance for compact stars. The
typical chemical potential is larger and the temperature is smaller in such
systems compared with those in nuclear collisions [176], though it may be
possible to have occasional baryonic dense spots in the latter through event-
by-event fluctuations. There have been extensive studies on the neutron star
equation of state – see, e.g., Refs. 25, 26 for recent reviews. In addition to
the intriguing observation regarding the Shapiro delay[177], the experimental
discovery of gravity waves has brought a plethora of new data, which can
constrain the nuclear equation of state in the cold and dense regime [24].
## 3 NEOS – hybrid QCD equation of state
We discuss the construction of the QCD equation of state at finite chemical
potentials of net baryon number, electric charge, and strangeness. neos is an
equation of state model, which takes one of the latest lattice QCD equations
of state at vanishing chemical potentials as a baseline. Following the Taylor
expansion method, the second- and fourth-order diagonal and off-diagonal
susceptibilities are implemented. This expansion method has the advantage of
being able to express the thermodynamic variables at finite density with those
at zero density. On the other hand, one needs an additional prescription at
low temperatures, because the Taylor expansion becomes less reliable when the
chemical potential over temperature ratio is large [178]. Thus, the hadron
resonance gas model, which is a framework to understand the low-temperature
QCD system in terms of stable hadrons and meta-stable resonances, is used at
lower temperatures, and its pressure is matched to the lattice-based pressure
near the crossover.
There are additional motivations for the connection procedure. First, all the
thermodynamic variables and second- and fourth- order susceptibilities of the
hadron resonance gas model are known to show excellent agreement with those of
lattice QCD. The fact that the hadron resonance gas model shares basic
thermodynamic properties with the first principle calculation motivates one to
assume that the model captures essential physics. Second, the success of
hydrodynamic modeling relies on the hadron resonance gas picture when the flow
field is converted into hadronic particles.111It should be noted that the
experimental data for particle spectra, chemical ratios, and chemical freeze-
out indicate that the concept of temperature is valid near particlization in
nuclear collisions, even though there are arguments regarding
hydrodynamization without thermalization at the earliest stage of hydrodynamic
evolution [179, 180, 181]. In anisotropic hydrodynamics thermodynamic
variables can receive modifications from local momentum anisotropies [182,
183, 184, 185, 186]. The Cooper-Frye prescription [137] for particlization
also requires that the equation of state in the hydrodynamic evolution is the
same as that in the hadronic transport model to allow for energy-momentum and
charge conservation. Direct use of the hadronic resonance gas equation of
state at low temperatures is the most practical way to achieve this.
### 3.1 Lattice QCD equation of state in the Taylor expansion method
The higher temperature side of the equation of state is constructed using the
Taylor-expanded pressure $P_{\mathrm{lat}}$,
$\displaystyle\frac{P_{\mathrm{lat}}}{T^{4}}$ $\displaystyle=$
$\displaystyle\frac{P_{0}}{T^{4}}+\sum_{l,m,n}\frac{\chi^{B,Q,S}_{l,m,n}}{l!m!n!}\bigg{(}\frac{\mu_{B}}{T}\bigg{)}^{l}\bigg{(}\frac{\mu_{Q}}{T}\bigg{)}^{m}\bigg{(}\frac{\mu_{S}}{T}\bigg{)}^{n},$
(1)
where $P_{0}$ and $\chi^{B,Q,S}_{l,m,n}$ are the pressure and $(l+m+n)$-th
order susceptibilities at zero chemical potentials, calculated in lattice QCD
simulations. $T$ is the temperature and $\mu_{B}$, $\mu_{Q}$, $\mu_{S}$ are
the chemical potentials for net baryon, electric charge, and strangeness,
respectively. They are related to the quark chemical potentials as
$\displaystyle\mu_{u}$ $\displaystyle=$
$\displaystyle\frac{1}{3}\mu_{B}+\frac{2}{3}\mu_{Q},$ (2)
$\displaystyle\mu_{d}$ $\displaystyle=$
$\displaystyle\frac{1}{3}\mu_{B}-\frac{1}{3}\mu_{Q},$ (3)
$\displaystyle\mu_{s}$ $\displaystyle=$
$\displaystyle\frac{1}{3}\mu_{B}-\frac{1}{3}\mu_{Q}-\mu_{S}.$ (4)
The susceptibilities can be expressed as
$\chi_{l,m,n}^{B,Q,S}=\frac{\partial^{l}\partial^{m}\partial^{n}P(T,\mu_{B},\mu_{Q},\mu_{S})/T^{4}}{\partial(\mu_{B}/T)^{l}\partial(\mu_{Q}/T)^{m}\partial(\mu_{S}/T)^{n}}\bigg{|}_{\mu_{B,Q,S}=0}.$
(5)
$l+m+n$ is constrained by the matter-antimatter symmetry to be even. One can
alternatively consider isospin instead of electric charge.
### 3.2 Hadron resonance gas equation of state
The hadron resonance gas picture is used for calculating the lower temperature
side of the equation of state. Its pressure reads
$\displaystyle P_{\mathrm{had}}$ $\displaystyle=$ $\displaystyle\pm
T\sum_{i}\int\frac{g_{i}d^{3}p}{(2\pi)^{3}}\ln[1\pm e^{-(E_{i}-\mu_{i})/T}]$
(6) $\displaystyle=$ $\displaystyle\sum_{i}\sum_{k}(\mp
1)^{k+1}\frac{1}{k^{2}}\frac{g_{i}}{2\pi^{2}}m_{i}^{2}T^{2}e^{k\mu_{i}/T}K_{2}\bigg{(}\frac{km_{i}}{T}\bigg{)},$
where $E_{i}=\sqrt{p^{2}+m_{i}^{2}}$ is the energy, $m_{i}$ is the mass,
$g_{i}$ is the degeneracy, and $\mu_{i}$ is the chemical potential of the
$i$-th hadronic species. $\mu_{i}$ can be expressed as
$\mu_{i}=B_{i}\mu_{B}+Q_{i}\mu_{Q}+S_{i}\mu_{S}$ using the quantum numbers
$B_{i}$, $Q_{i}$, and $S_{i}$ for net baryon, electric charge, and
strangeness. $K_{2}(x)$ is the modified Bessel function of the second kind.
The expansion with $k$ takes account of the correction of quantum statistics.
It is usually sufficient to consider the $k\leq 3$ terms for pions, the $k\leq
2$ terms for kaons, and the $k=1$ term for heavier particles. The Boltzmann
limit corresponds to the $k=1$ case. We treat hadrons as on-shell particles
and do not include spectral functions for the resonance states in our model
[187].
### 3.3 Hybrid equation of state
The neos equation of state is obtained by connecting the ones from lattice QCD
and the hadron resonance gas model. The pressure is given as
$\displaystyle\frac{P}{T^{4}}$ $\displaystyle=$
$\displaystyle\frac{1}{2}[1-f(T,\mu_{B},\mu_{Q},\mu_{S})]\frac{P_{\mathrm{had}}(T,\mu_{B},\mu_{Q},\mu_{S})}{T^{4}}$
(7) $\displaystyle+$
$\displaystyle\frac{1}{2}[1+f(T,\mu_{B},\mu_{Q},\mu_{S})]\frac{P_{\mathrm{lat}}(T,\mu_{B},\mu_{Q},\mu_{S})}{T^{4}},$
where the connecting function $f$ should satisfy $f\to 1$ and $f\to-1$ in the
high and low temperature limits, respectively. Here we choose a smooth
hyperbolic function
$f(T,\mu_{B},\mu_{Q},\mu_{S})=\tanh\bigg{[}\frac{T-T_{c}(\mu_{B})}{\Delta
T_{c}}\bigg{]}.$ (8)
$T_{c}(\mu_{B})$ is the connecting temperature for which we use
$T_{c}(\mu_{B})=0.16\ \mathrm{GeV}-0.4\,(0.139\
\mathrm{GeV}^{-1}\mu_{B}^{2}+0.053\ \mathrm{GeV}^{-3}\mu_{B}^{4})$ motivated
by the $\mu_{B}$ dependence of the chemical freeze-out line [188]. The
connecting width is chosen to be $\Delta T_{c}=0.1T_{c}(0)$. The dependencies
on electric charge and strangeness chemical potentials are assumed to be small
and neglected here. The choices of possible parameter values and their effects
are limited for the following reasons. First, the thermodynamic conditions
$\displaystyle\frac{\partial^{2}P}{\partial T^{2}}$ $\displaystyle=$
$\displaystyle\frac{\partial s}{\partial T}>0,$ (9)
$\displaystyle\frac{\partial^{2}P}{\partial\mu_{B,Q,S}^{2}}$ $\displaystyle=$
$\displaystyle\frac{\partial n_{B,Q,S}}{\partial\mu_{B,Q,S}}>0.$ (10)
have to be imposed near the connection range because they would no longer be
trivially satisfied when two different frameworks are being connected by
another function. The procedure leaves a narrow window for the possible choice
of parameters. Second, the fact that the lattice QCD and hadron resonance gas
equations of state match over a finite temperature range implies that the
overall thermodynamic properties of the system do not and should not depend on
the detailed parameter choice.
The above procedure gives a crossover equation of state by construction. One
may argue that there could be a critical point in the accessible range of the
QCD phase diagram. We consider the crossover-type equation of state here to
allow for baseline hydrodynamic calculations without critical behavior. Future
experimental observation of deviations from that baseline can then be analyzed
to deduce the existence and location of the QCD critical point. If one
introduces $f$ with a non-differentiable kink, an equation of state with a
first-order phase transition is easily obtained [69].
It is useful here to introduce basic thermodynamic relations for estimating
other macroscopic variables. The entropy density $s$, the net baryon, electric
charge, and strangeness densities $n_{B,Q,S}$, the energy density $e$, and the
sound velocity $c_{s}$ are obtained via
$\displaystyle s$ $\displaystyle=$ $\displaystyle\left.\frac{\partial
P}{\partial T}\right|_{\mu_{B},\mu_{Q},\mu_{S}},\ \ n_{B}=\left.\frac{\partial
P}{\partial\mu_{B}}\right|_{T,\mu_{Q},\mu_{S}},$ (11) $\displaystyle n_{Q}$
$\displaystyle=$ $\displaystyle\left.\frac{\partial
P}{\partial\mu_{Q}}\right|_{T,\mu_{B},\mu_{S}},\ \ n_{S}=\left.\frac{\partial
P}{\partial\mu_{S}}\right|_{T,\mu_{B},\mu_{Q}},$ (12) $\displaystyle e$
$\displaystyle=$ $\displaystyle Ts-P+\mu_{B}n_{B}+\mu_{Q}n_{Q}+\mu_{S}n_{S},$
(13) $\displaystyle c_{s}^{2}$ $\displaystyle=$
$\displaystyle\left.\frac{\partial P}{\partial
e}\right|_{n_{B},n_{Q},n_{S}}+\frac{n_{B}}{e+P}\left.\frac{\partial
P}{\partial n_{B}}\right|_{e,n_{Q},n_{S}}$ (14) $\displaystyle+$
$\displaystyle\frac{n_{Q}}{e+P}\left.\frac{\partial P}{\partial
n_{Q}}\right|_{e,n_{B},n_{S}}+\frac{n_{S}}{e+P}\left.\frac{\partial
P}{\partial n_{S}}\right|_{e,n_{B},n_{Q}},$
for the system with multiple conserved charges.
### 3.4 Multiple charges in nuclear collisions
The strangeness density in nuclear collisions on average is vanishing because
the colliding nuclei are net strangeness free. This is called the strangeness
neutrality condition. The condition leads to positive strangeness chemical
potential in the presence of positive baryon chemical potential, because the
number of strange quarks would exceed that of anti-quarks in the QGP phase if
$\mu_{S}=0$ was assumed. An interpretation based on the parton picture is that
$\mu_{S}\sim\mu_{B}/3$ follows from Eq. (4) when $\mu_{Q}\sim 0$. For the
hadronic phase, the strangeness chemical potential can be suppressed because
the lightest baryon with strangeness is $\Lambda$, the mass of which is
already large compared with the temperature of the system.
The electric charge density is related to the net baryon density via the
proton-to-nucleon number ratio $Z/A$. $Z/A$ of the nuclei used or planned in
the collider experiments at RHIC and LHC are listed in Table 3.4. The
primarily-used heavy ions Au and Pb have $Z/A\approx 0.4$. For neutron-rich
nuclei, the chemical potential of $d$ quarks is larger than that of $u$
quarks, i.e., $\mu_{d}=\mu_{B}/3-\mu_{Q}/3>\mu_{u}=\mu_{B}/3+2\mu_{Q}/3$,
which implies that $\mu_{Q}<0$ when $\mu_{B}>0$ in the QGP phase. The trend
remains in the hadronic phase because negative pions would be abundant
compared with positive pions, which leads to
$\mu_{\pi^{-}}=-\mu_{Q}>\mu_{\pi^{+}}=\mu_{Q}$. One would have the opposite
situation $\mu_{Q}>0$ for proton-rich nuclei, which are relevant in smaller
systems.
Number ratios of protons to nucleons $Z/A$ for the nuclei used or planned at
RHIC and LHC. Nucleus ${}^{1}_{1}$H ${}^{2}_{1}$H ${}^{3}_{2}$He
${}^{8}_{16}$O ${}^{27}_{13}$Al ${}^{63}_{29}$Cu $Z/A$ 1.000 0.500 0.667 0.500
0.481 0.460 Nucleus ${}^{96}_{40}$Zr ${}^{96}_{44}$Ru ${}^{127}_{54}$Xe
${}^{197}_{\ 79}$Au ${}^{208}_{\ 82}$Pb ${}^{238}_{\ 92}$U $Z/A$ 0.417 0.458
0.425 0.401 0.394 0.387
### 3.5 Numerical construction
Results of (2+1)-flavor lattice QCD simulations are used to evaluate the
pressure [48] and the second- and fourth-order susceptibilities [87, 88, 89,
90] at vanishing densities in the numerical construction of the hybrid
equation of state. In addition, $\chi_{6}^{B}$, $\chi_{5,1}^{B,Q}$, and
$\chi_{5,1}^{B,S}$ of the sixth-order susceptibilities are phenomenologically
introduced for a proper matching of the thermodynamic variables because the
results of the Taylor expansion method of lattice QCD simulations cannot be
naïvely used when they have large error bars, as small displacement of the
crossover temperature can lead to unphysical gaps in thermodynamic quantities
when $\mu_{B}/T$ is large. The Stefan-Boltzmann limits are used as anchors on
the high temperature side so the basic thermodynamic properties are preserved
when lattice QCD data points are scarce. Those treatments could be improved in
the future when more data become available. The functional forms for the
parametrization of all the susceptibilities used in the model are found in
Ref. 85.
The hadron resonance gas model includes all the hadrons and resonances which
have $u$, $d$ and/or $s$ as constituent components and have masses smaller
than 2 GeV in the Particle Data Group list [187]. The pressure and
susceptibilities up to the fourth order are found to agree well with those of
lattice QCD calculations.
The following three situations are simulated: (i) the conventional situation
$\mu_{S}=\mu_{Q}=0$ where only the net baryon number is considered as
conserved charge, (ii) the situation with the strangeness neutrality condition
$n_{S}=0$ and vanishing electric charge chemical potential $\mu_{Q}=0$, and
(iii) the realistic situation in collisions of heavy nuclei where $n_{S}=0$
and $n_{Q}=0.4\,n_{B}$. They are labeled as neos B, neos BS, and neos BQS,
respectively, in the article.
Figure 3: (a) The dimensionless pressure $P/T^{4}$ and (b) the dimensionless
strangeness density $-n_{S}/T^{3}$ of neos B, (c) the dimensionless pressure
$P/T^{4}$ and (d) the strangeness chemical potential $\mu_{S}$ of neos BS, and
(e) the dimensionless pressure $P/T^{4}$ and (f) the electric charge chemical
potential $-\mu_{Q}$ of neos BQS as functions of $T$ and $\mu_{B}$ [85]. The
solid, long-dashed, dash-dotted, and short-dashed lines indicate the constant
$s/n_{B}$ trajectories at 420, 144, 51, and 30, respectively.
The dimensionless pressure $P/T^{4}$ as a function of $T$ and $\mu_{B}$ is
shown in Fig. 3 (a) where $\mu_{S}=\mu_{Q}=0$ (neos B). The trajectories of
the constant entropy density to net baryon density ratio $s/n_{B}$ indicate
the typical trajectory in the $T$-$\mu_{B}$ plane explored by collider
experiments at each center-of-mass energy, because the net baryon density and
– in the ideal hydrodynamic approximation – entropy density are conserved
during the hydrodynamic evolution. $s/n_{B}=420,144,51$, and $30$ correspond
to $\sqrt{s_{NN}}=200,62.4,19.6$, and $14.5$ GeV [76], respectively. It should
be noted that there will be a range of $s/n_{B}$ for every collision since the
medium is spatially inhomogeneous. Also, event-by-event fluctuations further
smear the trajectories on the phase diagram. $\mu_{B}/T$ is fixed on those
trajectories when $s\sim T^{3}$ and $n_{B}\sim\mu_{B}T^{2}$ in the QGP phase.
Once the trajectories enter the hadronic phase, they are bent toward larger
$\mu_{B}$ because protons, the lightest baryons, are considerably heavier than
pions. While this situation leads to a thermodynamically consistent crossover
equation of state, it does not reflect the situation in nuclear collisions
because the strangeness neutrality condition is violated as shown in Fig. 3
(b). The negative strangeness density is consistent with the expectation that
positive $\mu_{B}$ leads to a system with more $s$ quarks and fewer $\bar{s}$
quarks. It approaches zero on the low temperature side because the lightest
hadrons with strangeness are kaons, whose mass is non-negligible in the
hadronic phase.
Once the strangeness neutrality condition $n_{S}=0$ is imposed, the pressure
is meaningfully modified in the region where the $\mu_{B}/T$ is relatively
large, as demonstrated in Fig. 3 (c). Figure 3 (d) shows that the strangeness
neutrality condition leads to positive strangeness chemical potentials. The
trajectories are shifted to the larger $\mu_{B}$ side by about 50% in the QGP
phase, because only $u$ and $d$ quarks contribute to $n_{B}$ in neos BS
instead of $u,d$, and $s$ quarks in neos B, because in strangeness neutral
systems strange quarks and antiquarks do not contribute to the net baryon
number. The larger values of $\mu_{B}$ can be important in the hydrodynamic
model because baryon diffusion, which is primarily driven by the spatial
gradient of $\mu_{B}/T$ [138, 189, 99], would be enhanced. The differences of
the trajectories in neos B and BS are smaller in the hadronic phase because,
as mentioned earlier, the lightest hadron with net baryon number and
strangeness is the $\Lambda$ baryon, which is already heavy compared with the
medium temperature.
Figure 4: The thick solid, thin solid, and thick dotted lines are the sound
velocity squared of neos B, BS, and BQS, respectively, as a function of
temperature along the trajectories of (a) $s/n_{B}=420$ and (b) $s/n_{B}=30$.
[85].
Finally, we study the situation of matter with fixed electric charge-to-baryon
ratio $n_{Q}/n_{B}=0.4$ and strangeness neutrality. Shown in Fig. 3 (e) and
(f) are the dimensionless pressure and electric chemical potential,
respectively. The pressure does not change much going from neos BS to BQS and
neither do the trajectories, because $\mu_{Q}=0$ implies $n_{Q}/n_{B}\sim 0.5$
which happens to be not too far from the more realistic situation. The
negative electric chemical potential, nevertheless, is important in heavy-ion
phenomenology, as it presents a quantitative explanation for the abundance of
negative pions over positive pions observed in the experiments [190, 191,
192].
Figure 4 shows the sound velocity as a function of temperature for neos B, BS,
and BQS. The sound velocity has a minimum because the pressure does not change
significantly as a function of the energy density in the vicinity of the
quark-hadron crossover (Eq. 14). Comparing the low baryon density
($s/n_{B}=420$) and high baryon density ($s/n_{B}=30$) results, the sound
velocity in the hadronic phase is found to be suppressed and the minimum is
shifted toward the lower temperature side for larger densities by a few MeV.
The strangeness neutrality condition slightly increases the sound velocity,
while the realistic electric charge-to-baryon ratio leads to negligible
change. The quantity approaches the Stefan-Boltzmann limit $c_{s}^{2}=1/3$ at
high temperatures, it reaches 94.8% at $T=0.6$ GeV and 97.2% at $T=0.8$ GeV of
the limit for $s/n_{B}=420$.
Figure 5: Isopressure planes in the chemical potential space in (a) the
hadronic phase where $P/T^{4}=0.8$ and $T=0.14$ GeV and (b) the QGP phase
where $P/T^{4}=2$ and $T=0.2$ GeV [85].
The isopressure surface at constant temperatures in the chemical potential
space is investigated to illustrate the interplay of multiple conserved
charges. The numerical result in the hadronic phase where $P/T^{4}=0.8$ and
$T=0.14$ GeV is shown in Fig. 5 (a). The intercepts can be defined as
$P(\mu_{B}^{\mathrm{int}},0,0)=P(0,\mu_{Q}^{\mathrm{int}},0)=P(0,0,\mu_{S}^{\mathrm{int}})$.
They are ordered as
$\mu_{B}^{\mathrm{int}}>\mu_{S}^{\mathrm{int}}>\mu_{Q}^{\mathrm{int}}$,
reflecting the mass ordering of the lightest hadrons to carry the respective
charges, $m_{p}>m_{K}>m_{\pi}$. The situation is different in the QGP phase as
shown in Fig. 5 (b) where $P/T^{4}=2$ and $T=0.2$ GeV. The intercept ordering
$\mu_{B}^{\mathrm{int}}>\mu_{Q}^{\mathrm{int}}>\mu_{S}^{\mathrm{int}}$ is
consistent with a parton gas interpretation that $\mu_{B}^{\mathrm{int}}/3\sim
2\mu_{Q}^{\mathrm{int}}/3\sim\mu_{S}^{\mathrm{int}}$, though
$\mu_{S}^{\mathrm{int}}$ is not as small owing to the fact that the strange
quark mass is not negligible at the chosen temperature.
Figure 6: The solid, long-dashed, dash-dotted, and short-dashed lines indicate
the constant $s/n_{B}$ trajectories at 420, 144, 51, and 30, respectively, in
the $\mu_{B}$-$\mu_{S}$-$\mu_{Q}$ space [85].
The constant $s/n_{B}$ trajectories of neos BQS are plotted in the chemical
potential space to illustrate typical regions explored in nuclear collisions
(Fig. 6). The end with larger values of $\mu_{S}$ and $|\mu_{Q}|$ corresponds
to the high temperature region. One can see that the trajectories form a
straight line in the QGP phase, because of the constraints $n_{S}=0$ and
$n_{Q}=0.4n_{B}$ under the leading order approximation of the partonic results
[85]
$\displaystyle\begin{pmatrix}n_{B}\\\ n_{Q}\\\
n_{S}\end{pmatrix}=T^{2}\begin{pmatrix}\chi_{2}^{B}&\chi_{1,1}^{B,Q}&\chi_{1,1}^{B,S}\\\
\chi_{1,1}^{B,Q}&\chi_{2}^{Q}&\chi_{1,1}^{Q,S}\\\
\chi_{1,1}^{B,S}&\chi_{1,1}^{Q,S}&\chi_{2}^{S}\end{pmatrix}\begin{pmatrix}\mu_{B}\\\
\mu_{Q}\\\ \mu_{S}\end{pmatrix},$ (15)
which leads to $\mu_{B}=4.6n_{B}/T^{2}$, $\mu_{Q}=-0.2n_{B}/T^{2}$, and
$\mu_{S}=1.6n_{B}/T^{2}$. As mentioned earlier, they deviate from the straight
line in the hadronic phase in the direction of larger baryon chemical
potentials because of the mass difference between protons and pions. The
second bend towards larger strangeness chemical potential near the low
temperature end is induced by the mass difference between kaons and pions. It
is important to note that one does not explore the $T$-$\mu_{B}$ plane but the
$T$-$\mu_{B}$-$\mu_{Q}$-$\mu_{S}$ space in nuclear collider experiments. This
should be taken into account when analyzing the experimental data to learn
about the phase structure of QCD.
### 3.6 Applications to nuclear collisions
The phenomenological consequences of the conditions of strangeness neutrality
and electric charge-to-baryon ratio of heavy nuclei are studied by using the
hydrodynamic model [193] of relativistic nuclear collisions. We consider Pb+Pb
collisions at $\sqrt{s_{NN}}=17.3$ GeV as conducted at the CERN SPS [191, 194,
195, 190, 196, 197].
The initial conditions for the hydrodynamic model are calculated using an
event-by-event dynamical Glauber model [98]. The Glauber model is a framework
that provides initial geometrical configurations in the transverse plane based
on the Woods-Saxon potential and inelastic nucleon-nucleon cross section
[198]. Its improved version, the Glauber-Lexus model [96], takes into account
the exchange of longitudinal momentum [199]. The dynamical Glauber model is
the four-dimensional version in the sense that the energy and net baryon
number densities are introduced to the system as each sub-collision of target
and projectile nucleons occurs over time.
The numerical implementation music [200, 28, 201] is used to perform the
three-dimensional hydrodynamic simulation. A simple choice of transport
coefficients, namely a shear viscosity of $\eta/s=0.08$ and vanishing bulk
viscosity and baryon diffusion, is employed to minimize ambiguities.
Particlization is assumed to occur on a surface of constant energy density,
defined by the switching energy density $e_{\textrm{sw}}$. Particles are then
further evolved using the hadron cascade model Ultra-relativistic Quantum
Molecular Dynamics (UrQMD) [202, 203].
Figure 7: (Left) The hadronic yields for the most central Pb+Pb collisions at
$\sqrt{s_{NN}}=17.3$ GeV calculated with neos B, BS, and BQS represented by
circular, triangular and cross symbols, respectively. [85]. (Right) The
particle-antiparticle ratios with the same conditions compared with the
experimental data [204].
List of hadronic chemical potentials. Hadrons Chemical potentials $\pi^{+}$
$\mu_{Q}$ $K^{+}$ $\mu_{Q}+\mu_{S}$ $\phi$ 0 $p$ $\mu_{B}+\mu_{Q}$ $\Lambda$
$\mu_{B}-\mu_{S}$ $\Xi^{-}$ $\mu_{B}-\mu_{Q}-2\mu_{S}$ $\Omega$
$\mu_{B}-\mu_{Q}-3\mu_{S}$
The simulated yields of particles and antiparticles in most central events
(using $e_{\textrm{sw}}=0.26$ GeV/fm3) are shown in Fig. 7 (left) and their
ratios in Fig. 7 (right) for the three different versions of the equation of
state. Comparison of neos B and BS results to the experimental data from SPS
shows that the strangeness neutrality condition improves the description of
the particle-antiparticle ratios of the hadrons with finite strangeness
chemical potential, $K$, $\Lambda$, $\Xi$, and $\Omega$. The antiproton-proton
ratio is also modified and moves closer to the data because of the
aforementioned enhancement in the baryon chemical potential, when imposing
strangeness neutrality. The differences between the neos BS and BQS results
are rather small because the electric chemical potential is small for the
collisions of heavy nuclei. As mentioned before, it is still
phenomenologically important because it explains the experimental result that
the anti-pion to pion ratio is greater than one. A list of particle species
along with the chemical potentials that affect their respective yields is
given in Table 3.6.
In Fig. 8, the switching energy density dependence is studied for particle
yields (left) and ratios (right) using neos BQS. Chemical equilibrium is
assumed down to lower temperatures when a lower switching energy density is
considered. The results mostly agree with the experimental data when
$e_{\textrm{sw}}=0.16$-$0.36$ GeV/fm3. The yields of antibaryons are most
affected as $e_{\textrm{sw}}$ decreases, possibly because of the interplay of
the enhancement of the baryon chemical potential at particlization (see Fig. 3
(e)) and the suppression of heavier particle production in the thermal bath.
The two effects tend to cancel for baryons while they are additive for
antibaryons.
Figure 8: (Left) The hadronic yields for the most central Pb+Pb collisions at
$\sqrt{s_{NN}}=17.3$ GeV estimated with neos BQS at
$e_{\textrm{sw}}=0.16,0.26$, and $0.36$ GeV/fm3 represented by circular,
triangular and cross symbols, respectively. [85]. (Right) The particle-
antiparticle ratios with the same conditions compared with experimental data
[204].
## 4 Comparison of equation of state models
We numerically compare different models of the QCD equation of state used in
hydrodynamic simulations of relativistic nuclear collisions. Then effects of
the differences on hydrodynamic evolution are investigated.
### 4.1 Thermodynamic properties
Shown in Fig. 9 (a) are the trace anomalies $(e-3P)/T^{4}$ from lattice QCD
simulations and several equation of state models alongside neos, that we
discussed in the previous section. neos and Duke [205] equations of state show
agreement with the continuum limit result of the HotQCD Collaboration, on
which their structure is based. Similarly, BEST [64] and University of Houston
(denoted as UH) [105] equations of state agree with the results from the
Wuppertal-Budapest (WB) Collaboration, utilized for their construction.
s95p-v1 (s95p) [132] is one of the earliest works on the hybrid equation of
state and the deviation from the rest of the models may be owing in part to
the difference in the lattice QCD data used. Its parametrization and matching
procedure are also different owing to the now-resolved discrepancy between the
resonance gas and early lattice data with non-physical pion mass. An updated
version, employing more recent lattice QCD results, s83s18, has recently been
released [135]. It is noteworthy that Duke and s95p results are very similar
in the hadronic phase. Finally, we point out that the two shown lattice QCD
results for the trace anomaly in the continuum limit are consistent.
The sound velocities are shown in Fig. 9 (b). The basic structure of having a
minimum of $c_{s}^{2}$ near the crossover is found in all models and lattice
simulations. The exact location of the minimum is sensitive to the details of
the construction of each model, such as the connecting temperature and width.
Again, by construction neos and Duke equations of state agree with the HotQCD
result – and BEST and UH equations of state with the WB result – at higher
temperatures.
Figure 9: (a) Comparison of the trace anomalies from lattice QCD simulations
and lattice QCD based equation of state models. (b) Comparison of the sound
velocities extracted from the equation of state models and lattice QCD
calculations.
The comparison of the trajectories on the phase diagram for constant
$s/n_{B}=94$, which approximately corresponds to the collision energy of
$\sqrt{s_{NN}}=39$ GeV [206], is shown in Fig. 10 (a) to illustrate the
properties of the equations of state at finite density. The phase trajectories
for neos B and BEST as well as those for neos BQS and UH BQS behave similarly
with small differences in the QGP phase, which may come from the difference in
the lattice QCD data employed. The effect of additional charges to the baryon
chemical potential is as discussed in the previous section. The difference
between neos BQS and UH BQS results in the hadronic phase may come from the
difference in the hadronic components used in the resonance gas [207] and the
structural difference that the UH BQS equation of state is expanded up to the
second and fourth order in $\mu_{B,Q,S}/T$ in the hadronic phase to perform
matching to the lattice data in the susceptibilities, whereas neos uses the
hadron resonance gas without truncation as the matching is done for the
pressure.
Figure 10 (b) shows the sound velocities at finite densities on the
$s/n_{B}=94$ trajectories. One can see that they are sensitive to the details
of the equation of state used. The results obtained with neos B and BQS are
similar as previously observed in Fig. 4. It should be noted that the small
wiggles in UH BQS and BEST equations of state at low temperature are artifacts
caused by a cut-off at $\mu_{B}=0.45$ GeV.
Figure 10: (a) Comparison of the phase trajectories for constant $s/n_{B}=94$
from BEST equation of state vs. neos B and UH BQS vs. neos BQS. (b) Comparison
of the sound velocities extracted from the equation of state models.
### 4.2 Hydrodynamic evolution
We now compare the hydrodynamic evolution in heavy ion collisions with
different equations of state. Figure 11 (a) shows the time-evolution of the
average time-like flow component $\langle u^{\tau}\rangle$ in one 30-40% Au+Au
collision at $\sqrt{s_{NN}}=200$ GeV with the IP-Glasma initial condition
[208, 209]. The quantity is closely related to radial flow, which affects the
slope of the particles’ transverse momentum spectra ($u^{\tau}$ is closely
related to the transverse flow velocity $u_{\perp}$ via the flow normalization
condition $u\cdot u=1$, particularly when neglecting longitudinal flow
$u^{\eta}$). $\langle u^{\tau}\rangle$ increases with time and exhibits
similar behavior in all cases. neos BQS and Duke equations of state lead to
similar results. The BEST and UH BQS equations of state lead to larger
$\langle u^{\tau}\rangle$, while $\langle u^{\tau}\rangle$ of the s95p
equation of state is smaller than that for neos BQS and Duke at later times,
but is slightly larger at earlier times before around $\tau=2$ fm. The
orderings are consistent with those of the sound velocity and trace anomaly,
considering that the average medium temperature is larger ($\sim 0.4$ GeV) at
earlier times and smaller ($\sim 0.2$ GeV) at later times.
The time evolution of the system’s averaged momentum anisotropy
$\varepsilon_{p}=\frac{\sqrt{\langle T^{xx}-T^{yy}\rangle^{2}+\langle
2T^{xy}\rangle^{2}}}{\langle T^{xx}+T^{yy}\rangle},$ (16)
in the same hydrodynamic setup is shown in Fig. 11 (b). This quantity is
closely related to the final elliptic momentum anisotropy of produced
particles. The differences in the momentum anisotropy between the equation of
state models are rather small. The s95p result rises and falls slightly
earlier than the others. The UH BQS equation of state has the largest momentum
anisotropy at later times, followed by the BEST equation of state. neos BQS
and Duke equations of state have similar $\varepsilon_{p}$, though the former
is slightly larger than the latter around $\tau=4$ fm.
Figure 11: (a) The time evolution of averaged $u^{\tau}$ for an Au+Au
collision in 30-40% centrality at 200 GeV with different equations of state.
(b) Similar comparison for the time evolution of the momentum anisotropy
$\varepsilon_{p}$.
We make a similar comparison at finite net baryon density by simulating (3+1)D
hydrodynamic evolution for 20-30% Au+Au collisions at 39 GeV with the event-
averaged initial condition [102]. Figure 12 shows that the four equations of
state produce a very similar evolution for the development of hydrodynamic
radial flow and the momentum anisotropy. Similar to the zero density case, the
larger speed of sound in the BEST equation of state leads to slightly stronger
radial flow compared to the neos and UH equations of state in Fig. 12(a). The
system’s momentum anisotropy at late time has the order BEST $>$ UH BQS $>$
neos.
Figure 12: (a) The time evolution of averaged $u^{\tau}$ for an Au+Au
collision in 20-30% centrality at 39 GeV with different equations of state.
(b) Similar comparison for the time evolution of the momentum anisotropy
$\varepsilon_{p}$.
Figure 13: The averaged phase trajectories for a fireball at mid-rapidity in
20-30% Au+Au collisions at 39 GeV.
Finally, we show the phase trajectories of mid-rapidity Au+Au collisions with
the four equations of state in Figure 13. These trajectories are averaged over
fluid cells from realistic (3+1)D hydrodynamic simulations. The difference
among the four equations of state are in qualitative agreement with the
difference in constant $s/n_{B}$ trajectories shown in Fig. 10. The
strangeness neutrality condition moves the trajectories towards larger
$\mu_{B}$ compared to those without this constraint. It indicates that, as
mentioned earlier, having multiple conserved charges is phenomenologically
important for the exploration of the QCD phase diagram, including the critical
point search, as well as for the estimation of dissipative processes, such as
baryon diffusion. The trajectory from the UH BQS has slightly larger $\mu_{B}$
values compared to the neos BQS in the QGP phase.
## 5 Conclusion and summary
We reviewed QCD equations of state at finite chemical potentials. All current
models for the equation of state generally agree at zero densities, because of
the advances in lattice QCD simulations, which all agree now that the quark-
hadron transition is a crossover at around $T=155$-$160$ MeV, and the
information from lattice QCD is used as input to determine parameters in the
various phenomenological models. On the other hand, the finite-density
structure of the QCD phase diagram, such as the critical point and first-order
phase transition, is less well understood. Going beyond zero density is not
directly possible on the lattice due to the fermion sign problem. Besides
lattice based methods such as Taylor expansion or the use of imaginary
chemical potentials, various approaches to obtain a finite density QCD
equation of state have been proposed, including the perturbative QCD method,
the Polyakov loop-extended Nambu-Jona-Lasinio model, and holographic
conjecture.
We have introduced the neos model where the three conserved charges in the
strongly-interacting medium – net baryon, electric charge and strangeness –
are explicitly considered. The model is built from a state-of-the-art lattice
QCD equation of state and the second- and fourth-order susceptibilities from
the lattice, together with the hadron resonance gas result, which includes all
known hadrons and resonances with masses below 2 GeV. Lattice and hadron gas
equations of state are connected near the quark-hadron transition in a
thermodynamically consistent way to obtain a crossover-type equation of state
at finite temperatures and chemical potentials. We have considered the
strangeness neutrality condition $n_{S}=0$ and the electric charge-to-baryon
ratio $n_{Q}=0.4n_{B}$, that reflect the situation in collisions of heavy
nuclei to elucidate the effects of multiple conserved charges.
The multi-dimensional neos QCD equation of state has been included in the
viscous hydrodynamic model of heavy-ion collisions music at intermediate
relativistic energies. We showed in the comparison of the theoretical
predictions with SPS experimental data of particle-antiparticle ratios that
the model description is visibly improved when the strangeness neutrality
condition is imposed for the hadrons with finite strangeness quantum numbers
and – through the interplay of conserved charges – also for those with finite
baryon number. The realistic electric charge-to-baryon ratio induces smaller
effect because the electric charge chemical potential is small when heavy
stable nuclei such as Au and Pb are used. Nevertheless, its inclusion leads to
a correct description of the antipion-to-pion number ratio exceeding one in
the observed data. Our results also clarify that in the beam energy scan one
is really exploring the $T$-$\mu_{B}$-$\mu_{Q}$-$\mu_{S}$ phase diagram,
instead of just one in the $T$-$\mu_{B}$ plane. This is important when
extracting the information of the QCD medium properties and phase structures
from experimental data.
We have then compared several models of the QCD equation of state used in
relativistic hydrodynamic studies of nuclear collisions. In the zero density
limit, the continuum limit results for the trace anomaly and sound velocity of
the Wuppertal-Budapest collaboration and the HotQCD collaboration agree within
their error bands. neos and Duke equations of state exhibit good agreement
with the latter results while BEST and UH equations of state with the former
results, as expected from their construction. The widely-used s95p-v1 has a
larger trace anomaly and smaller sound velocity in the crossover region, and
newer versions, such as s83s18, should be used. The comparison of the constant
$s/n_{B}$ trajectories in the phase diagram with different equations of state
demonstrated that neos B and BEST equations of state are close to each other
at finite density. Once the strangeness neutrality condition and the realistic
charge-to-baryon ratio is taken into account, the trajectories are shifted to
larger $\mu_{B}$. neos BQS and UH BQS equations of state behave similarly in
the QGP phase with a slight difference coming likely from the choice of
lattice data. The sound velocity differs between the two equations of state in
the vicinity of the crossover, possibly because of the difference in the
connection of the hadron resonance gas to the lattice QCD results.
We performed several hydrodynamic simulations of heavy ion collisions using
different equations of state and studied their effect on the time evolution of
flow velocities and momentum anisotropies. At $\sqrt{s_{NN}}=200$ GeV, the
time evolution of the averaged time-like flow component $\langle
u^{\tau}\rangle$ is in the order of BEST, UH BQS, neos BQS, Duke, and s95p,
from the fastest to the slowest buildup (and largest to smallest final
values), which is consistent with the ordering of the sound velocity in the
zero density case near the crossover temperature. Comparing the averaged
momentum anisotropies, final values for UH BQS and BEST are larger than for
neos BQS and Duke, which in turn are larger than those for s95p. The
differences in $\langle u^{\tau}\rangle$ at finite density at
$\sqrt{s_{NN}}=39$ GeV is rather small, but if closely observed, BEST produces
the largest, followed by UH BQS and neos at later times because of the
differences in the speed of sound. The average momentum anisotropy is also
ordered similarly, but BEST and UH BQS are closest to each other. We also
studied the trajectories of the average temperature and baryon chemical
potential of the system, and found them to be qualitatively consistent with
the constant $s/n_{B}$ trajectories of the corresponding equation of state
models.
Progress in determining the nuclear equation of state at finite densities as
been significant in the last several years. Advances in lattice QCD have
allowed to move towards realistic modeling with finite chemical potentials and
even studies including potential critical points are possible. Experimental
advances have also been tremendous, both on the front of heavy ion collision
beam energy scans and gravitational wave observations of neutron star (and
black hole) binary systems, which will allow for ever improving constraints on
the nuclear equation of state over a wide range in the phase diagram.
A public version of the neos tabulated results is available online [210] for
the use in relativistic hydrodynamic models and other related studies. Other
codes/data are also publicly available for BEST [211], UH BQS [212], s95p-v1
[213], s83s18 [214], and Duke [215] equations of state.
## Acknowledgments
The authors thank Frithjof Karsch, Swagato Mukherjee, and Sayantan Sharma for
useful discussions. AM is supported by JSPS KAKENHI Grant Number JP19K14722.
BPS is supported under DOE Contract No. DE-SC0012704. CS is supported in part
under DOE Contract No. DE-SC0013460 and in part by the National Science
Foundation (NSF) under grant number PHY-2012922. This research used resources
of the National Energy Research Scientific Computing Center, which is
supported by the Office of Science of the U.S. Department of Energy under
Contract No. DE-AC02-05CH11231. This work is supported in part by the U.S.
Department of Energy, Office of Science, Office of Nuclear Physics, within the
framework of the Beam Energy Scan Theory (BEST) Topical Collaboration.
## References
* [1] J. Letessier and J. Rafelski, Hadrons and quark - gluon plasma (Cambridge University Press, 2002).
* [2] J. Kapusta, B. Muller and J. Rafelski, Quark-Gluon Plasma: Theoretical Foundations (Elsevier, 2003).
* [3] K. Yagi, T. Hatsuda and Y. Miake, Quark-gluon plasma: From big bang to little bang (Cambridge University Press, 2005).
* [4] X.-N. Wang (ed.), Quark-Gluon Plasma 5 (World Scientific, New Jersey, 2016).
* [5] PHENIX Collaboration (K. Adcox et al.), Nucl. Phys. A 757, 184 (2005), arXiv:nucl-ex/0410003, 10.1016/j.nuclphysa.2005.03.086.
* [6] STAR Collaboration (J. Adams et al.), Nucl. Phys. A 757, 102 (2005), arXiv:nucl-ex/0501009, 10.1016/j.nuclphysa.2005.03.085.
* [7] PHOBOS Collaboration (B. Back et al.), Nucl. Phys. A 757, 28 (2005), arXiv:nucl-ex/0410022, 10.1016/j.nuclphysa.2005.03.084.
* [8] BRAHMS Collaboration (I. Arsene et al.), Nucl. Phys. A 757, 1 (2005), arXiv:nucl-ex/0410020, 10.1016/j.nuclphysa.2005.02.130.
* [9] ALICE Collaboration (K. Aamodt et al.), Phys. Rev. Lett. 105, 252302 (2010), arXiv:1011.3914 [nucl-ex], 10.1103/PhysRevLett.105.252302.
* [10] ATLAS Collaboration (G. Aad et al.), Phys. Lett. B 707, 330 (2012), arXiv:1108.6018 [hep-ex], 10.1016/j.physletb.2011.12.056.
* [11] CMS Collaboration (S. Chatrchyan et al.), Eur. Phys. J. C 72, 2012 (2012), arXiv:1201.3158 [nucl-ex], 10.1140/epjc/s10052-012-2012-3.
* [12] C. Gale, S. Jeon, B. Schenke, P. Tribedy and R. Venugopalan, Phys. Rev. Lett. 110, 012302 (2013), arXiv:1209.6330 [nucl-th], 10.1103/PhysRevLett.110.012302.
* [13] K. Fukushima and T. Hatsuda, Rept. Prog. Phys. 74, 014001 (2011), arXiv:1005.4814 [hep-ph], 10.1088/0034-4885/74/1/014001.
* [14] K. Fukushima and C. Sasaki, Prog. Part. Nucl. Phys. 72, 99 (2013), arXiv:1301.6377 [hep-ph], 10.1016/j.ppnp.2013.05.003.
* [15] P. de Forcrand, PoS LAT2009, 010 (2009), arXiv:1005.0539 [hep-lat], 10.22323/1.091.0010.
* [16] M. Asakawa and K. Yazaki, Nucl. Phys. A 504, 668 (1989), 10.1016/0375-9474(89)90002-X.
* [17] M. G. Alford, K. Rajagopal and F. Wilczek, Phys. Lett. B 422, 247 (1998), arXiv:hep-ph/9711395, 10.1016/S0370-2693(98)00051-3.
* [18] R. Rapp, T. Schäfer, E. V. Shuryak and M. Velkovsky, Phys. Rev. Lett. 81, 53 (1998), arXiv:hep-ph/9711396, 10.1103/PhysRevLett.81.53.
* [19] M. G. Alford, K. Rajagopal and F. Wilczek, Nucl. Phys. B 537, 443 (1999), arXiv:hep-ph/9804403, 10.1016/S0550-3213(98)00668-3.
* [20] T. Hatsuda, M. Tachibana, N. Yamamoto and G. Baym, Phys. Rev. Lett. 97, 122001 (2006), arXiv:hep-ph/0605018, 10.1103/PhysRevLett.97.122001.
* [21] M. Kitazawa, T. Koide, T. Kunihiro and Y. Nemoto, Prog. Theor. Phys. 108, 929 (2002), arXiv:hep-ph/0207255, 10.1143/PTP.108.929, [Erratum: Prog.Theor.Phys. 110, 185–186 (2003)].
* [22] L. McLerran and R. D. Pisarski, Nucl. Phys. A 796, 83 (2007), arXiv:0706.2191 [hep-ph], 10.1016/j.nuclphysa.2007.08.013.
* [23] LIGO Scientific, Virgo Collaboration (B. Abbott et al.), Phys. Rev. Lett. 119, 161101 (2017), arXiv:1710.05832 [gr-qc], 10.1103/PhysRevLett.119.161101.
* [24] LIGO Scientific, Virgo Collaboration (B. Abbott et al.), Phys. Rev. Lett. 121, 161101 (2018), arXiv:1805.11581 [gr-qc], 10.1103/PhysRevLett.121.161101.
* [25] G. Baym, T. Hatsuda, T. Kojo, P. D. Powell, Y. Song and T. Takatsuka, Rept. Prog. Phys. 81, 056902 (2018), arXiv:1707.04966 [astro-ph.HE], 10.1088/1361-6633/aaae14.
* [26] V. Dexheimer, J. Noronha, J. Noronha-Hostler, C. Ratti and N. Yunes (10 2020), arXiv:2010.08834 [nucl-th].
* [27] P. Kolb, P. Huovinen, U. W. Heinz and H. Heiselberg, Phys. Lett. B 500, 232 (2001), arXiv:hep-ph/0012137, 10.1016/S0370-2693(01)00079-X.
* [28] B. Schenke, S. Jeon and C. Gale, Phys. Rev. Lett. 106, 042301 (2011), arXiv:1009.3244 [hep-ph], 10.1103/PhysRevLett.106.042301.
* [29] J.-Y. Ollitrault, Phys. Rev. D 46, 229 (1992), 10.1103/PhysRevD.46.229.
* [30] A. M. Poskanzer and S. Voloshin, Phys. Rev. C 58, 1671 (1998), arXiv:nucl-ex/9805001, 10.1103/PhysRevC.58.1671.
* [31] J. Takahashi, B. Tavares, W. Qian, R. Andrade, F. Grassi, Y. Hama, T. Kodama and N. Xu, Phys. Rev. Lett. 103, 242301 (2009), arXiv:0902.4870 [nucl-th], 10.1103/PhysRevLett.103.242301.
* [32] B. Alver and G. Roland, Phys. Rev. C 81, 054905 (2010), arXiv:1003.0194 [nucl-th], 10.1103/PhysRevC.82.039903, [Erratum: Phys.Rev.C 82, 039903 (2010)].
* [33] S. Pratt, E. Sangaline, P. Sorensen and H. Wang, Phys. Rev. Lett. 114, 202301 (2015), arXiv:1501.04042 [nucl-th], 10.1103/PhysRevLett.114.202301.
* [34] E. Sangaline and S. Pratt, Phys. Rev. C 93, 024908 (2016), arXiv:1508.07017 [nucl-th], 10.1103/PhysRevC.93.024908.
* [35] J. E. Bernhard, J. S. Moreland, S. A. Bass, J. Liu and U. Heinz, Phys. Rev. C 94, 024907 (2016), arXiv:1605.03954 [nucl-th], 10.1103/PhysRevC.94.024907.
* [36] L.-G. Pang, K. Zhou, N. Su, H. Petersen, H. Stöcker and X.-N. Wang, Nature Commun. 9, 210 (2018), arXiv:1612.04262 [hep-ph], 10.1038/s41467-017-02726-3.
* [37] A. Monnai and J.-Y. Ollitrault, Phys. Rev. C 96, 044902 (2017), arXiv:1707.08466 [nucl-th], 10.1103/PhysRevC.96.044902.
* [38] J.-F. Paquet, C. Shen, G. Denicol, S. Jeon and C. Gale, Nucl. Phys. A 967, 429 (2017), 10.1016/j.nuclphysa.2017.06.024.
* [39] A. Chodos, R. Jaffe, K. Johnson, C. B. Thorn and V. Weisskopf, Phys. Rev. D 9, 3471 (1974), 10.1103/PhysRevD.9.3471.
* [40] A. Chodos, R. Jaffe, K. Johnson and C. B. Thorn, Phys. Rev. D 10, 2599 (1974), 10.1103/PhysRevD.10.2599.
* [41] A. De Rujula, H. Georgi and S. Glashow, Phys. Rev. D 12, 147 (1975), 10.1103/PhysRevD.12.147.
* [42] Y. Nambu and G. Jona-Lasinio, Phys. Rev. 122, 345 (1961), 10.1103/PhysRev.122.345.
* [43] Y. Nambu and G. Jona-Lasinio, Phys. Rev. 124, 246 (1961), 10.1103/PhysRev.124.246.
* [44] F. R. Brown, F. P. Butler, H. Chen, N. H. Christ, Z.-h. Dong, W. Schaffer, L. I. Unger and A. Vaccarino, Phys. Rev. Lett. 65, 2491 (1990), 10.1103/PhysRevLett.65.2491.
* [45] CP-PACS Collaboration (A. Ali Khan et al.), Phys. Rev. D 63, 034502 (2000), arXiv:hep-lat/0008011, 10.1103/PhysRevD.63.034502.
* [46] Y. Aoki, G. Endrodi, Z. Fodor, S. Katz and K. Szabo, Nature 443, 675 (2006), arXiv:hep-lat/0611014, 10.1038/nature05120.
* [47] S. Borsanyi, Z. Fodor, C. Hoelbling, S. D. Katz, S. Krieg and K. K. Szabo, Phys. Lett. B 730, 99 (2014), arXiv:1309.5258 [hep-lat], 10.1016/j.physletb.2014.01.007.
* [48] HotQCD Collaboration (A. Bazavov et al.), Phys. Rev. D 90, 094503 (2014), arXiv:1407.6387 [hep-lat], 10.1103/PhysRevD.90.094503.
* [49] H. Ding et al., Phys. Rev. Lett. 123, 062002 (2019), arXiv:1903.04801 [hep-lat], 10.1103/PhysRevLett.123.062002.
* [50] C. Shen, Nucl. Phys. A 1005, 121788 (2021), arXiv:2001.11858 [nucl-th], 10.1016/j.nuclphysa.2020.121788.
* [51] C. Shen and L. Yan, Nucl. Sci. Tech. 31, 122 (10 2020), arXiv:2010.12377 [nucl-th], 10.1007/s41365-020-00829-z.
* [52] C. Nonaka and M. Asakawa, Phys. Rev. C 71, 044904 (2005), arXiv:nucl-th/0410078, 10.1103/PhysRevC.71.044904.
* [53] M. Bluhm, B. Kampfer and G. Soff, Phys. Lett. B 620, 131 (2005), arXiv:hep-ph/0411106, 10.1016/j.physletb.2005.05.083.
* [54] M. Bluhm, B. Kampfer, R. Schulze, D. Seipt and U. Heinz, Phys. Rev. C 76, 034901 (2007), arXiv:0705.0397 [hep-ph], 10.1103/PhysRevC.76.034901.
* [55] V. Dexheimer and S. Schramm, Phys. Rev. C 81, 045201 (2010), arXiv:0901.1748 [astro-ph.SR], 10.1103/PhysRevC.81.045201.
* [56] J. Steinheimer, S. Schramm and H. Stocker, J. Phys. G 38, 035001 (2011), arXiv:1009.5239 [hep-ph], 10.1088/0954-3899/38/3/035001.
* [57] P. Huovinen and P. Petreczky, J. Phys. G 38, 124103 (2011), arXiv:1106.6227 [nucl-th], 10.1088/0954-3899/38/12/124103.
* [58] M. Hempel, V. Dexheimer, S. Schramm and I. Iosilevskiy, Phys. Rev. C 88, 014906 (2013), arXiv:1302.2835 [nucl-th], 10.1103/PhysRevC.88.014906.
* [59] M. Albright, J. Kapusta and C. Young, Phys. Rev. C 90, 024915 (2014), arXiv:1404.7540 [nucl-th], 10.1103/PhysRevC.90.024915.
* [60] M. Albright, J. Kapusta and C. Young, Phys. Rev. C 92, 044904 (2015), arXiv:1506.03408 [nucl-th], 10.1103/PhysRevC.92.044904.
* [61] R. Rougemont, R. Critelli, J. Noronha-Hostler, J. Noronha and C. Ratti, Phys. Rev. D 96, 014032 (2017), arXiv:1704.05558 [hep-ph], 10.1103/PhysRevD.96.014032.
* [62] R. Critelli, J. Noronha, J. Noronha-Hostler, I. Portillo, C. Ratti and R. Rougemont, Phys. Rev. D 96, 096026 (2017), arXiv:1706.00455 [nucl-th], 10.1103/PhysRevD.96.096026.
* [63] V. Vovchenko, J. Steinheimer, O. Philipsen and H. Stoecker, Phys. Rev. D 97, 114030 (2018), arXiv:1711.01261 [hep-ph], 10.1103/PhysRevD.97.114030.
* [64] P. Parotto, M. Bluhm, D. Mroczek, M. Nahrgang, J. Noronha-Hostler, K. Rajagopal, C. Ratti, T. Schäfer and M. Stephanov, Phys. Rev. C 101, 034901 (2020), arXiv:1805.05249 [hep-ph], 10.1103/PhysRevC.101.034901.
* [65] V. Vovchenko, J. Steinheimer, O. Philipsen, A. Pasztor, Z. Fodor, S. Katz and H. Stoecker, Nucl. Phys. A 982, 859 (2019), arXiv:1807.06472 [hep-lat], 10.1016/j.nuclphysa.2018.10.068.
* [66] W.-j. Fu, J. M. Pawlowski and F. Rennecke, SciPost Phys. Core 2, 002 (2020), arXiv:1808.00410 [hep-ph], 10.21468/SciPostPhysCore.2.1.002.
* [67] W.-j. Fu, J. M. Pawlowski and F. Rennecke, Phys. Rev. D 100, 111501 (2019), arXiv:1809.01594 [hep-ph], 10.1103/PhysRevD.100.111501.
* [68] A. Motornenko, V. Vovchenko, J. Steinheimer, S. Schramm and H. Stoecker, Nucl. Phys. A 982, 891 (2019), arXiv:1809.02000 [hep-ph], 10.1016/j.nuclphysa.2018.11.028.
* [69] C. J. Plumberg, T. Welle and J. I. Kapusta, PoS CORFU2018, 157 (2018), arXiv:1812.01684 [nucl-th], 10.22323/1.347.0157.
* [70] R. V. Gavai and S. Gupta, Phys. Rev. D 64, 074506 (2001), arXiv:hep-lat/0103013, 10.1103/PhysRevD.64.074506.
* [71] C. Allton, S. Ejiri, S. Hands, O. Kaczmarek, F. Karsch, E. Laermann, C. Schmidt and L. Scorzato, Phys. Rev. D 66, 074507 (2002), arXiv:hep-lat/0204010, 10.1103/PhysRevD.66.074507.
* [72] Z. Fodor and S. Katz, JHEP 04, 050 (2004), arXiv:hep-lat/0402006, 10.1088/1126-6708/2004/04/050.
* [73] Z. Fodor and S. Katz, Phys. Lett. B 534, 87 (2002), arXiv:hep-lat/0104001, 10.1016/S0370-2693(02)01583-6.
* [74] P. de Forcrand and O. Philipsen, Nucl. Phys. B 642, 290 (2002), arXiv:hep-lat/0205016, 10.1016/S0550-3213(02)00626-0.
* [75] M. D’Elia and M.-P. Lombardo, Phys. Rev. D 67, 014505 (2003), arXiv:hep-lat/0209146, 10.1103/PhysRevD.67.014505.
* [76] J. Guenther, R. Bellwied, S. Borsanyi, Z. Fodor, S. Katz, A. Pasztor, C. Ratti and K. Szabó, Nucl. Phys. A 967, 720 (2017), arXiv:1607.02493 [hep-lat], 10.1016/j.nuclphysa.2017.05.044.
* [77] G. Parisi, Phys. Lett. B 131, 393 (1983), 10.1016/0370-2693(83)90525-7.
* [78] J. R. Klauder and W. P. Petersen, J. Stat. Phys. 39, 53 (1985), https://doi.org/10.1007/BF01007974.
* [79] J. Ambjorn and S. Yang, Phys. Lett. B 165, 140 (1985), 10.1016/0370-2693(85)90708-7.
* [80] K. Nagata, J. Nishimura and S. Shimasaki, Phys. Rev. D 94, 114515 (2016), arXiv:1606.07627 [hep-lat], 10.1103/PhysRevD.94.114515.
* [81] F. Pham, Proc. Symp. Pure Math. 40, 319 (1983).
* [82] E. Witten, AMS/IP Stud. Adv. Math. 50, 347 (2011), arXiv:1001.2933 [hep-th].
* [83] Y. Mori, K. Kashiwa and A. Ohnishi, Phys. Rev. D 96, 111501 (2017), arXiv:1705.05605 [hep-lat], 10.1103/PhysRevD.96.111501.
* [84] Y. Mori, K. Kashiwa and A. Ohnishi, PTEP 2018, 023B04 (2018), arXiv:1709.03208 [hep-lat], 10.1093/ptep/ptx191.
* [85] A. Monnai, B. Schenke and C. Shen, Phys. Rev. C 100, 024907 (2019), arXiv:1902.05095 [nucl-th], 10.1103/PhysRevC.100.024907.
* [86] A. Monnai, B. Schenke and C. Shen, Nucl. Phys. A 1005, 121868 (2021), arXiv:2002.02661 [nucl-th], 10.1016/j.nuclphysa.2020.121868.
* [87] HotQCD Collaboration (A. Bazavov et al.), Phys. Rev. D 86, 034509 (2012), arXiv:1203.0784 [hep-lat], 10.1103/PhysRevD.86.034509.
* [88] H. T. Ding, S. Mukherjee, H. Ohno, P. Petreczky and H. P. Schadler, Phys. Rev. D 92, 074043 (2015), arXiv:1507.06637 [hep-lat], 10.1103/PhysRevD.92.074043.
* [89] A. Bazavov et al., Phys. Rev. D 95, 054504 (2017), arXiv:1701.04325 [hep-lat], 10.1103/PhysRevD.95.054504.
* [90] S. Sharma, private communications .
* [91] S. Borsanyi, Z. Fodor, S. D. Katz, S. Krieg, C. Ratti and K. Szabo, JHEP 01, 138 (2012), arXiv:1112.4416 [hep-lat], 10.1007/JHEP01(2012)138.
* [92] R. Bellwied, S. Borsanyi, Z. Fodor, S. Katz, A. Pasztor, C. Ratti and K. Szabo, Phys. Rev. D 92, 114505 (2015), arXiv:1507.04627 [hep-lat], 10.1103/PhysRevD.92.114505.
* [93] S. Borsanyi, Z. Fodor, J. N. Guenther, S. K. Katz, K. K. Szabo, A. Pasztor, I. Portillo and C. Ratti, JHEP 10, 205 (2018), arXiv:1805.04445 [hep-lat], 10.1007/JHEP10(2018)205.
* [94] S. Borsanyi, Z. Fodor, J. N. Guenther, R. Kara, S. D. Katz, P. Parotto, A. Pasztor, C. Ratti and K. K. Szabo, Phys. Rev. Lett. 125, 052001 (2020), arXiv:2002.02821 [hep-lat], 10.1103/PhysRevLett.125.052001.
* [95] G. Denicol, A. Monnai and B. Schenke, Phys. Rev. Lett. 116, 212301 (2016), arXiv:1512.01538 [nucl-th], 10.1103/PhysRevLett.116.212301.
* [96] A. Monnai and B. Schenke, Phys. Lett. B 752, 317 (2016), arXiv:1509.04103 [nucl-th], 10.1016/j.physletb.2015.11.063.
* [97] C. Shen, G. Denicol, C. Gale, S. Jeon, A. Monnai and B. Schenke, Nucl. Phys. A 967, 796 (2017), arXiv:1704.04109 [nucl-th], 10.1016/j.nuclphysa.2017.06.008.
* [98] C. Shen and B. Schenke, Phys. Rev. C 97, 024907 (2018), arXiv:1710.00881 [nucl-th], 10.1103/PhysRevC.97.024907.
* [99] G. S. Denicol, C. Gale, S. Jeon, A. Monnai, B. Schenke and C. Shen, Phys. Rev. C 98, 034916 (2018), arXiv:1804.10557 [nucl-th], 10.1103/PhysRevC.98.034916.
* [100] C. Shen and B. Schenke, Nucl. Phys. A 982, 411 (2019), arXiv:1807.05141 [nucl-th], 10.1016/j.nuclphysa.2018.08.007.
* [101] C. Gale, S. Jeon, S. McDonald, J.-F. Paquet and C. Shen, Nucl. Phys. A 982, 767 (2019), arXiv:1807.09326 [nucl-th], 10.1016/j.nuclphysa.2018.08.005.
* [102] C. Shen and S. Alzhrani, Phys. Rev. C 102, 014909 (2020), arXiv:2003.05852 [nucl-th], 10.1103/PhysRevC.102.014909.
* [103] D. Oliinychenko, C. Shen and V. Koch (9 2020), arXiv:2009.01915 [hep-ph].
* [104] W. Zhao, C. Shen, C. M. Ko, Q. Liu and H. Song, Phys. Rev. C 102, 044912 (2020), arXiv:2009.06959 [nucl-th], 10.1103/PhysRevC.102.044912.
* [105] J. Noronha-Hostler, P. Parotto, C. Ratti and J. Stafford, Phys. Rev. C 100, 064910 (2019), arXiv:1902.06723 [hep-ph], 10.1103/PhysRevC.100.064910.
* [106] K. Fukushima, Phys. Rev. D 79, 074015 (2009), arXiv:0901.0783 [hep-ph], 10.1103/PhysRevD.79.074015.
* [107] D. Toublan and J. B. Kogut, Phys. Lett. B 605, 129 (2005), arXiv:hep-ph/0409310, 10.1016/j.physletb.2004.11.018.
* [108] F. Xu, H. Mao, T. K. Mukherjee and M. Huang, Phys. Rev. D 84, 074009 (2011), arXiv:1104.0873 [hep-ph], 10.1103/PhysRevD.84.074009.
* [109] A. Ohnishi, H. Ueda, T. Nakano, M. Ruggieri and K. Sumiyoshi, Phys. Lett. B 704, 284 (2011), arXiv:1102.3753 [nucl-th], 10.1016/j.physletb.2011.09.018.
* [110] K. Kamikado, N. Strodthoff, L. von Smekal and J. Wambach, Phys. Lett. B 718, 1044 (2013), arXiv:1207.0400 [hep-ph], 10.1016/j.physletb.2012.11.055.
* [111] H. Ueda, T. Z. Nakano, A. Ohnishi, M. Ruggieri and K. Sumiyoshi, Phys. Rev. D 88, 074006 (2013), arXiv:1304.4331 [nucl-th], 10.1103/PhysRevD.88.074006.
* [112] A. Barducci, R. Casalbuoni, G. Pettini and L. Ravagli, Phys. Rev. D 69, 096004 (2004), arXiv:hep-ph/0402104, 10.1103/PhysRevD.69.096004.
* [113] Y. Nishida, Phys. Rev. D 69, 094501 (2004), arXiv:hep-ph/0312371, 10.1103/PhysRevD.69.094501.
* [114] D. Son and M. A. Stephanov, Phys. Rev. Lett. 86, 592 (2001), arXiv:hep-ph/0005225, 10.1103/PhysRevLett.86.592.
* [115] I. Karpenko, P. Huovinen, H. Petersen and M. Bleicher, Phys. Rev. C 91, 064901 (2015), arXiv:1502.01978 [nucl-th], 10.1103/PhysRevC.91.064901.
* [116] Y. Hatta, A. Monnai and B.-W. Xiao, Phys. Rev. D 92, 114010 (2015), arXiv:1505.04226 [hep-ph], 10.1103/PhysRevD.92.114010.
* [117] Y. Hatta, A. Monnai and B.-W. Xiao, Nucl. Phys. A 947, 155 (2016), arXiv:1507.04690 [hep-ph], 10.1016/j.nuclphysa.2015.12.009.
* [118] A. Andronic, P. Braun-Munzinger and J. Stachel, Nucl. Phys. A 772, 167 (2006), arXiv:nucl-th/0511071, 10.1016/j.nuclphysa.2006.03.012.
* [119] A. Bazavov et al., Phys. Rev. Lett. 109, 192302 (2012), arXiv:1208.1220 [hep-lat], 10.1103/PhysRevLett.109.192302.
* [120] J. Sollfrank, P. Huovinen, M. Kataja, P. Ruuskanen, M. Prakash and R. Venugopalan, Phys. Rev. C 55, 392 (1997), arXiv:nucl-th/9607029, 10.1103/PhysRevC.55.392.
* [121] C. Nonaka, E. Honda and S. Muroya, Eur. Phys. J. C 17, 663 (2000), arXiv:hep-ph/0007187, 10.1007/s100520000509.
* [122] D. Teaney, J. Lauret and E. V. Shuryak, Phys. Rev. Lett. 86, 4783 (2001), arXiv:nucl-th/0011058, 10.1103/PhysRevLett.86.4783.
* [123] K. Redlich and H. Satz, Phys. Rev. D 33, 3747 (1986), 10.1103/PhysRevD.33.3747.
* [124] J. Blaizot and J.-Y. Ollitrault, Phys. Rev. D 36, 916 (1987), 10.1103/PhysRevD.36.916.
* [125] D. H. Rischke and M. Gyulassy, Nucl. Phys. A 597, 701 (1996), arXiv:nucl-th/9509040, 10.1016/0375-9474(95)00447-5.
* [126] K. Kajantie, M. Laine, K. Rummukainen and M. E. Shaposhnikov, Nucl. Phys. B 503, 357 (1997), arXiv:hep-ph/9704416, 10.1016/S0550-3213(97)00425-2.
* [127] M. Laine and Y. Schroder, Phys. Rev. D 73, 085009 (2006), arXiv:hep-ph/0603048, 10.1103/PhysRevD.73.085009.
* [128] P. Romatschke and U. Romatschke, Phys. Rev. Lett. 99, 172301 (2007), arXiv:0706.1522 [nucl-th], 10.1103/PhysRevLett.99.172301.
* [129] T. S. Biro and J. Zimanyi, Phys. Lett. B 650, 193 (2007), arXiv:hep-ph/0607079, 10.1016/j.physletb.2007.04.057.
* [130] M. Chojnacki and W. Florkowski, Acta Phys. Polon. B 38, 3249 (2007), arXiv:nucl-th/0702030.
* [131] M. Chojnacki, W. Florkowski, W. Broniowski and A. Kisiel, Phys. Rev. C 78, 014905 (2008), arXiv:0712.0947 [nucl-th], 10.1103/PhysRevC.78.014905.
* [132] P. Huovinen and P. Petreczky, Nucl. Phys. A 837, 26 (2010), arXiv:0912.2541 [hep-ph], 10.1016/j.nuclphysa.2010.02.015.
* [133] M. Bluhm, P. Alba, W. Alberico, A. Beraudo and C. Ratti, Nucl. Phys. A 929, 157 (2014), arXiv:1306.6188 [hep-ph], 10.1016/j.nuclphysa.2014.06.013.
* [134] P. Alba, V. Mantovani Sarti, J. Noronha, J. Noronha-Hostler, P. Parotto, I. Portillo Vazquez and C. Ratti, Phys. Rev. C 98, 034909 (2018), arXiv:1711.05207 [nucl-th], 10.1103/PhysRevC.98.034909.
* [135] J. Auvinen, K. J. Eskola, P. Huovinen, H. Niemi, R. Paatelainen and P. Petreczky, Phys. Rev. C 102, 044911 (2020), arXiv:2006.12499 [nucl-th], 10.1103/PhysRevC.102.044911.
* [136] P. Huovinen, Nucl. Phys. A 761, 296 (2005), arXiv:nucl-th/0505036, 10.1016/j.nuclphysa.2005.07.016.
* [137] F. Cooper and G. Frye, Phys. Rev. D 10, 186 (1974), 10.1103/PhysRevD.10.186.
* [138] A. Monnai, Phys. Rev. C 86, 014908 (2012), arXiv:1204.4713 [nucl-th], 10.1103/PhysRevC.86.014908.
* [139] A. Monnai, Relativistic Dissipative Hydrodynamic Description of the Quark-Gluon Plasma, PhD thesis, Tokyo U. (2014).
* [140] A. M. Halasz, A. Jackson, R. Shrock, M. A. Stephanov and J. Verbaarschot, Phys. Rev. D 58, 096007 (1998), arXiv:hep-ph/9804290, 10.1103/PhysRevD.58.096007.
* [141] M. A. Stephanov, Prog. Theor. Phys. Suppl. 153, 139 (2004), arXiv:hep-ph/0402115, 10.1142/S0217751X05027965.
* [142] R. Guida and J. Zinn-Justin, Nucl. Phys. B 489, 626 (1997), arXiv:hep-th/9610223, 10.1016/S0550-3213(96)00704-3.
* [143] STAR Collaboration (L. Adamczyk et al.), Phys. Rev. Lett. 108, 202301 (2012), arXiv:1112.3930 [nucl-ex], 10.1103/PhysRevLett.108.202301.
* [144] STAR Collaboration (L. Adamczyk et al.), Phys. Rev. Lett. 112, 032302 (2014), arXiv:1309.5681 [nucl-ex], 10.1103/PhysRevLett.112.032302.
* [145] E. V. Shuryak, Sov. Phys. JETP 47, 212 (1978).
* [146] J. I. Kapusta, Nucl. Phys. B 148, 461 (1979), 10.1016/0550-3213(79)90146-9.
* [147] T. Toimela, Phys. Lett. B 124, 407 (1983), 10.1016/0370-2693(83)91484-3.
* [148] P. B. Arnold and C.-X. Zhai, Phys. Rev. D 50, 7603 (1994), arXiv:hep-ph/9408276, 10.1103/PhysRevD.50.7603.
* [149] P. B. Arnold and C.-x. Zhai, Phys. Rev. D 51, 1906 (1995), arXiv:hep-ph/9410360, 10.1103/PhysRevD.51.1906.
* [150] C.-x. Zhai and B. M. Kastening, Phys. Rev. D 52, 7232 (1995), arXiv:hep-ph/9507380, 10.1103/PhysRevD.52.7232.
* [151] K. Kajantie, M. Laine, K. Rummukainen and Y. Schroder, Phys. Rev. D 67, 105008 (2003), arXiv:hep-ph/0211321, 10.1103/PhysRevD.67.105008.
* [152] N. Su, Commun. Theor. Phys. 57, 409 (2012), arXiv:1204.0260 [hep-ph], 10.1088/0253-6102/57/3/12.
* [153] J.-P. Blaizot, E. Iancu and A. Rebhan, Thermodynamics of the high temperature quark gluon plasma (3 2003), pp. 60–122. arXiv:hep-ph/0303185.
* [154] U. Kraemmer and A. Rebhan, Rept. Prog. Phys. 67, 351 (2004), arXiv:hep-ph/0310337, 10.1088/0034-4885/67/3/R05.
* [155] J. O. Andersen and M. Strickland, Annals Phys. 317, 281 (2005), arXiv:hep-ph/0404164, 10.1016/j.aop.2004.09.017.
* [156] J. O. Andersen, E. Braaten and M. Strickland, Phys. Rev. Lett. 83, 2139 (1999), arXiv:hep-ph/9902327, 10.1103/PhysRevLett.83.2139.
* [157] J. O. Andersen, E. Braaten and M. Strickland, Phys. Rev. D 61, 014017 (2000), arXiv:hep-ph/9905337, 10.1103/PhysRevD.61.014017.
* [158] J. O. Andersen, E. Braaten and M. Strickland, Phys. Rev. D 61, 074016 (2000), arXiv:hep-ph/9908323, 10.1103/PhysRevD.61.074016.
* [159] A. Vuorinen, Phys. Rev. D 67, 074032 (2003), arXiv:hep-ph/0212283, 10.1103/PhysRevD.67.074032.
* [160] A. Vuorinen, Phys. Rev. D 68, 054017 (2003), arXiv:hep-ph/0305183, 10.1103/PhysRevD.68.054017.
* [161] N. Haque, M. G. Mustafa and M. Strickland, Phys. Rev. D 87, 105007 (2013), arXiv:1212.1797 [hep-ph], 10.1103/PhysRevD.87.105007.
* [162] N. Haque, J. O. Andersen, M. G. Mustafa, M. Strickland and N. Su, Phys. Rev. D 89, 061701 (2014), arXiv:1309.3968 [hep-ph], 10.1103/PhysRevD.89.061701.
* [163] J. O. Andersen, N. Haque, M. G. Mustafa and M. Strickland, Phys. Rev. D 93, 054045 (2016), arXiv:1511.04660 [hep-ph], 10.1103/PhysRevD.93.054045.
* [164] K. Fukushima, Phys. Lett. B 591, 277 (2004), arXiv:hep-ph/0310121, 10.1016/j.physletb.2004.04.027.
* [165] C. Ratti, M. A. Thaler and W. Weise, Phys. Rev. D 73, 014019 (2006), arXiv:hep-ph/0506234, 10.1103/PhysRevD.73.014019.
* [166] J. M. Maldacena, Int. J. Theor. Phys. 38, 1113 (1999), arXiv:hep-th/9711200, 10.1023/A:1026654312961.
* [167] S. Gubser, I. R. Klebanov and A. M. Polyakov, Phys. Lett. B 428, 105 (1998), arXiv:hep-th/9802109, 10.1016/S0370-2693(98)00377-3.
* [168] E. Witten, Adv. Theor. Math. Phys. 2, 253 (1998), arXiv:hep-th/9802150, 10.4310/ATMP.1998.v2.n2.a2.
* [169] P. Kovtun, D. T. Son and A. O. Starinets, Phys. Rev. Lett. 94, 111601 (2005), arXiv:hep-th/0405231, 10.1103/PhysRevLett.94.111601.
* [170] A. Buchel, Phys. Lett. B 663, 286 (2008), arXiv:0708.3459 [hep-th], 10.1016/j.physletb.2008.03.069.
* [171] M. Natsuume and T. Okamura, Phys. Rev. D 77, 066014 (2008), arXiv:0712.2916 [hep-th], 10.1103/PhysRevD.78.089902, [Erratum: Phys.Rev.D 78, 089902 (2008)].
* [172] P. Romatschke, Class. Quant. Grav. 27, 025006 (2010), arXiv:0906.4787 [hep-th], 10.1088/0264-9381/27/2/025006.
* [173] R. Rougemont, A. Ficnar, S. Finazzo and J. Noronha, JHEP 04, 102 (2016), arXiv:1507.06556 [hep-th], 10.1007/JHEP04(2016)102.
* [174] O. DeWolfe, S. S. Gubser and C. Rosen, Phys. Rev. D 83, 086005 (2011), arXiv:1012.1864 [hep-th], 10.1103/PhysRevD.83.086005.
* [175] O. DeWolfe, S. S. Gubser and C. Rosen, Phys. Rev. D 84, 126014 (2011), arXiv:1108.2029 [hep-th], 10.1103/PhysRevD.84.126014.
* [176] Y. Nara, N. Otuka, A. Ohnishi and T. Maruyama, Prog. Theor. Phys. Suppl. 129, 33 (1997), 10.1143/PTPS.129.33.
* [177] P. Demorest, T. Pennucci, S. Ransom, M. Roberts and J. Hessels, Nature 467, 1081 (2010), arXiv:1010.5788 [astro-ph.HE], 10.1038/nature09466.
* [178] F. Karsch, B.-J. Schaefer, M. Wagner and J. Wambach, Phys. Lett. B 698, 256 (2011), arXiv:1009.5211 [hep-ph], 10.1016/j.physletb.2011.03.013.
* [179] P. Romatschke, Eur. Phys. J. C 77, 21 (2017), arXiv:1609.02820 [nucl-th], 10.1140/epjc/s10052-016-4567-x.
* [180] M. P. Heller, A. Kurkela, M. Spaliński and V. Svensson, Phys. Rev. D 97, 091503 (2018), arXiv:1609.04803 [nucl-th], 10.1103/PhysRevD.97.091503.
* [181] W. Florkowski, M. P. Heller and M. Spalinski, Rept. Prog. Phys. 81, 046001 (2018), arXiv:1707.02282 [hep-ph], 10.1088/1361-6633/aaa091.
* [182] W. Florkowski and R. Ryblewski, Phys. Rev. C 83, 034907 (2011), arXiv:1007.0130 [nucl-th], 10.1103/PhysRevC.83.034907.
* [183] M. Martinez and M. Strickland, Nucl. Phys. A 848, 183 (2010), arXiv:1007.0889 [nucl-th], 10.1016/j.nuclphysa.2010.08.011.
* [184] D. Bazow, U. W. Heinz and M. Strickland, Phys. Rev. C 90, 054910 (2014), arXiv:1311.6720 [nucl-th], 10.1103/PhysRevC.90.054910.
* [185] M. Alqahtani, D. Almaalol and M. Strickland, MDPI Proc. 10, 38 (2019), arXiv:1811.01856 [hep-ph], 10.3390/proceedings2019010038.
* [186] M. Alqahtani and M. Strickland (8 2020), arXiv:2008.07657 [nucl-th].
* [187] Particle Data Group Collaboration (M. Tanabashi et al.), Phys. Rev. D 98, 030001 (2018), 10.1103/PhysRevD.98.030001.
* [188] J. Cleymans, H. Oeschler, K. Redlich and S. Wheaton, Phys. Rev. C 73, 034905 (2006), arXiv:hep-ph/0511094, 10.1103/PhysRevC.73.034905.
* [189] M. Li and C. Shen, Phys. Rev. C 98, 064908 (2018), arXiv:1809.04034 [nucl-th], 10.1103/PhysRevC.98.064908.
* [190] NA49 Collaboration (C. Alt et al.), Phys. Rev. C 77, 024903 (2008), arXiv:0710.0118 [nucl-ex], 10.1103/PhysRevC.77.024903.
* [191] NA49 Collaboration (S. Afanasiev et al.), Phys. Rev. C 66, 054902 (2002), arXiv:nucl-ex/0205002, 10.1103/PhysRevC.66.054902.
* [192] STAR Collaboration (L. Adamczyk et al.), Phys. Rev. C 96, 044904 (2017), arXiv:1701.07065 [nucl-ex], 10.1103/PhysRevC.96.044904.
* [193] B. Schenke, C. Shen and P. Tribedy, Phys. Rev. C 99, 044908 (2019), arXiv:1901.04378 [nucl-th], 10.1103/PhysRevC.99.044908.
* [194] NA49 Collaboration (C. Alt et al.), Phys. Rev. Lett. 94, 192301 (2005), arXiv:nucl-ex/0409004, 10.1103/PhysRevLett.94.192301.
* [195] NA49 Collaboration (C. Alt et al.), Phys. Rev. C 73, 044910 (2006), 10.1103/PhysRevC.73.044910.
* [196] NA49 Collaboration (C. Alt et al.), Phys. Rev. C 78, 034918 (2008), arXiv:0804.3770 [nucl-ex], 10.1103/PhysRevC.78.034918.
* [197] NA49 Collaboration (C. Alt et al.), Phys. Rev. C 78, 044907 (2008), arXiv:0806.1937 [nucl-ex], 10.1103/PhysRevC.78.044907.
* [198] M. L. Miller, K. Reygers, S. J. Sanders and P. Steinberg, Ann. Rev. Nucl. Part. Sci. 57, 205 (2007), arXiv:nucl-ex/0701025, 10.1146/annurev.nucl.57.090506.123020.
* [199] S. Jeon and J. I. Kapusta, Phys. Rev. C 56, 468 (1997), arXiv:nucl-th/9703033, 10.1103/PhysRevC.56.468.
* [200] B. Schenke, S. Jeon and C. Gale, Phys. Rev. C 82, 014903 (2010), arXiv:1004.1408 [hep-ph], 10.1103/PhysRevC.82.014903.
* [201] B. Schenke, S. Jeon and C. Gale, Phys. Rev. C 85, 024901 (2012), arXiv:1109.6289 [hep-ph], 10.1103/PhysRevC.85.024901.
* [202] S. Bass et al., Prog. Part. Nucl. Phys. 41, 255 (1998), arXiv:nucl-th/9803035, 10.1016/S0146-6410(98)00058-1.
* [203] M. Bleicher et al., J. Phys. G 25, 1859 (1999), arXiv:hep-ph/9909407, 10.1088/0954-3899/25/9/308.
* [204] `https://edms.cern.ch/document/1075059` .
* [205] J. S. Moreland and R. A. Soltz, Phys. Rev. C 93, 044913 (2016), arXiv:1512.02189 [nucl-th], 10.1103/PhysRevC.93.044913.
* [206] C. Ratti, J. Phys. Conf. Ser. 736, 012001 (2016), 10.1088/1742-6596/736/1/012001.
* [207] P. Alba et al., Phys. Rev. D 96, 034517 (2017), arXiv:1702.01113 [hep-lat], 10.1103/PhysRevD.96.034517.
* [208] B. Schenke, P. Tribedy and R. Venugopalan, Phys. Rev. Lett. 108, 252301 (2012), arXiv:1202.6646 [nucl-th], 10.1103/PhysRevLett.108.252301.
* [209] B. Schenke, C. Shen and P. Tribedy, Phys. Rev. C 102, 044905 (2020), arXiv:2005.14682 [nucl-th], 10.1103/PhysRevC.102.044905.
* [210] `https://sites.google.com/view/qcdneos/` .
* [211] `https://www.bnl.gov/physics/best/resources.php` .
* [212] `http://nsmn1.uh.edu/cratti/EoS_BQS.html` .
* [213] `http://www.ift.uni.wroc.pl/~huovinen/` .
* [214] `https://arxiv.org/src/2006.12499v2/anc` .
* [215] `https://github.com/j-f-paquet/eos_maker` .
|
11institutetext: Institute for Health and Society, Medical College of
Wisconsin, 8701 Watertown Plank Rd, Wauwatosa, WI, 53226 22institutetext:
Department of Statistics, University of Missouri, 600 S State St., Apt. 408
Bellingham, WA 98225
# Most Powerful Test Sequences with Early Stopping Options
Sergey Tarima Nancy Flournoy
(Received: date / Accepted: date)
###### Abstract
Sequential likelihood ratio testing is found to be most powerful in sequential
studies with early stopping rules when grouped data come from the one-
parameter exponential family. First, to obtain this elusive result, the
probability measure of a group sequential design is constructed with support
for all possible outcome events, as is useful for designing an experiment
prior to having data. This construction identifies impossible events that are
not part of the support. The overall probability distribution is dissected
into stage specific components. These components are sub-densities of interim
test statistics first described by Armitage, McPherson and Rowe (1969) that
are commonly used to create stopping boundaries given an $\alpha$-spending
function and a set of interim analysis times. Likelihood expressions
conditional on reaching a stage are given to connect pieces of the probability
anatomy together.
The reduction of the support caused by the adoption of an early stopping rule
induces sequential truncation (not nesting) in the probability distributions
of possible events. Multiple testing induces mixtures on the adapted support.
Even asymptotic distributions of inferential statistics are mixtures of
truncated distributions. In contrast to the classical result on local
asymptotic normality (Le Cam 1960), statistics that are asymptotically normal
without stopping options have asymptotic distributions that are mixtures of
truncated normal distributions under local alternatives with stopping options;
under fixed alternatives, asymptotic distributions of test statistics are
degenerate.
###### Keywords:
Adaptive designs adapted support group sequential designs local asymptotics
interim hypothesis testing likelihood ratio tests
## 1 Introduction
We define a _sequential experiment_ to be one in which the decision to stop
collecting data is based on data collected previously in the study. Wetherill
and Glazebrook (1986) emphasize that ” _two aspects of a sequential procedure
must be clearly distinguished, the stopping rule, and the manner in which
inferences are made once observations are stopped. $\ldots$ in the design
problem, it is important to know how probable are various possible results.”_
We distinguish the probability framework underpinning these two activities,
and also a third - the probability framework underlying interim hypothesis
tests.
Dodge and Romig (1929) proposed the first known sequential test procedure in
which a decision to stop or continue collecting data was based on prior data,
recognizing that decisions to stop or continue a trial made based on prior
observations could substantially reduce the expected numbers of required
subjects; it was a two-stage design. Bartky (1943) devised a multiple
sequential testing procedure for binomial data based on Neyman and Pearson
(1933)’s likelihood ratio test that Wald (1947) cites as a ”forerunner” to his
more general _sequential probability ratio test_ (SPRT) procedure in which the
probabilities of type I and II errors are controlled. Extensions with stopping
decisions based on groups of subjects [_Group Sequential Designs (GSDs)_] are
given in Jennison and Turnbull (1999).
Neyman and Pearson (1933) show that likelihood ratio tests are most powerful
for testing a simple null versus a simple alternative hypotheses. In Ferguson
(2014), the Karlin-Rubin theorem is viewed as an extension of Neyman-Pearson
approach to most powerful testing of composite hypotheses. The Karlin-Rubin
theorem applies to the one-dimensional exponential family. With sequential
stopping options, some elements of the sample space become impossible, which
changes the distributions of statistics. Sufficient statistics become
dependent on the random sample size; see Blackwell (1947). If statistics
belong to the exponential family without early stopping options, then they
belong to a curved exponential family when exposed to early stopping options
(Efron (1975); Liu and Hall (1999); Liu et al. (2006)). Section 3.2 shows that
despite being from a curved exponential family, the sequential tests based on
likelihood ratios continue to be most powerful for any $\alpha$-spending
function.
### 1.1 Notation and A Simple Example
This example demonstrates a couple important repercussions of sequential
stopping rules:
* •
The support is reduced,
* •
Bivariate normal random variables become non-observable; the observable
bivariate random variable is a mixture of truncated normal random variables,
Let $X_{1}$ and $X_{2}$ be $N(\theta,1)$ random variables with an unknown
location parameter $\theta$.
#### 1.1.1 Non-sequential experiments
If $X_{1}=x_{1}$ alone is observed, the log-likelihood
$l(\theta|X_{1}=x_{1})=-2\log\sqrt{2\pi}-0.5(x_{1}-\theta)^{2}$
is maximized at $\widehat{\theta}_{1}=x_{1}$.
If both $X_{1}=x_{1}$ and $X_{2}=x_{2}$ are observed independently, the random
variable $(X_{1},X_{2})$ is defined on the probability space
$\left(R^{2},{\cal{B}},P\right)$, where $R^{2}$ is the sample space
[$R=\left(-\infty,\infty\right)$], ${\cal{B}}$ is the Borel $\sigma$-algebra
on $R^{2}$ and $P$ is a bivariate normal distribution with mean vector
$(\theta,\theta)$, units variances and zero correlation.
The log-likelihood function is
$l(\theta|X_{1}=x_{1},X_{2}=x_{2})=-2\log\sqrt{2\pi}-0.5(x_{1}-\theta)^{2}-0.5(x_{2}-\theta)^{2}$
and the maximum likelihood estimator (MLE) of $\theta$ is
$\widehat{\theta}_{2}=(x_{1}+x_{2})/2$.
#### 1.1.2 Sequential experiments: likelihood, support and probability
measures
What happens in sequential settings when $X_{2}$ is only observed if
$X_{1}<2.18$? Let $D$ denote the random stopping stage. Then
$D=1+I(X_{1}<2.18)$, where $I(\cdot)$ is an indicator function. In this simple
example, $D$ is also the random sample size.
The joint distribution of (X,D) and the marginal distribution of X are the
same:
$\displaystyle
f_{\textbf{X}}\left(\textbf{x}|\theta\right)=f_{\textbf{X},D}\left(\textbf{x},d|\theta\right)$
$\displaystyle=\phi(x_{1}-\theta)\left[\phi(x_{2}-\theta)\right]^{I\left(d=2\right)}$
(1)
$\displaystyle=[\phi(x_{1}-\theta)]^{I(d=1)}\left[\phi(x_{1}-\theta)\phi(x_{2}-\theta)\right]^{I\left(d=2\right)},$
where $\phi(\cdot)$ is the standard normal density. The representation of
$f_{\textbf{X}}\left(\textbf{x}|\theta\right)$ in the first line of (1)
partitions the density according to data collection stages, while the
representation in the second line partitions the density according to stopping
stages. In canonical form,
$\displaystyle f_{\textbf{X},D}\left(\textbf{x},d|\theta\right)$
$\displaystyle=$ $\displaystyle
h(\mathbf{x})\exp\Big{(}\left[x_{1}+I(d=2)x_{2}\right]\theta-\left[1+I(d=2)\right]\frac{\theta^{2}}{2}\Big{)},$
(2)
where
$h(\mathbf{x})=\exp\Big{(}-\frac{1}{2}\left[x_{1}^{2}+I(d=2)x_{2}^{2}+I(d=2)\log\left(\sqrt{\pi}\right)+\log\left(\sqrt{\pi}\right)\right]\Big{)}$.
Thus, the density (2) belongs to the curved exponential family with a
sufficient statistic
$\left(\sum_{k=1}^{d}x_{i},d\right)=\left[x_{1}+I(d=2)x_{2},\,1+I(d=2)\right].$
Curved exponential families were defined by Efron (1975) and the sufficient
statistic with a random number of summands $\sum_{k=1}^{D}x_{k}$ was derived
by Blackwell (1947). Probability distribution (2) is a special case of the
exponential family derived in Liu et al. (2006) [see their formula (2.6)]. A
more general probability distribution of $\mathbf{X}$ is presented in Section
2.
The log-likelihood function
$\displaystyle l(\theta|X_{1}=x_{1},X_{2}=x_{2},D=d)$
$\displaystyle=\begin{cases}[\phi(x_{1}-\theta)]^{I(d=1)}&\textrm{ if }D=1,\\\
\left[\phi(x_{1}-\theta)\phi(x_{2}-\theta)\right]^{I\left(d=2\right)}&\textrm{
if }D=2\end{cases}$
is maximized at $\widehat{\theta}=I(d=1)x_{1}+I(d=2)(x_{1}+x_{2})/2.$
Consequently, the score function, the MLE and the observed information are the
same as for the non-sequential experiment. What changes?
The joint support of $X_{1}$ and $X_{2}$ changes because $X_{2}$ becomes
impossible (not just missing) when $X_{1}\geq 2.18$. The random variable
$(X_{1},X_{2})$ is non-observable when $X_{1}\geq 2.18$ and the joint
distribution of $X_{1}$ and $X_{2}$ is therefore truncated and not normal.
Formally, the support for joint density can be decomposed into support for the
experiment stopping with $X_{1}$ and support for the experiment continuing to
observe $X_{2}$:
${\cal{T}}=\\{x_{1}\geq 2.18\\}\ \cup\ \\{\\{x_{1}<2.18\\}\cap\\{x_{2}\in
R\\}\\}\ \subset R^{2}.$
This ${\cal{T}}$ is a special case of the support formalized in Liu et al.
(2006). If $A\in{\cal{T}}$, then
$\displaystyle P_{(D)}=\text{Pr}((X_{1},X_{2})\in A)$ $\displaystyle=$
$\displaystyle\text{Pr}(X_{1}\geq 2.18)\text{Pr}(X_{1}\in A|X_{1}\geq 2.18)$
(3) $\displaystyle+$
$\displaystyle\text{Pr}(X_{1}<2.18)\text{Pr}((X_{1},X_{2})\in A|X_{1}<2.18)$
is a probability measure on the $\sigma$-algebra $\sigma({\cal{T}})$. Thus,
the observable random variable
$\displaystyle\mathbf{X}_{{\cal{T}}}=\begin{cases}X_{1}&\textrm{ if }$D=1$\\\
(X_{1},X_{2})&\textrm{ if }$D=2$.\end{cases}$
is defined on the probability space
$\left({\cal{T}},\sigma({\cal{T}}),P_{(D)}\right).$
In contrast to $(X_{1},X_{2})$, $\mathbf{X}_{{\cal{T}}}$ is observable for
this sequential experiment. The MLE, is a random variable defined on
$\left({\cal{T}},\sigma({\cal{T}}),P_{(D)}\right)$ and its probability
distribution is a mixture of the left truncated normal random variable
$\\{X_{1}|D=1\\}$ and an average of the right truncated normal random variable
$\\{X_{1}|D=2\\}$ and the normal random variable $X_{2}$.
Thus, in this sequential experiment, MLEs are not normal random variables as
illustrated in Figure 1. The first column shows the distribution of the MLE if
the experiment stopped at stage 1, which is left truncated normal. The seconds
of histograms shows the distribution of the MLE if the experiment proceeded to
stage 2. This distribution is a mixture of right truncated data from stage 1
and untruncated normal data from stage 2. The final column shows unconditional
distribution of the MLE.
(a) $x_{1}:x_{1}\geq 2.18$, $\theta=0$ (b) $\frac{x_{1}+x_{2}}{2}:x_{1}<2.18$,
$\theta=0$ (c) $\widehat{\theta}$, $\theta=0$
(d) $x_{1}:x_{1}\geq 2.18$, $\theta=2.18$ (e)
$\frac{x_{1}+x_{2}}{2}:x_{1}<2.18$, $\theta=2.18$ (f) $\widehat{\theta}$,
$\theta=2.18$
Figure 1: Distribution of MLEs from the experiment described in Section 1.1.
$n_{1}=n_{2}=1$; $10^{6}$ Monte-Carlo simulations.
### 1.2 The Scope of this Paper
We consider sequential experiments having a small finite number of interim
decision points, that is, experimental set-ups for which Martingale central
limit theorems and Brownian theory are not suitable. Our interest is in
experiments that aim primarily on a hypothesis test of effect size. We focus
on characterizing the effect of sequential stopping rules on probability
distributions of test statistics.
In this manuscript, Section 2 introduces notation for GSDs with stopping rules
dependent on a parameter of interest [through the distributions of the test
statistics]. This section also presents distributions of cumulative test
statistics conditional on reaching a stage, conditional on stopping at a
stage, and unconditional defined on the probability space with truncation-
adapted support. All of these probability distributions are truncated or
truncated-mixtures. Section 3 presents likelihood-based inference on the
truncation-adapted probability space. In section 3.1 a local asymptotic
distribution of the MLEs is found to be non-degenerate; it is a mixture of
truncated normal distributions. Section 3.2 shows that the possibility of
early stopping does not change the monotonicity of likelihood ratios in the
one-parameter exponential family. Thus, stage-specific tests continue to be
uniformly most powerful by the Karlin-Rubin theorem, which makes sequential
tests based on monotone likelihood ratio uniformly most powerful. Throughout
Sections 2 and 3 theoretical results are illustrated by a two stage example,
Pocock (1977). Finally, Section 4 concludes this article with a summary and a
discussion of impact within the contemporary research environment.
## 2 Probability Distributions with Early Stopping
Let $X$ denote subjects’ outcome variable and assume that, when observed in
isolation, it has a probability distribution function or a probability mass
function $f_{X}=f_{X}\left(x|\theta\right)$. To simplify the material, the
term density is used to refer to probability measures without formally
distinguishing between them. Let a sequence of $X$s be observed with the
primary objective of testing the null hypothesis $H_{0}:\theta=0$ with overall
$\alpha$-level type 1 error and $1-\beta$ power at an alternative
$H_{1}:\theta=\theta_{1}$.
It is convenient to group the random sample into _stages_ separated by the
interim analysis times: $\mathbf{X}_{1},\mathbf{X}_{2},\ldots$, where
$\mathbf{X}_{k}=(X_{n_{(k-1)}+1},\ldots,X_{n_{(k)}})$;
$n_{(k)}=\sum_{j=1}^{k}n_{j}$; here for simplicity $n_{j}$ is a pre-specified
number of observations in stage $j$, $1\leq k\leq K$; $n_{(0)}=0$ and $K-1$ is
the maximum number of interim analyses permitted. Every stage is assumed to be
“reachable”, that is, there is a positive probability of reaching each stage.
Data collection at each stage is followed by a hypothesis test that results in
a decision to stop the study or to enroll a new group of patients; except that
if stage $K$ is reached, the experiment stops after $n_{(K)}$ observations
regardless of the last $n_{K}$ observations’ values.
### 2.1 Stopping Decisions
Let $T_{(k)}$ be a function of observations
$\mathbf{X}_{(k)}=(X_{1},\ldots,X_{n_{(k)}})$ that is compared against a
cutoff value $c_{k}$ to determine whether to stop at stage $k$ or continue
through stage $k+1$, $1\leq k<K-1$. These decisions are defined by the events
$\left\\{\cap_{j=1}^{k-1}\left\\{T_{(j)}\leq
c_{j}\right\\}\right\\}\cap\left\\{T_{(k)}>c_{k}\right\\}$ and
$\cap_{j=1}^{K-1}\left\\{T_{(j)}\leq c_{k}\right\\}$, respectively, and are
conveniently summarized by a random variable denoting the stopping stage:
$D=K\cdot I\left(\cap_{j=1}^{K-1}\left\\{T_{(j)}\leq
c_{j}\right\\}\right)+\sum_{k=1}^{K-1}k\cdot
I\left(\left(\cap_{j=1}^{k-1}\left\\{T_{(j)}\leq
c_{j}\right\\}\right)\cap\left\\{T_{(k)}>c_{k}\right\\}\right);$
$D\in\\{1,\ldots,K\\}$ will appear as random index such as in
$\mathbf{X}_{(D)}$ to emphasize that the stopping stage is unknown and is
described probabilistically though the random variable $D$. $\mathbf{X}_{(d)}$
is the random variable $\mathbf{X}_{(D)}$ conditioned on stopping with stage
$D=d,d=1,\ldots,K$.
It is important to account for $D$ in probability statements about $T_{(d)}$
because $D$ determines the observations’ probability support as illustrated in
Section 1.1.
### 2.2 After deciding to stop
(a) Sketch of disjoint support regions for subdensities by (scaleless) stage-
specific test statistics (b) Sketch of disjoint support regions for
subdensities by (scaleless) cumulative test statistics
Figure 2: Support associated with different stopping decisions when $K=3$;
$c_{k}$ is the critical value for stopping at stage $k$.
If study stops at stage $D=d<K$, $H_{0}$ is rejected. If $d=K$, the final
hypothesis test determines acceptance or rejection of $H_{0}$. At the time of
the final analysis, the density of observations conditional on stopping at
stage $d$ (i.e., the density of $\mathbf{X}_{(d)}=\mathbf{X}_{(D)}|\\{D=d\\}$)
is
$\displaystyle
f_{\mathbf{X}_{(d)}}^{C}=f_{\mathbf{X}_{(D)}}\left(\textbf{{x}}_{(D)}|D=d,\theta\right)=\frac{I\left(D=d\right)}{\text{Pr}_{\theta}\left(D=d\right)}f_{\mathbf{X}_{(d)}}\left(\textbf{{x}}_{(d)}|\theta\right)=\frac{f_{\mathbf{X}_{(d)}}^{sub}}{\text{Pr}_{\theta}\left(D=d\right)},$
(4)
where
$f_{\mathbf{X}_{(d)}}^{sub}=[f_{\mathbf{X}_{(d)}}(\textbf{{x}}_{(d)}|\theta)]^{I\left(D=d\right)}$
denotes the sub-density with support defined by $D=d$ (see the exemplary
sketches in Figure 2). Similarly, for a statistic
$T_{(d)}=T(\mathbf{X}_{(d)})$, the conditional on $D=d$ density is
$f_{T_{(d)}}^{C}$.
In contrast, if the stopping rule is not random and the experiment stops with
$n_{(d)}$ observations, $\text{Pr}_{\theta}(D=d)\equiv 1$ and the observations
have density $f_{\mathbf{X}_{(d)}}$.
### 2.3 At the Time of Experimental Design
Prior to data collection, both $D$ and $X_{(D)}$ are unknown and the joint
density of $X_{(D)}$ can be written in several ways:
$\displaystyle f_{\mathbf{X}_{(D)}}\left(\textbf{{x}}_{(D)}|\theta\right)$
$\displaystyle=$
$\displaystyle\sum_{d=1}^{K}[f_{\mathbf{X}_{(d)}}(\textbf{{x}}_{(d)}|\theta)]^{I\left(D=d\right)}$
(5) $\displaystyle=$
$\displaystyle\sum_{d=1}^{K}f_{\mathbf{X}_{(d)}}^{sub}=\sum_{d=1}^{K}\text{Pr}_{\theta}\left(D=d\right)f_{\mathbf{X}_{(d)}}^{C}.$
The joint density is a mixture of densities corresponding to possible outcome
vectors; i.e., these densities are defined on non-overlapping regions of the
density’s support.
When a test statistic $T_{d}$ summarizes $d$th stage data, the density of
$T_{(D)}$ can be written analogous to Equation (5) as
$\displaystyle f_{T_{(D)}}\left({\textit{t}}_{(D)}|\theta\right)$
$\displaystyle=$
$\displaystyle\sum_{d=1}^{K}[f_{{T}_{(d)}}({\textit{t}}_{(d)}|\theta)]^{I\left(D=d\right)}$
(6) $\displaystyle=$
$\displaystyle\sum_{d=1}^{K}f_{{T}_{(d)}}^{sub}=\sum_{d=1}^{K}\text{Pr}_{\theta}\left(D=d\right)f_{{T}_{(d)}}^{C}.$
Even if every stage-specific test statistic $T_{d}$ is normally distributed,
the distribution of $T_{(D)}$ is not.
### 2.4 At Interim Hypothesis Testing
Suppose at stage $d-1$, the decision was made to continue sampling, the
support for the $d$th test statistic is characterized by $D\geq d$. The
density of $T_{(D)}$ conditional on $D\geq d$ is
$\displaystyle f_{T_{(D)}}\left(t_{(D)}|D\geq d,\theta\right)$
$\displaystyle=$
$\displaystyle\frac{\sum_{k=d}^{K}f_{T_{(k)}}^{sub}}{\text{Pr}_{\theta}\left(D\geq
d\right)}=\sum_{k=d}^{K}\frac{\text{Pr}_{\theta}\left(D=k\right)}{\text{Pr}_{\theta}\left(D\geq
d\right)}f_{T_{(k)}}^{C}.$ (7)
Again, even if each $T_{k}$ is normal, the distributions of $T_{(D)}|D\geq d$
are not.
#### 2.4.1 Connection with Armitage’s algorithm
SAS’s popular SEQDESIGN procedure, R’s gsDesign package, Cytel’s EAST and
others assess type I and power properties using a recursive sub-density
formula (Armitage et al. (1969)) to evaluate the distribution of
$T_{(D)}|D\geq d$. Armitage’s subdensity is $f_{T_{(D)}|D\geq
d}^{sub}=\sum_{k=d}^{K}f_{T_{(k)}}^{sub}$ and
$\displaystyle f_{T_{(D)}}\left(t_{(D)}|D\geq d,\theta\right)$
$\displaystyle=$ $\displaystyle\frac{f_{T_{(D)}|D\geq
d}^{sub}}{\text{Pr}_{\theta}\left(D\geq d\right)}.$ (8)
For example, at $K=2$, the sub-density of $T_{(D)}|D\geq d$ is
$f^{sub}_{T_{(D)}|D\geq 2}(t|\theta)=\int_{-\infty}^{c_{1}}f_{T_{(D\geq
2)}|T_{(1)}}(t|t_{1},\theta)f_{T_{(1)}}(t_{1}|\theta)dt_{1}$ (9)
and its density is $f_{T_{(D)}|D\geq 2}(t|\theta)=f^{sub}_{T_{(D)}|D\geq
d}(t|\theta)\left(\int_{-\infty}^{c_{1}}f_{T_{(1)}}(t_{1}|\theta)dt_{1}\right)^{-1}$.
Recursively, the density conditional on reaching the $d$th interim analysis is
$f_{T_{(D)}}(t|D\geq d,\theta)=\frac{\int_{-\infty}^{c_{d-1}}f_{T_{(D\geq
d)}|T_{(d-1)}}(t|t_{d-1},\theta)f_{T_{(d-1)}}(t_{d-1}|\theta)dt_{d-1}}{\int_{-\infty}^{c_{d-1}}f_{T_{(d-1)}}(t_{d-1}|\theta)dt_{d-1}}.$
(10)
### 2.5 $\sigma-$fields and support defined by a set of critical values
At the time of $d$th hypothesis test, given $D\geq d$, the values
$\\{x_{j}:j>n_{(d)}\\}$ are not observable and hence do not contribute to the
density; indeed, they do not belong to the adaptation-rule driven sample
space, and consequently, they do not belong to a $\sigma$-field of the random
process being monitored: hence, they do not belong to the sequential
experiment as a whole. In this paper, by analogy with structural zeroes in
contingency tables, these values are excluded from the sample space.
The $d$th stage-specific test statistic $T_{d}=T\left(\textbf{X}_{d}\right)$
is defined on a probability space
$\left({\cal{T}}_{d},\sigma\left({\cal{T}}_{d}\right),P_{d}\right)$, where
${\cal{T}}_{d}$ is typically a real line (${\cal{R}}$),
$\sigma\left({\cal{T}}_{d}\right)$ is Borel $\sigma-$field and $P_{d}$ is a
probability measure on the measurable space
$\left({\cal{T}}_{d},\sigma\left({\cal{T}}_{d}\right)\right)$. A sequence of
nested $\sigma$-fields
${\cal{F}}_{(d)}:=\sigma\left({\cal{T}}_{1}\right)\times\cdots\times\sigma\left({\cal{T}}_{d}\right),$
$d=1,\ldots,K$, creates a filtration
$\textbf{F}=\left({\cal{F}}_{(d)}\right)_{d\leq K}$ on the product probability
space $\left({\cal{T}},{\cal{F}}_{(K)},P\right)$, where
${\cal{T}}={\cal{T}}_{1}\times\cdots\times{\cal{T}}_{K}$ and
$P=P_{1}\times\cdots\times P_{K}$. But in the presence of possible stopping,
not all combinations of $(T_{1},\ldots,T_{K})$ are possible. The cumulative
test statistics $T_{(d)}=T_{(d)}(T_{1},\ldots,T_{d})$ are defined only on a
subspace of the sample space ${\cal{T}}_{1}\times\cdots\times{\cal{T}}_{d}$.
Thus, probability environment substantially changes.
#### 2.5.1 Interim hypothesis testing
The statistic $T_{(1)}=T_{1}$ conditional on reaching stage $1$ $(D\geq 1)$ is
defined on the sample space ${\cal{T}}_{(1)}={\cal{T}}_{1}$; so
$\sigma\left({\cal{T}}_{(1)}\right)={\cal{F}}_{(1)}$. The statistic $T_{(2)}$
conditional on $D\geq 2$ is defined on
${\cal{T}}_{(2)}=\left[(t_{1},t_{2}):\\{t_{1}\leq c_{1}\\}\right],$
and $\sigma\left({\cal{T}}_{(2)}\right)\subset{\cal{F}}_{(2)}$. Further, for
$d\in\\{3,\ldots,K\\}$, $T_{(d)}$ conditional on $D\geq d$ is defined on
${\cal{T}}_{(d)}=\left((t_{1},\ldots,t_{d}):\left\\{t_{(j)}\leq
c_{j}\right\\},j=1,\ldots,d-1\right).$
For all $d$, the support ${\cal{T}}_{(d)}$ and the $\sigma$-field
$\sigma\left({\cal{T}}_{(d)}\right)$ is reduced by the possibility of early
stopping: $\sigma\left({\cal{T}}_{(d)}\right)\subset{\cal{F}}_{(d)}$. This
creates new measurable spaces
$\left({\cal{T}}_{(d)},\sigma\left({\cal{T}}_{(d)}\right)\right)$ for interim
tests at every possible stage $1,\ldots,K$. Armitage’s recursive sub-density
formula [Armitage et al. (1969)] is defined on this measurable space; see
Equation (9) for $K=2$. Re-scaling yields a density function [see Equation
(10)] which defines a probability measure to complete the probability space
$\left({\cal{T}}_{(d)},\sigma\left({\cal{T}}_{(d)}\right),P_{(d)}\right)$,
where the probability measure $P_{(d)}$ is determined by density (10):
$P_{(d)}(A)=\int_{A}f_{T_{(D)}}(t|D\geq d,\theta)dt$,
$A\in\sigma\left({\cal{T}}_{(d)}\right)$.
#### 2.5.2 After the stop decision is made
At each interim stage $d=1,\ldots,K-1$, the decision to reject or accept
$H_{0}$ splits ${\cal{T}}_{(d)}$ into two non-overlapping regions denoted
${\cal{T}}_{(d)}^{stop}$ and ${\cal{T}}_{(d)}^{cont}$, respectively. Since
${\cal{T}}_{(d)}^{stop}\cap{\cal{T}}_{(d+1)}=\emptyset$ and
${\cal{T}}_{(d)}^{cont}\subset{\cal{T}}_{(d+1)}$, then the sets
${\cal{T}}_{(1)}^{stop},\ldots,{\cal{T}}_{(K-1)}^{stop},$ and
${\cal{T}}_{(K)}$ make a partition of the sample space of $T_{(D)}$:
${\cal{T}}_{(K)}+\sum_{k=1}^{K-1}{\cal{T}}_{(k)}^{stop}=\left(\cup_{d=1}^{K-1}{\cal{T}}_{(d)}^{stop}\right)\cup{\cal{T}}_{(K)}=\cup_{d=1}^{K}{\cal{T}}_{(d)}\subset{\cal{T}}.$
We define a probability space for the observable random variable ${T}_{(d)}$
using the probability measures $P_{(d)}$ defined by $f_{T_{(d)}}^{C}$ on the
measurable space
$\left({\cal{T}}_{(d)}^{stop},\sigma\left({\cal{T}}_{(d)}^{stop}\right)\right)$.
At $D=K$, the probability measure $P_{(K)}$ defined by $f_{T_{(K)}}^{C}$
completes the probability space for a measurable space
$\left({\cal{T}}_{(K)},\sigma\left({\cal{T}}_{(K)}\right)\right)$.
Thus, at the end of the study, at $D=d$, the researcher operates with an
_observable_ random variable $T_{(d)}$.
#### 2.5.3 At the design stage
The conditional random variables, $T_{(d)}$, defined on non-overlapping
$\sigma$-fields are combined together into the unconditional random variable
$T_{(D)}\sim f_{T_{(D)}}$ defined on the sample space
$\left(\cup_{d=1}^{K-1}{\cal{T}}_{(d)}^{stop}\right)\cup{\cal{T}}_{(K)}$. The
probability distribution defined on the $\sigma$-field on this sample space is
$P_{(D)}=\sum_{d=1}^{K}\text{Pr}\left(D=d\right)P_{(d)}$.
#### 2.5.4 Impossible events
The set ${\cal{T}}_{0}={\cal{T}}\setminus\cup_{d=1}^{K}{\cal{T}}_{(d)}$
contains all impossible combinations of $(t_{1},\ldots,t_{K})$ under a chosen
stopping rule. If $K=2$, for example, then
${\cal{T}}\setminus\cup_{d=1}^{2}{\cal{T}}_{(d)}=\left\\{(t_{1},t_{2}):t_{1}>c_{1}\right\\}$.
Without the possibility of early stopping, all combinations of
$(t_{1},\ldots,t_{K})\in{\cal{T}}$ would be possible.
### 2.6 Pocock’s Example: One-Sided Two-Group Sequential Z-test
Pocock (1977) proposed a simple two-stage procedure for testing
$H_{0}:\theta=0$ with a pre-determined power at $H_{1}:\theta=\theta_{1}$ on
normal data. With $n_{1}=n_{2}$, $c_{1}=c_{2}=2.18$ is used to secure the
overall type 1 error rate $\alpha=0.025$ with a one-sided $z$ test. Let
$n_{0}=0$ and
$Z_{k}=\frac{1}{\sqrt{n_{k}}}\sum_{i=n_{(k-1)}+1}^{n_{(k)}}X_{i}=\sqrt{n_{k}}\cdot\bar{X}_{k}\overset{d}{=}\mathcal{N}(\theta,1),\quad
k=1,2.$
The study is stopped for efficacy at stage 1 if $Z_{1}\geq 2.18$ and proceeds
to stage 2 when $Z_{1}<2.18$. If the study is stopped at stage 1, the support
for $\bar{X}_{1}|Z_{1}\geq 2.18$ starts at $2.18/\sqrt{n_{1}}$ and stretches
to $+\infty$. If the study continues through stage 2, support for
$\bar{X}_{1}|Z_{1}<2.18$ ranges from $-\infty$ to $2.18/\sqrt{n_{1}}$ .
The test statistic for the second stage, under $Z_{1}<2.18$,
$Z_{(2)}=\frac{\sqrt{n_{1}}}{\sqrt{n_{1}+n_{2}}}Z_{1}+\frac{\sqrt{n_{2}}}{\sqrt{n_{1}+n_{2}}}Z_{2}=\frac{1}{\sqrt{n_{1}+n_{2}}}\sum_{i=1}^{n_{1}+n_{2}}X_{i}.$
The distribution of $Z_{(2)}$ is a mixture of a right-truncated normal
$Z_{1}|Z_{1}<2.18$ and the normal $Z_{2}$.
Figure 3 shows histograms of $Z_{1}|Z_{1}\geq 2.18$, $Z_{(2)}|Z_{1}<2.18$, and
$Z_{(D)}$ estimated from $100,000$ Monte-Carlo samples assuming $\theta=2.18$.
These histograms based on $n_{1}=n_{2}=100$ are almost identical to the
histograms in Figure 1 based on $n_{1}=n_{2}=1$. This highlights on the
important message that non-normality continues to be present even
asymptotically Tarima and Flournoy (2019a).
(a) $Z_{1}|Z_{1}\geq 2.18$, $\theta=2.18$ (b) $Z_{(2)}|Z_{1}<2.18,$
$\theta=2.18$ (c) $Z_{(D)},\theta=2.18$
Figure 3: Histograms of test statistics from a two-stage Pocock’s experiment
with critical value $c_{1}=2.18$; $f_{\mathbf{X}_{k}}=\mathcal{N}(2.18,1)$ and
$n_{k}=100,k=1,2$.
## 3 Likelihood-based Inference with Early Stopping
If the stopping rule is not random and the experiment stopped with $n_{(d)}$
observations, $\text{Pr}_{\theta}(D=d)\equiv 1$, the likelihood is
$\displaystyle{\mathcal{L}}^{fix}\left(\theta|k,\textbf{{x}}_{(d)}\right)=f_{\mathbf{X}_{(d)}}.$
(11)
Considering the joint density (5) conditional on the data
$\left(d,\textbf{{x}}_{(d)}\right)$ observed at the end of a sequential
experiment, the likelihood is
$\displaystyle{\mathcal{L}}\left(\theta|d,\textbf{{x}}_{(d)}\right)=f_{\mathbf{X}_{(d)}}^{sub}.$
(12)
The indicator in (4) emphasizes that support for the random variables is
reduced by the conditioning; see this illustrated in Figure 2. Note also in
Figure 2 that the support conditional on stopping at one stage is disjoint
from the support conditional on stopping at another stage.
${\cal{L}}$ is a function of $\theta$ and the observed data
$(d,\textbf{{x}}_{(d)})$, but ${\cal{L}}$ is a continuous function only of
$\theta$, as discontinuities in $(d,\textbf{{x}}_{(d)})$ arise from the
mixture distribution of $\mathbf{X}_{(D)}$. These discontinuities are
inherited by MLEs and other statistics derived from them. Conditional on
$(d,\textbf{{x}}_{(d)})$, MLEs maximizing ${\mathcal{L}}$ are
$\displaystyle\widehat{\theta}$
$\displaystyle=\arg\max_{\theta}f_{\mathbf{X}_{(d)}}^{sub}.$
For every $\left(d,\textbf{{x}}_{(d)}\right)$,
$f_{\mathbf{X}_{(d)}}^{sub}=f_{\mathbf{X}_{(d)}}$ for all $\theta$;
consequently,
$\widehat{\theta}^{fix}:=\arg\max_{\theta}f_{\mathbf{X}_{(d)}}=\arg\max_{\theta}f_{\mathbf{X}_{(d)}}^{sub}=\widehat{\theta},$
that is, as is well known, making stopping decisions does not alter maximum
likelihood point estimates; they are the same whether obtained by maximizing
${\mathcal{L}}$ or ${\mathcal{L}}^{fix}$ and other observed statistics derived
from the likelihood (e.g., the score function and the observed information)
are unaffected as well.
The following example extends the simple one in Section 1.1 with
$n_{1}=n_{2}=1$ to arbitrary $n_{1}$ and $n_{2}$ to illustrate (as is proven
later) that, although maximum likelihood point estimates and test statistics
are unaffected by early stopping decisions, their probability distribution
does not tend to normality even with larger sample sizes.
### 3.1 Large Sample Properties
The $d$th stage-specific MLEs $\widehat{\theta}_{d}$ of $\theta$ and their
statistical models are called _regular_ if, without the possibility of early
stopping,
$\xi_{d}=\sqrt{n_{d}}\left(\widehat{\theta}_{d}-\theta\right)\overset{d}{\to}{\cal{N}}(0,\sigma^{2}),$
(13)
where $0<\sigma<\infty$. Assumption (13) was described in Tarima and Flournoy
(2019a) to include the more specific assumptions for
* •
independent, identically distributed observations by Cramér [e.g., for
example, Ferguson (1996)],
* •
independent not identically distributed observations [e.g., Philippou et al.
(1973)],
* •
dependent observations [e.g., Crowder (1976)],
* •
and densities whose support depends on parameters [e.g., Wang et al. (2014)].
All these specific sets of assumptions include assumptions of the existence
and consistency of the MLE.
With large samples, the MLE at the stage $d$ analysis can be approximated
recursively by
$\displaystyle\widehat{\theta}_{(d)}$
$\displaystyle\approx\frac{n_{(d-1)}}{n_{(d)}}\widehat{\theta}_{(d-1)}+\frac{n_{d}}{n_{(d)}}\widehat{\theta}_{d}\approx\sum_{j=1}^{d}\frac{n_{j}}{n_{(d)}}\widehat{\theta}_{d},$
(14)
where $\widehat{\theta}_{(d-1)}$ is an MLE based on cumulative data from
stages $1$ to $d-1$ and $\widehat{\theta}_{d}$ is the $d$th stage-specific
MLE. The standardized $d$th stage-specific MLE is so,
$T_{(d)}=\sqrt{n_{(d)}}\sum_{j=1}^{d}\frac{n_{j}}{n_{(d)}}\left(\frac{\widehat{\theta}_{j}-\theta}{\sigma}\right)=\sum_{j=1}^{d}\sqrt{\frac{n_{j}}{n_{(d)}}}\
\xi_{j}\approx\sqrt{n_{(d)}}\left(\frac{\widehat{\theta}_{(d)}-\theta}{\sigma}\right),$
that is, $T_{(d)}$ is (approximately) the standardized $d$th stage-specific
MLE. The asymptotic properties of $T_{(D)}$ depend on the existence and
distribution of a limiting random variable $r_{(D)}$ defined by
$\sum_{d=1}^{K}I(D=d)\frac{n_{(d)}}{n_{d}}\overset{d}{\to}\sum_{d=1}^{K}I(D=d)r_{(d)}=r_{(D)},$
(15)
where $r_{(d)}=\lim_{n_{d}\to\infty}n_{(d)}/n_{d}$ is the asymptotic ratio of
the $d$th cumulative-stage and stage-specific sample sizes [Theorem 1 in
Tarima and Flournoy (2019a)]; $r_{(D)}$ is a multinomial random variable with
support on $r_{(d)}$.
Assume the limits $n_{(d)}/n_{j}\to r_{(d)j}\in(0,\infty)$, $j\leq d$ exist
with $r_{(d)d}=r_{(d)}$. Then given $D=d$,
$T_{(d)}\to\sum_{j=1}^{d}\frac{\xi_{j}}{\sqrt{r_{(d)j}}}.$
While $\xi_{j}\overset{d}{\to}{\cal{N}}(0,1)$, the distributions used for the
final analysis, the interim analysis and experimental design, respectively,
are mixtures of truncated distributions:
$\displaystyle{\text{Pr}_{\theta}}\left(T_{(D)}<v|D=d\right)$
$\displaystyle\to$
$\displaystyle\text{Pr}_{\theta}\left(\sum_{j=1}^{d}\frac{\xi_{j}}{\sqrt{r_{(d)j}}}<v\Big{|}D=d\right),$
$\displaystyle{\text{Pr}_{\theta}}\left(T_{(D)}<v|D\geq d\right)$
$\displaystyle\to$
$\displaystyle\sum_{k=d}^{K}{\text{Pr}_{\theta}}\left(D=k\right){\text{Pr}_{\theta}}\left(\sum_{j=1}^{k}\frac{\xi_{j}}{\sqrt{r_{(k)j}}}<v\Big{|}D=k\right),$
$\displaystyle\text{Pr}_{\theta}\left(T_{(D)}<v\right)$ $\displaystyle\to$
$\displaystyle\sum_{k=1}^{K}{\text{Pr}_{\theta}}\left(D=k\right){\text{Pr}_{\theta}}\left(\sum_{j=1}^{k}\frac{\xi_{j}}{\sqrt{r_{(k)j}}}<v\Big{|}D=k\right).$
### Pocock’s Example: Large Sample Properties
Under assumption (13) with large sample sizes,
$\widehat{\theta}_{(D)}=I(D=1)\widehat{\theta}_{(1)}+I(D=2)\widehat{\theta}_{(2)}$
are standardized as
$\displaystyle T_{(D)}$
$\displaystyle=\sqrt{n_{(D)}}\left(\widehat{\theta}_{(D)}-\theta\right)/\sigma$
$\displaystyle=I(D=1)\sqrt{n_{1}}\left(\widehat{\theta}_{(1)}-\theta\right)/\sigma+I(D=2)\sqrt{n_{1}+n_{2}}\left(\widehat{\theta}_{(2)}-\theta\right)/\sigma.$
(16)
Assume the limit
$I(D=1)+I(D=2)\frac{n_{1}+n_{2}}{n_{2}}\overset{d}{\to}r_{(D)}$ exists as
$n_{1}\to\infty$, Then, adopting the local alternative hypothesis
$\theta=h/\sqrt{n_{1}}$ yields a non-degenerate limiting distribution of
$T_{(D)}$ that models both stages of the experiment:
$\displaystyle\text{Pr}_{\theta}\left(T_{(D)}<v\right)$ $\displaystyle\to$
$\displaystyle p_{1}\Phi\left(\left.v\right|D=1\right)$ (17) $\displaystyle+$
$\displaystyle(1-p_{1})\int_{-\infty}^{c_{1}}\Phi\left(\sqrt{r_{(D)}}v-\sqrt{r_{(D)}-1}y\right)\phi\left(y|D=2\right)dy,$
where $p_{1}=\lim_{n_{1}\to\infty}{\text{Pr}_{\theta}}\left(D=1\right)$ is the
limiting stage 1 stopping probability; $\phi$ and $\Phi$ denote the standard
normal density and cumulative distribution function, respectively. With a
fixed alternative $\theta$,
$\text{Pr}_{\theta}\left(T_{(D)}<v\right)\to\Phi(v)$ because $p_{1}\to 1$ with
probability 1, and modeling related to stage 2 data is lost.
In Pocock’s example,
$T_{(D)}=I(D=1)Z_{1}+I(D=2)\left(Z_{1}+Z_{2}\right)/\sqrt{2}$ and
$\displaystyle\text{Pr}_{\theta}\left(T_{(D)}<v\right)$ $\displaystyle\to$
$\displaystyle\text{Pr}_{\theta}\left(Z_{1}>c_{1}\right)\Phi\left(\left.v\right|Z_{1}>c_{1}\right)$
(18) $\displaystyle+$ $\displaystyle\text{Pr}_{\theta}\left(Z_{1}\leq
c_{1}\right)\int_{-\infty}^{c_{1}}\Phi\left(\sqrt{2}v-y\right)\frac{\phi\left(y\right)}{\text{Pr}\left(y\leq
c_{1}\right)}dy$
Note if $Z_{1}\leq c_{1}$, then
$\text{Pr}_{\theta}\left(\frac{Z_{1}+Z_{2}}{\sqrt{2}}<v|Z_{1}\leq
c_{1}\right)=\int_{-\infty}^{c_{1}}\Phi\left(\sqrt{2}v-y\right)\frac{\phi\left(y\right)}{\text{Pr}\left(y\leq
c_{1}\right)}dy,$
which is a continuous mixture of distributions.
### 3.2 Most Powerful Group Sequential Tests
Let $X\sim f_{X}(\theta)$, where $f$ belongs to the one-parameter exponential
family. Without a possibility of early stopping, the likelihood for a
realization $\textbf{{x}}=\left(x_{1},\ldots,x_{n}\right)$ of a random sample
$\textbf{X}=\left(X_{1},\ldots,X_{n}\right)$ is
$\displaystyle{\cal{L}}\left(\theta|\textbf{{x}}\right)=\prod_{i=1}^{n}f_{X}\left(\textbf{{x}}\right)$
$\displaystyle=h\left(\textbf{{x}}\right)g\left(T\left(\textbf{{x}}\right)|\theta\right)=h\left(\textbf{{x}}\right)e^{\eta(\theta)T\left(\textbf{{x}}\right)+A\left(\theta\right)}$
where all relevant information about $\theta$ is absorbed by a sufficient
statistic $T\left(\textbf{{x}}\right)$. Assume the test statistic $Z$ is a
one-to-one transformation of $T$.
If $LR(t)=g(t|\theta)/g(t|\theta_{0})$ has monotone likelihood ratio (MLR) in
$t$, then the Karlin-Rubin theorem provides uniformly most powerful (UMP)
tests.
In sequential testing settings, when $d$th stage is reached $\left(D\geq
d\right)$, the interim likelihood
$\displaystyle{\cal{L}}\left(\theta|D\geq d,\textbf{{x}}_{(d)}\right)$
$\displaystyle=$ $\displaystyle I\left(D\geq
d\right)h\left(\textbf{{x}}_{(d)}\right)e^{\eta(\theta)T\left(\textbf{{x}}_{(d)}\right)+A_{d}\left(\theta\right)}$
(19)
has the associated interim likelihood ratio
$\displaystyle LR(t|D\geq d)=\frac{{\cal{L}}\left(\theta|D\geq
d,\textbf{{x}}_{(d)}\right)}{{\cal{L}}\left(\theta_{0}|D\geq
d,\textbf{{x}}_{(d)}\right)}.$
For every $d$th stage hypothesis test, $D\geq d$ and
$LR(t|D\geq
d)=\exp\left[\left(\eta(\theta_{1})-\eta(\theta_{0})\right)T_{(d)}+\left(A_{d}(\theta_{1})-A_{d}(\theta_{0})\right)\right]$
which means that the MLR property is preserved with early stopping.
Definition: A sequence of $\alpha_{d}$-level interim tests
$\\{T_{(1)}>c_{1}\\},\ldots,\\{T_{(D)}>c_{D}\\}$ will be called the sequential
test.
###### Theorem 3.1
For any fixed $(\alpha_{1},\ldots,\alpha_{K})$ and
$\left(n_{1},\ldots,n_{K}\right)$ $(1)$ the interim LR test
$\\{T_{(d)}>c_{d}\\}$ is a UMP $\alpha_{d}$-level test and $(2)$ no sequential
test is more powerful than the sequential test based interim LRs.
Feature | Mathematical Definition
---|---
overall type $1$ error | $\alpha=\sum_{k=1}^{K}\alpha_{k}\prod_{j=1}^{k-1}\text{Pr}_{0}\left(Z_{(j)}\leq u_{j}\right)$
stage-specific type 1 error | $\alpha_{d}=\text{Pr}_{0}\left(Z_{(d)}>u_{d}\big{|}\cap_{j=1}^{d-1}\left\\{Z_{(j)}\leq u_{j}\right\\}\right)$
$\alpha$-spending function | $\alpha_{(d)}=\sum_{j=1}^{d}\alpha_{j}\prod_{i=1}^{j-1}\text{Pr}_{0}\left(Z_{(j)}\leq u_{j}\right)$
overall type 2 error | $\beta(\theta)=\sum_{j=1}^{K}\beta_{j}(\theta)\prod_{i=1}^{j-1}\text{Pr}_{\theta}\left(Z_{(j)}\leq u_{j}\right)$
stage-specific type 2 error | $\beta_{d}(\theta)=\text{Pr}_{\theta}\left(Z_{(d)}>u_{d}\big{|}\cap_{j=1}^{d-1}\left\\{Z_{(j)}\leq u_{j}\right\\}\right)$
$\beta$-spending function | $\beta_{(d)}\left(\theta\right)=\sum_{j=1}^{d}\beta_{j}(\theta)\prod_{i=1}^{j-1}\text{Pr}_{\theta}\left(Z_{(j)}\leq u_{j}\right)$
overall power | $1-\beta(\theta)$
stage-specific power | $1-\beta_{d}(\theta)$
cumulative power | $1-\beta_{(d)}\left(\theta\right)$
Table 1: Definitions of Operational Characteristics for Sequential Tests
$d=1,\ldots,K$; $\prod_{i=1}^{0}[\cdot]=1$.
Proof. $(1)$ As shown above, conditional on reaching stage $d$, $(D\geq d)$,
$T_{(d)}$ continues to be sufficient in exponential families and the MLR
property is preserved. By the Karlin-Rubin theorem, the test based on
$T_{(d)}$ is uniformly most powerful at $d^{th}$ stage.
$(2)$ Using Table 1 notation,
$1-\beta(\theta)=\sum_{k=1}^{K}\left(\prod_{j=1}^{k-1}\left[\beta_{j}(\theta)\right]\right)[1-\beta_{k}(\theta)]=1-\prod_{k=1}^{K}\beta_{k}(\theta).$
At $K=1$, $\alpha_{1}$ and $n_{1}$ uniquely define $c_{1}$ and
$\left\\{T_{(1)}>c_{1}\right\\}$ is a UMP by part 1 with a stage-specific
power curve $1-\beta_{1}(\theta)$. At an arbitrary stage $d$, $\alpha_{d}$ and
$n_{d}$ uniquely define $c_{d}$ and, by part 1 of the theorem, the stage-
specific power $1-\beta_{d}(\theta)$, is the highest. Thus, for any choice of
$(\alpha_{1},\ldots,\alpha_{K})$ and $\left(n_{1},\ldots,n_{K}\right)$, and
consequently $\\{T_{(d)}>c_{d}\\}$, the power of the sequential test based on
LRs is $1-\beta(\theta)=1-\prod_{k=1}^{K}\beta_{k}(\theta)$. This power is the
highest for any given $\theta$, because the stage specific type 2 errors
$\beta_{d}(\theta)$ are the lowest for each $d$ at any $\theta$. Q.E.D.
### Pocock’s Example: Likelihood Ratio
For $X_{i}\sim{\cal{N}}(\theta,1)$, $i=1,\ldots,n$, the likelihood ratio for
testing $H_{0}:\theta=\theta_{0}$ vs $H_{A}:\theta=\theta_{1}>\theta_{0}$ is
$\log LR(t)=-n\left(\log\theta_{0}-\log\theta_{1}\right)-t/\theta_{1},$
which has rejection region $\left\\{t>c\right\\}$, where
$t=\sum_{i=1}^{n}X_{i}$ is a sufficient statistic. Using $Z=t/\sqrt{n}\sim
f_{Z}={\cal{N}}(\theta_{1},1)$, the likelihood ratio test is
$\left\\{Z>c/\sqrt{n}\right\\}$.
With Pocock’s example, under $D\geq 1$, $Z_{1}$ is a normal random variable
used for stage 1 hypothesis testing, and
$\log LR(Z_{1}|D\geq
1)=-n_{1}\left(\log\theta_{0}-\log\theta_{1}\right)-\sqrt{n_{1}}Z_{1}/\theta_{1}.$
Then,
$\log LR(Z_{(2)}|D\geq
2)=-(n_{1}+n_{2})\left(\log\theta_{0}-\log\theta_{1}\right)-\sqrt{n_{1}+n_{2}}Z_{(2)}/\theta_{1}+C,$
where
$C=\log\text{Pr}\left(Z_{1}<c_{1}|\theta_{0}\right)-\log\text{Pr}\left(Z_{1}<c_{1}|\theta_{1}\right)$
is independent of data given $D\geq 2$. By Theorem 3.1, $\\{Z_{(d)}\geq
c_{d}\\}$ is the UMP test at stage $d$ given $\alpha_{d}$ and $n_{d}$, and no
other sequential test is more powerful overall.
## 4 Impact and Summary
To establish that a sequence of likelihood ratio tests are most powerful, we
began by constructing the joint probability distribution over the set of
possible events. The use of an early stopping criterion eliminates the
possibility of some realizations of cumulative test statistics
$\mathbf{Z}=\left(Z_{(1)},\ldots,Z_{(K)}\right)$. This makes an otherwise
normal random process $\mathbf{Z}$ unobservable. On its true stopping rule
adapted support, the distribution of $\mathbf{Z}$ is a mixture of truncated
distributions at each stage. Thus, unadapted distributional assumptions should
be modified to take into account the planned adaptation scheme. Liu and Hall
(1999) and Liu et al. (2006) recognized change in support in the one-parameter
exponential family and investigated bias estimation. Schou and Marschner
(2013) recognized presence of truncation in the joint distribution of stage-
specific test statistics, but they mostly focused on bias in meta-analytic
studies. The adapted support is critical for derivation MLEs’ and test
statistics’ distributions in Section 2.
In Section 2, distinct probability measures are derived for design
[unconditional], interim hypothesis testing [conditional on collecting data up
to the time of interim testing] and when the study is completed [conditional
on deciding to stop at a particular stage]. These probability distributions
formalized a new probabilistic framework (Section 2) which is used in Section
3.2 to show no testing sequence is more powerful than sequential likelihood
ratio tests.
Likelihood ratio tests are most powerful for testing simple hypotheses and,
under monotone likelihood ratio, for testing composite hypotheses; see Neyman
and Pearson (1933) and Karlin-Rubin theorem in Ferguson (2014). The Karlin-
Rubin theorem works with the one-parameter exponential family and Section 3.2
shows that most powerful testing continues to hold for sequential tests. Even
though distributions of sample-size-dependent sufficient statistics (Blackwell
(1947)) belong to a curved exponential family (Efron (1975); Liu and Hall
(1999); Liu et al. (2006)), the distributions conditional on the sample size
still belong to one-parameter exponential family. This fact is used in Section
3.2 to show that likelihood ratio sequential tests are most powerful tests
with any pre-determined $\alpha$-spending function. This result is applicable
to many common group sequential designs.
Treatment-effect-dependent (non-ancillary) stopping rules are part of many
common GSDs including Pocock Pocock (1977), O’Brien & Fleming O’Brien and
Fleming (1979), and Haybittle-Peto designs Haybittle (1971); Peto et al.
(1976). If the data follow a normal distribution with known variance (without
possibility of early stopping), then these designs are most powerful for their
$\alpha$-spending functions. Similarly, application of Simes test (see Simes
(1986)) and its recently proposed modification for sequential testing (see
Tamhane et al. (2020)) cannot have higher power than group sequential tests
from Jennison and Turnbull (1999) with the same $\alpha$-spending function.
Historically, many researchers have relied, and currently rely, on joint
normality (e.g., Jennison and Turnbull (1999), Proschan et al. (2006), Kunz et
al. (2020)). Critical adjustments ate often done with recursive sub-density
estimation; see Armitage et al. (1969). Section 2.4.1 places Armitage’s sub-
density formula in the broader probability framework of possible events and
confirms, from the prospective of this $\sigma$-algebra, that the common
current practice of using Armitage’s formula for calculating distributions of
interim test statistics is appropriate.
There is plenty of evidence that adaptive designs make statistics non-normally
distributed. Demets and Lan (1994) point out that the distribution of stage-
specific test statistics is not normal and should be estimated recursively.
Jennison and Turnbull (1999) plot the density of a normal test statistic used
in GSD settings [pages 174-177], where discontinuity points clearly show non-
normality. Li et al. (2002) find the joint density of stage 1 and stage 2
standardized test statistics not to be bivariate normal. Local asymptotic non-
normality was established following sample size recalculations (SSRs) that
depend on an interim observed treatment effect (Tarima and Flournoy (2019a,
b)); and a GSD with a single interim analysis can be viewed as a special case
of an SSR. MLEs converging to random mixtures of normal variables have been
found in other adaptive designs (Ivanova et al. (2000), Ivanova and Flournoy
(2001), May and Flournoy (2009), Lane and Flournoy (2012), Flournoy et al.
(2018)).
Milanzi et al. (2015) developed a likelihood approach that applies when the
early stopping rule does not depend on the parameter of interest. In this
case, sample size adaptation is ancillary to the treatment effect and
asymptotic normality of MLEs holds. Gnedenko and Korolev (1996); Bening et al.
(2012); Christoph et al. (2020); Korolev and Zeifman (2019) assume that the
distribution of the random sample size does not depend on previosly collected
data; Molenberghs et al. (2012) assumes this asymptotically.
Nevertheless, convergence of sample means to non-normal random variables was
shown even for ancillary random sample sizes. In Gnedenko and Korolev (1996),
Gnedenko and Korolev show convergence of standardized sums with random number
of summands of infinitely divisible random variables to mixtures of stationary
distributions. They give conditions for convergence to a mixture of normal
distributions. Bening et al. Bening et al. (2012) and Christoph et al.
Christoph et al. (2020) explore convergence to mixtures of normal
distributions and to Student’s limit distribution. When convergence is _mixed_
(see, for example Häusler and Luschgy (2015)), Lin et al. (2020) shows how
norming with the observed information can result in a normal limit. However,
the requirement for mixed convergence appears strong, and it does not cover
the limiting mixtures obtained in this paper.
The impact of early stopping is pervasive. It affects the probabilistic
characterisation of the tests (e.g., type I error and Fisher Information) as
well as the distributions of MLEs and test statistics. Its effect on Fisher
information, when stopped at different stages, is not widely recognized.
During the design phase, before observations are taken, the full form of the
joint density (5), accounting for all possible events, is appropriate. This
contrasts with the current practice of using the density assuming the
experiment will continue through stage $K$. More details on the differences
resulting from these two design approaches will be the subject of another
paper.
However, normality and asymptotic normality assumptions continue to be
directly used with non-normally distributed statistics. We identify two main
reasons for this.
1. 1.
Many researchers consider large sample properties against a fixed treatment
effect independent of the sample size. From one point of view, a treatment
effect is a population quantity which does not change with sample size. But if
one develops an asymptotic approximation to the testing environment using a
fixed treatment effect, the statistical experiment stops at the first interim
analysis with probability one for any consistent test; test statistics
degenerate to a point mass; see Section 7.4 in Fleming and Harrington (1991).
Under a fixed treatment effect, the power converges to one and cannot be used
to compare different testing procedures. This issue triggered development of
various descriptions of asymptotic relative efficiency. The most popular
approach is Pitman asymptotic relative efficiency Pitman (1948), where
asymptotic power is evaluated under local alternatives; see Nikitin (1995). In
addition, local alternatives clearly reflect actual practice for experiment
planning. Experiments are never planned for a statistical power = 1. Small
sample size studies (pre-clinical, animal studies) are planned to detect large
effect sizes, moderate sample sizes (typical phase 3 studies) are used to
detect moderate effect sizes, and large sample sizes (epidemiological studies,
like vaccine studies) are used to detect small differences.
Koopmeiners et al. (2012) explored MLEs conditional on stopping, but assumed
asymptotic normality to evaluate their uncertainty. Martens and Logan (2018)
relied on asymptotic normality for evaluating regression coefficients under
the Fine–Gray model in GSD settings. Asendorf et al. (2019) evaluated
asymptotic properties with SSR under a fixed alternative for negative binomial
random variables.
2. 2.
Some researchers investigate local asymptotic properties when early stopping
is not possible or is ancillary to the treatment effect. This also leads to
asymptotic normality. Scharfstein et al. (1997) show that without possibility
of early stopping “time-sequential joint distributions of many statistics
$\ldots$ are multivariate normal with an independent increments covariance
structure” under local alternatives. These results are generally consistent
with the classical results on local asymptotic normality of Le Cam (1960),
where both mean and variance of the limiting normal distribution depend on the
parameter of interest; see Chapter 7 of van der Vaart (1998). However, Section
3.1 shows that the possibility of early stopping destroys local asymptotic
normality: the limiting distribution of standardized test statistics is a
mixture of truncated normal distributions. Similar findings were previously
proved for non-ancillary sample size recalculations; see Tarima and Flournoy
(2019a).
Gao et al. (2013) is a rare exception in not making a normality assumption;
these authors mostly deal with set operations and probabilities and, using
stage-wise ordering of events, they calculate P-values, confidence intervals,
and a median unbiased estimate of the parameter of interest.
It is recommended that the full form of the joint density (5), accounting for
all possible events, be used for study design before observations are taken.
This contrasts with the current practice of using the density assuming the
experiment will continue through stage $K$.
## Conflict of interest
The authors declare that they have no conflict of interest.
## References
* Armitage et al. (1969) Armitage, P., C. McPherson, and B. Rowe (1969). Repeated significance tests on accumulating data. Journal of the Royal Statistical Society: Series A 132, 235–244.
* Asendorf et al. (2019) Asendorf, T., R. Henderson, H. Schmidli, and T. Friede (2019). Sample size re-estimation for clinical trials with longitudinal negative binomial counts including time trends: Sample size re-estimation with longitudinal negative binomial counts. Statistics in Medicine 38(9), 1503–1528.
* Bartky (1943) Bartky, W. (1943). Multiple sampling with constant probability. The Annals of Matheamtical Staistics 14, 363–377.
* Bening et al. (2012) Bening, V., N. Galieva, and R. Korolev (2012). On rate of convergence in distribution of asymptotically normal statistics based on samples of random size. In Annales Mathematicae et Informaticae, pp. 17–28.
* Blackwell (1947) Blackwell, D. (1947, 03). Conditional expectation and unbiased sequential estimation. Ann. Math. Statist. 18(1), 105–110.
* Christoph et al. (2020) Christoph, G., M. M. Monakhov, and V. V. Ulyanov (2020). Second-order chebyshev–edgeworth and cornish–fisher expansions for distributions of statistics constructed from samples with random sizes. Journal of Mathematical Sciences, 1–29.
* Crowder (1976) Crowder, M. J. (1976). Maximum likelihood estimation for dependent observations. Journal of the Royal Statistical Society. Series B (Methodological) 38(1), 45–53.
* Demets and Lan (1994) Demets, D. L. and K. K. G. Lan (1994). Interim analysis: The alpha spending function approach. Statistics in Medicine 13(13–14), 1341–1352.
* Dodge and Romig (1929) Dodge, H. and H. Romig (1929). A method of sampling inspection. The Bell System Technical Journal 8, 613–631.
* Efron (1975) Efron, B. (1975, 11). Defining the curvature of a statistical problem (with applications to second order efficiency). Ann. Statist. 3(6), 1189–1242.
* Ferguson (1996) Ferguson, T. (1996). A Course in Large Sample Theory. New York: Routledge.
* Ferguson (2014) Ferguson, T. S. (2014). Mathematical statistics: A decision theoretic approach, Volume 1. Academic press.
* Fleming and Harrington (1991) Fleming, T. R. and D. P. Harrington (1991). Counting Processes and Survival Analysis. John Wiley & Sons.
* Flournoy et al. (2018) Flournoy, N., C. May, and C. Tommasi (2018). The effects of adaptation on inference for non-linear regression models with normal errors. arXiv preprint arXiv:1812.03970.
* Gao et al. (2013) Gao, P., L. Liu, and C. Mehta (2013). Exact inference for adaptive group sequential designs. Statistics in medicine 32(23), 3991–4005.
* Gnedenko and Korolev (1996) Gnedenko, B. V. and V. Y. Korolev (1996). Random summation: limit theorems and applications. CRC press.
* Häusler and Luschgy (2015) Häusler, E. and H. Luschgy (2015). Stable convergence and stable limit theorems, Volume 74. Springer.
* Haybittle (1971) Haybittle, J. (1971). Repeated assessment of results in clinical trials of cancer treatment. The British journal of radiology 44(526), 793–797.
* Ivanova and Flournoy (2001) Ivanova, A. and N. Flournoy (2001). A birth and death urn for ternary outcomes: stochastic processes applied to urn models. Chapman and Hall/CRC, Boca Raton.
* Ivanova et al. (2000) Ivanova, A., W. F. Rosenberger, S. D. Durham, and N. Flournoy (2000). A birth and death urn for randomized clinical trials: Asymptotic methods. Sankhy, Series B 62(1), 104–118.
* Jennison and Turnbull (1999) Jennison, C. and B. Turnbull (1999). Group Sequential Methods with Applications to Clinical Trials. Chapman & Hall/CRC Interdisciplinary Statistics. CRC Press.
* Koopmeiners et al. (2012) Koopmeiners, J. S., Z. Feng, and M. S. Pepe (2012). Conditional estimation after a two-stage diagnostic biomarker study that allows early termination for futility. Statistics in Medicine 31(5), 420–435.
* Korolev and Zeifman (2019) Korolev, V. and A. Zeifman (2019). From asymptotic normality to heavy-tailedness via limit theorems for random sums and statistics with random sample sizes. In Probability, Combinatorics and Control, pp. XX–XX. IntechOpen.
* Kunz et al. (2020) Kunz, C. U., S. Jörgens, F. Bretz, N. Stallard, K. V. Lancker, D. Xi, S. Zohar, C. Gerlinger, and T. Friede (2020). Clinical trials impacted by the covid-19 pandemic: Adaptive designs to the rescue? Statistics in Biopharmaceutical Research 0(0), 1–17.
* Lane and Flournoy (2012) Lane, A. and N. Flournoy (2012). Two-stage adaptive optimal design with fixed first-stage sample size. Journal of Probability and Statistics 2012, n/a–n/a.
* Le Cam (1960) Le Cam, L. M. (1960). Locally asymptotically normal families of distributions. Univ. California Publ. Statist. 3, 37–98.
* Li et al. (2002) Li, G., W. J. Shih, T. Xie, and J. Lu (2002). A sample size adjustment procedure for clinical trials based on conditional power. Biostatistics 3(2), 277–287.
* Lin et al. (2020) Lin, Z., N. Flournoy, and W. F. Rosenberger (2020). Random norming aids analysis of non-linear regression models with sequential informative dose selection. Journal of Statistical Planning and Inference 206, 29–42.
* Liu and Hall (1999) Liu, A. and W. Hall (1999). Unbiased estimation following a group sequential test. Biometrika 86(1), 71–78.
* Liu et al. (2006) Liu, A., W. Hall, K. F. Yu, and C. Wu (2006). Estimation following a group sequential test for distributions in the one-parameter exponential family. Statistica Sinica 16(1), 165–181.
* Martens and Logan (2018) Martens, M. J. and B. R. Logan (2018). A group sequential test for treatment effect based on the Fine–Gray model. Biometrics 74(3), 1006–1013.
* May and Flournoy (2009) May, C. and N. Flournoy (2009). Asymptotics in response-adaptive designs generated by a two-color, randomly reinforced urn. The Annals of Statistics 37(2), 1058–1078.
* Milanzi et al. (2015) Milanzi, E., G. Molenberghs, A. Alonso, M. G. Kenward, A. A. Tsiatis, M. Davidian, and G. Verbeke (2015). Estimation after a group sequential trial. Statistics in Biosciences 7(2), 187–205.
* Molenberghs et al. (2012) Molenberghs, G., M. G. Kenward, M. Aerts, G. Verbeke, A. A. Tsiatis, M. Davidian, and D. Rizopoulos (2012). On random sample size, ignorability, ancillarity, completeness, separability, and degeneracy: sequential trials, random sample sizes, and missing data. Statistical Methods in Medical Research 23, 11–41.
* Neyman and Pearson (1933) Neyman, J. and E. S. Pearson (1933). Ix. on the problem of the most efficient tests of statistical hypotheses. Philosophical Transactions of the Royal Society of London. Series A, Containing Papers of a Mathematical or Physical Character 231(694-706), 289–337.
* Nikitin (1995) Nikitin, Y. (1995). Asymptotic efficiency of nonparametric tests. Cambridge University Press.
* O’Brien and Fleming (1979) O’Brien, P. C. and T. R. Fleming (1979). A multiple testing procedure for clinical trials. Biometrics, 549–556.
* Peto et al. (1976) Peto, R., M. Pike, P. Armitage, N. Breslow, D. Cox, S. V. Howard, N. Mantel, K. McPherson, J. Peto, and P. Smith (1976). Design and analysis of randomized clinical trials requiring prolonged observation of each patient. i. introduction and design. British journal of cancer 34(6), 585–612.
* Philippou et al. (1973) Philippou, A. N., G. Roussas, et al. (1973). Asymptotic distribution of the likelihood function in the independent not identically distributed case. The Annals of Statistics 1(3), 454–471.
* Pitman (1948) Pitman, E. (1948). Lecture Notes on Nonparametric Statistics. New York: Columbia University.
* Pocock (1977) Pocock, S. J. (1977). Group sequential methods in the design and analysis of clinical trials. Biometrika 64(2), 191–199.
* Proschan et al. (2006) Proschan, M. A., K. K. G. Lan, and J. T. Wittes (2006). Statistical Monitoring of Clinical Trials: A Unified Approach. Springer.
* Scharfstein et al. (1997) Scharfstein, D. O., A. A. Tsiatis, and J. M. Robins (1997). Semiparametric efficiency and its implication on the design and analysis of group-sequential studies. Journal of the American Statistical Association 92(440), 1342–1350.
* Schou and Marschner (2013) Schou, I. M. and I. C. Marschner (2013). Meta-analysis of clinical trials with early stopping: an investigation of potential bias. Statistics in Medicine 32(28), 4859–4874.
* Simes (1986) Simes, R. J. (1986). An improved bonferroni procedure for multiple tests of significance. Biometrika 73(3), 751–754.
* Tamhane et al. (2020) Tamhane, A. C., J. Gou, and A. Dmitrienko (2020). Some drawbacks of the simes test in the group sequential setting. Statistics in Biopharmaceutical Research 12(3), 390–393.
* Tarima and Flournoy (2019a) Tarima, S. and N. Flournoy (2019a). Asymptotic properties of maximum likelihood estimators with sample size recalculation. Statistical Papers 60, 23–44.
* Tarima and Flournoy (2019b) Tarima, S. and N. Flournoy (2019b). Distribution theory following blinded and unblinded sample size re-estimation under parametric models. Communications in Statistics - Simulation and Computation 0(0), 1–12.
* van der Vaart (1998) van der Vaart, A. (1998). Asymptotic Statistics. Cambridge University Press.
* Wald (1947) Wald, A. (1947). Sequential Analysis. New York: Wiley.
* Wang et al. (2014) Wang, H., N. Flournoy, and E. Kpamegan (2014). A new bounded log-linear regression model. Metrika 77(5), 695–720.
* Wetherill and Glazebrook (1986) Wetherill, G. and K. Glazebrook (1986). Sequential Methods in Statistics (3 ed.). New York: Chapman and Hall.
|
# Dexterous Manipulation Primitives for the Real Robot Challenge
Claire Chen
&Krishnan Srinivasan 11footnotemark: 1
&Jeffrey Zhang 11footnotemark: 1
&Junwu Zhang 11footnotemark: 1
Stanford University, Palo Alto, CA
###### Abstract
This report describes our approach for Phase 3 of the Real Robot Challenge. To
solve cuboid manipulation tasks of varying difficulty, we decompose each task
into the following primitives: moving the fingers to the cuboid to grasp it,
turning it on the table to minimize orientation error, and re-positioning it
to the goal position. We use model-based trajectory optimization and control
to plan and execute these primitives. These grasping, turning, and re-
positioning primitives are sequenced with a state-machine that determines
which primitive to execute given the current object state and goal. Our method
shows robust performance over multiple runs with randomized initial and goal
positions. With this approach, our team placed second in the challenge, under
the anonymous name “sombertortoise” on the leaderboard. Example runs of our
method solving each of the four levels can be seen in this video.
Advised by: Lin Shao11footnotemark: 1, Shenli Yuan11footnotemark: 1, Preston
Culbertson11footnotemark: 1, Hongkai Dai ***Toyota Research Institute, Los
Altos, CA, Mac Schwager11footnotemark: 1, Jeannette Bohg 11footnotemark: 1
## 1 Real Robot Challenge Overview
Figure 1: The Phase 3 setup consists of the TriFinger robot and a cuboid
object.
The Real Robot Challenge [1] invited teams to design and implement methods for
performing dexterous manipulation tasks using the TriFinger robotic platform
[2]. The TriFinger robot has three fingers, each with three degrees of
freedom. Phases 1 and 2 of the challenge involve manipulating a cube with a
6cm side length, first in simulation and then with the real robot. Phase 3
involves manipulating a 2cm x 2cm x 8cm cuboid object. Each phase is divided
into the following 4 levels of increasing difficulty:
* •
Level 1: Move object to randomly sampled goal position on the table.
Orientation not considered.
* •
Level 2: Move object to a fixed goal position above the table. Orientation not
considered.
* •
Level 3: Move object to randomly sampled goal position. Orientation not
considered.
* •
Level 4: Move object to randomly sampled goal pose where both position and
orientation are considered.
In all phases, we have access to robot proprioception including joint angles
and fingertip forces, as well as an estimate of the object pose obtained from
three cameras. We are also given full information about the shape and mass of
the object, as well as the mass properties of the robot.
## 2 Method
Our approach rests on the idea that in-hand manipulation tasks can be broken
down into manipulation primitives [3]. We decompose the task of moving the
object to a goal pose into three primitives:
* •
grasp: Move fingertips from initial positions to pre-defined contact points on
cuboid. Before executing the re-position and turn primitives, the state
machine must choose the grasp primitive.
* •
re-position: With three fingers on the cuboid, translate and lift to goal
position.
* •
turn: With three fingers on the cuboid, rotate it on the table and return
fingers to initial positions above object.
We use model-based trajectory optimization and control to plan and execute
these primitives, the details of which are described in the following sub-
sections. We design a state machine to sequence the primitives. At the start
of a task, the state machine selects the grasp primitive to bring the fingers
to desired contact points on the cuboid. Once the cuboid has been grasped, the
state machine determines whether to re-position or turn the cuboid. For Levels
1-3, the state machine always chooses the re-position primitive, as
orientation error is not considered in the score. For Level 4, the state
machine chooses to turn the object to align its yaw rotation with that of the
goal pose before choosing the re-position primitive to move it to the goal
position.
To grasp the object, we plan trajectories for each fingertip to reach desired
contact points using trajectory optimization and track these trajectories
using the simplified impedance controller from [2]. To re-position and turn
the object, we first compute a trajectory for the object pose assuming that
contact forces are applied at the pre-defined contact points from the grasping
primitive, and then use this trajectory to compute the corresponding fingertip
trajectories.
In addition to assuming that the fingers remain fixed at the pre-defined
contact points, we also assume that we have access to prior knowledge of robot
kinematics and dynamics, physical parameters of the object, as well as
accurate state feedback on object pose and finger joints. Given these
assumptions, a model-based trajectory optimizer and simplified impedance
controller to track fingertip trajectories are sufficient for executing all of
the primitives.
Figure 2: A diagram of the state machine. A primitive is selected based on the
required level of difficulty and the difference between the object’s current
and goal pose. The state machine begins with the grasp primitive to move the
fingers to contact points on the object. For Levels 1-3, the state machine
always chooses the re-position primitive, as orientation error is not
considered in the score. For Level 4, the state machine chooses to turn the
object to align the yaw rotation before choosing the re-position primitive to
move it to the goal position. The re-position primitive is always assumed to
be successful and is thus the terminal state.
### 2.1 Grasping
To grasp the cuboid securely, we use a three-fingered grasp to pinch the
cuboid across its short axis, as shown in Figure 2. This grasp is used for
both the re-position and turn primitives. We pre-define the contact points
that compose this grasp, with one finger placed on a long face of the cuboid
and the other two fingers placed the opposite long face of the cuboid,
equidistant from the center of the face. For each finger, we choose which face
to contact based on the face that is closest to that finger. In our tests, we
find that this simple face assignment strategy is enough to ensure that all
fingers are able to reach the contact points without collision, since the
cuboid always starts near the center of the arena. Since the object is very
narrow, we perform a two-stage grasping motion by first lowering the fingers
to the height of the object before then moving them towards the object to
pinch it. We use trajectory optimization to first plan a trajectory to lower
the fingers, and then again to pinch the cuboid.
For the cube manipulation task in Phase 2, we included a collision penalty
term in the trajectory optimization problem to prevent the fingertips from
making unwanted contact with the cube. While this worked well for grasping the
cube, we found it much more challenging to tune the collision penalty term for
a slimmer object like the cuboid. Ultimately, the trajectory optimization we
use for the cuboid tasks in Phase 3 does not include collision avoidance, as
the two-stage grasping motion was more reliable for preventing unwanted
collisions. We acknowledge that without the need for collision avoidance, an
operational space controller could have been used instead of trajectory
optimization for the simple task of moving the fingertips to desired goal
positions; however, we already had the trajectory optimization method
implemented.
#### Fingertip position trajectory optimization:
We formulate a direct collocation trajectory optimization problem to compute
collision-free fingertip trajectories to the desired goal positions in
Cartesian space. The optimization problem, shown below, considers only the
kinematics of the fingers and assumes that the pose of the cuboid does not
change. We find trajectories of length $T$ for joint angles $q$ and joint
velocities $\dot{q}$ and use forward kinematics to compute the corresponding
fingertip positions $x$ and velocities.
$\displaystyle\underset{q,\dot{q},\alpha}{\text{minimize}}\;\;\;\sum_{t=0}^{T}(x_{\text{goal}}-x_{t})^{T}Q(x_{\text{goal}}-x_{t})+\sum_{t=0}^{T-1}\dot{q}^{T}R\dot{q}+\sum\alpha$
(1) subject to: $\displaystyle q_{k+1}-q_{k}$
$\displaystyle=\frac{1}{2}(t_{k+1}-t_{k})(\dot{q}_{k+1}+\dot{q}_{k})$
$\displaystyle\;\;\;\forall k\in(0,1,2,...,T-1)$ (2)
$\displaystyle||x_{k}||_{2}$ $\displaystyle<r_{\text{arena}}$
$\displaystyle\;\;\;\forall k\in(0,1,2,...,T-1)$ (3)
$\displaystyle(x_{\text{goal}}-x_{T})^{2}$ $\displaystyle>\alpha$ (4)
$\displaystyle\alpha$ $\displaystyle>0$ (5) $\displaystyle q$
$\displaystyle\in[q_{\text{min}},q_{\text{max}}]$ (6) $\displaystyle\dot{q}$
$\displaystyle\in[\dot{q}_{\text{min}},\dot{q}_{\text{max}}]$ (7)
The cost function includes the distance between the fingertip positions and
goal positions at each time step weighted with weight matrix $Q$, joint
velocities at each time step weighted with weight matrix $R$, and the sum of
all the slack variables $\alpha$. The weight matrices were tuned to produce
reasonable-looking solution trajectories for the fingers. Equation 2 enforces
joint velocity constraints, Equation 3 ensures that the fingertips remain
within the arena radius, and Equations 4 and 5 constrain the final fingertip
positions to be equal to the goal fingertip positions, relaxed with slack
variables $\alpha$. We use IPOPT [4] to solve all trajectory optimization
problems.
### 2.2 Turning and re-positioning
Given an initial and goal pose for the object’s center of mass (CoM), we use
direct collocation to plan a trajectory for the object, as well as the
associated forces that need to be applied at fixed contact points to move the
object along this trajectory. The contact points are assumed to be fixed at
the pre-defined positions specified in the grasp primitive.
#### Object CoM trajectory optimization:
The following optimization problem finds trajectories of length $T$ for object
pose $o$, object velocity $\dot{o}$, and contact forces $\lambda^{cf}$
expressed in the corresponding local contact point reference frames $cf$.
$\displaystyle\underset{o,\dot{o},\lambda^{cf},\alpha}{\text{minimize}}\;\;\;\sum_{t=0}^{T}(o_{\text{goal}}-o_{t})^{T}Q(o_{\text{goal}}-o_{t})+\sum_{t=0}^{T-1}(\lambda^{cf}_{\text{target}}-\lambda^{cf}_{t})^{T}R(\lambda^{cf}_{\text{target}}-\lambda^{cf}_{t})+\sum\alpha$
(8) subject to: $\displaystyle o_{k+1}-o_{k}$
$\displaystyle=\frac{1}{2}(t_{k+1}-t_{k})(\dot{o}_{k+1}+\dot{o}_{k})$
$\displaystyle\;\;\;\forall k\in(0,1,2,...,T-1)$ (9)
$\displaystyle\dot{o}_{k+1}-\dot{o}_{k}$
$\displaystyle=\frac{1}{2}(t_{k+1}-t_{k})(\ddot{o}_{k+1}+\ddot{o}_{k})$
$\displaystyle\;\;\;\forall k\in(0,1,2,...,T-1)$ (10) where:
$\displaystyle\ddot{o}=M_{\text{obj}}^{-1}(G\lambda^{cf}+g_{\text{obj}})$ (11)
$\displaystyle(o_{\text{goal}}-o_{T})^{2}$ $\displaystyle>\alpha$ (12)
$\displaystyle\alpha$ $\displaystyle>0$ (13)
We use Equation 11 to define object dynamics, where $M_{\text{obj}}$ is the
mass matrix of the object, $G$ is the grasp matrix, and $g_{\text{obj}}$ is
the vector of gravitational forces on the object. The cost function includes
the distance between the object positions and goal positions at each time step
weighted with weight matrix $Q$, the difference between the contact forces and
target contact forces $\lambda^{cf}_{\text{target}}$ at each time step
weighted with weight matrix $R$, and the sum of all the slack variables
$\alpha$. The weight matrices were tuned to produce trajectories where the
object would move towards the goal pose gradually, rather than too quickly, as
we found that it was most reliable to move the object slowly. The normal force
components of $\lambda^{cf}_{\text{target}}$ are set to some desired normal
force. We can also constrain the contact forces to lie within linear
approximations of friction cones to prevent contact slippage, but in practice,
we found that applying sufficient normal force was enough to prevent grasp
slippage. Equations 9 \- 11 enforce the object’s dynamic constraints, and
Equations 12 and 13 constrain the final object position to be equal to the
goal object pose $o_{\text{goal}}$, relaxed with slack variables $\alpha$.
Given the trajectories for object pose and contact forces expressed in local
contact frames and given our assumption of fixed contact points, we compute
the corresponding trajectories for fingertip positions $x$ and contact forces
expressed in robot frame $\lambda^{rf}$.
Compared to re-positioning the object, re-orienting the object as required in
Level 4 is considerably less straight-forward. While re-positioning the cuboid
can be achieved with a single grasp, rotating the cuboid could potentially
require re-grasping. To enable this, we introduce additional logic into our
state machine. Given the difficult nature of re-orienting the object in mid-
air, we only attempt to turn the object while it rests on the table. To ensure
that the contact points remain reachable by each finger while rotating the
cuboid, we chose a simple heuristic: we turn the object in 45 degree
increments, resetting and re-grasping the object between each rotation. For
example, a 120 degree rotation would be broken down into two 45 degree
rotations followed by a 30 degree rotation.
### 2.3 Impedance Controller
We compute the joint torques necessary for tracking the desired fingertip
trajectories in Cartesian space using the simplified impedance controller from
[2] with additional gravity compensation for the fingers (time index omitted
for clarity):
$\tau=J^{T}(k_{p}(x_{\text{ref}}-x)+k_{v}(\dot{x}_{\text{ref}}-\dot{x})+\lambda^{rf})+g_{\text{hand}}$
(14)
where $\tau\in\rm I\\!R^{9}$ is the vector of joint torques to be applied to
each finger, $x_{\text{ref}}$ are the desired fingertip positions from the
reference trajectory, $\lambda^{rf}$ is the vector of desired contact forces
to be applied by each finger in robot frame, $J$ is the Jacobian of the 3
fingers, $g_{\text{hand}}$ is the gravity compensation vector, and $k_{p}$ and
$k_{v}$ are hand-tuned controller gains.
While we were able to grasp and move the object by just taking into account
feedback on fingertip positions, we also tried incorporating object pose
feedback to further reduce the steady state error between current and goal
object pose. We follow the method presented in [2], which uses a PD law to
compute the wrench that needs to be applied to the object to track a desired
trajectory. The noisy object orientation estimates made it difficult for us to
obtain stable performance with this additional PD law, so we chose to only
track fingertip trajectories.
### 2.4 Reinforcement Learning
To improve the score of the model-based method described above, we explored
using several deep reinforcement learning algorithms for learning other
primitives, such as pushing and turning the cuboid. We used Proximal Policy
Optimization (PPO) to optimize a neural network policy $\pi_{\phi}$,
parametrized by $\phi$, to push the object from an arbitrary starting position
and orientation on the table to the center of the table with a desired goal
orientation. The composite reward function integrates the objectives of
minimizing position and orientation error, and is computed by the following
equation:
$R(o_{\text{pos}},o_{\text{ori}},g_{\text{pos}},g_{\text{ori}})=r_{\text{pos}}(\|o_{\text{pos}}-g_{\text{pos}}\|)+r_{\text{ori}}(\|o_{\text{ori}}^{\text{yaw}}-g_{\text{ori}}^{\text{yaw}}\|).$
(15)
The position reward function, $r_{\text{pos}}$, is
$r_{\text{pos}}(d)=\begin{cases}1,&\text{if $d<0.05$}\\\
\frac{1}{9(20d)^{2}+1}&\text{$0.05<d<0.075$}\\\ 0&\text{o.w.}\end{cases}$
and similarly, the orientation reward function $r_{\text{ori}}$, is computed
by
$r_{\text{ori}}(d)=\begin{cases}1,&\text{if $d<\frac{\pi}{8}$}\\\
\frac{1}{9(8^{-1}\pi d)^{2}+1}&\text{if $\frac{\pi}{8}<d<\frac{3\pi}{16}$}\\\
0&\text{o.w.}\end{cases}$
The objective from [5] is then optimized to learn a policy:
$L(s,a,\phi_{k},\phi)=\min\bigg{(}\frac{\pi_{\phi}(a|s)}{\pi_{\phi_{k}}(a|s)}A^{\pi_{\phi_{k}}}(s,a),\;\text{clip}\bigg{(}\frac{\pi_{\phi}(a|s)}{\pi_{\phi_{k}}(a|s)},1-\epsilon,1+\epsilon\bigg{)}A^{\pi_{\phi_{k}}}(s,a)\bigg{)},$
(16)
where $s$ and $a$ are an arbitrary state and action, $A^{\pi_{\phi}}(s,a)$ is
the advantage function, and $\phi_{k}$ is the policy parameters at the
previous optimization step $k$.
The resulting policy forms a new _Push-Turn_ primitive and pushes the object
to a position that is a) easier for the grasp primitive to grasp the object
from and b) avoids calling the _turn_ primitive, which requires more time to
complete. While this policy was able to perform the rotation task in
simulation, due to robustness issues when transferred to the real robot, we
opted to use the optimal controller alone in the final submission for phase 3,
as it showed less variance in the Level 4 task score.
## 3 Results and Discussion
| Level 1 | Level 2 | Level 3 | Level 4
---|---|---|---|---
Goal # | 1 | 1 | 1 | 2 | 3 | 1 | 2 | 3
Median score | -3455 | -7889 | -12241 | -8676 | -11775 | -49773 | -26250 | -39230
Table 1: The median scores of our method for different goals and levels. Per
level and goal, we execute three runs each starting from a random initial
position. For Levels 1 and 2, we show results for one randomly sampled goal
position each. For Levels 3 and 4, we show results for three randomly sampled
goal poses.
Our current method is able to turn the cuboid on the table and re-position it
to a goal pose. We evaluate our method on each level. Table 1 shows the median
scores over three runs per goal. The score for an episode is computed by
summing the negated error between current and goal object pose over each time
step of a two minute episode, i.e. scores closer to zero are better. As
expected, the scores depend heavily on the goal pose; goal poses closer to the
center of the arena, where the cuboid starts, result in better scores, as it
takes fewer time steps for the object to reach the goal. For example, goals #1
and #2 in Level 3 are approximately 15cm and 10cm away from the center of the
arena, respectively. Although the median final error to the goal position is
approximately 1cm for both goals, the final scores are quite different. On the
real robot, the initial pose of the cuboid is always near the center of the
arena, but with some randomness due to the robot initialization procedure. As
a consequence, we observe that the varied initial poses of the cuboid can
cause large variations in scores across runs for the same goal pose. This was
particularly true for Level 4 goals, since the initial orientations varied
greatly.
Although our method is able to complete Levels 1, 2, and 3, as well as make
good progress in Level 4, there are several improvements that could be made in
the future. Firstly, our method assumes accurate knowledge of object pose, and
is therefore not robust to the noisy object orientation measurements from the
visual tracker. Secondly, our method still suffered from steady state errors
when re-positioning the cuboid, which could be improved by taking into account
object pose feedback with either another feedback law or a learned policy. For
instance, a hybrid control policy combining RL with impedance control in a
hierarchical or residual policy could be helpful in reducing some of these
errors, and aid in planning a trajectory that requires reorienting the object.
Finally, more sophisticated reasoning could be used to choose grasps and
sequence rotations for re-orienting the object.well
## 4 Acknowledgements
We thank the challenge organizers for providing us with the opportunity to use
the TriFinger robot. We would especially like to thank Felix Widmaier for the
very prompt responses to our many questions. We look forward to using the
TriFinger platform for more exciting projects in the future!
## References
* Rea [2020] Real robot challenge, 2020. URL https://real-robot-challenge.com/en.
* Wüthrich et al. [2020] M. Wüthrich, F. Widmaier, F. Grimminger, J. Akpo, S. Joshi, V. Agrawal, B. Hammoud, M. Khadiv, M. Bogdanovic, V. Berenz, et al. Trifinger: An open-source robot for learning dexterity. _arXiv preprint arXiv:2008.03596_ , 2020.
* Okamura et al. [2000] A. M. Okamura, N. Smaby, and M. R. Cutkosky. An overview of dexterous manipulation. In _Proceedings 2000 ICRA. Millennium Conference. IEEE International Conference on Robotics and Automation. Symposia Proceedings (Cat. No. 00CH37065)_ , volume 1, pages 255–262. IEEE, 2000.
* Wächter and Biegler [2006] A. Wächter and L. T. Biegler. On the implementation of an interior-point filter line-search algorithm for large-scale nonlinear programming. _Mathematical Programming_ , 2006.
* Schulman et al. [2017] J. Schulman, F. Wolski, P. Dhariwal, A. Radford, and O. Klimov. Proximal policy optimization algorithms. _ArXiv_ , abs/1707.06347, 2017.
|
# The FRAM robotic telescope for atmospheric monitoring at the Pierre Auger
Observatory
The Pierre Auger Collaboration et al
###### Abstract
FRAM (F/Photometric Robotic Atmospheric Monitor) is a robotic telescope
operated at the Pierre Auger Observatory in Argentina for the purposes of
atmospheric monitoring using stellar photometry. As a passive system which
does not produce any light that could interfere with the observations of the
fluorescence telescopes of the observatory, it complements the active
monitoring systems that use lasers. We discuss the applications of stellar
photometry for atmospheric monitoring at optical observatories in general and
the particular modes of operation employed by the Auger FRAM. We describe in
detail the technical aspects of FRAM, the hardware and software requirements
for a successful operation of a robotic telescope for such a purpose and their
implementation within the FRAM system.
## 1 Introduction
The Pierre Auger Observatory [1], located near Malargüe, Argentina, is
currently the largest ultra-high energy cosmic ray observatory in the world.
It measures the properties of extensive air showers induced by cosmic rays in
the atmosphere. The observatory uses a hybrid detection technique and consists
in its basic configuration of an array of surface particle detectors and five
stations with optical telescopes overlooking this array from each side. These
telescopes detect the faint fluorescence light emitted in the atmosphere due
to the passage of the secondary particles of the extensive air showers. While
the surface detectors work continuously, the fluorescence telescopes can be
operated only in dark conditions during nights with low Moon illumination
(approximately 15% of the time). The measurement of the longitudinal profile
of the fluorescence light by the fluorescence telescopes provides an estimate
of the energy deposit of the showers applying the known air fluorescence yield
and its dependence on local atmospheric conditions [2, 3] and the
consideration of the so-called invisible energy [4]. The total energy of the
primary cosmic ray is then determined as the sum of the calorimetric energy
and the invisible energy. The longitudinal fluorescence profile also carries
information about the character of the primary particle, and thus about the
mass composition of the primary particle beam, and potentially also about the
high-energy hadronic interactions that took place during the development of
the shower.
In order to fully exploit the fluorescence technique, many aspects of
atmospheric conditions must be continually monitored. The production yield of
the fluorescence light depends on vertical profiles of temperature, humidity,
and pressure. The transmission of this light is affected by scattering by
molecules and aerosols (the contribution of absorption is negligible in the
wavelength range used by the fluorescence telescopes) over distances up to
tens of kilometres to the fluorescence telescopes. While the Rayleigh
scattering by molecules can be determined from the state variables of the
atmosphere, the Mie scattering by aerosols requires frequent dedicated
measurements [5, 6]. Using hourly vertical aerosol profiles, instead of an
average profile, significantly improves the precision of the determination of
both the energy of the primary particle and the depth of shower maximum [7].
The presence of any clouds can also affect the determination of the
aforementioned observables as well as of any other properties inferred from
the shape of the profile (such as those used for particle physics studies), as
the presence of the clouds can significantly distort the apparent profile. A
series of cloud cuts from different instruments is applied to select events
for physics analysis. For these reasons, the operation of fluorescence
telescopes requires a sophisticated atmospheric monitoring system.
At the Pierre Auger Observatory, the primary devices to assess the
transparency of the atmosphere are two laser installations near the center of
the array, called the CLF (Central Laser Facility) and the XLF (eXtreme Laser
Facility) [8], used as reference light sources. The scattered light is then
observed with the fluorescence telescopes themselves to infer the vertical
profiles of aerosol extinction [1]. Additionally, a Raman lidar is operated at
site of the CLF, sharing the laser beam with the CLF [9]. Furthermore, elastic
lidars at the sites of the fluorescence sites primarily provide information on
cloud cover and height [10]; the cloud cover is also monitored using infra-red
cloud cameras [11] and visible-light All-Sky Cameras that detect clouds by
star counting [12]. The laser systems actively produce light which means that
when used in the field of view of the fluorescence telescopes, the data taking
is affected. The FRAM (F/Photometric Robotic Atmospheric Monitor) robotic
telescope has been developed as an additional atmospheric monitoring device
using stars as reference light sources. The first FRAM hardware was installed
at the site of the first fluorescence telescope station (Los Leones) in 2005.
Since then, the hardware, measurement methods, and applications of the system
have gone through several major improvements. The layout of the observatory
with the described devices is illustrated in Fig. 1.
Figure 1: Schematic overview of selected atmospheric monitoring devices
installed at the Pierre Auger Observatory. At each FD site, there is, among
other devices, a lidar station, an infrared camera for cloud cover detection
and a visible-light All-Sky Camera. Two laser facilities (CLF and XLF) are
installed close to the center of the surface detector array. The FRAM
telescope is located at the Los Leones FD site (with a second FRAM planned for
the Cohuieco FD site in the future).
Using stars as reference sources allows for determining the integral optical
depth using a passive system. Indeed, the FRAM complements the laser-based
methods where data are needed with high temporal and spacial resolution
without interfering with the operation of the fluorescence telescopes.
Currently the main application of FRAM at the Auger Observatory is the rapid
monitoring of atmospheric conditions along the apparent path of showers that
may have an anomalous longitudinal profile [13]. Knowing actual atmospheric
conditions for those showers enables us to exclude the possibility that such
anomaly was caused by the presence of any clouds along the shower track
("Shoot-the-Shower") and hadronic interactions models can be tested and
further constrained [14]. The successful development of the Auger FRAM led to
the proposal to include a similar instrument in the design of the Cherenkov
Telescope Array (CTA), the future largest ground-based gamma-ray observatory
in the world. The CTA FRAMs have a modified design tailored to the
characteristics of the CTA observatory where they will monitor the changes in
atmospheric transparency across the whole field of view of the Cherenkov
telescopes simultaneously [15]. Based on data from the wide-field telescopes
installed at the future sites of the Cherenkov Telescope Array, it has been be
shown that wide-field photometry can be used to measure the integral
atmospheric extinction with a precision better than 0.01 optical depths [16].
We plan to implement this method at the FRAM setup at the Pierre Auger
Observatory in the near future.
## 2 Applications of stellar photometry for atmospheric monitoring
The basic idea of using stars as reference sources for atmospheric monitoring
is simple: the difference between the observed brightness of a star and the
predicted value based on the star’s known properties depends on the
transparency of the atmosphere between the detector and the upper edge of the
atmosphere in the given direction. More specifically, for each star observed,
a model brightness is calculated from the catalogue value (in magnitudes)
$m_{\mathrm{cat}}$ as
$m_{\mathrm{model}}=M\left(m_{\mathrm{cat}}+Z+f(C,x,y)+g(A,k,C)\right),$ (2.1)
where $M$ accounts for possible non-linearity of the whole system (including
the photometric method), $Z$ is the calibration constant of the system (so-
called zeropoint), $f$ describes corrections due to the different response of
the system to stars of different colors through some color index $C$ and in
different parts of the field of view, and $g$ is a model of extinction
depending on the airmass $A$, the color of the star and the extinction
constant $k$, which is equal to the extinction measured at the zenith [17].
The airmass $A$ is a dimensionless quantity expressing the integral mass of
atmosphere (or a component thereof) encountered by the light of a star at a
given altitude above the horizon relative to the same quantity at the zenith.
If stars at different values of airmass are observed at once or in a short
time window, the parameters of the instrument and of the atmosphere can be
determined simultaneously as the former do not depend on the airmass while the
latter do and can thus be separated using a fitting procedure, akin to the
well-known Langley method [18]. As some instrumental parameters change only
slowly in time, they may be fitted over a larger set of observations globally.
In reality, both the observed and the predicted brightness are affected by
uncertainties which determine the applicability of the method for different
purposes and the optimal detection setup.
The availability of precise photometric data for stars varies from star to
star as there are many catalogs with varying levels of sky coverage. Moreover,
typically a brightness value measured in a bandpass that does not match the
spectral response of a given setup and thus data from several bandpasses are
required for a reliable prediction of the brightness of the star. In standard
photometric fields, there are many stars in a few small fields of the sky for
which precise data in many wavelength bands exist, but using only these would
severely limit the possibility to do measurements in an arbitrary direction.
Furthermore, these isolated fields are typically unsuitable for any method
that uses the Langley calibration, because the only way to observe them at
various values of airmass is to wait for them to move due to the rotation of
the Earth – however during this time, the conditions may change. For the
purpose described here, it is thus better to use a suitable all-sky catalog,
such as Tycho2 [19]. Despite the precision of Tycho2 for individual stars
being lower than that of dedicated photometric surveys, the homogeneity of its
data over the entire sky is extremely valuable as it prevents the introduction
of any biases when processing measurements of stars from large areas of the
sky simultaneously. Using such a homogeneous catalog allows measurements of
transparency in arbitrary directions, although the low precision for
individual stars requires the observation of a large number of stars at once.
The uncertainty of the catalog brightness in Tycho2 increases quickly above
magnitudes of roughly 10. Since there are relatively few sufficiently bright
stars in the sky, a wide-field setup (on the order of degrees) has to be
preferred, although this property comes along with a relatively small
aperture.
The uncertainty in the measured brightness of the stars depends on the
specific instrumental setup. For a long time, photoelectric photometers were
considered the gold standard in astronomical photometry thanks to their
stability and absolute calibration, but they have been mostly superseded by
CCD cameras for most applications. For a CCD camera, the uncertainty in a
photometric measurement depends on the noise level and stability of the camera
electronics and other hardware effects and then chiefly on the amount of light
registered from the star. That in turn depends on exposure length and the
aperture of the optics; for a given aperture, the maximal field of view is
limited by practical limits of optics and thus there is always a trade-off
between the precision of measurement for individual stars and the number of
stars observed.
Based on these considerations and from practical experience, we have
identified several basic operation modes in which a small robotic optical
telescope can be employed in atmospheric monitoring for an optical observatory
(not necessarily a set of fluorescence or Cherenkov telescopes).
Figure 2: An example of an altitude scan taken by the Auger FRAM. Each
individual point represents a single star and the y-value is equal to the
observed extinction for the star: the difference between its catalog and
measured brightness, corrected for various instrumental and atmospheric
effects. The fit of the extinction as a function of altitude (shown by the
dashed line) allows the determination of the vertical aerosol optical depth.
This particular scan has been originally taken in Mode B, triggered by a
cosmic ray shower, but the same data can be used for precise aerosol
measurements. In this mode, besides the continuous coverage of the field of
view of the fluorescence detector between 1.5 and 31.5 degrees of altitude,
also an image of the arrival direction of the shower is taken, which is also
included in the fit to improve the lever arm. The error bars include the
statistical and systematic uncertainties of both the measurement and the
catalog values.
#### Mode A – Precise aerosol measurement using altitude scans.
Taking a series of images at different altitudes above the horizon and thus at
different values of airmass $A$ allows for the simultaneous determination of
the atmospheric extinction and the instrument calibration constant $Z$. This
is achieved by a fit to the altitude dependence of the difference between
predicted and observed brightness of the stars when assuming horizontal
stratification of aerosols in the atmosphere as described in detail in [16] –
see Fig. 2. This method provides the most precise value of the integral
vertical aerosol optical depth, which is obtained from the shape of the
altitude dependence and thus does not require prior knowledge of the
calibration constant $Z$. In fact it can be used to provide calibration for
measurements in other modes. However in order for the fit to work, the
extinction must follow the assumed altitude dependence, which is violated in
particular when clouds are present. In the presence of some clouds on the sky,
the probability for successful measurement can be improved by using external
data to select a cloud-free path for the scan on the sky; this is implemented
on all the current FRAMs using data from the All-Sky Cameras that are present
both at the Auger Observatory and at the future CTA sites. Potentially it can
be also problematic when aerosols are distributed in a non-stratified manner.
Such instances are clearly visible from the data as deviations from the fitted
shape and can be excluded, so the result is not an invalid value, but no value
at all. Therefore, in order to produce results, the method requires some level
of homogeneity over several kilometers of distance from the telescope towards
the azimuth of the scan, which is the typical distance where most of the
observed extinction happens (usually only stars more than 7 degrees above the
horizon are taken into account in altitude scans). However this is still a
much smaller scale than that of the whole Auger Observatory ($65\times 45$
km2). The method can also be used to investigate small inhomogeneities in the
aerosol distribution by taking scans in different azimuth directions and this
can also be repeated with high temporal resolution in order to quantify the
changes in the aerosol conditions. At the Auger FRAM, the altitude scans are
taken as a series of 30 s exposures taken with the photometric Johnson B
filter (see Fig. 3 for an example) at an azimuth based on the available real-
time cloud data from other instruments and the position of the Moon in the
sky. The Johnson B filter has been chosen for its proximity in wavelength to
the near-UV region in which the Auger fluorescence telescopes operate, as
photometry directly in this region is complicated by the spectral properties
of the used optics (cf. Fig. 8) and the lack of suitable all-sky stellar
catalogs for comparison. The measured extinction can be translated to the
near-UV region using Eq. 2.2.
Figure 3: One of the images taken by the Auger FRAM during a Mode A altitude
scan, in this case covering a well-known region of the Milky Way around the
star $\eta$ Carinae and its surrounding nebula (seen left of the center). A
single 30-second exposure in the Johnson B filter can be used to measure the
brightness of hundreds of stars simultaneously in such a rich area.
#### Mode B – Triggered operation in a large field of view.
For a system such as the fluorescence detector of the Pierre Auger Observatory
with a large field of view ($180^{\circ}\times 30^{\circ}$ for each of the
four main FD telescope stations), it is not feasible to permanently monitor
the whole field of view, but a robotic telescope can be used for timely
determination of conditions in a selected part of the field of view. For
triggering such a dedicated measurement, quasi-online reconstructed data from
the fluorescence telescopes, indicating “interesting” events, can be used,
based on a set of configurable cuts, as the rate of detection of showers by
the fluorescence telescopes is too high to follow-up on all of them. When a
shower is selected for observation by FRAM, the geometric parameters of its
trajectory are passed to the FRAM system, which generates a set of fields to
observe so that the trajectory is well covered, see Fig. 4. The investigated
trajectory then also forms a scan in altitude and thus the method of mode A
can be applied directly to obtain a calibrated measurement. However if the aim
is to identify inhomogeneities in the investigated region (such as in the case
of the Shoot-the-Shower program at Auger), it is sufficient to just search for
deviations from the theoretical altitude dependence of extinction and thus the
method is extremely sensitive to any such disturbances as those are completely
independent of the calibration of the system. On the other hand a horizontally
uniform layer is completely undetectable through stellar photometry, but may
be important for the subject at hand as an extensive air shower for example
may pass through the layer making its light affected by the extinction in the
layer only for a part of its trajectory. In principle, a similar method could
be used for any other purpose where a suitable trigger can be found to define
an area of interest within a larger field of view. If only deviations from the
theoretical altitude dependence are of interest, the area can be of any shape;
if precise aerosol measurements are needed, the area either needs to span a
large range of airmass, or separate altitude scans (mode A) shall be made for
calibration.
Figure 4: The apparent trajectory of an extensive air shower can be described
by several geometric parameters (left) that are then passed to FRAM, where a
series of observation fields (right) is generated in order to cover the
trajectory across the whole FD field of view between 1.5 and 31.5 degrees of
altitude as well as the direction of the arrival of the shower as a
serendipity observation for the case that the cosmic ray was associated with a
transient phenomenon. The observation depicted here is the same that produced
the data shown in Fig. 2.
#### Mode C – Continuous monitoring of a small field of view.
If the area of interest on the sky is sufficiently small so that it can be
contained entirely within a field of view of a robotic telescope (such as is
the case with the field of view of the CTA telescopes), it can be monitored
continuously (with temporal resolution given by the length of exposure and
readout) for changes in transparency. As in the previous case (Mode B), it is
much easier to detect changes in transparency than to measure the absolute
value of it in a single field. However even the latter is readily possible
using the calibration provided by taking altitude scans (Mode A) at regular
intervals, even though the precision of the extinction measured is smaller
than that of the value obtained from a full scan, as it combines the
uncertainty of the measurement in the field and that of the calibration
determined from the scans. This is the operation mode foreseen for the CTA
FRAMs during their future operation. The field of view itself will be cut into
smaller areas of comparable numbers of stars using adaptive Voronoi
tessellation in order to provide a 2D view of possible changes in the
transparency across the field of view; when large changes are detected, the
observation of the Cherenkov telescopes will be interrupted and the vertical
profile will be assessed with a lidar [20].
#### Mode D – Measurement of wavelength dependence of aerosol extinction.
The aerosol extinction is usually assumed to be inversely proportional to a
power of the wavelength with an Ångström exponent $\alpha$ between 0 and 2,
which varies due to changes in the physical composition of the aerosols, so
that for the optical lengths $\tau_{1}$ and $\tau_{2}$ measured at wavelengths
$\lambda_{1}$ and $\lambda_{2}$, we find
$\frac{\tau_{1}}{\tau_{2}}=\left(\frac{\lambda_{1}}{\lambda_{2}}\right)^{-\alpha}.$
(2.2)
This can be determined in principle by taking altitude scans (Mode A) in
several different wavebands (typically defined by photometric filters, such as
the Johnson BVRI system used by both Auger and CTA FRAMs). Very precise values
are needed to obtain a reasonable precision in $\alpha$ when the aerosol
content of the atmosphere is small, as the absolute error of $\alpha$ is
proportional to the relative errors of the individual measurements. For
example when measuring at two mean wavelengths $\lambda_{1}$, $\lambda_{2}$,
we receive
$\Delta\alpha=\frac{\ln(1+\delta\tau_{1})+\ln(1+\delta\tau_{2})}{\ln\lambda_{1}-\ln\lambda_{2}}\approx\frac{\delta\tau_{1}+\delta\tau_{2}}{\ln\lambda_{1}-\ln\lambda_{2}}.$
(2.3)
Note that the effective wavelength of the measured extinction depends on the
spectrum of the individual stars and the Ångström exponent itself; the former
effect is accounted for within Eq. (2.1), where the dependence of $k$ on the
color of the star is explicitly modeled, the latter can be dealt with in an
iterative manner – first processing the data while assuming e.g. $\alpha=1$ in
Eq. (2.1), then using Eq. (2.2) to improve the estimate of $\alpha$ and
processing the data again.
For other modes of operation, the bandpass defined by the Johnson B filter
(centered around 445 nm [21]) can be used, in which the molecular contribution
to the extinction is dominated by Rayleigh scattering which is easily
calculated from the overall column density of the atmosphere as provided by
e.g. a global model, such as the GDAS [22]. For other Johnson filters
(typically V, centered around 551 nm and R, centered around 658 nm), the
contribution of molecular absorption mainly by water and ozone is both
important and highly variable due to meteorological effects and thus must be
carefully considered when subtracting the molecular component from the
measured extinction in order to extract the aerosol optical depth. The Tycho2
catalog only has data in two filters, which are close to Johnson B and V which
provide only a limited lever arm for the determination of $\alpha$ and thus
other catalogs must be considered. For atmospheric monitoring at Auger and
CTA, the effect of changing $\alpha$ is small as their primary calibration
methods (lasers) operate at wavelengths close to those of the majority of the
observed light – for example the contribution of the unknown wavelength
dependence of the aerosol scattering to the uncertainty in shower energy is
estimated to be only 0.5% [4]. Nevertheless the knowledge of physical
properties of aerosols may be of general interest [23, 24, 25]. For the Auger
FRAM, a measurement in this mode consists of a scan in altitude, during which
a 30-second image is taken in each of the B, V and R filters for every field
of the scan.
Figure 5: A simplified schema of the major components of the FRAM setup at
the Pierre Auger Observatory. Lines between components indicate either
mechanical binding (dotted) or electronic communication (solid).
## 3 Hardware
### 3.1 General requirements
Even though the Auger FRAM went through several major hardware changes, it has
generally consisted of the same basic elements: a weatherproof enclosure, a
German equatorial astronomical mount, two light detection systems (carried
jointly by the same mount) – a small (20–30 cm) telescope first equipped with
a photometer and later with a CCD camera (a “narrow-field”, NF, system) and a
photographic lens with a CCD camera (a “wide-field", WF, system) – and a set
of control and auxiliary devices. For a schematic overview of the system, see
Fig. 5.
The WF system allows both the measurement of a large number of stars for the
purposes of precision aerosol measurement and the quick coverage of a large
uninterrupted band of the sky for the purposes of the detection of clouds in
the Shoot-the-Shower program and it is thus the primary tool used for
atmospheric monitoring. The NF system currently provides the opportunity for
additional astronomical observations (such as astrometry and/or photometry of
asteroids, comets and variable stars, and gamma-ray burst follow-up [26])
within the spare time allowed by the atmospheric monitoring requirements, but
it is also being developed for use in the atmospheric monitoring program. Such
a possibility is particularly promising after the latest upgrade where the
apparent field of view of the NF system has been considerably expanded,
increasing the number of stars that can be measured at once.
One of the benefits of using stellar photometry for atmospheric monitoring is
the possibility to assemble a highly capable device from affordable off-the-
shelf products. The unique operating conditions of FRAM impose specific
requirements on those devices for several reasons: (i) The robotic nature of
the telescope means that the observations are carried out continuously during
every night, without breaks unless the weather conditions do not allow
observations. Moreover, the exposures are typically short and in many
observation modes, the target area on the sky is changed between each pair of
exposures, resulting in many movements of the camera shutters and the mount.
(ii) The remote locations where astroparticle experiments (such as the Auger
Observatory) are located require that the instruments are as autonomous as
possible and most issues are solved remotely without local intervention.
Regular on-site maintenance also should be kept to a minimum (typically once
per year in the case of the Auger FRAM). (iii) The environmental conditions
at the sites are demanding, with considerable annual and diurnal changes in
temperature, torrential rains, snow, periods of high humidity, but also of
blowing dust. Additionally to the technical considerations, further
requirements stem from the needs of the analysis of the atmospheric monitoring
data and evolve with the progress of this analysis.
The long experience with operating FRAM allows us to identify key issues, take
appropriate steps to mitigate them, and select suitable products during
hardware upgrades. We now describe these considerations for each part of the
FRAM setup.
Figure 6: The FRAM enclosure with one half open and one closed. Inside is an
older version of the FRAM setup with the Meade SCT and Paramount ME mount.
### 3.2 Enclosure
The enclosure protects the FRAM setup from the environment when it is not
operating. As FRAM is expected to quickly react to triggers and to observe
various parts of the sky in short sequence, a solution with a roof that fully
opens to give access to the full sky at once is preferable to a more
traditional rotating dome with a slit opening. In the FRAM case, the pyramidal
roof consists of two independently operated halves (Fig. 6), each moved by a
pair of hydraulic cylinders. Mounted in the corners of the dome, the cylinders
automatically balance the load (they are connected in parallel), and can open
each half to almost 180°. The hydraulic system is designed for operation with
considerable wind-forces acting on the dome halves in transition. The
mechanical linkages are also designed with sufficient safety margin as they
must hold the dome closed against the most extreme wind forces that might be
experienced at the site of the Auger Observatory. The hydraulic pressure is
provided by a single pump driven by a single-phase AC motor. The whole system
is now controlled by a Schneider Zelio smart-relay (PLC) module [27] using
custom firmware. Since a custom-made board has proven to be not reliable
enough, the industry-standard PLC solution has been implemented.
The PLC has outputs that control the position of the hydraulic valves, the
power to the pump motor, the source of this power (mains or backup) and the
on/off state of the backup power inverter. Additionally, it controls the power
to the mount, cameras and other devices to provide a reliable way of
restarting them and it also provides a special output for the on/off control
of the 10micron mount (see Sec. 3.3) For each half of the roof, mechanical
end-switches are connected to the PLC inputs to indicate the position of a
fully opened and fully closed roof. Further inputs provide the voltage of the
(normally constantly charged) battery used to provide the 12 V DC voltage for
the valves and the status of the mains AC power.
The PLC communicates with the control PC via Modbus over TCP/IP. To achieve a
fail-safe design to open the roof, the PC transmits a heartbeat signal
(periodically changes a register value in the PLC). To close the roof, the PC
simply stops the heartbeat. In case the communication with the PC is lost, the
roof closes automatically, as this is simply equivalent to a command to close.
On remote sites such as the Auger Observatory, a mains power cut is not
uncommon. As the Auger Observatory does not provide global backup power to all
its devices, and the Auger FRAM is fully dependent on this external power
(unlike for example the solar-powered FRAMs at the southern CTA site), it has
to be able to safely shut down in case of a power cut. Significant effort has
been devoted to ensure the ability of the roof to close in such a situation.
The pump motor creates a significant power spike on startup, which requires
the use of a powerful inverter providing backup power from four lead-acid
batteries. The inverter cannot be left powered during normal operation,
because it causes electromagnetic interference in the CCD cameras, but it also
must be given at least 30 seconds to stabilize after power on before the pump
draws power from it – this is ensured through the logic in the PLC. During a
power outage, the PLC runs from a UPS which also backs up the control PC, but
the internal network is disconnected on purpose (the network switch is not
backed up by the UPS) so that commands from the PC cannot interfere with the
emergency closing of the roof.
The hydraulic roof system is generally reliable, but regular maintenance (on
the scale of once per year) is needed to check the status of the hydraulic
connections and the oil level. The pump and all the hydraulic hoses had to be
replaced after roughly 10 years of operation and a similar frequency of
maintenance is needed in the future. The emergency closing system is a
critical safety measure with several single points of failure and thus its
integrity should be checked after any changes in the FRAM setup.
### 3.3 Mount
The mount points the cameras to the target area in the sky and follows the
movement of the target due to the rotation of the Earth. For a remote robotic
observatory like FRAM, the mount must be able to resume observation after a
power cut without human intervention, which requires at least that it is able
to find a “home” position autonomously – the lack of this ability has proven
the original Losmandy G11 mount [28] of FRAM as unsuitable.
The replacement – Paramount ME [29] – had a simple optical sensor in each axis
allowing it to return to its home position. All movements are evaluated
relative to the home position from the movement of the stepper motors in each
axis. After a power cut or mount restart, it must be ensured that the homing
procedure is executed first before any attempt to reach a position on the sky,
because commanding a move from a “desynchronized” state may cause movements of
the mount beyond its physical boundaries. This does not harm the hardware, but
requires human intervention, as reaching the physical stops prevents further
autonomous movement of the mount in any direction. This is an intrinsic
feature of the stepper motors used in Paramount ME, which need freedom of at
least one step in both directions to properly initialize. A software daemon is
installed at the control PC for ensuring the homing procedure. The Paramount
ME required regular maintenance even beyond that prescribed by the vendor, due
to the dusty environment of the Auger Observatory and the constant load of the
robotic operation. Cleaning and greasing of the worm gear mechanisms had to be
carried out at least once per year and the worms on both axes had to be
replaced after five years of use. When the worms were in bad condition, the
mount had a tendency to slip during movement and slew slightly off target –
while this could be corrected using the astrometry of the star images, it
occasionally led to the aforementioned problem of desynchronization and
reaching the physical stops, requiring on-site intervention.
The current equatorial mount – 10micron GM2000 HPS [30] – installed in
September 2018, has absolute position sensors and its position is thus
precisely known at any moment. Moreover, the manufacturer claims at least 10
years of maintenance-free operation. The GM2000 contains its own control
computer, which needs to be booted and shut down properly. Normally this is
done using a switch – for remote operation, the switch is driven by one of the
outputs of the roof PLC. The switch is stateless and there is no output from
the mount to easily allow the PLC to determine whether it is on or off. Thus
the control PC checks the state of the mount (over TCP/IP) continually and
stores this information in a register of the PLC so that the PLC can
autonomously handle the safe shutdown of the mount in case of a power cut.
Since commissioning, the GM2000 has operated reliably and without maintenance.
The internal communication between parts of the mount seems to be quite
sensitive to electromagnetic disturbance and requires careful grounding.
### 3.4 Light sensors
We have tested a photoelectric photometer as the main light detector, but we
have not found the technology practical for the purposes of atmospheric
monitoring: the measurement of a large number of stars is slow and even for
individual stars, the reproducibility of the measurement was poor in the
context of a remote robotic telescope due to issues with centering the star
image on the aperture of the photometer automatically and various other issues
that do not occur when an observer is physically present on site. On the other
hand, CCD photometry easily allows simultaneous measurements of a large number
of stars, even though it brings its own issues, especially regarding the
stability of absolute calibration due to a multitude of effects including
temperature dependences of the behavior of the camera electronics and focusing
of the optics. With the combination of a careful data analysis and a well
designed operational mode, the CCD cameras still give much better results than
when using the photometer. CCD cameras are thus currently used in both Auger
and CTA FRAMs. In the near future, we plan to test the performance in this
application of CMOS detectors which, although superficially similar (as a
large matrix of micron-sized pixels), are technologically very different from
CCD cameras. The possibility to use SiPMs as detectors has been also
considered but never explored further.
The Finger Lakes Instrumentation (FLI) [31] cameras used in the first
iterations of FRAM were not suitable for such a heavy use – their three-bladed
iris shutters suffered from the repeated short exposures and became quickly
unusable despite maintenance. Since 2012, all cameras used at FRAM are
supplied by Moravian Instruments (MI) [32] and use a rotating shutter made of
a single butterfly-shaped piece of metal which does not show any degradation
in function even after years of use. Since one of the cameras failed without
an apparent reason, the vendor now provides a special version of the cameras
with protective coating on all electronics and the problem has not appeared
again. However, care must be taken to regularly (at least yearly) replace the
desiccant in the cold chamber as one of the CTA FRAM cameras was considerably
damaged most likely due to corrosion caused by internal humidity.
The most problematic part of the cameras is the filter wheel, which typically
holds at least BVR Johnson photometric filters. For most applications, it
would be sufficient to include a stationary B filter, but for the measurements
of the Ångström exponent and for various astronomical applications, the
ability to change filters is highly desirable. The MI filter wheel contains a
large circular filter holder with a rubber band at the edge which is rotated
by a small driving wheel; the positions of the filters are marked by holes in
the holder that are detected by an optical gate, with the first position
indicated by two holes as a “home” location. This setup has the tendency to
become worn out and slip, causing a de-synchronization which then causes the
applied filters to be shifted by one position from the requested ones. The
vendor has delivered several improvements to the rubber band as well as
detailed procedures for the setting of the pressure between the driving wheel
and the rubber band, but the problem still occasionally appears. We have
developed a software routine that checks for possible problems and restarts
the filter wheel if needed. Most of the filter changes occur during the
Ångström measurements (Mode D), which is always done as a series of scans with
different filters – the procedure is such that in case of a filter wheel
malfunction, it usually still obtains a series of consistent scans, just in a
different set of filters and a large fraction of affected data can be salvaged
automatically because from a full scan, the filter used can be relatively
easily inferred.
Figure 7: The current FRAM setup with the narrow-field system on the top and
the wide-field system on the bottom.
### 3.5 Wide-field system
Originally using a Pentacon 200/2.8 lens with a small FLI camera and then a MI
G2-1600, the wide-field system was significantly upgraded in 2013 to the
current configuration of a Nikkor 300/2.8 lens and a large-format 36$\times$36
mm2 MI G4-16000 camera (Fig. 7). Even though the focal length of the Nikkor is
longer, the large format camera covers an area of $7^{\circ}\times 7^{\circ}$,
significantly larger than before the upgrade. The optical quality of the
Nikkor lens degrades towards the corners of the image, as expected because the
CCD chip is even larger than a traditional film frame for which the lens has
been originally designed. The correction for this effect in Eq. (2.1) is
acceptable in the whole field for the purposes of cloud detection in the
Shoot-the-Shower program, but for precision aerosol measurements, only the
inner circular part with a diameter of roughly 6.3 degrees, where this
correction is less dependent on the momentary state of focus, is used; this
still amounts to more than 60% of the area of the CCD chip.
The fast f/2.8 lens has a narrow depth of focus. Moreover, it loses its
focusing due to temperature changes and slippage due to the movement of the
mount. Also due to residual color aberrations, the optimal focus position is
slightly different for different filters, even though the filters themselves
are confocal. Keeping proper focus is important in particular for operation in
Mode C (monitoring of a single field) because changes in focusing have a
strong effect on the calibration constant for stellar photometry as it affects
the distribution of light around the center of the stellar image. This mode is
not commonly employed at the Auger Observatory, but even for cloud detection
or self-calibrated scans, it is necessary to have some means of controlling
the focus as otherwise the lens eventually de-focuses to a useless state. The
lens has only a manual focusing ring; to be able to focus it remotely, we thus
use a focusing kit by Rigel Systems [33] which consists of a plastic cogwheel
attached to the focus ring of the lens which is driven by a USB-controlled
stepper motor.
Ideally, the lens should be aligned so that its optical axis intersects the
CCD chip perpendicularly in the center of the chip as then the correction on
the position of the star on the chip would be strictly radial, reducing the
number of free parameters. The alignment can be checked easily by measuring
the star shapes across the imaged field. A custom-made holding system for the
lens and camera was designed and then gradually improved. Eventually such a
task proves near impossible to perform flawlessly as there are many combined
effects causing optical misalignment and not all of them are due to the mutual
alignment of the camera and the lens. However it turns out that if a rough
alignment is achieved, the remaining variations in star shape across the field
of view can be satisfactorily treated during data processing, even if they are
not radial, without introducing a bias in the measured atmospheric properties.
Thus, further improvement is not required.
Figure 8: Spectral characteristics of the key components of the wide-field
system – transmissivity of the lens and the filters and quantum efficiency of
the CCD – as measured in an optical laboratory.
Fig. 8 shows the spectral characteristics of the key components of the wide-
field system.
### 3.6 Narrow-field system
The narrow-field system is a standard small astronomical telescope. The
original 20-centimeter Cassegrain telescope had significant problems with
mechanical stability and has been replaced by a 30cm Meade [34] Schmidt-
Cassegrain (on loan from Instituto de Astrofisica Andalusia) and later by a
30-centimeter Orion UK [35] Dall-Kirkham (Fig. 7). The change to the Dall-
Kirkham system was accompanied by a change from a small camera with a
14$\times$10 mm2 chip to another G4-16000 with a 36$\times$36 mm2 chip, which
the Dall-Kirkham fully covers, albeit with significant vignetting; the setup
achieves a full square degree of field of view. To avoid contributing to the
vignetting, the system is equipped with a 3-inch focuser from Astro Systeme
Austria [36].
## 4 Control, software and operation
All components of the FRAM system are controlled from a single PC running
Linux. Apart from reliable basic components, a key consideration is the
exclusive use of SSDs as spinning disks universally succumb to the
environmental conditions on site. The requirements for local storage are
substantial, as the Pierre Auger Observatory (as is typical for astroparticle
experiments) is located in a remote area and transferring large amounts of
low-priority data, such as the raw images from the CCD cameras, is not always
easy. The PC is equipped with an IPMI interface allowing remote management and
power control using a dedicated second ethernet interface. The most
problematic aspect of the PC control lies in the USB communication with the
various devices. Even though the problems could be solved by applying custom-
made kernel patches, recently problems in USB communication appeared again,
presumably because of the specific situation of a telescope setup, where many
USB devices are mounted together in such a manner that their enclosures are
conductively connected, possibly causing multiple ground loops. One possible
solution, which is currently under investigation, could be the installation of
optical isolators on the USB cables to prevent the ground loops and other
sources of interference.
The whole FRAM operation is conducted within the RTS2 software framework, an
open-source package for robotic observatories [37]. This modular framework was
originally developed primarily for follow-ups of gamma-ray bursts (GRBs), but
over time it has been extended for many purposes and is deployed on many
observatories around the world. The Auger FRAM has been one of the test-beds
for its development for years and some features of the package are tailored to
the needs of the project [38]. RTS2 can run an astronomical telescope with all
its accessories (including the enclosure) autonomously. It is based on the
elementary concept of a “target”, which can be an astronomical object
described by its coordinates, or a more complicated recipe for observation
defined via a Python interface using a specific API. During the night, the
targets can be chosen from a database based on priorities assigned to them or
the telescope set to follow nightly queues of planned observations in a given
order. Thanks to the GRB background, RTS2 also naturally supports reactive
observations triggered by external sources of information. For the Shoot-the-
Shower program, a dedicated module has been developed to receive and process
near real-time data from the Central Data Acquisition System of the Auger
Observatory and take a decision to follow up an observation of a cosmic ray
shower based on configurable sets of parameters.
Despite the autonomous capabilities of RTS2, the best results for FRAM in
terms of uptime and data quality are still achieved with human supervision.
Since 2012, we additionally rely on a dedicated observer, who checks the
status of FRAM (and later also of the other FRAMs at CTA) every day before and
sometimes during observations and resolves various issues or, if necessary,
calls for on-site intervention. Over time, we have been able to identify the
most common issues (such as race conditions among devices leading to data
corruption, improper metadata recording, incorrect handling of the meridian
flip of the mount etc.) and have either resolved them with improvements in
hardware or by software workarounds, thus slowly eliminating the need for
nightly human oversight. It is still not clear how to judge the quality of
data taken automatically due to the variability of atmospheric conditions (how
to determine whether any problems with data are due to unfavourable conditions
or system issues) and thus regular checking of the outputs by an experienced
observer is still desirable.
Apart from a multitude of bug fixes, the most common among the software
solutions to the operational issues is the use of watchdogs over the drivers
for the necessary devices. Each of the drivers runs as a separate process and
in case of a crash, it is restarted, sometimes including an automatic power-
cycle of the associated device. An important improvement is the development of
a routine for automatic focusing, that can reliably decide if it is actually
detecting stars and if the results are meaningful. It is implemented by
analyzing a set of stars that are automatically detected in a sequence of
images acquired with different focus positions. The images are cross-matched
in order to reject spurious detections, and then a common minimum is found in
the measured sequence of stellar FWHMs. If not enough star sequences are
detected or if no clear minima are seen in their FWHMs, the routine properly
identifies the focusing failure and does not update an optimal focus
estimation, thus avoiding artificial focus drifts in bad or unstable weather
conditions.
## 5 Current FRAM setup and observation modes
To summarize, the current FRAM setup at the Pierre Auger Observatory consists
of a wide-field imager (MI G4-16000 CCD on a Nikkor 300/2.8 with a stepper
motor external focuser) and a narrow-field imager (MI G4-16000 CCD on a
30-centimeter Orion UK Dall-Kirkham with 3” ASA focuser) jointly carried by a
10micron GM2000 HPS German equatorial mount, housed in a custom enclosure with
a hydraulically opened roof controlled by a PLC and controlled from a PC
running the RTS2 software with local modifications. The operation is mostly
automated, but the status is checked almost nightly by a dedicated observer.
The basis of the usual observation program is taking regular altitude scans
(Mode A) taken as a series of 30 second exposures in the B filter in azimuths
selected taking into account the data from the All-sky Camera and the position
of the Moon (a too close passage to the Moon is avoided as such data are
difficult to process properly); less frequently, scans are conducted in B, V
and R filters for the Ångström coefficient measurements (Mode D). The scans
are implemented using the Python API interface and a server-client
configuration which allows multiple cameras (two in our case, NF and WF) to
take data simultaneously and also allows clean resumption of scans in case of
an interruption. In between these scans, some selected astronomical targets
can be observed. The Shoot-the-Shower module is continually listening for
triggers from the Central Data Acquisition System of the Auger Observatory and
in case of an extensive air shower passing a set of predefined cuts, any
ongoing observation is immediately stopped and the observation along the
apparent path of the shower is commenced (Mode B). Regular runs of the auto-
focusing routine are scheduled through the night to keep the focus consistent
as the nightly decrease of outside temperatures causes contraction of the
optical paths of the instruments.
## 6 Conclusions
Stellar photometry can be used for atmospheric monitoring in various modes,
several of which are employed by the Auger FRAM. The Shoot-the-Shower program
is the most valuable for further applications in physics analyses using the
data of the Pierre Auger Observatory. Currently, the FRAM data on triggered
air showers are being integrated into an analysis of cloud-free anomalous air
shower events for studies of aspects of the mass composition of the primary
beam and particle physics. The highest quality dataset comes from the latest
WF setup installed in 2013, thus comprising more than 7 years so far. The
expected number of truly anomalous events for the limited field of view of the
fluorescence telescopes is still very low (at the order of 1 ‰ of all detected
events, according to the rough estimates of [14]) and thus another FRAM
installation is planned at the Cohuieco FD station. At that station, the low-
energy extension HEAT [39] is also located, further increasing the number of
detected showers. This second Auger FRAM will consist of two WF setups with
identical optics, but different light sensors – alongside the traditional CCD
camera, a CMOS detector will be tested to assess its viability for atmospheric
monitoring. The methods to process the aerosol data from Mode A scans and
Ångström coefficients from Mode D observations from both the existing FRAM and
the future one are being finalized. All images taken by the Auger FRAM are
calibrated and stored in a database. As part of the Open Data project at FZU -
Institute of Physics of the Czech Academy of Sciences, all images were made
publicly available through a web interface111https://pc048b.fzu.cz/archive/,
where they can be searched using various criteria for a range of astronomical
applications, such as pre-discovery data on newly discovered Solar System
objects or the study of specific variable stars.
## Acknowledgments
The successful installation, commissioning, and operation of the Pierre Auger
Observatory would not have been possible without the strong commitment and
effort from the technical and administrative staff in Malargüe. We are very
grateful to the following agencies and organizations for financial support:
Argentina – Comisión Nacional de Energía Atómica; Agencia Nacional de
Promoción Científica y Tecnológica (ANPCyT); Consejo Nacional de
Investigaciones Científicas y Técnicas (CONICET); Gobierno de la Provincia de
Mendoza; Municipalidad de Malargüe; NDM Holdings and Valle Las Leñas; in
gratitude for their continuing cooperation over land access; Australia – the
Australian Research Council; Brazil – Conselho Nacional de Desenvolvimento
Científico e Tecnológico (CNPq); Financiadora de Estudos e Projetos (FINEP);
Fundação de Amparo à Pesquisa do Estado de Rio de Janeiro (FAPERJ); São Paulo
Research Foundation (FAPESP) Grants No. 2019/10151-2, No. 2010/07359-6 and No.
1999/05404-3; Ministério da Ciência, Tecnologia, Inovações e Comunicações
(MCTIC); Czech Republic – Grant No. MSMT CR LTT18004, LM2015038, LM2018102,
CZ.02.1.01/0.0/0.0/16_013/0001402, CZ.02.1.01/0.0/0.0/18_046/0016010 and
CZ.02.1.01/0.0/0.0/17_049/0008422; France – Centre de Calcul IN2P3/CNRS;
Centre National de la Recherche Scientifique (CNRS); Conseil Régional Ile-de-
France; Département Physique Nucléaire et Corpusculaire (PNC-IN2P3/CNRS);
Département Sciences de l’Univers (SDU-INSU/CNRS); Institut Lagrange de Paris
(ILP) Grant No. LABEX ANR-10-LABX-63 within the Investissements d’Avenir
Programme Grant No. ANR-11-IDEX-0004-02; Germany – Bundesministerium für
Bildung und Forschung (BMBF); Deutsche Forschungsgemeinschaft (DFG);
Finanzministerium Baden-Württemberg; Helmholtz Alliance for Astroparticle
Physics (HAP); Helmholtz-Gemeinschaft Deutscher Forschungszentren (HGF);
Ministerium für Innovation, Wissenschaft und Forschung des Landes Nordrhein-
Westfalen; Ministerium für Wissenschaft, Forschung und Kunst des Landes Baden-
Württemberg; Italy – Istituto Nazionale di Fisica Nucleare (INFN); Istituto
Nazionale di Astrofisica (INAF); Ministero dell’Istruzione, dell’Universitá e
della Ricerca (MIUR); CETEMPS Center of Excellence; Ministero degli Affari
Esteri (MAE); México – Consejo Nacional de Ciencia y Tecnología (CONACYT) No.
167733; Universidad Nacional Autónoma de México (UNAM); PAPIIT DGAPA-UNAM; The
Netherlands – Ministry of Education, Culture and Science; Netherlands
Organisation for Scientific Research (NWO); Dutch national e-infrastructure
with the support of SURF Cooperative; Poland -Ministry of Science and Higher
Education, grant No. DIR/WK/2018/11; National Science Centre, Grants No.
2013/08/M/ST9/00322, No. 2016/23/B/ST9/01635 and No. HARMONIA
5–2013/10/M/ST9/00062, UMO-2016/22/M/ST9/00198; Portugal – Portuguese national
funds and FEDER funds within Programa Operacional Factores de Competitividade
through Fundação para a Ciência e a Tecnologia (COMPETE); Romania – Romanian
Ministry of Education and Research, the Program Nucleu within MCI
(PN19150201/16N/2019 and PN19060102) and project PN-
III-P1-1.2-PCCDI-2017-0839/19PCCDI/2018 within PNCDI III; Slovenia – Slovenian
Research Agency, grants P1-0031, P1-0385, I0-0033, N1-0111; Spain – Ministerio
de Economía, Industria y Competitividad (FPA2017-85114-P and
PID2019-104676GB-C32, Xunta de Galicia (ED431C 2017/07), Junta de Andalucía
(SOMM17/6104/UGR, P18-FR-4314) Feder Funds, RENATA Red Nacional Temática de
Astropartículas (FPA2015-68783-REDT) and María de Maeztu Unit of Excellence
(MDM-2016-0692); USA – Department of Energy, Contracts No. DE-AC02-07CH11359,
No. DE-FR02-04ER41300, No. DE-FG02-99ER41107 and No. DE-SC0011689; National
Science Foundation, Grant No. 0450696; The Grainger Foundation; Marie Curie-
IRSES/EPLANET; European Particle Physics Latin American Network; and UNESCO.
## References
* [1] Pierre Auger collaboration, _The Pierre Auger Cosmic Ray Observatory_ , _Nucl. Instrum. Meth. A_ 798 (2015) 172 [1502.01323].
* [2] F. Arquerosa, J. R. Hörandel, B. Keilhauer, _Nucl. Instrum. Meth. A_ 597 (2008) 1-22 [0807.3760].
* [3] M. Ave et al., _Astroparticle Physics_ 28 (2007) 41 [astro-ph/0703132].
* [4] V. Verzi for the Pierre Auger Collaboration, _The Energy Scale of the Pierre Auger Observatory_ , 33rd International Cosmic Ray Conference (ICRC2013), _Proc. ICRC2013_ 1, 376 (2013) [1307.5059].
* [5] B. Keilhauer for the Pierre Auger Collaboration and M. Will, _Atmospheric monitoring and model applications at the Pierre Auger Observatory_ , Atmospheric Monitoring for High Energy AstroParticle Detectors (AtmoHEAD) 2014, _European Physical Journal Web of Conferences_ 89 (2015) 02001.
* [6] Pierre Auger collaboration, _A study of the effect of molecular and aerosol conditions in the atmosphere on air fluorescence measurements at the Pierre Auger Observatory_ , _Astroparticle Physics_ 33 (2010) 108 [1002.0366].
* [7] V. M. Harvey for the Pierre Auger Collaboration, _Real-Time Measurements with Atmospheric Instruments at the Pierre Auger Observatory_ , in _36th International Cosmic Ray Conference (ICRC2019)_ , PoS(ICRC2019)283 [1909.09073].
* [8] Pierre Auger collaboration, _Techniques for measuring aerosol attenuation using the Central Laser Facility at the Pierre Auger Observatory_ , _Journal of Instrumentation_ 8 (2013) P04009 [1303.5576].
* [9] C. Medina-Hernandez for the Pierre Auger Collaboration, _Atmospheric monitoring at the Pierre Auger Observatory using the upgraded Central Laser Facility_ , in _34th International Cosmic Ray Conference (ICRC2015)_ , PoS(ICRC2015)624 [1909.09073].
* [10] S. Y. Benzvi, R. Cester, M. Chiosso, B. M. Connolly, A. Filipčič, B. García et al., _The Lidar system of the Pierre Auger Observatory_ , _Nucl. Instrum. Meth. A_ 574 (2007) 171 [astro-ph/0609063].
* [11] J. Chirinos, _Remote Sensing of Clouds using Satellites, Lidars, CLF/XLF and IR Cameras at the Pierre Auger Observatory_ , Atmospheric Monitoring for High Energy AstroParticle Detectors (AtmoHEAD) 2014, _European Physical Journal Web of Conferences_ 89 (2015) 03012.
* [12] D. Mandat, M. Pech, M. Hrabovsky, P. Schovanek, M. Palatka, P. Travnicek et al., _All Sky Camera instrument for night sky monitoring_ , Atmospheric Monitoring for High Energy AstroParticle Detectors (AtmoHEAD) 2013 [1402.4762].
* [13] Pierre Auger collaboration, _The rapid atmospheric monitoring system of the Pierre Auger Observatory_ , _Journal of Instrumentation_ 7 (2012) P09001 [1208.1675].
* [14] J. Blazek, _Searching for Anomalous Longitudinal Profiles with the FRAM Telescope_ , Atmospheric Monitoring for High Energy Astroparticle Detectors (AtmoHEAD) 2016, _European Physical Journal Web of Conferences_ 144, 01009 (2017)
* [15] B. S. Acharya, et al. (CTA Consortium), _Introducing the CTA concept_ , _Astroparticle Physics_ 43 (2013) 3.
* [16] J. Ebr, S. Karpov, J. Eliášek, J. Blažek, R. Cunniffe, I. Ebrová et al., _A new method for aerosol measurement using wide-field photometry_ , submitted to _Astronomical Journal_ (2020) [2101.03074].
* [17] J. Ebr, J. Blažek, M. Prouza, D. Mandát, M. Pech, M. Mašek et al., _Aerosol Measurements with the FRAM Telescope_ , Atmospheric Monitoring for High Energy Astroparticle Detectors (AtmoHEAD) 2016, _European Physical Journal Web of Conferences_ 144, 01011 (2017)
* [18] G. E. Shaw, _Sun photometry_ , _Bulletin of the American Meteorological Society_ 64 (1983) 4.
* [19] E. Høg, C. Fabricius, V. V. Makarov, S. Urban, T. Corbin, G. Wycoff et al., _The Tycho-2 catalogue of the 2.5 million brightest stars_ , _Astronomy & Astrophysics_ 355 (2000) L27.
* [20] J. Ebr, T. Bulik, L. Font, M. Gaug, P. Janecek, J. Jurysek et al., _Atmospheric calibration of the Cherenkov Telescope Array_ , in _35th International Cosmic Ray Conference (ICRC2017)_ , PoS(ICRC2017)833, [1709.04273].
* [21] J. Binney and M. Merrifield, _Galactic Astronomy_ , Princeton University Press (1998).
* [22] Pierre Auger collaboration, _Description of atmospheric conditions at the Pierre Auger Observatory using the Global Data Assimilation System (GDAS)_ , _Astroparticle Physics_ 35 (2012) 591 [1201.2276].
* [23] A. R. Kolhe, G. V. Pawar, S. R. Varpe, P. P. Kumar, P. C. S. Devara and G. R. Aher, _Multi-year analysis of aerosol properties retrieved from the Ångström parameters for different spectral ranges over Pune_ , _Aerosol and Air Quality Research_ 16 (2017) 3266.
* [24] D. M. Giles, B. N. Holben, T. F. Eck, A. Sinyuk, I. Slutsker, A. Smirnov et al., _Using the Aerosol Single Scattering Albedo and Angstrom Exponent from AERONET to Determine Aerosol Origins and Mixing States over the Indo-Gangetic Plain_ , American Geophysical Union, Fall Meeting 2012, abstract id. A13M-04.
* [25] A. A. Floutsi, M. B. Korras Carraca, C. Matsoukas, N. Hatzianastassiou and G. Biskos, _The regime of Aerosol Optical Depth and Ångström exponent over Central and South Asia_ , Central Asian DUst Conference (CADUC 2019), _E3S Web of Conferences_ 99, 01003 (2019).
* [26] J. Ebr, P. Janeček, M. Prouza, P. Kubánek, M. Jelínek, M. Mašek et al., _FRAM: showers, comets, GRBs and popular science_ , III Workshop on Robotic Autonomous Observatories (2013), _Revista Mexicana de Astronomia y Astrofisica Conference Series_ 45, 114 (2014).
* [27] _Schneider Electric SE_ , _http://www.se.com_ .
* [28] _Hollywood General Machining Inc._ , _http://www.losmandy.com_ .
* [29] _Software Bisque, Inc._ , _http://www.bisque.com_ .
* [30] _COMEC snc_ , _http://www.10micron.eu_ .
* [31] _IDEX Health and Science LLC._ , _http://www.flicamera.com_ .
* [32] _Moravian Instruments Cameras for Astronomy_ , _http://www.gxccd.com_ .
* [33] _Rigel Systems_ , _http://www.rigelsys.com/_ .
* [34] _Meade Instruments Corp._ , _http://www.meade.com_ .
* [35] _Orion Optics UK_ , _http://www.orionoptics.co.uk_ .
* [36] _ASA Astrosysteme GmbH_ , _http://www.astrosysteme.com_ .
* [37] P. Kubánek, _RTS2—The Remote Telescope System_ , Workshop on Robotic Autonomous Observatories 2009, _Advances in Astronomy_ 2010 (2010) 902484 [1005.1014].
* [38] P. Kubánek, M. Jelínek, S. Vítek, A. de Ugarte Postigo, M. Nekola, J. French et al., _Status of robotics telescopes driven by RTS2 (BART, BOOTES, FRAM and Watcher)_ , _Nuovo Cimento B Serie_ 121 (2006) 1501.
* [39] C. Meurer, N. Scharf for the Pierre Auger Collaboration, _HEAT - a low energy enhancement of the Pierre Auger Observatory_ , 22nd European Cosmic Ray Symposium (ECRS 2010), _Astrophysics and Space Sciences Transactions_ 7 (2011) 183 [1106.1329].
$\bullet$
## The Pierre Auger Collaboration
A. Aab80, P. Abreu72, M. Aglietta52,50, J.M. Albury12, I. Allekotte1, A.
Almela8,11, J. Alvarez-Muñiz79, R. Alves Batista80, G.A. Anastasi61,50, L.
Anchordoqui87, B. Andrada8, S. Andringa72, C. Aramo48, P.R. Araújo Ferreira40,
J. C. Arteaga Velázquez66, H. Asorey8, P. Assis72, G. Avila10, A.M. Badescu75,
A. Bakalova30, A. Balaceanu73, F. Barbato43,44, R.J. Barreira Luz72, K.H.
Becker36, J.A. Bellido12,68, C. Berat34, M.E. Bertaina61,50, X. Bertou1, P.L.
Biermannb, T. Bister40, J. Biteau35, J. Blazek30, C. Bleve34, M. Boháčová30,
D. Boncioli55,44, C. Bonifazi24, L. Bonneau Arbeletche19, N. Borodai69, A.M.
Botti8, J. Brackd, T. Bretz40, P.G. Brichetto Orchera8, F.L. Briechle40, P.
Buchholz42, A. Bueno78, S. Buitink14, M. Buscemi45, K.S. Caballero-Mora65, L.
Caccianiga57,47, F. Canfora80,82, I. Caracas36, J.M. Carceller78, R.
Caruso56,45, A. Castellina52,50, F. Catalani17, G. Cataldi46, L. Cazon72, M.
Cerda9, J.A. Chinellato20, K. Choi13, J. Chudoba30, L. Chytka31, R.W. Clay12,
A.C. Cobos Cerutti7, R. Colalillo58,48, A. Coleman93, M.R. Coluccia46, R.
Conceição72, A. Condorelli43,44, G. Consolati47,53, F. Contreras10, F.
Convenga54,46, D. Correia dos Santos26, C.E. Covault85, S. Dasso5,3, K.
Daumiller39, B.R. Dawson12, J.A. Day12, R.M. de Almeida26, J. de Jesús8,39,
S.J. de Jong80,82, G. De Mauro80,82, J.R.T. de Mello Neto24,25, I. De
Mitri43,44, J. de Oliveira26, D. de Oliveira Franco20, F. de Palma54,46, V. de
Souza18, E. De Vito54,46, M. del Río10, O. Deligny32, A. Di Matteo50, C.
Dobrigkeit20, J.C. D’Olivo67, R.C. dos Anjos23, M.T. Dova4, J. Ebr30, R.
Engel37,39, I. Epicoco54,46, M. Erdmann40, C.O. Escobara, A. Etchegoyen8,11,
H. Falcke80,83,82, J. Farmer92, G. Farrar90, A.C. Fauth20, N. Fazzinia, F.
Feldbusch38, F. Fenu52,50, B. Fick89, J.M. Figueira8, A. Filipčič77,76, T.
Fodran80, M.M. Freire6, T. Fujii92,e, A. Fuster8,11, C. Galea80, C.
Galelli57,47, B. García7, A.L. Garcia Vegas40, H. Gemmeke38, F. Gesualdi8,39,
A. Gherghel-Lascu73, P.L. Ghia32, U. Giaccari80, M. Giammarchi47, M. Giller70,
J. Glombitza40, F. Gobbi9, F. Gollan8, G. Golup1, M. Gómez Berisso1, P.F.
Gómez Vitale10, J.P. Gongora10, J.M. González1, N. González13, I. Goos1,39, D.
Góra69, A. Gorgi52,50, M. Gottowik36, T.D. Grubb12, F. Guarino58,48, G.P.
Guedes21, E. Guido50,61, S. Hahn39,8, P. Hamal30, M.R. Hampel8, P. Hansen4, D.
Harari1, V.M. Harvey12, A. Haungs39, T. Hebbeker40, D. Heck39, G.C. Hill12, C.
Hojvata, J.R. Hörandel80,82, P. Horvath31, M. Hrabovský31, T. Huege39,14, J.
Hulsman8,39, A. Insolia56,45, P.G. Isar74, P. Janecek30, J.A. Johnsen86, J.
Jurysek30, A. Kääpä36, K.H. Kampert36, B. Keilhauer39, J. Kemp40, H.O.
Klages39, M. Kleifges38, J. Kleinfeller9, M. Köpke37, N. Kunka38, B.L. Lago16,
R.G. Lang18, N. Langner40, M.A. Leigui de Oliveira22, V. Lenok39, A.
Letessier-Selvon33, I. Lhenry-Yvon32, D. Lo Presti56,45, L. Lopes72, R.
López62, L. Lu94, Q. Luce37, A. Lucero8, J.P. Lundquist76, A. Machado
Payeras20, G. Mancarella54,46, D. Mandat30, B.C. Manning12, J. Manshanden41,
P. Mantscha, S. Marafico32, A.G. Mariazzi4, I.C. Mariş13, G. Marsella59,45, D.
Martello54,46, H. Martinez18, O. Martínez Bravo62, M. Mastrodicasa55,44, H.J.
Mathes39, J. Matthews88, G. Matthiae60,49, E. Mayotte36, P.O. Mazura, G.
Medina-Tanco67, D. Melo8, A. Menshikov38, K.-D. Merenda86, S. Michal31, M.I.
Micheletti6, L. Miramonti57,47, S. Mollerach1, F. Montanet34, C. Morello52,50,
M. Mostafá91, A.L. Müller8, M.A. Muller20, K. Mulrey14, R. Mussa50, M.
Muzio90, W.M. Namasaka36, A. Nasr-Esfahani36, L. Nellen67, M. Niculescu-
Oglinzanu73, M. Niechciol42, D. Nitz89, D. Nosek29, V. Novotny29, L. Nožka31,
A Nucita54,46, L.A. Núñez28, M. Palatka30, J. Pallotta2, P. Papenbreer36, G.
Parente79, A. Parra62, M. Pech30, F. Pedreira79, J. Pȩkala69, R. Pelayo64, J.
Peña-Rodriguez28, E.E. Pereira Martins37,8, J. Perez Armand19, C. Pérez
Bertolli8,39, M. Perlin8,39, L. Perrone54,46, S. Petrera43,44, T. Pierog39, M.
Pimenta72, V. Pirronello56,45, M. Platino8, B. Pont80, M. Pothast82,80, P.
Privitera92, M. Prouza30, A. Puyleart89, S. Querchfeld36, J. Rautenberg36, D.
Ravignani8, M. Reininghaus39,8, J. Ridky30, F. Riehn72, M. Risse42, V.
Rizi55,44, W. Rodrigues de Carvalho19, J. Rodriguez Rojo10, M.J. Roncoroni8,
M. Roth39, E. Roulet1, A.C. Rovero5, P. Ruehl42, S.J. Saffi12, A. Saftoiu73,
F. Salamida55,44, H. Salazar62, G. Salina49, J.D. Sanabria Gomez28, F.
Sánchez8, E.M. Santos19, E. Santos30, F. Sarazin86, R. Sarmento72, C.
Sarmiento-Cano8, R. Sato10, P. Savina54,46,32, C.M. Schäfer39, V. Scherini46,
H. Schieler39, M. Schimassek37,8, M. Schimp36, F. Schlüter39,8, D. Schmidt37,
O. Scholten81,14, P. Schovánek30, F.G. Schröder93,39, S. Schröder36, J.
Schulte40, S.J. Sciutto4, M. Scornavacche8,39, A. Segreto51,45, S. Sehgal36,
R.C. Shellard15, G. Sigl41, G. Silli8,39, O. Sima73,f, R. Šmída92, P.
Sommers91, J.F. Soriano87, J. Souchard34, R. Squartini9, M. Stadelmaier39,8,
D. Stanca73, S. Stanič76, J. Stasielak69, P. Stassi34, A. Streich37,8, M.
Suárez-Durán28, T. Sudholz12, T. Suomijärvi35, A.D. Supanitsky8, J. Šupík31,
Z. Szadkowski71, A. Taboada37, A. Tapia27, C. Taricco61,50, C.
Timmermans82,80, O. Tkachenko39, P. Tobiska30, C.J. Todero Peixoto17, B.
Tomé72, A. Travaini9, P. Travnicek30, C. Trimarelli55,44, M. Trini76, M.
Tueros4, R. Ulrich39, M. Unger39, L. Vaclavek31, M. Vacula31, J.F. Valdés
Galicia67, L. Valore58,48, E. Varela62, V. Varma K.C.8,39, A. Vásquez-
Ramírez28, D. Veberič39, C. Ventura25, I.D. Vergara Quispe4, V. Verzi49, J.
Vicha30, J. Vink84, S. Vorobiov76, H. Wahlberg4, C. Watanabe24, A.A. Watsonc,
M. Weber38, A. Weindl39, L. Wiencke86, H. Wilczyński69, T. Winchen14, M.
Wirtz40, D. Wittkowski36, B. Wundheiler8, A. Yushkov30, O. Zapparrata13, E.
Zas79, D. Zavrtanik76,77, M. Zavrtanik77,76, L. Zehrer76, A. Zepeda63 and R.
Cunniffe30, J. Eliášek30, I. Ebrová,g M. Jelínek,h S. Karpov30, P. Kubánek,i
M. Mašek30
1
Centro Atómico Bariloche and Instituto Balseiro (CNEA-UNCuyo-CONICET), San
Carlos de Bariloche, Argentina
2
Centro de Investigaciones en Láseres y Aplicaciones, CITEDEF and CONICET,
Villa Martelli, Argentina
3
Departamento de Física and Departamento de Ciencias de la Atmósfera y los
Océanos, FCEyN, Universidad de Buenos Aires and CONICET, Buenos Aires,
Argentina
4
IFLP, Universidad Nacional de La Plata and CONICET, La Plata, Argentina
5
Instituto de Astronomía y Física del Espacio (IAFE, CONICET-UBA), Buenos
Aires, Argentina
6
Instituto de Física de Rosario (IFIR) – CONICET/U.N.R. and Facultad de
Ciencias Bioquímicas y Farmacéuticas U.N.R., Rosario, Argentina
7
Instituto de Tecnologías en Detección y Astropartículas (CNEA, CONICET,
UNSAM), and Universidad Tecnológica Nacional – Facultad Regional Mendoza
(CONICET/CNEA), Mendoza, Argentina
8
Instituto de Tecnologías en Detección y Astropartículas (CNEA, CONICET,
UNSAM), Buenos Aires, Argentina
9
Observatorio Pierre Auger, Malargüe, Argentina
10
Observatorio Pierre Auger and Comisión Nacional de Energía Atómica, Malargüe,
Argentina
11
Universidad Tecnológica Nacional – Facultad Regional Buenos Aires, Buenos
Aires, Argentina
12
University of Adelaide, Adelaide, S.A., Australia
13
Université Libre de Bruxelles (ULB), Brussels, Belgium
14
Vrije Universiteit Brussels, Brussels, Belgium
15
Centro Brasileiro de Pesquisas Fisicas, Rio de Janeiro, RJ, Brazil
16
Centro Federal de Educação Tecnológica Celso Suckow da Fonseca, Nova Friburgo,
Brazil
17
Universidade de São Paulo, Escola de Engenharia de Lorena, Lorena, SP, Brazil
18
Universidade de São Paulo, Instituto de Física de São Carlos, São Carlos, SP,
Brazil
19
Universidade de São Paulo, Instituto de Física, São Paulo, SP, Brazil
20
Universidade Estadual de Campinas, IFGW, Campinas, SP, Brazil
21
Universidade Estadual de Feira de Santana, Feira de Santana, Brazil
22
Universidade Federal do ABC, Santo André, SP, Brazil
23
Universidade Federal do Paraná, Setor Palotina, Palotina, Brazil
24
Universidade Federal do Rio de Janeiro, Instituto de Física, Rio de Janeiro,
RJ, Brazil
25
Universidade Federal do Rio de Janeiro (UFRJ), Observatório do Valongo, Rio de
Janeiro, RJ, Brazil
26
Universidade Federal Fluminense, EEIMVR, Volta Redonda, RJ, Brazil
27
Universidad de Medellín, Medellín, Colombia
28
Universidad Industrial de Santander, Bucaramanga, Colombia
29
Charles University, Faculty of Mathematics and Physics, Institute of Particle
and Nuclear Physics, Prague, Czech Republic
30
Institute of Physics of the Czech Academy of Sciences, Prague, Czech Republic
31
Palacky University, RCPTM, Olomouc, Czech Republic
32
CNRS/IN2P3, IJCLab, Université Paris-Saclay, Orsay, France
33
Laboratoire de Physique Nucléaire et de Hautes Energies (LPNHE), Sorbonne
Université, Université de Paris, CNRS-IN2P3, Paris, France
34
Univ. Grenoble Alpes, CNRS, Grenoble Institute of Engineering Univ. Grenoble
Alpes, LPSC-IN2P3, 38000 Grenoble, France
35
Université Paris-Saclay, CNRS/IN2P3, IJCLab, Orsay, France
36
Bergische Universität Wuppertal, Department of Physics, Wuppertal, Germany
37
Karlsruhe Institute of Technology (KIT), Institute for Experimental Particle
Physics, Karlsruhe, Germany
38
Karlsruhe Institute of Technology (KIT), Institut für Prozessdatenverarbeitung
und Elektronik, Karlsruhe, Germany
39
Karlsruhe Institute of Technology (KIT), Institute for Astroparticle Physics,
Karlsruhe, Germany
40
RWTH Aachen University, III. Physikalisches Institut A, Aachen, Germany
41
Universität Hamburg, II. Institut für Theoretische Physik, Hamburg, Germany
42
Universität Siegen, Department Physik – Experimentelle Teilchenphysik, Siegen,
Germany
43
Gran Sasso Science Institute, L’Aquila, Italy
44
INFN Laboratori Nazionali del Gran Sasso, Assergi (L’Aquila), Italy
45
INFN, Sezione di Catania, Catania, Italy
46
INFN, Sezione di Lecce, Lecce, Italy
47
INFN, Sezione di Milano, Milano, Italy
48
INFN, Sezione di Napoli, Napoli, Italy
49
INFN, Sezione di Roma “Tor Vergata”, Roma, Italy
50
INFN, Sezione di Torino, Torino, Italy
51
Istituto di Astrofisica Spaziale e Fisica Cosmica di Palermo (INAF), Palermo,
Italy
52
Osservatorio Astrofisico di Torino (INAF), Torino, Italy
53
Politecnico di Milano, Dipartimento di Scienze e Tecnologie Aerospaziali ,
Milano, Italy
54
Università del Salento, Dipartimento di Matematica e Fisica “E. De Giorgi”,
Lecce, Italy
55
Università dell’Aquila, Dipartimento di Scienze Fisiche e Chimiche, L’Aquila,
Italy
56
Università di Catania, Dipartimento di Fisica e Astronomia, Catania, Italy
57
Università di Milano, Dipartimento di Fisica, Milano, Italy
58
Università di Napoli “Federico II”, Dipartimento di Fisica “Ettore Pancini”,
Napoli, Italy
59
Università di Palermo, Dipartimento di Fisica e Chimica ”E. Segrè”, Palermo,
Italy
60
Università di Roma “Tor Vergata”, Dipartimento di Fisica, Roma, Italy
61
Università Torino, Dipartimento di Fisica, Torino, Italy
62
Benemérita Universidad Autónoma de Puebla, Puebla, México
63
Centro de Investigación y de Estudios Avanzados del IPN (CINVESTAV), México,
D.F., México
64
Unidad Profesional Interdisciplinaria en Ingeniería y Tecnologías Avanzadas
del Instituto Politécnico Nacional (UPIITA-IPN), México, D.F., México
65
Universidad Autónoma de Chiapas, Tuxtla Gutiérrez, Chiapas, México
66
Universidad Michoacana de San Nicolás de Hidalgo, Morelia, Michoacán, México
67
Universidad Nacional Autónoma de México, México, D.F., México
68
Universidad Nacional de San Agustin de Arequipa, Facultad de Ciencias
Naturales y Formales, Arequipa, Peru
69
Institute of Nuclear Physics PAN, Krakow, Poland
70
University of Łódź, Faculty of Astrophysics, Łódź, Poland
71
University of Łódź, Faculty of High-Energy Astrophysics,Łódź, Poland
72
Laboratório de Instrumentação e Física Experimental de Partículas – LIP and
Instituto Superior Técnico – IST, Universidade de Lisboa – UL, Lisboa,
Portugal
73
“Horia Hulubei” National Institute for Physics and Nuclear Engineering,
Bucharest-Magurele, Romania
74
Institute of Space Science, Bucharest-Magurele, Romania
75
University Politehnica of Bucharest, Bucharest, Romania
76
Center for Astrophysics and Cosmology (CAC), University of Nova Gorica, Nova
Gorica, Slovenia
77
Experimental Particle Physics Department, J. Stefan Institute, Ljubljana,
Slovenia
78
Universidad de Granada and C.A.F.P.E., Granada, Spain
79
Instituto Galego de Física de Altas Enerxías (IGFAE), Universidade de Santiago
de Compostela, Santiago de Compostela, Spain
80
IMAPP, Radboud University Nijmegen, Nijmegen, The Netherlands
81
KVI – Center for Advanced Radiation Technology, University of Groningen,
Groningen, The Netherlands
82
Nationaal Instituut voor Kernfysica en Hoge Energie Fysica (NIKHEF), Science
Park, Amsterdam, The Netherlands
83
Stichting Astronomisch Onderzoek in Nederland (ASTRON), Dwingeloo, The
Netherlands
84
Universiteit van Amsterdam, Faculty of Science, Amsterdam, The Netherlands
85
Case Western Reserve University, Cleveland, OH, USA
86
Colorado School of Mines, Golden, CO, USA
87
Department of Physics and Astronomy, Lehman College, City University of New
York, Bronx, NY, USA
88
Louisiana State University, Baton Rouge, LA, USA
89
Michigan Technological University, Houghton, MI, USA
90
New York University, New York, NY, USA
91
Pennsylvania State University, University Park, PA, USA
92
University of Chicago, Enrico Fermi Institute, Chicago, IL, USA
93
University of Delaware, Department of Physics and Astronomy, Bartol Research
Institute, Newark, DE, USA
94
University of Wisconsin-Madison, Department of Physics and WIPAC, Madison, WI,
USA
—–
a
Fermi National Accelerator Laboratory, Fermilab, Batavia, IL, USA
b
Max-Planck-Institut für Radioastronomie, Bonn, Germany
c
School of Physics and Astronomy, University of Leeds, Leeds, United Kingdom
d
Colorado State University, Fort Collins, CO, USA
e
now at Hakubi Center for Advanced Research and Graduate School of Science,
Kyoto University, Kyoto, Japan
f
also at University of Bucharest, Physics Department, Bucharest, Romania
g
Nicolaus Copernicus Astronomical Center, Polish Academy of Sciences, Warsaw,
Poland
h
Astronomical Institute of the Czech Academy of Sciences, Ondřejov, Czech
Republic
i
Vera C. Rubin Observatory, La Serena, Chile
|
# Sojourn times of Gaussian related random fields
Krzysztof Dȩbicki Krzysztof Dȩbicki, Mathematical Institute, University of
Wrocław, pl. Grunwaldzki 2/4, 50-384 Wrocław, Poland
<EMAIL_ADDRESS>, Enkelejd Hashorva Enkelejd Hashorva,
Department of Actuarial Science, University of Lausanne, UNIL-Dorigny, 1015
Lausanne, Switzerland<EMAIL_ADDRESS>, Peng Liu Peng Liu,
Department of Mathematical Sciences, University of Essex, Colchester, UK
<EMAIL_ADDRESS>and Zbigniew Michna Zbigniew Michna, Department of
Logistics, Wrocław University of Economics and Business, Poland
<EMAIL_ADDRESS>
Abstract: This paper is concerned with the asymptotic analysis of sojourn
times of random fields with continuous sample paths. Under a very general
framework we show that there is an interesting relationship between tail
asymptotics of sojourn times and that of supremum. Moreover, we establish the
uniform double-sum method to derive the tail asymptotics of sojourn times. In
the literature, based on the pioneering research of S. Berman the sojourn
times have been utilised to derive the tail asymptotics of supremum of
Gaussian processes. In this paper we show that the opposite direction is even
more fruitful, namely knowing the asymptotics of supremum o f random processes
and fields (in particular Gaussian) it is possible to establish the
asymptotics of their sojourn times. We illustrate our findings considering i)
two dimensional Gaussian random fields, ii) chi-process generated by
stationary Gaussian processes and iii) stationary Gaussian queueing processes.
Key Words: sojourn/occupation times; exact asymptotics; generalized Berman-
type constants; Gaussian random fields; queueing process; chi-process.
AMS Classification: Primary 60G15; secondary 60G70
## 1\. Introduction & First Result
Let $X(t),t\in E$ be a random field with compact parameter set
$E\subset\mathbb{R}^{d},d\geq 1$ and almost surely continuous sample paths.
For a given level $u\in\mathbb{R}$ define the excursion set of $X$ above the
level $u$ by
$A_{u}(X)\coloneqq\\{t\in E:X(t)>u\\}.$
The probability that $A_{u}$ is not empty
$\mathbb{P}\left\\{A_{u}(X)\not=\emptyset\right\\}=\mathbb{P}\left\\{\exists
t\in E:X(t)>u\right\\}=\mathbb{P}\left\\{\sup_{t\in E}X(t)>u\right\\}=:p_{u}$
is widely studied in the literature under the asymptotic regime $u\to\infty$,
and the assumption that $X$ has marginals with infinite upper endpoint; see,
e.g., [26, 1] for $X$ being Gaussian processes and related random fields.
Define the Lebesgue volume of $A_{u}(X)$ by
$Vol(A_{u}(X))=\int_{E}\mathbb{I}(X(t)>u)dt.$
For specific cases, commonly $d=1$ and $X$ is stationary, asymptotic results
as $u\to\infty$ are also known for the probability that the volume of the
excursion set (occupation time or sojourn time) exceeds $v(u)z$, i.e.,
approximations of
$r_{u}(z)\coloneqq\mathbb{P}\left\\{Vol(A_{u}(X))>v(u)z\right\\},\quad
u\to\infty$
for some specific positive scale function $v$ and $z\geq 0$ are available, see
the seminal contribution [4].
The non-stationary case has been considered in [5, 6]. See also [7] for the
comprehensive introduction of extremes of sojourns for Gaussian processes.
In this contribution we are mainly interested in the formalisation of the
uniform double-sum method for sojourns of random processes and fields focusing
on the multidimensional case $d\geq 2$, for which no asymptotic results for
$r_{u}(z)$ are available in the literature.
The first question of our study is whether we can determine a positive scaling
functions $v(u),u>0$ and some survival function $\bar{F}$ such that
(1)
$\lim_{u\to\infty}\mathbb{P}\left\\{Vol(A_{u}(X))>v(u)z\Bigl{\lvert}Vol(A_{u}(X))>0\right\\}=\lim_{u\to\infty}\mathbb{P}\left\\{Vol(A_{u}(X))>v(u)z\Bigl{\lvert}\sup_{t\in
E}X(t)>u\right\\}=\bar{F}(z)$
is valid for all $z\geq 0$.
If (1) holds for some $z$ positive such that $\bar{F}(z)>0$ the asymptotics of
$r_{u}(z)$ is proportional to that of $p_{u}$, i.e.,
$r_{u}(z)\sim\bar{F}(z)p_{u},\quad u\to\infty.$
Here $a(t)\sim b(t)$ means asymptotic equivalence of two real-valued functions
$a(t)$ and $b(t)$ when the argument $t$ tends to infinity or zero. For a given
index set $K$ we write $\sharp K$ for the cardinality of $K$.
The following theorem states tractable conditions that imply (1) for $X$ as
above and $E=E_{u}$. In order to avoid repetition, all Gaussian processes
hereafter are assume to have almost surely continuous sample paths.
###### Theorem 1.1.
Let $E_{u},u>0$ be compact set of $\mathbb{R}^{d}$ such that
$\lim_{u\to\infty}\mathbb{P}\left\\{\sup_{t\in E_{u}}X(t)>u\right\\}=0$.
Suppose that there exist collections of Lebesgue measurable disjoint compact
sets $I_{k}(u,n),k\in K_{u,n}$ with $K_{u,n}$ non-empty countable index sets
such that
$E(u,n)\coloneqq\bigcup_{k\in
K_{u,n}}I_{k}(u,n)\subset{\color[rgb]{0,0,0}E_{u}},$
then (1) holds with $E=E_{u}$ if the following three conditions are satisfied:
A1) (Reduction to relevant sets)
$\lim_{n\rightarrow\infty}\limsup_{u\rightarrow\infty}\frac{\mathbb{P}\left\\{\sup_{t\in{\color[rgb]{0,0,0}E_{u}}\setminus
E(u,n)}X(t)>u\right\\}}{\mathbb{P}\left\\{\sup_{t\in{\color[rgb]{0,0,0}E(u,n)}}X(t)>u\right\\}}=0.$
A2) (Uniform single-sum approximation) There exists $v(u)>0$ and
$\bar{F}_{n},n\geq 1$ such that
(2) $\lim_{u\to\infty}\sup_{k\in
K_{u,n}}\biggl{\lvert}\frac{\mathbb{P}\left\\{Vol(\\{t\in
I_{k}(u,n):X(t)>u\\})>v(u)x\right\\}}{\mathbb{P}\left\\{\sup_{t\in
I_{k}(u,n)}X(t)>u\right\\}}-\bar{F}_{n}(x)\biggr{\rvert}=0,\quad x\geq 0,\
{n\geq 1}$
and for all $x\geq 0$
(3)
$\displaystyle\bar{F}(x)\coloneqq\lim_{n\rightarrow\infty}\bar{F}_{n}(x)\in(0,1].$
A3) (Double-sum negligibility) For all large $n$ and large $u$, $\sharp
K_{u,n}\geq 2$ and
$\lim_{n\rightarrow\infty}\limsup_{u\rightarrow\infty}\frac{\sum_{i\neq
j,i,j\in K_{u,n}}\mathbb{P}\left\\{\sup_{t\in I_{i}(u,n)}X(t)>u,\sup_{t\in
I_{j}(u,n)}X(t)>u\right\\}}{\sum_{k\in K_{u,n}}\mathbb{P}\left\\{\sup_{t\in
I_{k}(u,n)}X(t)>u\right\\}}=0.$
For $X(t),t\in\mathbb{R}$ being a Gaussian process, [10] shows that conditions
A1)-A3) are satisfied under very general assumptions on $X$. From [10], we can
formulate some general conditions on $X$ that imply
(4) $\lim_{u\to\infty}\sup_{k\in
K_{u,n}}\biggl{\lvert}\frac{\mathbb{P}\left\\{\sup_{t\in
I_{k}(u,n)}X(t)>u\right\\}}{\Xi_{k}(u)}-C_{n}\biggr{\rvert}=0$
for some known deterministic functions $\Xi_{k}(u)$, $k\in K_{u,n}$ and
$C_{n}$ positive constants such that $\lim_{n\to\infty}C_{n}=C\in(0,\infty)$.
In order to prove (2) if (4) holds, we shall prove that
(5) $\lim_{u\to\infty}\sup_{k\in
K_{u,n}}\biggl{\lvert}\frac{\mathbb{P}\left\\{Vol(\\{t\in
I_{k}(u,n):X(t)>u\\})>v(u)x\right\\}}{\Xi_{k}(u)}-D_{n}(x)\biggr{\rvert}=0,\quad$
where $D_{n},\ n\geq 1$ are deterministic functions such that
$\lim_{n\to\infty}D_{n}(x)=D(x)>0$, $x\geq 0$. This then in turn implies that
(3) holds with
$\bar{F}(x)=\frac{D(x)}{C}.$
Note that in case that $D$ is continuous at $x=0$ we also expect that $C=D(0)$
for all $z\geq 0$.
In the literature various results are known for supremum of functions of
Gaussian vector processes, for instance for chi-square processes, chaos of
Gaussian processes, order statistics of Gaussian processes, (see, e.g., [25,
26, 20, 2]) or reflected Gaussian processes modelling a queueing process with
Gaussian input (see, e.g., [24, 21, 17, 27, 15, 22, 19, 23, 12, 13]). In
Section 3 we illustrate the applicability of Theorem 1.1 by the analysis of
three diverse families of stochastic processes: 1) Gaussian random fields
(GRF’s), 2) chi-processes and 3) reflected fractional Brownian motions. For
all this families of stochastic processes the available results in the
literature show that both A1) and A3) hold under quite general conditions; see
Section 2. Hence, in view of Theorem 1.1, in order to get (1) it suffices to
determine $\bar{F}$ in A2).
Except the above examples, our findings can also be applied to many other
GRF’s. For instance, multi-dimensional GRF’s with $d\geq 3$, non-stationary
chi-process or chi-square process, Gaussian chaos process, non-stationary
Gaussian fluid queues and so on. However, we shall not analyze these random
processes or fields in this paper.
Brief organisation of the rest of the paper. In Section 2 we introduce some
notation and Berman-type constants that play the core role in the description
of $\bar{F}$. In Section 3, we provide examples that illustrate the derived in
Theorem 1.1 technique for getting (1). Some technical lemmas are given in
Section 4; their proofs are deferred to Section 6. The proofs of the main
contributions of this paper are presented in Section 5.
## 2\. Berman-type constants
We begin with the introduction of the Berman-type constants for given
independent fBm’s $B_{\alpha_{i}}(s),s\in\mathbb{R}$ with Hurst index
$\alpha_{i}/2\in(0,1]$, $i=1,2$. For given continuous functions $h_{1},h_{2}$
set
$W_{\alpha_{1},\alpha_{2},h_{1},h_{2}}(t)\coloneqq\sum_{i=1}^{2}(W_{\alpha_{i}}(t_{i})-h_{i}(t_{i})),\quad
t=(t_{1},t_{2})\in\mathbb{R}^{2},\quad
W_{\alpha_{i}}(t_{i})=\sqrt{2}B_{\alpha_{i}}(t_{i})-\left|t_{i}\right|^{\alpha_{i}}\,.$
For simplicity, let $B_{0}(s)\equiv 0,s\in\mathbb{R}$. For
$\alpha_{i}\in[0,2],i=1,2$, $x\geq 0$ and $E\subset\mathbb{R}^{2}$ a compact
set, let
$\mathcal{B}_{\alpha_{1},\alpha_{2}}^{h_{1},h_{2}}(x,E)=\int_{\mathbb{R}}\mathbb{P}\left\\{\int_{E}\mathbb{I}(W_{\alpha_{1},\alpha_{2},h_{1},h_{2}}(t)>z)dt>x\right\\}e^{z}dz$
and if the limit exists, define
$\mathcal{B}_{\alpha_{1},\alpha_{2}}^{b_{1}|t_{1}|^{\beta_{1}},b_{2}|t_{2}|^{\beta_{2}}}(x)\coloneqq\lim_{S\rightarrow\infty}\frac{\mathcal{B}_{\alpha_{1},\alpha_{2}}^{b_{1}|t_{1}|^{\beta_{1}},b_{2}|t_{2}|^{\beta_{2}}}(x,G(S,\alpha_{1},\beta_{1},\alpha_{2},\beta_{2}))}{S^{\mathbb{I}(\alpha_{1}<\beta_{1})+\mathbb{I}(\alpha_{2}<\beta_{2})}},$
where
$G(S,\alpha_{1},\beta_{1},\alpha_{2},\beta_{2})=\left\\{\begin{array}[]{cc}[0,S]^{2},&\alpha_{1}<\beta_{1},\alpha_{2}<\beta_{2},\\\
{[-S,S]\times[0,S]},&\alpha_{1}\geq\beta_{1},\alpha_{2}<\beta_{2},\\\
{[0,S]\times[-S,S]},&\alpha_{1}<\beta_{1},\alpha_{2}\geq\beta_{2},\\\
{[-S,S]^{2}}&\alpha_{1}\geq\beta_{1},\alpha_{2}\geq\beta_{2}.\end{array}\right.$
We omit superscripts $h_{i}$’s if $h_{1}(s)=h_{2}(s)=0,s\in\mathbb{R}$ and
then we put in our notation $\beta_{1}=\beta_{2}=\infty$ (this implies that
$\alpha_{1}<\beta_{1}$ and $\alpha_{2}<\beta_{2}$). Notice that for $x=0$,
$\mathcal{B}_{\alpha_{1},\alpha_{2}}^{h_{1},h_{2}}(x)$ reduces to the
classical Pickands or Piterbarg constants, see e.g., [26]. The one-dimensional
Berman type constant is given by
$\mathcal{B}_{\alpha}(x,[a,b])=\int_{\mathbb{R}}\mathbb{P}\left\\{\int_{[a,b]}\mathbb{I}(W_{\alpha}(s)>z)ds>x\right\\}e^{z}dz$
for $\alpha\in(0,2],a<b,a,b\in\mathbb{R}$, and
$\mathcal{B}_{\alpha}(x)=\lim_{S\to\infty}\frac{\mathcal{B}_{\alpha}(x,[0,S])}{S}.$
One can refer to [16] and [14] for the existence and properties of one-
dimensional Berman constants. For $x=0$,
$\mathcal{H}_{\alpha}\coloneqq\mathcal{B}_{\alpha}(0)$ reduces to the
classical Pickands constant; see, e.g., [26].
The next lemma deals with properties of
$\displaystyle\widehat{\mathcal{B}}_{\alpha_{1},\dots,\alpha_{m}}\left(x,\prod_{i=1}^{m}[0,n_{i}]\right)\coloneqq\int_{\mathbb{R}}\mathbb{P}\left\\{\int_{[0,n_{1}]}\mathbb{I}\left\\{\sup_{t_{i}\in[0,n_{i}],i=2,\dots,m}\sum_{i=1}^{m}W_{\alpha_{i}}(t_{i})>s\right\\}dt_{1}>x\right\\}e^{s}ds$
for $\alpha_{i}\in(0,2],i=1,\dots,m$ and $m\geq 1$.
###### Lemma 2.1.
For any $x\geq 0$, and $n_{1}>0$
(6)
$\displaystyle\widehat{\mathcal{B}}_{\alpha_{1},\dots,\alpha_{m}}(x,n_{1})$
$\displaystyle\coloneqq$
$\displaystyle\lim_{n_{i}\rightarrow\infty,i=2,\dots,m}\frac{\widehat{\mathcal{B}}_{\alpha_{1},\dots,\alpha_{m}}\left(x,\prod_{i=1}^{m}[0,n_{i}]\right)}{\prod_{i=2}^{m}n_{i}}$
$\displaystyle=$
$\displaystyle\prod_{i=2}^{m}\mathcal{H}_{\alpha_{i}}\int_{\mathbb{R}}\mathbb{P}\left\\{\int_{[0,n_{1}]}\mathbb{I}\left\\{W_{\alpha_{1}}(t)>s\right\\}dt>x\right\\}e^{s}ds\in(0,\infty)$
and
(7)
$\displaystyle\widehat{\mathcal{B}}_{\alpha_{1},\dots,\alpha_{m}}(x)\coloneqq\lim_{n\rightarrow\infty}\frac{\widehat{\mathcal{B}}_{\alpha_{1},\dots,\alpha_{m}}(x,n)}{n}=\mathcal{B}_{\alpha_{1}}(x)\prod_{i=2}^{m}\mathcal{H}_{\alpha_{i}}\in(0,\infty).$
###### Remark 2.2.
The limits in (6) are finite and positive and
$\widehat{\mathcal{B}}_{\alpha_{1},\dots,\alpha_{m}}(x,n_{1})$ is a continuous
function of $x$ over $[0,n_{1})$ which follows from the combination of Lemma
2.1 and Lemma 4.1 in [14]. The claim of Lemma 2.1 still holds if we replace
$B_{\alpha_{i}}$ by $X_{i}$ being independent centered Gaussian processes with
stationary increments and variance function satisfying some regular conditions
as e.g. in [17].
## 3\. Illustrating examples
In this section we shall apply Theorem 1.1 to three classes of processes: i)
GRF’s , ii) chi-process generated by a stationary Gaussian process and iii)
stationary reflected fractional Brownian motions with drift.
### 3.1. Sojourns of GRF’s
Although numerous results for the tail asymptotics of supremum of GRF’s are
available for both stationary and non-stationary cases (see e.g., [26, 28]),
sojourns have not been treated so far in the literature. It follows from the
available results in the literature, that A1) holds under quite general
conditions, for instance when the variance function has a unique point of
maximum and $X$ satisfies a global Hölder continuity condition, see e.g.,
[26]. The main tool for proving A1) is the so-called Piterbarg inequality, see
[26][Thm 8.1] and the recent contribution [9]. Under some further weak
assumptions on the variance/covariance function of $X$, also A3) has been
shown to hold for a wide collection of cases of interest, see [26, 8]. Thus,
in light of Theorem 1.1, in order to prove (1) for GRF’s the main task is the
explicit calculation of $\bar{F}$.
#### 3.1.1. GRF’s with constant variance
First we consider $X$ being a centred GRF with $Var(X(t))=1,t\in
E\subset\mathbb{R}^{2}$ and the correlation function $r(t,s)$,
$t,s\in\mathbb{R}^{2}$ satisfying
(8) $\displaystyle 1-r(t_{1},t_{2},s_{1},s_{2})\sim
a_{1}|t_{1}-s_{1}|^{\alpha_{1}}+a_{2}|t_{2}-s_{2}|^{\alpha_{2}},\quad(t_{1},t_{2}),(s_{1},s_{2})\in
E,|t_{i}-s_{i}|\rightarrow 0,i=1,2,$
with $a_{i}>0$ and $\alpha_{i}\in(0,2]$, $i=1,2$. Moreover,
(9) $\displaystyle
r(t_{1},t_{2},s_{1},s_{2})<1,\quad(t_{1},t_{2}),(s_{1},s_{2})\in
E,(t_{1},t_{2})\neq(s_{1},s_{2}).$
For notational simplicity we shall consider $E=[0,T_{1}]\times[0,T_{2}]$, the
results for general hypercubes in $\mathbb{R}^{d}$ follows with similar
calculations. The case that $T_{i}=T_{i,u},i=1,2$ depend on $u$ needs some
extra care. $T_{i,u}$’s should not be too small, i.e.,
$\lim_{u\to\infty}T_{i,u}u^{2/\alpha_{i}}=\infty,i=1,2.$
On the other side $T_{i,u}$’s cannot be too large too. If the GRF is
stationary, then for some $\beta\in(0,1)$ we should require that
$\lim_{u\to\infty}T_{1,u}T_{2,u}e^{-\beta u^{2}/2}=0.$
In the more complex situation that we are looking at below the existence of
$\beta$ is not clear. We suppress the discussion for long intervals in order
to avoid further complications.
###### Proposition 3.1.
Let $X(t),t\in E=[0,T_{1}]\times[0,T_{2}]$ be a centred GRF which satisfies
(8) and (9) and assume that
$v(u)=a_{1}^{-1/\alpha_{1}}a_{2}^{-1/\alpha_{2}}u^{-2/\alpha_{1}-2/\alpha_{2}}$.
Then for all $x\geq 0$
$\displaystyle\lim_{u\rightarrow\infty}\mathbb{P}\left\\{\int_{E}\mathbb{I}(X(t)>u)dt>v(u)x\Bigl{\lvert}\sup_{t\in
E}X(t)>u\right\\}=\frac{\mathcal{B}_{\alpha_{1},\alpha_{2}}(x)}{\mathcal{B}_{\alpha_{1},\alpha_{2}}(0)}.$
#### 3.1.2. GRF’s with non-constant variance
Denote by $\sigma(t)=\sqrt{Var(X(t))}$ and assume that
$t^{*}={(t^{*}_{1},t^{*}_{2})}\in E=[-T_{1},T_{1}]\times[-T_{2},T_{2}]$ is the
inner point of $E$, which is the unique point such that
$\sigma(t^{*})=\sup_{t\in E}\sigma(t)=1$ satisfying
(10) $\displaystyle 1-\sigma(t)\sim
b_{1}|t_{1}-t_{1}^{*}|^{\beta_{1}}+b_{2}|t_{2}-t_{2}^{*}|^{\beta_{2}},\quad
t=(t_{1},t_{2})\in E,\,{\lVert t-t^{*}\rVert\rightarrow 0},$
with $b_{i}>0,\beta_{i}>0$, $i=1,2$. Here $\lVert\cdot\rVert$ denotes the
Euclidean norm. Moreover, let
(11) $\displaystyle 1-r(t,s)\sim
a_{1}|t_{1}-s_{1}|^{\alpha_{1}}+a_{2}|t_{2}-s_{2}|^{\alpha_{2}}$
as $t,s\in E,\lVert t-t^{*}\rVert,\lVert s-t^{*}\rVert\rightarrow 0$ with
$a_{i}>0$ and $\alpha_{i}\in(0,2]$, $i=1,2$, $s=(s_{1},s_{2})$, where $r(t,s)$
is the correlation function of the random field $X$. In the notation below we
interpret $\infty\cdot 0$ as $0$.
###### Proposition 3.2.
If $X(t),t\in E$ is a centered GRF which satisfies (10) and (11) and
$v(u)=\prod_{i=1}^{2}\left(a_{i}^{-1/\alpha_{i}^{*}}u^{-2/\min(\alpha_{i},\beta_{i})}\right)$
with
$\alpha_{i}^{*}=\alpha_{i}\mathbb{I}(\alpha_{i}\leq\beta_{i})+\infty\mathbb{I}(\alpha_{i}>\beta_{i})$,
then for all $x\geq 0$
$\lim_{u\rightarrow\infty}\mathbb{P}\left\\{\int_{E}\mathbb{I}(X(t)>u)dt>v(u)x\Bigl{\lvert}\sup_{t\in
E}X(t)>u\right\\}=\frac{\mathcal{B}_{\hat{\alpha}_{1},\hat{\alpha}_{2}}^{\bar{a}_{1}b_{1}\left|t_{1}\right|^{\beta_{1}},\bar{a}_{2}b_{2}\left|t_{2}\right|^{\beta_{2}}}(x)}{\mathcal{B}_{\hat{\alpha}_{1},\hat{\alpha}_{2}}^{\bar{a}_{1}b_{1}\left|t_{1}\right|^{\beta_{1}},\bar{a}_{2}b_{2}\left|t_{2}\right|^{\beta_{2}}}(0)},$
where
$\bar{a}_{i}=\left\\{\begin{array}[]{cc}0&\alpha_{i}<\beta_{i}\\\
\frac{1}{a_{i}}&\alpha_{i}=\beta_{i}\\\
1&\alpha_{i}>\beta_{i}\end{array}\right.,\quad\hat{\alpha}_{i}=\left\\{\begin{array}[]{cc}\alpha_{i}&\alpha_{i}\leq\beta_{i}\\\
0&\alpha_{i}>\beta_{i}\end{array}\right.,i=1,2.$
### 3.2. Sojourns of chi-processes
Let $X(t),t\in[0,T]$ be a centered stationary Gaussian process with unit
variance and correlation function satisfying
$\displaystyle 1-r(s,t)\sim a|t-s|^{\alpha},\quad|s-t|\rightarrow 0,$
where $\alpha\in(0,2]$ and for all $s\neq t$, $s,t\in[0,T]$
$\displaystyle r(s,t)<1.$
Define the chi-process of degree $m\geq 1$ by
(12) $\displaystyle\chi(t)\coloneqq\sqrt{\sum_{i=1}^{m}X_{i}^{2}(t)},\quad
t\in\mathbb{R},$
where $X_{i},1\leq i\leq m$ are iid copies of $X$. The exact asymptotics of
$\mathbb{P}\left\\{\sup_{t\in[0,T]}\chi(t)>u\right\\}$ has been investigated
in [25, 26, 20]. In the following theorem we consider the sojourn time of
$\chi$.
###### Proposition 3.3.
Let $\chi$ be defined as in (12). If $v(u)=a^{-1/\alpha}u^{-2/\alpha}$, then
for all $x\geq 0$
$\lim_{u\rightarrow\infty}\mathbb{P}\left\\{\int_{[0,T]}\mathbb{I}(\chi(t)>u)dt>v(u)x\Bigl{\lvert}\sup_{t\in[0,T]}\chi(t)>u\right\\}=\frac{{\mathcal{B}}_{\alpha}(x)}{{\mathcal{B}}_{\alpha}(0)}.$
### 3.3. Sojourns of stationary reflected fractional Brownian motion with
drift
Consider a stationary reflected fractional Brownian motion with drift
$Q(t),t\geq 0$, i.e.,
$\displaystyle Q(t)\coloneqq\sup_{s\geq
t}\left(B_{\alpha}(s)-B_{\alpha}(t)-c(s-t)\right),$
where $B_{\alpha}$ is an fBm with Hurst parameter $\alpha/2\in(0,1)$ and
$c\in(0,\infty)$. Motivated by some applications of $Q(t)$ to queueing models,
the seminal paper [21] studied the tail asymptotics of $Q(0)$. Later on, [27]
considered the tail asymptotics of the supremum of $Q(t)$ over a time horizon.
Recently, the findings of Piterbarg have been extended to Gaussian processes
with stationary increments [12]. We consider next the case of fBm and note
that a more general case of Gaussian processes with stationary increments can
be also dealt with using results from [12]. In the following we consider
$E_{u}=[0,T_{u}]$, where $T_{u}$ is a non-negative function of $u>0$.
###### Proposition 3.4.
Let
$v(u)=u^{\frac{2(\alpha-1)}{\alpha}}\left(\frac{\sqrt{2}(\tau^{*})^{\alpha}}{1+c\tau^{*}}\right)^{2/\alpha}$
with $\tau^{*}=\frac{\alpha}{c(2-\alpha)}$ and $\alpha\in(0,2)$.
i) If $\lim_{u\to\infty}\frac{T_{u}}{v(u)}=T\in(0,\infty)$, then for $T>x\geq
0$
$\lim_{u\rightarrow\infty}\mathbb{P}\left\\{\int_{[0,T_{u}]}\mathbb{I}(Q(t)>u)dt>v(u)x\Bigl{\lvert}\sup_{t\in[0,T_{u}]}Q(t)>u\right\\}=\frac{\mathcal{B}_{\alpha}(x,[0,T])}{\mathcal{B}_{\alpha}(0,[0,T])}.$
ii) If $\lim_{u\to\infty}\frac{T_{u}}{v(u)}=\infty$ and $T_{u}<e^{\beta
u^{2-\alpha}}$ with
$\beta\in\left(0,\left(\frac{1+c\tau^{*}}{\sqrt{2}(\tau^{*})^{\alpha/2}}\right)^{2}\right)$,
then for all $x\geq 0$
$\lim_{u\rightarrow\infty}\mathbb{P}\left\\{\int_{[0,T_{u}]}\mathbb{I}(Q(t)>u)dt>v(u)x\Bigl{\lvert}\sup_{t\in[0,T_{u}]}Q(t)>u\right\\}=\frac{{\mathcal{B}}_{\alpha}(x)}{{\mathcal{B}}_{\alpha}(0)}.$
###### Remark 3.5.
1) Note that $\lim_{u\rightarrow\infty}v(u)=\infty$ for $\alpha>1$, and
$\lim_{u\rightarrow\infty}v(u)=0$ for $\alpha<1$.
2) Conclusion in i) of Proposition 3.4 still holds for $x>T$ since both sides
in the equality of i) are $0$. However, it becomes tricky for the case $T=x$.
We consider two special cases for $T=x$. If $T=x$ and $T_{u}\leq xv(u)$ for
$u$ sufficiently large, then both sides in the equality of i) are $0$. If
$T=x$ and $T_{u}>xv(u)$ for sufficiently large $u$, we get, as $u\to\infty$
$\mathbb{P}\left\\{\int_{[0,T_{u}]}\mathbb{I}(Q(t)>u)dt>v(u)x\Bigl{\lvert}\sup_{t\in[0,T_{u}]}Q(t)>u\right\\}\sim\frac{\mathbb{P}\left\\{\inf_{t\in[0,T_{u}]}Q(t)>u\right\\}}{\mathbb{P}\left\\{\sup_{t\in[0,T_{u}]}Q(t)>u\right\\}}.$
Combining the above two cases for $T=x$, we conclude that the limit for $T=x$
generally does not exist.
## 4\. Auxiliary lemmas
In this section we collect some lemmas that play important, although mostly
technical role in the proofs of results given in Sections 1-3. Their proofs
are deferred to Section 6. We begin with a lemma which is an extension of
Theorem 2.1 from [10]. Suppose that for a compact $d-$dimensional
hyperrectangle $K\subset\mathbb{R}^{d}$ we have
$I_{k}(u,n)=\\{t_{u,n,k}+(v_{1}(u)t_{{1}},\dots v_{d}(u)t_{d}):\,t\in K\\},$
where $v_{i}(u)>0$, $i=1,\ldots,d$ and
$t=(t_{1},\ldots,t_{d})\in\mathbb{R}^{d}$. Then, by transforming time, we have
$\displaystyle\mathbb{P}\left(Vol(\\{t\in I_{k}(u,n):X(t)>u\\})>v(u)z\right)$
$\displaystyle=$
$\displaystyle\mathbb{P}\left(\int_{I_{k}(u,n)}\mathbb{I}(X(t)>u)dt>v(u)z\right)$
$\displaystyle=$
$\displaystyle\mathbb{P}\left(\int_{K}\mathbb{I}(X(t_{u,n,k}+(v_{1}(u)t_{{1}},\dots
v_{d}(u)t_{d}))>u)dt>z\right),$
where $v(u)=\prod_{i=1}^{d}v_{i}(u)$.
Motivated by these calculations, we consider next $\xi_{u,j}(t),t\in E_{1},\
j\in S_{u},\ {u\geq 0}$ a family of centered GRF’s with continuous sample
paths and variance function $\sigma_{u,j}^{2}$.
Suppose in the following that $S_{u}$ is a countable set for all $u$ large.
For simplicity in the following we assume that ${0}\in E_{1}$. For a random
variable $Z$, we set $\overline{Z}=\frac{Z}{\sqrt{Var(Z)}}$ if $Var(Z)>0$.
We introduce next three assumptions:
* C0:
$\\{g_{u,j},j\in S_{u}\\}$ is a sequence of deterministic functions of $u$
satisfying
$\displaystyle\lim_{u\to\infty}\inf_{j\in S_{u}}g_{u,j}=\infty.$
* C1:
$Var(\xi_{u,j}({0}))=1$ for all large $u$ and any $j\in S_{u}$ and there
exists some bounded continuous function $h$ on $E_{1}$ such that
$\displaystyle\lim_{u\to\infty}\sup_{{s}\in E_{1},j\in
S_{u}}\left|g_{u,j}^{2}\left(1-\sigma_{u,j}({s})\right)-h({s})\right|=0.$
* C2:
There exists a centered GRF $\zeta({s}),{s}\in\mathbb{R}^{d}$ with a.s.
continuous sample paths such that
(13) $\displaystyle\lim_{u\to\infty}\sup_{s,s^{\prime}\in E_{1},j\in
S_{u}}\left|g_{u,j}^{2}\big{(}Var(\overline{\xi}_{u,j}({s})-\overline{\xi}_{u,j}({s}^{\prime}))\big{)}-2Var(\zeta(s)-\zeta(s^{\prime}))\right|=0.$
* C3:
There exist positive constants $C,\nu,u_{0}$ such that
$\displaystyle\sup_{j\in
S_{u}}g_{u,j}^{2}Var(\overline{\xi}_{u,j}({s})-\overline{\xi}_{u,j}({s}^{\prime}))\leq
C\lVert{s}-{s}^{\prime}\rVert^{\nu}$
holds for all ${s},{s}^{\prime}\in E_{1},u\geq u_{0}$.
Denote by $C(E_{i}),i=1,2$ the Banach space of all continuous functions
$f:E_{i}\mapsto\mathbb{R}$, with $E_{i}\subset\mathbb{R}^{d_{i}},d_{i}\geq
1,i=1,2$ being compact rectangles equipped with the sup-norm.
Let $\Gamma:C(E_{1})\rightarrow C(E_{2})$ be a continuous functional
satisfying
F1: For any $f\in C(E_{1})$, and $a>0,b\in\mathbb{R}$,
$\Gamma(af+b)=a\Gamma(f)+b$;
F2: There exists $c>0$ such that
$\sup_{t\in E_{2}}\Gamma(f)(t)\leq c\sup_{s\in E_{1}}f(s),\ \ \forall f\in
C(E_{1}).$
Hereafter, $Q_{i},i\in\mathbb{N}$ are some positive constants which might be
different from line to line and $f(u,n)\sim g(u),u\to\infty,n\to\infty$ means
that
$\lim_{n\to\infty}\lim_{u\to\infty}\frac{f(u,n)}{g(u)}=1.$
###### Lemma 4.1.
Let $\\{\xi_{u,j}({s}),{s}\in E_{1},j\in S_{u},{u\geq 0}\\}$ be a family of
centered GRF’s defined as above satisfying C0-C3 and let $\Gamma$ satisfy
F1-F2. Let $\eta$ be a positive $\sigma$-finite measure on $E_{2}$ being
equivalent with the Lebesgues measure on $E_{2}$. If for all large $u$ and all
$j\in S_{u}$
$\mathbb{P}\left\\{\sup_{t\in E_{2}}\Gamma(\xi_{u,j})(t)>g_{u,j}\right\\}>0,$
then for all $x\in[0,\eta(E_{2}))$
(14) $\displaystyle\lim_{u\to\infty}\sup_{j\in
S_{u}}\biggl{\lvert}\frac{\mathbb{P}\left\\{\int_{E_{2}}\mathbb{I}\left(\Gamma(\xi_{u,j})({t})>g_{u,j}\right)\eta(dt)>x\right\\}}{\Psi(g_{u,j})}-\mathcal{B}^{\Gamma,h,\eta}_{\zeta}(x,E_{2})\biggr{\rvert}=0,$
where $\Psi$ is the tail of the standard normal distribution and
$\displaystyle\mathcal{B}^{\Gamma,h,\eta}_{\zeta}(x,E_{2})\coloneqq\int_{\mathbb{R}}\mathbb{P}\left\\{\int_{E_{2}}\mathbb{I}\big{(}\Gamma(\sqrt{2}\zeta-
Var(\zeta)-h)(t)+y>0\big{)}\eta(dt)>x\right\\}e^{-y}dy$
and the constant $\mathcal{B}^{\Gamma,h,\eta}_{\zeta}(x,E_{2})$ is continuous
at $x\in{(}0,\eta(E_{2}))$.
###### Lemma 4.2.
Let $x\geq 0$. Then
(i)
$\mathcal{B}_{\alpha_{1},\alpha_{2}}(x)=\lim_{n\rightarrow\infty}\frac{\mathcal{B}_{\alpha_{1},\alpha_{2}}(x,[0,n]^{2})}{n^{2}}\in(0,\infty),$
(ii)
$\lim_{n\rightarrow\infty}\frac{\mathcal{B}_{\alpha_{1},\alpha_{2}}^{a_{1}^{-1}b_{1}|t_{1}|^{\alpha_{1}},0}(x,[-n,n]\times[0,n])}{n}\in(0,\infty),$
(iii)
$\lim_{n\rightarrow\infty}\mathcal{B}_{\alpha_{1},\alpha_{2}}^{a_{1}^{-1}b_{1}|t_{1}|^{\alpha_{1}},a_{2}^{-1}b_{2}|t_{2}|^{\alpha_{2}}}(x,[-n,n]^{2})\in(0,\infty).$
## 5\. Proofs
### 5.1. Proof of Theorem 1.1
Let next $A_{u}(X)\coloneqq\\{t\in E_{u}:X(t)>u\\}.$ For all $x\geq 0$ and all
$u$ positive, since $v(u)$ is non-negative we have
$\displaystyle\pi(u)$ $\displaystyle\coloneqq$
$\displaystyle\mathbb{P}\left\\{Vol(A_{u}(X))>v(u)x\Bigl{\lvert}Vol(A_{u}(X))>0\right\\}$
$\displaystyle=$
$\displaystyle\mathbb{P}\left\\{Vol(A_{u}(X))>v(u)x\Bigl{\lvert}\sup_{t\in
E_{u}}X(t)>u\right\\}$ $\displaystyle=$
$\displaystyle\frac{\mathbb{P}\left\\{\int_{E}\mathbb{I}(X(t)>u)dt>v(u)x\right\\}}{\mathbb{P}\left\\{\sup_{t\in{\color[rgb]{0,0,0}E_{u}}}X(t)>u\right\\}}$
and further for all $n\geq 1$
$\displaystyle\pi(u)$ $\displaystyle\geq$
$\displaystyle\frac{\mathbb{P}\left\\{\int_{E(u,n)}\mathbb{I}(X(t)>u)dt>v(u)x\right\\}}{\mathbb{P}\left\\{\sup_{t\in
E(u,n)}X(t)>u\right\\}+\mathbb{P}\left\\{\sup_{t\in{\color[rgb]{0,0,0}E_{u}}\setminus
E(u,n)}X(t)>u\right\\}},$ $\displaystyle\pi(u)$ $\displaystyle\leq$
$\displaystyle\frac{\mathbb{P}\left\\{\int_{E(u,n)}\mathbb{I}(X(t)>u)dt>v(u)x\right\\}}{\mathbb{P}\left\\{\sup_{t\in
E(u,n)}X(t)>u\right\\}}+\frac{\mathbb{P}\left\\{\sup_{t\in{\color[rgb]{0,0,0}E_{u}}\setminus
E(u,n)}X(t)>u\right\\}}{\mathbb{P}\left\\{\sup_{t\in{\color[rgb]{0,0,0}E(u,n)}}X(t)>u\right\\}}.$
Applying A1, it follows that
$\displaystyle\pi(u)\sim\frac{\mathbb{P}\left\\{\int_{E(u,n)}\mathbb{I}(X(t)>u)dt>v(u)x\right\\}}{\mathbb{P}\left\\{\sup_{t\in
E(u,n)}X(t)>u\right\\}}=:\pi(u,n),\quad
u\rightarrow\infty,n\rightarrow\infty.$
For the case that $\sharp K_{u,n}=1$ for $u$ and $n$ sufficiently large, the
claim can be established straightforwardly by A2. Thus let us suppose that
$\sharp K_{u,n}\geq 2$ for $n$ and $u$ sufficiently large. In order to proceed
we shall apply the standard scheme utilising Bonferroni inequality. Set
therefore
$\Sigma_{u,n}\coloneqq\sum_{k\in K_{u,n}}\mathbb{P}\left\\{\sup_{t\in
I_{k}(u,n)}X(t)>u\right\\},\quad\Sigma\Sigma_{u,n}\coloneqq\sum_{i\neq
j,i,j\in K_{u,n}}\mathbb{P}\left\\{\sup_{t\in I_{i}(u,n)}X(t)>u,\sup_{t\in
I_{j}(u,n))}X(t)>u\right\\}.$
By the Bonferroni inequality
$\displaystyle\Sigma_{u,n}-\Sigma\Sigma_{u,n}\leq\mathbb{P}\left\\{\sup_{t\in
E(u,n)}X(t)>u\right\\}\leq\Sigma_{u,n}.$
The asymptotic behaviour of the probability of interest in the above
inequality can be derived if the following two-step procedure is successful
(which will work in our settings here). First we determine the exact
asymptotics of the upper bound and then in a second step we show that the
correction in the lower bound is asymptotically negligible.
Now we want to apply the same idea for the sojourn functional, here the
analysis is however more involved. Observe first that for any $u>0$
$\displaystyle\mathbb{P}\left\\{\int_{E(u,n)}\mathbb{I}(X(t)>u)dt>v(u)x\right\\}$
$\displaystyle\quad\leq\mathbb{P}\left\\{\sum_{k\in
K_{u,n}}\int_{I_{k}(u,n)}\mathbb{I}(X(t)>u)dt>v(u)x\right\\}$
$\displaystyle\quad\leq\mathbb{P}\left\\{\exists k\in
K_{u,n},\int_{I_{k}(u,n)}\mathbb{I}(X(t)>u)dt>v(u)x\right\\}$
$\displaystyle\quad\quad+\mathbb{P}\left\\{\exists i,j\in K_{u,n},i\neq
j,\int_{I_{i}(u,n)}\mathbb{I}(X(t)>u)dt>0,\int_{I_{j}(u,n)}\mathbb{I}(X(t)>u)dt>0\right\\}$
$\displaystyle\quad\leq\hat{\pi}(u,n)+\Sigma\Sigma_{u,n},$
where
$\hat{\pi}(u,n)=\sum_{k\in
K_{u,n}}\mathbb{P}\left\\{\int_{I_{k}(u,n)}\mathbb{I}(X(t)>u)dt>v(u)x\right\\}.$
Using Bonferroni inequality again we have
$\displaystyle\mathbb{P}\left\\{\int_{E(u,n)}\mathbb{I}(X(t)>u)dt>v(u)x\right\\}$
$\displaystyle\geq$ $\displaystyle\mathbb{P}\left\\{\exists k\in
K_{u,n},\int_{I_{k}(u,n)}\mathbb{I}(X(t)>u)dt>v(u)x\right\\}$
$\displaystyle\geq$ $\displaystyle\hat{\pi}(u,n)-\Sigma\Sigma_{u,n}.$
The sojourn integral can then be approximated by $\hat{\pi}(u,n)$ if we show
the correction in the lower bound is negligible. We have
$\displaystyle\limsup_{u\rightarrow\infty}\pi(u,n)$ $\displaystyle\leq$
$\displaystyle\limsup_{u\rightarrow\infty}\frac{\hat{\pi}(u,n)+\Sigma\Sigma_{u,n}}{\Sigma_{u,n}-\Sigma\Sigma_{u,n}}=\limsup_{u\rightarrow\infty}\frac{\hat{\pi}(u,n)}{\Sigma_{u,n}}\times\frac{1+\limsup_{u\rightarrow\infty}\frac{\Sigma\Sigma_{u,n}}{\hat{\pi}(u,n)}}{1-\limsup_{u\rightarrow\infty}\frac{\Sigma\Sigma_{u,n}}{\Sigma_{u,n}}},$
$\displaystyle\liminf_{u\rightarrow\infty}\pi(u,n)$ $\displaystyle\geq$
$\displaystyle\liminf_{u\rightarrow\infty}\frac{\hat{\pi}(u,n)-\Sigma\Sigma_{u,n}}{\Sigma_{u,n}}=\liminf_{u\rightarrow\infty}\frac{\hat{\pi}(u,n)}{\Sigma_{u,n}}-\limsup_{u\rightarrow\infty}\frac{\Sigma\Sigma_{u,n}}{\Sigma_{u,n}}.$
By (2) in A2 for any $n\geq 1$ and $x\geq 0$
$\displaystyle\limsup_{u\rightarrow\infty}\frac{\hat{\pi}(u,n)}{\Sigma_{u,n}}=\liminf_{u\rightarrow\infty}\frac{\hat{\pi}(u,n)}{\Sigma_{u,n}}=\bar{F}_{n}(x)$
implying
(15)
$\displaystyle\bar{F}_{n}(x)-\limsup_{u\rightarrow\infty}\frac{\Sigma\Sigma_{u,n}}{\Sigma_{u,n}}\leq\liminf_{u\rightarrow\infty}\pi(u,n)\leq\limsup_{u\rightarrow\infty}\pi(u,n)\leq\bar{F}_{n}(x)\times\frac{1+\limsup_{u\rightarrow\infty}\frac{\Sigma\Sigma_{u,n}}{\bar{F}_{n}(x)\Sigma_{u,n}}}{1-\limsup_{u\rightarrow\infty}\frac{\Sigma\Sigma_{u,n}}{\Sigma_{u,n}}}.$
In view of A3, letting $n\rightarrow\infty$ in the above inequalities we have
that for $x\geq 0$
$\lim_{n\rightarrow\infty}\lim_{u\rightarrow\infty}\pi(u,n)=\bar{F}(x)\in{(0,1]}.$
This completes the proof. $\Box$
### 5.2. Proof of Lemma 2.1
By the independence of $W_{\alpha_{i}}$’s for any positive
$n_{1},\ldots,n_{m}$
$\displaystyle\widehat{\mathcal{B}}_{\alpha_{1},\dots,\alpha_{m}}\left(x,\prod_{i=1}^{m}[0,n_{i}]\right)$
$\displaystyle=$
$\displaystyle\mathbb{E}\left\\{\int_{\mathbb{R}}\mathbb{I}(\int_{[0,n_{1}]}\mathbb{I}\left\\{\sup_{t_{i}\in[0,n_{i}],i=2,\dots,m}\sum_{i=1}^{m}W_{\alpha_{i}}(t_{i})>s\right\\}dt_{1}>x)e^{s}ds\right\\}$
$\displaystyle=$
$\displaystyle\mathbb{E}\left\\{e^{\sum_{i=2}^{m}\sup_{t_{i}\in[0,n_{i}]}W_{\alpha_{i}}(t_{i})}\int_{\mathbb{R}}\mathbb{I}(\int_{[0,n_{1}]}\mathbb{I}\left\\{W_{\alpha_{1}}(t_{1})>s\right\\}dt_{1}>x)e^{s}ds\right\\}$
$\displaystyle=$
$\displaystyle\prod_{i=2}^{m}\mathbb{E}\left\\{\sup_{t_{i}\in[0,n_{i}]}e^{W_{\alpha_{i}}(t_{i})}\right\\}\int_{\mathbb{R}}\mathbb{P}\left\\{\int_{[0,n_{1}]}\mathbb{I}\left\\{W_{\alpha_{1}}(t_{1})>s\right\\}dt_{1}>x\right\\}e^{s}ds.$
Hence the claim follows by the definition of Pickands and Berman constants.
$\Box$
### 5.3. Proof of Proposition 3.1
The proof will be established by checking that A1-A3 in Theorem 1.1 are
satisfied. We begin with the introduction of partition
$I_{k_{1},k_{2}}(u,n)=\prod_{i=1}^{2}[a_{i}^{-1/\alpha_{i}}u^{-2/\alpha_{i}}k_{i}n,a_{i}^{-1/\alpha_{i}}u^{-2/\alpha_{i}}(k_{i}+1)n],$
for
$0\leq
k_{i}\leq[T_{i}a_{i}^{1/\alpha_{i}}u^{2/\alpha_{i}}n^{-1}]-1=:N_{i}(u,n),\
i=1,2.$
Let
$K_{u,n}=\\{(k_{1},k_{2}):0\leq k_{1}\leq N_{1}(u,n),0\leq k_{2}\leq
N_{2}(u,n)\\}$
and $E(u,n)=\bigcup_{(k_{1},k_{2})\in K_{u,n}}I_{k_{1},k_{2}}(u,n)$. Then
$E(u,n)\subset E$.
Condition A1. It follows straightforwardly from Lemma 7.1 in [26] that
(16) $\displaystyle\mathbb{P}\left\\{\sup_{t\in
E}X(t)>u\right\\}\sim\sum_{0\leq k_{i}\leq
N_{i}(u,n),i=1,2}\mathbb{P}\left\\{\sup_{t\in
I_{k_{1},k_{2}}(u,{n})}X(t)>u\right\\},\quad u\rightarrow\infty,n\to\infty,$
which implies that condition A1 holds.
Condition A2. Let for $t=(t_{1},t_{2})$
$\xi_{u,n,k_{1},k_{2}}(t)=X(a_{1}^{-1/\alpha_{1}}u^{-2/\alpha_{1}}(k_{1}n+t_{1}),a_{2}^{-1/\alpha_{2}}u^{-2/\alpha_{2}}(k_{2}n+t_{2})),\quad
v(u)=a_{1}^{-1/\alpha_{1}}a_{2}^{-1/\alpha_{2}}u^{-2/\alpha_{1}-2/\alpha_{2}}.$
We derive the uniform asymptotics, as $u\rightarrow\infty$, of
$\mathbb{P}\left\\{Vol(\\{t\in
I_{k_{1},k_{2}}(u,n):X(t)>u\\})>v(u)x\right\\}=\mathbb{P}\left\\{\int_{[0,n]^{2}}\mathbb{I}(\xi_{u,n,k_{1},k_{2}}(t)>u)dt>x\right\\},$
with $x\geq 0$. For this, we check conditions C0-C3 of Lemma 4.1 with
$\Gamma(f)=f,f\in C([0,n]^{2})$. First note that C0-C1 follow trivially with
$h=0$ and $g_{u,j}=u$. Moreover, by (8), we have
$\displaystyle\lim_{u\rightarrow\infty}\sup_{0\leq k_{i}\leq
N_{i}(u,n),i=1,2}\sup_{s,t\in[0,n]^{2}}\left|u^{2}Var(\xi_{u,n,k_{1},k_{2}}(t)-\xi_{u,n,k_{1},k_{2}}(s))-2Var\left(\sum_{i=1}^{2}B_{\alpha_{i}}(t_{i})-\sum_{i=1}^{2}B_{\alpha_{i}}(s_{i})\right)\right|=0,$
with $B_{\alpha_{i}},i=1,2$ being two independent fBms’ with indices
$\alpha_{i}/2$, respectively. This implies that C2 is satisfied with
$\zeta(t)=\sum_{i=1}^{2}B_{\alpha_{i}}(t_{i}).$ Additionally, in light of (8),
we have that
$\sup_{0\leq k_{i}\leq
N_{i}(u,n)+1,i=1,2}u^{2}Var(\xi_{u,n,k_{1},k_{2}}(t)-\xi_{u,n,k_{1},k_{2}}(s))\leq
C\lVert{t}-{s}\rVert^{\min(\alpha_{1},\alpha_{2})},\quad{s},{t}\in[0,n]^{2}.$
This means that C3 holds. Thus, by Lemma 4.1,
(17) $\displaystyle\lim_{u\rightarrow\infty}\sup_{0\leq k_{i}\leq
N_{i}(u,n),i=1,2}\left|\frac{\mathbb{P}\left\\{Vol(\\{t\in
I_{k_{1},k_{2}}(u,n):X(t)>u\\})>v(u)x\right\\}}{\Psi(u)}-\mathcal{B}_{\alpha_{1},\alpha_{2}}(x,[0,n]^{2})\right|=0.$
Therefore, by Lemma 6.1 in [26] we obtain
$\lim_{u\rightarrow\infty}\sup_{0\leq k_{i}\leq
N_{i}(u,n),i=1,2}\left|\frac{\mathbb{P}\left\\{Vol(\\{t\in
I_{k_{1},k_{2}}(u,n):X(t)>u\\})>v(u)x\right\\}}{\mathbb{P}\left\\{\sup_{t\in
I_{k_{1},k_{2}}(u,n)}X(t)>u\right\\}}-\frac{\mathcal{B}_{\alpha_{1},\alpha_{2}}(x,[0,n]^{2})}{\mathcal{B}_{\alpha_{1},\alpha_{2}}(0,[0,n]^{2})}\right|=0.$
Since, by (i) of Lemma 4.2, for any $x\geq 0$ we have
(18)
$\displaystyle\mathcal{B}_{\alpha_{1},\alpha_{2}}(x)=\lim_{n\rightarrow\infty}\frac{\mathcal{B}_{\alpha_{1},\alpha_{2}}(x,[0,n]^{2})}{n^{2}}\in(0,\infty),$
then
(19)
$\displaystyle\frac{\mathcal{B}_{\alpha_{1},\alpha_{2}}(x)}{\mathcal{B}_{\alpha_{1},\alpha_{2}}(0)}=\lim_{n\rightarrow\infty}\frac{\mathcal{B}_{\alpha_{1},\alpha_{2}}(x,[0,n]^{2})}{\mathcal{B}_{\alpha_{1},\alpha_{2}}(0,[0,n]^{2})}\in(0,{1]},\quad
x\geq 0,$
which confirms that A2 holds with
$\bar{F}(x)=\frac{\mathcal{B}_{\alpha_{1},\alpha_{2}}(x)}{\mathcal{B}_{\alpha_{1},\alpha_{2}}(0)}$.
Condition A3. By (7.4) in the proof of Lemma 7.1 in [26], for all large $u$
and $n$
$\displaystyle\sum_{0\leq k_{i},k_{i}^{\prime}\leq
N_{i}(u,n),i=1,2,(k_{1},k_{2})\neq(k_{1}^{\prime},k_{2}^{\prime})}\mathbb{P}\left\\{\sup_{t\in
I_{k_{1},k_{2}}(u,n)}X(t)>u,\sup_{t\in
I_{k_{1}^{\prime},k_{2}^{\prime}}(u,n)}X(t)>u\right\\}\leq\left(\frac{\mathbb{C}_{2}}{\sqrt{n}}+e^{-\mathbb{C}_{1}n^{\mathbb{C}}}\right)\mathbb{P}\left\\{\sup_{t\in
E}X(t)>u\right\\},$
where $\mathbb{C},\mathbb{C}_{1}$ and $\mathbb{C}_{2}$ are some positive
constants, which gives that A3 is satisfied.
This completes the proof. $\Box$
### 5.4. Proof of Proposition 3.2
Without loss of generality, we assume that $t^{*}=(0,0)$. The proof relies on
verification that A1-A3 in Theorem 1.1 are satisfied. We begin by introducing
some notation. Let
(20) $\displaystyle
I_{k_{1},k_{2}}(u,n)=\prod_{i=1}^{2}[k_{i}v_{i}(u)n,(k_{i}+1)v_{i}(u)n],\quad
v_{i}(u)=a_{i}^{-1/\alpha_{i}^{*}}u^{-2/\min(\alpha_{i},\beta_{i})},i=1,2,\quad
v(u)=v_{1}(u)v_{2}(u),$
where
$\alpha_{i}^{*}=\alpha_{i}\mathbb{I}(\alpha_{i}\leq\beta_{i})+\infty\mathbb{I}(\alpha_{i}>\beta_{i})$.
Additionally, let
$e(t)=\frac{1-{\sigma}(t)}{\sum_{i=1}^{2}b_{i}|t_{i}|^{\beta_{i}}}-1,|t|\neq
0,~{}e_{u}=\sup_{0<|t_{i}|<\left(\frac{\ln u}{u}\right)^{2/\beta_{i}}}|e(t)|,$
and set
$N_{i}^{\prime}(u,n)=\left[\frac{(e_{u}^{-1/4}\wedge\ln
u)^{2/\beta_{i}}}{u^{2/\beta_{i}}v_{i}(u)n}\right],\ i=1,2.$
We distinguish different scenarios according to the values of
$\alpha_{i},\beta_{i},i=1,2$.
Case $\alpha_{i}<\beta_{i},i=1,2$. In this scenario
$v_{i}(u)=a_{i}^{-1/\alpha_{i}}u^{-2/\alpha_{i}},i=1,2,\ \
K_{u,n}=\\{(k_{1},k_{2}):0\leq|k_{i}|\leq N_{i}^{\prime}(u,n),i=1,2\\}$
and $E(u,n)=\bigcup_{(k_{1},k_{2})\in K_{u,n}}I_{k_{1},k_{2}}(u,n)$.
Conditions A1 and A3. Following the same reasoning as in the proof of
Proposition 3.1, the validity of conditions A1 and A3 follows
straightforwardly from (34), (40) and (41) in [18].
Condition A2. Let
$\xi_{u,n,k_{1},k_{2}}(t)=\overline{X}(v_{1}(u)(k_{1}n+t_{1}),v_{2}(u)(k_{2}n+t_{2})),$
(21) $\displaystyle u_{n,k_{1},k_{2}}^{-}=u\inf_{t\in
I_{k_{1},k_{2}}(u,n)}\frac{1}{{\sigma(t)}},\quad
u_{n,k_{1},k_{2}}^{+}=u\sup_{t\in I_{k_{1},k_{2}}(u,n)}\frac{1}{{\sigma(t)}}.$
Then
$\displaystyle\mathbb{P}\left\\{Vol(\\{t\in
I_{k_{1},k_{2}}(u,n):X(t)>u\\})\geq
v(u)x\right\\}\leq\mathbb{P}\left\\{\int_{[0,n]^{2}}\mathbb{I}(\xi_{u,n,k_{1},k_{2}}(t)>u_{n,k_{1},k_{2}}^{-})dt>x\right\\},$
$\displaystyle\mathbb{P}\left\\{Vol(\\{t\in
I_{k_{1},k_{2}}(u,n):X(t)>u\\})\geq
v(u)x\right\\}\geq\mathbb{P}\left\\{\int_{[0,n]^{2}}\mathbb{I}(\xi_{u,n,k_{1},k_{2}}(t)>u_{n,k_{1},k_{2}}^{+})dt>x\right\\}.$
In order to derive the uniform asymptotics of the above terms we check
conditions C0-C3 of Lemma 4.1 with $\Gamma(f)=f,$ $f\in C([0,n]^{2})$ for
$\xi_{u,n,k_{1},k_{2}}(t),\ (k_{1},k_{2})\in K_{u,n}.$
Note that C0-C1 holds with $h=0$ and $g_{u,j}=u_{n,k_{1},k_{2}}^{\pm}$. By
(10) and (11), we have
$\displaystyle\lim_{u\rightarrow\infty}\sup_{s,t\in[0,n]^{2},(k_{1},k_{2})\in
K_{u,n}}\left|(u_{n,k_{1},k_{2}}^{\pm})^{2}(Var(\xi_{u,n,k_{1},k_{2}}(t)-\xi_{u,n,k_{1},k_{2}}(s)))-2Var\left(\sum_{i=1}^{2}B_{\alpha_{i}}(t_{i})-\sum_{i=1}^{2}B_{\alpha_{i}}(s_{i})\right)\right|=0,$
where $B_{\alpha_{i}},i=1,2$ are two independent fBm’s with indices
$\alpha_{i},i=1,2$ respectively. This confirms that C2 holds with
$\zeta(t_{1},t_{2})=B_{\alpha_{1}}(t_{1})+B_{\alpha_{2}}(t_{2})$. By (11), we
have
$\sup_{(k_{1},k_{2})\in
K_{u,n}}(u_{n,k_{1},k_{2}}^{\pm})^{2}(Var(\xi_{u,n,k_{1},k_{2}}(t)-\xi_{u,n,k_{1},k_{2}}(s)))\leq
Q||s-t||^{\min(\alpha_{1},\alpha_{2})},\quad s,t\in[0,n]^{2}.$
Thus C3 is satisfied.
Therefore, by Lemma 4.1, we have that for $0\leq x<n^{2}$,
(22) $\displaystyle\lim_{u\rightarrow\infty}\sup_{(k_{1},k_{2})\in
K_{u,n}}\left|\frac{\mathbb{P}\left\\{\int_{[0,n]^{2}}\mathbb{I}(\xi_{u,n,k_{1},k_{2}}(t)>u_{n,k_{1},k_{2}}^{\pm})dt>x\right\\}}{\Psi(u_{n,k_{1},k_{2}}^{\pm})}-\mathcal{B}_{\alpha_{1},\alpha_{2}}(x,[0,n]^{2})\right|=0.$
Since
(23) $\displaystyle\lim_{u\rightarrow\infty}\sup_{(k_{1},k_{2})\in
K_{u,n}}\left|\frac{\Psi(u_{n,k_{1},k_{2}}^{-})}{\Psi(u_{n,k_{1},k_{2}}^{+})}-1\right|=0$
(see Section 6 for the validation of (23)), by (22) we obtain for $0\leq
x<n^{2}$
$\lim_{u\rightarrow\infty}\sup_{(k_{1},k_{2})\in
K_{u,n}}\left|\frac{\mathbb{P}\left\\{Vol(\\{t\in
I_{k_{1},k_{2}}(u,n):X(t)>u\\})\geq
v(u)x\right\\}}{\Psi(u_{n,k_{1},k_{2}}^{-})}-\mathcal{B}_{\alpha_{1},\alpha_{2}}(x,[0,n]^{2})\right|=0.$
Therefore, (2) holds with
$\bar{F}_{n}(x)=\frac{\mathcal{B}_{\alpha_{1},\alpha_{2}}(x,[0,n]^{2})}{\mathcal{B}_{\alpha_{1},\alpha_{2}}(0,[0,n]^{2})},\quad
x\geq 0.$
Finally, by (19), we have that A2 holds. Thus the claim is established with
$\bar{F}(x)=\frac{\mathcal{B}_{\alpha_{1},\alpha_{2}}(x)}{\mathcal{B}_{\alpha_{1},\alpha_{2}}(0)}.$
Case $\alpha_{1}=\beta_{1},\alpha_{2}<\beta_{2}$. In this case
$v_{i}(u)=a_{i}^{-1/\alpha_{i}}u^{-2/\alpha_{i}},i=1,2.$ Let
(24) $\displaystyle\hat{I}_{k_{2}}(u,n)=I_{-1,k_{2}}(u,n)\cup
I_{0,k_{2}}(u,n),\quad E_{1}(u,n)=\bigcup_{k_{2}\in
K_{u,n}}\hat{I}_{k_{2}}(u,n),$
where $K_{u,n}\coloneqq\\{k_{2}\in\mathbb{Z}:|k_{2}|\leq
N_{2}^{\prime}(u,n)\\}$.
Conditions A1 and A3. Analogously to the previous case, conditions A1 and A3
hold with $E(u,n)\coloneqq E_{1}(u,n)$ and
${I}_{k}(u,n)\coloneqq\hat{I}_{k_{2}}(u,n)$, by (34), (46), (48) and (49) of
[18].
Condition A2. Rewrite (10) as
$\frac{1}{\sigma(t)}=\left(1+(1+e_{1}(t_{1}))b_{1}|t_{1}|^{\beta_{1}}\right)\left(1+(1+e_{2}(t_{2}))b_{2}|t_{2}|^{\beta_{2}}\right),$
for some functions $e_{1}(t_{1})$ and $e_{2}(t_{2})$ which satisfy
$\lim_{u\rightarrow\infty}\sup_{t\in E_{1}(u,n)}|e_{i}(t_{i})|=0,\quad i=1,2.$
Let
$\xi_{u,n,k_{2}}(t)=\frac{\overline{X}(v_{1}(u)t_{1},v_{2}(u)(k_{2}n+t_{2}))}{1+b_{1}|v_{1}(u)t_{1}|^{\beta_{1}}(1+e_{1}(v_{1}(u)t_{1}))},\quad
v(u)=a_{1}^{-1/\alpha_{1}}a_{2}^{-1/\alpha_{2}}u^{-2/\alpha_{1}-2/\alpha_{2}},$
$u_{k_{2},n}^{-}=u\inf_{t\in\hat{I}_{k_{2}}(u,n)}(1+b_{2}|t_{2}|^{\beta_{2}}(1+e_{2}(t_{2}))),\quad
u_{k_{2},n}^{+}=u\sup_{t\in\hat{I}_{k_{2}}(u,n)}(1+b_{2}|t_{2}|^{\beta_{2}}(1+e_{2}(t_{2}))).$
Then it follows that
$\displaystyle\mathbb{P}\left\\{Vol(\\{t\in\hat{I}_{k_{2}}(u,n):X(t)>u\\})>v(u)x\right\\}\leq\mathbb{P}\left\\{\int_{[-n,n]\times[0,n]}\mathbb{I}(\xi_{u,n,k_{2}}(t)>u_{k_{2},n}^{-})dt>x\right\\},$
$\displaystyle\mathbb{P}\left\\{Vol(\\{t\in\hat{I}_{k_{2}}(u,n):X(t)>u\\})>v(u)x\right\\}\geq\mathbb{P}\left\\{\int_{[-n,n]\times[0,n]}\mathbb{I}(\xi_{u,n,k_{2}}(t)>u_{k_{2},n}^{+})dt>x\right\\}.$
Straightforward application of Lemma 4.1 with $\Gamma(f)=f,$ $f\in
C([-n,n]\times[0,n])$ and $h(t)=a_{1}^{-1}b_{1}|t_{1}|^{\alpha_{1}}$ in C1,
gives that for $0\leq x<2n^{2}$,
$\displaystyle\lim_{u\rightarrow\infty}\sup_{k_{2}\in
K_{u,n}}\left|\frac{\mathbb{P}\left\\{\int_{[-n,n]\times[0,n]}\mathbb{I}(\xi_{u,n,k_{2}}(t)>u_{k_{2},n}^{\pm})dt>x\right\\}}{\Psi(u_{k_{2},n}^{\pm})}-\mathcal{B}_{\alpha_{1},\alpha_{2}}^{a_{1}^{-1}b_{1}|t_{1}|^{\alpha_{1}},0}(x,[-n,n]\times[0,n])\right|=0.$
Similarly to (23), we have
(25) $\displaystyle\lim_{u\rightarrow\infty}\sup_{k_{2}\in
K_{u,n}}\left|\frac{\Psi(u_{k_{2},n}^{-})}{\Psi(u_{k_{2},n}^{+})}-1\right|=0.$
Consequently, for $0\leq x<2n^{2}$
$\displaystyle\lim_{u\rightarrow\infty}\sup_{k_{2}\in
K_{u,n}}\left|\frac{\mathbb{P}\left\\{Vol(\\{t\in\hat{I}_{k_{2}}(u,n):X(t)>u\\})>v(u)x\right\\}}{\Psi(u_{k_{2},n}^{-})}-\mathcal{B}_{\alpha_{1},\alpha_{2}}^{a_{1}^{-1}b_{1}|t_{1}|^{\alpha_{1}},0}(x,[-n,n]\times[0,n])\right|=0.$
Thus (2) holds with
$\bar{F}_{n}(x)=\frac{\mathcal{B}_{\alpha_{1},\alpha_{2}}^{a_{1}^{-1}b_{1}|t_{1}|^{\alpha_{1}},0}(x,[-n,n]\times[0,n])}{\mathcal{B}_{\alpha_{1},\alpha_{2}}^{a_{1}^{-1}b_{1}|t_{1}|^{\alpha_{1}},0}(0,[-n,n]\times[0,n])}.$
By (ii) of Lemma 4.2 it follows that
(27)
$\lim_{n\to\infty}\bar{F}_{n}(x)=\frac{\mathcal{B}_{\alpha_{1},\alpha_{2}}^{a_{1}^{-1}b_{1}|t_{1}|^{\alpha_{1}},0}(x)}{\mathcal{B}_{\alpha_{1},\alpha_{2}}^{a_{1}^{-1}b_{1}|t_{1}|^{\alpha_{1}},0}(0)}\in(0,{1]},$
which confirms that A2 holds. Thus, applying Theorem 1.1, we establish the
claim with
$\bar{F}(x)=\frac{\mathcal{B}_{\alpha_{1},\alpha_{2}}^{a_{1}^{-1}b_{1}|t_{1}|^{\alpha_{1}},0}(x)}{\mathcal{B}_{\alpha_{1},\alpha_{2}}^{a_{1}^{-1}b_{1}|t_{1}|^{\alpha_{1}},0}(0)}.$
Case $\alpha_{1}=\beta_{1},\alpha_{2}=\beta_{2}$. In this case we have
$v_{i}(u)=a_{i}^{-1/\alpha_{i}}u^{-2/\alpha_{i}},i=1,2.$ Let
(28) $\displaystyle
E(u,n)\coloneqq\hat{I}(u,n)\coloneqq\bigcup_{i,j=-1,0}I_{i,j}(u,n).$
Conditions A1 and A3. It follows from (34) and (52) in the proof of theorem
3.1 of [18] that A1 holds. Since we take only one interval $I_{1}(u,n)$,
condition A3 is not applicable to this case.
Condition A2. Let
$\displaystyle\xi_{u,n}(t)=X(v_{1}(u)t_{1},v_{2}(u)t_{2})),\quad
v(u)=a_{1}^{-1/\alpha_{1}}a_{2}^{-1/\alpha_{2}}u^{-2/\alpha_{1}}u^{-2/\alpha_{2}}.$
Then
$\mathbb{P}\left\\{Vol(\\{t\in\hat{I}(u,n):X(t)>u\\})\geq
v(u)x\right\\}=\mathbb{P}\left\\{\int_{[-n,n]^{2}}\mathbb{I}(\xi_{u,n}(t)>u)dt>x\right\\}.$
In order to derive the asymptotics of the above term, similarly to the
previous cases, we observe that C1 in Lemma 4.1 holds with
$h(t)=a_{1}^{-1}b_{1}|t_{1}|^{\alpha_{1}}+a_{2}^{-1}b_{2}|t_{2}|^{\alpha_{2}}$
while C2 and C3 have been checked in the case of $\alpha_{i}<\beta_{i},i=1,2$.
Hence we have
$\displaystyle\lim_{u\rightarrow\infty}\left|\frac{\mathbb{P}\left\\{\int_{[-n,n]^{2}}\mathbb{I}(\xi_{u,n}(t)>u)dt>x\right\\}}{\Psi(u)}-\mathcal{B}_{\alpha_{1},\alpha_{2}}^{a_{1}^{-1}b_{1}|t_{1}|^{\alpha_{1}},a_{2}^{-1}b_{2}|t_{2}|^{\alpha_{2}}}(x,[-n,n]^{2})\right|=0.$
Combining the above with the fact that, by (iii) of Lemma 4.2,
$\displaystyle\lim_{n\rightarrow\infty}\mathcal{B}_{\alpha_{1},\alpha_{2}}^{a_{1}^{-1}b_{1}|t_{1}|^{\alpha_{1}},a_{2}^{-1}b_{2}|t_{2}|^{\alpha_{2}}}(x,[-n,n]^{2})\in(0,\infty)$
we conclude that A2 holds with
$\bar{F}(x)=\frac{\mathcal{B}_{\alpha_{1},\alpha_{2}}^{a_{1}^{-1}b_{1}|t_{1}|^{\alpha_{1}},a_{2}^{-1}b_{2}|t_{2}|^{\alpha_{2}}}(x)}{\mathcal{B}_{\alpha_{1},\alpha_{2}}^{a_{1}^{-1}b_{1}|t_{1}|^{\alpha_{1}},a_{2}^{-1}b_{2}|t_{2}|^{\alpha_{2}}}(0)}\in(0,{1]}.$
Hence we establish the claim.
For the cases $\alpha_{1}>\beta_{1},\alpha_{2}=\beta_{2}$, and
$\alpha_{1}>\beta_{1},\alpha_{2}>\beta_{2}$, we can establish the claim
similarly to the case of $\alpha_{1}=\beta_{1},\alpha_{2}=\beta_{2}$. For the
case $\alpha_{1}>\beta_{1},\alpha_{2}<\beta_{2}$, the proof is similar to the
case of $\alpha_{1}=\beta_{1}$, $\alpha_{2}<\beta_{2}$. This completes the
proof. $\Box$
### 5.5. Proof of Proposition 3.3
In order to apply Theorem 1.1, we introduce some useful notation. Let
$\displaystyle I_{k}(u,n)=[kv(u)n,(k+1)v(u)n],\quad
N(u,n)=\left[\frac{T}{v(u)n}\right]-1,$
and $E(u,n)=\bigcup_{k\in K_{u,n}}I_{k}(u,n),$ with
$K_{u,n}=\\{k\in\mathbb{N}:0\leq k\leq N(u,n)\\}$ and
$v(u)=a^{-1/\alpha}u^{-2/\alpha}.$ We denote by
$Z(t,\theta)=\sum_{i=1}^{m}X_{i}(t)v_{i}(\theta),\quad
A=[0,\pi]^{m-2}\times[0,2\pi),$
where $\theta=(\theta_{1},\ldots,\theta_{m-1})$ and
$v_{1}(\theta)=\cos\theta_{1},v_{2}(\theta)=\sin\theta_{1}\cos\theta_{2},v_{3}(\theta)=\sin\theta_{1}\sin\theta_{2}\cos\theta_{3},\dots,v_{m-1}(\theta)=(\prod_{i=1}^{m-2}\sin\theta_{i})\cos\theta_{m-1},v_{m}(\theta)=\prod_{i=1}^{m-1}\sin\theta_{i}.$
In this proof, we will use that
$\chi(t)=\sup_{\theta\in A}Z(t,\theta).$
We split the set $A$ into (setting $k=(k_{1},\dots,k_{m-1})$)
$A=\bigcup_{k\in\Lambda}A_{k},\quad\Lambda=\\{(k_{1},\dots,k_{m-1}):1\leq
k_{i}\leq L,1\leq i\leq m-2,1\leq k_{m-1}\leq 2L\\},$
where
$A_{k}=\prod_{i=1}^{m-1}\left[\frac{(k_{i}-1)\pi}{L},\frac{k_{i}\pi}{L}\right],\quad
k_{m-1}\leq 2L-1,$
$A_{k_{1},\dots,k_{m-2},2L}=\left(\prod_{i=1}^{m-2}\left[\frac{(k_{i}-1)\pi}{L},\frac{k_{i}\pi}{L}\right]\right)\times\left[2\pi-\frac{\pi}{L},2\pi\right),$
and $L$ is a positive integer. Moreover, let
(29) $\displaystyle\pi_{1}(u)$ $\displaystyle\coloneqq$
$\displaystyle\sum_{k\neq
k^{\prime},k,k^{\prime}\in\Lambda}\mathbb{P}\left\\{\sup_{t\in[0,v(u)n],\theta\in
A_{k}}Z(t,\theta)>u,\sup_{t\in[0,v(u)n],\theta\in
A_{k^{\prime}}}Z(t,\theta)>u\right\\},$ $\displaystyle\Sigma\Sigma_{u,n}$
$\displaystyle\coloneqq$ $\displaystyle\sum_{0\leq k_{1}<k_{2}\leq
N(u,n)}\mathbb{P}\left\\{\sup_{t\in I_{k_{1}}(u,n)}\chi(t)>u,\sup_{t\in
I_{k_{2}}(u,n)}\chi(t)>u\right\\}$ $\displaystyle=$ $\displaystyle\sum_{0\leq
k_{1}<k_{2}\leq N(u,n)}\mathbb{P}\left\\{\sup_{(t,\theta)\in
I_{k_{1}}(u,n)\times A}Z(t,\theta)>u,\sup_{(t,\theta)\in I_{k_{2}}(u,n)\times
A}Z(t,\theta)>u\right\\}$ $\displaystyle\leq$ $\displaystyle\sum_{0\leq
k_{1}<k_{2}\leq N(u,n),i,j\in\Lambda}\mathbb{P}\left\\{\sup_{(t,\theta)\in
I_{k_{1}}(u,n)\times A_{i}}Z(t,\theta)>u,\sup_{(t,\theta)\in
I_{k_{2}}(u,n)\times A_{j}}Z(t,\theta)>u\right\\}.$
Denote by (with $k=(k_{1},\dots,k_{m-1}),l=(l_{1},\dots,l_{m-1})$)
$J_{k,l}(u)=\prod_{i=1}^{m-1}\left[\frac{(k_{i}-1)\pi}{L}+l_{i}u^{-1}n_{1},\frac{(k_{i}-1)\pi}{L}+(l_{i}+1)u^{-1}n_{1}\right],\quad\Lambda_{1}(u)=\left\\{l:0\leq
l_{i}\leq\left[\frac{\pi u}{Ln_{1}}\right],1\leq i\leq m-1\right\\},$
and let
(30) $\displaystyle p_{k}^{*}(u)=\sum_{l,l^{\prime}\in\Lambda_{1}(u),l\neq
l^{\prime}}\mathbb{P}\left\\{\sup_{t\in[0,v(u)n],\theta\in
J_{k,l}(u)}Z(t,\theta)>u,\sup_{t\in[0,v(u)n],\theta\in
J_{k,l^{\prime}}(u)}Z(t,\theta)>u\right\\}.$
Conditions A1 and A3. Condition A1 follows from Corollary 7.3 in [26] while A3
can be deduced from equations (7.4), (7.6) and (7.18) in the proofs of Lemma
7.1 and Theorem 7.1 in [26].
Condition A2. Let us put
$\displaystyle\pi(n,u)\coloneqq\mathbb{P}\left\\{\int_{[0,v(u)n]}\mathbb{I}(\chi(t)>u)dt>v(u)x\right\\}=\mathbb{P}\left\\{\int_{[0,v(u)n]}\mathbb{I}\left(\sup_{\theta\in
A}Z(t,\theta)>u\right)dt>v(u)x\right\\}.$
To verify A2, by stationarity we have to find the asymptotics of $\pi{(n,u)}$
as $u\to\infty$, which is given in the following lemma.
###### Lemma 5.1.
For $n>x$
$\displaystyle\pi(n,u)\sim\frac{\widehat{\mathcal{B}}_{\alpha,2,\dots,2}(x,n)}{\widehat{\mathcal{B}}_{\alpha,2,\dots,2}(0,n)}\mathbb{P}\left\\{\sup_{[0,v(u)n]}\chi(t)>u\right\\},\quad
u\rightarrow\infty.$
Proof of Lemma 5.1. Let $D_{k}=\\{t\in[0,v(u)n]:\sup_{\theta\in
A_{k}}Z(t,\theta)>u\\}.$ Then we have
$\displaystyle\int_{[0,v(u)n]}\mathbb{I}\left(\sup_{\theta\in
A}Z(t,\theta)>u\right)dt$ $\displaystyle=$
$\displaystyle\int_{[0,v(u)n]}\mathbb{I}_{\bigcup_{k\in\Lambda}D_{k}}(t)dt$
$\displaystyle\leq$
$\displaystyle\sum_{k\in\Lambda}\int_{[0,v(u)n]}\mathbb{I}_{D_{k}}(t)dt$
and
$\displaystyle\int_{[0,v(u)n]}\mathbb{I}\left(\sup_{\theta\in
A}Z(t,\theta)>u\right)dt$ $\displaystyle\geq$
$\displaystyle\sum_{k\in\Lambda}\int_{[0,v(u)n]}\mathbb{I}_{D_{k}}(t)dt-\sum_{k\neq
k^{\prime},k,k^{\prime}\in\Lambda}\int_{[0,v(u)n]}\mathbb{I}_{D_{k}\bigcap
D_{k^{\prime}}}(t)dt.$
Note that
$\displaystyle\pi(n,u)$ $\displaystyle\geq$
$\displaystyle\mathbb{P}\left(\sum_{k\in\Lambda}\int_{[0,v(u)n]}\mathbb{I}_{D_{k}}(t)dt-\sum_{k\neq
k^{\prime},k,k^{\prime}\in\Lambda}\int_{[0,v(u)n]}\mathbb{I}_{D_{k}\bigcap
D_{k^{\prime}}}(t)dt>v(u)x\right)$ $\displaystyle\geq$
$\displaystyle\mathbb{P}\left(\sum_{k\in\Lambda}\int_{[0,v(u)n]}\mathbb{I}_{D_{k}}(t)dt>v(u)(x+\epsilon),\sum_{k\neq
k^{\prime},k,k^{\prime}\in\Lambda}\int_{[0,v(u)n]}\mathbb{I}_{D_{k}\bigcap
D_{k^{\prime}}}(t)dt\leq v(u)\epsilon\right)$ $\displaystyle\geq$
$\displaystyle\mathbb{P}\left(\sum_{k\in\Lambda}\int_{[0,v(u)n]}\mathbb{I}_{D_{k}}(t)dt>v(u)(x+\epsilon)\right)-\mathbb{P}\left(\sum_{k\neq
k^{\prime},k,k^{\prime}\in\Lambda}\int_{[0,v(u)n]}\mathbb{I}_{D_{k}\bigcap
D_{k^{\prime}}}(t)dt>v(u)\epsilon\right)$ $\displaystyle\geq$
$\displaystyle\mathbb{P}\left(\sum_{k\in\Lambda}\int_{[0,v(u)n]}\mathbb{I}_{D_{k}}(t)dt>v(u)(x+\epsilon)\right)-\pi_{1}(u)$
$\displaystyle\geq$
$\displaystyle\sum_{k\in\Lambda^{*}}p_{k}(x+\epsilon,u)-2\pi_{1}(u),$
where $\epsilon>0$ and $\pi_{1}(u)$ is given in (29) and
$\displaystyle p_{k}(x,u)$ $\displaystyle=$
$\displaystyle\mathbb{P}\left\\{\int_{[0,v(u)n]}\mathbb{I}\left(\sup_{\theta\in
A_{k}}Z(t,\theta)>u\right)dt>v(u)x\right\\},$ $\displaystyle\Lambda^{*}$
$\displaystyle=$ $\displaystyle\\{k\in\Lambda,1<k_{i}<L,1\leq i\leq
m-2,k_{m-1}\neq 1,L,2L\\}.$
Similarly we get
$\pi(n,u)\leq\sum_{k\in\Lambda}p_{k}(x,u)+\pi_{1}(u).$
Hence
(31)
$\displaystyle\sum_{k\in\Lambda^{*}}p_{k}(x+\epsilon,u)-2\pi_{1}(u)\leq\pi(n,u)\leq\sum_{k\in\Lambda}p_{k}(x,u)+\pi_{1}(u).$
$\diamond$ Upper bound for $p_{k}(x,u)$. A direct calculations show
$\displaystyle Var(Z(t,\theta))$ $\displaystyle=$ $\displaystyle 1,$
$\displaystyle Corr(Z(t,\theta),Z(t^{\prime},\theta^{\prime}))$
$\displaystyle=$ $\displaystyle
Corr(X(t),X(t^{\prime}))\left(\cos(\theta_{1}-\theta_{1}^{\prime})-\sin\theta_{1}\sin\theta_{1}^{\prime}(1-\cos(\theta_{2}-\theta_{2}^{\prime}))\right.$
$\displaystyle\quad\left.-\dots-\left(\prod_{i=1}^{m-2}\sin\theta_{i}\sin\theta_{i}^{\prime}\right)\left(1-\cos(\theta_{m-1}-\theta_{m-1}^{\prime})\right)\right).$
Hence
(32) $\displaystyle 1-Corr(Z(t,\theta),Z(t^{\prime},\theta^{\prime}))$
$\displaystyle\sim$ $\displaystyle
a|t-t^{\prime}|^{\alpha}+\frac{1}{2}(\theta_{1}-\theta_{1}^{\prime})^{2}+\frac{\sin^{2}\theta_{1}}{2}(\theta_{2}-\theta_{2}^{\prime})^{2}$
$\displaystyle\quad+\frac{1}{2}\left(\prod_{i=1}^{m-2}\sin^{2}\theta_{i}\right)(\theta_{m-1}-\theta_{m-1}^{\prime})^{2},\quad|t-t^{\prime}|\rightarrow
0,||\theta-\theta^{\prime}||\rightarrow 0.$
We have
(33) $\displaystyle
p_{k}(x,u)\leq\sum_{l\in\Lambda_{1}(u)}\mathbb{P}\left\\{\int_{[0,v(u)n]}\mathbb{I}\left(\sup_{\theta\in
J_{k,l}(u)}Z(t,\theta)>u\right)dt>v(u)x\right\\}+p_{k}^{*}(u),$
where $p_{k}^{*}(u)$ is given in (30). Let
$Z_{u,k,l}(t,\theta)=Z\left(v(u)t,\frac{(k_{1}-1)\pi}{L}+l_{1}u^{-1}n_{1}+u^{-1}c_{1}^{-1}(\theta_{k,l}(u))\theta_{1},\dots,\frac{(k_{m-1}-1)\pi}{L}+l_{m-1}u^{-1}n_{1}+u^{-1}c_{m-1}^{-1}(\theta_{k,l}(u))\theta_{m-1}\right),$
and $G_{l}=\prod_{i=1}^{m-1}[0,c_{i}(\theta_{k,l}(u))n_{1}],$ where
$c_{k}(\theta)=2^{-1/2}\prod_{i=1}^{k-1}|\sin\theta_{i}|,2\leq k\leq m-1,\quad
c_{1}(\theta)=2^{-1/2},\quad\theta_{k,l}(u)=\left(\frac{(k_{1}-1)\pi}{L}+l_{1}u^{-1}n_{1},\dots,\frac{(k_{m-1}-1)\pi}{L}+l_{m-1}u^{-1}n_{1}\right).$
Noting that
$G_{l}=\prod_{i=1}^{m-1}[0,c_{i}(\theta_{l}(u))n_{1}]\subset\prod_{i=1}^{m-1}[0,c_{k,i}^{+}n_{1}]=:G_{k}^{+},\quad
c_{k,i}^{+}=\sup_{\theta\in A_{k}}c_{i}(\theta),$
we have
$\displaystyle\mathbb{P}\left\\{\int_{[0,v(u)n]}\mathbb{I}\left(\sup_{\theta\in
J_{k,l}(u)}Z(t,\theta)>u\right)dt>v(u)x\right\\}$ $\displaystyle=$
$\displaystyle\mathbb{P}\left\\{\int_{[0,n]}\mathbb{I}\left(\sup_{\theta\in
G_{l}}Z_{u,k,l}(t,\theta)>u\right)dt>v(u)x\right\\}$ $\displaystyle\leq$
$\displaystyle\mathbb{P}\left\\{\int_{[0,n]}\mathbb{I}\left(\sup_{\theta\in
G_{k}^{+}}Z_{u,k,l}(t,\theta)>u\right)dt>v(u)x\right\\}.$
A straightforward application of Lemma 4.1 for $\Gamma:C([0,n]\times
G_{k}^{+})\rightarrow C([0,n])$ defined by $\Gamma(f)=\sup_{\theta\in
G_{k}^{+}}f(t,\theta),\quad f\in C([0,n]\times G_{k}^{+})$, where $h=0$ in C1
and $\zeta(t,\theta)=B_{\alpha}(t)+\sum_{i=1}^{m-1}N_{i}\theta_{i},$ with
$N_{i},i=1,\dots,m-1$ being independent standard normal random variables
independent of $B_{\alpha}$, implies that for all $x\geq 0$ we have
(34)
$\displaystyle\lim_{u\rightarrow\infty}\sup_{l\in\Lambda_{1}(u)}\left|\frac{\mathbb{P}\left\\{\int_{[0,n]}\mathbb{I}\left(\sup_{\theta\in
G_{l}^{+}}Z_{u,k,l}(t,\theta)>u\right)dt>v(u)x\right\\}}{\Psi(u)}-\widehat{\mathcal{B}}_{\alpha,2,\dots,2}(x,[0,n]\times
G_{l}^{+})\right|=0.$
By (7.18) in the proof of Theorem 7.1 in [26], we have
(35) $\displaystyle p_{k}^{*}(u)=o\left(u^{m-1}\Psi(u)\right),\quad
u\rightarrow\infty,n_{1}\rightarrow\infty.$
Hence, by (33)-(35) and using Lemma 2.1 we have
$\displaystyle p_{k}(x,u)$ $\displaystyle\leq$
$\displaystyle\limsup_{n_{1}\rightarrow\infty}\frac{\widehat{\mathcal{B}}_{\alpha,2,\dots,2}(x,[0,n]\times
G_{l}^{+})}{(n_{1})^{m-1}}\left(\frac{\pi}{L}\right)^{m-1}u^{m-1}\Psi(u)$
$\displaystyle\leq$
$\displaystyle\widehat{\mathcal{B}}_{\alpha,2,\dots,2}(x,n)\prod_{i=1}^{m-1}c_{k,i}^{+}\left(\frac{\pi}{L}\right)^{m-1}u^{m-1}\Psi(u),\quad
u\rightarrow\infty,n_{1}\rightarrow\infty.$
$\diamond$ Lower bound for $p_{k}(x,u)$. By (31), we have that for
$\epsilon>0$
$\displaystyle p_{k}(x,u)$ $\displaystyle\geq$
$\displaystyle\sum_{l\in\Lambda_{2}(u)}\mathbb{P}\left\\{\int_{[0,v(u)n]}\mathbb{I}\left(\sup_{\theta\in
J_{{k,l}}(u)}Z(t,\theta)>u\right)dt>v(u)(x+\epsilon)\right\\}-2p_{k}^{*}(u)$
$\displaystyle\geq$
$\displaystyle\sum_{l\in\Lambda_{2}(u)}\mathbb{P}\left\\{\int_{[0,n]}\mathbb{I}\left(\sup_{\theta\in
G_{k}^{-}}Z_{u,k,l}(t,\theta)>u\right)dt>v(u)(x+\epsilon)\right\\}-2p_{k}^{*}(u),$
where
$\Lambda_{2}(u)=\left\\{l:0\leq l_{i}\leq\left[\frac{\pi
u}{Ln_{1}}\right]-1,1\leq i\leq m-1\right\\},$
$G_{l}=\prod_{i=1}^{m-1}[0,c_{i}(\theta_{l}(u))n_{1}]\supset\prod_{i=1}^{m-1}[0,c_{k,i}^{-}n_{1}]=:G_{k}^{-},\quad
c_{k,i}^{-}=\min_{\theta\in A_{k}}c_{i}(\theta).$
By (34), (35), Lemma 2.1 and Remark 2.2 we have for $n>x$
$\displaystyle p_{k}(x,u)$ $\displaystyle\geq$
$\displaystyle\liminf_{n_{1}\rightarrow\infty}\frac{\widehat{\mathcal{B}}_{\alpha,2,\dots,2}(x+\epsilon,[0,n]\times
G_{l}^{-})}{(n_{1})^{m-1}}\left(\frac{\pi}{L}\right)^{m-1}u^{m-1}\Psi(u)$
$\displaystyle\geq$
$\displaystyle\widehat{\mathcal{B}}_{\alpha,2,\dots,2}(x+\epsilon,n)\prod_{i=1}^{m-1}c_{k,i}^{-}\left(\frac{\pi}{L}\right)^{m-1}u^{m-1}\Psi(u)$
$\displaystyle\geq$
$\displaystyle\widehat{\mathcal{B}}_{\alpha,2,\dots,2}(x,n)\prod_{i=1}^{m-1}c_{k,i}^{-}\left(\frac{\pi}{L}\right)^{m-1}u^{m-1}\Psi(u),\quad
u\rightarrow\infty,\epsilon\rightarrow 0.$
$\diamond$ Asymptotics for $\pi(u,n)$. By (7.6) in [26]
$\pi_{1}(u)=o\left(u^{m-1}\Psi(u)\right),\quad
u\rightarrow\infty,L\rightarrow\infty.$
Therefore, in view of (31),
$\displaystyle\limsup_{u\to\infty}\frac{\pi(n,u)}{u^{m-1}\Psi(u)}$
$\displaystyle\leq$
$\displaystyle\limsup_{L\rightarrow\infty}\sum_{k\in\Lambda}\left(\prod_{i=1}^{m-1}c_{k,i}^{+}\right)\left(\frac{\pi}{L}\right)^{m-1}\widehat{\mathcal{B}}_{\alpha,2,\dots,2}(x,n),$
$\displaystyle\liminf_{u\to\infty}\frac{\pi(n,u)}{u^{m-1}\Psi(u)}$
$\displaystyle\geq$
$\displaystyle\liminf_{L\rightarrow\infty}\sum_{k\in\Lambda^{*}}\left(\prod_{i=1}^{m-1}c_{k,i}^{-}\right)\left(\frac{\pi}{L}\right)^{m-1}\widehat{\mathcal{B}}_{\alpha,2,\dots,2}(x,n).$
Using the fact that
$\limsup_{L\rightarrow\infty}\sum_{k\in\Lambda}\left(\prod_{i=1}^{m-1}c_{k,i}^{+}\right)\left(\frac{\pi}{L}\right)^{m-1}=\liminf_{L\rightarrow\infty}\sum_{k\in\Lambda^{*}}\left(\prod_{i=1}^{m-1}c_{k,i}^{-}\right)\left(\frac{\pi}{L}\right)^{m-1}=Vol(S_{m-1}),$
it follows that
$\displaystyle\pi(n,u)\sim\frac{\widehat{\mathcal{B}}_{\alpha,2,\dots,2}(x,n)}{\widehat{\mathcal{B}}_{\alpha,2,\dots,2}(0,n)}\mathbb{P}\left\\{\sup_{[0,v(u)n]}\chi(t)>u\right\\},\quad
u\rightarrow\infty.$
This completes the proof of Lemma 5.1. $\Box$
Condition A2 continued. Lemma 2.1 yields that for $x\geq 0$
$\frac{\widehat{\mathcal{B}}_{\alpha,2,\dots,2}(x)}{\widehat{\mathcal{B}}_{\alpha,2,\dots,2}(0)}=\lim_{n\rightarrow\infty}\frac{\widehat{\mathcal{B}}_{\alpha,2,\dots,2}(x,[0,n])}{\widehat{\mathcal{B}}_{\alpha,2,\dots,2}(0,[0,n])}\in(0,{1]}.$
Hence A2 holds with
$\bar{F}(x)=\frac{\widehat{\mathcal{B}}_{\alpha,2,\dots,2}(x)}{\widehat{\mathcal{B}}_{\alpha,2,\dots,2}(0)},\quad
x\geq 0.$
Thus we establish the claim and hence the proof is complete. $\Box$
### 5.6. Proof of Proposition 3.4
We first apply Theorem 1.1 to derive the asymptotics for case ii) of
Proposition 3.4. Let
$E(u,n)=\bigcup_{i=0}^{N(u,n)}I_{i}(u,n),\quad
I_{i}(u,n)=[iv(u)n,(i+1)v(u)n],~{}N(u,n)=\left[\frac{T_{u}}{nv(u)}\right]-2,$
and
$v(u)=u^{\frac{2(\alpha-1)}{\alpha}}\left(\frac{(\tau^{*})^{\alpha/2}}{1+c\tau^{*}}\right)^{2/\alpha},\quad\tau^{*}=\frac{\alpha}{c(2-\alpha)}.$
Let
$Z(s,t)=\frac{B_{\alpha}(s)-B_{\alpha}(t)}{1+c(s-t)},~{}I_{i}^{\prime}(u,n)=[iq(u)n,(i+1)q(u)n],~{}q(u)=u^{-1}v(u),$
and
$\displaystyle\Sigma\Sigma(u,n)$ $\displaystyle\coloneqq\sum_{i\neq j,0\leq
i,j\leq N(u,n)}\mathbb{P}\left\\{\sup_{t\in I_{i}(n,u)}Q(t)>u,\sup_{t\in
I_{j}(n,u)}Q(t)>u\right\\}$ $\displaystyle{=}\sum_{i\neq j,0\leq i,j\leq
N(u,n)}\mathbb{P}\left\\{\sup_{t\in I_{i}(n,u),s\geq
t}(B_{\alpha}(s)-B_{\alpha}(t)-c(s-t))>u,\sup_{t\in I_{j}(n,u),s\geq
t}(B_{\alpha}(s)-B_{\alpha}(t)-c(s-t))>u\right\\}$ $\displaystyle=\sum_{i\neq
j,0\leq i,j\leq N(u,n)}\mathbb{P}\left\\{\sup_{t\in I_{i}^{\prime}(n,u),s\geq
t}Z(s,t)>u^{1-\alpha/2},\sup_{t\in I_{j}^{\prime}(n,u),s\geq
t}Z(s,t)>u^{1-\alpha/2}\right\\},$
where in the last equality we use the self-similarity of fBm. Moreover, let
$L_{i}(u)=[\tau^{*}+iq(u)n_{1},\tau^{*}+(i+1)q(u)n_{1}],\quad
M(u)=\left[\frac{u^{\alpha/2}\ln u}{v(u)n_{1}}\right],$
$G(u)=\\{s:|s-\tau^{*}|<u^{\alpha/2-1}\ln u\\},\quad
G^{c}(u)=[0,\infty)\setminus G(u),$
and
(36) $\displaystyle\pi_{2}(u)=\sum_{-M(u)-1\leq i<j\leq
M(u)+1}\mathbb{P}\left\\{\sup_{t\in[0,q(u)n],s\in
L_{i}(u)}Z(s,t)>u^{1-\alpha/2},\sup_{t\in[0,q(u)n],s\in
L_{j}(u)}Z(s,t)>u^{1-\alpha/2}\right\\}.$
Conditions A1 and A3. Condition A1 follows from Theorems 3.1-3.3 of [12] while
A3 is due to Lemma 5.6 of [12] and the upper bounds of
$\Sigma_{i}(u),i=1,2,3,4$ in the proof of Theorem 3.1 in [12].
Condition A2. Due to stationarity of the process $Q$, in order to show (2) it
suffices to find the exact asymptotics of
$\mathbb{P}\left\\{\int_{[0,v(u)n]}\mathbb{I}(Q(t)>u)dt>v(u)x\right\\}$ as
$u\to\infty$. By the self-similarity of $B_{\alpha}$, we have
$\displaystyle\mathbb{P}\left\\{\int_{[0,v(u)n]}\mathbb{I}(Q(t)>u)dt>v(u)x\right\\}$
$\displaystyle=$
$\displaystyle\mathbb{P}\left\\{\int_{[0,v(u)n]}\mathbb{I}\left(\sup_{s\geq
t}(B_{\alpha}(s)-B_{\alpha}(t)-c(s-t))>u\right)dt>v(u)x\right\\}$
$\displaystyle=$
$\displaystyle\mathbb{P}\left\\{\int_{[0,q(u)n]}\mathbb{I}\left(\sup_{s\geq
t}Z(s,t)>u^{1-\alpha/2}\right)dt>q(u)x\right\\}.$
###### Lemma 5.2.
For $n>x$
(37)
$\displaystyle\mathbb{P}\left\\{\int_{[0,q(u)n]}\mathbb{I}\left(\sup_{s\geq
t}Z(s,t)>u^{1-\alpha/2}\right)dt>q(u)x\right\\}\sim\widehat{\mathcal{B}}_{\alpha,\alpha}(x,n)\sqrt{\frac{2A}{B}}\frac{u}{m(u)v(u)}\Psi(u),\quad
u\rightarrow\infty.$
Proof. Upper bound. Using the fact that
$\mathbb{I}\left(\sup_{s\geq
t}Z(s,t)>u^{1-\alpha/2}\right)\leq\mathbb{I}\left(\sup_{s\in
G(u)}Z(s,t)>u^{1-\alpha/2}\right)+\mathbb{I}\left(\sup_{s\in
G^{c}(u)}Z(s,t)>u^{1-\alpha/2}\right)$
we obtain
$\displaystyle\mathbb{P}\left\\{\int_{[0,q(u)n]}\mathbb{I}\left(\sup_{s\geq
t}Z(s,t)>u^{1-\alpha/2}\right)dt>q(u)x\right\\}$
$\displaystyle\quad\leq\mathbb{P}\left\\{\int_{[0,q(u)n]}\left(\mathbb{I}\left(\sup_{s\in
G(u)}Z(s,t)>u^{1-\alpha/2}\right)+\mathbb{I}\left(\sup_{s\in
G^{c}(u)}Z(s,t)>u^{1-\alpha/2}\right)\right)dt>q(u)x\right\\}$
$\displaystyle\quad\leq\mathbb{P}\left\\{\int_{[0,q(u)n]}\mathbb{I}\left(\sup_{s\in
G(u)}Z(s,t)>u^{1-\alpha/2}\right)dt>q(u)x\right\\}$
$\displaystyle\quad\quad+\mathbb{P}\left\\{\int_{[0,q(u)n]}\mathbb{I}\left(\sup_{s\in
G^{c}(u)}Z(s,t)>u^{1-\alpha/2}\right)dt>q(u)x\right\\}$
$\displaystyle\quad\quad+\mathbb{P}\left\\{\int_{[0,q(u)n]}\mathbb{I}\left(\sup_{s\in
G(u)}Z(s,t)>u^{1-\alpha/2}\right)dt>0,\int_{[0,q(u)n]}\mathbb{I}\left(\sup_{s\in
G^{c}(u)}Z(s,t)>u^{1-\alpha/2}\right)dt>0\right\\}$
$\displaystyle\quad\leq\mathbb{P}\left\\{\int_{[0,q(u)n]}\mathbb{I}\left(\sup_{s\in
G(u)}Z(s,t)>u^{1-\alpha/2}\right)dt>q(u)x\right\\}+2\mathbb{P}\left\\{\sup_{t\in[0,q(u)n],s\in
G^{c}(u)}Z(s,t)>u^{1-\alpha/2}\right\\}.$
Moreover, since
$\displaystyle\mathbb{I}\left(\sup_{s\in
G(u)}Z(s,t)>u^{1-\alpha/2}\right)\leq\sum_{|i|\leq
M(u)+1}\mathbb{I}\left(\sup_{s\in L_{i}(u)}Z(s,t)>u^{1-\alpha/2}\right)$
we have
$\displaystyle\mathbb{P}\left\\{\int_{[0,q(u)n]}\mathbb{I}\left(\sup_{s\in
G(u)}Z(s,t)>u^{1-\alpha/2}\right)dt>q(u)x\right\\}\leq\pi_{1}(u)+\pi_{2}(u),$
where $\pi_{2}(u)$ is given in (36) and
(38) $\displaystyle\pi_{1}(u)=\sum_{|i|\leq
M(u)+1}\mathbb{P}\left\\{\int_{t\in[0,q(u)n]}\mathbb{I}\left(\sup_{s\in
L_{i}(u)}Z(s,t)>u^{1-\alpha/2}\right)dt>q(u)x\right\\}.$
By Lemma 5.6 of [12] we obtain
$\mathbb{P}\left\\{\sup_{t\in[0,q(u)n],s\in
G^{c}(u)}Z(s,t)>u^{1-\alpha/2}\right\\}=o\left(\mathbb{P}\left\\{\sup_{t\in[0,v(u)n]}Q(t)>u\right\\}\right),\quad
u\rightarrow\infty,$
and in light of the upper bounds of $\Lambda_{i}(u),i=1,2,3,4$ in the proof of
Theorem 3.1 of [12]
(39)
$\displaystyle\pi_{2}(u)=o\left(\mathbb{P}\left\\{\sup_{t\in[0,v(u)n]}Q(t)>u\right\\}\right),\quad
u\rightarrow\infty,n_{1}\rightarrow\infty.$
Next we focus on $\pi_{1}(u)$. We denote
$m(u)=\frac{1+c\tau^{*}}{(\tau^{*})^{\alpha/2}}u^{1-\alpha/2},\quad\tau^{*}=\frac{\alpha}{c(2-\alpha)},$
$A=\left(\frac{\alpha}{c(2-\alpha)}\right)^{-\alpha/2}\frac{2}{2-\alpha},\quad
B=\left(\frac{\alpha}{c(2-\alpha)}\right)^{-\alpha/2-1}\frac{\alpha}{2}.$
Rewrite
$\displaystyle\mathbb{P}\left\\{\int_{t\in[0,q(u)n]}\mathbb{I}\left(\sup_{s\in
L_{i}(u)}Z(s,t)>u^{1-\alpha/2}\right)dt>q(u)x\right\\}=\mathbb{P}\left\\{\int_{t\in[0,n]}\mathbb{I}\left(\sup_{s\in[0,n_{1}]}Z_{u,i}(s,t)>m(u)\right)dt>x\right\\},$
where
$Z_{u,i}(s,t)=\frac{B_{\alpha}(\tau^{*}+q(u)(in_{1}+s))-B_{\alpha}(q(u)t)}{1+c(\tau^{*}+q(u)(in_{1}+s-t))}\cdot\frac{1+c\tau^{*}}{(\tau^{*})^{\alpha/2}}.$
Let for $0<\epsilon<1$
$m_{i}^{\pm}(u)=m(u)\left(1+\left(\frac{B}{2A}\pm\epsilon\right)q(u)(in_{1}\pm
n)^{2}\right).$
A direct calculation shows (see also Lemmas 5.3-5.4 in [12]) that
(40) $\displaystyle m_{i}^{-}(u)\leq m(u)(Var(Z_{u,i}(s,t)))^{-1/2}\leq
m_{i}^{+}(u),\quad|i|\leq M(u)+1$
and
(41) $\displaystyle\lim_{u\rightarrow\infty}\sup_{|i|\leq
M(u)+1}\sup_{(s,t)\neq(s^{\prime},t^{\prime}),(s,t),(s^{\prime},t^{\prime})\in[0,n_{1}]\times[0,n]}\left|(m_{i}^{\pm}(u))^{2}\frac{1-Corr(Z_{u,i}(s,t),Z_{u,i}(s^{\prime},t^{\prime}))}{|t-t^{\prime}|^{\alpha}+|s-s^{\prime}|^{\alpha}}-1\right|=0.$
Hence
$\displaystyle\mathbb{P}\left\\{\int_{t\in[0,n]}\mathbb{I}\left(\sup_{s\in[0,n_{1}]}Z_{u,i}(s,t)>m(u)\right)dt>x\right\\}\leq\mathbb{P}\left\\{\int_{t\in[0,n]}\mathbb{I}\left(\sup_{s\in[0,n_{1}]}\overline{Z}_{u,i}(s,t)>m_{i}^{-}(u)\right)dt>x\right\\}.$
Next, by Lemma 4.1 applied to $\Gamma:C([0,n]\times[0,n_{1}])\rightarrow
C([0,n])$ defined by $\Gamma(f)=\sup_{t\in[0,n_{1}]}f(s,t),f\in
C([0,n]\times[0,n_{1}])$, with $h=0$ in C0-C1 and C2 satisfied with
$\zeta(s,t)=B_{\alpha}(s)+B_{\alpha}^{\prime}(t)$, we have
(42) $\displaystyle\lim_{u\rightarrow\infty}\sup_{|i|\leq
M(u)+1}\left|\frac{\mathbb{P}\left\\{\int_{t\in[0,n]}\mathbb{I}\left(\sup_{s\in[0,n_{1}]}\overline{Z}_{u,i}(s,t)>m_{i}^{-}(u)\right)dt>x\right\\}}{\Psi(m_{i}^{-}(u))}-\widehat{\mathcal{B}}_{\alpha,\alpha}(x,[0,n]\times[0,n_{1}])\right|=0,$
and in light of Lemma 2.1, we have
(43) $\displaystyle\pi_{{1}}(u)$ $\displaystyle\leq$
$\displaystyle\widehat{\mathcal{B}}_{\alpha,\alpha}(x,[0,n]\times[0,n_{1}])\sum_{|i|\leq
M(u)+1}\Psi(m_{i}^{-}(u))$ $\displaystyle\leq$
$\displaystyle\widehat{\mathcal{B}}_{\alpha,\alpha}(x,[0,n]\times[0,n_{1}])\Psi(u)\sum_{|i|\leq
M(u)+1}e^{-m^{2}(u)\left(\frac{{\color[rgb]{0,0,0}B}}{2A}-\epsilon\right)\left(u^{-1}v(u)(in_{1})\right)^{2}}$
$\displaystyle\leq$
$\displaystyle\frac{\widehat{\mathcal{B}}_{\alpha,\alpha}(x,[0,n]\times[0,n_{1}])}{n_{1}}\sqrt{\frac{2A\pi}{B}}\frac{u}{m(u)v(u)}\Psi(u)$
$\displaystyle\sim$
$\displaystyle\widehat{\mathcal{B}}_{\alpha,\alpha}(x,n)\sqrt{\frac{2A\pi}{B}}\frac{u}{m(u)v(u)}\Psi(u),$
as $u\rightarrow\infty,n_{1}\rightarrow\infty,\epsilon\rightarrow 0$.
Therefore, we conclude that
$\displaystyle\mathbb{P}\left\\{\int_{[0,q(u)n]}\mathbb{I}\left(\sup_{s\geq
t}Z(s,t)>u^{1-\alpha/2}\right)dt>q(u)x\right\\}\leq\widehat{\mathcal{B}}_{\alpha,\alpha}(x,n)\sqrt{\frac{2A\pi}{B}}\frac{u}{m(u)v(u)}\Psi(u),\quad
u\rightarrow\infty.$
Lower bound. Observe that for u sufficiently large, $s>t$ holds for all $s\in
G(u),t\in[0,q(u)n]$. Therefore,
$\displaystyle\mathbb{P}\left\\{\int_{[0,q(u)n]}\mathbb{I}\left(\sup_{s\geq
t}Z(s,t)>u^{1-\alpha/2}\right)dt>q(u)x\right\\}\geq\mathbb{P}\left\\{\int_{[0,q(u)n]}\mathbb{I}\left(\sup_{s\in
G(u)}Z(s,t)>u^{1-\alpha/2}\right)dt>q(u)x\right\\}.$
By the fact that
$\displaystyle\mathbb{I}\left(\sup_{s\in G(u)}Z(s,t)>u^{1-\alpha/2}\right)$
$\displaystyle\geq$ $\displaystyle\sum_{|i|\leq
M(u)}\mathbb{I}\left(\sup_{s\in L_{i}(u)}Z(s,t)>u^{1-\alpha/2}\right)$
$\displaystyle-\sum_{-M(u)\leq i<j\leq M(u)}\mathbb{I}\left(\sup_{s\in
L_{i}(u)}Z(s,t)>u^{1-\alpha/2},\sup_{s\in
L_{j}(u)}Z(s,t)>u^{1-\alpha/2}\right)$ $\displaystyle=:$ $\displaystyle
A_{1}(u,t)-A_{2}(u,t),$
it follows that for $\epsilon>0$ (recall $q(u)=u^{-1}v(u)$)
$\displaystyle\mathbb{P}\left\\{\int_{[0,q(u)n]}\mathbb{I}\left(\sup_{s\geq
t}Z(s,t)>u^{1-\alpha/2}\right)dt>q(u)x\right\\}$
$\displaystyle\geq\mathbb{P}\left\\{\int_{[0,q(u)n]}\left(A_{1}(u,t)-A_{2}(u,t)\right)dt>q(u)x\right\\}$
$\displaystyle\geq\mathbb{P}\left\\{\int_{[0,q(u)n]}A_{1}(u,t)dt>q(u)(x+\epsilon),\int_{[0,q(u)n]}A_{2}(u,t)dt<q(u)\epsilon\right\\}$
$\displaystyle\geq\mathbb{P}\left\\{\int_{[0,q(u)n]}A_{1}(u,t)dt>q(u)(x+\epsilon)\right\\}-\mathbb{P}\left\\{\int_{[0,q(u)n]}A_{2}(u,t)dt\geq
q(u)\epsilon\right\\}$
$\displaystyle\geq\mathbb{P}\left\\{\exists{\color[rgb]{0,0,0}i:}\,|i|\leq
M(u),\int_{[0,q(u)n]}\mathbb{I}\left(\sup_{s\in
L_{i}(u)}Z(s,t)>u^{1-\alpha/2}\right)dt>q(u)(x+\epsilon)\right\\}-\pi_{2}(u)$
(44) $\displaystyle\geq\sum_{|i|\leq
M(u)}\mathbb{P}\left\\{\int_{t\in[0,q(u)n]}\mathbb{I}\left(\sup_{s\in
L_{i}(u)}Z(s,t)>u^{1-\alpha/2}\right)dt>q(u)(x+\epsilon)\right\\}-2\pi_{2}(u),$
where $\pi_{2}(u)$ is defined in (36). Similarly as in (43) and in light of
(39), we have
$\displaystyle\mathbb{P}\left\\{\int_{[0,q(u)n]}\mathbb{I}\left(\sup_{s\geq
t}Z(s,t)>u^{1-\alpha/2}\right)dt>q(u)x\right\\}$ $\displaystyle\geq$
$\displaystyle\widehat{\mathcal{B}}_{\alpha,\alpha}(x+\epsilon,n)\sqrt{\frac{2A}{B}}\frac{u}{m(u)v(u)}\Psi(u)$
$\displaystyle\geq$
$\displaystyle\widehat{\mathcal{B}}_{\alpha,\alpha}(x,n)\sqrt{\frac{2A}{B}}\frac{u}{m(u)v(u)}\Psi(u),\quad
u\rightarrow\infty,\epsilon\rightarrow 0.$
Consequently for $n>x$
(45)
$\displaystyle\mathbb{P}\left\\{\int_{[0,q(u)n]}\mathbb{I}\left(\sup_{s\geq
t}Z(s,t)>u^{1-\alpha/2}\right)dt>q(u)x\right\\}\sim\widehat{\mathcal{B}}_{\alpha,\alpha}(x,n)\sqrt{\frac{2A}{B}}\frac{u}{m(u)v(u)}\Psi(u),\quad
u\rightarrow\infty.$
$\Box$
Moreover, by Lemma 2.1
$\displaystyle\frac{\mathcal{B}_{\alpha}(x)}{\mathcal{B}_{\alpha}(0)}=\frac{\widehat{\mathcal{B}}_{\alpha,\alpha}(x)}{\widehat{\mathcal{B}}_{\alpha,\alpha}(0)}=\lim_{n\rightarrow\infty}\frac{\widehat{\mathcal{B}}_{\alpha,\alpha}(x,n)}{\widehat{\mathcal{B}}_{\alpha,\alpha}(0,n)}\in(0,{1]}.$
Thus A2 holds with
$\bar{F}(x)=\frac{\mathcal{B}_{\alpha}(x)}{\mathcal{B}_{\alpha}(0)},\quad
x\geq 0.$
This completes the proof of case ii).
For case i), note that if $x=0$, the claim clearly holds. Next we suppose that
$0<x<T$. By (45) for any $0<\epsilon<\min(x/2,(T-x)/2)$,
$\displaystyle\mathbb{P}\left\\{\int_{[0,T_{u}]}\mathbb{I}(Q(t)>u)dt>v(u)x\right\\}$
$\displaystyle\leq$
$\displaystyle\mathbb{P}\left\\{\int_{[0,v(u)(T+\epsilon)]}\mathbb{I}(Q(t)>u)dt>v(u)x\right\\}$
$\displaystyle\leq$
$\displaystyle\mathbb{P}\left\\{\int_{[0,v(u)T]}\mathbb{I}(Q(t)>u)dt>v(u)(x-\epsilon)\right\\}$
$\displaystyle\sim$
$\displaystyle\frac{\widehat{\mathcal{B}}_{\alpha,\alpha}(x-\epsilon,T)}{\widehat{\mathcal{B}}_{\alpha,\alpha}(0,T)}\mathbb{P}\left\\{\sup_{t\in[0,v(u)T]}Q(t)>u\right\\},\quad
u\rightarrow\infty.$
Analogously,
$\displaystyle\mathbb{P}\left\\{\int_{[0,T_{u}]}\mathbb{I}(Q(t)>u)dt>v(u)x\right\\}\geq\frac{\widehat{\mathcal{B}}_{\alpha,\alpha}(x+\epsilon,T)}{\widehat{\mathcal{B}}_{\alpha,\alpha}(0,T)}\mathbb{P}\left\\{\sup_{t\in[0,v(u)T]}Q(t)>u\right\\},\quad
u\rightarrow\infty.$
In light of Remark 2.2 we establish the claim by letting $\epsilon\rightarrow
0$ in the above inequalities. This completes the proof. $\Box$
## 6\. Appendix
Proof of Lemma 4.1 For notational simplicity denote by $\rho_{u,j}$ the
correlation function of the random field $\xi_{u,j}$. Further set
$\chi_{u,j}(s)\coloneqq
g_{u,j}(\overline{\xi}_{u,j}(s)-\rho_{u,j}(s,{0})\overline{\xi}_{u,j}({0})),\quad
s\in E_{1}$
and
$f_{u,j}(s,y)\coloneqq
y\rho_{u,j}(s,{0})-g^{2}_{u,j}\left(1-\rho_{u,j}(s,{0})\right)-g^{2}_{u,j}\frac{1-\sigma_{u,j}(s)}{\sigma_{u,j}(s)},\
s\in E_{1},y\in\mathbb{R}.$
Conditioning on $\xi_{u,j}({0})$, by F1 and using that
$\overline{\xi}_{u,j}({0})$ and
$\overline{\xi}_{u,j}(s)-\rho_{u,j}(s,{0})\overline{\xi}_{u,j}({0})$ are
mutually independent we obtain
$\displaystyle\mathbb{P}\left\\{\int_{E_{2}}\mathbb{I}\left\\{\Gamma\left({g_{u,j}(\xi_{u,j}(s)-g_{u,j})}\right)(t)>0\right\\}\eta(dt)>x\right\\}$
$\displaystyle=\frac{e^{-g^{2}_{u,j}/2}}{\sqrt{2\pi}g_{u,j}}\int_{\mathbb{R}}\exp\left(-y-\frac{y^{2}}{2g^{2}_{u,j}}\right)\mathbb{P}\left\\{\int_{E_{2}}\mathbb{I}\left\\{\Gamma\left(g_{u,j}(\xi_{u,j}(s)-g_{u,j})\right)(t)>0\right\\}\eta(d{t})>x|\xi_{u,j}({0})=g_{u,j}+yg_{u,j}^{-1}\right\\}dy$
$\displaystyle=\frac{e^{-g^{2}_{u,j}/2}}{\sqrt{2\pi}g_{u,j}}\int_{\mathbb{R}}\exp\left(-y-\frac{y^{2}}{2g^{2}_{u,j}}\right)\mathbb{P}\left\\{\int_{E_{2}}\mathbb{I}\left\\{\Gamma\left(\sigma_{u,j}(s)\left(\chi_{u,j}(s)+f_{u,j}(s,y)\right)\right)(t)>0\right\\}\eta(dt)>x\right\\}dy$
$\displaystyle=\frac{e^{-g^{2}_{u,j}/2}}{\sqrt{2\pi}g_{u,j}}\int_{\mathbb{R}}\exp\left(-y-\frac{y^{2}}{2g^{2}_{u,j}}\right)\mathcal{I}_{u,j}(y;x)dy,$
where
$\mathcal{I}_{u,j}(y;x)\coloneqq\mathbb{P}\left\\{\int_{E_{2}}\mathbb{I}\left\\{\Gamma\left(\sigma_{u,j}(s)\left(\chi_{u,j}(s)+f_{u,j}(s,y)\right)\right)(t)>0\right\\}\eta(dt)>x\right\\}.$
Noting that
$\lim_{u\rightarrow\infty}\sup_{j\in
S_{u}}\left|\frac{\frac{e^{-g^{2}_{u,j}/2}}{\sqrt{2\pi}g_{u,j}}}{\Psi(g_{u,j})}-1\right|=0$
in order to show the claim it suffices to prove that
(46) $\displaystyle\lim_{u\to\infty}\sup_{j\in
S_{u}}\left|\int_{\mathbb{R}}\exp\left(-y-\frac{y^{2}}{2g^{2}_{u,j}}\right)\mathcal{I}_{u,j}(y;x)dy-\mathcal{B}^{\Gamma,h,\eta}_{\zeta}(x,E_{2})\right|=0$
for all $x\geq 0$. In view of C3 it follows that that for $u>u_{0}$
$Var(\chi_{u,j}({s})-\chi_{u,j}({s}^{\prime}))\leq
g^{2}_{u,j}\mathbb{E}\left\\{\overline{\xi}_{u,j}({s})-\overline{\xi}_{u,j}({s^{\prime}})\right\\}^{2}\leq
Q_{1}\lVert{s}-{s^{\prime}}\rVert^{\nu},\quad s,s^{\prime}\in E_{1},$
with $\nu>0$. Further, by C0-C2 for each $y\in\mathbb{R}$
(47) $\displaystyle\lim_{u\to\infty}\sup_{j\in S_{u},s\in
E_{1}}\left|f_{u,j}(s,y)-y+\sigma^{2}_{\zeta}(s)+h(s)\right|=0.$
Hence, by F2
(48) $\displaystyle\sup_{j\in S_{u}}e^{-y}\mathcal{I}_{u,j}(y;x)$
$\displaystyle\leq$ $\displaystyle e^{-y}\sup_{j\in
S_{u}}\mathbb{P}\left\\{\sup_{t\in
E_{2}}\Gamma\left(\chi_{u,j}(s)+f_{u,j}(s,y)\right)(t)>0\right\\}$
$\displaystyle\leq$ $\displaystyle e^{-y}\sup_{j\in
S_{u}}\mathbb{P}\left\\{\sup_{s\in
E_{1}}\\{\chi_{u,j}(s)+f_{u,j}(s,y)\\}>0\right\\}$ $\displaystyle\leq$
$\displaystyle e^{-y}\sup_{j\in S_{u}}\mathbb{P}\left\\{\sup_{s\in
E_{1}}\chi_{u,j}(s)>Q_{2}\left|y\right|-Q_{3}\right\\}$ $\displaystyle\leq$
$\displaystyle Q_{4}\left|y\right|^{2n/\nu-1}e^{-Q_{5}y^{2}-y},\quad y<-M,$
where in the last inequality we used Piterbarg inequality and $M>0$. Moreover,
it follows trivially that for all $x\geq 0$
(49) $\displaystyle\sup_{j\in S_{u}}e^{-y}\mathcal{I}_{u,j}(y;x)\leq
e^{-y},\quad y\in\mathbb{R}.$
Therefore by the dominated convergence theorem and assumption C0
$\displaystyle\sup_{j\in
S_{u}}\left|\int_{\mathbb{R}}\exp\left(-y-\frac{y^{2}}{2g^{2}_{u,j}}\right)\mathcal{I}_{u,j}(y;x)dy-\int_{\mathbb{R}}e^{-y}\mathcal{I}_{u,j}(y;x)dy\right|$
$\displaystyle\leq\int_{\mathbb{R}}\sup_{j\in
S_{u}}\left(e^{-y}\mathcal{I}_{u,j}(y;x)({1-e^{-y^{2}/(2g^{2}_{u,j})}})\right)dy\rightarrow
0,\quad u\to\infty.$
Hence in order to prove the convergence in (46) it suffices to show that
(50) $\displaystyle\lim_{u\to\infty}\sup_{j\in
S_{u}}\left|\int_{\mathbb{R}}e^{-y}\mathcal{I}_{u,j}(y;x)dy-\mathcal{B}^{\Gamma,h,\eta}_{\zeta}(x,E_{2})\right|=0$
for all $x\in[0,\eta(E_{2}))$.
Weak convergence. The claim follows from the same arguments as in [11][Lem
4.3,4.7], where the precise meaning of uniform weak convergence is also given.
Thus let $C(E_{1})$ denote the Banach space of all continuous functions on the
compact set $E_{1}$ equipped with supremum norm. For any ${s},{s}^{\prime}\in
E_{1}$, by C2 we have
$Var(\chi_{u,j}(s)-\chi_{u,j}({s}^{\prime}))=g^{2}_{u,j}\left(\mathbb{E}\left\\{\overline{\xi}_{u,j}(s)-\overline{\xi}_{u,j}({s}^{\prime})\right\\}^{2}-\left(\rho_{u,j}(s,{0})-\rho_{u,j}({s}^{\prime},{0})\right)^{2}\right)\rightarrow
2Var(\zeta(s)-\zeta({s}^{\prime}))$
uniformly with respect to $j\in S_{u}$ as $u\to\infty$. Hence, the finite-
dimensional distributions of $\chi_{u,j}(s),s\in E_{1}$ weakly converge to
that of $\sqrt{2}\zeta(s),s\in E_{1}$ uniformly with respect to $j\in S_{u}$.
In view of C3, we know that the measures on $C(E_{1})$ induced by
$\\{\chi_{u,j}(s),s\in E_{1},j\in S_{u}\\}$ are uniformly tight for large $u$,
and by C1, $\sigma_{u,j}(s)$ converges to $1$ uniformly for $s\in E_{1}$ and
$j\in S_{u}$ as $u\to\infty$. Therefore, $\\{\sigma_{u,j}(s)\chi_{u,j}(s),s\in
E_{1}\\}$ converge weakly to $\\{\sqrt{2}\zeta(s),s\in E_{1}\\}$ as
$u\to\infty$ uniformly with respect to $j\in S_{u}$, which together with (47)
implies that for each $y\in\mathbb{R}$, the probability measures on $C(E_{1})$
induced by $\\{\chi_{u,j}^{f}(s,y),s\in E_{1}\\}$ converges weakly as
$u\to\infty$ to that induced by $\\{\zeta_{h}(s)+y,{t}\in E_{1}\\}$ uniformly
with respect to $j\in S_{u}$, where
$\chi_{u,j}^{f}({s},y)\coloneqq\sigma_{u,j}(s)\left(\chi_{u,j}(s)+f_{u,j}(s,y)\right)\quad\textrm{and}\quad\zeta_{h}(s)\coloneqq\sqrt{2}\zeta(s)-\sigma^{2}_{\zeta}(t)-h(s).$
Continuous mapping theorem implies that for each $y\in\mathbb{R}$, the push-
forward probability measures $P_{u,y}$ on $C(E_{2})$ induced by
$\\{\Gamma\left(\chi_{u,j}^{f}(\cdot,y)\right)(t),t\in E_{2}\\}$ converges
weakly the push-forward probability measure $P_{y}$ induced by
$\\{\Gamma\left(\zeta_{h}\right)(t)+y,t\in E_{2}\\}$ as $u\to\infty$ uniformly
with respect to $j\in S_{u}$.
The continuity of the sojourn functional is also discussed in [3][Lem 4.2]. A
sequence of functions $f_{n}\in C(E_{2})$ converges to $f\in C(E_{2})$ as
$n\to\infty$ with respect to uniform topology if $f_{n}\to f$ uniformly as
$n\to\infty$. Since $\eta$ is absolutely continuous with respect to Lebesgue
measure on $E_{2}$ we can define the set
$A_{*}=\left\\{f\in
C(E_{2}):\int_{E_{2}}\mathbb{I}(f(t)=0)\eta(dt)>0\right\\},$
which is measurable in the completion $\mathcal{C}^{\mu}$ of $\mathcal{C}$
with respect to $\nu$, where $\mathcal{C}$ is the Borel $\sigma$-field of
$C_{2}(E)$. Its complement belongs to ${\color[rgb]{0,0,0}\mathcal{C}^{\mu}}$,
i.e.,
$A_{*}^{c}=C(E_{2})\setminus A_{*}\in{\color[rgb]{0,0,0}\mathcal{C}^{\mu}}.$
Any function $f\in A_{*}^{c}$ is a continuity point of the sojourn functional
$J:C(E_{2})\mapsto[0,\eta(E_{2})],$ where
$J(f)=\int_{E_{2}}1(f(t)>0)\eta(dt),f\in C(E_{2}).$
This functional is measurable $\mathcal{C}/\mathcal{B}(\mathbb{R})$ by the
assumption on $\eta$. We shall show that it is continuous at any $f\in
A_{*}^{c}$. Let such $f$ be given. By the definition of the integral such $f$
is not equal to zero on any compact interval of $\mathbb{R}$. Let $f_{n}\to f$
uniformly as $n\to\infty$. Then $1(f_{n}(t)>0)\to 1(f(t)>0)$ as $n\to\infty$
for almost all $t\in\mathbb{R}$ (with respect to Lebesgue measure). Hence by
dominated convergence theorem we have $J(f_{n})\to J(f)$ as $n\to\infty$,
which means that the functional is continuous for all $f\in A_{*}^{c}$. Recall
that $P_{y}$ is the push-forward (image measure) on $C(E_{2})$ with respect to
$\Gamma(\xi_{h})+y$. We claim that
$P_{y}(A_{*})>0$
is possible only for $y$ in a countable set of $\mathbb{R}$. Indeed, any $f\in
A_{*}$ is such that it is constant equal to zero on a compact interval.
Consequently, $P_{y}(A_{*})>0$ means that the functions $f\in A_{*}$ are
constant equal to $-y$ on some interval of $\mathbb{R}$. If this is true for
two different $y$’s, then the intervals where $f$ is constant equal $-y$ must
be disjoint, therefore this can be true only for countable $y$’s.
Alternatively, using the fact that
$\mathbb{P}\left\\{\Gamma(\zeta_{h})(t)+y=0\right\\}=0$ a.e.,
$y\in\mathbb{R}$, by the $\sigma$-finiteness of $\eta$, Fubini-Tonelli theorem
yields
$\displaystyle\int_{\mathbb{R}}\mathbb{E}\left\\{\int_{E_{2}}\mathbb{I}(\Gamma(\zeta_{h})(t)+y=0)\eta(dt)\right\\}dy=\int_{E_{2}}\int_{\mathbb{R}}\mathbb{P}\left\\{\Gamma(\zeta_{h})(t)+y=0\right\\}dy\eta(dt)=0.$
Hence for almost all $y\in\mathbb{R}$
$\mathbb{E}\left\\{\int_{E_{2}}\mathbb{I}(\Gamma(\zeta_{h})(t)+y=0)\eta(dt)\right\\}=0,$
which means that, for almost all $y\in\mathbb{R}$
$P_{y}(A_{*})=\mathbb{P}\left(\int_{E_{2}}\mathbb{I}(\Gamma(\zeta_{h})(t)+y=0)\eta(dt)>0\right)=0.$
Consequently, since $J(f)$ is continuous for $f\in A_{*}^{c}$, by continuous
mapping theorem, as $u\to\infty$
(51)
$\displaystyle\int_{E_{2}}\mathbb{I}\left(\Gamma\left(\chi_{u,j}^{f}(\cdot,y)\right)(t)>0\right)\eta(dt)$
weakly converges to
$\int_{E_{2}}\mathbb{I}\left(\Gamma\left(\zeta_{h}\right)(t)+y>0\right)\eta(dt)$
uniformly with respect to $j\in S_{u}$ for almost all $y\in\mathbb{R}$.
Convergence on continuity points. Define
$\mathcal{I}(y;x)\coloneqq\mathbb{P}\left\\{\int_{E_{2}}\mathbb{I}\left(\Gamma(\zeta_{h})(t)+y>0\right)\eta(d{t})>x\right\\}.$
We draw a similar argument as in Theorem 1.3.1 of [7] to verify (50) for all
continuity points $x\in(0,\eta(E_{2}))$ of
$\mathcal{B}^{\Gamma,h,\eta}_{\zeta}(x,E_{2})$. Let $x_{0}\in(0,\eta(E_{2}))$
be such a continuity point, that is
$\lim_{\varepsilon\to
0}\int_{\mathbb{R}}\left(\mathcal{I}(y;x_{0}+\varepsilon)-\mathcal{I}(y;x_{0}-\varepsilon)\right)e^{-y}dy=0.$
Since for large $M$ and all $x\geq 0$ by F2 as in the derivation of (48) we
have
(52) $\displaystyle e^{-y}\mathcal{I}(y;x)\leq
Q_{4}^{\prime}\left|y\right|^{2n/\nu-1}e^{-Q_{5}y^{2}-y},\quad y<-M$
it follows from the dominated convergence theorem that
$\int_{\mathbb{R}}\left(\mathcal{I}(y;x_{0}+)-\mathcal{I}(y;x_{0}-)\right)e^{-y}dy=0$
and thus by the monotonicity of $\mathcal{I}(y;x)$ in $x$ for each fixed $y$,
$x_{0}$ is a continuous point of $\mathcal{I}(y;x)$ for a.e. $y\in\mathbb{R}$.
Thus by (51) for a.e. $y\in\mathbb{R}$
(53) $\displaystyle\lim_{u\to\infty}\sup_{j\in
S_{u}}\left|\mathcal{I}_{u,j}(y;x_{0})-\mathcal{I}(y;x_{0})\right|=0.$
As shown in (48), (49) and (52) it follows from the dominated convergence
theorem that
$\displaystyle\sup_{j\in
S_{u}}\left|\int_{\mathbb{R}}e^{-y}\mathcal{I}_{u,j}(y;x_{0})dy-\int_{\mathbb{R}}e^{-y}\mathcal{I}(y;x_{0})dy\right|$
(54) $\displaystyle\leq\int_{\mathbb{R}}\sup_{j\in
S_{u}}\left|\mathcal{I}_{u,j}(y;x_{0})-\mathcal{I}(y;x_{0})\right|e^{-y}dy\rightarrow
0,\quad u\to\infty$
establishing the proof for all continuity points $x\in(0,\eta(E_{2}))$.
Moreover, for the case that $x=0$, (6) also holds by replacing sojourn with
supremum. This can be shown directly without any continuity requirement for
$\mathcal{B}^{\Gamma,h,\eta}_{\zeta}(x,E_{2})$ at $x=0$.
Continuity of $\mathcal{B}^{\Gamma,h,\eta}_{\zeta}(x,E_{2})$. Next we show
that $\mathcal{B}^{\Gamma,h,\eta}_{\zeta}(x,E_{2})$ is continuous at any
$x\in(0,\eta(E_{2}))$ using that $\eta$ is equivalent with Lebesgue measure on
$E_{2}$. Note that $\mathcal{B}^{\Gamma,h,\eta}_{\zeta}(x,E_{2})$ is clearly
right continuous at $0$. Next we show the continuity at $x\in(0,E_{2})$. The
claimed continuity at $x$ follows if we show
$\int_{\mathbb{R}}\mathbb{P}\left\\{A_{y}\right\\}e^{-y}dy=0,\quad
A_{y}=\left\\{\int_{E_{2}}\mathbb{I}\big{(}\Gamma(\zeta_{h})(t)+y>0\big{)}\eta(d{t})=x\right\\},\quad
y\in\mathbb{R}.$
If
$\int_{E_{2}}\mathbb{I}\big{(}\Gamma(\zeta_{h})(t)+y>0\big{)}\eta(d{t})=x,$
with $0<x<\eta(E_{2})$, then using the fact that $\Gamma(\zeta_{h})(t)$ is
continuous over $E_{2}$ and the Lebesgue measure is absolutely continuous with
respect to $\eta$, we have that for any $y^{\prime}>y$
$\int_{E_{2}}\mathbb{I}\big{(}\Gamma(\zeta_{h})(t)+y^{\prime}>0\big{)}\eta(d{t})>x.$
This implies that $A_{y}\cap A_{y^{\prime}}=\emptyset,y\neq
y^{\prime},y,y^{\prime}\in\mathbb{R}.$ Noting that the continuity of
$\Gamma(\zeta_{h})$ guarantees the measurability of $A_{y}$, and
$\\{y:y\in\mathbb{R}\quad\text{such
that}\quad\mathbb{P}\left\\{A_{y}\right\\}>0\\}$
is a countable set because if it were not we would find countably many
(disjoint) $A_{y}$ such that $\sum\mathbb{P}\left\\{A_{y}\right\\}=\infty$.
Thus we get $\int_{\mathbb{R}}\mathbb{P}\left\\{A_{y}\right\\}e^{-y}dy=0,$
hence $\mathcal{B}^{\Gamma,h,\eta}_{\zeta}(x,E_{2})$ is continuous on
$(0,\eta(E_{2}))$, establishing the claim. $\Box$
Before proceeding to the proof of Lemma 4.2, under notation introduced in the
proof of Proposition 3.1, we denote and analyze
(55) $\displaystyle\Sigma\Sigma_{1}(u,n)$ $\displaystyle\coloneqq$
$\displaystyle\sum_{0\leq k_{i},k_{i}^{\prime}\leq
N_{i}(u,n),i=1,2,(k_{1},k_{2})\neq(k_{1}^{\prime},k_{2}^{\prime})}\mathbb{P}\left\\{\sup_{t\in
I_{k_{1},k_{2}}(u,n)}X(t)>u,\sup_{t\in
I_{k_{1}^{\prime},k_{2}^{\prime}}(u,n)}X(t)>u\right\\},$ (56)
$\displaystyle\Sigma\Sigma_{2}(u,n)$ $\displaystyle\coloneqq$
$\displaystyle\sum_{0\leq 2k_{i},2k_{i}^{\prime}\leq
N_{i}(u,n),i=1,2,(k_{1},k_{2})\neq(k_{1}^{\prime},k_{2}^{\prime})}\mathbb{P}\left\\{\sup_{t\in
I_{2k_{1},2k_{2}}(u,n)}X(t)>u,\sup_{t\in
I_{2k^{\prime}_{1},2k^{\prime}_{2}}(u,n)}X(t)>u\right\\},$ (57)
$\displaystyle\Theta(u)$ $\displaystyle\coloneqq$ $\displaystyle
T_{1}T_{2}a_{1}^{1/\alpha_{1}}a_{2}^{1/\alpha_{2}}u^{2/\alpha_{1}+2/\alpha_{2}}\Psi(u).$
Moreover, following notation introduced in the proof of Proposition 3.2, let
$\Sigma\Sigma_{3}^{\prime\prime}(u,n)\coloneqq\sum_{|k_{i}|,|k_{i}^{\prime}|\leq
N_{i}^{\prime}(u,n),i=1,2,(k_{1},k_{2})\neq(k_{1}^{\prime},k_{2}^{\prime})}\mathbb{P}\left\\{\sup_{t\in
I_{k_{1},k_{2}}(u,n)}X(t)>u,\sup_{t\in
I_{k_{1}^{\prime},k_{2}^{\prime}}(u,n)}X(t)>u\right\\}.$ (58)
$\displaystyle\hat{I}_{k_{2}}(u,n)\coloneqq I_{-1,k_{2}}(u,n)\cup
I_{0,k_{2}}(u,n),\quad E_{1}(u,n)\coloneqq\bigcup_{|k_{2}|\leq
N_{2}^{\prime}(u,n)}\hat{I}_{k_{2}}(u,n),$
and
(59) $\displaystyle\Sigma_{3}^{\prime}(u,n)$ $\displaystyle\coloneqq$
$\displaystyle\sum_{|k_{i}|\leq
N_{i}^{\prime}(u,n)+1,i=1,2,~{}k_{1}\neq-1,0}\mathbb{P}\left\\{\sup_{t\in
I_{k_{1},k_{2}}(u,n)}X(t)>u\right\\},$ (60)
$\displaystyle\Sigma\Sigma_{3}(u,n)$ $\displaystyle\coloneqq$
$\displaystyle\sum_{|k_{2}|,|k_{2}^{\prime}|\leq N_{2}^{\prime}(u,n),k_{2}\neq
k_{2}^{\prime}}\mathbb{P}\left\\{\sup_{t\in\hat{I}_{k_{2}}(u,n)}X(t)>u,\sup_{t\in\hat{I}_{k_{2}^{\prime}}(u,n)}X(t)>u\right\\},$
(61) $\displaystyle\Sigma\Sigma_{4}(u,n)$ $\displaystyle\coloneqq$
$\displaystyle\sum_{|2k_{2}|,|2k_{2}^{\prime}|\leq
N_{2}^{\prime}(u,n)-1,k_{2}\neq
k_{2}^{\prime}}\mathbb{P}\left\\{\sup_{t\in\hat{I}_{2k_{2}}(u,n)}X(t)>u,\sup_{t\in\hat{I}_{2k_{2}^{\prime}}(u,n)}X(t)>u\right\\}.$
###### Lemma 6.1.
Under the assumptions of Proposition 3.1
(62) $\displaystyle\mathbb{P}\left\\{\sup_{t\in
E}X(t)>u\right\\}\sim\sum_{0\leq k_{i}\leq
N_{i}(u,n),i=1,2}\mathbb{P}\left\\{\sup_{t\in
I_{k_{1},k_{2}}(u,{n})}X(t)>u\right\\}\sim\mathbb{C}_{0}\Theta(u),\quad
u\rightarrow\infty,n\to\infty,$
where $\mathbb{C}_{0}>0$. Moreover, for all large $u$ and $n$
$\displaystyle\Sigma\Sigma_{1}(u,n)$ $\displaystyle\leq$
$\displaystyle\left(\frac{\mathbb{C}_{2}}{\sqrt{n}}+e^{-\mathbb{C}_{1}n^{\mathbb{C}}}\right)\Theta(u),\quad\Sigma\Sigma_{2}(u,n)\leq
e^{-\mathbb{C}_{1}n^{\mathbb{C}}}\Theta(u),$
where $\mathbb{C},\mathbb{C}_{1}$ and $\mathbb{C}_{2}$ are some positive
constants.
Proof of Lemma 6.1 Asymptotics (62) follow from Lemma 7.1 in [26], while the
bounds can be deduced from equations (7.4) and (7.6) in the proof of Lemma 7.1
in [26]. $\Box$
###### Lemma 6.2.
Under the assumptions of Proposition 3.2, for $\alpha_{i}<\beta_{i},i=1,2$,
$\mathbb{P}\left\\{\sup_{t\in E\setminus
E(u,n)}X(t)>u\right\\}=o\left(\mathbb{P}\left\\{\sup_{t\in
E}X(t)>u\right\\}\right)$
as $u\to\infty,n\to\infty$, and
$\Sigma\Sigma_{3}^{\prime\prime}(u,n)=o\left(\sum_{0\leq k_{i}\leq
N_{i}^{\prime}(u,n),i=1,2}\mathbb{P}\left\\{\sup_{t\in
I_{k_{1},k_{2}}(u,n)}X(t)>u\right\\}\right),$
as $u\to\infty,n\to\infty$. For $\alpha_{1}=\beta_{1},\alpha_{2}<\beta_{2}$
$\mathbb{P}\left\\{\sup_{t\in E\setminus
E_{1}(u,n)}X(t)>u\right\\}=o\left(\mathbb{P}\left\\{\sup_{t\in
E}X(t)>u\right\\}\right),$
as $u\to\infty$, $n\to\infty$, and for $u$ and $n$ sufficiently large
$\displaystyle\Sigma\Sigma_{3}(u,n)$
$\displaystyle\leq\left(\frac{\mathbb{C}_{2}}{\sqrt{n}}+e^{-\mathbb{C}_{1}n^{\mathbb{C}}}\right)\mathbb{P}\left\\{\sup_{t\in
E}X(t)>u\right\\},$ $\displaystyle\Sigma_{3}^{\prime}(u,n)$ $\displaystyle\leq
e^{-\mathbb{C}_{1}n^{\mathbb{C}}}\mathbb{P}\left\\{\sup_{t\in
E}X(t)>u\right\\},$ $\displaystyle\Sigma\Sigma_{4}(u,n)$ $\displaystyle\leq
e^{-\mathbb{C}_{1}n^{\mathbb{C}}}\mathbb{P}\left\\{\sup_{t\in
E}X(t)>u\right\\}.$
For $\alpha_{1}=\beta_{1}$ and $\alpha_{2}=\beta_{2}$
$\mathbb{P}\left\\{\sup_{t\in
E\setminus\bigcup_{i,j\in\\{-1,0\\}}I_{i,j}(u,n)}X(t)>u\right\\}=o\left(\mathbb{P}\left\\{\sup_{t\in
E}X(t)>u\right\\}\right),$
as $u\to\infty,n\to\infty$.
Proof of Lemma 6.2 The proof of Lemma 6.2 follows from [18]. Specifically, the
first one follows from (34), the second one from (40) and (41), the third one
from (34) and (46), the fourth one from (48) and (49), the fifth one from
(46), the six one from (48), and the last one from (34) and (52) in the proof
of Theorem 3.1 of [18]. $\Box$
Now we are in the position to prove Lemma 4.2.
Proof of Lemma 4.2 Ad (i). We follow notation introduced in the proof of
Proposition 3.1. For any $n,n_{1}>\sqrt{x}$, we have
(63)
$\displaystyle\Sigma_{1}^{-}(u,n_{1})-\Sigma\Sigma_{1}(u,n_{1})\leq\mathbb{P}\left\\{\int_{E(u,n)}\mathbb{I}(X(t)>u)dt>v(u)x\right\\}\leq\Sigma_{1}^{+}(u,n)+\Sigma\Sigma_{1}(u,n),$
where $\Sigma\Sigma_{1}(u,n)$ is given in (55) and
$\displaystyle\Sigma_{1}^{\pm}(u,n)=\sum_{0\leq k_{i}\leq N_{i}(u,n)\pm
1,i=1,2}\mathbb{P}\left\\{\int_{I_{k_{1},k_{2}}(u,n)}\mathbb{I}(X(t)>u)dt>v(u)x\right\\}.$
By (17), it follows that
$\displaystyle\Sigma_{1}^{+}(u,n)$ $\displaystyle\leq$
$\displaystyle\sum_{0\leq k_{i}\leq
N_{i}(u,n),i=1,2}\mathcal{B}_{\alpha_{1},\alpha_{2}}(x,[0,n]^{2})\Psi(u)$
$\displaystyle\leq$
$\displaystyle\frac{\mathcal{B}_{\alpha_{1},\alpha_{2}}(x,[0,n]^{2})}{n^{2}}\Theta(u),\quad
u\rightarrow\infty,$
where $\Theta(u)$ is defined in (57). Analogously, we obtain the lower bound
$\displaystyle\Sigma_{1}^{-}(u,n)$ $\displaystyle\geq$
$\displaystyle\frac{\mathcal{B}_{\alpha_{1},\alpha_{2}}(x,[0,n]^{2})}{n^{2}}\Theta(u),\quad
u\rightarrow\infty.$
Lemma 6.1 shows that for $u$ and $n$ sufficiently large
$\Sigma\Sigma_{1}(u,n)\leq\left(\frac{\mathbb{C}_{2}}{\sqrt{n}}+e^{-\mathbb{C}_{1}n^{\mathbb{C}}}\right)\Theta(u).$
Dividing both sides of (63) by $\Theta(u)$ and letting $u\rightarrow\infty$,
we have
$\displaystyle\frac{\mathcal{B}_{\alpha_{1},\alpha_{2}}(x,[0,n_{1}]^{2})}{n_{1}^{2}}-\frac{\mathbb{C}_{2}}{\sqrt{n_{1}}}-e^{-\mathbb{C}_{1}n_{1}^{\mathbb{C}}}\leq\frac{\mathcal{B}_{\alpha_{1},\alpha_{2}}(x,[0,n]^{2})}{n^{2}}+\frac{\mathbb{C}_{2}}{\sqrt{n}}+e^{-\mathbb{C}_{1}n^{\mathbb{C}}}.$
The above implies that
$\displaystyle\limsup_{n\rightarrow\infty}\frac{\mathcal{B}_{\alpha_{1},\alpha_{2}}(x,[0,n]^{2})}{n^{2}}$
$\displaystyle=$
$\displaystyle\liminf_{n\rightarrow\infty}\frac{\mathcal{B}_{\alpha_{1},\alpha_{2}}(x,[0,n]^{2})}{n^{2}}<\infty.$
Next we show that
$\liminf_{n\rightarrow\infty}\frac{\mathcal{B}_{\alpha_{1},\alpha_{2}}(x,[0,n]^{2})}{n^{2}}>0.$
Observe that
(64)
$\displaystyle\mathbb{P}\left\\{\int_{E}\mathbb{I}(X(t)>u)dt>v(u)x\right\\}\geq\Sigma_{2}(u,n)-\Sigma\Sigma_{2}(u,n),$
where $\Sigma\Sigma_{2}(u)$ is given in (56) and
$\displaystyle\Sigma_{2}(u,n)=\sum_{0\leq 2k_{i}\leq
N_{i}^{\prime}(u,n),i=1,2}\mathbb{P}\left\\{\int_{I_{2k_{1},2k_{2}}(u,n)}\mathbb{I}(X(t)>u)dt>v(u)x\right\\}.$
In light of (17), we have
$\displaystyle\Sigma_{2}(u,n)$ $\displaystyle\geq$ $\displaystyle\sum_{0\leq
2k_{i}\leq
N_{i}^{\prime}(u,n),i=1,2}\mathcal{B}_{\alpha_{1},\alpha_{2}}(x,[0,n]^{2})\Psi(u)$
$\displaystyle\geq$
$\displaystyle\frac{\mathcal{B}_{\alpha_{1},\alpha_{2}}(x,[0,n]^{2})}{4n^{2}}\Theta(u),\quad
u\rightarrow\infty.$
Moreover, by Lemma 6.1 we have, for $u$ and $n$ large enough
$\Sigma\Sigma_{2}(u,n)\leq e^{-\mathbb{C}_{1}n^{\mathbb{C}}}\Theta(u).$
Combination of upper bound in (63) and lower bound in (64) leads to
(65)
$\displaystyle\liminf_{n\rightarrow\infty}\frac{\mathcal{B}_{\alpha_{1},\alpha_{2}}(x,[0,n]^{2})}{n^{2}}\geq\frac{\mathcal{B}_{\alpha_{1},\alpha_{2}}(x,[0,n_{1}]^{2})}{4n_{1}^{2}}-e^{-\mathbb{C}_{1}n_{1}^{\mathbb{C}}}.$
For $n_{1}>\sqrt{x}$
$\displaystyle\mathcal{B}_{\alpha_{1},\alpha_{2}}(x,[0,n_{1}]^{2})$
$\displaystyle=$
$\displaystyle\int_{\mathbb{R}}\mathbb{P}\left\\{\int_{[0,n_{1}]^{2}}\mathbb{I}\left(\sum_{i=1}^{2}(\sqrt{2}B_{\alpha_{i}}(t_{i})-\left|t_{i}\right|^{\alpha_{i}})>s\right)dt>x\right\\}e^{s}ds$
$\displaystyle\geq$
$\displaystyle\int_{\mathbb{R}}\mathbb{P}\left\\{\inf_{t\in[0,n_{1}]^{2}}\sum_{i=1}^{2}\left(\sqrt{2}B_{\alpha_{i}}(t_{i})-\left|t_{i}\right|^{\alpha_{i}}\right)>s\right\\}e^{s}ds>0,$
which combined with the monotonicity of
$\mathcal{B}_{\alpha_{1},\alpha_{2}}(x,[0,n_{1}]^{2})$ in $n_{1}$ and (65)
implies that for sufficiently large $n_{1}$
$\liminf_{n\rightarrow\infty}\frac{\mathcal{B}_{\alpha_{1},\alpha_{2}}(x,[0,n]^{2})}{n^{2}}\geq\frac{\mathcal{B}_{\alpha_{1},\alpha_{2}}(x,[0,n_{1}]^{2})-4n_{1}^{2}e^{-\mathbb{C}_{1}n_{1}^{\mathbb{C}}}}{4n_{1}^{2}}>0,$
establishing the proof of (i).
Ad (ii). We follow notation introduced in the proof of Proposition 3.2 for the
case $\alpha_{1}=\beta_{1}$ and $\alpha_{2}<\beta_{2}$. Let next for $u>0$
$E_{2}(u)\coloneqq\left[-\left(\frac{e_{u}^{-1/4}\wedge\ln
u}{u}\right)^{2/\beta_{1}},\left(\frac{e_{u}^{-1/4}\wedge\ln
u}{u}\right)^{2/\beta_{1}}\right]\times\left[-\left(\frac{e_{u}^{-1/4}\wedge\ln
u}{u}\right)^{2/\beta_{2}},\left(\frac{e_{u}^{-1/4}\wedge\ln
u}{u}\right)^{2/\beta_{2}}\right],$
$I_{k_{1},k_{2}}(u,n)\coloneqq[k_{1}v_{1}(u)n,(k_{1}+1)v_{1}(u)n]\times[k_{2}v_{2}(u)n,(k_{2}+1)v_{2}(u)n],$
$\Theta_{1}(u)\coloneqq
2\hat{\Gamma}(1/\beta_{2}+1)a_{2}^{1/\alpha_{2}}b_{2}^{-1/\beta_{2}}u^{2/\alpha_{2}-2/\beta_{2}}\Psi(u),$
where $\hat{\Gamma}(\cdot)$ is the gamma function and
$e_{u}=\sup_{0<|t_{i}|<\left(\frac{\ln
u}{u}\right)^{2/\beta_{i}},i=1,2}|e(t)|,\ \
~{}e(t)=\frac{1-{\sigma}(t)}{\sum_{i=1}^{2}b_{i}|t_{i}|^{\beta_{i}}}-1,|t|\neq
0.$
Observe that
$\displaystyle\mathbb{P}\left\\{\int_{E_{2}(u)}\mathbb{I}(X(t)>u)dt>v(u)x\right\\}$
$\displaystyle\geq$
$\displaystyle\mathbb{P}\left\\{\int_{E_{1}(u,n)}\mathbb{I}(X(t)>u)dt>v(u)x\right\\},$
$\displaystyle\mathbb{P}\left\\{\int_{E_{2}(u)}\mathbb{I}(X(t)>u)dt>v(u)x\right\\}$
$\displaystyle\leq$ $\displaystyle\mathbb{P}\left\\{\int_{\bigcup_{|k_{2}|\leq
N_{2}^{\prime}(u,n)+1}\hat{I}_{k_{2}}(u,n)}\mathbb{I}(X(t)>u)dt>v(u)x\right\\}$
$\displaystyle\quad+\mathbb{P}\left\\{\sup_{E_{2}(u)\setminus(\bigcup_{|k_{2}|\leq
N_{2}^{\prime}(u,n)+1}\hat{I}_{k_{2}}(u,n))}X(t)>u\right\\}.$
Hence it follows that
(66)
$\displaystyle\Sigma_{3}^{-}(u,n_{1})-\Sigma\Sigma_{3}(u,n_{1})\leq\mathbb{P}\left\\{\int_{E_{2}(u)}\mathbb{I}(X(t)>u)dt>v(u)x\right\\}\leq\Sigma_{3}^{+}(u,n)+\Sigma_{3}^{\prime}(u,n),$
with
$\displaystyle\Sigma_{3}^{\pm}(u,n)=\sum_{|k_{2}|\leq N_{2}^{\prime}(u,n)\pm
1}\mathbb{P}\left\\{\int_{\hat{I}_{k_{2}}(u,n)}\mathbb{I}(X(t)>u)dt>v(u)x\right\\},$
where $I_{k_{1},k_{2}}(u,n)$ is defined in (20) and $\Sigma_{3}^{\prime}$ and
$\Sigma\Sigma_{3}$ are given in (59) and (60) respectively. Noting that (5.4)
also holds for $|k_{2}|\leq N_{2}^{\prime}(u,n)+1$, we have for $x\geq 0$
$\displaystyle\Sigma_{3}^{\pm}(u,n)$ $\displaystyle\sim$
$\displaystyle\mathcal{B}_{\alpha_{1},\alpha_{2}}^{a_{1}^{-1}b_{1}|t_{1}|^{\alpha_{1}},0}(x,[-n,n]\times[0,n])\sum_{|k_{2}|\leq
N_{2}^{\prime}(u,n)+1}\Psi(u_{k_{2},n}^{\pm})$ $\displaystyle\sim$
$\displaystyle\mathcal{B}_{\alpha_{1},\alpha_{2}}^{a_{1}^{-1}b_{1}|t_{1}|^{\alpha_{1}},0}(x,[-n,n]\times[0,n])\Psi(u)\sum_{|k_{2}|\leq
N_{2}^{\prime}(u,n)+1}e^{-u^{2}b_{2}(|k_{2}|v_{2}(u)n)^{\beta_{2}}}$
$\displaystyle\sim$
$\displaystyle\frac{\mathcal{B}_{\alpha_{1},\alpha_{2}}^{a_{1}^{-1}b_{1}|t_{1}|^{\alpha_{1}},0}(x,[-n,n]\times[0,n])}{n}\Theta_{1}(u),\quad
u\rightarrow\infty.$
In light of Lemma 6.2, we have that for $u$ and $n$ sufficiently large
$\Sigma\Sigma_{3}(u,n)+\Sigma_{3}^{\prime}(u,n)\leq\left(\frac{\mathbb{C}_{2}}{\sqrt{n}}+e^{-\mathbb{C}_{1}n^{\mathbb{C}}}\right)\Theta_{1}(u).$
Dividing both sides of (66) by $\Theta_{1}(u)$ respectively and letting
$u\rightarrow\infty$, we have that
$\frac{\mathcal{B}_{\alpha_{1},\alpha_{2}}^{a_{1}^{-1}b_{1}|t_{1}|^{\alpha_{1}},0}(x,[-n_{1},n_{1}]\times[0,n_{1}])}{n_{1}}-\frac{\mathbb{C}_{2}}{\sqrt{n_{1}}}-e^{-\mathbb{C}_{1}n_{1}^{\mathbb{C}}}\leq\frac{\mathcal{B}_{\alpha_{1},\alpha_{2}}^{a_{1}^{-1}b_{1}|t_{1}|^{\alpha_{1}},0}(x,[-n,n]\times[0,n])}{n}+\frac{\mathbb{C}_{2}}{\sqrt{n}}+e^{-\mathbb{C}_{1}n^{\mathbb{C}}},$
which gives that
$\displaystyle\liminf_{n\rightarrow\infty}\frac{\mathcal{B}_{\alpha_{1},\alpha_{2}}^{a_{1}^{-1}b_{1}|t_{1}|^{\alpha_{1}},0}(x,[-n,n]\times[0,n])}{n}=\limsup_{n\rightarrow\infty}\frac{\mathcal{B}_{\alpha_{1},\alpha_{2}}^{a_{1}^{-1}b_{1}|t_{1}|^{\alpha_{1}},0}(x,[-n,n]\times[0,n])}{n}<\infty.$
Moreover, we have
$\displaystyle\mathbb{P}\left\\{\int_{E_{2}(u)}\mathbb{I}(X(t)>u)dt>v(u)x\right\\}\geq\Sigma_{4}(u,n)-\Sigma\Sigma_{4}(u,n),$
where $\Sigma\Sigma_{4}(u,n)$ is defined in (61) and
$\displaystyle\Sigma_{4}(u,n)=\sum_{|2k_{2}|\leq
N_{2}^{\prime}(u,n)-1}\mathbb{P}\left\\{\int_{\hat{I}_{2k_{2}}(u,n)}\mathbb{I}(X(t)>u)dt>v(u)x\right\\}.$
By (5.4), for $x\geq 0$ we have
$\displaystyle\Sigma_{4}(u,n)$ $\displaystyle\sim$
$\displaystyle\mathcal{B}_{\alpha_{1},\alpha_{2}}^{a_{1}^{-1}b_{1}|t_{1}|^{\alpha_{1}},0}(x,[-n,n]\times[0,n])\sum_{|2k_{2}|\leq
N_{2}^{\prime}(u,n)-1}\Psi(u_{k_{2},n}^{-})$ $\displaystyle\sim$
$\displaystyle\frac{\mathcal{B}_{\alpha_{1},\alpha_{2}}^{a_{1}^{-1}b_{1}|t_{1}|^{\alpha_{1}},0}(x,[-n,n]\times[0,n])}{2n}\Theta_{1}(u),\quad
u\rightarrow\infty.$
By Lemma 6.2, for $u$ and $n$ sufficiently large, we have
$\Sigma\Sigma_{4}(u,n)\leq e^{-\mathbb{C}_{1}n^{\mathbb{C}}}\Theta_{1}(u).$
In view of (66) for the upper bound, we have
$\displaystyle\liminf_{n\rightarrow\infty}\frac{\mathcal{B}_{\alpha_{1},\alpha_{2}}^{a_{1}^{-1}b_{1}|t_{1}|^{\alpha_{1}},0}(x,[-n,n]\times[0,n])}{n}$
$\displaystyle\geq$
$\displaystyle\frac{\mathcal{B}_{\alpha_{1},\alpha_{2}}^{a_{1}^{-1}b_{1}|t_{1}|^{\alpha_{1}},0}(x,[-n_{1},n_{1}]\times[0,n_{1}])}{n_{1}}-e^{-\mathbb{C}_{1}n_{1}^{\mathbb{C}}}.$
Noting that for $n>\sqrt{x}$
$\displaystyle\mathcal{B}_{\alpha_{1},\alpha_{2}}^{a_{1}^{-1}b_{1}|t_{1}|^{\alpha_{1}},0}(x,[-n,n]\times[0,n])$
$\displaystyle=$
$\displaystyle\int_{\mathbb{R}}\mathbb{P}\left\\{\int_{[-n,n]\times[0,n]}\mathbb{I}\left(\sum_{i=1}^{2}(B_{\alpha_{i}}(t_{i})-\left|t_{i}\right|^{\alpha_{i}})-a_{1}^{-1}b_{1}|t_{1}|^{\alpha_{1}}>s\right)dt>x\right\\}e^{s}ds$
$\displaystyle\geq$
$\displaystyle\int_{\mathbb{R}}\mathbb{P}\left\\{\inf_{t\in[-n,n]\times[0,n]}\left(\sum_{i=1}^{2}\left(B_{\alpha_{i}}(t_{i})-\left|t_{i}\right|^{\alpha_{i}}\right)-a_{1}^{-1}b_{1}|t_{1}|^{\alpha_{1}}\right)>s\right\\}e^{s}ds>0,$
and by the monotonicity of
$\mathcal{B}_{\alpha_{1},\alpha_{2}}^{a_{1}^{-1}b_{1}|t_{1}|^{\alpha_{1}},0}(x,[-n,n]\times[0,n])$
with respect to $n$, we have, for $n_{1}$ sufficiently large,
$\displaystyle\liminf_{n\rightarrow\infty}\frac{\mathcal{B}_{\alpha_{1},\alpha_{2}}^{a_{1}^{-1}b_{1}|t_{1}|^{\alpha_{1}},0}(x,[-n,n]\times[0,n])}{n}$
$\displaystyle\geq$
$\displaystyle\frac{\mathcal{B}_{\alpha_{1},\alpha_{2}}^{a_{1}^{-1}b_{1}|t_{1}|^{\alpha_{1}},0}(x,[-n_{1},n_{1}]\times[0,n_{1}])}{n_{1}}-e^{-\mathbb{C}_{1}n_{1}^{\mathbb{C}}}>0.$
This completes the proof of (ii).
Ad (iii). We follow notation introduced in the proof of Proposition 3.2 for
the case $\alpha_{i}=\beta_{i}$, $i=1,2$ Observe that
(67)
$\displaystyle\Sigma_{5}(u,n)\leq\mathbb{P}\left\\{\int_{E^{\prime}(u,n)}\mathbb{I}(X(t)>u)dt>v(u)x\right\\}\leq\Sigma_{5}(u,n)+\Sigma\Sigma_{5}(u,n),$
where $E^{\prime}(u,n)=\bigcup_{(k_{1},k_{2})\in K_{u,n}}I_{k_{1},k_{2}}(u,n)$
and
$\Sigma_{5}(u,n)=\mathbb{P}\left\\{\int_{\hat{I}(u,n)}\mathbb{I}(X(t)>u)dt>v(u)x\right\\},$
$\Sigma\Sigma_{5}(u,n)=\sum_{|k_{i}|\leq
N_{i}^{\prime}(u,n),k_{i}\neq-1,0,i=1,2}\mathbb{P}\left\\{\sup_{t\in
I_{k_{1},k_{2}}(u,n)}\overline{X}(t)>u_{n,k_{1},k_{2}}^{-}\right\\},$
with $u_{n,k_{1},k_{2}}^{-}$ defined in (21) and $\hat{I}(u,n)$ in (28). In
light of (22) and (10), we have that for $u$ sufficiently large
$\displaystyle\Sigma\Sigma_{5}(u,n)$ $\displaystyle\leq$
$\displaystyle\mathcal{B}_{\alpha_{1},\alpha_{2}}(x,[0,n]^{2})\sum_{|k_{i}|\leq
N_{i}^{\prime}(u,n),k_{i}\neq-1,0,i=1,2}\Psi(u_{n,k_{1},k_{2}}^{-})$
$\displaystyle\leq$
$\displaystyle\mathcal{B}_{\alpha_{1},\alpha_{2}}(x,[0,n]^{2})\Psi(u)\sum_{|k_{i}|\leq
N_{i}^{\prime}(u,n),k_{i}\neq-1,0,i=1,2}e^{-a_{1}^{-1}b_{1}|k_{1}^{*}n|^{\beta_{1}}-a_{2}^{-1}b_{2}|k_{2}^{*}n|^{\beta_{2}}}$
$\displaystyle\leq$
$\displaystyle\mathcal{B}_{\alpha_{1},\alpha_{2}}(x,[0,n]^{2})e^{-Q_{1}(n^{\beta_{1}}+n^{\beta_{2}})}\Psi(u),$
where $k_{i}^{*}=k_{i}I_{\\{k_{i}>0\\}}+(|k_{i}|-1)I_{\\{k_{i}<0\\}},i=1,2.$
Hence dividing (67) by $\Psi(u)$ and letting $u\rightarrow\infty$, we have for
any $n,n_{1}>\sqrt{x}$
$\displaystyle
0<\mathcal{B}_{\alpha_{1},\alpha_{2}}^{a_{1}^{-1}b_{1}|t_{1}|^{\alpha_{1}},a_{2}^{-1}b_{2}|t_{2}|^{\alpha_{2}}}(x,[-n,n]^{2})\leq\mathcal{B}_{\alpha_{1},\alpha_{2}}^{a_{1}^{-1}b_{1}|t_{1}|^{\alpha_{1}},a_{2}^{-1}b_{2}|t_{2}|^{\alpha_{2}}}(x,[-n_{1},n_{1}]^{2})+\mathcal{B}_{\alpha_{1},\alpha_{2}}(x,[0,n_{1}]^{2})e^{-Q_{1}(n_{1}^{\beta_{1}}+n_{1}^{\beta_{2}})}.$
Letting $n\rightarrow\infty$ with $n_{1}$ fixed in the above inequality, we
complete the proof. $\Box$
Proof of (23): Observe that
$\displaystyle\frac{\Psi(u_{n,k_{1},k_{2}}^{-})}{\Psi(u_{n,k_{1},k_{2}}^{+})}\sim
e^{\frac{\left(u_{n,k_{1},k_{2}}^{+}\right)^{2}-\left(u_{n,k_{1},k_{2}}^{-}\right)^{2}}{2}},\quad
u\rightarrow\infty$
uniformly with respect to $0\leq|k_{i}|\leq N_{i}{{}^{\prime}}(u,n),i=1,2$.
Furthermore, by (10), for $u$ sufficiently large
$\displaystyle\left(u_{n,k_{1},k_{2}}^{+}\right)^{2}-\left(u_{n,k_{1},k_{2}}^{-}\right)^{2}$
$\displaystyle=$ $\displaystyle u^{2}\left(\sup_{t\in
I_{k_{1},k_{2}}(u,n)}\frac{1}{{\sigma^{2}(t)}}-\inf_{t\in
I_{k_{1},k_{2}}(u,n)}\frac{1}{{\sigma^{2}(t)}}\right)$ $\displaystyle=$
$\displaystyle u^{2}\sup_{s,t\in
I_{k_{1},k_{2}}(u,n)}\frac{{\sigma^{2}(t)}-{\sigma^{2}(s)}}{{\sigma^{2}(t)}{\sigma^{2}(s)}}$
$\displaystyle\leq$ $\displaystyle 4u^{2}\sup_{s,t\in
I_{k_{1},k_{2}}(u,n)}|{\sigma(t)}-{\sigma(s)}|$ $\displaystyle=$
$\displaystyle 4u^{2}\sup_{s,t\in
I_{k_{1},k_{2}}(u,n)}\left|(1+e(t))\sum_{i=1}^{2}b_{i}|t_{i}|^{\beta_{i}}-(1+e(s))\sum_{i=1}^{2}b_{i}|s_{i}|^{\beta_{i}}\right|$
$\displaystyle\leq$ $\displaystyle 4u^{2}\sup_{s,t\in
I_{k_{1},k_{2}}(u,n)}\left|\sum_{i=1}^{2}b_{i}|t_{i}|^{\beta_{i}}-\sum_{i=1}^{2}b_{i}|s_{i}|^{\beta_{i}}\right|+8u^{2}\sup_{t\in
I_{k_{1},k_{2}}(u,n)}|e(t)|\sum_{i=1}^{2}b_{i}|t_{i}|^{\beta_{i}}$
$\displaystyle\leq$ $\displaystyle
4u^{2}\sum_{i=1}^{2}b_{i}\beta_{i}|\theta_{i}|^{\beta_{i}-1}v_{i}(u)n+8u^{2}\sup_{t\in
I_{k_{1},k_{2}}(u,n)}|e(t)|\sum_{i=1}^{2}b_{i}|t_{i}|^{\beta_{i}},$
where
$e(t)=\frac{1-{\sigma(t)}}{\sum_{i=1}^{2}b_{i}|t_{i}|^{\beta_{i}}}-1,|t|\neq
0$ and $\theta_{i}\in(k_{i}v_{i}(u)n,(k_{i}+1)v_{i}(u)n)$. Using the fact that
$N_{i}^{\prime}(u,n)=\left[\frac{(e_{u}^{-1/4}\wedge\ln
u)^{2/\beta_{i}}}{u^{2/\beta_{i}}v_{i}(u)n}\right]~{}\text{and}~{}\lim_{u\rightarrow\infty}e_{u}=0,$
we have that
$u^{2}\sup_{t\in
I_{k_{1},k_{2}}(u,n)}|e(t)|\sum_{i=1}^{2}b_{i}|t_{i}|^{\beta_{i}}\leq
2e_{u}\sum_{i=1}^{2}b_{i}(e_{u}^{-1/4}\wedge\ln u)^{2}\rightarrow 0,$
as $u\rightarrow\infty$ uniformly with respect to $0\leq|k_{i}|\leq
N_{i}{{}^{\prime}}(u,n),i=1,2$. For $\beta_{i}\geq 1,i=1,2,$
$\displaystyle
u^{2}\sum_{i=1}^{2}b_{i}\beta_{i}|\theta_{i}|^{\beta_{i}-1}v_{i}(u)n$
$\displaystyle\leq$ $\displaystyle
u^{2}\sum_{i=1}^{2}b_{i}\beta_{i}\left(\frac{\ln
u}{u}\right)^{\frac{2(\beta_{i}-1)}{\beta_{i}}}v_{i}(u)n$ $\displaystyle\leq$
$\displaystyle\sum_{i=1}^{2}2a_{i}^{-1/\alpha_{i}}b_{i}\beta_{i}u^{2/\beta_{i}-2/\alpha_{i}}(\ln
u)^{\frac{2(\beta_{i}-1)}{\beta_{i}}}n\rightarrow 0,\quad u\rightarrow\infty$
uniformly with respect to $0\leq|k_{i}|\leq N_{i}{{}^{\prime}}(u,n),i=1,2$,
where $(\theta_{1},\theta_{2})\in I_{k_{1},k_{2}}(u,n)$. For
$0<\beta_{i}<1,i=1,2,$
$\displaystyle u^{2}\sup_{s,t\in
I_{k_{1},k_{2}}(u,n)}\left|\sum_{i=1}^{2}b_{i}|t_{i}|^{\beta_{i}}-\sum_{i=1}^{2}b_{i}|s_{i}|^{\beta_{i}}\right|$
$\displaystyle\leq$ $\displaystyle
u^{2}\sum_{i=1}^{2}b_{i}\beta_{i}|\theta_{i}|^{\beta_{i}-1}v_{i}(u)n$
$\displaystyle\leq$ $\displaystyle
u^{2}\sum_{i=1}^{2}b_{i}\beta_{i}|v_{i}(u)n|^{\beta_{i}}\rightarrow 0,\quad
u\rightarrow\infty,$
holds uniformly for $0\leq|k_{i}|\leq
N_{i}{{}^{\prime}}(u,n),k_{i}\neq-1,0,i=1,2.$ For
$0<\beta_{i}<1,k_{i}=-1,0,i=1,2$
$\displaystyle u^{2}\sup_{s,t\in
I_{k_{1},k_{2}}(u,n)}\left|\sum_{i=1}^{2}b_{i}|t_{i}|^{\beta_{i}}-\sum_{i=1}^{2}b_{i}|s_{i}|^{\beta_{i}}\right|$
$\displaystyle\leq$ $\displaystyle u^{2}\sup_{s,t\in
I_{k_{1},k_{2}}(u,n)}\left(\sum_{i=1}^{2}b_{i}|t_{i}|^{\beta_{i}}+\sum_{i=1}^{2}b_{i}|s_{i}|^{\beta_{i}}\right)$
$\displaystyle\leq$ $\displaystyle
2u^{2}\sum_{i=1}^{2}b_{i}|v_{i}(u)n|^{\beta_{i}}$ $\displaystyle=$
$\displaystyle
2\sum_{i=1}^{2}a_{i}^{-\beta_{i}/\alpha_{i}}b_{i}n^{\beta_{i}}u^{2-2\beta_{i}/\alpha_{i}}\rightarrow
0,\quad u\rightarrow\infty.$
Therefore, we can conclude that
$\displaystyle\left(u_{n,k_{1},k_{2}}^{+}\right)^{2}-\left(u_{n,k_{1},k_{2}}^{-}\right)^{2}\rightarrow
0$
as $u\rightarrow\infty$ uniformly with respect to $0\leq|k_{i}|\leq
N_{i}^{\prime}(u,n),i=1,2$ establishing the proof. $\Box$
Acknowledgement: K. Dȩbicki was partially supported by NCN Grant No
2018/31/B/ST1/00370 (2019-2022). Partial financial support from Swiss National
Science Foundation Grant 200021-196888 is kindly acknowledged. Additionally,
this project is financed by the Ministry of Science and Higher Education in
Poland under the programme ”Regional Initiative of Excellence” 2019 - 2022
project number 015/RID/2018/19.
## References
* [1] R. J. Adler and J. E. Taylor. Random fields and geometry. Springer Monographs in Mathematics. Springer, New York, 2007.
* [2] L. Bai. Extremes of Gaussian chaos procsses with trend. J. Math. Anal. Appl., 473(2):1358–1376, 2019.
* [3] S. M. Berman. Excursions of stationary gaussian processes above high moving barriers. The Annals of Probability, pages 365–387, 1973.
* [4] S. M. Berman. Sojourns and extremes of stationary processes. Ann. Probab., 10(1):1–46, 1982.
* [5] S. M. Berman. The maximum of a Gaussian process with nonconstant variance. Ann. Inst. H. Poincaré Probab. Statist., 21(4):383–391, 1985\.
* [6] S. M. Berman. Sojourns above a high level for a Gaussian process with a point of maximum variance. Comm. Pure Appl. Math., 38(5):519–528, 1985.
* [7] S. M. Berman. Sojourns and Extremes of Stochastic Processes. The Wadsworth & Brooks/Cole Statistics/Probability Series. Wadsworth & Brooks/Cole Advanced Books & Software, Pacific Grove, CA, 1992.
* [8] K. Dȩbicki, E. Hashorva, and L. Ji. Extremes of a class of nonhomogeneous Gaussian random fields. Ann. Probab., 44(2):984–1012, 2016.
* [9] K. Dȩbicki, E. Hashorva, and P. Liu. Extremes of $\gamma$-reflected Gaussian process with stationary increments. ESAIM Probab. Stat., 21:495–535, 2017.
* [10] K. Dȩbicki, E. Hashorva, and P. Liu. Uniform tail approximation of homogenous functionals of Gaussian fields. Adv. Applied Probab, 49(4):1037–1066, 2017.
* [11] K. Dȩbicki, E. Hashorva, and L. Wang. Extremes of vector-valued Gaussian processes. Stochastic Process. Appl., 130(9):5802–5837, 2020.
* [12] K. Dȩbicki and P. Liu. Extremes of stationary Gaussian storage models. Extremes, 19(2):273–302, 2016.
* [13] K. Dȩbicki and P. Liu. Extrems of nonstationary Gaussian fluid queues. Adv. Applied Probab, 50(3):887–917, 2018.
* [14] K. Dȩbicki, P. Liu, and Z. Michna. Sojourn times of Gaussian process with trends. J. Theoret. Probab., 33:2119–2166, 2020.
* [15] K. Dȩbicki and M. Mandjes. Exact overflow asymptotics for queues with many Gaussian inputs. J. Appl. Probab., 40(3):704–720, 2003.
* [16] K. Dȩbicki, Z. Michna, and X. Peng. Approximation of sojourn times of Gaussian processes. Methodol. Comput. Appl. Probab., 21(4):1183–1213, 2019.
* [17] K. Dȩbicki. Ruin probability for Gaussian integrated processes. Stochastic Process. Appl., 98(1):151–174, 2002.
* [18] K. Dȩbicki, E. Hashorva, and P. Liu. Extremes of Gaussian processes with regularly varying dependence structure. Extremes, 20(2):333–392, 2017.
* [19] A. B. Dieker. Extremes of Gaussian processes over an infinite horizon. Stochastic Process. Appl., 115(2):207–248, 2005.
* [20] E. Hashorva and L. Ji. Piterbarg theorems for chi-processes with trend. Extremes, 18(1):37–64, 2015.
* [21] J. Hüsler and V. I. Piterbarg. Extremes of a certain class of Gaussian processes. Stochastic Process. Appl., 83(2):257–271, 1999.
* [22] J. Hüsler and V. I. Piterbarg. On the ruin probability for physical fractional Brownian motion. Stochastic Process. Appl., 113(2):315–332, 2004.
* [23] M. Mandjes. Large deviations for Gaussian queues. John Wiley & Sons, Ltd., Chichester, 2007. Modelling communication networks.
* [24] I. Norros. A storage model with self-similar input. Queueing Systems Theory Appl., 16(3-4):387–396, 1994.
* [25] V. I. Piterbarg. High excursions for nonstationary generalized chi-square processes. Stochastic Process. Appl., 53(2):307–337, 1994.
* [26] V. I. Piterbarg. Asymptotic methods in the theory of Gaussian processes and fields, volume 148 of Translations of Mathematical Monographs. American Mathematical Society, Providence, RI, 1996.
* [27] V. I. Piterbarg. Large deviations of a storage process with fractional Browanian motion as input. Extremes, 4:147–164, 2001.
* [28] V. I. Piterbarg. Twenty Lectures About Gaussian Processes. Atlantic Financial Press, London, New York, 2015.
|
# Shape or Texture: Understanding
Discriminative Features in CNNs
Md Amirul Islam1,6, Matthew Kowal1, Patrick Esser3, Sen Jia2, Björn Ommer3,
Konstantinos G. Derpanis1,5,6 & Neil Bruce4,6
1Department of Computer Science, Ryerson University, Canada
2University of Waterloo, Canada
3IWR, HCI, Heidelberg University, Germany
4School of Computer Science, University of Guelph, Canada
5Samsung AI Centre Toronto, Canada
6Vector Institute for AI, Canada
<EMAIL_ADDRESS><EMAIL_ADDRESS>
<EMAIL_ADDRESS><EMAIL_ADDRESS>
###### Abstract
Contrasting the previous evidence that neurons in the later layers of a
Convolutional Neural Network (CNN) respond to complex object shapes, recent
studies have shown that CNNs actually exhibit a ‘texture bias’: given an image
with both texture and shape cues (e.g., a stylized image), a CNN is biased
towards predicting the category corresponding to the texture. However, these
previous studies conduct experiments on the final classification output of the
network, and fail to robustly evaluate the bias contained (i) in the latent
representations, and (ii) on a per-pixel level. In this paper, we design a
series of experiments that overcome these issues. We do this with the goal of
better understanding what type of shape information contained in the network
is discriminative, where shape information is encoded, as well as when the
network learns about object shape during training. We show that a network
learns the majority of overall shape information at the first few epochs of
training and that this information is largely encoded in the last few layers
of a CNN. Finally, we show that the encoding of shape does not imply the
encoding of localized per-pixel semantic information. The experimental results
and findings provide a more accurate understanding of the behaviour of current
CNNs, thus helping to inform future design choices.
## 1 Introduction
Convolutional neural networks (CNNs) have achieved unprecedented performance
in various computer vision tasks, such as image classification (Krizhevsky et
al., 2012; Simonyan & Zisserman, 2015; He et al., 2016), object detection (Ren
et al., 2015; He et al., 2017) and semantic segmentation (Long et al., 2015;
Chen et al., 2017; Islam et al., 2017). Despite their black box nature,
various studies have shown that early layers in CNNs activate for low-level
patterns, like edges and blobs, while deeper layers activate for more complex
and high-level patterns (Zeiler & Fergus, 2014; Springenberg et al., 2014).
The intuition is that this hierarchical learning of latent representations
allows CNNs to recognize complex object shapes to correctly classify images
(Kriegeskorte, 2015). In contrast, recent works (Brendel & Bethge, 2019;
Hermann & Lampinen, 2020) have argued that CNNs trained on ImageNet (IN) (Deng
et al., 2009) classify images mainly according to their texture, rather than
object shape. These conflicting results have large implications for the field
of computer vision as it suggests that CNNs trained for image classification
might be making decisions based largely off spurious correlations rather than
a full understanding of different object categories.
One example of these spurious correlations is how the Inception CNN (Szegedy
et al., 2015) recognizes the difference between ‘Wolf’ and ‘Husky’, based on
whether there is snow in the background (Tulio Ribeiro et al., 2016).
Recognizing object shapes is important for the generalization to out-of-domain
examples (e.g., few-shot learning), as shape is more discriminative than
texture when
Figure 1: A shape biased model (trained on Stylized ImageNet) makes
predictions based on the object’s shape, or does it? Extracting binary
($3^{\text{rd}}$ column) and semantic ($4^{\text{th}}$ col.) segmentation maps
with a one convolutional layer read-out module shows that, while the model
classifies the image level shape label correctly as a ‘bird’, it fails to
encode the full object shape ($3^{\text{rd}}$ col.) as well as fails to
categorically assign every object pixel to the ‘bird’ class ($4^{\text{th}}$
col.).
texture-affecting phenomena arise, such as lighting, shading, weather, motion
blur, or when switching between synthetic and real data. In addition to
performance, identifying the discriminative features that CNNs use for
decision making is critical for the transparency and further improvements of
computer vision models. While the model may achieve good performance for a
certain task, it cannot communicate to the user about the reasons it made
certain predictions. In other words, successful models need to be good, and
interpretable (Lipton, 2019). This is crucial for many domains where causal
mechanisms should play a significant role in short or long-term decision
making such as healthcare (e.g., what in the MRI indicates a patient has
cancer?). Additionally, if researchers intend for their algorithms to be
deployed, there must be a certain degree of trust in the decision making
algorithm.
One downside of the increasing abstraction capabilities of deep CNNs is the
lack of interpretability of the latent representations due to hidden layer
activations coding semantic concepts in a distributed fashion (Fong & Vedaldi,
2018). It has therefore been difficult to precisely quantify the type of
information contained in the latent representations of CNNs. Some methods have
looked at ways to analyze the latent representations of CNNs on a neuron-to-
neuron level. For instance, (Bau et al., 2017) quantify the number of
interpretable neurons for a CNN by evaluating the semantic segmentation
performance of an individual neuron from an upsampled latent representation.
Later work (Fong & Vedaldi, 2018) then removed the assumption that each neuron
encodes a single semantic concept. These works successfully quantify the
number of filters that recognize textures or specific objects in a CNN, but do
not identify shape information within these representations.
The most similar works to ours are those that aim to directly quantify the
shape information in CNNs. For example, (Geirhos et al., 2018) analyzed the
outputs of CNNs on images with conflicting shape and texture cues. By using
image stylization (Huang & Belongie, 2017), they generated the Stylized
ImageNet dataset (SIN), where each image has an associated shape and texture
label. They then measured the ‘shape bias’ and ‘texture bias’ of a CNN by
calculating the percentage of images a CNN predicts as either the shape or
texture label, respectively. They conclude that CNNs are ‘texture biased’ and
make predictions mainly from texture in an image. This metric has been used in
subsequent work exploring shape and texture bias in CNNs (Hermann & Kornblith,
2019); however, the method only compares the output of a CNN, and fails to
robustly quantify the amount of shape information contained in the latent
representations (note that they refer to ‘shape’ as the entire 3D form of an
object, including contours that are not part of the silhouette, while in our
work, we define ‘shape’ as the 2D class-agnostic silhouette of an object).
Thus, the method from (Hermann & Kornblith, 2019) cannot answer a question of
focus in our paper: ‘What fraction of the object’s shape is actually encoded
in the latent representation?’. Further, as their metric for shape relies
solely on the semantic class label, it precludes them from evaluating the
encoded shape and associated categorical information on a per-pixel level. For
instance, we show in Fig. 1 that shape biased models (i.e., trained on
stylized images) do not classify images based on the entire object shape: even
though the CNN correctly classifies the image as a bird, only the partial
binary mask (i.e., ‘shape’) can be extracted from the latent representations
and it cannot attribute the correct class label to the entire object region
(i.e., semantic segmentation mask).
Contributions. To address these issues, we perform an empirical study on the
ability of CNNs to encode shape information on a neuron-to-neuron and per-
pixel level. To quantify these two aspects, we first approximate the mutual
information of latent representations between pairs of semantically related
images which allows us to estimate the number of dimensions in the feature
space dedicated to encoding shape and texture. We then propose a simple
strategy to evaluate the amount of shape information contained in the internal
representations of a CNN, on a per-pixel level. The latter technique is
utilized to distinguish the quality of different shape encodings, regardless
of the number of neurons used in each encoding. After showing the efficacy of
the two methods, we reveal a number of meaningful properties of CNNs with
respect to their ability to encode shape information, including the following:
(i) Biasing a CNN towards shape predominantly changes the number of shape
encoding neurons in the last feature encoding stage. (ii) When a CNN is
trained on ImageNet, the majority of shape information is learned during the
first few epochs. (iii) A significant amount of shape is encoded in the early
layers of CNNs, which can be utilized to extract additional shape information
from the network, by combining with shape encodings from deeper layers. (iv)
Encoding the shape and class of an object does not imply the useful encoding
of localized per-pixel categorical information. All code will be released to
reproduce data and results.
## 2 Do CNNs Spend More Learning Capacity on Shape or Texture?
With the goal of revealing the characteristics of where, when, and how much
shape information is encoded in CNNs, we first aim to quantify the number of
dimensions which encode shape in a CNN’s latent representation. This analysis
on the latent representations will allow us to determine where the network
spends learning capacity on shape, while other methods that focus solely on
the network outputs have difficulty measuring the difference in shape
information between convolutional layers.
Figure 2: Illustration of the techniques used to quantify shape in this paper.
(A) Estimating the dimensionality of semantic concepts in latent
representations: We stylize each image with five textures to generate image
pairs which share the semantic concepts shape (right pair) and texture (left
pair). We feed these image pairs (shown is shape) to an encoder, $E(\cdot)$,
and calculate the mutual information between the two latent representations,
$z^{a}$ and $z^{b}$, to estimate the dimensionality, $|z_{\text{shape}}|$. (B)
We quantify the shape information encoded in a convolutional neural network by
freezing the weights, and then training a small read-out module (i.e., three
3$\times$3 convolutional layers) on the latent representation to predict
either a binary or semantic segmentation map.
### 2.1 Estimating shape and texture dimensionality
Previous works (Bau et al., 2017; Esser et al., 2020) proposed various
mechanisms to reveal the semantic concepts encoded in latent representations
of CNNs. To quantify the amount of texture and shape information, we follow
the approach of (Esser et al., 2020), where the number of neurons that
represent a certain semantic concept is estimated. Given a pretrained CNN
encoder, $E(I)=z$, where $z$ is a latent representation, we aim to estimate
the dimensionality of the semantic concepts shape and texture within $z$. The
main idea is that the mutual information between image pairs, $I^{a}$ and
$I^{b}$, which are similar in a semantic concept, will be preserved in a
neuron $z_{i}$ only if the neuron encodes that specific semantic concept.
Hence, the mutual information between the corresponding neuron pairs,
$z_{i}^{a}=E(I^{a})$ and $z_{i}^{b}=E(I^{b})$, can be used to quantify the
degree to which a semantic concept is represented by the neuron. A simple and
efficient estimate for their mutual information
$\operatorname{MI}(z_{i}^{a},z_{i}^{b})$ can be obtained based on the
correlation coefficient $\rho_{i}$. Indeed, under the assumption that the
marginal distribution of the neuron $z_{i}$ is Gaussian, the correlation
coefficient $\rho_{i}$ provides a lower bound on the true mutual information
through the following relationship which becomes tight for jointly Gaussian
$z_{i}^{a},z_{i}^{b}$ (Kraskov et al., 2004; Foster & Grassberger, 2011).
$\operatorname{MI}(z_{i}^{a},z_{i}^{b})\geq-\frac{1}{2}\log(1-\rho_{i}^{2}),\quad\text{where
}\rho_{i}=\frac{\text{Cov}\bigl{(}z_{i}^{a},z_{i}^{b}\bigr{)}}{\sqrt{\text{Var}(z_{i}^{a})\;\text{Var}(z_{i}^{b})}}.$
(1)
To quantify how well a concept $k$ is represented in terms of the number of
neurons $|z_{k}|$ that encode the concept, we compute a score for each concept
and the relative number of neurons is determined with a softmax over these
scores and a baseline score. The latter is given by the number of neurons
$|z|$, and shape and texture scores are given by the sum of their respective
correlation coefficients $\rho_{i}^{\text{shape}}$ and
$\rho_{i}^{\text{texture}}$, which are computed according to Eq. 1 with
statistics taken over image pairs that are similar in shape and texture,
respectively. Note that $k\in\\{1,2\\}$ in our case, and the remaining
dimensions not captured in any of the two semantic factors are allocated to
the residual semantic factor, which by definition captures all other
variability in the latent representation, $z$.
Stylized PASCAL VOC 2012 Dataset. Our goal is to estimate the dimensionality
of two semantic concepts: (i) shape and (ii) texture, and analyze pixel-wise
shape information. Therefore we must generate a dataset that we can sample
image pairs which share the semantic factors shape or texture, and have per-
pixel object annotations. To accomplish this goal, we create the Stylized
PASCAL VOC 2012 (SVOC) dataset. Similar to SIN, we use the AdaIN style
transfer algorithm (Huang & Belongie, 2017) to generate stylized images from
the PASCAL VOC 2012 dataset (Everingham et al., 2010) with the same settings
and hyperparameters as in the original paper (Huang & Belongie, 2017). We
choose five random textures from the Describable Textures Dataset (Cimpoi et
al., 2014) as the styles and we stylize every PASCAL VOC image with all five
of these textures. For a fair comparison with models trained on ImageNet
variants, we take only the images from PASCAL VOC which contain a single
object. With the SVOC dataset, we can now sample image pairs which are similar
in texture, by using two images from different categories but stylized with
the same texture (Fig. 2(A) left), or shape, by using the same image stylized
with two different textures (Fig. 2(A) right).
Table 1: Dimensionality estimation of semantic factors $|z_{k}|$ for the
stage-5 latent representation. Note that the total dimension of the latent
representation, $|z|$, is 2048 for all networks, and that the remaining
dimensions are allocated to the ‘residual’ factor. (a) ResNet50 compared with
BagNets. BagNets have more neurons which encode texture than shape due to
their restricted receptive field. (b) Networks with varying levels of shape
bias. The number of neurons which encode shape correlates with shape bias. (c)
Deeper networks contain more shape encoding neurons.
Model | Factor $|z_{k}|$
---|---
Shape | Texture
ResNet50 | 349 | 692
BagNet33 | 284 | 825
BagNet17 | 278 | 839
BagNet9 | 276 | 841
Training Data | Factor $|z_{k}|$
---|---
Shape | Texture
IN | 349 | 692
SIN | 536 | 477
(SIN+IN)$\rightarrow$IN | 376 | 640
Model | Factor $|z_{k}|$
---|---
Shape | Texture
ResNet50 | 349 | 692
ResNet101 | 365 | 667
ResNet152 | 371 | 661
### 2.2 Results
We now evaluate the efficacy of the dimensionality estimation method by
comparing two networks which differ significantly in their ability to encode
shape information. The first is a standard ResNet50 architecture and the
second is the recently proposed BagNet (Brendel & Bethge, 2019). BagNets are a
modified version of ResNet50 that restrict the height and width of the
effective receptive field of the CNN to be a fixed maximum, i.e., either 9,
17, or 33 pixels. This patch-based construction precludes BagNets from
classifying images based on extended shape cues. The results of this
comparison are presented in Table 1(a) where both the ResNet50 and the BagNet
variants are trained on IN. Note that ‘Stage’ refers to a residual block in a
ResNet (i.e., there are five stages in ResNet) and all experiments in Table 1
use the stage-5 features. As expected, BagNets have more neurons encoding
texture than the ResNet50 and there is a clear correlation between the
receptive field of the network and the amount of shape encoded. As the
receptive field decreases, the number of neurons encoding texture increases
even further, while the number of neurons encoding shape decreases.
We now examine whether the ‘shape bias’ metric (Geirhos et al., 2018)
correlates with the number of shape encoding neurons. Table 1(b) compares the
estimated dimensionality of a ResNet50 trained on ImageNet against networks
which are biased towards shape using two different training strategies: (i)
training solely on SIN and (ii) training on SIN and IN simultaneously followed
by fine-tuning on IN (denoted as (SIN+IN)$\rightarrow$IN, which achieves the
best accuracy on ImageNet top-1% out of the three variations (Geirhos et al.,
2018)). ResNet50 trained on IN has far more neurons dedicated to encoding
texture than shape. There is a large difference when training a ResNet solely
on SIN, where it has less neurons which encode texture than shape. When
trained and fine-tuned on (SIN+IN) and IN, respectively, there is an increase
in the number of neurons which encode shape compared to IN. We consider if
there is any pattern in the number of neurons encoding shape or texture as the
network depth increases.
Table 2: Comparing dimensionality estimations for the texture factor between
two methods. Both methods show that more texture neurons are found in
representations with smaller receptive fields.
Method | Res152 | Res101 | Res50 | Bag33 | Bag17 | Bag9
---|---|---|---|---|---|---
Net. Diss. | 433 | 481 | 499 | 592 | 623 | 537
Dim. est. | 661 | 667 | 692 | 825 | 839 | 841
As can be seen in Table 1(c), networks have more shape and less texture
neurons as depth increases. This may be due to the increase in learning
capacity of the deeper networks, as more hierarchical representations allow
for the network to learn increasingly complex shapes compared to the shallower
networks. Further, deeper networks have stronger long range connections due to
a larger effective receptive field potentially resulting in additional shape
encoding neurons. The increase in shape understanding could be one of the
reasons why deeper networks achieve better performance on various tasks, e.g.,
image classification. Finally, we assess the consistency between the
dimensionality estimation technique and network dissection (Bau et al., 2017),
another method which estimates the number of neurons representing different
concepts (described in Sec. A.2). Since network dissection cannot estimate
shape dimensionalities, the comparison is limited to the texture dimensions
shown in Table 2. Except for the case of BagNet9 and a difference in the
absolute numbers of neurons (discussed in Sec. A.2), both methods agree about
the correlation between texture dimensionality and the receptive field, which
provides further evidence that dimensionality estimates quantify the relevance
of semantic concepts faithfully.
Stage-Wise Analysis of Shape and Texture Dimensionality. We now explore where
CNNs encode shape by applying the dimensionality estimation method with the
latent representations from ResNet50 stages one to five with different amounts
of shape bias. Due to the different dimensions
Table 3: Percentage of neurons ($|z_{k}|/|z|$) encoding different semantic
concepts, $k$, for different stages of ResNet50 trained for various levels of
shape bias. While a moderate percentage of neurons encode shape in stages
$f_{1}$, $f_{2}$, and $f_{3}$, the majority of shape neurons are found in
stage $f_{5}$. Networks with shape bias learn additional shape information in
stage $f_{5}$.
Stage | IN | SIN | (SIN+IN)$\rightarrow$IN
---|---|---|---
Factor $|z_{k}|/|z|$ | Factor $|z_{k}|/|z|$ | Factor $|z_{k}|/|z|$
Shape | Texture | Shape | Texture | Shape | Texture
$f_{1}$ | 12.5% | 42.2% | 12.5% | 42.2% | 12.5% | 42.2%
$f_{2}$ | 14.1% | 40.2% | 14.1% | 40.6% | 14.1% | 40.6%
$f_{3}$ | 14.6% | 39.5% | 14.8% | 39.5% | 14.8% | 39.5%
$f_{4}$ | 15.3% | 37.9% | 17.7% | 34.7% | 17.7% | 34.8%
$f_{5}$ | 17.0% | 33.8% | 26.2% | 23.3% | 18.4% | 31.2%
at each of these stages, we present the results as the percentage of
dimensions encoding the particular semantic factor, $|z_{k}|/|z|$, where $|z|$
refers to length of the latent representation. Table 3 shows that all stages
of the network encode shape with an increase in the last two stages. Further,
biasing the models towards shape only changes the percentage of shape encoding
in the final two stages. Beginning at the fourth stage, there is a significant
jump in the number of shape dimensions for all three models with the shape
biased models having a larger increase. At the final stage, latent
representations encode even more shape, where SIN in particular has a large
increase of 8.5%. This indicates that biasing a model towards shape mainly
affects the last two stages of the network, suggesting that future work could
focus on improving the shape bias of earlier layers. An increase in shape
dimensions is inversely proportional to the amount of texture dimensions.
Notably, from stage four to stage five, there is a large drop in the amount of
texture dimensions for all networks.
When Does Shape Become Relevant During Training? To answer the question ‘When
do models learn to encode shape and texture during training?’, we capture the
changes of shape and texture occurring over the course of training a
classifier on ImageNet (IN) and Stylized ImageNet (SIN). We obtain 18
different instances of a ResNet50 model during training on IN and SIN, each
representing a checkpoint between epochs 0 and 90 (equally distributed). For
each checkpoint, we measure the dimensionality of shape and texture semantic
factors and plot the results in Fig. 3. The shape factor in the stage-5 latent
representations for both IN (Fig. 3 middle) and SIN (Fig. 3 right) models
become increasingly more relevant during the course of training, however the
percentage of dimensions grows much larger and faster in the case of the SIN
trained model. The texture factor decreases as the training progresses in both
cases as well. For the stage-4 representation in a model trained on IN (Fig. 3
left), note that the shape encoding neurons increase only marginally over the
course of training. This further reveals that a large proportion of shape
information is encoded at the deepest layer.
Figure 3: Analyzing the number of dimensions in a ResNet50 which encode shape
($|z_{shape}|$) and texture ($|z_{texture}|$) over the course of ImageNet (IN,
left two) and Stylized ImageNet (SIN, right) training. Dimensions are
estimated using stage four, $|z^{(4)}|=1024$, and stage five,
$|z^{(5)}|=2048$, latent representations. When training begins, $z$ is very
sensitive to texture but over the course of training learns to focus on the
shape instead (faster in SIN case). The vertical lines represent multiplying
the learning rate by a factor of 0.1. Note that the estimated dimensions
differ slightly from Table 1 as we trained the IN and SIN models used in this
figure from scratch.
## 3 How Much Shape Information do CNNs Encode?
The previous section measured the dimensionality of shape encodings for
various CNNs and settings. We now aim to evaluate the quality of these
encodings, and whether more shape encoding neurons implies that more robust
shape information can be extracted from these latent representations. We also
conduct a set of experiments by targeting the shape and texture-specific
neurons (see Sec. 3.3 and Sec. A.1 for results and discussion), revealing an
additional link between the two techniques used in our paper (i.e.,
dimensionality estimation and read-out module). Hermann & Kornblith (2019)
measured the quality of shape encodings in a CNN’s latent representations by
training a linear classifier on the CNN’s late-stage features to predict the
shape label of SIN images. Quantifying shape information by using image level
labels does not allow for the per-pixel evaluation of the encoded shape, and
its relation to the associated categorical label, two key components for fully
evaluating the characteristics of shape information contained in a particular
encoding.
### 3.1 Quantifying Shape Information in CNN Latent Representations
To overcome the aforementioned issues, we consider two tasks which require a
detailed understanding of object shape: binary and semantic segmentation. A
‘shape encoding network’ (SEN), the network being analyzed, consists of a CNN
with fixed weights. We then train a shallow read-out module that takes a
latent representation from the SEN, to predict a segmentation map (i.e.,
binary or semantic). If the read-out module can accurately segment objects
with a binary mask, we conclude the SEN encodes the precise shape of the
objects of interest. Further, the read-out modules ability to perform semantic
segmentation, measures how much of this encoded shape is successfully
localized with per-pixel categorical information. We use ResNet networks of
various depths (i.e., 34, 50, and 101) as SENs with a readout module
containing either one or three convolution layers with 3$\times$3 kernels.
Table 4: Left: We measure the amount of shape encoded in frozen CNN by
training a read-out module on either binary (Bin) or semantic segmentation
(Sem) under different training settings. ‘None’: random initialization, ‘End-
to-End’: network is not frozen and trained with the read-out module, ‘IN’:
pre-trained on ImageNet. Right: Shape information contained in various shape
biased models.
Training | 1 Layer Readout | 3 Layers Readout
---|---|---
ResNet34 | ResNet50 | ResNet101 | ResNet34 | ResNet50 | ResNet101
Bin | Sem | Bin | Sem | Bin | Sem | Bin | Sem | Bin | Sem | Bin | Sem
None | 46.5 | 5.2 | 48.0 | 6.1 | 44.9 | 5.1 | 58.0 | 7.2 | 58.0 | 6.0 | 55.0 | 4.8
End-to-End | 80.2 | 63.4 | 80.2 | 62.7 | 81.0 | 65.8 | 82.1 | 67.7 | 82.2 | 68.1 | 82.9 | 71.5
IN | 66.3 | 48.1 | 70.6 | 50.9 | 72.1 | 51.9 | 78.9 | 59.1 | 79.8 | 61.6 | 80.4 | 63.4
Training | ResNet50
---|---
Bin | Sem
IN | 79.8 | 61.6
SIN | 76.4 | 53.7
(SIN+IN)$\rightarrow$IN | 77.8 | 58.0
### 3.2 Results
We use the trainaug and val split of the VOC 2012 dataset to train and test
the read-out module, respectively. The binary segmentation ground truth labels
are generated by converting all semantic categories to a single ‘object’
class. Note that the binary segmentation and semantic segmentation experiments
are done completely independently of one another.
Table 4 presents the results in terms of mean-Intersection-over-Union (mIoU)
under different initialization settings with; ‘IN’: a SEN trained for ImageNet
classification, ‘None’: a SEN with random weight initialization and without
any training, ‘End-to-End’: the SEN and readout module trained in an end-to-
end manner on either the binary (Bin) or semantic (Sem) segmentation ground
truth. The None and End-to-End networks represent lower and upper bounds for
encoding shape, respectively. All read-out modules in this section are trained
on the last layer’s latent representations. Interestingly, three convolutional
layers can extract similar amounts of shape information from the IN-SEN as the
End-to-End-SEN. For example, training the ResNet101 End-to-End-SEN for Bin
improves the mIoU by merely 2.5% compared to the IN-SEN. ImageNet trained CNNs
also contain shape encodings which successfully localize per-pixel categorical
information as well, which can be seen when comparing the performance of the
IN-SEN and End-to-End SEN, e.g., for ResNet50, the IN-SEN and End-to-End-SEN
achieve 61.6% and 68.1%, respectively. This is an interesting result
considering the difficulty of semantic segmentation and that none of the IN-
SEN weights are trained for pixel-wise objectives. Shape information also
increases relative to the depth of the network which supports the results
presented in Table 1(c). As expected, the End-to-End-SEN and IN-SEN contain
significantly more shape information in their latent representations than the
baseline None-SEN.
We now evaluate the shape information encoded in networks which have different
levels of shape bias. We compare the Bin and Sem performance of the read-out
module trained on the features of three different SENs trained on IN, SIN, and
(SIN+IN)$\rightarrow$IN. As the validation is on non-stylized images, SIN-SEN
has slightly lower performance for Bin, and significantly less performance on
Sem. Such a large difference in performance implies that while the boundary of
the object is known, it is difficult for the network to correctly assign per-
pixel categorical information, a phenomenon further explored in Sec. 3.2.1.
Interestingly, the (SIN+IN)$\rightarrow$IN-SEN also has slightly lower
performance than the IN-SEN for Bin, but does not suffer in performance as
much as the SIN-SEN in the case of Sem.
Where is Shape Information Stored?
Table 5: Shape encoding results for different stages of ResNet networks
trained on ImageNet. Combining features from early stages increases shape
encoding.
$f_{1}$ | $f_{2}$ | $f_{3}$ | $f_{4}$ | $f_{5}$ | ResNet50 | ResNet101
---|---|---|---|---|---|---
Bin | Sem | Bin | Sem
| | | | | 44.7 | 4.6 | 42.3 | 4.7
| | | | | 52.7 | 6.4 | 53.0 | 5.6
| | | | | 59.6 | 10.9 | 57.7 | 9.4
| | | | | 70.8 | 33.9 | 73.2 | 43.6
| | | | | 70.6 | 50.9 | 72.1 | 51.9
| | | | | 66.0 | 16.6 | 63.8 | 13.5
| | | | | 74.5 | 42.2 | 76.8 | 49.8
| | | | | 77.3 | 53.7 | 78.2 | 56.2
| | | | | 77.3 | 52.9 | 78.2 | 55.2
Figure 4: Stage-wise predictions of read-out module on binary (Bin) and
semantic (Sem) segmentation.
We now examine if the large amount of shape information contained in ImageNet
pretrained models is equally distributed across different stages of the CNN.
In this experiment, we train one layer read-out modules on features from
different stages, ($f_{1},f_{2},f_{3},f_{4},f_{5}$), of the SEN to examine
which stage of a CNN encodes shape information. As shown in Table 5, the read-
out module trained on the last stage features, ($f_{4},f_{5}$), achieves
higher performance compared to the earlier stage features,
($f_{1},f_{2},f_{3}$), for both Bin and Sem. This is to be expected, as
feature maps from later stages have higher channel dimensions and larger
effective receptive fields compared to the feature maps extracted from earlier
layers. A surprising amount of shape information (i.e., Bin) can be extracted
from stages $f_{1},f_{2}$ and $f_{3}$; however, these features lack high-level
semantics to correlate with this shape information, which can be observed as
the corresponding Sem performance is much lower. Figure 4 reveals this
phenomenon; the horse and person are outlined even for the early stage binary
masks, but are only labelled with correct per-pixel categorical assignments in
the later stages. Considering the non-trivial amount of shape information
contained in the early stages, we investigate if aggregating multi-stage
features encodes more shape compared to the last stage feature, $f_{5}$. Table
5 (bottom) shows that training a readout module on multi-stage features
significantly improves the Bin and Sem performance, suggesting that tasks
requiring shape information may benefit from hypercolumn style architectures
(Hariharan et al., 2015). This indicates that some shape information is
encoded in earlier layers but not captured in the late stages, which agrees
with the dimensionality estimation results in Table 3, as around 12.5% and 14%
of neurons encode shape in the first stage and second stage, respectively.
Figure 5: Quantifying the shape and semantic information encoded by a CNN over
the course of ImageNet training. Vertical lines represent the learning rate
decay.
When do CNNs Encode Shape During ImageNet Training? Now we quantify the amount
of shape encoded in the latent representations over the same ImageNet training
snapshots as in Sec. 2.1. Fig. 5 shows the performance of the read-out module
trained on the frozen SEN every five epochs, as well as the ResNet50’s
validation accuracy on ImageNet classification. Note that we train a separate
read-out module for every snapshot. Similar to the findings in Fig. 3, we see
that the majority of both binary and semantic shape information is learned
within the first 10 epochs i.e., $\frac{77.1\%}{79.4\%}=97.1\%$ of the final
Bin mIoU and $\frac{52.9\%}{60.0\%}=88.2\%$ of the final Sem mIoU is obtained
by the read-out module after only 10 epochs. Contrasting purely shape related
information (i.e., Bin), a small but significant portion of per-pixel
categorical information is learned after the initial 10 epochs, when the
learning rate decay is employed.
Table 6: Shape encoding results for shape biased ResNet50s on stylized VOC12
validation set.
Training Data | ResNet50
---|---
Bin. | Sem.
IN | 45.7 | 8.5
SIN | 60.3 | 26.4
(SIN+IN)$\rightarrow$ IN | 44.5 | 9.0
Figure 6: Binary and semantic segmentation masks extracted from CNNs trained
on ImageNet (IN) and Stylized ImageNet (SIN).
#### 3.2.1 Does Knowing an Object’s Shape Imply Knowing Its Semantic Class?
We now explore whether a CNN encoding an object’s shape necessarily implies
that it also encodes the correct semantic category on a per-pixel level. In
other words, for a frozen CNN, can a read-out module (trained for binary
segmentation) successfully extract the binary mask while another read-out
module (trained for semantic segmentation) cannot successfully extract the
semantic segmentation mask? Previous results (e.g., Table 4, Table 5) show
that, for certain layers and networks, the binary segmentation performance of
a read-out module is much higher relative to the semantic segmentation
performance. This suggests that shape information (i.e., the binary mask) and
semantic information can be encoded in a mutually exclusive manner, i.e., a
CNN can encode the silhouette of the object without encoding the semantic
category of each pixel of the silhouette belongs to.
To this end, we validate various SENs and their read-out modules on stylized
VOC12 val images as this ensures the networks must encode per-pixel semantic
information based solely on the object’s shape (note that stylization removes
all texture information, see Sec. 2.2). The difference in performance between
the Bin mIoU and Sem mIoU can therefore approximate the amount of shape
information that is not correlated to its corresponding semantic class. As
shown in Table 6, the large difference in performance between Bin and Sem
suggests that these SENs capture the shape (i.e., Bin mask) of the object but
lack the ability to correctly assign per-pixel semantic labels to these
objects. Qualitative results are presented in Fig. 6; note how the binary mask
(presented as likelihood heatmaps) for the SIN trained model reasonably
segments the objects, while the semantic masks fail to resemble realistic
predictions, i.e., multiple object categories are placed spuriously over the
object of interest.
### 3.3 Targeting Shape and Texture Neurons
7$10$$20$$50$$100$$40$$60$$80$Top $X$% of Neurons Kept (%)mean IoU (%)Binary
Segmentation - IN$10$$20$$50$$100$$20$$40$$50$$60$Percentage of Neurons Kept
(%)Semantic Segmentation - IN$10$$20$$50$$100$$40$$60$$80$Percentage of
Neurons Kept (%)Binary Segmentation -
SIN$10$$20$$50$$100$$20$$40$$50$Percentage of Neurons Kept (%)Semantic
Segmentation - SIN
Figure 7: Shape encoding results by means of training a read-out module on the
latent representations of an ImageNet (i.e., texture-biased model) (left two)
and stylized ImageNet (shape-biased model)(right two) trained ResNet-50 for
binary and semantic segmentation when removing all but top $X\%$ shape or
texture-specific neurons.
In Sec. 2, we used a dimensionality estimation technique to estimate the
number of dimensions which encode shape and texture in a CNNs latent
representations. Given these neurons, we now validate if the most shape-
specific, or texture-specific, neurons can influence the performance of a
read-out module when keeping these specific neurons during training. We
hypothesize that the network trained on Stylized ImageNet (i.e., a shape
biased model) will be more reliant on the shape neurons than a network trained
on ImageNet (i.e., a texture biased model) which are known to naturally
exhibit a texture bias. See Appendix A.1 for additional experiments where we
remove the targeted neurons instead of keeping them to further validate if the
most shape or texture-specific neurons can influence the performance of a
read-out module during inference. To asses this hypothesis, we conduct a
series of read-out module experiment and the same settings as Sec. 3 are
imposed. However, during this experiment we manipulate the latent
representation as an image passes through the ResNet-50, before it is fed
through the read-out module. More specifically, we rank the neurons by mutual
information for both the shape and texture semantic factors, and then identify
the top $X\%$ of neurons from either the shape or texture neurons. Then, we
train read-out modules on the latent representations of two frozen ResNet-50s,
one trained on ImageNet (IN) and another model trained on Stylized ImageNet
(SIN). Before the latent representation is fed through the read-out module, we
remove all other neurons except for the top $X\%$ of shape or texture-specific
neurons. This forces the read-out modules to learn to perform binary
segmentation and semantic segmentation solely from the top $X\%$ of neurons
for either semantic factor, and we can identify which neurons are more heavily
relied on for each network (i.e., the shape biased or texture biased model).
Results. Figure 7 illustrates the binary and semantic segmentation results in
terms of mIoU obtained from training read-out modules on IN (left two) and SIN
((right two)) trained ResNet50s, respectively. It is clear that for the model
biased towards texture (IN pretrained), keeping texture neurons while removing
all other neurons results in a better performance than keeping only the shape
neurons. In contrast, Fig. 7 (right two) shows that for shape-biased model
(SIN pretrained), keeping shape-specific neurons achieves better performance
than keeping only texture-specific neurons. These results support the
hypothesis that the network trained on Stylized ImageNet (i.e., a shape biased
model) is not only biased towards making predictions based on object shape,
but more reliant on shape-specific neurons than a network trained on IN.
## 4 Conclusion
In this paper, we presented a systematic study of the capacity and quality of
shape encoded in a CNNs latent representations. Approximating the mutual
information between stylized PASCAL VOC images allowed us to estimate the
dimensionality of the semantic concepts shape and texture (Sec. 2.1). We also
designed a simple strategy for determining how much shape information is
encoded in these latent representations, by training a read-out module on per-
pixel binary segmentation ground truth labels. Additionally, we perform
semantic segmentation to quantify how much of this shape encoding can be
correctly attributed to per-pixel categorical information. We showed that a
model pre-trained on ImageNet has weights that contain almost all the shape
and categorical information needed to perform binary or semantic segmentation
from the late stage features. We showed that CNNs encode a surprising amount
of shape information at all stages of the network, but correctly assigning
categorical labels to the corresponding shape only occurs at the last layers
of the network, and that removing the image’s texture information severely
hurts this correspondence. Finally, we showed how removing all but a certain
number of targeted shape or texture-specific neurons affects performance
differently depending on the reliance on these neurons. These findings reveal
important mechanisms which characterize a network’s ability to encode shape
information. We anticipate these findings will be valuable for designing more
robust and trustworthy computer vision algorithms.
## Acknowledgements
The authors gratefully acknowledge financial support from the Canadian NSERC
Discovery Grants, Ontario Graduate Scholarship, and Vector Institute Post-
graduate Affiliation award. K.G.D. contributed to this work in his personal
capacity as an Associate Professor at Ryerson University. We also thank the
NVIDIA Corporation for providing GPUs through their academic program.
## References
* Bau et al. (2017) David Bau, Bolei Zhou, Aditya Khosla, Aude Oliva, and Antonio Torralba. Network dissection: Quantifying interpretability of deep visual representations. In _CVPR_ , 2017.
* Brendel & Bethge (2019) Wieland Brendel and Matthias Bethge. Approximating CNNs with bag-of-local-features models works surprisingly well on imagenet. In _ICLR_ , 2019.
* Chen et al. (2017) Liang-Chieh Chen, George Papandreou, Florian Schroff, and Hartwig Adam. Rethinking atrous convolution for semantic image segmentation. _arXiv:1706.05587_ , 2017.
* Cimpoi et al. (2014) Mircea Cimpoi, Subhransu Maji, Iasonas Kokkinos, Sammy Mohamed, and Andrea Vedaldi. Describing textures in the wild. In _CVPR_ , 2014.
* Deng et al. (2009) Jia Deng, Wei Dong, Richard Socher, Li-Jia Li, Kai Li, and Li Fei-Fei. Imagenet: A large-scale hierarchical image database. In _CVPR_ , 2009.
* Esser et al. (2020) Patrick Esser, Robin Rombach, and Björn Ommer. A disentangling invertible interpretation network for explaining latent representations. In _CVPR_ , 2020.
* Everingham et al. (2010) M. Everingham, L. Van Gool, C. K. I. Williams, J. Winn, and A. Zisserman. The PASCAL Visual Object Classes Challenge 2010 (VOC2010) Results. http://www.pascal-network.org/challenges/VOC/voc2010/workshop/index.html, 2010\.
* Fong & Vedaldi (2018) Ruth Fong and Andrea Vedaldi. Net2vec: Quantifying and explaining how concepts are encoded by filters in deep neural networks. In _CVPR_ , 2018.
* Foster & Grassberger (2011) David Foster and Peter Grassberger. Lower bounds on mutual information. _Physical review. E, Statistical, nonlinear, and soft matter physics_ , 2011.
* Geirhos et al. (2018) Robert Geirhos, Patricia Rubisch, Claudio Michaelis, Matthias Bethge, Felix A Wichmann, and Wieland Brendel. Imagenet-trained CNNs are biased towards texture; increasing shape bias improves accuracy and robustness. In _ICLR_ , 2018.
* Hariharan et al. (2015) Bharath Hariharan, Pablo Arbeláez, Ross Girshick, and Jitendra Malik. Hypercolumns for object segmentation and fine-grained localization. In _CVPR_ , 2015.
* He et al. (2016) Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. In _CVPR_ , 2016.
* He et al. (2017) Kaiming He, Georgia Gkioxari, Piotr Dollár, and Ross Girshick. Mask R-CNN. In _ICCV_ , 2017.
* Hermann & Kornblith (2019) Katherine L Hermann and Simon Kornblith. Exploring the origins and prevalence of texture bias in convolutional neural networks. _arXiv preprint arXiv:1911.09071_ , 2019.
* Hermann & Lampinen (2020) Katherine L Hermann and Andrew K Lampinen. What shapes feature representations? exploring datasets, architectures, and training. In _NeurIPS_ , 2020.
* Huang & Belongie (2017) Xun Huang and Serge Belongie. Arbitrary style transfer in real-time with adaptive instance normalization. In _ICCV_ , pp. 1501–1510, 2017.
* Islam et al. (2017) Md Amirul Islam, Mrigank Rochan, Neil DB Bruce, and Yang Wang. Gated feedback refinement network for dense image labeling. In _CVPR_ , 2017.
* Kraskov et al. (2004) Alexander Kraskov, Harald Stögbauer, and Peter Grassberger. Estimating mutual information. _Phys. Rev. E_ , 69:066138, 2004.
* Kriegeskorte (2015) Nikolaus Kriegeskorte. Deep neural networks: a new framework for modeling biological vision and brain information processing. _Annual review of vision science_ , 1:417–446, 2015.
* Krizhevsky et al. (2012) Alex Krizhevsky, Ilya Sutskever, and Geoffrey E Hinton. ImageNet classification with deep convolutional neural networks. In _NIPS_ , 2012.
* Lipton (2019) ZC Lipton. The mythos of model interpretability. arxiv 2016. _arXiv preprint arXiv:1606.03490_ , 2019.
* Long et al. (2015) Jonathan Long, Evan Shelhamer, and Trevor Darrell. Fully convolutional networks for semantic segmentation. In _CVPR_ , 2015.
* McAllester & Stratos (2020) David McAllester and Karl Stratos. Formal limitations on the measurement of mutual information. In _ICAIS_ , 2020.
* Poole et al. (2019) Ben Poole, Sherjil Ozair, Aaron Van Den Oord, Alex Alemi, and George Tucker. On variational bounds of mutual information. PMLR, 2019.
* Ren et al. (2015) Shaoqing Ren, Kaiming He, Ross Girshick, and Jian Sun. Faster R-CNN: Towards real-time object detection with region proposal networks. In _NIPS_ , 2015.
* Simonyan & Zisserman (2015) Karen Simonyan and Andrew Zisserman. Very deep convolutional networks for large-scale image recognition. In _ICLR_ , 2015.
* Springenberg et al. (2014) Jost Tobias Springenberg, Alexey Dosovitskiy, Thomas Brox, and Martin Riedmiller. Striving for simplicity: The all convolutional net. In _ICLRW_ , 2014.
* Szegedy et al. (2015) Christian Szegedy, Wei Liu, Yangqing Jia, Pierre Sermanet, Scott Reed, Dragomir Anguelov, Dumitru Erhan, Vincent Vanhoucke, and Andrew Rabinovich. Going deeper with convolutions. In _CVPR_ , 2015.
* Tschannen et al. (2020) Michael Tschannen, Josip Djolonga, Paul K. Rubenstein, Sylvain Gelly, and Mario Lucic. On mutual information maximization for representation learning. In _ICLR_ , 2020.
* Tulio Ribeiro et al. (2016) Marco Tulio Ribeiro, Sameer Singh, and Carlos Guestrin. ” Why Should I Trust You?”: Explaining the predictions of any classifier. _arXiv_ , pp. arXiv–1602, 2016.
* Zeiler & Fergus (2014) Matthew D Zeiler and Rob Fergus. Visualizing and understanding convolutional networks. In _ECCV_ , 2014.
* Zhao et al. (2020) Hengshuang Zhao, Jiaya Jia, and Vladlen Koltun. Exploring self-attention for image recognition. In _CVPR_ , 2020.
## Appendix A Appendix
### A.1 Removing Shape and Texture Neurons
In Sec. 3.3, we performed an experiment where we kept the top $X\%$ shape and
texture-specific neurons to compare how much different CNNs relied on these
neurons to encode shape. We now perform a similar experiment, but instead of
keeping the top $X\%$ neurons, we first identify the top $N$ of neurons from
either the shape or texture-specific neurons. We then remove these neurons
before passing the latent representation to the read-out module by simply
setting the features at other dimensions to zero. This allows us to identify
which neurons in each network (i.e., the shape biased or texture biased model)
are relied on more heavily to encode shape and semantic information. Note that
for this experiment, no training occurs. The goal is to simply measure the
difference in inference performance when removing $N$ shape-specific neurons,
or $N$ texture-specific neurons. Therefore we simply take the trained models
and read-out modules from Sec. 3 to perform inference while masking out the
targeted neurons. Note that validation is done on the val split from (non-
stylized) VOC 2012.
Experimental Details. For this experiment, the dimensions sharing the most
mutual information with respect to shape and texture are obtained from the
same experiments from Sec. 2. We then rank the dimensions for each semantic
factor by mutual information. Training and inference are done with the trainug
and val split, respectively, from the (non-stylized) PASCAL VOC 2012
(Everingham et al., 2010) dataset.
Table 7: Shape encoding results for ResNet50’s trained on ImageNet (IN) and
stylized ImageNet (SIN) based read-out modules when the top $N$ shape or
texture-specific neurons are removed from the latent representation during
inference. Removing the top $N$ shape specific neurons from the SIN-read-out
hurts the network’s shape-recognition abilities more compared to the IN-read-
out model.
$N$ | Shape | Texture | Residual
---|---|---|---
IN | SIN | IN | SIN | IN | SIN
Bin | Sem | Bin | Sem | Bin | Sem | Bin | Sem | Bin | Sem | Bin | Sem
0 | 70.6 | 50.9 | 76.4 | 53.7 | 70.6 | 50.9 | 76.4 | 53.7 | 70.6 | 50.9 | 76.4 | 53.7
100 | 68.8 | 46.4 | 64.6 | 37.1 | 69.3 | 44.7 | 65.7 | 39.2 | 68.3 | 46.7 | 64.7 | 37.3
200 | 67.9 | 40.0 | 57.3 | 31.4 | 67.9 | 39.7 | 62.5 | 32.7 | 66.1 | 44.2 | 64.9 | 34.8
300 | 61.6 | 38.0 | 58.6 | 25.8 | 66.3 | 37.6 | 55.0 | 27.9 | 62.9 | 40.0 | 62.3 | 30.1
#### A.1.1 Results
Table 7 presents the binary and semantic segmentation results in terms of
mIoU. We report the results under three different settings; (i) top $N$ shape-
specific neurons removed, (ii) top $N$ texture-specific neurons removed, and
(iii) top $N$ residual neurons removed. Note that for this experiment, we do
not train the read-out module. Instead, we first remove the specified neurons,
and then run inference using the pretrained IN and SIN models as well as the
already trained read-out modules. Interestingly, we find that gradually
removing the shape-specific neurons from SIN pretrained model more
significantly hurts performance than the IN pretrained model. For instance,
removing 100 shape-specific neurons from SIN achieves 37.1% sem mIoU, while
the performance dropped to 25.8% sem IoU when the top 300 shape-specific
neurons are removed (i.e., an 11.3% drop). When comparing this to the
performance drop of the IN trained model, we see that the difference is lower,
from 46.4% to 38.0% (i.e., an 8.4% drop). This further supports the hypothesis
that SIN trained models are more reliant on the individual shape encoding
neurons than the texture encoding neurons. In addition, removing shape-
specific neurons from SIN pretrained model hurts performance more than
removing the texture neurons. For example, when removing 300 shape neurons for
the SIN trained model, the performance drops to 25.8%, while removing 300
texture-specific neurons decreases the performance to only 27.9%. Finally, we
observe that removing shape or texture specific neurons hurts performance more
than removing the residual neurons. This suggests that shape and texture are
the two most important semantic factors for a network to encode and that other
semantic factors contained in the residual (e.g., color, lighting) are not as
discriminative for the task of image classification.
### A.2 Consistency of Dimensionality Estimation
Sec. 2.1 analyzed the consistency between the dimensionality estimate of
(Esser et al., 2020) and that of (Bau et al., 2017). In order to quantify
interpretability, the latter evaluates the alignment between neurons and
semantic concepts using the Broden dataset, which consists of images with
pixel-wise labelings of different semantic concepts. For each neuron $z_{i}$,
it determines the top quantile level $T_{i}$ such that the activation value of
$z_{i}$ exceeds $T_{i}$ only in $0.5\%$ of all observed cases over the
dataset, i.e. $p(z_{i}>T_{i})=0.005$. Feature maps are then upsampled to the
original image resolution and each neuron is thresholded according to its
quantile $T_{i}$ to obtain a binary segmentation mask. A neuron is then termed
a detector for a concept, if its segmentation mask has the highest
intersection over union score (IoU) for this concept, and the IoU exceeds a
threshold of $0.04$.
Because the Broden dataset contains no shape concepts, the comparison in Table
2 is limited to estimates on the number of neurons encoding texture, which is
determined by the number of detectors for concepts from the categories
material, texture and color of the Broden validation dataset.
We observe that both methods predict an inverse relationship between the
estimated texture dimensionalities and the receptive field, except for the
case of BagNet9, where a sudden drop in the number of texture detectors is
observed for network dissection. Besides this qualitative agreement, the
predicted absolute numbers differ. There are two main sources for the
incompatibility in the absolute number of neurons. First, both approaches rely
on hyperparameters,u i.e. the baseline score and the choice of the
normalization function in the case of dimensionality estimation, and the
chosen quantile and IoU threshold in the case of network dissection. Second,
the semantic meaning of texture depends on the data, i.e. dimensionality
estimation relies on the image pairs of SVOC, whereas network dissection
relies on texture images of Broden. This might also explain the drop in
detectors for BagNet9 if its receptive field is too small for some of the
Broden textures. While absolute numbers depend on hyperparameters, results
obtained with both methods are comparable across networks.
The dimensionality estimate relies on an estimate of mutual information from
samples. This remains a challenging problem, and even powerful variational
bounds exhibit either high-bias or high-variance and suffer from sensitivity
to batch sizes (Poole et al., 2019). Besides statistical limitations on the
ability to accurately estimate mutual information (McAllester & Stratos,
2020), even estimates which give neither upper nor lower bounds or those which
give loose bounds are still useful in practice. For example, (Tschannen et
al., 2020) demonstrate that loose bounds can lead to better representations
when they are learned by mutual information maximization. For dimensionality
estimation, potential biases of estimates will cancel out when comparing them
between shape and texture neurons, hence an estimate based on the correlation
is a suitable and efficient choice for our purposes.
### A.3 Estimating Shape and Texture Dimensionality of Different Networks
Trained on Stylized ImageNet
We further estimate the dimensionality of shape and texture semantic concepts
of different networks in Table 8 to test the consistency of the results
reported in Table 1 on different architectures. We run the dimensionality
estimation experiment (see Sec. 2.1) on AlexNet (Krizhevsky et al., 2012) and
VGG-16 (Simonyan & Zisserman, 2015), trained on IN and SIN. Consistent with
the findings for ResNet50, Table 8 shows that training on SIN increases the
number of dimensions encoding shape and concurrently decreases the number of
dimensions encoding texture: AlexNet-IN: [Shape=729, Texture=1299], AlexNet-
SIN: [1119, 870]. VGG-16-IN: [710, 1321], VGG-16-SIN: [1090, 879]. The
dimensionality estimation is done on the final representation before the last
linear layer for all networks.
Table 8: Comparison of shape bias and shape dimensionality for different
networks.
Network | IN | SIN
---|---|---
Factor $|z_{k}|$ | Bias | Factor $|z_{k}|$ | Bias
Shape | Texture | Shape | Texture | Shape | Texture | Shape | Texture
ResNet-50 | 14.1% | 40.2% | 22.1% | 77.9% | 26.2% | 23.3% | 81.0% | 19.0%
AlexNet | 18.0% | 30.6% | 42.9% | 57.1% | 26.0% | 21.5% | 75.5% | 24.5%
VGG-16 | 15.3% | 37.9% | 17.2% | 82.8% | 26.6% | 21.5% | 77.4% | 22.6%
### A.4 Estimating Shape and Texture Dimensionality of Self-Attention
Networks
We also experiment with the recently proposed Self-Attention Networks (Zhao et
al., 2020), which replace convolutional layers with self-attention layers.
Three different depths of Self-Attention Networks (SANs) were proposed. For
fair comparison, we experiment with SAN-19, which the authors claim is the
most similar size to ResNet50, in terms of the number of parameters in the
network. Additional, SANs come with two types of layer operations, patch-based
and pair-based. The patch-based SAN compares patches of pixels within the
attention operations, while the pair-based SAN compares individual pixels and
achieves lower performance on ImageNet. Due to the lower effective receptive
field of the pair-based SAN compared to the patch-based SAN, we expect to see
a larger number of neurons encoding shape in the patch-based SAN. The results
are shown in Table 9. The patch-based SAN19 has the largest number of shape
encoding dimensions and lowest number of texture encoding dimensions, when
compared to the pair-based SAN19 and ResNet50.
Table 9: Comparing the number of shape encoding neurons and texture encoding neurons for self-attention networks (Zhao et al., 2020). Model | Factor $|z_{k}|$ | Factor $|z_{k}|/|z|$
---|---|---
Shape | Texture | Shape | Texture
ResNet50 | 349 | 692 | 17.0% | 33.8%
SAN-19 (patch) | 384 | 610 | 18.8% | 29.8%
SAN-19 (pair) | 304 | 764 | 14.9% | 37.3%
### A.5 Layer-Wise Dimensionality Estimation on AlexNet
We now explore where another CNN encodes shape and texture at each layer of
the network. More specifically, we apply the dimensionality estimation
technique from Sec. 2.1 on AlexNet (Krizhevsky et al., 2012) on a number of
different layers. Due to the different dimensions
Table 10: Percentage of neurons ($|z_{k}|/|z|$) encoding different semantic
concepts, $k$, for different stages of AlexNet (Krizhevsky et al., 2012)
trained for various levels of shape bias.
Stage | IN | SIN
---|---|---
Factor $|z_{k}|/|z|$ | Factor $|z_{k}|/|z|$
Shape | Texture | Shape | Texture
conv1 | 17.0% | 35.0% | 16.6% | 35.3%
pool1 | 19.0% | 31.7% | 18.8% | 31.8%
conv2 | 20.7% | 27.2% | 21.1% | 27.7%
pool2 | 20.7% | 26.1% | 21.1% | 25.9%
conv3 | 20.6% | 27.0% | 23.9% | 23.2%
conv4 | 21.2% | 25.8% | 25.4% | 21.7%
conv5 | 21.3% | 24.5% | 25.2% | 21.0%
pool3 | 21.7% | 23.8% | 25.4% | 20.9%
fc6 | 18.8% | 28.7% | 24.5% | 21.9%
fc7 | 18.0% | 30.6% | 26.0% | 21.5%
at each of these stages, we present the results as the percentage of
dimensions encoding the particular semantic factor, $|z_{k}|/|z|$, where $|z|$
refers to length of the latent representation. The results are presented in
Table 10 and Fig. 8. Note that the output from the convolutional layers also
include the ReLU activation function.
8
conv1pool1conv2pool2conv3conv4conv5pool3fc6fc7$15$$20$$25$$30$LayerDimensions
(%)Shapeconv1pool1conv2pool2conv3conv4conv5pool3fc6fc7$20$$25$$30$$35$$40$LayerTexture
Figure 8: Shape (left) and texture (right) encoding dimensions estimated on
each layer of AlexNet (Krizhevsky et al., 2012). Shape biased AlexNet trained
on Stylized ImageNet (Geirhos et al., 2018) encode more shape at the later
layers of the network which is consistent with the findings for ResNets (He et
al., 2016).
|
# Reduced Order and Surrogate Models for Gravitational Waves
Manuel Tiglio111email<EMAIL_ADDRESS>Aarón Villanueva222email:
<EMAIL_ADDRESS>
(Facultad de Matemática, Astronomía, Física y Computación,
Universidad Nacional de Córdoba, Córdoba (5000), Argentina
)
###### Abstract
We present an introduction to some of the state of the art in reduced order
and surrogate modeling in gravitational wave (GW) science. Approaches that we
cover include Principal Component Analysis, Proper Orthogonal Decomposition,
the Reduced Basis approach, the Empirical Interpolation Method, Reduced Order
Quadratures, and Compressed Likelihood evaluations. We divide the review into
three parts: representation/compression of known data, predictive models, and
data analysis. The targeted audience is that one of practitioners in GW
science, a field in which building predictive models and data analysis tools
that are both accurate and fast to evaluate, especially when dealing with
large amounts of data and intensive computations, are necessary yet can be
challenging. As such, practical presentations and, sometimes, heuristic
approaches are here preferred over rigor when the latter is not available.
This review aims to be self-contained, within reasonable page limits, with
little previous knowledge (at the undergraduate level) requirements in
mathematics, scientific computing, and other disciplines. Emphasis is placed
on optimality, as well as the curse of dimensionality and approaches that
might have the promise of beating it. We also review most of the state of the
art of GW surrogates. Some numerical algorithms, conditioning details,
scalability, parallelization and other practical points are discussed. The
approaches presented are to large extent non-intrusive and data-driven and can
therefore be applicable to other disciplines. We close with open challenges in
high dimension surrogates, which are not unique to GW science.
###### Contents
1. 1 Introduction
2. 2 Reduced Order Modeling and Gravitational Waves
3. 3 Mathematical Preliminaries
4. I Representation and Compression
1. 4 Principal Component Analysis
2. 5 Proper Orthogonal Decomposition
1. 5.1 PCA and POD
3. 6 Spectral expansions
1. 6.1 Spectral methods
1. 6.1.1 Fourier Expansions
2. 6.1.2 Jacobi Polynomials
4. 7 Parametrized problems and optimal approximations
5. 8 Reduced Basis
1. 8.1 Introduction
2. 8.2 The Training set
3. 8.3 Greedy algorithms
1. 8.3.1 Convergence rates and near-optimality
2. 8.3.2 Complexity, scaling, computational aspects
5. II Predictive Models
1. 9 Polynomial interpolation
1. 9.1 Representation versus prediction
2. 9.2 Prediction through polynomial interpolation
3. 9.3 Convergence rates, collocation points and Runge’s phenomenon
4. 9.4 Discrete expansions and interpolation
5. 9.5 Multiple dimensions
2. 10 The Empirical Interpolation Method
1. 10.1 From projection to interpolation
2. 10.2 EIM algorithm
3. 10.3 Accuracy and conditioning of the EIM
3. 11 Surrogate models
1. 11.1 Surrogate models for components
2. 11.2 Empirical interpolant based surrogates
4. 12 Surrogates of compact binaries
1. 12.1 Numerical relativity binary black holes
2. 12.2 Numerical relativity hybrid binary black holes
3. 12.3 Extreme mass ratio inspirals
1. 12.3.1 Eccentric inspirals
4. 12.4 Effective One Body
5. 12.5 Post-Newtonian (PN)
6. 12.6 Ringdown
6. III Data Analysis
1. 13 Reduced Order Quadratures
1. 13.1 ROQ, other quadrature methods, dependence on dimensionality
2. 14 Accelerating Parameter Estimation with ROQ
1. 14.1 Constructing the Reduced Order Quadrature
2. 14.2 Implementation in LALInference
3. 15 Challenges, open issues and comments
4. 16 Acknowledgments
#### Notation
* •
$h_{\lambda}$: a function (for the purposes of this article, a GW waveform)
associated with parameter value $\lambda$. It can be, for example, a time
series $h_{\lambda}=h_{\lambda}(t)$, or a frequency domain function
$h_{\lambda}=h_{\lambda}(f)$.
* •
$\lambda$: a (usually multi-dimensional) parameter for $h_{\lambda}$.
* •
$\tt dim$: number of parameters of $\lambda$.
* •
$L$: number of time/frequency samples.
* •
${\cal F}$: an abstract space of functions of interest. Typically, ${\cal
F}:=\\{h_{\lambda}\,|\,\lambda\in\Phi\\}$, with $\lambda$ in some compact
region (continuous or discrete) $\Phi$. We also refer to ${\cal F}$ as the
fiducial or underlying model.
* •
$N$: size of the training set. Also, the number of points in standard
quadratures, polynomial interpolation, etc.
* •
${\cal T}$: training set of parameter points ${\cal
T}:=\\{\lambda_{i}\\}_{i=1}^{N}$. It is assumed to be compact.
* •
${\cal K}$: training set of functions in ${\cal F}$, ${\cal
K}:=\\{f_{i}\\}_{i=1}^{N}=\\{f(\lambda_{i},\cdot)\\}_{i=1}^{N}\subset{\cal
F}$.
* •
RB: Reduced Basis as a framework.
* •
$\Lambda_{i}$: selected (typically, by a greedy algorithm) parameter values.
* •
rb: a specific reduced basis.
* •
$t,f$: time, frequency domains.
* •
EIM: The Empirical Interpolation Method.
* •
$T,F$: selected (typically, by the EIM) time/frequency interpolation nodes.
* •
$n$: number of basis elements and of EIM nodes (they are equal by
construction). The goal of ROM, if the problem is amenable to dimensional
reduction, is to find an accurate basis such that $n\ll N$.
* •
$\Lambda_{n}$: the Lebesgue constant for an approximation with $n$ basis
elements/interpolation points. This involves some ambiguity: we use the same
symbol for greedy selected points, but it should be clear from context which
one we refer to.
* •
$h_{\tt s}$: a surrogate function for $h$.
* •
Boldfaces are used for matrices, for example $\bf A$ has elements $A_{ij}$.
* •
$\langle a,b\rangle$: scalar product between two vectors or functions $a,b$.
### 1 Introduction
Gravitational wave (GW) science has reached a level of maturity in which many
tools from areas such as modern approximation theory, data science, machine
learning, and artificial intelligence are being incorporated into the field.
These attempt to address challenges such as dealing with complex modeling,
analysis, and handling of big data. A common feature of these challenges is
the computational cost involved, which in many cases can be prohibitive, to
the point that it cannot be solely overcome with larger or faster
(super)computers, specialized hardware such as GPUs, or software optimization.
This is particularly the case for parametrized problems, where each query
depends on multiple input parameters that might only be known at run time.
This is exacerbated as the number of parameters grow, usually resulting in the
curse of dimensionality. This refers to the complexity of the problem (here
leaving the term complexity ambiguous on purpose) growing fast, sometimes
exponentially, with the number of parameters. In the case of gravitational
waves from binary systems, parameters can be intrinsic or extrinsic. The
former relate to parameters such as the mass and spin of the binary
components, the initial separation and eccentricity of the system, and
equations of state if matter is present. Extrinsic parameters include distance
of the source to Earth, sky position, orientation, time and phase of arrival.
One of the first challenges of a parametrized problem is sampling it. With
standard methods (for example, equally spaced, using the metric approach in
GWs, or stochastic sampling) the accuracy of such catalogs for GW detection
increases in many cases at best linearly with the number of samples, and their
sizes typically increase exponentially with the number of parameter dimensions
dim as ${\cal O}(N^{\tt dim})$. In fact, for the metric approach, the number
of templates grows as $(1-\text{MM})^{-\text{dim}/2}$, with MM the minimal
match. Producing such catalogs, as well as their storage and analysis, can
become challenging and, again, even not tractable through raw computational
power and storage.
One approach to this problem is decreased fidelity: the modeling of the
problem is simplified by approximations to Einstein’s General Relativity
equations which are cheaper to solve for and analyze. Decreased fidelity is a
delicate approach, though, since the accuracy of the approximations might not
be known without access to the high fidelity models for an arbitrary query,
which if not available could become an issue when attempting to assign error
bars and a given precision needs to be guaranteed for statistical purposes.
Decreased fidelity models can also lead to missed signals or biases in
parameter estimation of detected GWs.
Thus, another challenge is to simulate the GWs emitted by, for example, the
coalescence of compact binary objects in real time, without any physical
approximation. Here the gold standard is high accuracy numerical relativity
(NR) solutions to the full Einstein field equations. It might appear
unattainable to replace online (that is, not evaluated in advance), for an
arbitrary query, a supercomputer simulation which might take per query
hundreds to tens of thousands of hours of computing time with a substitute or
surrogate model of equal accuracy but which can be evaluated in real time on a
commodity laptop. As we show throughout this review, this can be – and is
being – done, though there are still outstanding challenges left and the
problem is not completely solved.
Finally, another challenge of equal importance is to perform parameter
estimation on the sources of any detected GW signal in quasi real time.
Meaning, fast enough so as to process the large number of detections by modern
GW laser interferometers. And, most important, in the case of sources with
electromagnetic counterparts, fast enough to allow for rapid telescope
followups. This requires both online fast evaluation of the GW waveforms, as
well as rapid likelihood evaluations.
In this review we discuss some state of the art approaches which attempt to
(and in some, but not all, cases manage to) obtain accurate and fast-to-
evaluate and analyze surrogate models of gravitational waves emitted by binary
systems, accomplishing some of the aforementioned challenges. The common
aspect underlying all these methods is reduced order modeling (ROM). The
approaches here reviewed are Principal Component Analysis (PCA), Singular
Value Decompositions (SVD), Reduced Basis (RB), the Empirical Interpolation
Method (EIM), and derivatives, such as Reduced Order Quadratures (ROQ). The
goal is intendedly not to be a definite survey, since it is a very active
field. Instead, we attempt to provide an introduction to some of the
approaches that are being used in practice and do deliver on some or several
of the above challenges, along with some basic theory underlying each of them.
The review is divided into three parts, each of which builds upon the previous
one:
1. 1.
Representation
One example is the generation of compact catalogs or banks of gravitational
waves for searches. Another one is to analyze a system and look for
redundancies.
2. 2.
Predictive Models
The most ambitious goal here is to build surrogate models that can be
evaluated in real time and are indistinguishable from numerical relativity
supercomputer simulations of the Einstein equations, without any physical
approximation. The target is the evaluation of a waveform in the order of
milliseconds per angular mode; that is, a speed up of at least $\sim 10^{8}$
with respect to NR and without loss of accuracy.
3. 3.
Data Analysis
One of the main goals of ROM and other efforts in GW science is to achieve
very fast parameter estimations, in particular so that real time alerts can be
sent for searches of electromagnetic counterparts. From an astrophysical point
of view, the target is from months using standard methods to around $10$
minutes using Focused Reduced Order Quadratures (discussed in Section 14),
including millions to tens of millions waveform evaluations and likelihood
evaluations.
We encourage the reader to provide us with feedback, including topics to cover
in future versions of this Living Review article. For briefness we have to
skip many references, we apologize in advance for any and all omissions in
such an active field.
### 2 Reduced Order Modeling and Gravitational Waves
This article reviews approaches which attempt to solve many of the
aforementioned problems in GW science through Reduced Order Modeling (ROM),
also known as Dimensional or Complexity Reduction, and the related field of
Surrogate Modeling. ROM as a field has been around for a long time, but over
the last decade and a half there have been major theoretical advances followed
by a rapid raise in the number of applications and pace at which they both
take place, in many areas of science, engineering and technology.
This is in part, again, due to powerful new approaches and results, from
approximation theory to numerical analysis and scientific computing, but also
due to the recognition of the power of dimensional reduction in many important
problems, some of them long-standing. That is, problems which are only now
being seen in the light of the latest developments of ROM. On top of that,
fields such as data science (DS), Machine Learning (ML) and Artificial
Intelligence (AI) are considerably benefiting from ROM to either eliminate
redundancies in big data, making them more amenable to analysis, and/or
identifying relevant features for further studies using techniques from these
disciplines. In fact, depending on the definitions of these fields, some
include ROM as a subdiscipline.
There is a big difference, though. There are many books on established DS, ML,
and AI practices and theory, while the literature on modern approaches to ROM
for parametrized problems, with some important exceptions that we mention in
this review, is largely composed of technical papers, which could be difficult
to grasp for practitioners as introductory material. Also, these notable
exceptions usually focus on time-independent partial differential equations
(PDEs), and even as introductory material they might be hard to absorb by non-
mathematicians. In contrast, this review covers approaches which are purely
data-driven and attempts to be amenable as an introduction for GW science
practitioners, keeping theory to the minimum necessary to build intuition on
why a given approach works as it does, what are the challenges left, and
possible approaches to them.
One of the reasons why ROM for parametrized system has been largely devoted to
time-independent PDEs is because in that case many rigorous results, such as a
priori error bounds on the reduced model, can be proved; these are called
certified approaches. Here we move beyond that particular arena to ROM for
general parametrized systems. This includes problems which might not involve
any differential equation, be time-dependent, to purely data-driven problems
such as analysis and handling of large amounts of data. This broadening is at
the expense, in many cases, of more heuristics and less rigor, such as a
posteriori validation as opposed to rigorous a priori error bounds. The
rationale for this is simply that many problems of interest in GW science are
far too complex for existing detailed rigorous theorems. On the other hand, GW
problems share many similarities with others for which approaches with proven
properties have been developed and therefore certain algorithms can be adapted
to cases of interest in GWs. In other instances it is quite the opposite:
available rigorous results are quite generic, and can be abstracted from any
previous application in which they were introduced.
Half-rigorous, half-heuristic approaches are not as pessimistic as it might
seem. Many widely used techniques in DS, ML and AI are somewhat heuristic in
nature and require a posteriori validation. Similarly with half-rigorous,
half-heuristic cutting-edge ROM approaches as those discussed here: they can
(and should) always be a posteriori validated.
Our presentation keeps technicalities and analyses to the minimum necessary
for building intuition of why a technique works as it does, under which
conditions it does so, and what can be expected given the properties of the
problem of interest. At the same time, it is not a survey of all the work done
in the field.
There are two important books about Reduced Basis; though focused on partial
differential equations there is a lot of valuable information (well beyond the
scope of this short introduction): [70] and [111]. A website for ROM with
several resources is [9], while [11] has the speakers slides from a workshop
on ROM in General Relativity; even though the latter is several years old and
there has been much progress since then, the topics covered are still highly
relevant. A very recent and lengthier, year-long program at ICERM (Brown
University) focused on ROM in GR can be found on the website [6], including
many of the slides and talks of each workshop.
### 3 Mathematical Preliminaries
This section serves as a brief recap of some basic mathematics and to
introduce some notation used hereon. Since in this review we deal with model
reduction for parametrized problems, we next consider an example amenable to
it.
Example 1.
Consider the model
$h(t;\lambda)=\exp(\lambda t)\,.$ (1)
where $\lambda$ is a complex parameter and $\operatorname{Im}({\lambda})>0$.
This represents oscillatory, exponentially damped functions; it can be seen as
a toy model for the ringdown of a black hole (discussed in Section 12.6).
In the notation that we use throughout this review, $h$ is a function of the
physical variable $t$ parametrized by $\omega$ and for compactness we often
use the following type of shortcuts:
$h=h_{\lambda}=h(\lambda)=h(t;\lambda)\,.$
Suppose now that we have a set of $N$ samples of $\lambda$, leading to a
training set
${\cal
K}=\\{h_{i}\\}=\\{h_{\lambda_{i}}\\}=\\{h(\lambda_{i})\\}=\\{h(t;\lambda_{i})\\}\,,\quad
i=1\ldots N.$
Since the functions of interest (1) are known in closed-form, building such a
training set is straightforward, though in general this is not the case (as
when solving the Einstein equations is required).
One of the goals of ROM is to find a reduced basis, that is, a subset of
${\cal K}$ with number of elements $n\leq N$ – with, hopefully, $n\ll N$ –,
such that its span (the span of a set of vectors is its set of linear
combinations) represents ${\cal K}$ with arbitrarily high accuracy. The
compression rate is then
$C_{r}:=N/n\,.$
Furthermore, in practice one also needs to discretize time (or frequency). So
let’s sample these training set elements at an arbitrary set of times
$\\{t_{i}\\}_{i=1}^{L}$, where $L$ stands for Length. Using, for example, the
Empirical Interpolation Method (EIM, discussed in Section 10), one can
subsample these time samples to a subset $l\leq L$. In fact, in EIM, if the
number of reduced basis is $n$, then by construction $l=n$.
That is, the initial set being of size $N\times L$ can now be reduced to
$n^{2}$, with a double compression rate (both in parameter and time domains),
$C_{r}:=\frac{N\times L}{n^{2}}\,.$
#### Inner products and norms for functions or vectors
Let $\Omega$ denote the physical domain. Throughout this review it represents
time or frequency interval, $[t_{\rm min},t_{\rm max}]$ or $[f_{\rm
min},f_{\rm max}]$ respectively, though in general it could be space, space-
time, or some more abstract arena. For the sake of discussion we consider time
intervals. For any two complex-valued functions, $h(t)$ and $g(t)$, we
consider inner/scalar/dot products and their corresponding norms of the form
$\langle
h_{1},h_{2}\rangle:=\int_{\Omega}\bar{h}_{1}(t)h_{2}(t)\omega(x)dx,\quad\|f\|^{2}:=\langle
f,f\rangle\,,$ (2)
where the bar over $h_{1}$ denotes complex conjugation and $\omega(x)$ is a
generic weight function. These are referred to as weighted $L_{2}$ scalar
products. In data analysis/signal processing and usually in the frequency
domain, $S:=\omega^{-1}$ characterizes the sensitivity of the detector and is
referred to as the power spectral density (PSD).
We consider also discrete inner products and norms of the form
$\langle
h_{1},h_{2}\rangle:=\sum_{i=1}^{L}\bar{h}_{1}(t_{i})h_{2}(t_{i})\omega_{i}\,\quad\|h\|^{2}:=\langle
h,h\rangle\,.$ (3)
This discrete version of an $L_{2}$ scalar product is usually denoted as
$\ell_{2}$, with a lowercase to distinguish it from the continuum case.
Whenever (3) is the discrete approximation of an integral (2) the values
$t_{i}\in\Omega$ and $\omega_{i}$ are respectively referred to as quadrature
nodes and weights. Together, $\\{t_{i},\omega_{i}\\}_{i=1}^{L}$ is called a
quadrature rule. When $\omega_{i}=1$, (3) is referred to as the Euclidean
scalar product. Throughout this review we also use the infinity, or max, norm
$\|h\|_{\infty}:=\max_{1\leq i\leq L}|h(t_{i})|\,.$
of a vector $f\in\mathbb{C}^{L}$. This norm is not induced by an inner
product. That is, it can be shown that there is no scalar product
$\langle\cdot,\cdot\rangle$ for which $\|h\|_{\infty}=\langle h,h\rangle$.
Formally speaking, (2) is defined for functions while (3) is for vectors. We
use the same notation $\langle\cdot,\cdot\rangle$ for both cases; with the
hope that the distinction is clear from the context. Some proofs are sometimes
more convenient in a continuous setting while numerical computations are
restricted to discrete values. Depending on the context, we sometimes switch
between these two settings using the same notations for both of them.
Example 2. Polynomials.
Consider the space of degree $n$ polynomials defined on $\Omega=[-1,1]$. Any
element
$f(t)=\sum_{i=0}^{n}c_{i}t^{i}\,,$ (4)
can be written as a sum of $(n+1)$ terms and such space is a linear one of
dimension $(n+1)$. The first $(n+1)$ normalized Legendre polynomials
$\\{P_{i}(t)\\}_{i=0}^{n}$ form an orthonormal basis with respect to the
scalar product (2) with $\omega(t)\equiv 1$.
#### Matrices
Consider a complex $L\times N$ matrix $\textbf{H}\in\mathbb{C}^{L\times N}$,
where $L$ is the number of rows (“length of the time series”) of H and $N$ is
the number of columns (“number of parameter samples”),
${\bf H}:=\begin{pmatrix}h_{11}&h_{12}&\cdots&h_{1N}\\\
h_{21}&h_{22}&\cdots&h_{2N}\\\ \vdots&\vdots&\ddots&\vdots\\\
h_{L1}&h_{L2}&\cdots&h_{LN}\end{pmatrix}\,.$ (5)
As one might imagine, for gravitational waves $h_{ij}=h_{j}(t_{i})$. Each
column is referred to as a snapshot. A matrix ${\bf H}$ of shape $L\times N$
appears often throughout this review, where $N$ is the number of samples in
the training set (or $n$, the dimensionality of the reduced basis) and $L$ is
the length of each time series, so it is a very concrete example.
A square matrix ${\bf H}$ is said to be non-singular if its inverse – denoted
by ${\bf H}^{-1}$ – exists, i.e. ${\bf H}{\bf H}^{-1}={\bf H}^{-1}{\bf H}={\bf
I}$, where
${\bf I}=\begin{pmatrix}1&0&\cdots&0\\\ 0&1&\cdots&0\\\
\vdots&\vdots&\ddots&\vdots\\\
0&0&\cdots&1\end{pmatrix}=:diag(1,1,\ldots,1)\,.$
The (Hermitian) transpose of ${\bf H}$, denoted by $H^{\dagger}$ is defined as
the matrix with elements
$\left({\bf H}\right)^{\dagger}_{ij}=\bar{{\bf H}}_{ji}\,,$
the bar again indicates complex conjugation, and we say that ${\bf H}$ is
symmetric (hermitian or self-adjoint, in the current context), if ${\bf
H}={\bf H}^{\dagger}$.
Eigenvectors and eigenvalues:
A vector $x\neq 0$ is an eigenvector of ${\bf H}$ with eigenvalue $\alpha$ if
${\bf H}x=\alpha x\,.$
Eigenvectors are defined up to a normalization constant; that is, if $x$ is an
eigenvector with eigenvalue $\alpha$ so is $ax$, for any non-zero $a$.
##### Matrix Norms
The Frobenius norm. Imagine “unpacking” ${\bf H}$ into a long vector of size
$L\times N$. Measuring this vector with the Euclidean norm defines the
Frobenius norm
$\|{{\bf H}}\|^{2}_{F}:=\sum_{i}^{L}\sum_{j}^{N}\,\left|h_{ij}\right|^{2}$ (6)
of a matrix ${\bf H}$.
Later we will see that the Frobenius norm plays a key role when discussing
approximation by proper orthogonal decompositions.
Induced norm.
The matrix can be viewed as a linear operator mapping vectors from
$\mathbb{C}^{N}$ to $\mathbb{C}^{L}$. Given norms for both $\mathbb{C}^{L}$
and $\mathbb{C}^{N}$, which need not be the same, the induced norm of a matrix
is defined as
$\|{{\bf H}}\|=\max_{\|x\|=1}\|{{\bf H}}x\|\,.$ (7)
This norm characterizes the maximal possible “amplification” from the
application of ${\bf H}$ to $x$. More precisely: the value of $\|{{\bf H}}\|$
is the smallest positive number $c$ such that
$\|{{\bf H}}x\|\leq c\|x\|\,.$ (8)
Notice that, following standard practice, we use the same notation – namely,
$||\cdot||$ – for both the norm of a vector ($\|x\|$ and $\|{\bf H}x\|$ in the
above definition) and a matrix ($\|{\bf H}\|$). It should be clear from the
context, though, which one we are referring to.
##### Rank and Kernel
The range of a complex $L\times N$ matrix ${\bf H}$ is
$\mbox{range}({\bf H})=\\{y\in\mathbb{C}^{L}|\,y={\bf H}x\;\;\mbox{for some
}x\in\mathbb{C}^{N}\\}\,,$
its rank
$\mbox{rank}({\bf H})=\dim\left(\mbox{range}({\bf H})\right)\,,$
its kernel
$\mbox{ker}({\bf H})=\\{x\in\mathbb{C}^{L}|\,{\bf H}x=0\\}\,,$
and its nullity
$\text{nul}({\bf H})=\text{dim}(\text{ker}({\bf H}))\,.$
As a linear algebra exercise it can be shown that
$\text{rank}({\bf H})+\text{nul}({\bf H})=N\,.$
For square $N\times N$ matrices, the following properties are equivalent:
1. 1.
${\bf A}$ is non-singular
2. 2.
$\mbox{det}({\bf A})\neq 0$
3. 3.
$\mbox{ker}({\bf A})=\\{{\bf 0}\in\mathbb{C}^{N}\\}$
4. 4.
$\mbox{rank}({\bf A})=N$
5. 5.
${\bf A}$ has linearly independent columns and vectors.
#### Approximation by projection
##### The least squares problem
It is rather easy to approximate one function by other functions, such as
Fourier, wavelets, polynomial expansions, or somewhat physically inspired
bases. In this article we focus on representing gravitational waves by
themselves. Intuitively, this should be (and is) more efficient. In this
subsection we briefly introduce projection-based approximations, with more
details given in Section 6.
To motivate the problem consider an $n$-dimensional vector space $W_{n}$ which
is itself a subspace of a Hilbert’s one ${\cal H}$ (for the purposes of this
article, this means that a scalar product is assumed to exist). A common
approximation criteria is a least squares (LS) one. That is, one seeks to
approximate $h\in{\cal H}$ by $h^{(n)}\in W_{n}$ which is the solution to
$h^{(n)}={\tt argmin}_{\tilde{h}}\|h-\tilde{h}\|^{2}\,,$ (9)
where $\tilde{h}\in W_{n}$. Equation (9) means finding the element of $W_{n}$
that minimizes the squared norm of the representation error.
If $\\{e_{i}\\}_{i=1}^{n}$ is an orthonormal basis of $W_{n}$ and
$\langle.,.\rangle$ the scalar product associated with ${\cal H}$, the
following is the (unique) solution to the LS problem,
$h^{(n)}=\sum_{i=1}^{n}c_{i}e_{i}\,,\quad\text{with}\;c_{i}=\langle
e_{i},h\rangle.$ (10)
The solution $h^{(n)}$ is the orthogonal projection of $h$ onto $W_{n}$,
therefore denoted as
${\cal P}_{n}h:=h^{(n)}\,.$
What it means is that if $h\in W_{n}$ then
${\cal P}_{n}h=h\,,$ (11)
and the residual $\delta h={\cal P}_{n}h-h$ satisfies
$\langle{\cal P}_{n}h,\delta h\rangle=0\,.$ (12)
The solution in 10 is basis-independent and $h^{(n)}$ is uniquely defined. The
orthogonal projection onto any linear space is a geometric construction, so it
is independent of the basis used to represent it: If $\\{e_{i}\\}_{i=1}^{n}$
and $\\{\tilde{e}_{i}\\}_{i=1}^{n}$ are any two orthonormal bases, and the
projection coefficients $c_{i}$ and $\tilde{c}_{i}$ are computed according to
(10) (replacing in the second case $e_{j}$ by $\tilde{e}_{j}$), then
${\cal
P}_{n}h=\sum_{i=1}^{n}c_{i}e_{i}=\sum_{i=1}^{n}\tilde{c}_{i}\tilde{e}_{i}\,.$
$\sum_{i=1}^{n}\langle e_{i},v\rangle
e_{i}=\sum_{i=1}^{n}\langle\tilde{e}_{i},v\rangle\tilde{e}_{i}\,,$ (13)
leading to the same solution of the LS problem. One might ask why the emphasis
on orthogonal or orthonormal bases; it is due to a conditioning issue which we
discuss this in Section 6. For the time being the summary is that one should
always use orthogonal or orthonormal bases. An exception is when building
predictive models, as discussed in Section 11.
##### Representing a set of elements: collective error
Suppose we seek to approximate not one function but a set of them:
${\cal F}=\\{h_{1},h_{2},\dots,h_{N}|h_{i}\in{\cal H}\\}$ (14)
From the above discussion in Section 3 we know that ${\cal P}_{n}h_{i}$ is the
the best representation of any particular $h_{i}$. The question of how one
measures the approximation error of the collective set naturally arises. Two
reasonable and standard criteria are
$\frac{1}{N}\sum_{i=1}^{N}\|h_{i}-{\cal P}_{n}h_{i}\|^{2}\,,$ (15)
and
$\max_{i}\|h_{i}-{\cal P}_{n}h_{i}\|^{2}\,,$ (16)
which have the interpretations of the mean and maximum errors, respectively.
Whether the notion of error ${\rm(1)}$ or ${\rm(2)}$ is chosen for an error
minimization criteria will lead to the main two ROM approaches discussed in
this review. Namely, Proper Orthogonal/Singular Value Decompositions, or the
Reduced Basis-greedy approach.
#### Further reading
The material discussed can be found in standard linear algebra/numerical
analysis books. There are many introductory good ones, we suggest [129] and
[130].
## Part I Representation and Compression
### 4 Principal Component Analysis
Principal Component Analysis (PCA) is perhaps one of the most used tools when
seeking for redundancy or a hierarchy of importance of variables in
statistical analysis. It has a close relationship with Proper Orthogonal
Decomposition (POD) as we discuss in Section 5.1. With this approach one seeks
to determine the most statistically relevant variables and to potentially
dimensionally discard from the problem the least relevant ones. In order to do
this, we recall the definition of the covariance between two stochastic
variables $X,Y$ as given by
$\displaystyle{\rm Cov}(X,Y)$ $\displaystyle=\big{\langle}(X-\langle
X\rangle)(Y-\langle Y\rangle)\big{\rangle}$ $\displaystyle=\langle
XY\rangle-\langle X\rangle\langle Y\rangle\,,\,$ (17)
where the brackets denote expectation values. The covariance between two
variables provides a measure of the degree to which their fluctuations are
correlated. A smaller (larger) covariance implies lower (higher) correlation.
In particular, the covariance of a variable with itself is its variance (i.e.,
the standard deviation squared) and measures deviations from the mean value.
When there are multiple stochastic variables $X_{i}$ ($i=1,\ldots,n$) one can
construct their associated covariance matrix $\bf{C}$ with components
$C_{ij}={\rm Cov}(X_{i},X_{j})$. This matrix is symmetric, non-negative
definite and can therefore be diagonalized with an orthogonal transformation.
Consider the $i^{\rm th}$ normalized eigenvector $\hat{{\bf V}}_{i}$. If we
set ${\vec{\bf X}}:=(X_{1},\ldots,X_{n})$ then the principal components (PCs)
are the associated eigenmodes,
${\cal E}_{i}={\vec{\bf X}}\cdot\hat{\bf V}_{i}\,.$ (18)
The PCs are new variables representing directions in which the $X_{i}$ have
different levels of variance. They are uncorrelated, a consequence of the
orthogonality of the eigenvectors, and their associated eigenvalues $\lambda$
are their variances,
${\rm Cov}({\cal E}_{i},{\cal E}_{j})=\lambda_{i}\delta_{ij}\,.$ (19)
The fact that, by construction, principal components are uncorrelated with
each other is important since they provide independent pieces of statistical
information.
The smaller an eigenvalue $\lambda_{i}$, the more likely that the
corresponding linear combination ${\cal E}_{i}$ will not deviate from its
average value for a randomly set of variables. Therefore, if there exist small
eigenvalues then the associated PCs are largely conserved in a statistical
sense. Conversely, the larger an eigenvalue is then the more relevant the
associated PC is in describing the dynamics and variations in the problem.
There are two related but different senses in which for a parametrized time
series a principal component with small variance can be semi-conserved. The
first is being constant as a function of time for an arbitrary but fixed set
of parameters. The second one is in the statistical sense that deviations of a
principal component from the mean value are small for arbitrary but fixed
initial and final times over a set of runs with the initial configurations. A
small variance automatically implies approximate conservation in the second
sense but not necessarily in the first one. The interest is not only in those
PCs which have the smallest variances (and thus identify semi-conserved
quantities in the second sense) but also in those with the largest variances,
which encode the most information about the system dynamics. This should be
clarified through the following example.
Example 3. PCA for spin dynamics.
One may wonder if given a uniform initial spin orientation distribution, after
a while those orientations turn into some preferred orientation, such as being
aligned into some preferred direction. This was studied in [57] through
massive numerical simulations. We only briefly review some of the results of
that reference. The interest at the time was motivated by the unexpectedly
large kicks found in numerical relativity simulations and whether they were of
a generic nature in a statistical sense; for a review on recent comprehensive
studies on this topic see Section 12.1. In order to perform a large enough
analysis the Post-Newtonian equations were used, up to 3.5PN order in the
angular frequency and 2PN with the covariant spin supplementary condition.
Next, in any PCA study one has to define which variables to analyze. In order
for the quantities to be invariant under rotations of the system of reference,
at least in a Newtonian sense, the first quantities chosen were scalar
products between the normalized spins of each binary component and the orbital
angular momentum of the system. Moreover, their differences between initial
and final values:
$\displaystyle\Delta({\bf\hat{S}}_{1}\cdot{\bf\hat{L}})$ $\displaystyle=$
$\displaystyle{\bf\hat{S}}_{1}\cdot{\bf\hat{L}}|_{f}-{\bf\hat{S}}_{1}\cdot{\bf\hat{L}}|_{i}=:{{\Delta}_{1\text{L}}}$
(20) $\displaystyle\Delta({\bf\hat{S}}_{2}\cdot{\bf\hat{L}})$ $\displaystyle=$
$\displaystyle{\bf\hat{S}}_{2}\cdot{\bf\hat{L}}|_{f}-{\bf\hat{S}}_{2}\cdot{\bf\hat{L}}|_{i}=:{{\Delta}_{2\text{L}}}$
(21) $\displaystyle\Delta({\bf\hat{S}}_{1}\cdot{\bf\hat{S}}_{2})$
$\displaystyle=$
$\displaystyle{\bf\hat{S}}_{1}\cdot{\bf\hat{S}}_{2}|_{f}-{\bf\hat{S}}_{1}\cdot{\bf\hat{S}}_{2}|_{i}=:{{\Delta}_{12}}\,,$
(22)
where hats stand for unitary vectors. The choice of initial and final values
is something to define and is discussed in [57], the summary is that the main
results do not depend on these choices, which gives insight into the fact that
the binary problem in GR is highly redundant.
The orbital angular momentum and spin orientations naturally become correlated
due to spin-orbit and spin-spin interactions as each of these binary black
hole configurations evolve in time. However, at least within the PN
approximation here considered, the orbital angular momentum and spin vectors
remain perfectly uniformly distributed [68]. For example, a Kolmogorov-Smirnov
test for a representative configuration returns a p-value of $\sim 10^{-5}$
when testing for lack of uniformness [68]. Higher PN expansions might
introduce small biases [89] but if so they appear to be at a level in which
approximating the mean of the above scalar products at any instant of time by
zero is a very good approximation.
Ref. [57] starts with a simple case and makes contact with previous
conservation results. The authors start building towards the more general case
by first doing a PCA using only the two spin-orbit (SO) variables in (20) and
(21),
${{\Delta}_{1\text{L}}}=\Delta({\bf\hat{S}}_{1}\cdot{\bf\hat{L}})\,,\quad{{\Delta}_{2\text{L}}}\mathrel{\mathop{:}}=\Delta({\bf\hat{S}}_{2}\cdot{\bf\hat{L}})\,.$
(23)
However, spin-spin interactions in both the numerical simulations and in the
analytical calculations are included in the PN equations of motion used when
solving for the evolution of each configuration.
For mass and spin magnitudes ($m_{j},\chi_{j}$) of each black hole the
covariance matrix for the variables (23) is
${\bf C}=\left(\begin{array}[]{cc}{\rm
Cov}({{\Delta}_{1\text{L}}},{{\Delta}_{1\text{L}}})&{\rm
Cov}({{\Delta}_{1\text{L}}},{{\Delta}_{2\text{L}}})\\\ {\rm
Cov}({{\Delta}_{2\text{L}}},{{\Delta}_{1\text{L}}})&{\rm
Cov}({{\Delta}_{2\text{L}}},{{\Delta}_{2\text{L}}})\end{array}\right)\,,$ (24)
where the entries can come either from numerical simulations or from what the
authors call the instantaneous approximation. The matrix $\bf C$ is then
diagonalized to find the principal components.
From numerical simulations the authors find that, sampling across many random
initial spin orientations, each of the principal components has zero mean over
time (to numerical accuracy), $\langle\Delta{\cal
E}_{j}^{\text{SO}}\rangle=0$, a consequence of the spin orientation
distributions remaining highly uniform during the inspiral.
Furthermore, they find $\lambda_{2}$ to be in the range $\sim 10^{-9}-10^{-4}$
for the parameters sampled and that it grows with both spin magnitudes, which
is expected from physical intuition, but also that it increases as the equal
mass case is approached, which was unexpected.
As an example with $(m_{1},m_{2},\chi_{1},\chi_{2})=(0.4,0.6,1.0,1.0)$, Figure
1 shows a graphical representation of the principal components overlaid on a
scatter plot of the ${{\Delta}_{1\text{L}}}$ and ${{\Delta}_{2\text{L}}}$ data
from $1,\\!000$ out of $100,\\!000$ numerical simulations using random initial
spin orientations. Notice that the first PC, which points along the direction
of the eigenvector ${\bf\hat{V}}_{1}$ with the largest eigenvalue
$\lambda_{1}$, captures the largest variation in the data while the second PC,
pointing along ${\bf\hat{V}}_{2}$, indicates that there is very little spread
in the data in that direction, which is also implied by the smallness of
$\lambda_{2}$ relative to $\lambda_{1}$. Therefore, for the time interval
considered, the second PC is largely irrelevant. This figure is almost an
ideal example of dimensionality reduction through PCA. For a detailed analysis
see [57].
Figure 1: A graphical representation of the principal components for the spin-
orbit variables and the numerical data of ${{\Delta}_{1\text{L}}}$ and
${{\Delta}_{2\text{L}}}$ for a binary black hole system with, as illustration,
$m_{1}=0.4$ and maximal spin magnitudes. One can see that PC2 is largely
irrelevant. For details see [57].
#### Further Reading
Principal Component Analysis is a very well known tool to look for
redundancies or categorize the relevance of different variables. One of its
weaknesses is that it depends on which variables to look for. The case study
that we chose is due to the fact that if looking in more detail at the
original reference reviewed, there are strong indications that there are at
least three (out of eight) redundant quantities in the problem, and therefore
the problem is amenable to an unsupervised dimensional reduction approach, as
presented throughout this review. Reference [76] is a very pedagogical book
solely on PCA, with many case studies and modern developments.
### 5 Proper Orthogonal Decomposition
Given the snapshots $H_{1},\ldots,H_{N}\in\mathbb{C}^{L}$ (the training
samples), the snapshot matrix is
${\bf H}=[H_{1},\ldots H_{N}]\in\mathbb{C}^{L\times N}$
with, in general, $L\geq N$. The goal here is to find $n\leq N$ orthonormal
vectors $\\{\phi_{i}\\}_{i=1}^{n}$ in $\mathbb{C}^{L}$ minimizing the average
approximation error (15)
$\frac{1}{N}\sum_{i=1}^{N}\|H_{i}-{\cal P}_{n}H_{i}\|^{2}\,,$ (25)
where ${\cal P}_{n}$ is the orthogonal projector onto
$W_{n}:=\text{span}\\{\phi_{i}\\}_{i=1}^{n}$ and $\|\cdot\|$ is the Euclidean
norm. The minimizer subspace of (25) is called a Proper Orthogonal
Decomposition (POD) of rank $n$, and its generating basis the POD one.
Notice that in general the POD basis elements $\phi_{i}$ are not members of
$\\{H_{i}\\}_{i=1}^{N}$. In the gravitational wave case, this means that the
reduced basis is not a subset of waveforms, neither does the method provide a
set of the “most representative” points in the space of waveforms for building
a posteriori bases through numerical relativity simulations.
Another disadvantage is that the problem, involving a potentially very large
linear algebra system, is not easily parallelizable in distributed memory
architectures. Still, the method is straightforward to implement and can, for
example, quickly provide insight into whether a problem is amenable to
dimensionality reduction: after performing a POD decomposition of the snapshot
matrix, one can look at the decay rate of singular values as a function of $n$
in order to know if the problem can be codified in a space of lower
dimensionality than the original one. This can be turned into an actual
strategy: a POD for a small subset of the problem can provide a quick insight,
and if the singular values decay fast enough with $n$, the larger problem of
interest can be tackled with a greedy approach, as discussed in Section 8.3.
The optimal solution to the minimization problem defined by Eq. (25) can be
accomplished by means of a Singular Value Decomposition (SVD) procedure. In
this framework the POD basis is given by the first $n$ left singular vectors
of the snapshot matrix $\bf{H}$ (this can be proved by considering the first-
order optimality conditions for the minimization problem).
Consider then the following singular value decomposition of H:
$\bf{H}=\bf{U\Sigma V}^{\dagger}\,,$ (26)
where ${\bf U}=[u_{1},\ldots,u_{L}]\in\mathbb{C}^{L\times L}$, ${\bf
V}=[v_{1},\ldots,v_{N}]\in\mathbb{C}^{N\times N}$ are orthogonal matrices and
${\bf\Sigma}=\begin{bmatrix}{\bf D}\\\ {\bf
0}\end{bmatrix}\in\mathbb{C}^{L\times N},{\bf
D}=diag(\sigma_{1},\ldots,\sigma_{N})\in\mathbb{C}^{N\times
N},\sigma_{1}\geq\ldots\geq\sigma_{N}\geq 0.$ (27)
Then, the rank-$n$ POD basis is given by $\phi_{i}=u_{i},\,i=1,\ldots,n$. This
basis provides a low-rank approximation to H and represents the minimizer
basis for the minimization problem stated in (25). To see this, consider from
(26) the relation
$u_{i}^{\dagger}{\bf H}=\sigma_{i}v_{i}^{\dagger}\,,$
valid for all column vectors of ${\bf U}$. Next, multiply by $u_{i}$ on the
left and sum
$\sum_{i=1}^{n}u_{i}u_{i}^{\dagger}{\bf
H}=\sum_{i=1}^{n}\sigma_{i}u_{i}v_{i}^{\dagger}\,.$ (28)
The l.h.s. of (28) is exactly the orthogonal projector associated to the basis
$\\{u_{i}\\}_{i=1}^{n}$ acting on $\bf H$ and the r.h.s. represents its
rank-$n$ approximation ${\bf H}_{n}$. Therefore we can rewrite (28) as
${\cal P}_{n}{\bf H}={\bf H}_{n}\,.$ (29)
Remarks on the SVD decomposition
* •
The numbers $\sigma_{i}$ are known as the singular values of the matrix H and
correspond to the positive square roots of the eigenvalues of the associated
matrix $\bf{K}=\bf{H}^{\dagger}\bf{H}$. They are usually chosen in descending
order in practice to facilitate the recognition of the most relevant principal
components.
* •
For the matrices U and V, only the first $\mathrm{rank}(\mathbf{H})$ columns
are unique, whereas the remaining ones are arbitrarily extended such that
orthogonality is maintained. Since $\sigma_{i}=0$ whenever
$i>\mathrm{rank}(\mathbf{H})$ the factorization (26) does not depend on the
choice of this extension, since such extension is annihilated by null entries
of $\bf\Sigma$.
* •
The first $N$ columns of U and V are known as the left- and right-singular
vectors of H respectively. The right-singular vectors, $v_{i}$, are the
normalized eigenvectors of ${\bf H}^{\dagger}{\bf H}$ and the left-singular
vectors, $u_{i}$, are the normalized eigenvectors of ${\bf H}{\bf
H}^{\dagger}$.
* •
The number $K$ of non-zero elements in $\mathbf{\Sigma}$ is exactly
$\text{rank}(\mathbf{H})$. Consequently, Eq. ((26)) is sometimes called a
rank-revealing factorization.
Error of a POD approximation
The accuracy of the low rank-$n$ approximation ${\bf H}_{n}={\cal P}_{n}{\bf
H}=[{\cal P}_{n}H_{1},\ldots,{\cal P}_{n}H_{N}]$ is given by the following
lemma.
Lemma 1. Define the approximation error by
$\epsilon=\frac{1}{N}\sum_{i=1}^{N}\|H_{i}-{\cal P}_{n}H_{i}\|^{2}.$ (30)
It can be shown that $\epsilon$ satisfies
$\epsilon=\frac{1}{N}\sum_{i=n+1}^{N}\sigma_{i}^{2}.$ (31)
In terms of the Frobenius norm,
$\epsilon=\frac{1}{N}\|{\bf H}-{\bf H}_{n}\|^{2}_{F}.$ (32)
Intuitively, the square of the Frobenius norm of the difference between a
matrix and its low-rank approximation represents the total squared difference
of the rows of ${\bf H}$ due to omitting the last $(N-n)$ singular values when
forming ${\bf H}$. As discussed above, the projection ${\bf H}_{n}$ represents
the best rank-$n$ approximation of H in the Frobenius norm.
In summary, going back to section 3, one can see that the minimization problem
(1) related to the average error of a set of functions ${\cal F}$ is optimally
solved by the POD/SVD decomposition of the matrix associated to those
functions.
In the context of a gravitational-wave template bank the rows of the snapshot
matrix are the waveforms evaluated at different time/frequency values, and the
different rows correspond to different intrinsic parameters. The inner product
of rows with themselves, in turn, correspond to the total power in the bank.
Thus, the squared Frobenius norm of a template bank is the total power
contained within all the templates in it. For normalized templates, this is
simply the size of the bank. It can be seen how Eq. (32) directly corresponds
to the total power in a template bank lost from a low-rank approximation to
it. More precisely, the error measure is directly related to the average
fractional SNR loss which, up to a constant, is the squared Frobenius norm of
the difference between the full template bank and its low-rank approximation.
Fixing the average fractional SNR loss thus determines the total number of
non-zero singular values which must be retained to guarantee that the rank-
reduced bank remains effective. This is discussed in practical terms through
the following case study.
Example 4. SVD for gravitational waves.
In Ref. [35] the authors studied the application of SVD to gravitational wave
templates to reduce the redundancies in the bank and build an orthogonal basis
to represent the whole set. As a proof of concept, the authors applied a SVD
approach to a set of CBC waveforms corresponding to a sliver of the BNS
parameter space. Next, we summarize some of the results of this reference,
closely following its notation, which might be different from the rest of this
review.
In order to detect a GW signal, the common choice of the minimal match between
an arbitrary point in parameter space and its nearest point of the template
bank is $97\%$.
In order to compare and filter the data against the entire template bank, an
approximation to the matched filtering
$\rho_{\alpha}=\langle\bar{h}_{\alpha},s\rangle$ is sought for, where
$h_{\alpha}$ is a complex waveform vector and $s$ is the data vector (the
presumed signal). In this way the number of evaluation of inner products can
be reduced as well as its computational cost. Let’s define the $N\times L$,
where $N=2M$ and $M$ stands for the number of complex waveform (there is no
obvious reason why to perform an SVD on real and imaginary parts of a
waveform, given that the POD approach can handle complex snapshots) template
matrix
${\bf H}=\\{h_{1}^{R},h_{1}^{I},\ldots,h_{M}^{R},h_{M}^{I}\\}\,,$ (33)
where $h_{i}^{R,I}$ are the real and imaginary part of the $i$-template
waveform, each one corresponding to the rows of $\bf H$. Applying an SVD
decomposition to ${\bf H}$ and writing it in component form (here we follow
the notation of [35])
${\bf
H}_{\mu\nu}=\sum_{\kappa=1}^{N}v_{\mu\kappa}\sigma_{\kappa}u_{\kappa\nu}\,,$
(34)
one can define the truncated sum
${\bf
H}^{\prime}_{\mu\nu}=\sum_{\kappa=1}^{N^{\prime}}v_{\nu\kappa}\sigma_{\kappa}u_{\kappa\nu}\,,$
(35)
and approximate $\rho_{\alpha}$ by
$\rho^{\prime}_{\alpha}=\langle
H^{\prime}_{2\alpha-1}-iH^{\prime}_{2\alpha},s\rangle=\sum_{\nu=1}^{N^{\prime}}(v_{{2\alpha-1}\,\nu}\sigma_{\nu}-iv_{{2\alpha}\,\nu}\sigma_{\nu})\langle
u_{\nu},s\rangle\,,$ (36)
where the $H^{\prime}_{j}$ are the vector rows of ${\bf H}^{\prime}$.
Figure 2 shows a representation of the matrix of waveforms H and its
associated SVD-basis. One disadvantage of this kind of dimensional reduction
is that the remaining basis barely resembles the structure of the original
template. This can be fixed with the RB-greedy approach that is presented in
Section 8.
Figure 2: Left: Representation of template waveforms corresponding to $\bf H$
against time. Right: Representation of the first four SVD-basis. Notice that
since the SVD basis functions do not correspond to actual waveforms, they
display non-physical behavior which might be difficult to fit for when
building a surrogate model as discussed in Section 11. Figures taken from
[35].
Accuracy. An approximation to the data snapshot matrix is sought such that
$\|{\bf H}_{\mu}-{\bf H}^{\prime}_{\mu}\|\sim 1\%$. Following this
requirement, the fractional SNR
$\Big{\langle}\frac{\delta\rho}{\rho}\Big{\rangle}:=\frac{1}{M}\sum_{\alpha=1}^{M}\frac{\delta\rho_{\alpha}}{\rho_{\alpha}}$
(37)
can be approximated by:
$\Big{\langle}\frac{\delta\rho}{\rho}\Big{\rangle}=\frac{1}{2N}\sum_{\mu=N^{\prime}+1}^{N}\sigma_{\mu}^{2}\,,$
(38)
resulting from a Taylor expansion valid in the range
$\langle\delta\rho/\rho\rangle<10\%$. Note that this approximation is
proportional to the squared Frobenius norm of the truncation error of ${\bf
H}$,
$\Big{\langle}\frac{\delta\rho}{\rho}\Big{\rangle}=\frac{1}{2N}\|{\bf H}-{\bf
H}^{\prime}\|_{F}^{2}\,.$ (39)
As a case of application, [35] shows an SVD analysis to gravitational waves
emitted by a CBC-BNS, with chirp masses $1.125M_{\odot}\leq
M_{c}<1.240M_{\odot}$ (that is, a rather small sliver in parameter space) and
component masses $1M_{\odot}\leq m_{1},m_{2}<3M_{\odot}$. In order to satisfy
a minimal match of $96.8\%$, a number of templates $M=456$ ($N=912$) to cover
the parameter space was found. In Fig. 3 (plot of
$\langle\delta\rho/\rho\rangle$ vs. # of SVD-basis elements) it can be seen
that, to obtain $\langle\delta\rho/\rho\rangle=10^{-3}$, the number of basis
elements needed to reconstruct the whole template bank to that accuracy can be
reduced from $N=912$ to $N^{\prime}=118$. Though POD/SVD is a good starting
point for ROM, in following sections we will develop a more modern framework
for modeling reduction.
Figure 3: SNR (Eq. (37)) for a CBC-BNS gravitational waves as a function of
the number of SVD-basis elements – valid for $SNR<10\%$ –. Figure taken from
[35].
#### 5.1 PCA and POD
Principal Component Analysis and Proper Orthogonal Decomposition are closely
related from a mathematical point of view, though usually applied in different
contexts. PCA is in general used in statistical analysis. It represents a
solution to an optimization problem; namely, that one of finding the
uncorrelated directions of maximum statistical variability in a stochastic
data set. POD is usually applied in dimension reduction modeling. As in PCA,
POD also represents a solution to an optimization problem: finding the best
low-rank approximation to a data set matrix. This is achieved by minimizing
the Frobenius norm of the error matrix (Eq. (32)).
Suppose we have a data set matrix of size $L\times N$, $N$ being the number of
variables and $L$, the number of samples corresponding to each variable. In
PCA one is not interested in the “length” of each variable and their
individual components, but rather on their correlations. Put differently, if
there are $N$ stochastic variables, the covariance matrix, regardless of the
number of components in each stochastic variable, is in all cases of size
$N\times N$. Compression can be achieved by ignoring the directions in which
the stochastic data have little variance.
Mathematical relationship between PCA and POD
Although conceptually different, both methods are mathematically related. To
see this, we make a simple observation concerning the snapshot matrix and its
covariance matrix. Let ${\bf X}:=[X_{1},\ldots,X_{N}]\in\mathbb{C}^{L\times
N}$ be the snapshot matrix with each column $i$ storing $L$ independent
observations of the random variable $X_{i}$. We assume that each $X_{i}$ is
identically distributed. Subtracting to each column $X_{i}$ the corresponding
mean value $\langle X_{i}\rangle$, $X_{s,i}\rightarrow X_{s,i}-\langle
X_{i}\rangle\,,\,s=1,\ldots,L$, and replacing $\bf{X}$ by this “centered”
matrix version, the associated covariance matrix $\bf{C}$ can be written as
${\bf C}=\frac{1}{L-1}{\bf X}^{\dagger}{\bf X}\,.$ (40)
Without loss of generality, we will suppose $\bf X$ has full rank. Therefore,
being an hermitian, non-negative definite matrix, $\bf C$ can be diagonalized
as
${\bf C}={\bf V}{\bf\Lambda}{\bf V}^{\dagger}\,,$
where ${\bf\Lambda}=\text{diag}(\lambda_{1},...,\lambda_{N})$ is the matrix of
non-zero descending-ordered real eigenvalues and $\bf V$ is the unitary matrix
storing eigenvectors of $\bf C$. The columns of $\bf V$ are called principal
axes or principal directions in the PCA framework. PCs correspond to the
projection of the data onto these principal directions, the columns of the
matrix product ${\bf X}{\bf V}$.
Now, perform a POD/SVD decomposition of the snapshot matrix X
${\bf X}=\mathbf{U}\mathbf{\Sigma}\mathbf{\tilde{V}}^{\dagger}$
and rewrite Eq. (40) as
${\bf
C}=\frac{1}{L-1}\,\,(\mathbf{U}\mathbf{\Sigma}\mathbf{\tilde{V}}^{\dagger})^{\dagger}\mathbf{U}\mathbf{\Sigma}\mathbf{\tilde{V}}^{\dagger}=\mathbf{\tilde{V}}\,\frac{\mathbf{\Sigma}^{\top}\mathbf{\Sigma}}{L-1}\,\mathbf{\tilde{V}}^{\dagger}\,.$
(41)
The similarity relation between $\bf C$ and
$\frac{\mathbf{\Sigma}^{\top}\mathbf{\Sigma}}{L-1}$ implies that both have the
same eigenvalues:
$\lambda_{i}=\frac{\sigma^{2}_{i}}{L-1}\,,\quad i=1,\ldots,N\,.$
Another benefit of this mathematical correspondence is that PC’s are
straigthforward to compute as ${\bf X}{\bf V}={\bf U}\mathbf{\Sigma}$.
#### Further reading
The reader may consult [76, 140] for clear expositions about the classical
matrix factorization methods presented in this section. QR decompositions form
another class of matrix factorizations which can be used for low-rank
approximations. A theoretical framework for rank revealing QR decompositions
of a matrix was developed in the the late 1980’s and 1990’s [38, 73],
motivated partly by the SVD’s high computational cost. QR decompositions form
the basis for many modern fast algorithms [66, 64, 47, 42]. Actually,
motivated by GW research an important resuIt has been shown: the RB-greedy
algorithm is completely equivalent to a certain type of QR decomposition [16].
### 6 Spectral expansions
In this section we discuss fast converging classical linear approximations,
since they naturally lead to Reduced Basis when considering parametrized
problems.
Consider a set of complex functions,
${\cal F}:=\\{h:\mathbb{C}\rightarrow\mathbb{C}\\}\,.$
A standard linear approximation consists of the following: a basis with $n$
(non necessary orthogonal) functions is somehow chosen,
$\\{e_{i}\\}_{i=1}^{n}\,.$
Next, a function $h$ in ${\cal F}$ is approximated by a linear combination of
the chosen basis elements,
$h(z)\approx h^{(n)}(z):=\sum_{i=1}^{n}c_{i}e_{i}(z)\,,\quad
c_{i}\in\mathbb{C}.$ (42)
There are several criteria to choose the expansion coefficients $\\{c_{i}\\}$.
We recap least-squares (LS), in Section 10 we will discuss interpolation. In
the LS approach the coefficients $\\{c_{i}\\}$ are chosen such that the
representation error is minimized with respect to some chosen norm,
$\left\|h-\sum_{i=1}^{n}c_{i}e_{i}\right\|^{2}\,.$ (43)
As was stated in Section 3, the solution to the LS problem (Eq. (9), Section
3) is the orthogonal projection onto the span of the basis,
$h^{(n)}={\cal P}_{n}h\,.$ (44)
That is, the approximation (42) minimizes the error (43) when the coefficients
$\\{c_{i}\\}$ are chosen such that the residual
$\delta^{(n)}h=h-h^{(n)}$
satisfies $\langle\delta h,h^{(n)}\rangle=0$, which implies
$\left\langle
e_{j}(\cdot),h(\cdot)-\sum_{i=1}^{n}c_{i}e_{i}(\cdot)\right\rangle=0\quad\quad\text{
for }j=1,\ldots,n\,.$ (45)
The solution to (45) is
$c_{i}=\sum_{j=1}^{n}({\bf G}^{-1})_{ij}\langle e_{j},h\rangle\,,$ (46)
where ${\bf G}^{-1}$ is the inverse of the Gram matrix or Gramian ${\bf G}$,
with entries
${\bf G}_{ij}:=\langle e_{i},e_{j}\rangle\,.$
If the basis is orthonormal, this matrix is the identity and one recovers the
familiar expression
${\cal P}_{n}h=\sum_{i=1}^{n}\langle e_{i},h\rangle e_{i}\,.$ (47)
The Gramian matrix can be very ill-conditioned in general, meaning that the
calculation of its inverse, needed in (46), can have large numerical errors
(see, for example [136]). Therefore, from a numerical conditioning point of
view, it is convenient to work with an orthonormal (or orthogonal) basis since
the Gram matrix is the identity.
In iterative approximations one defines the convergence rate as the rate at
which the representation error
$\|\delta^{(n)}h\|=\left\|h-{\cal P}_{n}h\right\|$ (48)
decreases as $n$ increases for any given $h$. In the context of RB, when one
is dealing with parametrized systems, this error will depend also on the
parameters of the system.
#### 6.1 Spectral methods
In terms of accuracy and optimal convergence rates for approximation of a
space ${\cal F}$, we have not yet discussed two related aspects:
1. 1.
The choice of a “good” basis or, more precisely, the approximation space
$W_{n}$.
2. 2.
The optimal choice of how many basis elements to use. One would think that the
larger, the better. This is related to the regularity of the functions in
${\cal F}$. That is, how smooth or differentiable they are.
The first comment might be puzzling, after all we have emphasized that given
an approximation space $W_{n}$, the LS approximation is uniquely defined in a
geometric way. The question really is what the approximation space $W_{n}$
should be to minimize the representation error.
##### 6.1.1 Fourier Expansions
We start with the simplest and most familiar case: that one of periodic
functions in $[0,2\pi]$, unit weight, $\omega(x)=1$ 333The choice of norm is
crucial, see the next Section about Jacobi polynomials for more insights about
this., and Fourier modes as (orthonormal) basis,
$e_{j}(x)=\frac{1}{\sqrt{2\pi}}e^{ijx}\,,\quad j\in\mathbb{Z}$
Assuming for simplicity that $n$ is even, we have
${\cal P}_{n+1}h(x)=\sum_{-n/2}^{n/2}\hat{h}_{j}e_{j}(x)\,,$
where $\hat{h}_{j}$ are the Fourier coefficients of $h$.
One can show that if $h$ has $s$ derivatives, then there exists a constant
$C>0$ independent of $n$ such that
$\|h-{\cal P}_{n+1}h\|\leq C(n+1)^{-s}\left\|\frac{d^{s}h}{dx^{s}}\right\|$
(49)
for all $n\geq 0$. In particular, if $h$ is smooth ($h\in C^{\infty}$), then
the representation error decays faster than any power law with
$n\rightarrow\infty$, which is referred to as spectral convergence. Under
further conditions, in particular the case in which $h$ is analytic, the error
decay is actually exponential. The Fourier case gives a very intuitive way of
how this happens just by using integration by parts, for more details see
Chapter 9 of [120].
The summary here is that Fourier modes are not a good basis just because they
are periodic functions, but they provide fast convergence as a representation
space, as fast as the regularity of the function(s) being represented has.
This is the core idea of spectral methods and, for parametrized systems,
reduced basis. A brief summary of the non-periodic case follows.
##### 6.1.2 Jacobi Polynomials
In the case of non periodic functions a similar result to that one discussed
for periodic functions holds, whether the domain is bounded or infinite.
Again, the choice of weight is crucial. Any interval can be mapped into
$[-1,1]$ or $(-1,1)$, that is why we usually refer to those intervals. The
open interval case is because in some cases (such as Chebyshev) the weights
actually diverge at the $\pm 1$ end points – there is a reason for this but we
shall skip it.
Polynomials are a natural basis to use. Why? Just because after centuries we
understand them well; going beyond that is one way of looking at Reduced
Basis. Given a maximum degree, the span of polynomials is the same, so the
question is only what kind of weights guarantee fast convergence upon
regularity of the function to be represented. The following family of weights
is a sufficient class, thought not a necessary one,
$\omega(x)=(1-x)^{\alpha}(1+x)^{\beta}\,\quad x\in(-1,1)$ (50)
with $\alpha,\beta>-1$.
Under these assumptions, the following holds in a Jacobi Polynomial
approximation:
$\|h-{\cal P}_{n+1}h\|\leq
C(n+1)^{-s}\left\|(1-x^{2})^{s/2}\frac{d^{s}h}{dx^{s}}\right\|$ (51)
for all $n>(s-1)$ and $C$ independent of $n$. If $h$ is smooth there is
asymptotic spectral convergence. Under additional conditions on the smoothness
(or, but not necessarily, analyticity), the convergence is in fact
exponential. In this review we will loosely associate smoothness with
asymptotic exponential convergence of application-specific spectral
expansions, without discussing these additional assumptions.
Most of spectral methods literature is based on solving numerical problems,
prominently differential equations. In this review we shift this focus to that
one of a more fundamental representation problem, after which all the calculus
for approximating quadratures, taking derivatives, and solving differential
equations follows in a rather straightforward way. This is not just a matter
of taste, but will make the introduction of Reduced Basis and all its
associated calculus and applications very natural and almost trivial.
#### Further reading
The weights (50) lead to Jacobi polynomials, which are solutions to a singular
Sturm-Liouville problem, the properties of which guarantee the above mentioned
spectral convergence. Standard examples are Legendre and Chebyshev
polynomials. Their span is the same, it is with respect to which scalar
product they guarantee fast convergence and their discrete version and
relation to interpolation, discussed in Section 10. They are also
orthonormalized with respect to their own scalar products, which helps avoid
the typical conditioning problem of inverting the associated Gramian matrix.
For more details on and a quick glimpse at spectral methods, from a physicist
or practitioner perspective, see Chapter 9 of Ref. [120]. Reference [139] is a
surprisingly compact and efficient book to get started with spectral methods
(despite its title, MATLAB is not necessary at all to digest the content). A
long classic reference, especially for practitioners, is [29]. The book is
legally available for free from the author’s webpage, though with some typos.
A great book, in some sense targeted at ordinary differential equations is
[56]. For a modern presentation and the latest results, from theory to current
approaches to beat Gibb’s phenomena, see [70] and references therein.
### 7 Parametrized problems and optimal approximations
We are interested in approximating some abstract space of parametrized
functions ${\cal F}$, which in general is not linear (the sum of two waveforms
does not need to be a waveform) but we assume that it can be embedded in a
Hilbert one ${\cal H}$ (for example, the one of integrable functions in the
$L_{2}$ sense). We denote the underlying space of parameters as $\Phi$, which
we assume to be compact. For example, ${\cal F}$ can be the space of
$\lambda$-parametrized solutions $u_{\lambda}(x)$ of a partial differential
equation representing the dynamics of a physical system. In this review we
place emphasis in gravitational waves, parametrized for example by the mass
and spin of each black hole in a binary collision.
The question discussed next is how well one can theoretically approximate all
of ${\cal F}$ by a set of $n$ basis elements of ${\cal H}$ in a linear and
most compact way. This leads to the Kolmogorov n-width 444There are other
widths, see [107]. $d_{n}$ of ${\cal F}$ with respect to ${\cal H}$,
$\displaystyle d_{n}=d_{n}({\cal F},{\cal
H}):=\min_{\\{e_{i}\\}_{i=1}^{n}\in{\cal
H}}\max_{\lambda\in\Phi}\min_{c_{i}\in\mathbb{C}}\bigg{\|}h(\cdot;\lambda)-\sum_{i=1}^{n}c_{i}(\lambda)e_{i}(\cdot)\bigg{\|}^{2}\,.$
(52)
Comment 1. We explain the meaning of (52), from right to left, with respect
to the $\min,\max,\min$ properties.
* •
The first $\min$ implies that the optimal representation with respect to the
(so far arbitrary) basis $\\{e_{i}\\}_{i=1}^{n}$ is used. That is, given a
basis, the best representation (in the induced norm $\|.\|$) is considered. We
already discussed that this is the orthogonal projection ${\cal P}_{n}$ onto
the span of the basis $\\{e_{i}\\}_{i=1}^{n}$, so we can replace (52) by
$\displaystyle d_{n}:=\min_{\\{e_{i}\\}_{i=1}^{n}\in{\cal
H}}\max_{\lambda\in\Phi}\bigg{\|}h(\cdot;\lambda)-{\cal
P}_{n}h(\cdot;\lambda)\bigg{\|}^{2}\,.$ (53)
* •
Next, the largest error in parameter space for such a best approximation is
picked. That is the “worst best” approximation given a choice of basis.
* •
Finally, a choice of basis (more precisely, the subspace $W_{n}$) which
minimizes this worst best error is chosen and the associated error is by
definition the n-width. The compactness of $\Phi$ is important to guarantee
that these minimum and maximum exist in the search through the parameter
space.
In other words, the $n$-width is an upper bound of “what is the best that one
can do” if one could optimally choose an optimal basis.
This is mostly a theoretical problem, since solving for such a basis is
impractical (or, actually, in most cases, intractable) from a computational
point of view because it carries combinatorial complexity (all elements of the
basis have to be simultaneously chosen). It is more of a theoretical upper
bound against which to benchmark the quality of any computable approximation
which seeks for an approximate solution to the $n$-width problem.
Parameter regularity and fast convergence
Exact expressions for the $n$-width in general can only be achieved in a few
cases (see, for example, [91]). In general, the best one can do is to set up
upper bounds under specific assumptions and infer the dependence of the
$n$-width with respect to $n$. Indeed, Kolmogorov calculated this distance for
a special class of functions [91] with its first $(r-1)$ derivatives with
respect to parameter variation absolutely continuous and obtained a power law
dependency for the $n$-width,
$d_{n}\sim n^{-r}\,.$
If the set of functions considered were $C^{\infty}$ with respect to parameter
variation – the functions themselves can be discontinuous – the rate of
convergence of its optimal basis representation would be better than any power
law. This is called spectral convergence and will be relevant in next chapters
at the moment of quantifying the fast convergence of a Reduced Basis approach.
This is exactly the case in many scenarios of GWs, since they depend smoothly
on the parameters of their sources and one therefore expects an optimal
approach to have very fast (in fact, exponential) convergence, as opposed to
random sampling, for example, for which the convergence is sublinear. This is
the main reason underlying the extreme high accuracy of very compact surrogate
models based on reduced bases. That is, one expects in most GW scenarios
$d_{n}\sim e^{-an^{b}}\,,$
and an optimal basis should be extremely accurate and compact; in fact in some
cases super-exponential convergence ($b>1$) has been found in the GW context
[53]. The question then turns into how to find an approximate basis which is
not only computable but is also nearly optimal with respect to the n-width.
It is “common knowledge” that parametrized problems with regularity with
respect to parameter variation show spectral convergence of the n-width in
practice. One might ask how this observation can be possible at all when we
have mentioned that the n-width is not computable in practice: we will return
to this point when we discuss the greedy algorithm. But in fact, up to our
knowledge there is no rigorous proof of this expectation for general
parametrized systems. But there is a very compelling argument. Asymptotic
exponential convergence is also observed in practice in GWs and, being
physicists, that is good enough for us. The argument is as follow (courtesy of
Albert Cohen):
The convergence rate of the n-width depends on the smoothness/regularity of
the functions with respect to the parameters of the problem. We now argue that
if there is regularity with respect to parameter variation, as in many cases
of interest, then one can expect fast (in fact, up to exponential or even
super-exponential) decay of the approximation error. The argument is a
spectral standard one (spectral methods were discussed in the previous
section) applied to parameter variation and is as follows:
If the parametric map
$\lambda\rightarrow h_{\lambda}(\cdot)$
is smooth enough with respect to $\lambda$, then it can be very well
approximated in some appropriate basis in the $\lambda$ parameter variable,
for instance by Fourier or Jacobi polynomials. This means that there is an
expansion of the form
$c_{1}(\lambda)h_{1}(\cdot)+\ldots+c_{n}(\lambda)h_{n}(\cdot)+\ldots\,,$ (54)
with $h_{i}\in{\cal F}$, that converges fastly towards $h_{\lambda}(\cdot)$ in
${\cal F}$ uniformly in $\lambda$.
Now, a partial sum of the form (54) is a member of the span of
$\\{h_{1},\ldots,h_{n}\\}$, which means that this linear space approximates
well all functions in ${\cal F}$.
What “well” means depends on the approximation result used. If the number of
parameters is finite and the dependence is analytic, then exponential rates of
the form $exp(-n^{1/d})$ can easily be proved. If the dependence is only
$C^{s}$, then a rate $n^{-s/d}$ can also easily be proved. This argument
establishes, then, that there are bases in which expansions of the form (54)
converge very fast, and as a consequence the n-width can only decay faster.
To summarize, for problems with smooth parametric dependence, fast convergence
in terms of greedy reduced bases can be expected and it is not surprising that
such global methods outperform local ones.
We mentioned that the functions themselves can be discontinuous, which might
be confusing so it is worthwhile explaining it in more detail, even if
qualitatively. Consider a problem of fluid dynamics, where the solutions might
develop shocks at a finite time, and imagine that one is solving the partial
differential equations for a parametrized family of initial data. This is
completely fine in terms of the fast convergence of the n-width, since the
location, shape, etc., of the shock depend smoothly on the initial data; the
time series (in this case) itself does not need to have regularity with
respect to the physical variable (time in this example) but with respect to
parameter variation.
### 8 Reduced Basis
In this section we present the RB-greedy framework for parametrized systems in
order to address the resolution of the Kolmogorov problem in a quasi-optimal
way. First mathematical conventions related to the RB-scheme are presented,
followed by a discussion of its near-optimality with respect to different
$n$-width behaviors.
#### 8.1 Introduction
In previous sections we have discussed linear approximation of functions by
orthogonal projection onto the span of a basis. The latter was taken to be a
generic, problem-independent one, such as Fourier modes or polynomials. The
Reduced Basis (RB) approach is a framework for efficiently solving
parametrized problems, representing the solutions in a compact way, and
predicting new ones based on an offline-online decomposition.
In parametrized problems one is interested in functions of the form
$h_{\lambda}=h_{\lambda}(\cdot)=h(\cdot;\lambda)\,.$
where is (in general a multidimensional) parameter $\lambda$, the
discretization of which will define the training space.
In fields related to scientific computing and data science one is usually
interested in multiple evaluations and analyses of functions in real time. The
approach of RB is especially tailored to that one in which some numerical
problem has to be solved in order to obtain each function, and obtaining such
numerical solutions is very expensive. The approach is also very powerful when
large existing data sets are known and a sparse representation is needed
and/or multiple, fast operations on them.
Then a minimal, nearly optimal set of representative such solutions
(waveforms, for the purpose of this review) is sought for in order to
construct a reduced basis (rb) for the whole solution set. These rb solutions
constitute an application-specific basis. That is, a set
$\\{h_{\Lambda_{i}}\\}_{i=1}^{n}$ (55)
of functions in the space of interest for carefully chosen parameter values
$\Lambda_{i}$ is used as a basis itself, as opposed to generic basis such as
polynomials or Fourier modes.
The scalar product in this parametrized space is taken at fixed parameter
values; that is, with respect to the physical dimension(s),
$\langle h_{i},h_{j}\rangle=\langle
h_{\lambda_{i}},h_{\lambda_{j}}\rangle=\langle
h({\cdot;\lambda_{i}}),h({\cdot;\lambda_{j}})\rangle\;.$
The same results of Section 6 follow through for the parametrized and
application-specific case. Namely, the optimal solution to the least-squares
approximation is
$h(\cdot;\lambda)\approx{\cal
P}_{n}h(\cdot;\lambda)=\sum_{i=1}^{n}c_{i}(\lambda)e_{i}(\cdot)\,,$ (56)
with the coefficients $c_{i}$ given by the equivalent of Eq. (46), in the
physical dimension(s) and the reduced basis 55 relabeled as
$\\{e_{1},\ldots,e_{n}\\}$. Namely,
$c_{i}(\lambda)=\sum_{j=1}^{n}({\bf G}^{-1})_{ij}\langle
e_{j}(\cdot),h(\cdot;\lambda)\rangle\,,$ (57)
where the entries of the Gramian matrix are
${\bf G}_{ij}:=\langle e_{i},e_{j}\rangle\,.$
Clearly, the Gramian coefficients depend on the special parameters
$\Lambda_{i}$ associated to each basis element.
So far, the span of the reduced basis uniquely determines the rb
representation. This reduced space depends primarily on the choice of the
selected parameter points in order to define a starting basis. As of a
convenient basis itself, for the same reasons discussed in Section 6, from a
conditioning numerical perspective, it is convenient to work with an
orthonormal basis built out of the rb solutions through a simple
orthonormalization procedure. This does not change the span of the rb and is
numerically convenient.
#### 8.2 The Training set
A commonly used approach to construct a basis, described below in Section 8.3,
is through a greedy algorithm. In its simplest version, the algorithm
identifies a set of $n$ points in parameter space out of a representative
enough set of functions of interest that are actually known. We call this set
of functions the training set:
${{\cal K}}:=\\{h_{\lambda_{i}}\\}_{i=1}^{N}\,.$ (58)
The subset $\\{h_{\Lambda_{i}}\\}_{i=1}^{n}$ constitutes a nearly optimal
basis for application-specific spectral expansions of any function in ${\cal
K}$ in a precise sense discussed in Section 8.3.1. If there is partial
redundancy/similarity in the latter, then $n<N$ or even $n\ll N$. Note here
that we have used the symbol ${\cal K}$ to represent a discretization of the
space of interest ${\cal F}$ in order to perform actual computer calculations.
The training set can be constructed by any means, including simple random or
uniform sampling, more sophisticated stochastic methods [95], the metric
approach [102], or those of Ref. [92], for example. For large problems,
resampling it while constructing the basis can be critical; for an application
in the case of GWs see for example [28]. Regardless of the method used to
populate the training set, the RB-greedy formalism produces a compact and
highly accurate representation of the training space catalog. The reduced
basis is used to approximate other functions in the space of interest, whether
they were in the training set catalog or not, through linear combinations that
represent an orthogonal projection onto its span,
$h_{\lambda}={\cal P}_{n}h_{\lambda}+\delta h_{\lambda}\,,$ (59)
where the RB approximation is ${\cal P}_{n}h_{\lambda}$ and satisfies, by
construction, $\langle{\cal P}_{n}h_{\lambda},\delta h_{\lambda}\rangle=0$.
#### 8.3 Greedy algorithms
We mentioned that a commonly used way to generate a reduced basis is through a
greedy algorithm. In its simplest form, such as when the waveforms are
inexpensive to compute or the data is already somehow known, the greedy
algorithm, outlined in Algorithm 1, has as input a discretization of the
parameter and solution space,
${\cal T}:=\\{\lambda_{i}\\}_{i=1}^{N}\quad{{\cal
K}}=\\{h_{\lambda_{i}}\\}_{i=1}^{N}$ (60)
with the elements of ${\cal T}$ usually called training points. To put
emphasis on structure instead of form, all functions in the training set in
this review are normalized. How to recover the norm, particularly for
parameter estimation of any detected signal, is discussed in Section 14.1.
The scheme needs an arbitrary seed $\Lambda_{1}\in{\cal T}$ to initialize it,
and a threshold error $\epsilon$ for a target representation accuracy (or
greedy error). Part of the output of the algorithm is a sequential selection
of $n$ parameter greedy points
$\\{\Lambda_{1},\Lambda_{2},\ldots,\Lambda_{n}\\}\subset{\cal T}$
and their associated waveforms
$\\{h_{\Lambda_{1}},h_{\Lambda_{2}},\ldots,h_{\Lambda_{n}}\\}\subset\\{h_{\lambda_{i}}\\}_{i=1}^{N}\,.$
The set of waveforms $\\{h_{\Lambda_{i}}\\}_{i=1}^{n}$ constitutes the reduced
basis. As was already discussed, for numerical conditioning purpose it is
sometimes convenient (see Eq. (10)) to work with an orthonormalized set
$\\{e_{i}\\}_{i=1}^{n}$ instead of directly the
$\\{h_{\Lambda_{i}}\\}_{i=1}^{n}$.
Another output of the algorithm is precisely the set of projection
coefficients for functions in the training space catalog, whereas coefficients
for any other known function can be computed through projection onto the
basis. In Section 10 we discuss how to approximate these coefficients through
interpolation in the physical dimension (time in the case of gravitational
waveforms), and in Section 11 a predictive approach for accurate and fast
evaluation of new (unknown) functions in ${\cal F}$ through surrogate models.
These are predictive models: as opposed to a known function being projected
into a compact basis, they predict new solutions (waveforms).
Algorithm 1 Greedy algorithm for reduced basis
1:Input: $\\{\lambda_{i}\,,h(\cdot;\lambda_{i})\\}_{i=1}^{N}$, $\epsilon$
2:Initialize $i=0$ and define $\sigma_{0}=1$
3:Seed choice (arbitrary): $\Lambda_{1}\in{\cal T}$,
$e_{1}=h(\cdot;\Lambda_{1})$
4:rb = $\\{e_{1}\\}$
5:while $\sigma_{i}\geq\epsilon$ do
6: $i=i+1$
7: $\sigma_{i}=\max_{\lambda\in{\cal T}}\|h(\cdot;\lambda)-{\cal
P}_{i}h(\cdot;\lambda)\|^{2}$
8: $\Lambda_{i+1}=\text{argmax}_{\lambda\in{\cal T}}\|h(\cdot;\lambda)-{\cal
P}_{i}h(\cdot;\lambda)\|^{2}$
9: $e_{i}=h(\cdot;\Lambda_{i+1})-{\cal P}_{i}h(\cdot;\Lambda_{i+1})$ (Gram-
Schmidt)
10: $e_{i+1}=e_{i+1}/\|e_{i+1}\|$ (normalization)
11: rb = rb $\cup\,e_{i+1}$
12:end while
13:Output: rb $\\{e_{i}\\}_{i=1}^{n}$ and greedy points
$\\{\Lambda_{i}\\}_{i=1}^{n}$
Comment 2.
* •
Greedy-type algorithms are global optimization procedures used in contexts
outside reduced basis or even dimensional reduction.
* •
Being a global optimization algorithm, the choice of the seed $\Lambda_{1}$ is
largely irrelevant. This was explicitly discussed as a sidenote in Ref. [36],
see Fig. 4 below. We also refer to Section 12.6 for a discussion of Ref. [36]
in the context of RB for multi-mode black hole ringdown.
* •
What the greedy algorithm does in Step 8 is to select the waveform for which
its representation error onto the existing basis with $i-1$ elements is worst,
and in Step 11 adding it to the enrichment of the rb representation.
* •
In steps 9 and 10 the rb waveforms are orthonormalized to avoid ill-
conditioning of the computation of the projection (see the discussion in
Section 6, before 6.1).
* •
Given an arbitrary user-defined tolerance error $\epsilon$, the algorithm
stops when the approximation (56) meets the error tolerance – introduced as an
input in the greedy algorithm –,
$\|h_{\lambda}-{\cal
P}_{n}{h_{\lambda}}\|^{2}\leq\epsilon\,\,\,\,\forall\,\,\lambda\in{\cal T}.$
* •
The expected exponential convergence of the method for the problems of
interest implies that $\epsilon$ can be made arbitrarily small with a
relatively small number $n$ of basis element, with $n<N$ and, in many cases,
$n\ll N$.
Figure 4: Taken from [36]. This figure shows the representation error as a
function of the number of reduced basis waveforms for a single QNM catalog.
The authors iterate over all possible seed waveforms in the training set. The
dark line represents the average, and the shaded area the maximum dispersion
around it. This numerical experiment confirms that the seed choice becomes
nearly irrelevant due to the global nature of the the greedy algorithm at each
step, as intuitively expected being a global optimization.
.
In practice, one uses a better conditioned orthogonalization algorithm in step
$9$ than the standard (i.e., “classical”) Gram-Schmidt one. The naive
implementation of the classical Gram-Schmidt procedure is actually ill-
conditioned. This is related to the fact that the Gramian matrix, which would
have to be inverted, can become nearly singular [136]. To overcome this one
can use an iterated Gram-Schmidt algorithm or a QR decomposition. A popular
alternative to the classical Gram-Schmidt, largely used, and called the
modified Gram-Schmidt, is also ill-conditioned, contrary to common perception
since its ill conditioning only becomes apparent for large large data sets.
See [117, 63, 131] for discussions about the conditioning and numerical
stability of different orthonormalization procedures.
Error definitions. We define the greedy error of the rb
$\\{{e_{i}}\\}_{i=1}^{n}$ as
$\displaystyle\sigma_{n}:=\max_{\lambda}\bigg{\|}h_{\lambda}-{\cal
P}_{n}h_{\lambda}\bigg{\|}^{2}\leq\epsilon\,,$ (61)
where $\epsilon$ is the user-defined tolerance error, which depends on the
number of basis elements.
The quantity $\sigma_{n}$ represents the largest error in the parameter space
of the best approximation by the greedy-reduced basis.
As discussed in [53], in the limit of sufficiently dense training spaces the
greedy error is comparable to the minimal match (${\mathrm{MM}}$) through
$\sigma_{n}\sim 1-{\mathrm{MM}}{\rm~{}~{}as~{}~{}}N\rightarrow\infty.$ (62)
Since the RB framework allows us to compress the information presented in the
training set, it is useful to introduce the quantity
$C_{r}:=N/n\,,$ (63)
called the compression ratio [119], to measure it.
##### 8.3.1 Convergence rates and near-optimality
The greedy algorithm chooses, in a precise sense, a nearly-optimal basis for
the function spaces ${\cal K}$ or ${\cal F}$, depending on whether one is
discussing the discrete or continuum cases. In order to quantify this near-
optimality, recall the definition of the $n$-width from Section 7, Eq. (53),
but write it in a more geometric way:
$\displaystyle d_{n}=d_{n}({\cal F},{\cal H})=\min_{W_{n}\in{\cal
H}}\max_{\lambda}\bigg{\|}h_{\lambda}-{\cal P}_{n}h_{\lambda}\bigg{\|}^{2}\,.$
(64)
The search space here is the whole Hilbert space and one looks for an optimal
$n$-dimensional subspace $W_{n}$ for approximating the function manifold
${\cal F}$. As was stated in Section 7, $d_{n}$ is a theoretical upper bound
to any practical algorithm to perform a linear approximation. In practice,
solving such optimization problem becomes unfeasible due to its intrinsic
combinatorial complexity. The nested nature of the greedy algorithm becomes
crucial for reducing the complexity of the Kolmogorov problem. This means that
each $H_{m}={\tt Span}\\{e_{i}\\}_{i=1}^{m}$ satisfies $H_{1}\subset
H_{2}\subset\ldots\subset H_{n}$ and this feature dramatically reduces the
search space in (64). One expects that this reduction of the search space is
at the expense of losing completely – if not a modicum – the $n$-width
optimality but, as we discuss in the next paragraph, this expectation is the
opposite from being true.
In which sense does the reduced basis-greedy procedure degrade the optimality
in the Kolmogorov sense? To answer this, lets summarize two important results
in relation with the greedy algorithm: if $d_{n}$ decays exponentially with
$n$ then so does $\sigma_{n}$ [48],
$d_{n}\leq
D\mathrm{e}^{-an^{\alpha}}\implies\sigma_{n}\leq\sqrt{2D}\gamma^{-1}\mathrm{e}^{-a^{\prime}_{\alpha}n^{\alpha}}\,,$
(65)
where $D$, $a$, $\alpha$ are positive constants.
Similarly, if the n-width has polynomial decay then so does the greedy error,
$d_{n}\leq Dn^{-\alpha}\implies\sigma_{n}\leq
D^{\prime}_{\alpha}n^{-\alpha}\,,$ (66)
where $D,\alpha>0$. More generally, for any decay rate of the n-width,
$\sigma_{n}\leq 2\gamma^{-1}d_{n/2}\,.$ (67)
The factor $\gamma$ in Eqs. (65) and (67) is a constant in $(0,1]$ [48]. In
recent years there have been efforts to improve these bounds. For details see
[148, 100].
In light of these results, we see that the reduced basis-greedy procedure
inherites the optimality of the $n$-width. If the latter has exponential
convergence, so it does the greedy error. In this precise sense the reduced
basis-greedy approach is nearly-optimal: we cut down the complexity of the
Kolmogorov problem at the very low expense of losing quality which, in most
practical applications, becomes insignificant.
The convergence rate of the $n$-width (and, in consequence, of the greedy
error) depends on the parametric smoothness/regularity of the
functions/waveforms. Indeed, if the functions are analytic and can be extended
to a complex region, exponential decay for the greedy error can be proven (see
[111], section 5.5, and citations therein). Therefore, the rb expansion is
expected to have very fast convergence to the original waveform as the number
of basis elements is increased. This is similar in spirit to the standard fast
convergence of spectral methods, but here smoothness in the parameters of the
problem is exploited, and the basis are elements themselves of the space of
functions of interest. For this reason RB is sometimes referred to as an
application-specific spectral expansion.
To summarize, for problems with smooth parametric dependence, fast convergence
in terms of greedy bases can be expected and it is not surprising that such
global methods outperform local ones.
Comment 3. Not only the choice of basis is important, but a also global
expansion. To illustrate this consider again the example of the space of
smooth periodic functions from section 6.1.1. If one is interested in a family
of them, and is able to choose $n$ Fourier modes to represent them as best as
possible, one could:
* •
build a discretization of the space of periodic functions and choose $n$
Fourier modes which when compared against the whole training set gives the
best global approximation error; this approach is a global variation of the
well known best m-term approximation problem, in which one is interested in
approximating some function and wants the best set of $m$ elements chosen from
a dictionary of functions to do it;
* •
follow the standard approach, in which the functions of interest are projected
onto the span of the first $n$ Fourier modes. If the functions of interest are
smooth, this approach gives exponentially decaying representation errors with
the number of Fourier modes. This is, of course, not a predictive model, since
the function to be projected has to be known.
The first path would return a global and problem-dependent basis by
construction. The second one, on the contrary, will return a standarized basis
with no further feeling for the problem than the minimization of the maximum
error due to the projection.
##### 8.3.2 Complexity, scaling, computational aspects
Besides its near-optimality the greedy algorithm has several implementation
advantages:
1. 1.
It is simple to implement (see Algorithm 1).
2. 2.
It follows a stricter aproximation criteria (see section 3) than POD/SVD,
namely, the minimization of the worst error, as opposed to the average one, in
parameter space due to the projection onto the rb basis. It provides a
stricter error control.
3. 3.
It is embarrassingly parallel. Each greedy sweep, defined by Step 7 of Alg. 1,
can be carried out simultaneously and independently for different values of
$\lambda\in{\cal T}$.
4. 4.
As we remarked in Sec. 8.3.1, the greedy points and associated basis are
hierarchical (nested). This is particularly important if the algorithm is to
be used to guide which numerical relativity simulations should be carried out.
Since they are so expensive, it is desirable that if further accuracy is
needed, then it should be possible to achieve it by adding more points (as
opposed to starting the process from scratch).
5. 5.
Low complexity. That is, the cost of all steps 6-12 in Alg. 1 are independent
of $i$, leading to a total cost linear in $n$, the total number of basis
elements. This observation follows from the fact that
${\cal P}_{i+1}h_{\lambda}={\cal P}_{i}h_{\lambda}+\langle
e_{i+1},h_{\lambda}\rangle e_{i+1}\,.$ (68)
So, if the projection coefficients are stored while building the basis, at
each sweep only the dot product between the elements of the training set and
the last basis element has to be computed.
#### Further reading
A comprehensive review of reduced basis is [112]. For textbooks see [70] and
[111].
## Part II Predictive Models
### 9 Polynomial interpolation
There are several reasons why we find it useful to discuss classical
polynomial interpolation:
* •
It is used in ROM for gravitational wave physics and other fields; we discuss
some advantages and disadvantages.
* •
There is a close relationship between interpolation, quadrature rules,
spectral methods and Reduced Basis. It is easier to introduce the standard
theory first, followed by the application specific version at the heart of ROM
for parametrized systems.
The emphasis of this section is on approximation quality quantified through
convergence rates and error bounds. Except for some side comments pointing out
implementations known to be poorly conditioned, we shall not discuss numerical
algorithms or their implementation which can be found in standard references.
#### 9.1 Representation versus prediction
In standard approximation theory, discussed in Sections 5 and 6, one expands a
function in some basis. Assuming that the approximation quality is measured by
some weighted $L_{2}$ norm, the optimal approximation is the orthogonal
projection onto the representation space. This projection is independent of
the basis, in the sense that it only depends on its span. Assuming these are
orthonormal for simplicity and conditioning purposes, the orthogonal
projection is given by
$h(x)\approx{\cal P}_{n}h(x)=\sum_{i=1}^{n}\langle e_{i},h\rangle e_{i}(x)\,.$
(69)
Approximating $h$ by its projection as in (69), while perhaps useful by
itself, is not a predictive tool: we need to know the values of $h(x)$ for all
$x$ in order to calculate the projection coefficients in Eq. (69). Here, by
prediction it is meant a new evaluation within the range used to build the
approximating model. This means, for example, within the same parameter and
physical (time/frequency for example) ranges in the case of GWs. We are not
dealing with a much more difficult problem: namely, that one of prediction of
time series, for example, based on previous history. Then, to get a predictive
power we need to first turn to an interpolation scheme.
Building an interpolant for $h(x)$ requires only the knowledge of $n$ values
of $h(x)$ whereas the unknown values of the function are predicted from this
interpolant. As previously mentioned, the applications are many, including
numerical schemes to solve partial differential equations, the numerical
computation of integrals and building predictive surrogate models. As we will
see in Section 9.4 these apparently radically different approaches for
approximation, such as interpolation and projection, are closely related, and
in some cases both are equivalent within certain frameworks.
#### 9.2 Prediction through polynomial interpolation
Suppose we have a collection of $(n+1)$ different points
$\\{x_{i}\\}_{i=0}^{n}$555Note that we count from $i=0$, only in this Section,
in order to match our notation with the common definition of Lagrange
polynomials., and a set of associated function values
$\\{f_{i}:=f(x_{i})\\}_{i=0}^{n}$. We assume that $f$ is a real valued, one-
dimensional function. We will discuss the more challenging multivariate case
in Section 9.5.
The interpolation problem consists of using the partial function sampling
$\\{f_{i}\\}_{i=0}^{n}$ to approximate $f(x)$ by another function which agrees
with $f$ at the interpolating nodes $\\{x_{i}\\}_{i=0}^{n}$. To be concrete,
let’s focus on linear interpolation: we require the interpolant of $f$ to be
expressible as a linear combination of the $(n+1)$ independent basis functions
$\\{e_{i}(x)\\}_{i=0}^{n}$,
${\cal I}_{n}[f](x):=\sum_{i=0}^{n}C_{i}e_{i}(x)\,.$ (70)
We already discussed in section 6 that the optimal least-squares-sense
approximation of the above form is provided by orthogonal projection onto the
basis $\\{e_{i}(x)\\}_{i=0}^{n}$. In the interpolation scheme one trades the
least-squares criteria for one which requires less information about $f$.
Let’s define this criteria by making the interpolant being equal to the
function at the nodes,
${\cal I}_{n}[f](x_{j})=f(x_{j})\,,\quad j=0,\ldots,n\,.$ (71)
When the basis are linearly independent polynomials of degree $\leq n$ we call
this special case Polynomial interpolation. One can show by explicit
construction that there is a unique polynomial which satisfies the
interpolation problem (70,71). Before doing so we give an example.
Example 5.
Suppose we have two points, $\\{x_{0},x_{1}\\}$, sampling a function $f(x)$.
Let the sample values be $f_{0}=f(x_{0})$ and $f_{1}=f(x_{1})$. The functions
$\\{1,x\\}$ comprise a suitable basis for our $n=1$ polynomial interpolant.
Then the interpolating polynomial is of degree $\leq 1$, i.e. a straight line
joining the two points:
${\cal I}_{1}[f](x)=ax+b$
We explicitly solve for $(a,b)$ by writing the interpolation problem (71)
$\displaystyle{\cal I}_{1}[f](x_{0})$ $\displaystyle=$ $\displaystyle
ax_{0}+b=f_{0}$ (72) $\displaystyle{\cal I}_{1}[f](x_{1})$ $\displaystyle=$
$\displaystyle ax_{1}+b=f_{1}$ (73)
which has as solutions
$a=\frac{f_{0}-f_{1}}{x_{0}-x_{1}}\,\,,\,\,b=\frac{x_{0}f_{1}-x_{1}f_{0}}{x_{0}-x_{1}}\,.$
(74)
Thus,
${\cal
I}_{1}[f](x)=\left(\frac{f_{0}-f_{1}}{x_{0}-x_{1}}\right)x+\left(\frac{x_{0}f_{1}-x_{1}f_{0}}{x_{0}-x_{1}}\right)\,.$
(75)
General solution: Lagrange polynomials
Existence. The existence of a polynomial which solves the interpolation
problem (71) can be shown by explicit construction. Suppose we have access to
a basis $e_{i}(x)=\ell_{i}^{(n)}(x)$ which satisfies
${\cal I}_{n}[f](x)=\sum_{j=0}^{n}f_{j}l_{j}^{(n)}(x)$ (76)
where, for each $j=0\ldots n$, $l_{j}^{(n)}(x)$ is a polynomial of degree
$\leq n$ such that
$l_{j}^{(n)}(x_{i})=\delta_{ij}\,\,\,\mbox{for }i=0\ldots n$ (77)
and $\delta_{ij}$ is the Kronecker delta ($\delta_{ij}=1$ if $i=j$ and zero
otherwise). Then ${\cal I}_{n}[f](x)$ as given by Eq. (76) would be a solution
to the interpolation problem.
The Lagrange polynomials are precisely those with the property (77). They are
given by
$l_{j}^{(n)}(x)=\left(\prod_{k=0,k\neq
j}^{n}(x-x_{k})\right)/\left(\prod_{k=0,k\neq j}^{n}(x_{j}-x_{k})\right)\,.$
(78)
Example 6.
We look at the same case ($n=1$) of Example 9.2, but now we construct an
interpolant through Lagrange polynomials. By Eq. (78) these are
$\l_{0}^{(1)}(x)=\frac{x-x_{1}}{x_{0}-x_{1}}\;\;\;,\;\;\;\l_{1}^{(1)}(x)=\frac{x-x_{0}}{x_{1}-x_{0}}\,,$
(79)
and using (76) the Lagrange form of the interpolant becomes
$\displaystyle{\cal I}_{1}[f](x)$ $\displaystyle=$ $\displaystyle
f_{0}l_{0}^{(1)}(x)+f_{1}l_{1}^{(1)}(x)$ (80) $\displaystyle=$
$\displaystyle\left(\frac{f_{0}-f_{1}}{x_{0}-x_{1}}\right)x+\left(\frac{x_{0}f_{1}-x_{1}f_{0}}{x_{0}-x_{1}}\right)$
The result (80) is the same as (75) due to the uniqueness of the polynomial
interpolant, which we now prove.
Uniqueness. Suppose there are two polynomials $P_{n}$ and $R_{n}$ of degree
$\leq n$ satisfying the interpolation condition
$P_{n}(x_{j})=R_{n}(x_{j})=f_{j},\,j=0,\ldots,n$. The polynomial
$Q(x):=P_{n}(x)-R_{n}(x)$ then satisfies $Q(x_{j})=0,\,j=0,\ldots,n$. This
means that $Q(x)$ has $(n+1)$ zeros, in contradiction with the hypothesis that
it is at most of degree $n$. Ergo, $Q(x)=0$ and $P_{n}(x)=R_{n}(x)$.
The Vandermonde form
Yet another way of showing existence and uniqueness for the polynomial
interpolant in a constructive way is the following. We write down the
polynomial interpolant using monomial basis functions,
${\cal I}_{n}[f](x)=\sum_{j=0}^{n}a_{j}x^{j}\,.$
The interpolation conditions (71) become a set of linear equations for the
coefficients $a_{j}$
${\bf V}a=f\,,$ (81)
where ${\bf V}_{ij}=x_{i}^{j-1}$, $a=[a_{0},\ldots,a_{n}]^{\top}$ and
$f=[f(x_{0}),\ldots,f(x_{n})]^{\top}$. The matrix ${\bf V}$ is called the
Vandermonde matrix and its determinant, also called the Vandermonde
polynomial, takes the form
$\det({\bf V})=\prod_{0\leq i<j\leq n}(x_{j}-x_{i})\,.$
In order to have one and only one solution to Eq. (81) this determinant has to
be non-zero. In one spatial dimension this condition is automatically
satisfied whenever the interpolation points are unique. There are ways of
inverting ${\bf V}$ in order to solve Eq. (81) using Lagrange polynomials,
which we will not pursue here. The main reason for briefly mentioning the
Vandermonde matrix is because a generalization of it will play a prominent
role in the Empirical Interpolation Method, described in Section 10.
#### 9.3 Convergence rates, collocation points and Runge’s phenomenon
Here we discuss the behavior of the error of the polynomial interpolant ${\cal
I}_{n}[f](x)$ of some function $f(x)$ at $(n+1)$ nodes, to illustrate the
difficulties in obtaining an accurate interpolation scheme on a unstructured
set of nodes which, from the point of view of ROM, one wishes to be sparse.
Leaving aside both sparsity and unstructured meshes, one might imagine that
for “nice enough” functions (for example, infinite differentiable), this error
decreases as the number of nodes and, therefore, degree of the polynomial
basis $n$ increase. This is not the case, the classical counterexample is
Runge’s one, here described, and the overall problem is referred to as Runge’s
phenomenon.
The distinction is between local approximations of fixed degree and global
ones. Because these two concepts play a key role in reduced order modeling we
discuss them in this context with some detail. Suppose that one has
$\\{x_{i}\\}_{i=0}^{n}$ nodes at which to interpolate a function.
Local interpolation
We consider local interpolants of degree one because we have already
explicitly discussed them in Example 75. One would partition the interval of
interest, where on each subinterval
$I_{j}:=[x_{j},x_{j+1}]\,,\quad j=0,\ldots,n-1$ (82)
one builds a local interpolant of degree one using the end points as
interpolating nodes,
${\cal
I}[f](x)=\left(\frac{f_{j}-f_{j+1}}{x_{j}-x_{j+1}}\right)x+\left(\frac{x_{j}f_{j+1}-x_{j+1}f_{j}}{x_{j}-x_{j+1}}\right)\,,\quad\text{for
}x\in I_{j}\,,\quad j=0,\ldots,n-1\,.$ (83)
Such fixed order interpolants, where their degrees are independent of the
number of nodes $n$, is – as explained later below – guaranteed to converge to
the interpolated function as the number of nodes in a given fixed physical
interval increases (the gridspacing decreases). Furthermore, such convergence
is guaranteed regardless of the structure/location of the interpolating nodes,
they can even be randomly located. Even more, to some degree, convergence is
also guaranteed regardless of the smoothness of the data; this is why fixed
order methods, especially very low order ones, are in general used for noisy
data.
This class of methods is usually very robust, but at the expense of accuracy.
The latter is exacerbated in the absence of a dense amount of data. On the
contrary, in Reduced Order Modeling one seeks very sparse representations of
very high accuracy. This combination of, and even trade-off between,
robustness (convergence) and accuracy is a delicate and challenging issue in
ROM.
Global interpolation
Instead, one could use a global interpolant, where all $(n+1)$ nodes and
function values are used to build a single interpolant,
${\cal I}_{n}[f](x)=\sum_{i=0}^{n}\ell_{i}^{(n)}(x)f_{i}\,.\quad\forall
x\in[x_{0},x_{n}]\,.$
As the density of points and $n$ increases, one would keep using all of them,
thereby increasing the degree of the interpolant. This might seem desirable
and more accurate than a local approach of fixed degree, since more
information is used. The classical counterexample is due to Runge [118], and
it consists of interpolating the function
$f(x)=\frac{1}{1+(5x)^{2}}\;\,,\quad x\in[-1,1]\,.$ (84)
This function is $C^{\infty}$ (has infinite derivatives) yet the error of a
global polynomial interpolant at equally spaced points diverges near the
boundaries [52] at higher orders (see Fig. 5). For unstructured points, such
as those usually appearing in reduced models, the lack of convergence of
global approximations is in general worse.
Figure 5: Runge’s phenomenon in polynomial interpolation. In this example the
approximation uses 19 equispaced nodes over the interval $[-1,1]$ and the
interpolant polynomial is of degree 18. The interpolation tends to diverge
near the boundaries. This issue cannot be solved by increasing the resolution
of the sampling but the opposite: divergencies get worse for equispaced nodes.
As discussed below, this can be resolved and fast convergence achieved if the
interpolation nodes can be appropriately chosen. However, standard choices for
such special nodes are in general not suitable for reduced order modeling. In
particular, they are neither necessarily sparse nor adapted to the problem of
interest. They are generic and in general (and most important) not
hierarchical, requiring recomputing from scratch the sought approximation if
higher accuracy is required.
Error analysis and convergence I
In what follows we assume that $f$ is smooth enough as needed from the
context. Denoting the interpolation error by
$E_{n}(x):=\left|f(x)-{\cal I}_{n}[f](x)\right|$
the following result holds:
$E_{n}(x)=\left|\frac{1}{(n+1)!}f^{(n+1)}(\xi)\omega_{n+1}(x)\right|$ (85)
where
$\omega_{n+1}(x):=\prod_{i=0}^{n}(x-x_{i})$
is called the nodal polynomial of order $(n+1)$, and $\xi$ is in the smallest
interval $I_{x}$ containing $x_{0}\ldots x_{n}$ and $x$. Equation (85) can be
proven as follows. Let’s define the function
$F(t)=f(t)-{\cal I}_{n}[f](t)-\alpha(x)\omega_{n+1}(t)$
where
$\alpha(x)=\frac{f(x)-{\cal I}_{n}[f](x)}{\omega_{n+1}(x)}$
This function is well defined since $x\neq x_{i}$, and it is as smooth as
$f(t)$. Observe that $F(x)=0$ and, in particular, $F(x_{i})=0$ ($i=1,...,n$).
This means that $F(t)$ has $(n+2)$ distinct zeros in the interval in which
$f(t)$ is smooth. Now, the Mean Value Theorem tells us that, between any
couple of zeros, there is one zero for the derivative $F^{\prime}(t)$. Then
the function $F^{\prime}(t)$ has $(n+1)$ zeros. Repeating the previous
argument, we conclude that $F^{(n+1)}(t)$ has exactly one zero. Let’s call it
$\xi$. Thus $F^{(n+1)}(\xi)=f^{n+1}(\xi)-\alpha(\xi)(n+1)!=0$. The statement
in (85) follows directly from this relation.
Let’s make a comment here. The point $x$ at which the interpolant is evaluated
when computing the error (85) does not need to be in the smallest interval
containing $x_{0}\ldots x_{n}$. Assuming an ordering $x_{0}<\ldots<x_{n}$,
sometimes the process of approximating $f(x)$ by ${\cal I}_{n}[f](x)$ is
called interpolation only if $x\in[x_{0},x_{n}]$ and extrapolation otherwise.
We get back to the error formula (85) and analyze its behavior as $n$
increases. In particular, we ask ourselves under what conditions does
$E_{n}(x)\rightarrow 0\;\;\;\mbox{as}\;\;\;n\rightarrow\infty\,.$
The error (85) can be decomposed in two terms, one related to the behavior of
the derivatives of $f$,
$\left|\frac{f^{(n+1)}(\xi)}{(n+1)!}\right|$
and another one related to the distribution of nodes through
$\omega_{n+1}(x)$.
We first discuss why fixed order interpolants are guaranteed to converge,
regardless of the distribution of nodes (assuming the interpolated function is
smooth enough), since the argument is simple.
Discussion 1. Local interpolants
For definiteness consider the case of equally spaced points (the
generalization to an arbitrary set is straightforward),
$x_{j}=x_{0}+j\Delta x\,,\quad j=0,\ldots n\,.$
Then the error of the interpolant (83) on each subinterval (82) is bounded by
$E\leq\frac{1}{(m+1)!}\max_{x\in I_{j}}\left|f^{(m+1)}(x)\right|(\Delta
x)^{m}\,,$ (86)
where $m$ is the (fixed) number of interpolation nodes and degree of the
interpolants (clearly, $m\leq n$). Since $m$ is fixed, the convergence rate is
determined by $\Delta x\rightarrow 0$, and is of order $m$.
Discussion 2. Global interpolants
In this case $m=n$. One can see from the interpolation error (85) that
convergence is guaranteed if the derivatives of $f$ decay with $n$
sufficiently fast. In fact, the interpolant converges in the infinity norm if
all derivatives $f^{(s)}$ are bounded by the same constant in the interval of
interest. However, it is not obvious from (85) to find out in a
straightforward way the allowed growth rate of the derivatives of $f$ such
that convergence/divergence of the error takes place.
In fact, it is not easy to see how such error formula explains Runge’s
phenomenon without resorting to complex calculus [52]. However, we can ask
ourselves the next related questions:
(Q1) What can be said about the maximum of $|\omega_{n+1}(x)|$ in the interval
of interest?
(Q2) If the physical problem allows such freedom, how can one choose the nodes
so as to minimize such maximum?
For simplicity and without loss of generality we focus on the interval
$x\in[-1,1]$. We can answer these questions as:
Answer to (Q1): For all choices of nodes
$\max_{[-1,1]}|\omega_{n+1}(x)|\geq 2^{-n}$ (87)
Answer to (Q2): If
$\omega_{n+1}(x)=2^{-n}T_{n+1}(x)\,,$
where $T_{j}$ denotes the Chebyshev polynomial of degree $n$, then
$\max_{[-1,1]}|\omega_{n+1}(x)|=2^{-n}\,.$
In other words, using the roots of the Chebyshev polynomial as interpolating
nodes minimizes the error associated with the location of such nodes. It is in
this sense that Chebyshev nodes are many times referred to as optimal points
for interpolation. They have the disadvantage, though, that they are note
hierarchical/nested, or adapted to the problem of interest. The former failure
means, from a ROM perspective, that any model relying on them, including
perhaps expensive numerical simulations, have to be completely built from
scratch if higher accuracy is required. This is solved by a nearly-optimal set
of nested nodes for application-specific interpolants designed with ROM
problems in mind, discussed in Section 10.
Error analysis and convergence II
In principle, direct approximation through projection requires the knowledge
of the continuum waveform, while interpolation only needs information at a
(specified by the user) set of nodes. We next address what is the relationship
between these two strategies. In terms of representation accuracy, we focus on
the $L_{2}$ norm, since it is directly related to the overlap error commonly
used in data analysis. Since, as discussed, given a basis and linear
approximation, projection is optimal in the least squares sense, one knows
that
$\|h-{{\cal P}}_{n}h\|^{2}\leq\|h-{\cal I}_{n}[h]\|^{2}\,.$ (88)
The question is how “sub-optimal” in the above sense interpolation is. The
answer obviously depends on the basis and interpolation nodes. A general
framework to study this is through the definition of the Lebesgue constant.
For any arbitrary basis and interpolation scheme one can derive the more
precise bound related to (88):
$\|h-{\cal I}_{n}[h]\|^{2}\leq\Lambda_{n}\|h-{{\cal P}}_{n}h\|^{2}\,,$ (89)
where
$\Lambda_{n}:=\|{\cal I}_{n}\|^{2}\,.$ (90)
Comment 4.
1. 1.
The $\Lambda_{n}$’s are referred as the Lebesgue constants and depend, given a
vector norm, purely on the interpolation operator ${\cal I}_{n}$.
2. 2.
In terms of accuracy, the strategy is to build an interpolant (by choosing
both the basis and nodes) such that the Lebesgue constant grows as slow as
possible with $n$. It is a computable quantity, so one can actually quantify
how much is “lost” with respect to projection through Eq. (89).
3. 3.
Strictly speaking, the Lebesgue constant is usually referred to in the context
of the infinity norm,
$\|u\|_{\infty}:=\max_{i}|u_{i}|\,,$
but throughout this review, unless otherwise stated, we refer to it in the
$L_{2}$ norm.
As an example, and going back to Runge’s phenomenon at equally spaced versus
Chebyshev nodes for global polynomial interpolation, the following illustrates
the advantages of the latter and explains why it is such a popular choice –
when it is a feasible approach at all within the problem of interest, which is
not always the case in practice.
Example 7.
Here the $\|.\|_{\infty}$ norm is used. For large $n$ one has the following
behavior (see, for example, [71]).
1. 1.
For equally spaced nodes, the Lebesgue constant grows exponentially, as
$\Lambda_{n}\sim\frac{2^{n+1}}{n\log n}\,.$ (91)
2. 2.
For Chebyshev nodes, it only grows logarithmically,
$\Lambda_{n}\sim\log{(n)}\,.$ (92)
In the context of polynomial interpolation, it can be proved that the Lebesgue
constant grows at least logarithmically in the infinity norm for any selection
of nodes. In particular, Chebyshev nodes belong to a family that induce a
nearly-optimal behavior on the Lebesgue constant, since the growth (92) is
only logarithmic. For thorough discussions around this topic see [71].
#### 9.4 Discrete expansions and interpolation
In approximation through projection one in principle needs full knowledge of
the function $f$ to be approximated through the projection coefficients
$\langle e_{i},h\rangle$ in Eq. (47). In interpolation, instead, one only
needs knowledge of the function at the interpolating nodes.
In practice, the integrals to compute $\langle e_{i},h\rangle$ are replaced by
a quadrature rule, i.e. a numerical approximation (we discuss quadrature rules
and specifically those built using ROM in Section 13). These are called then
discrete projection approximations. Here it suffices to say that, if the
quadrature nodes can be chosen, one can find an optimal family of quadrature
rules that maximizes the degree for which the quadrature is exact for
polynomials. These are called Gaussian quadratures and the nodes Gaussian
nodes; examples of the latter are Legendre and Chebyshev nodes. The relevant
result here relating approximation by discrete projection and interpolation is
the following:
Result 1. Discrete projection and interpolation
If the projection coefficients are approximated by Gaussian quadratures and
are used Gaussian nodes for interpolation, then the resulting approximation
${\cal P}_{n}h$ exactly equals the polynomial interpolant ${\cal I}[h]$ [71].
Through this result one can guarantee fast convergence of interpolation using
Gauss nodes. The problem is that in many parametrized problems of interest,
one cannot choose Gaussian nodes in parameter space as representative
solutions. Either because data is not available (for example, if taken from
experiments), or because it is not efficient to do so from a modeling
perspective. In Section 10 we discuss a generalization for ROM and
parametrized systems that attempts to generalize Result 9.4.
#### 9.5 Multiple dimensions
Evaluation cost of multivariate polynomials
Even when counting with enough data for multi-dimensional fits, enough for
accurate local approximations which avoid Runge’s phenomenom, a perhaps not so
well known fact is that the evaluation cost of multivariate polynomials in
general grows exponentially with the dimensionality of the problem.
Naive evaluation of a $1$-dimensional polynomial of degree $n$, that is,
evaluating all the monomials in the standard format
$p_{n}(x)=a_{0}+a_{1}x+\ldots+a_{n}x^{n}$ (93)
carries an operation count of ${\cal O}(n^{2})$. Through a simple
factorization, Horner’s rule (see [132]), the operation count can be reduced
to $2n$. Suppose we want to evaluate $p_{n}(x)$. First note that $p_{n}(x)$
can be decomposed in a nested way
$p_{n}(x)=a_{0}+x(a_{1}+x(a_{2}+\cdots x(a_{n-1}+a_{n}x)\cdots)\,.$
Horner’s rule consists in computing iteratively the recursion
$\begin{split}b_{0}=&a_{n}\\\ b_{k}=&a_{n-k}+b_{k-1}x\quad
k=1,\cdots,n\end{split}$
reaching $p_{n}(x)=b_{n}$ in $2n$ operations (multiplication and addition, $n$
times). It can be shown that this is the minimum number of operations to
evaluate a one-dimensional problem unless some offline factorization is
carried out for multiple online evaluations, in which case the computational
cost can be slightly reduced.
Next, consider multivariate polynomial evaluation. Suppose one attempts to
evaluate the generalization of (93) to the $d$-dimensional case. For
definiteness, the form in the ${\tt dim}=2$ case would take the form
$p_{n}(x,y)=\sum_{i=0}^{n_{x}}\sum_{j=0}^{n_{y}}a_{ij}x^{i}y^{j}\,.$
The evaluation cost of such an approach in the $d$-dimensional case, assuming
for simplicity the same degree $n$ in each dimension, is of order ${\cal
O}(n^{2d})$. Ideally, one would wish a generalization of Horner’s rule with,
say, evaluation cost of order ${\cal O}(d\times n)$. Unfortunately not such
algorithm is known, neither a proof of which could be the optimal number of
operations needed to perform a multivariate polynomial evaluation. In recent
years there have been several proposals to face this problem, ranging from
rigorous mathematical attempts (see e.g. [106, 98, 141, 81, 18, 18, 77, 142])
to the proposal of heuristic greedy algorithms [37] aiming to reduce the
number of calculations.
Multivariate polynomial evaluations scale exponentially with the
dimensionality of the problem. This is a severe issue in most cases of
interest which cannot be underestimated. In the context of ROM, the
limitations of multivariate polynomial approximations appear in almost every
possible context. This point is discussed in more detail and emphasized in
Section 15.
#### Further reading
Reference [58] reviews multivariate polynomial interpolation, and [99]
presents a proposal for polynomial interpolation on unstructured grids in
multiple dimensions.
There are schemes for fast multiple evaluations of multi-variate polynomials
though. They are based on an offline factorization of the given polynomial
(which can be expensive, but done only once), with a total cost of ${{\cal
O}}(n^{{\tt dim}+1})$ for ${\cal O}(n^{\tt dim})$ evaluations, leading to an
average of ${\cal O}(n)$ for each evaluation [84]. To our knowledge this
problem of multi-dimensionality has not been yet tackled in ROM for
gravitational waves and is still largely an open problem in approximation
theory. Instead, in GW science local interpolants in the form of splines are
being used, which results in degradation of the accuracy of the surrogate
models being built. In terms of higher (than one) dimensions, sparse grids are
being used. One of their problems is that they are not necessarily
hierarchical, leading to a problem when building training sets from numerical
relativity to later apply the RB framework.
### 10 The Empirical Interpolation Method
The Empirical Interpolation Method (EIM) was proposed in 2004 [20] as a way of
identifying a good set of interpolation nodes on multi-dimensional
unstructured meshes and has since found numerous applications (for a very
short list, see [1, 39, 40, 51, 90]). When the basis is application-specific,
so is the EIM.
Here, perhaps contrary to what might be suspected, interpolation is in the
physical dimension(s), not on the parametric ones. For definiteness, in the
case of gravitational waves this is either the time or frequency domain.
Before getting into technical details, we highlight some of the properties of
the EIM:
* •
It is hierarchical (nested). The “most relevant” points in the physical
dimension(s) and associated interpolants are selected and built, respectively.
This follows a process that is dual, in a precise sense, to the construction
of a reduced basis using a greedy algorithm.
* •
It is designed to overcome the difficulties of stability, accuracy and
evaluation cost on sparse and scattered grids, especially in multiple
dimensions.
* •
It is highly accurate. By design the algorithm attempts to control the
behavior of the Lebesgue constant (see Section 10.3) at each greedy EIM sweep.
* •
It provides an affine parametrization which has multiple consequences. In
particular, it allows for the design of Reduced Order Quadratures, described
in Section 13, which lead in particular to fast likelihood evaluations, as
discussed in Section 14.
The material below relies on standard interpolation ideas which we have
summarized in Section 9. We will refer to the physical variable(s) as $x$, the
nodes selected by the EIM as $X_{i}\,(i=1\ldots n)$ 666Except for the GW case,
in which we will explicitly use $t$ and $T$ – or $f$ and $F$, in the case of
frequency., the (in general multidimensional) parameter as $\lambda$ and its
corresponding greedy points as $\Lambda_{i}\,(i=1\ldots n)$. It is not a
coincidence that there is the same number of EIM nodes and greedy points, in
fact they go hand by hand.
#### EIM: algorithm and properties
Within the EIM method one seeks to find an empirical (that is, problem-
dependent) global interpolant, which also provides an affine parametrization
of functions. That is, the EIM approximation of a function $h(x;\lambda)$ is
of the form
$h(x;\lambda)\approx{\cal
I}_{n}[h](x;\lambda)=\sum_{i=1}^{n}B_{i}(x)h(X_{i};\lambda)\,.$ (94)
The affine parametrization of ${\cal I}_{n}[h](x;\lambda)$ refers to the fact
that the r.h.s. of (94) is a sum of terms which depend only on $x$ (the
$B_{i}$ coefficients) multiplied by other ones which depend only on $\lambda$
(the $h(X_{i};\lambda)$); one might also want to refer to this as
“separability”.
Below in Section 10.1 we discuss how one arrives at (94), how to compute in
the offline stage the $B_{i}$ coefficients, and the quality of this
approximation.
In addition, even though it is perhaps not usually viewed this way, or not
emphasized enough compared to online evaluations to solutions of parametrized
PDEs, the EIM also provides an application-specific down- and upsampling that,
for practical purposes, beats Nyquist downsampling, with implications for
signal or more generally data processing. See for example Section 11.2 and
discussion around Figures 12 and 13.
#### 10.1 From projection to interpolation
In Section 3 we discussed that the optimal linear representation in any
weighted $L_{2}$ norm, given a basis (assuming for simplicity and conditioning
that it is orthonormal) of cardinality $n$ is through an expression of the
form
$h(x;\lambda)\approx{\cal
P}_{n}h(x;\lambda):=\sum_{i=1}^{n}c_{i}(\lambda)e_{i}(x)\,,$ (95)
where
$c_{i}(\lambda)=\langle e_{i}(\cdot),h(\cdot;\lambda)\rangle\,.$ (96)
This is the standard approximation by projection, the most common one being
using polynomials or Fourier modes as basis. Now, approximation through
projection requires knowledge of the function to be represented at sufficient
values of $x$ so as to accurately compute the coefficients (96). In Section 9
we discussed how this is usually replaced by interpolation, and the strong
relationship between projection and interpolation in standard spectral theory.
Next we describe the EIM strategy, which mimics the spectral approach.
The approximation through projection onto a reduced basis, Eq. (95), is
replaced by interpolation (in the physical dimension(s), $x$) as follows.
First a rb $\\{e_{i}(x)\\}$ is chosen, for example through a POD or greedy
approach. The interpolant then is sought to have the form
${\cal I}_{n}[h](x;\lambda):=\sum_{i=1}^{n}C_{i}(\lambda)e_{i}(x)\,,$ (97)
where on purpose the expansion, which are not projection, coefficients $C_{i}$
are denoted by capital letters to distinguish them from the projection ones
(96). Instead, they are defined to be solutions of the interpolation problem;
namely that the interpolant exactly agrees with the function at the EIM nodes
(the construction of which we discuss below)
$\displaystyle{\cal
I}_{n}[h](X_{i};\lambda)=h(X_{i};\lambda),\qquad\forall\,i=1,\dots,n.$ (98)
For the moment, we shall assume that the EIM nodes $X_{i}$ are known and
proceed to describe how to use them to find the EIM interpolant. This can be
somewhat misleading, since it is not the way it works in practice, which is:
the first EIM node is found, its associated interpolant built, the second EIM
node found, the interpolant enriched to take it into account, and so on. They
are not disjoint processes, unlike (again) standard spectral methods, where
all the Gaussian nodes are found and afterwards the interpolant built. This is
not only a procedural difference, but highlights a big difference of the EIM:
namely that the approach (nodes and interpolant) is hierarchical. Hopefully
this gradual presentation is intuitive and pedagogical, later in this section
we will present the full, coupled algorithm together with its numerical
intricacies.
Equation (98) is equivalent to solving the $n$-by-$n$ system
$\sum_{i=1}^{n}{\bf V}_{ji}C_{i}(\lambda)=h(X_{j};\lambda)\,,\quad
j=1,\ldots,n$ (99)
for the coefficients $\\{C_{i}\\}_{i=1}^{n}$, where the interpolation matrix
$\bf{V}:=\left(\begin{array}[]{cccc}e_{1}(X_{1})&e_{2}(X_{1})&\cdots&e_{n}(X_{1})\\\
e_{1}(X_{2})&e_{2}(X_{2})&\cdots&e_{n}(X_{2})\\\
e_{1}(X_{3})&e_{2}(X_{3})&\cdots&e_{n}(X_{3})\\\
\vdots&\vdots&\ddots&\vdots\\\
e_{1}(X_{n})&e_{2}(X_{n})&\cdots&e_{n}(X_{n})\\\ \end{array}\right)$ (100)
is a generalization of the Vandermonde matrix – see Eq. (81) – when polynomial
basis are used. We have already discussed that in that case the Vandermonde
matrix can easily be very ill-conditioned if the nodes are chosen, for
example, equally spaced, not to mention if they are scattered. So one can
anticipate that the solution to the problem (97) is already non-trivial and
one of the goals of the EIM is to make sure that it does not lead to an ill
conditioned problem, as well as providing a high accuracy interpolant.
The choice of empirical nodes given by the EIM together with the linear
independence of the reduced basis ensure that $\bf{V}$ (Eq. (100)) is
invertible so that
$C_{i}=\sum_{j=1}^{n}\left({\bf V}^{-1}\right)_{ij}h(X_{j};\lambda)$ (101)
is the unique solution to (98). It then follows upon substituting (101) into
(97) that the empirical interpolant is
${\cal I}_{n}[h](x;\lambda)=\sum_{j=1}^{n}B_{j}(x)h(X_{j};\lambda)$ (102)
where
$B_{j}(x):=\sum_{i=1}^{n}e_{i}(x)\left({\bf V}^{-1}\right)_{ij}$ (103)
is independent of $\lambda$.
Note that (102) is a linear combination of the waveform itself evaluated at
the empirical nodes. The coefficients $\\{B_{i}\\}_{i=1}^{n}$ satisfy
$B_{i}(X_{j})=\delta_{ij}$ and are built directly from the reduced basis. They
provide a clean offline/online separation. Because of this the
$\\{B_{i}\\}_{i=1}^{n}$ functions can be pre-computed offline once the reduced
basis is generated while the (fast) interpolation is evaluated during the
online stage from (102) when the parameter $\lambda$ is specified by the user.
Evaluations of the waveform at the EIM nodes are still needed at the
arbitrarily chosen parameter $\lambda$ in order to construct the interpolant
in (102). One can wonder how this can be of any use. In Section 11 we explain
how to build surrogate predictive models for the evaluations of
$h(X_{j};\lambda)$ for any $\lambda$ (that is, not present in the original
training space).
Next we discuss how to compute the EIM nodes in a rather qualitative way,
before presenting the general algorithm later on.
Example 8. Finding the EIM nodes and building the interpolant.
Consider for definiteness a time series. The algorithm takes as input the
basis set $\\{e_{i}\\}_{i=1}^{n}$ and an arbitrary number and choice of time
samples $\\{t_{i}\\}_{i=1}^{L}$ from which the empirical interpolation nodes
$\\{T_{i}\\}_{i=1}^{n}$ are to be selected. The EIM algorithm proceeds as
follows
1. 1.
The first time node is chosen to maximize the value of $|e_{1}(t_{i})|$; that
is, $\left|e_{1}(T_{1})\right|\geq\left|e_{1}(t_{i})\right|$ for all time
samples $t_{i}$.
2. 2.
Next, an empirical interpolant for the second basis function is built using
only the first basis function: From Eqs. (97,98) we have ${\cal
I}_{1}[e_{2}](t)=C_{1}e_{1}(t)$ where $C_{1}=e_{2}(T_{1})/e_{1}(T_{1})$ has
been found from Eq. (98) with $i=1$.
3. 3.
The second empirical interpolation node is chosen to maximize the value of the
pointwise interpolation error of ${\cal I}_{1}[e_{2}](t)-e_{2}(t)$; that is,
$\left|{\cal I}_{1}[e_{2}](T_{2})-e_{2}(T_{2})\right|\geq\left|{\cal
I}_{1}[e_{2}](t_{i})-e_{2}(t_{i})\right|$ for all data samples.
4. 4.
Steps $2$ and $3$ are then repeated to select the remaining $(n-2)$ nodes. The
full algorithm for generating the EIM nodes is shown in Algorithm 2.
Comment 5. As described, the EIM follows a greedy approach, albeit somewhat
different from that one usually used to build a reduced basis. While a greedy
algorithm to build a RB selects the most relevant points in parameter space,
the EIM selects the most relevant points in the physical dimension(s). In
addition, the former uses a (possibly weighted) $L_{2}$ norm while the EIM
uses the infinity one.
Comment 6. Notice also that even though constructing the EIM is also done
offline, it has a much smaller computational cost than constructing the basis,
since it operates only on the latter, the training set is not involved
anymore.
The following example graphically shows the first three iterations of the EIM
algorithm applied to Legendre polynomials.
Example 9.
Consider the set $\\{P_{i}(x)\\}_{i=0}^{n}$ of $(n+1)$ normalized Legendre
polynomials defined on $[-1,1]$. These form an orthonormal basis for the space
of degree $n$ polynomials wrt the weight function $\omega(x)\equiv 1$. Their
ordering is important for approximations: expanding, by orthogonal projection,
a smooth function using the first $n$ Legendre polynomials typically results
in exponential convergence [71, 146]. Said another way, $P_{0}$ is the most
important basis function, $P_{1}$ the next most important, and so on.
Given a convergence rate for the aforementioned Legendre projection-based
approximation we might wonder how much accuracy is lost by trading it for the
interpolation (97) and how to optimally choose the nodes $\\{X_{i}\\}$. When
the relevant error measurement is the maximum pointwise error, Chebyshev nodes
are known to be well suited for interpolation, bringing an additional error
which is bounded by $\log(n)$ [108, 113], as discussed in Section 9.3 (see
Example 5). For this reason, Chebyshev nodes are specially tailored to
benchmark the nodes selected by the EIM in this example.
We select the EIM nodes according to the same preferential ordering as the
Legendre polynomials themselves; the first six normalized polynomials are
shown in Fig. 6.
Figure 6: The first six normalized Legendre polynomials.
For this example we chose the first 24 Legendre polynomials as the initial
basis for feeding the EIM algorithm. Figure 7 shows the process of the EIM
algorithm to choose the interpolation nodes for the Legendre basis. The first
EIM node is defined by the location of max$(|P_{0}|)$. Since $P_{0}$ is a
constant function there is no preference in the search, so we simply select
the middle point $x=0$. To identify the second node we
1. 1.
build the empirical interpolant ${\cal I}_{0}[P_{1}]$ of $P_{1}$ using $P_{0}$
as the basis and $x=0$ as interpolation node (the first EIM node),
2. 2.
compute the pointwise error $r_{1}=P_{1}-{\cal I}_{0}[P_{1}]$, and
3. 3.
select the second EIM node by the location of max$(\left|r_{1}\right|)$. In
this case we find $r_{1}=P_{1}$ since ${\cal I}_{0}[P_{1}]$ is zero due to the
fact that $P_{1}(0)=0$.
The process continues until the number of EIM nodes equals the number of basis
elements.
Figure 7: Representation of the EIM algorithm acting on the first 24
normalized Legendre basis. We show the bases, residuals and EIM nodes involved
in the iterations. The first three figures (upper-left, upper-right and
bottom-left) illustrate the two steps of the algorithm that maximize residuals
and select the EIM nodes. The last panel (bottom-right) shows the latest
iteration of the algorithm, which corresponds to the $24$th basis. The
vertical line corresponds to the residual’s maximum.
In Fig. 8 we show the $24$ EIM nodes found by the algorithm and compare them
to the first 24 Chebyshev nodes in the interval $[-1,1]$. Both node
distributions are qualitatively similar. Note that the EIM distribution
emulates the clustering of the Chebyshev nodes at the edges $x\pm 1$.
Reference [90], which has inspired our numerical experiment with Legendre
polynomials, also compares the Lebesgue constant(s) for the EIM nodes of this
example and the and Chebyshev ones.
Figure 8: A direct comparison between the distribution of the 24 EIM nodes for
the Example 10.1 and the first 24 Chebyshev nodes.
Example 10.
This example, taken from Ref. [34], graphically shows the EIM algorithm’s
application to basis functions for which a good set of interpolation nodes is
unknown. This is a sine-Gaussian family of the form
$h(t;\lambda):=Ae^{-(t-t_{c})^{2}/(2\alpha^{2})}\sin(2\pi f_{0}(t-t_{c}))\,,$
(104)
which models gravitational waves from generic burst processes. Here $A$,
$f_{0}$ and $\alpha$ are the amplitude, frequency and width of the waveform
$h$ respectively, $t_{c}$ is the arrival time of the GW-burst signal, and
$t\in[-\infty,\infty]$. The Fourier transform (FT) of this waveform is given
by
$\displaystyle{\tilde{h}}(f,t_{c};\lambda)=e^{i2\pi
ft_{c}}{\tilde{h}}(f;\lambda)\,,$ (105)
where ${\tilde{h}}(f;\lambda)$ is the FT of the GW-burst at $t_{c}=0$:
$\displaystyle{\tilde{h}}(f;\lambda)=i2A\alpha\sqrt{2\pi}\sinh(4\pi^{2}\alpha^{2}f_{0}f)e^{-2\pi^{2}\alpha^{2}(f_{0}^{2}+f^{2})}.$
(106)
This waveform family is, then, described by four free parameters
$\lambda=(\alpha,f_{0},t_{c},A)$. We focus on the two parameters
$(\alpha,f_{0})$, since the others are extrinsic and can be handled
differently. The parameter space considered is
$\alpha=[.02,2]\sec\quad\,,\quad f_{0}=[.01,1]{\rm Hz}\,,$ (107)
Figure 9 provides a graphical illustration of the EIM first iterations.
Figure 9: Iterations 1 (left) and 2 (right) of the EIM algorithm for Example
10.1. The first EIM node is defined by the location of max$(|e_{1}|)$. To
identify the second node one: i) builds the empirical interpolant ${\cal
I}_{1}[e_{2}]$ of $e_{2}$ using $e_{1}$ and the EIM node $F_{1}$, ii) computes
the pointwise error ${\cal I}_{1}[e_{2}]-e_{2}$; iii) the second EIM node is
then defined by the location of max$(\left|{\cal I}_{1}[e_{2}]-e_{2}\right|)$.
The process continues until all $n$ empirical interpolation nodes are found.
Figures taken from [34].
#### 10.2 EIM algorithm
We present the EIM algorithm as used/introduced in Ref. [54].
Algorithm 2 The Empirical Interpolation Method
1:Input: rb = $\\{e_{i}\\}_{i=1}^{n}$
2:$X_{1}=\text{argmax}_{x}|e_{1}|$
3:for $j=2\to n$ do
4: Build ${\cal I}_{j-1}[e_{j}](x)$
5: $r_{j}(x)=e_{j}(x)-{\cal I}_{j-1}[e_{j}](x)$ ($r$ stands for residual)
6: $X_{j}=\text{argmax}_{x}|r_{j}|$
7:end for
8:Output: EIM nodes $\\{X_{i}\\}_{i=1}^{n}$ and interpolant ${\cal I}_{n}$
Comment 7. We make some remarks on the algorithm:
* •
It uses a greedy algorithm in the infinity norm.
* •
The iteration is performed only on the reduced basis, and there is no need of
a training set as it was the case in the construction of the basis itself.
* •
At each greedy sweep it chooses the node that differs most from the
interpolant built so far.
* •
The selection of EIM nodes is basis-dependent: different choices of the basis
should lead to different nodes. However, once the nodes are computed, the
interpolant becomes basis independent [40]. From a geometric viewpoint, one
realizes that the interpolant is a projector (though not orthogonal) onto the
span of the basis. Once the reduced space and the nodes are chosen, the
interpolant becomes uniquely defined.
#### 10.3 Accuracy and conditioning of the EIM
In Section 9.3 we discussed how the quality of an interpolant in a weighted
$L_{2}$ norm is related to the optimal projection representation.
Interpolation in general loses accuracy but at the advantage of needing less
information for computing the approximation. One can bound the interpolation
error defining a Lebesgue constant (see Eq. (89)):
$\|h_{\lambda}-{\cal
I}_{n}[h_{\lambda}]\|^{2}\leq\Lambda_{n}\|h_{\lambda}-{{\cal
P}}_{n}h_{\lambda}\|^{2}\,,\quad\Lambda_{n}:=\|{\cal I}_{n}\|^{2}\,.$ (108)
Taking the maximum in both sides of the inequality, and recalling the
definition of the greedy projection error in Eq. (61), we obtain
$\max_{\lambda}\|h{{}_{\lambda}}-{\cal
I}_{n}[h_{\lambda}]\|^{2}\leq\Lambda_{n}\sigma_{n}\,.$ (109)
Comment 8. The maxima in (109) involve some ambiguities, as they might refer
to those with respect to the training set or those of the underlying space of
interest. It is problem and user-dependent to clarify which is the case, but
the above results apply to both of them.
Comment 9. As discussed below, the Lebesgue constant is an a posteriori
computable number, so if one has an estimate of the projection error
$\sigma_{n}$, also has an estimation of the interpolation one. Assuming there
is enough training data, an estimate for $\sigma_{n}$ can be accomplished by
standard validation tests. Otherwise, if the data is sparse, one may apply
other standard methods from machine learning such as k-fold or leave-one-out
cross-validations.
In the context of the EIM, some obvious questions arise from these
observations: how is the EIM related to the Lebesgue constant? Does it
optimize for accuracy in some way? These questions were addressed in Ref.
[138].
First notice that, at each step of the EIM, when building the interpolant, a
Vandermonde matrix ${\bf V}_{n}$ (Eq. (100)) has to be inverted for computing
the interpolation coefficients. The nature of the algorithm already ensures
the invertibility of ${\bf V}_{n}$ at all steps. It turns out that the EIM in
fact optimizes with respect to this invertibility.
Theorem. Define $\textup{det}({\bf V}_{j}):=V_{j}(X_{1},\ldots,X_{j})$. Then,
the residual $r_{j}(x)$ computed in the $(j-1)$-iteration of the EIM-loop
satisfies
$r_{j}(x)=\frac{V_{j}(X_{1},\ldots,X_{j-1},x)}{V_{j-1}(X_{1},\ldots,X_{j-1})}\quad
j=2,3,\ldots n\,.$ (110)
In consequence, once $X_{j}$ is chosen in Step 6, the residual at this node
becomes
$r_{j}(X_{j})=\frac{V_{j}(X_{1},\ldots,X_{j-1},X_{j})}{V_{j-1}(X_{1},\ldots,X_{j-1})}=\frac{\text{det}({\bf
V}_{j})}{\text{det}({\bf V}_{j-1})}\,.$ (111)
The proof can be found in [138]. This result says that, at each $j$-step, the
EIM algorithm selects a new node $X_{j}$ in order to maximize the module of
the determinant of ${\bf V}_{j}$, making the Vandermonde matrix as invertible
as possible at each iteration.
The following is a useful identity to compute Lebesgue constants in practice
[17]: if the basis vectors are orthonormal in the 2–norm $\|\cdot\|_{2}$, then
$\Lambda_{n}=\|{\bf V}_{n}^{-1}\|^{2}_{2}\,.$ (112)
We use this relation and the formulation of the inverse of a matrix in terms
of its adjoint to rewrite the Lebesgue constant as
$\Lambda_{n}=\frac{\|\text{adj}({\bf V}_{n})\|^{2}_{2}}{|\text{det}({\bf
V}_{n})|^{2}}\,.$ (113)
In doing this, we see that when maximizing the determinant of the Vandermonde
matrix at each step the EIM attempts to partially control the growth of
$\Lambda_{n}$, in the sense of making the denominator of (113) as large as
possible, but without controlling its numerator.
So, the algorithm does not solve for the optimization problem of finding
global (since it is a nested approach) nodes that maximize the determinant of
the Vandermonde matrix, but neither the partial problem of minimizing the
whole Lebesgue constant at each step, as is usually thought. Analogous
observations can be made for the conditioning of ${\bf V}_{n}$, since both the
Lebesgue constant and the condition number of ${\bf V}_{n}$ are directly
related by
$\kappa_{n}=\|{\bf V}_{n}\|_{2}\Lambda_{n}\,.$
For further discussions around these topics in the context of GWs, see [138].
Finally, we point out that, in the context of the EIM, a rather pessimistic
bound for the Lebesgue constant can be derived [40]:
$\Lambda_{n}\leq(1+\sqrt{2L})^{n-1}\|e_{1}\|_{\infty}^{-1}\,,$
where $L$ stands for the size of the sampling of the $x$ variable. As an a
priori estimate, this bound is not useful in practice since it does not
reflect the observed growth of $\Lambda_{n}$, which is, in most cases of
interest, much slower. Sharper a priori estimates are difficult to prove. In
practice, one computes the Lebesgue constant a posteriori.
#### Further reading
The original papers introducing the EIM are [20] and [90]. Reference [40]
presents a very clearly explained discrete version with emphasis on solutions
to non-linear time-dependent parametrized equations. In the latter case one is
interested in approximating non-linear terms, in similarity with Galerkin
versus collocation methods when solving partial differential equations. An hp-
refinement of the standard EIM approach is presented in [51]; this in general
should be of very practical importance for several scenarios, such as when
there are discontinuities with respect to parameter variation, but it has so
far not been used in GW science. The study of conditioning of Vandermonde
matrices can be tracked to [59], where orthogonal polynomials are used to
improve the condition number. See also [60, 61], and, for a short review,
[72]. For different studies around Vandermonde-like matrices see [78, 46, 104,
115].
### 11 Surrogate models
The first kind of surrogate models for gravitational waves followed the lines
of fitting for the projection coefficients of a reduced basis, be it obtained
through a POD or greedy approach. That is, given a basis
$\\{e_{i}\\}_{i=1}^{n}$ the RB approximation for a waveform $h$ is
$h(t;\lambda):=\sum_{i=1}^{n}c(\lambda_{i})e_{i}(t)\,,$ (114)
where the basis elements $\\{e_{i}\\}$ can be waveforms themselves as in the
RB-greedy approach, or linear combinations of them as in the POD approach.
This representation requires knowledge of the projection coefficients
$\\{c_{i}\\}$, which are only known if the waveform itself is known. That is,
as it stands, Eq. (114) is not a predictive tool but a representation one. A
natural step to predict new waveforms is to build a surrogate by simply
fitting in some way for new values of the parameter $\lambda$:
$h_{\tt s}(t,\lambda):=\sum_{i=1}^{n}c_{{\tt s}\,,i}(\lambda)e_{i}(t)\,,$
(115)
where $\\{c_{{\tt s}\,,i}\\}$ are $n$ approximations to the true projection
coefficients $\\{c_{i}\\}$. There is a large variety of ways to fit for these
coefficients, such as polynomial interpolation, splines, radial basis, or ML
regression approaches. A number of them (without much success) were studied in
Ref. [2].
One of the problems found when the basis elements are not gravitational waves
themselves, as in the POD case, or if using the auxiliary orthonormal basis in
the RB-greedy approach, is that they have non-physical complex structure in
their parameter dependence that is difficult to represent, this is also
discussed in Appendix F of [54]. One solution is to use as basis elements the
waveforms themselves, something that is not possible within the POD approach,
since it does not provide a set of the most representative waveforms to use.
Another approach, if a dense training space is available, is to use local fits
(such as splines) to find these unknown projection coefficients. The
limitation of needing a dense training set is precisely being able to generate
it.
The main source of errors in these approaches is the fitting procedure, which
in general has slow convergence. Therefore, fitting only at the EIM nodes, as
described below, serves at least two goals: i) to minimize the source of
fitting errors by doing so only at the EIM nodes, ii) to provide very fast to
evaluate surrogate models. Impressive results have been achieved, which we
discuss in Section 12, using this approach. Other approaches follow non-
algorithmic schemes and the literature in this field is rather new and in
constant evolution. To mention only a few, see for example [69, 147] in which
neural networks are combined with reduced order modeling or [41] for the same
approach in the context of GWs. We leave for future updates of this review a
thorough discussion of these topics.
#### 11.1 Surrogate models for components
Even though we have emphasized data-driven approaches, some domain knowledge
is still (very) useful. For example, the GWs emitted by two black holes have
an apparent complexity, but can be decomposed into components with simple
structure and easier to build models for. For example, in the non-precessing
case the GWs are essentially oscillatory functions of time with an increasing
amplitude, until around the time of merger, followed by the ringdown regime.
Thus, it is advantageously to decompose them into phase and amplitude:
$h(t;\lambda)=A(t,\lambda)e^{i\Phi(t,\lambda)}\,.$ (116)
The structure of the waveforms themselves as well as phase and amplitude for
the case of two black holes initially in quasi-circular orbit and without
spin) are shown in Figures 10 and 11. One can notice the simplification in
structure when decomposing into phase and amplitude.
Figure 10: Representation of the $h_{+}$ polaritazion of a waveform
corresponding to the mode $(2,2)$ for a BBH coalescence process initially in
quasi-circular orbit without spin and mass ratio $q=m_{1}/m_{2}=1.7$. This
picture corresponds to the surrogate model SpEC_q1_10_NoSpin [25], and was
generated using the gwsurrogate Python package [5].
Figure 11: Representation of amplitude $A(t,\lambda=q)$ and phase
$\phi(t,\lambda=q)$ surfaces for the waveform mode $(2,2)$ corresponding to a
symbolic model [137] for the surrogate plotted in Fig. 10 above.
Similarly, in the precessing case using a co-precessing frame simplifies the
structure of the waveforms, and it is advantageous to include the coordinate
transformations themselves as components to model for. A question arises of
whether to build a reduced basis for the waveforms and later build surrogate
models for the system components, or to start by building bases for the latter
and then reconstruct the waveform from these surrogate components or pieces.
In general, the latter approach should be more accurate, even if at the
expense of higher offline cost.
#### 11.2 Empirical interpolant based surrogates
The surrogate approach here described was introduced in Ref. [54], the
presentation of which we follow closely. It represents, with some conceptually
relatively minor variations, the state-of-the art as of this writing. The
approach has three offline stages, which are described below in decreasing
order of computational cost, and a fourth, very fast to evaluate online
surrogate model.
Offline stages
1. 1.
Select the most relevant $n$ points in parameter space (shown as red dots in
Fig. 12) using, for example, a greedy algorithm as described in Section 8.3.
The waveforms/functions associated with these selections (shown as red lines)
provide a nearly optimal reduced basis for the space of interest ${\cal F}$.
2. 2.
Identify $n$ time (or frequency) samples of the full time series using the
EIM, to build an interpolant that accurately reconstructs any fiducial
waveform for all times if it is known at the EIM nodes. These nodes are shown
as blue dots on the vertical axis in Fig. 12.
3. 3.
At each EIM node perform a fit with the method of choice in the parameter
dimension for the amplitude and phase of the waveform using the greedy data
from Step 1. The fits are indicated by blue horizontal lines in Fig. 12.
Figure 12: A schematic of the method for building and evaluating the surrogate
model, taken from [54]. The red dots show the greedy selection of parameter
points for building the reduced basis (Step 1, offline), the blue dots (Step
2, offline) show the associated empirical nodes in time from which a waveform
can be reconstructed by interpolation with high accuracy, and the blue lines
(Step 3, offline) indicate a fit for the waveform’s parametric dependence at
each EIM node. The yellow dot shows a generic parameter, which is predicted at
the yellow diamonds and filled in between for arbitrary times using the EIM,
represented as a dotted black line (Step 4, online).
For further illustration, we show in Fig. 13 the distribution of EIM nodes for
gravitational waves corresponding to an EOB model, taken from [54]. Notice the
sparsity of the distribution of EIM subsamples (nodes): only 19 empirical
nodes are needed to reconstruct the whole time series consisting of roughly
$65-70$ waveform cycles for the entire parameter space. From a signal
processing perspective, this may sound quite strange, since the EIM sampling
does not seem to meet the minimal Nyquist sampling needed to reconstruct the
time series. This apparent contradiction is resolved by realizing that the
reduced basis already encodes the physical dependency of the model, so, there
is indeed information in the background working for a faithful reconstruction
of the waveforms.
Figure 13: Example of a distribution of the EIM nodes corresponding to a
fiducial model composed by EOB waveforms. Only $19$ nodes are needed to
represent $65-70$ cycles within about machine accuracy. Figure taken from
[54].
Step by step offline construction
Put in a more specific way, the surrogate model is built as follows. Given
$\\{\Lambda_{i}\\}_{i=1}^{n}$ greedy parameters and functions (waveforms) the
empirical interpolant for an arbitrary parameter value $\lambda$ has the form
(Section 10)
${\cal I}[h](t,\lambda)=\sum_{i=1}^{n}B_{i}(t)h(T_{i},\lambda)\approx
h(t,\lambda)\,,$ (117)
where the $\\{B_{i}\\}$ functions are computed offline from the basis, as well
as the EIM nodes $\\{T_{i}\\}$.
In the case of non-spinning binaries, it is convenient to decompose all
waveforms $h$ into their phase and amplitudes as in Eq. (116) and, in
particular at the EIM nodes,
$h(T_{i},\lambda)=A(T_{i},\lambda)e^{i\Phi(T_{i},\lambda)}\,.$ (118)
For each $i=1\ldots n$, build fits for amplitude and phase,
$A(T_{i},\lambda)\approx A^{\tt
fit}(T_{i},\lambda)\quad,\quad\Phi(T_{i},\lambda)\approx\Phi^{\tt
fit}(T_{i},\lambda)\quad\forall\lambda\,.$
Replacing these approximants, which are so far only valid for the EIM nodes
$\\{T_{i}\\}$, as opposed to arbitrary values of $t$, into (119) yields
$h_{\tt s}(T_{i},\lambda):=A^{\tt fit}(T_{i},\lambda)e^{i\Phi^{\tt
fit}(T_{i},\lambda)}\,.$ (119)
Finally replacing $h$ by $h_{\tt s}$ in Eq. (117) leads to the final surrogate
model for all values of $t$ and $\lambda$,
$h_{\tt s}(t,\lambda):=\sum_{i=1}^{n}B_{i}(t)A^{\tt
fit}(T_{i},\lambda)e^{i\Phi^{\tt fit}}(T_{i},\lambda)\quad\forall
t,\lambda\,.$ (120)
Expressions similar to Eq. (120) are the final expressions of the surrogates
used so far within numerical relativity, while decomposing the waveforms into
more components in the case of spin and precession. Importantly, once the
training set has been processed to build the bases, the surrogate model only
relies on the knowledge of the bases, the training set is no longer involved.
Its accuracy (error estimates) and evaluation costs are discussed next.
Online stage, evaluation cost
We now discuss the cost, in terms of operation counts, to evaluate a surrogate
model. For simplicity we count every arithmetic operation as a single one
regardless of their actual complexity.
Also for simplicity and definiteness, still restricting the discussion to the
non-spinning case, the complete surrogate model is given by Eq. (120), where
the $n$ coefficients $B_{i}(t)$ and the $2n$ fitting functions
$\\{A_{i}(\lambda)\\}_{i=1}^{n}$ and $\\{\phi_{i}(\lambda)\\}_{i=1}^{n}$ are
assembled offline as described above.
In order to evaluate the surrogate model for some parameter $\lambda_{0}$ we
only need to evaluate each of those $2n$ fitting functions at $\lambda_{0}$,
recover the $n$ complex values
$\\{A_{i}(\lambda_{0})e^{-i\phi_{i}(\lambda_{0})}\\}_{i=1}^{n}$, and finally
perform the summation in (120). Each $B_{i}(t)$ is a complex-valued time
series with $L$ samples, where $L$ is the desired number of time or frequency
outputs.
Therefore, the overall operation count to evaluate (120) at each $\lambda_{0}$
is $(2n-1)L$ plus the cost to evaluate the fitting functions. At least in one-
dimension, the evaluation cost of polynomials of order $n_{\tt fit}$ using
Horner’s method (which is known to be optimal in terms of operation counts
[103]) is $2n_{\tt fit}$. We discussed the exponential cost of evaluating
polynomials in higher (than one) dimensions in Section 9.5 and the final
comments of that whole Section. We present some further discussions on this
and related topics related to the fitting stage in Section 11. To summarize,
the cost of evaluating the surrogate (120) is
${\tt cost_{s}}=\left[(2n-1)+c_{\tt fit}\right]L$ (121)
where at least in $1$-dimension $c_{\tt fit}\leq 2n$ and therefore
${\tt cost_{s}}\leq\left(4n-1\right)L\sim{\cal O}(L)$ (122)
Comment 10.
* •
The evaluation of the surrogate model (120) is embarrassingly parallel in the
number of outputs $L$ (each output evaluation is completely independent of the
other ones) and therefore the total computational time can be speed up as
wished. In contrast, time dependent equations (ordinary or differential) are
very difficult to parallelize in time.
* •
The cost of most ordinary and partial differential equations solvers is at
best linear in $L$. The speedup of the described surrogate approach for
evaluation, if parallelized, is actually independent of the number of desired
outputs $L$.
* •
In Eq. (122), the linearity in $L$ is simply due to the number of independent
output sample evaluations. The optimal cost would be ${\tt cost_{s}}=L$,
corresponding to one arithmetic operation to evaluate the surrogate at any
output $x$ (time or frequency, in the context of this review) of interest. If
$n$ is small enough, as it happens in practice for GWs, the operation count
$(\ref{eq:cost_surrogate})$ is in a precise sense, very efficient.
Error estimates
One of the errors of interest for the complete surrogate model is a discrete
version of the $L_{2}$ normed difference between a fiducial waveform and its
surrogate. For definiteness we consider equally spaced $L$ time output
samples,
$\|h(\cdot,\lambda)-h_{\tt S}(\cdot,\lambda)\|^{2}:=\Delta
t\sum_{i=1}^{L}\left|h(t_{i};\lambda)-h_{\rm S}(t_{i};\lambda)\right|^{2}\,,$
(123)
where $\Delta t=(t_{\mathrm{max}}-t_{\mathrm{min}})/(L-1)$. Other errors of
interest are the pointwise ones for the phase and amplitude,
$\left|\frac{A(t;\lambda)-A_{\rm
S}(t;\lambda)}{A(t;\lambda)}\right|\,,\quad|\phi(t;\lambda)-\phi_{\rm
S}(t;\lambda)|\,.$ (124)
The following error bound for the discrete error (123) can be derived:
$\|h(\cdot,\lambda)-h_{\tt
S}(\cdot,\lambda)\|^{2}\leq\Lambda_{n}\sigma_{n}+\Lambda_{n}\Delta
t\sum_{i=1}^{n}\left[h(T_{i},\lambda)-h_{\rm S}(T_{i},\lambda)\right]^{2}\,.$
(125)
For a proof, see Ref. [54]. This bound identifies contributions from two
sources. The first term in the r.h.s. of (125) describes how well the EIM
interpolant (i.e., the basis and EIM nodes) represents $h(t;\lambda)$. The
expected exponential decay of the greedy error $\sigma_{n}$ with $n$ along
with a slowly growing Lebesgue constant $\Lambda_{n}$ results in this term
being very small. The second term in the r.h.s of (125) is related to the
quality of the fit. Currently, the fitting step is the dominant source of
error in the surrogate models constructed, compared to the first two steps of
generating the reduced basis and building the empirical interpolant. Improving
on this source of error is still a remaining goal.
#### Further reading
Certified approaches refer to those which can a priori guarantee error bounds
for the reduced model compared to the ground truth. Up to our knowledge these
have been so far restricted to elliptic (coercive) partial differential
equations. We again refer to [112], [70] and [111]. For an implementation of
the Reduced Basis and Empirical Interpolation Methods in a concise and user-
friendly API, see the Arby [3] Python package, a fully data-driven module for
building surrogate models, reduced bases and empirical interpolants from
training data.
### 12 Surrogates of compact binaries
To date the most accurate binary black hole (BBH) numerical relativity (NR)
surrogates have used a reduced bases-greedy approach, the EIM, and surrogate
construction as discussed in this review, with some variations, and the SpEC
code [13] for numerical relativity training simulations. However, other public
catalogs of NR waveforms such as that one of the RIT group could be equally
used [67, 12]. The surrogate catalog can be found in [4] and can be evaluated
with the gwsurrogate [5] Python package. Another package for surrogate
evaluations is surfinBH [14], built on top of gwsurrogate. What is found in
all these surrogates is that when compared to numerical relativity, they are
at least an order of magnitude more accurate than other existing models such
as hybrid, phenomenological or effective ones.
In addition to NR surrogates, we also discuss other ones using as fiducial
models post-Newtonian, Effective One Body, and Ringdown approximations.
#### 12.1 Numerical relativity binary black holes
1. 1.
Non spinning (1 dim):
The first BBH NR surrogate was presented in [25], and referred to as
SpEC_q1_10_NoSpin. The black holes are initially non-spinning with initial
orbital eccentricity smaller than $10^{-3}$, in the time range $[-2750,100]M$
where, as is common practice, the waveforms have been aligned so that $t=0$
stands for the peak of the amplitude (which is around the merger of the two
black holes), corresponding to about $15$ orbits before merger. This is a one-
dimensional parameter problem, with the mass ratio $q$ chosen in the range
$q:=m_{1}/m_{2}\in[1,10]$ .
The selected $17$ greedy values were taken from an EOB model, seeded with
existing five parameters corresponding to numerical relativity simulations
using the Einstein equations. Next, numerical relativity was used to solve for
the remaining $17$ parameters and later the reduced basis built, of $22$
elements total. Impressively, the resulting surrogate model includes all
spherical harmonic modes up to $\ell=8$.
One could also ask for a symbolic model representing the numerical surrogate
one. In this line, in [137] the authors constructed the first ab initio free-
form symbolic model (that is, analytical expressions in terms of elementary
functions) for gravitational waves using symbolic regression (SR) through
genetic programming. The fiducial model corresponds to the principal
$(2,2)$-mode of the surrogate SpEC_q1_10_NoSpin described in [25], which is
taken as ground truth solution of the Einstein field equations, since it is
practically indistinguishable (meaning, within the numerical relativity errors
themselves) from supercomputer numerical simulations. The approach is ab
initio, meaning no approximations to the Einstein equations are taken, such as
stitching PN waveforms with EOB, NR and ringdown ones. The search for closed
expressions is completely free, meaning that no prior hypothesis related to
the type of functions is made. This is at the heart of genetic programming:
successive models evolve under evolutionary pressure until reaching a
tolerance error (or another stopping condition) without incurring in any human
bias. To generate the training set of waveforms for SR, instead of performing
naive samplings such as equally spaced grids (which showed prohibitive in
terms of convergence times), the key step was to sample the time domain with
the $22$ empirical interpolation nodes (EIM nodes) used in the assembly of the
fiducial model SpEC_q1_10_NoSpin. This sampling represented the minimal set of
pieces of information to represent the whole fiducial model without loss of
structure, then accomplishing fast convergence in the regression instance and
ending with closed-form expressions of maximum overlap error of 1% with
respect to the NR surrogate model. One of the salient features of these closed
expressions is that they are not divided into expressions for the inspiral
regime, for the merger, and finally for the ringdown, but cover the whole
inspiral-merger-ringdown regime.
2. 2.
Spinning, motivated by GW150914 (4-5 dim):
Continuing the work of Refs. [25], Ref. [26] presented the first spinning BBH
surrogate model, with the parameter region motivated by the first GW
detection, GW150914.
Two surrogates were built: NRSur4d2s_TDROM_grid12 and NRSur4d2s_FDROM_grid12.
The physical range is $q\leq 2$, dimensionless spin magnitudes $\chi_{1,2}\leq
0.8$, and the initial spin of the smaller black hole along the axis of the
orbital angular momentum. The parameter region includes LIGO’s first
detection, GW150914, though with less cycles (using for training NR
simulations with $25-35$ cycles before merger, while GW150914 has around $55$
cycles).
NRSur4d2s_TDROM_grid12 was built using $276$ NR simulations as training set
and PN to find the greedy points, for which the PN surrogate reaches a floor
error of $10^{-3}$. The parametric fits, used in the surrogate assembly, are
fixed to a particular “order” – not necessarily polynomial order, since
trigonometric functions were included as part of the dictionary of fitting
functions . The authors attribute the reason for this floor to that fixing,
the general problem of this issue is discussed in Section 15. Next, NR
simulations were performed for those PN greedy points using a minimally
rotating and co-precessing frame and, together with the coordinate
transformations to an inertial frame, reduced bases for the different “pieces”
or components of the waveform were built.
The initial spin of the smaller black hole by construction lays along the axis
of the orbital angular momentum, reducing the parameter dimensionality to $5$.
Next, the azimuthal component of the spin of the larger black hole at a
reference time $t_{0}=t_{\text{peak}}-4,500M$ is included through an
analytical approximation, thus effectively reducing the parameter
dimensionality to 4, while modeling $5$ dimensions. Thus, the model does
include a physical approximation and is not purely based on NR simulations,
while being able to include precession in the modeling.
Model NRSur4d2s_FDROM_grid12 is built not from the 276 NR training set
simulations but from the time domain surrogate NRSur4d2s_TDROM_grid12, which
can be quickly evaluated at a much larger number of points to populate a new,
more dense, training set.
3. 3.
Spinning (7d):
The first surrogate model, named NRSur7dq2, for the full 7-D parameter space
of GWs emitted for a non-eccentric BBH coalescence was presented in [27]. It
used $744$ NR simulations to construct the training set with parameter ranges
$q\leq 2$, $\chi_{1},\chi_{2}\leq 0.8$, for about $20$ orbits prior to merger,
and $\ell\leq 4$. The authors “recycled” the $276$ NR simulations used in the
spinning 4-D case described above and complete the total of $744$ NR
simulations with a metric-based population criteria to select the remaining
parameter points. Then the authors extend the number of simulations by means
of symmetry arguments to $886$. As in [26], but with certain improvements,
they decompose waveforms in data pieces and proceed to construct the surrogate
following as a guideline the 4-D case [26].
The in-sample errors computed for the $886$ NR waveforms show that the largest
surrogate errors are comparable to the largest NR resolution errors ($\sim
10^{-2}$). For estimating the out-of-sample errors, the authors performed
cross-validation over the training set by randomly dividing it in 20 sets of
$44-45$ waveforms. They left out $1$ set at each step and built a trial
surrogate for the remaining $19$ sets to compare it against the one that left
out. This resulted in mistmatches similar to those of the in-sample case. The
authors also compare the mismatches for a fully precessing EOB model (SEOBNRv3
[105]) and for a phenomenological waveform model which includes some effects
of precession (IMRPhenomPv2 [65]). Mismatches more than an order of magnitude
larger than the NRSur7dq2 surrogate model are found.
In [143] the 7-D parameter space is covered, for about $20$ orbits before
merger, mass ratios $q\leq 4$, arbitrary spin orientations with dimensionless
magnitudes $\chi_{1},\chi_{2}\leq 0.8$, $\ell\leq 4$ multipole modes, and
initial orbital eccentricity also less than $10^{-3}$. This model, NRSur7dq4,
extends the previous 7-D one and is to date the most exhaustive and general
surrogate model for BBHs from NR. The length of these simulations is
sufficient to represent some but not all the BBH signals measured by LIGO and
Virgo in the first two observation runs. It is proposed in the reference to
hybridize the NR waveforms with PN approximations for higher values of mass
ratio and spin magnitudes (see Sec. 12.2). In contrast to the above Ref. [27],
here $1,528$ NR simulations were used for the training set. In addition, a
surrogate model, named NRSur7dq4Remnant, is built for the mass, spin, and
recoil kick velocity of the remnant black hole. To test their models, the
authors perform a 20-fold cross-validation study on the training simulations.
First they randomly divide the $1,528$ training simulations into 20 groups of
$\sim 76$ simulations each. For each one, they built a trial surrogate using
the $\sim 1,452$ remaining training simulations and test against these $\sim
76$ validation ones. They show that the mismatches for NRSur7dq4 against NR,
computed with the Advanced LIGO design sensitivity noise curve, are always
$\lesssim 8\times 10^{-3}$ at the 95 percentile level over the mass range
$50-200M$. For NRSur7dq4Remnant the 95th percentile errors are $\sim 5\times
10^{-4}M$ for mass, $\sim 2\times 10^{-3}$ for spin magnitude, and $\sim
4\times 10^{-4}c$ for kick magnitude. Compared to the spinning EOB waveform
model SEOBNRv3[105], they found in the two models an improvement of errors of
at least one order of magnitude.
Name | mass ratio | $\chi_{1,2}$ | precession | dim | $\ell$ | GW cyles
---|---|---|---|---|---|---
SpEC_q1_10_NoSpin | $\leq 10$ | 0 | no | 1 | $\leq 8$ | $\sim 25-31$
NRSur4d2s_TDROM_grid12 | $\leq 2$ | $\leq 0.8$ | yes | 4 | $\leq 3$ | $\sim 25-35$
NRSur7dq2 | $\leq 2$ | $\leq 0.8$ | yes | 7 | $\leq 4$ | $\sim 40$
NRSur7dq4 | $\leq 4$ | $\leq 0.8$ | yes | 7 | $\leq 4$ | $\sim 40$
Table 1: Characterization of the waveform surrogates described so far
corresponding to NR binary black holes.
##### Kicks
Before the release of accurate surrogates for GWs covering the full 7-D
parameter space of BBH dynamics, black hole kicks were mostly modeled with
fitting formulas based on PN theory with the subsequent calibration based on
NR simulations. Reference [62] presented the first effort to determine remnant
properties from BH mergers using a ROM-based surrogate model. More
specifically, the authors used the NRSur7dq2 model [27] to generate a waveform
template and analyze the linear momentum dissipation due to the emission of
GWs. Their procedure provides the velocity accumulation profile ${\bf v}(t)$
and the final kick speed $v_{k}$ of the remnant black hole. The comparison for
recoil speeds between NR simulations and NRSur7dq2 shows well agreement within
an order of $\sim 10^{-4}c$, with some outliers in the order of $\sim
10^{-3}$c. The authors suggest that, even in the case where the surrogate
accurately models post-merger strains, small errors might propagate to the
phase of the center-of-mass-oscillation causing a relatively large error on
the final kick velocity.
More recently, in [145] the first surrogate models for remnant oscillations
were constructed using Gaussian process regression (a type of machine learning
regression method, see, for example, [49]) and NR simulations for training.
These fits are able to provide remnant mass, spin vector and recoil kick
vector with high accuracy for 1) precessing BHs with mass ratio $q\leq 2$ and
spin magnitudes $\chi_{1},\chi_{2}\leq 0.8$; 2) non-precessing BHs with mass
ratio $q\leq 8$ and anti-aligned spin magnitudes $\chi_{1},\chi_{2}\leq 0.8$.
#### 12.2 Numerical relativity hybrid binary black holes
The previous surrogate models do not cover the entire LIGO band. To remedy
this, in [144] a hybridized non-precessing model named NRHybSur3dq8 was
presented. In that work NR waveforms are “stitched” at early times with PN and
EOB ones, thus being able to cover the entire band of advanced LIGO with a
starting frequency of $20$Hz and for systems with mass as low as
$2.25M_{\odot}$. This model is based on $104$ NR simulations for the 3-D
parameter region $q\leq 8$, $|\chi_{z1}|,|\chi_{z2}|\leq 0.8$ for modes
$\ell\leq 4$ and $(5,5)$, excluding $(4,1)$ and $(4,0)$. To populate the
training space, the authors performed $91$ NR simulations and completed for a
total of $104$ with $13$ waveforms added using BHs exchange symmetry (equal
mass, unequal spin). The parameters for the $91$ NR simulations were selected
by a greedy procedure, iteratively constructing a PN surrogate model, testing
it with a dense validation set and selecting the next greedy-parameter for the
largest model error. At the hybridization stage, the early-inspiral waveforms
were stitched with NR ones minimizing a cost function by varying the time and
frame shifts between waveforms in an appropiate matching region. The matching
region was settled to start at $1,000M$ after the start of the NR waveform and
end after $3$ orbits of the binary inspiral. At the end, the authors find that
its hybridized surrogate model NRHybSur3dq8 performs well within NR truncation
errors and outperforms the SEOBNRv4HM spinning-EOB model [43] by about two
orders of magnitude.
As an application case, NRHybSur3dq8 was used in [19] to generate GWs to study
tidal effects by means of a PN tidal splicing method. The resulting model was
named NRHybSur3dq8Tidal. It was also added to the last update of the SXS
Collaboration catalog [30] of numerical simulations for BBH coalescences.
#### 12.3 Extreme mass ratio inspirals
In [116] the authors built a surrogate model, named EMRISur1dq1e4, for extreme
mass ratio inspirals using non-spinning point particle black hole perturbation
theory (ppBHPT) through numerical solutions of the Teukolsky equation with a
point particle as source. The trajectory of the particle was determined by an
adiabatic inspiral at early times, a late-stage geodesic plunge, and a
transition region. The mass ratio was taken to be in the range $[3,10^{3}]$.
After a mass rescaling the surrogate model agrees remarkably well with NR
waveforms (solving the full Einstein equations), which are available for mass
ratios $q\leq 10$. The mass rescaling was empirical, in the sense that it was
chosen as a function of the mass ratio and numerically chosen to minimize the
difference with NR waveforms for the $(2,2)$ mode. Even so, the degree of
agreement after rescaling is surprisingly good and unexpected since a priori
there is no reason why such a good agreement should be present at all. As the
authors point out, however, their result seems to be in line with growing
evidence that suggests perturbation theory with self-force corrections might
be applicable to nearly equal mass systems.
##### 12.3.1 Eccentric inspirals
In [21] the authors introduce a SVD-based ROM technique to model waveforms
emitted by the coalescence of compact binaries with any residual orbital
eccentricity. They apply this framework to eccentric-PN waveforms generated
with the CBwaves open-source software [44] and build a reduced order model for
a 3-D subset of waveforms of the full 8-D parameter space corresponding to
total mass $M$, mass ratio $q$ and eccentricity $e_{0}$. The ranges covered by
the template bank were $2.15M_{\odot}\leq M\leq 215M_{\odot}$, $0.01\leq q\leq
1$ and $0\leq e_{0}\leq 0.96$.
The speedup in evaluating the surrogate model is $2-3$ orders of magnitude
faster than generating the corresponding CBwaves waveforms, reaching a factor
of several thousand around $10-50M_{\odot}$.
#### 12.4 Effective One Body
In [54] accurate surrogate models for EOB waveforms of non-spinning BBH
coalescences were constructed using the Reduced Basis (RB) framework,
corresponding to modes $(2,1),(2,2),(3,3),(4,4)$ and $(5,5)$ with mass ratios
from $1$ to $10$. The authors benchmarked the surrogate model against a
fiducial one generated with the EOB solver of the LAL software package. For a
sampling rate of $2048$ Hz they found a speedup of $\approx 2,300$, about
three orders of magnitude faster than the LAL waveform model.
Reference [79] presented a surrogate model of a non-spinning EOB waveform
model with $l=2,3,4$ tidal multipole moments that reproduces binary neutron
star (BNS) numerical simulation waveforms up to merger. The authors find,
within the RB framework, that $12$ amplitude and $7$ phase basis elements are
sufficient to reconstruct any BNS waveform with a starting frequency of $10$
Hz. The surrogate has maximum errors of $3.8$ in amplitude ($0.04$ excluding
the last 100M before merger) and $0.043$ radians in phase. Following a
different trend, Ref. [80] implemented Gaussian process regression to build a
frequency-domain surrogate version for an aligned-spin BNS waveform model
using the EOB formalism. The resulting surrogate has a maximum mismatch of
$4.5\times 10^{-4}$ and a speedup $O(10^{3})$ with respect to the original
model.
As an alternative to a greedy-based method for ROM, Singular Value
Decomposition (SVD) can be used to generate a reduced basis for the GWs. Along
this line, Refs. [109] and [110] presented two frequency-domain reduced order
models for EOB models SEOBNRv1 [135] and SEOBNRv2 [134], respectively. These
surrogates are built upon an SVD-based method to construct reduced basis and
implement tensor product splines as interpolation method. The surrogate for
SEOBNRv2 is a spin-aligned model for the GW dominant $(2,2)$ mode and extends
the spin range of the first surrogate to almost the entire Kerr range. It also
covers the entire parameter space in which the first one is defined: symmetric
mass ratios $0.01\leq\eta\leq 0.25$ and spin magnitudes $-1\leq\chi_{i}\leq
0.99$. In general, the mistmatches are better than $\sim 0.1\%$ against
SEOBNRv2 except in regions of parameter space in which the original model
presents discontinuities, inducing mistmatches $\sim 1\%$ in the surrogate.
#### 12.5 Post-Newtonian (PN)
The first contact between the Reduced Basis method with gravitational waveform
modeling occurred in [53]. The authors built a reduced basis for 2PN waveforms
corresponding to a 2-D space of non-spinning BNS inspirals with mass
components in the range $[1,3]M_{\odot}$. They found that, remarkably, to
machine precision error, only $921$ basis elements are needed to represent the
full template bank used as fiducial model. Also the greedy-approach here used
was compared against a metric template placement method, finding exponential
decay of the greedy error with the number of bases as opposed to the
approximately linear convergence rate of the metric approach.
Later on, in [55] a reduced basis was built for the non-precessing case of BBH
inspirals (4-D parameter space: $2$ masses, $2$ aligned or anti-aligned spins)
using the restricted TaylorF2 PN approximation with component masses in the
range $[3,30]M_{\odot}$ and dimensionless spin magnitudes in the full range
$[-1,1]$. The authors found that, for a tolerance error of $10^{-11}$, when
increasing the dimensionality of the parameter space from 2-D to 4-D the
number of basis elements needed to span the whole space of waveforms increased
only in $6.6\%$ with respect to the $1,725$ bases needed for the 2-D case.
Furthermore, going from the 3-D to the 4-D case implied adding only $15$ more
basis elements. This opened the possibility that the curse of dimensionality
could be beaten for the complete 8/7-D case, as discussed in Section 15.
In [28] the problem of building a reduced basis for the full 7-D case, where
there are no closed-form expressions and ordinary differential equations need
to be solved, was tackled. The waveforms correspond to 3.5PN precessing
inspirals (mass ratio $q\in[1,10]$ and dimensionless spin magnitudes
$||\chi_{i}||\leq 0.9$). It used a modified version of the standard greedy
algorithm to randomly resample the 7-D parameter space at each iteration with
a fixed number $K$ of waveforms. This was crucial to overcome the limitations
imposed by the curse of dimensionality against the construction of a densely
populated training space. The sampling number $K$ was increased in each run
until reaching $K=36,000$ for which the number of RB waveforms became
independent of $K$. Choosing to work in the binary’s precessing frame, the
authors exploited the fact that in this frame the waveforms have a weaker
parametric dependence than they do in the inertial one. Another important
ingredient was the choice of waveform parametrization in phase instead of time
or frequency, taking advantage of the smooth dependence of waveforms on this
variable. They find that with all these modifications, only $50$ waveforms are
needed to represent the entire 7-D space with an error of $10^{-7}$.
#### 12.6 Ringdown
With the advent of GW astronomy since the first detection of the GW150914
event in 2015 a new era for testing general relativity and alternative
theories of gravity in strong regimes [74, 24, 151, 114] has opened. In the
case of compact binary coalescences different techniques were developed in
recent years [133, 45, 101] to improve the extraction of information after
merger. As it is well known, the account of post-merger properties is
accomplished through the study of the quasinormal modes (QNMs) of the remnant
Kerr black hole. Different models for ringdown waveforms have been constructed
in rather recent years through several techniques. See for example [87, 88,
86, 85] for recent applications of greedy and regression methods in the
construction of QNM models.
In practice, single-mode ringdown searches can limit the number of measured
events in advanced ground-based detectors such as LIGO. Moreover, parameter
estimation errors can become large for such single-mode searches when the
actual waveform contains a second mode. Besides this, there are further
motivations on multi-mode searches, such as consistency tests of GR (e.g., the
no-hair theorem [22, 50, 23]) and feature inference about the progenitors of
the final black hole. Looking forward for future multi-mode ringdown searches,
the Reduced Basis scheme was implemented in [36] to construct a compact
representation of multi-mode QNM catalogs. A single QNM waveform has the form
$h_{lmn}(t)={\cal
A}_{lmn}(\Omega)\frac{M}{r}e^{-\pi\frac{f_{lmn}}{Q_{lmn}}t}\cos{(2\pi
f_{lmn}t)}\,,$ (126)
where the different symbols are defined as ${\cal A}_{lmn}(\Omega):=$
orientation-dependent dimensionless amplitude; $r:=$ distance to the source;
$M:=$ black hole mass; $f_{lmn}:=$ central frequency; $Q_{lmn}:=$ quality
factor, and geometric units $G=c=1$ are taken for granted. Starting from the
one-mode case, a template bank of ringdown waveforms was constructed for the
$(l,m,n)=(2,2,0)$ and $(3,3,0)$ QNMs. Asking for several minimal matches
$(MM)$ between template and signal the parameter space is filled using a
metric-based population criteria over the 2-D space $(f,Q)$. For example, a
minimal match $MM=0.99$ corresponds to a catalog of $2,213$ waveforms with
spin magnitude of the remnant black hole being in the range $[0,0.9947]$ and
mass $M$ in the range $[2.9744,3025.7]M_{\odot}$. Table 14 shows the number of
RB elements needed for each case.
Figure 14: Taken from [36]. Number of reduced basis waveforms ($N_{RB}$) and
metric-based templates $N_{metric}$ needed to represent one-mode QNM training
spaces with $(l,m,n)=(2,2,0)$ and (3, 3, 0) for different minimal matches MM.
The training space representation error is $\epsilon=10^{-12}$. For $\tt
dim=2$, $N_{metric}$ scales with MM as $N_{metric}\propto(1-MM)^{-1}$.
The RB representation was validated through a Monte Carlo simulation by
randomly sampling the parameter space with $10^{7}$ waveforms. An average
waveform representation error $||h_{\lambda}-{\cal
P}_{N}h_{\lambda}||^{2}\approx 4.21\times 10^{-13}$ was found, one order of
magnitude better than the maximum training space representation error
$\epsilon^{2}=10^{-12}$. For the two-mode case, the authors linearly compose
the previous single modes $(2,2,0)$ and $(3,3,0)$ and find similar results
than in the one-mode case, since the algorithm takes advantage of the
linearity of the QNMs. Following this observation, the authors finally propose
a simple method to represent unconstrained multi-mode waveforms by
approximating it by the sum of individual one-mode projections.
## Part III Data Analysis
### 13 Reduced Order Quadratures
We have so far presented a method to produce a reduced model which can be used
as a representation for an underlying set of functions. This combines the RB
and EIM frameworks to produce accurate representations. Finally, one builds a
surrogate predictive model through interpolation in parameter space at EIM
nodes. Next we discuss how this framework can be used to compute numerical
approximations of integrals (quadratures).
Reduced order quadratures (ROQ) were introduced, at least in the context of
GWs, in [17]. They use the EIM to build a set of nodes and weights to
construct the integral, so the main difference with standard quadratures is
that they are application-specific for parametrized problems. This results in
fast online quadrature evaluations, which are at the core of data analysis
when computing correlations, likelihoods, etc.
Comment 11. Reduced Order Quadratures choose a nearly optimal subset of nodes
from the training points in the physical domain (time, frequency, space) at
which any given basis is known. These training points can be arbitrarily
located: they can be equally spaced, randomly distributed, etc. This is an
important aspect for many practical purposes, such as experimental data, where
one might not be able to dictate when signals are measured, and it is in sharp
contrast with fast converging Gaussian quadratures, which do dictate the time
at which data should be measured. The latter might not only be impractical but
also unfeasible in many experimental scenarios.
Comment 12. Being based on the EIM, the method naturally applies to multiple
dimensions, unstructured data and meshes of arbitrary shapes.
Comment 13. By design, also due to being based on the EIM, very fast
convergence with the number of ROQ nodes is observed, typically exponentially
in the cases of interest. This would not be possible using standard methods
which rely on smoothness of the functions to be integrated. More precisely, in
the case of GWs, in order for Gaussian quadratures to converge fast the signal
would have to be smooth as a function of time, which is never the case for GW
signals due to the presence of measurement noise. In contrast, ROQ lift this
requirement and in practice achieve exponential convergence even for noisy
data.
The idea of ROQ is remarkably simple for its impact:
1. 1.
Step 1. Build a reduced basis for the problem of interest. The basis can be
built, for example, using a POD or greedy approach.
2. 2.
Step 2. Build an empirical interpolant, based on the previous basis, as
discussed in Section 10. This approximates a parametrized function
$f(x,\lambda)$ in the space of interest by an empirical interpolant of the
form
${\cal I}_{n}[f](x,\lambda):=\sum_{i=1}^{n}B_{i}(x)f(X_{i},\lambda)\approx
f(x,\lambda)$ (127)
In GW science $f$ would typically be a waveform, and $x$ time or frequency but
the method is generic.
The affine parameterization in $\lambda$ and physical dimensions $x$ achieved
by (127) is one of the critical ingredients that allows for ROQ. By affine
parameterization one means a decomposition in terms of products of functions
which depend only on parameter and physical variables, as in (127).
3. 3.
Step 3. Reduced order quadratures follow what would otherwise be the standard
procedure when building quadratures, but replacing standard polynomial
interpolants with EIM based ones. Namely,
$I_{n}[f](\lambda):=\int f(x,\lambda)dx\approx\int{\cal
I}_{n}[f](x,\lambda)dx=\sum_{i=0}^{n}w_{i}f(X_{i},\lambda)\,,$ (128)
where the ROQ weights
$\omega_{i}:=\int B_{i}(x)dx$ (129)
are computed offline, with any quadrature method of choice or availability for
computing (approximating) the integrals in Eq. (129).
Comment 14. If the number of available (say, time or frequency) samples is
$L$, then one would typically use them to precompute the weights (129) in the
offline stage. After this offline work, the online evaluation is decreased to
$n$ ROQ nodes, with no practical loss of accuracy and usually $n\ll L$, as we
discuss in Section 14, leading to dramatic speedups in likelihood
computations, among other applications.
#### 13.1 ROQ, other quadrature methods, dependence on dimensionality
Gaussian quadratures (such as Chebyshev and Legendre) are considered some of
the most efficient methods for integrating smooth generic functions. This is
not just a perception: there is a whole theory of why this is generically the
case. Compared to ROQ, though, they do suffer from several disadvantages,
mentioned in the Discussion 5 of Section 9.3:
1. 1.
The location of their nodes, at which the integrand has to be known, is
dictated by the method, which is unrealistic for any experiment or application
based on observations.
2. 2.
They are not hierarchical. That is, more nodes cannot simply be added for
higher accuracy but each new quadrature has to be built from scratch.
3. 3.
Their fast convergence and near-optimality is subject to very specific
conditions: smoothness of the integrand being the most important one after the
imposition of the nodes location. This is an unrealistic assumption for
signals taken from experiments.
4. 4.
As with any polynomial-based approach, their efficiency in general (as when
using grids based on tensor products of one-dimensional ones) decreases with
the dimensionality of the problem, unlike ROQ for cases of interest in this
review.
As we discuss next, ROQ can beat these best known generic methods while
working under more relaxed conditions. This is at the expense of offline work.
Obstacles (1) and (2) are well known, so there is not much need to comment on
them. In Refs. [33, 34] point (3) is discussed and how ROQ can lead to fast
convergence even in the presence of noisy data.
So next we discuss point (4), by considering two simple examples, in one and
two dimensions (both physically and parametrically). They are taken from
reference [17].
Example 11.
The example function family of interest here is (there is nothing particular
about this choice)
$f(x,\lambda)=\left[\left(x-\lambda_{1}\right)^{2}+0.1^{2}\right]^{-1/2}\,,$
(130)
where both $x$ and $\lambda$ are real and we want to compute an approximation
to its integral,
$\displaystyle I(\lambda)=\int_{\Omega}f(x,\lambda)dx\,,\quad$ (131)
with, out of arbitrariness, $\Omega=[-1,1]$ and $\lambda_{1}\in[-0.1,0.1]$. We
use two approaches to approximate (131):
1. 1.
Gaussian quadratures: Numerical integration is carried out using up to $n=150$
nodes and a Gauss-Legendre rule. Recall that for each value of $\lambda$ a
separate quadrature for each value of $n$ has to be performed since the method
is not hierarchical.
2. 2.
Reduced Order Quadratures: A reduced basis (in this case using a greedy
approach) is first built for the family of functions (130), subsequently the
empirical interpolant and nodes and, finally, the ROQ. Since ROQ are
hierarchical the additional cost for showing a convergence test from $n$ to
$(n+1)$ is independent of $n$, only an extra node needs to be added. This,
again, is in contrast to regenerating each quadrature rule as in the Gaussian
case.
The results are shown in Figure (15), as black solid and dashed lines. Even
for an error of at most $10^{-4}$, ROQ provide a factor of $\sim 4$ savings,
with the savings dramatically increasing with higher accuracy. We recall once
again that we are comparing ROQ with one of the best general purpose
quadrature rules.
Figure 15: Error curves for the 1-D and 2-D dimensional cases of Examples
13.1 and 15 from Ref. [17], respectively, using Gauss-Legendre and ROQ rules.
Errors are computed by taking the maximum over the entire training set. The
ROQ savings increase from $\sim 4$ to $\sim 12$ as the number of spatial
dimensions is increased from one to two. Further savings are expected as the
number of spatial dimensions increases. Note that the non-monotonicity of the
ROQ curves is because the EIM does not optimize for accuracy at each step,
this is discussed in Section 10.3.
Approximating functions by products of functions (in particular of
polynomials) in each physical dimension leads to evaluation costs of the
approximants which scale exponentially with the dimentionality. As with higher
accuracy, increasing the dimensionality of the problem leads to ROM providing
larger savings when compared to generic methods. We next give an example.
Example 12.
The family of functions of interest is now $2$-dimensional, both in space and
parameters,
$f(x,\lambda)=\left[\left(x_{1}-\lambda_{1}\right)^{2}+\left(x_{2}-\lambda_{2}\right)^{2}+0.1^{2}\right]^{-1/2}\,,$
(132)
where the physical domain is $x=(x_{1},x_{2})\in\Omega=[-1,1]\times[-1,1]$ and
for the parametric one $\lambda\in[-0.1,0.1]\times[-0.1,0.1]$, and we are
interested in multiple evaluations of the parametrized integral
$I(\lambda)=\int_{\Omega}f(x,\lambda)dx\,.$
When considering Gaussian quadratures the curse of dimensionality already in
two dimensions becomes apparent. In standard multidimensional quadratures the
integrand is approximated by the product of one-dimensional polynomials.
Therefore we consider up to $150^{2}$ Gauss nodes to integrate functions from
the family (132). And, again, because Gaussian quadratures are not
hierarchical, this requires a set of $150^{2}$ quadrature rules for each
quadrature evaluation in a convergence test.
Next, we build a reduced greedy basis, EIM, and ROQ. The convergence of both
methods, Gaussian and ROQ, are displayed as solid and dashed blue lines in
Fig. 15. Instead of a factor $\sim 4$ savings as in the 1-D case for a modest
error of $10^{-4}$, for this same accuracy the savings of ROQ are $\sim 12$.
As we will show in the next section, for realistic problems in GW physics, the
savings are much larger than those of these test problems.
### 14 Accelerating Parameter Estimation with ROQ
Once a detection of a gravitational wave is made one would like to infer the
astrophysical properties of the source which emitted it. The goal here is to
do a full Bayesian analysis to compute the posterior probability distribution
function (PDF) of a set of astrophysical parameters ($\lambda$) which describe
the source. Being able to do so in real time (or nearly so) is particularly
important for rapid followups of electromagnetic counterparts enabling multi-
messenger astronomy, among other motivations. For this reason we place
emphasis on binary neutron stars, or mixed pairs.
Assuming that the detector data $d$ contains the source’s signal
$h(\lambda_{\tt true})$ and stationary Gaussian noise $n$ in an additive way
777Here the noise is denoted by $n$, as is the dimensionality of a reduced
basis throughout this review. Hopefully, from context this should not cause
confusion.,
$d=h(\lambda_{\tt true})+n\,,$
the likelihood of data $d$ corresponding to a parameter $\lambda$ is given by
[75]
$\mathcal{L}(d|\lambda)=A\exp[-\frac{1}{2}(d-h(\lambda),d-h(\lambda)]\,.$
(133)
The evaluations of likelihoods for parameter estimation (PE) is a high
dimensional one. For binary systems, even excluding equations of state in the
presence of matter, it is already a $15$-dimensional problem, with $8$
parameters being intrinsic (mass and spin vector of each binary component) and
$7$ extrinsic ones such as luminosity distance, coalescence time, inclination
of the binary plane, and sky localization.
The bulk of the computational cost in (133) comes from two sources:
1. 1.
Computing the gravitational wave candidates $h(\lambda)$.
2. 2.
Evaluating the inner products in (133) .
Surrogate models can decrease the computational cost of (1), and ROQ those of
(2). In this section we briefly review (2); that is, the use of ROQ and ROM in
general for accelerating parameter estimation (PE). Therefore, this section is
intendently limited in scope, and it is not intended to be a general review of
approaches for accelerating PE, which is a field by itself given its
importance.
Our presentation follows closely that one of Ref. [34].
In Eq. (133) $(a|b)$ is a weighted inner product for discretely sampled data.
More explicitly, and working in the frequency domain, of the form
$(d|h(\lambda))=4\mathbb{R}\ \Delta
f\sum_{i=1}^{L}\frac{\bar{d}(f_{i})h(f_{i},\lambda)}{S_{n}(f_{i})}\,,$ (134)
where $d(f_{i})$ and $h(f_{i},\lambda)$ are the (discrete) Fourier transforms
at frequencies $\\{f_{i}\\}_{i=1}^{L}$, a bar denotes complex conjugation, and
the power spectral density (PSD) $S_{n}(f_{i})$ characterizes the detector’s
noise.
For a given observation time $T=1/\Delta f$ and detection frequency window
$(f_{\tt high}-f_{\tt low})$ there are
$L={\tt int}\left(\left[f_{\tt high}-f_{\tt low}\right]T\right)$ (135)
sampling points in the sum (134). Usually $L$ is large, which determines the
second bulk of the computation when evaluating the likelihood (133).
#### 14.1 Constructing the Reduced Order Quadrature
The ROQ scheme for parameter estimation requires three steps, as described in
Section 13. Here we slightly reformulate them in order to bring up some
practical issues. These steps are:
1. 1.
Construct a reduced basis, i.e. a set $n$ elements whose span reproduces the
GW model with a desired precision.
2. 2.
Using the EIM, construct the empirical interpolant and its nodes.
3. 3.
The ROQ weights (138) are computed, and are used to replace, without loss of
accuracy, inner product evaluations (134) by ROQ compressed ones.
Comment 15.
* •
Step 1. The reduced basis only needs to be built over the space of intrinsic
parameters for the waveform family. Furthermore, if the basis is generated
using a PSD $\equiv 1$ the representation of the waveform family can be used
with any PSD whenever the weights are built as in Eq. ((138)).
In practice, for the purposes of this Section, basis generation proceeds in
two stages. A reduced basis is first constructed, which can be evaluated for
any value of. Next, this basis is evaluated at $L$ equally spaced frequency
samples appropriate for the detector.
* •
Step 2. Given an accurate basis of dimension $n$, it is possible to uniquely
and accurately reconstruct any waveform $h(\lambda)$ from only $n$ subsamples
$\\{h(F_{k},\lambda)\\}_{k=1}^{n}$ using the EIM. The frequency nodes
$\\{F_{k}\\}_{k=1}^{n}$, selected from the full set $\\{f_{i}\\}_{i=1}^{L}$,
are also obtained by the EIM.
This step provides a near-optimal compression strategy in frequency which is
complimentary to the parameter one of Step (1). The model’s empirical
interpolant, valid for all parameters, can be written as (cf. Eq. (19) of Ref.
[54])
${\cal
I}[f](f_{i},\lambda)=\exp(-2\pi\mathrm{i}t_{c}f_{i})\sum_{j=1}^{n}B_{j}(f_{i})h(t_{c}=0,F_{j},\lambda)\approx
h(f_{i},\lambda)\,,$ (136)
where, for the sake of the discussion below, we have temporarily isolated the
coalescence time $t_{c}$ from the other extrinsic parameters.
* •
Step 3. Except for $t_{c}$, no extrinsic parameter affects the construction of
the ROQs. The coalescence time, however, requires special treatment. One can
see this by substituting Eq. (136) into Eq. (134),
$(d|h(t_{c},\lambda))=\sum_{k=1}^{n}\omega_{k}(t_{c})h(t_{c}=0,F_{k},\lambda)\,,$
(137)
with the ROQ weights given by
$\omega_{k}(t_{c})=4\mathbb{R}\Delta
f\sum_{i=1}^{L}\frac{\bar{d}(f_{i})B_{k}(f_{i})}{S_{n}(f_{i})}\exp(-2\pi\mathrm{i}t_{c}f_{i})\,.$
(138)
In practice, the dependence of (138) on $t_{c}$ can be achieved through a
simple domain decomposition, using that an estimate for the time window $W$
centered around the coalescence time $t_{\tt trigger}$ is given by the GW
search pipeline. This suggests a prior interval $[t_{\tt trigger}-W,t_{\tt
trigger}+W]$ to be used for $t_{c}$. This prior interval is then split into
$n_{c}$ equal subintervals of size $\Delta t_{c}$. The number of subintervals
is chosen so that the discretization error is below the measurement
uncertainty on the coalescence time. Finally, on each subinterval a unique set
of ROQ weights is constructed.
Step (3) is currently implemented in the LALInference pipeline as summarized
in Algorithm (3). The offline steps (1) and (2) are carried out independently.
By construction, the approach guarantees that these offline steps need be to
carried out only once for each waveform family model.
Algorithm 3 Computing the ROQ weights for parameter estimation
1:Input: $d,S_{n},\\{B_{j}\\}_{j=1}^{N},\Delta f,t_{\tt trigger},W,\Delta
t_{c}$.
2:Set $n_{c}={\tt int}\left(\left(2W\right)/\Delta t_{c}\right)+1$
3:for $j=1\to n_{c}$ do
4: $T_{j}=t_{\tt trigger}-W+\left(j-1\right)\Delta t_{c}$
5: for $k=1\to N$ do
6: Compute $\omega_{k}(T_{j})$ via Eq. (138)
7: end for
8:end for
9:Output:
$\\{T_{j}\\}_{j=1}^{n_{c}},\\{\\{\omega_{k}(T_{j})\\}_{k=1}^{N}\\}_{j=1}^{n_{c}}$.
##### Compressed Norm Evaluations
There are two more terms to consider for fast computations of inner products,
which can be seen from Eq. (133). One of them is $(h(\lambda)|h(\lambda))$.
Unless the GW model is a closed-form expression, one needs to build a fast
online evaluation for this norm for values of $\lambda$ that are only known at
run time. This can be achieved by constructing a reduced basis for $(h,h)$,
then its empirical interpolant, and finally its ROQ. The other term, $(d|d)$,
only needs to be computed once per parameter estimation or search analysis, so
it does not require any special treatment to speed up its calculation.
#### Total speedup
Notice that even though the ROQ weights (Alg. 3) are computed in the online
stage, they only depend on the detection-triggered data $d$ and not on any
$h(\lambda)$. Therefore, they can be computed in what can be referred to as
the startup stage, which requires $n$ full (of size $L$) inner product (134)
evaluations for each $t_{c}$ interval. As discussed below, in practice this
cost is negligible, while each likelihood is subsequently calculated millions
of times, leading to significant speedups in parameter estimation studies,
resulting in observed speedups in the whole PE study equal to $L/n$, which is
the reduction in the number of terms needed to compute (137) instead of (134).
This assumes that $n<L$ but, as we see below, this is indeed the case in
problems of interest (furthermore, usually $n\ll L$).
This speedup comes from operation counts, but it has also been observed in
practice in actual implementations in the LIGO Algorithm Library (LAL)
pipeline [7], as discussed below. For BNS, for the early advanced detectors’
configuration [83] ROQ showed to provide a factor of $\sim 30\times$ speedup
in PE for low-frequency sensitivity of $40$Hz, and $\sim 70\times$ and $\sim
150\times$ as the sensitivity band is lowered to $20$Hz and $10$Hz,
respectively; in all cases without practical loss of accuracy or systematic
biases.
Example 13. The material for this example is taken from Ref. [33].
The majority of a binary neutron star’s GW signal can be expected to be in the
inspiral regime [121], which can be described by the closed-form TaylorF2
approximation [32]. While TaylorF2 does not incorporate spins or the merger-
ringdown phases of the binary’s evolution, these might not be important for
BNS parameter estimation and can therefore be neglected, at least in a first
approximation [125]. For a thorough study on this point using the SpinTaylorF2
approximation, see [96].
Even for this simple to evaluate waveform family, inference on a single data
set used to require significant computational wall-time with standard
parameter estimation methods [15].
In Ref. [33] the authors first computed the observation time $T$ required to
contain a typical BNS signal. Next, a reduced basis of dimensionality $n$
needed to represent this model for any pair of BNS masses was constructed. The
upper frequency $f_{\tt high}$ was fixed to $1024$Hz while $f_{\tt low}$
varied between $10$Hz and $40$Hz.
The time taken for a BNS system with an initial GW frequency of $f_{\tt low}$
to inspiral to $1024$Hz,
$\displaystyle T_{\tt BNS}=\left[6.32+2.07\times\frac{10^{6}}{\left(f_{\tt
low}/\text{Hz}\right)^{3}+5.86\left(f_{\tt
low}/\text{Hz}\right)^{2}}\right]{\text{s}}\,,$ (139)
was empirically found by generating a $\left(1+1\right)\,M_{\odot}$ waveform
(directly given in the frequency domain) and Fourier transforming it to the
time domain where the duration up to when the waveform’s evolution terminates
is measured. Equation (139) and subsequent fits were found using a genetic
algorithm-based symbolic regression software, Eureqa [10, 122, 123]. The
length $L$, as implied by Eq. (135), is plotted in the top panel of Fig. 16.
Figure 16: Top: Length $L$ (red dots) of a typical binary neutron star
inspiral waveform, with the solid black curve connecting this data implied by
the fit (139). Middle: Number of reduced basis waveforms (red crosses), with
the solid black curve given by the fit (140). Bottom: Speedup implied by
operation counts, as given by equation (141). Figure from [33], where what
here is denoted as $n$ (the number of basis) in that reference is $N$ (here
used for the size of the initial training set).
As discussed, each basis only needs to be constructed over the space of
intrinsic parameters – in this case the two-dimensional space of component
masses, chosen to be in the range $\left[1,4\right]M_{\odot}$. This range is
wider than expected for neutron stars, but ensures that the resulting PDFs do
not have sharp cut-offs [93]. The number of reduced basis required to
represent the TaylorF2 model within this range with a representation error
around double precision ($\sim 10^{-14}$) can be fit by
$\displaystyle n_{\tt BNS}=3.12\times 10^{5}\left(f_{\tt
low}/\text{Hz}\right)^{-1.543}\,,$ (140)
and is shown in the middle panel of Fig. 16. It was found that increasing the
high-frequency cutoff to $4096$ Hz only adds a handful of basis elements,
while $L$ changes by a factor of $4$, indicating that the speedup for an
inspiral-merger-ringdown model might be higher, especially given that not many
EIM nodes are needed for the merger and ringdown regimes [54]. This was indeed
shown in [128], as discussed below.
Recalling equation (135), the expected speedup from standard to ROQ-compressed
likelihood evaluations is given by
$\displaystyle\frac{L}{n}\approx\left(1024\text{Hz}-f_{\tt
low}\right)\frac{T_{\tt BNS}}{n_{\tt BNS}}\,,$ (141)
with $T_{\tt BNS}$ and $n_{\tt BNS}$ given by Eqs. (139) and (140),
respectively. As reported in [128], this speedup is indeed observed using
LALInference, and is shown in Fig. 16 (bottom), with a reduction in
computational cost and time of $\sim 30$ for the initial detectors (with a
cutoff of $f_{\tt low}=40\,$Hz) and $\sim 150$ once the advanced detectors
reach $f_{\tt low}\sim 10\,$Hz.
#### 14.2 Implementation in LALInference
Reduced order quadratures for compressed likelihood evaluations and Algorithm
(3) were originally implemented in the LAL parameter estimation pipeline,
known as LALInference [7, 15], by Vivien Raymond and Rory Smith. The resulting
variation is called LALInference$\\_$ROQ. This section provides unpublished
details which were not included in [33], courtesy also of VR and RS. Below is
a comparison between MCMC parameter estimation results using the standard
version of LALInference and ROQ accelerated studies using LALInference$\\_$ROQ
for the previous example, where TaylorF2 is the waveform model. Synthetic
signals embedded in simulated Gaussian noise were injected into the LAL
pipeline, for settings anticipating at the time the initial configuration of
aLIGO, using the zero detuned high power PSD [82] and $f_{\tt low}=40$Hz.
The time window was taken to be $W=0.1$s about the coalescence time $t_{c}$ of
a binary neutron star signal [126, 15]. Following the discussed procedure,
LALInference$\\_$ROQ discretizes this prior into $n_{c}=2,000$ sub-intervals,
each of size $\Delta t_{c}=10^{-5}$s, for which it constructs a unique set of
ROQ weights on each sub-interval. A width of $10^{-5}$s ensures that this
discretization error is below the measurement uncertainty on the coalescence
time, which is typically $\sim 10^{-3}$s [15].
As expected, the ROQ and standard likelihood approaches produce statistically
indistinguishable results for posterior probability density functions over the
full $9$-dimensional parameter space ($2$ intrinsic dimensiones and the full
$7$ extrinsic ones). Figure 17 and Table 2 describe results for the nine
parameters obtained in one particular MCMC simulation; other simulations were
qualitatively similar.
| $\mathcal{M}_{c}\,(M_{\odot})$ | $\eta$ | $D\,$(Mpc) | $t_{c}$(s) | $\alpha$ (rad) | $\delta$ (rad) | $\phi_{c}$ (rad) | $\psi$ (rad) | $\iota$ (rad) | SNR
---|---|---|---|---|---|---|---|---|---|---
injection | 1.2188 | 0.25 | 172 | 0 | 4.91 | -0.981 | 1.46 | 0 | 1.89 | 11.4
standard | $1.2188^{1.2189}_{1.2184}$ | $0.249^{0.250}_{0.243}$ | $153^{107}_{334}$ | $0^{0}_{0}$ | $4.89_{1.79}^{4.98}$ | $-0.978_{-1.01}^{1.20}$ | $3.35_{0.309}^{6.07}$ | $1.58_{1.54\times 10^{-2}}^{3.13}$ | $1.82_{1.32}^{2.91}$ | 12.9
ROQ | $1.2188^{1.2189}_{1.2184}$ | $0.249^{0.250}_{0.243}$ | $149_{100}^{352}$ | $0_{0}^{0}$ | $4.89_{1.82}^{4.98}$ | $-0.978_{-1.01}^{1.19}$ | $3.01_{0.543}^{5.97}$ | $1.58_{2.85\times 10^{-2}}^{3.11}$ | $1.79_{1.39}^{2.82}$ | 12.9
Table 2: Chirp mass $\mathcal{M}_{c}$, symmetric mass ratio $\eta$, source
distance (from Earth) $D$, coalescence time $t_{c}$, right ascension $\alpha$,
declination $\delta$, coalescence phase $\phi_{c}$, polarization phase $\psi$,
inclination $\iota$ and Signal-to-Noise Ratio (SNR) of the analysis from
Figure 17. Median value and 90% credible intervals are provided for both the
standard likelihood (second line) and the ROQ compressed likelihood (third
line). The SNR is empirically measured from
$\mathrm{Likelihood_{max}}\approx\mathrm{SNR}^{2}/2$. The differences between
the two methods are dominated by statistics from computing intervals with a
finite number of samples. Credit: Vivien Raymond
It is also useful to quantify the fractional difference in the 9-D likelihood
function computed using ROQ and the standard approach. This fractional error
has found to be
$\Delta\log\,\mathcal{L}=1-\left(\frac{\log\,\mathcal{L}}{\log\,\mathcal{L}_{\tt{ROQ}}}\right)\lesssim
10^{-6}$
in all cases. That is, both approaches are indistinguishable for all practical
purposes.
In addition to providing indistinguishable results, ROQ accelerated inference
is significantly faster: The ROQ-based MCMC study with the discussed settings
takes $\sim 1$ hour, compared to $\sim 30$ hours using the standard likelihood
approach, in remarkable agreement with the expected savings based on operation
counts. The wall-time of the analysis is proportional to the total number of
posterior samples of the MCMC simulation, which in this case was $\sim
10^{7}$. The startup stage required to build the ROQ weights has negligible
cost and is completed in near real-time, $\sim 30$s, which is equivalent to
$\sim 0.028\%$ of the total cost of a standard likelihood parameter estimation
study.
Figure 17: Probability density function for the chirp mass $\mathcal{M}_{c}$,
symmetric mass ratio $\eta$, inclination $\iota$, coalescence phase
$\phi_{c}$, right ascension $\alpha$, declination $\delta$, polarization phase
$\psi$, source distance (from Earth) $D$ and coalescence time $t_{c}$ of a
simulated event in LIGO/Virgo data. In green as obtained in $\sim 30$ hours by
the standard likelihood, and in blue as obtained in $1$ hour with the ROQ. The
injection values are in red, and are listed in Table 2. The overlap region of
the sets of PDFs is the hatched region. Credit: Vivien Raymond.
For a lower cutoff frequency of $f_{\tt low}\sim 20$Hz, the speedup reduction
is from a couple of weeks to hours. For advanced detectors with $f_{\tt
low}\sim 10$Hz, the longest BNS signals last around $2048$s in duration.
Assuming a fiducial high frequency cut-off of $1024\,$Hz, which is approaching
the upper limit of the sensitivity of aLIGO/AdV, datasets can be as large as
$L\sim 1024\text{Hz}^{-1}\times 2048\text{s}\sim 10^{6}$. Assuming that the
advanced detectors require at least $\sim 10^{7}$ posterior samples, this
implies runtimes upwards of $\sim 100$ days and one Petabyte worth of model
evaluations using the standard approach. On the other hand, ROQ reduces this
to hours. Remarkably, this approach when applied to the advanced detectors
operating at design sensitivity is faster than even the standard likelihood
one used for the initial detectors. Additionally, with parallelization of the
sum in each likelihood evaluation essentially real-time full Bayesian analysis
might be achieved. More details can be found on the website [8].
#### Further reading
Even though ROQs decrease the computational cost of likelihood evaluations for
binary neutron stars for advanced detectors by about two orders of magnitude,
this is still in the order of hours, which does not meet the need for real-
time followup of GW detections through electromagnetic counterparts.
Therefore, other ideas, beyond or on top of ROQ are needed. One approach,
named Focused Reduced Order Quadratures (FROQ) [97], is to restrict the
parameter estimation search based on trigger values from the detection
pipeline, which further reduces the cost to around 10 minutes, which is a good
target from an astrophysical point of view for searches of electromagnetic
counterparts. Even though the study of [97] uses Post-Newtonian closed-form
approximations, if NR surrogates are used for other scenarios, waveform
evaluation cost should not be an issue for extending FROQ.
In Reference [128] the authors built the first ROQ for precessing inspiral-
merger-ringdown compact binaries, using the IMRPhenomPv2 model, which includes
all $7$ intrinsic parameters. For low cutoff frequencies of $f_{\tt
low}=20$Hz, the authors found speedups of up to between $\sim 4\times$ and
$\sim 300\times$ (for short, BBH, and long, BNS systems, respectively),
leading to an estimate of $6-12$ hours for a full PE study, as opposed to
$\sim 1$day instead of $\sim 6$ months. The resulting code is also available
from LALInference. This study was extended in [94] adding a parametrization
for deviations from General Relativity using IMRPhenomPv2.
We mentioned that one of the computational burdens of PE is the computation of
the gravitational waveforms and that this can be solved using surrogate
models. This speedup was analyzed in detail using SVD interpolated waveforms
in [126].
Even though in this Section we have focused on the use of ROQ for parameter
estimation, the procedure for GW searches is similar.
As a proof of the generality of ROQ, it was used in Ref. [31] in the context
of studies of laser light scattering with a speedup of $\sim 2,750$ with
essentially no loss in precision.
As emphasized in Section 15, an important topic that we have not touched in
this review is that one of statistical machine learning approaches (as opposed
to algorithmic ones as in this review) for GWs prediction and analysis. An
exception in order here is that one of [41], since it bridges ROM with
Artificial Neural Networks (ANN) and addresses an important problem. Namely,
the challenge of interpolating the projection coefficients in a Reduced Basis
approach. Instead of algorithmically interpolating them, the authors use ANN
to map the relationship from parameters to projection coefficients. Though the
focus of the reference is on parameter estimation of GWs, the fundamental idea
is applicable to the predictive step itself; in fact it was proposed around
the same time in a different context in [147].
### 15 Challenges, open issues and comments
To discuss one of the main challenges left in reduced order and surrogate
models for gravitational waves we review the case of inspiral spinning non-
precessing binary PN black hole waveforms, taken from Ref. [55]. The parameter
space is 4-D and the sector covered corresponds to $m_{i}\in[3-30]M_{\odot}$,
with $\chi_{i}\in[-1,1]$ ($i=1,2$). The authors study, in the context of
Reduced Basis, the rb for parameter dimensions ${\tt dim}=2,3,4$, which
correspond to the non-spinning, spinning-aligned, and spinning non-precessing
cases, respectively. In Figure 18 there is a histogram corresponding to the
4-D case showing the greedy parameters count for mass ratio
$m_{1}/m_{2}\in[0.1,10]$ for a representation error of $\sim 10^{-14}$ of any
waveform in the considered parameter range. In Figure 19, for visual
representation, we show the distribution of the greedy points in parameter
space for the 3-D case where the spins are aligned or anti-aligned, with
$\chi_{1}=\chi_{2}=\chi$. The lower cutoff frequency being considered at the
time was $40$Hz.
Figure 18: The greedy parameter choices for the mass ratio
$m_{1}/m_{2}\in[0.1,10]$. Notice that the greedy points cluster in the near-
equal mass sector. Figure taken from [55]. Figure 19: The greedy parameter
choices for the mass ratio $m_{1}/m_{2}$ in the 4-D case. Right: the greedy
parameter choices for the 3-D $(m_{1},m_{2},\chi)$ for the aligned spin case.
Figure taken from [55]. Figure 20: The reduced basis representation error as a
function of the dimensionality of the reduced bases for ${\tt dim}=2,3,4$. The
number of reduced basis elements barely grows with the dimensionality of the
problem $\tt dim$, suggesting that at least in the context of GWs from compact
binaries, the curse of dimensionality might be beaten by the reduced basis
approach. Figure taken from [55].
Comment 16.
1. 1.
From Fig. 18 it can be seen that the largest number of selected greedy points
is near the equal mass case. The reason for this is that since a fixed
frequency range is considered, the longer waveforms correspond to nearly equal
mass components, have a larger number of cycles and therefore more structure
and representative information.
2. 2.
In the non-spinning case around $1,700$ basis waveforms are needed for machine
precision representation, see Fig. 20. This is a rather large number of data
points to fit each waveform phase and amplitude at the EIM nodes for the
considered mass range when building a surrogate as proposed in Section 11.2.
Therefore, a low order local polynomial or spline interpolation would probably
be accurate enough for practical purposes when the fiducial model is fast
enough to evaluate – as it is in this example – while avoiding Runge’s
phenomenon (discussed in Section 9).
3. 3.
The reason why as increasing the dimensionality of the problem the reduced
basis approach seems to beat the curse of dimensionality can be seen from the
right panel of Fig. 19. Namely, the data becomes extremely sparse, and the
number of basis elements remains approximately constant while the volume of
the parameter space grows exponentially with the number of dimensions (Figure
20). This sparseness is a very much desired condition indeed for any ROM
approach, but it introduces difficult challenges, as discussed below.
In the case of projection and when the waveforms are known in closed form, or
can be computed in a relatively inexpensive way, there are no major issues: in
order to evaluate the representation of the waveform one simply computes the
projection coefficients onto the basis. Similarly for predictive models, since
in those cases there are an “arbitrary” number of local waveforms to use for
fitting in Step 3 of an EIM-based interpolant (Section 11.2): a local fit of
low order (be it in the form of interpolation, least squares, splines, etc.)
can be used with high accuracy and while avoiding Runge’s phenomena.
A very different situation is that one in which the training set is
computationally expensive to build, as when it involves solving the full
Einstein equations through numerical relativity and only a sparse training set
can be constructed and there are not enough points nearby in parameter space
to perform local fits with enough accuracy.
One solution is to keep running NR simulations to enrich the training set.
This approach might be affordable, with enough computational resources, for
binary black holes. For binary neutron stars or mixed pairs, on the other
hand, where the parameter dimensionality grows considerably, this might be
impractical if not just impossible in reality.
This challenge follows directly from the agnostic, data-driven, purposely
design of the surrogate approach discussed in this review, in which no
differential equations are invoked. The motivation for this design is that it
is a major effort to build a production Einstein solver for realistic problems
of interest, and an intrusive approach would require major algorithmic and
software changes and as a consequence the NR community might be reluctant to
it. Yet, that might be the only feasible approach in the long run. An
intermediate solution might be doing further research in state of the art of
global approaches to approximations in multiple dimensions which are fast to
evaluate and do not suffer from Runge’s phenomena, be it in the form of
deterministic algorithms or statistical machine learning/artificial
intelligence type.
In this review we have covered mostly algorithms of the first kind and
tangentially mentioned some results of the second kind. But there is a lot of
ongoing activity and interesting results on the ML/AI side, though it appears
too premature to cover them here, so we will defer their discussion to a
future, updated version of this Living Review. For a recent preliminary
comparative study, though, on ML/AI surrogates, see [124]. In terms of
parameter estimation and within reduced order modeling, the use of Focused
Reduced Order Quadratures, discussed at the end of Section 14, might be a good
enough solution for near real-time or fast-enough followups of electromagnetic
counterparts.
A related active field but not in GW science is that one of Uncertainty
Quantification (UQ) [127] and, in particular, generalized polynomial chaos
[149], which we have not discussed in this review but are closely related to
Reduced Basis and can remove the usual assumption (and its consequences) of
the noise being Gaussian and stationary noise, which is not the case in GW
science. The application of these areas of research in GW science are unknown
to us, so we also defer them to a further version of this review. However, it
has to be highlighted that other techniques to deal with non-Gaussian noise
have found new detections from public LIGO data [150], so the importance of
the topic cannot be overemphasized.
### 16 Acknowledgments
We thank, for valuable discussions about topics covered in this review through
collaborations and/or interactions throughout the years: Harbir Antil, Peter
Binev, Jonathan Blackman, Priscilla Canizares, Sarah Caudill, Wojciech Czaja,
Albert Cohen, Scott Field, Jonathan Gair, Chad Galley, Chad Hanna, Frank
Herrmann, Jan Hesthaven, Stephen Lau, Jason Kaye, David Knezevic, Tom Loredo,
Akil Narayan, Ricardo Nochetto, Evan Ochsner, Dianne O’Leary, Michael Pürrer,
Vivien Raymond, Gianluigi Rozza, Rory Smith, Benjamin Stamm, Béla Szilágyi,
Eitan Tadmor, Michele Vallisneri, and Alan Weinstein. We apologize in advance
for any omissions.
We especially thank Scott Field, Chad Galley and Rory Smith for initial
contributions to an initial draft of this review years ago, which we have
attempted to write from scratch, given the results over the last years in the
field, but some imprints from the original conception might still be left.
Special credits are due to Frank Herrmann, who played a key role in the
foundational efforts of reduced basis in gravitational wave science since
2009, bringing it from exploratory analyses and initial proofs of concept to a
production-level stage. Special credits are also due to Jan S. Hesthaven, who
introduced one of us (MT) to the idea of using the Reduced Basis framework for
GW science back in 2009 at Brown University, and Ben Stamm for a short but
intensive course at UMD that he gave us to get started.
We also thank Jorge Pullin, the Horace Hearne Institute and the Center for
Computation and Technology at LSU, for hospitality while part of this review
was written. This work was supported partially by CONICET and EVC-CIN
(Argentina).
### References
* [1] Tor O. Aanonsen, Empirical Interpolation with Application to Reduced Basis Approximations, PhD Thesis, Norwegian University of Science and Technology, 2009,http://hdl.handle.net/11250/258487.
* [2] J. Kaye, The interpolation of gravitational waveforms, Thesis, Brown University, 2012, www.dam.brown.edu/scicomp/reports/2013-8/.
* [3] “Arby: A Python Package for reduced order modeling”. URL (cited on 26 January 2021):
https://github.com/aaronuv/arby.
* [4] “Binary black-hole surrogate waveform catalog”. URL (cited on 18 January 2021):
https://zenodo.org/record/2669459#.XNQ6VC2ZPjA.
* [5] “gwsurrogate·PyPI”. URL (cited on 18 January 2021):
https://pypi.org/project/gwsurrogate/.
* [6] “ICERM sp-s20 program”URL:
https://icerm.brown.edu/programs/sp-s20/.
* [7] “LAL”. URL (cited on 18 January 2021):
https://www.lsc-group.phys.uwm.edu/daswg/
projects/lalsuite.html.
* [8] “LAL-ROQ”. URL (cited on 18 January 2021):
http://www.ligo.caltech.edu/~vraymond.
* [9] “Model Reduction for Parametrized Systems”. URL (cited on 18 January 2021):
https://www.morepas.org.
* [10] “Nutonian—DataRobot”. URL (cited on 18 January 2021):
https://www.datarobot.com/nutonian/.
* [11] “Reduced Order Modeling in General Relativity”. URL (cited on 18 January 2021):
http://www.tapir.caltech.edu/~rom-gr.
* [12] “RIT Waveform Catalog”. URL (cited on 18 January 2021):
https://ccrg.rit.edu/content/data/rit-waveform-catalog.
* [13] “Spectral Einstein Code”. URL (cited on 18 January 2021):
https://www.black-holes.org/code/SpEC.html.
* [14] “surfinBH 1.1.1”. URL (cited on 18 January 2021):
https://pypi.org/project/surfinBH/.
* [15] Aasi, J., et al. (LIGO Collaboration, Virgo Collaboration), “Parameter estimation for compact binary coalescence signals with the first generation gravitational-wave detector network”, Phys.Rev., D88, 062001, (2013). [DOI], [arXiv:1304.1775 [gr-qc]].
* [16] Antil, H., Chen, D., and Field, S., “A Note on QR-Based Model Reduction: Algorithm, Software, and Gravitational Wave Applications”, Computing in Science Engineering, 20(4), 10–25, (2018). [DOI].
* [17] Antil, H., Field, Scott E., Herrmann, F., Nochetto, Ricardo H., and Tiglio, M., “Two-Step Greedy Algorithm for Reduced Order Quadratures”, Journal of Scientific Computing, 57(3), 604–637, (2013). [DOI].
* [18] Ballico, E., Elia, M., and Sala, M., “On the evaluation of multivariate polynomials over finite fields”, Journal of Symbolic Computation, 50, 255 – 262, (2013). [DOI].
* [19] Barkett, Kevin, Scheel, Mark A., Haas, Roland, Ott, Christian D., Bernuzzi, Sebastiano, Brown, Duncan A., Szilágyi, Béla, Kaplan, Jeffrey D., Lippuner, Jonas, Muhlberger, Curran D., Foucart, Francois, and Duez, Matthew D., “Gravitational waveforms for neutron star binaries from binary black hole simulations”, Phys. Rev. D, 93, 044064, (Feb 2016). [DOI].
* [20] Barrault, Maxime, Maday, Yvon, Nguyen, Ngoc Cuong, and Patera, Anthony T., “An ‘empirical interpolation’ method: application to efficient reduced-basis discretization of partial differential equations”, Comptes Rendus Mathematique, 339(9), 667 – 672, (2004). [DOI].
* [21] Barta, Dániel, and Vasúth, Mátyás, “Fast prediction and evaluation of eccentric inspirals using reduced-order models”, Phys. Rev. D, 97(12), 124011, (2018). [DOI].
* [22] Berti, Emanuele, Cardoso, Jaime, Cardoso, Vitor, and Cavaglia, Marco, “Matched-filtering and parameter estimation of ringdown waveforms”, Phys. Rev., D76, 104044, (2007). [DOI].
* [23] Berti, Emanuele, Cardoso, Vitor, and Will, Clifford M., “On gravitational-wave spectroscopy of massive black holes with the space interferometer LISA”, Phys. Rev., D73, 064030, (2006). [DOI].
* [24] Berti, Emanuele, Yagi, Kent, and Yunes, Nicolás, “Extreme Gravity Tests with Gravitational Waves from Compact Binary Coalescences: (I) Inspiral-Merger”, Gen. Rel. Grav., 50(4), 46, (2018). [DOI].
* [25] Blackman, Jonathan, Field, Scott E., Galley, Chad R., Szilágyi, Béla, Scheel, Mark A., Tiglio, Manuel, and Hemberger, Daniel A., “Fast and Accurate Prediction of Numerical Relativity Waveforms from Binary Black Hole Coalescences Using Surrogate Models”, Phys. Rev. Lett., 115, 121102, (Sep 2015). [DOI].
* [26] Blackman, Jonathan, Field, Scott E., Scheel, Mark A., Galley, Chad R., Hemberger, Daniel A., Schmidt, Patricia, and Smith, Rory, “A Surrogate model of gravitational waveforms from numerical relativity simulations of precessing binary black hole mergers”, Phys. Rev. D, 95, 104023, (May 2017). [DOI].
* [27] Blackman, Jonathan, Field, Scott E., Scheel, Mark A., Galley, Chad R., Ott, Christian D., Boyle, Michael, Kidder, Lawrence E., Pfeiffer, Harald P., and Szilágyi, Béla, “Numerical relativity waveform surrogate model for generically precessing binary black hole mergers”, Phys. Rev. D, 96, 024058, (Jul 2017). [DOI].
* [28] Blackman, Jonathan, Szilagyi, Bela, Galley, Chad R., and Tiglio, Manuel, “Sparse Representations of Gravitational Waves from Precessing Compact Binaries”, Phys. Rev. Lett., 113(2), 021101, (2014). [DOI].
* [29] Boyd, J.P., Chebyshev and Fourier Spectral Methods: Second Revised Edition, Dover Books on Mathematics, (Dover Publications, 2001).
* [30] Boyle, Michael, Hemberger, Daniel, Iozzo, Dante A B, Lovelace, Geoffrey, Ossokine, Serguei, Pfeiffer, Harald P, Scheel, Mark A, Stein, Leo C, Woodford, Charles J, Zimmerman, Aaron B, Afshari, Nousha, Barkett, Kevin, Blackman, Jonathan, Chatziioannou, Katerina, Chu, Tony, Demos, Nicholas, Deppe, Nils, Field, Scott E, Fischer, Nils L, Foley, Evan, Fong, Heather, Garcia, Alyssa, Giesler, Matthew, Hebert, Francois, Hinder, Ian, Katebi, Reza, Khan, Haroon, Kidder, Lawrence E, Kumar, Prayush, Kuper, Kevin, Lim, Halston, Okounkova, Maria, Ramirez, Teresita, Rodriguez, Samuel, Rüter, Hannes R, Schmidt, Patricia, Szilagyi, Bela, Teukolsky, Saul A, Varma, Vijay, and Walker, Marissa, “The SXS collaboration catalog of binary black hole simulations”, Classical and Quantum Gravity, 36(19), 195006, (sep 2019). [DOI].
* [31] Brown, D., Smith, R. J. E., and Freise, A., “Fast simulation of Gaussian-mode scattering for precision interferometry”, Journal of Optics, 18(2), 025604, (jan 2016). [DOI].
* [32] Buonanno, Alessandra, Iyer, Bala, Ochsner, Evan, Pan, Yi, and Sathyaprakash, B.S., “Comparison of post-Newtonian templates for compact binary inspiral signals in gravitational-wave detectors”, Phys.Rev., D80, 084043, (2009). [DOI].
* [33] Canizares, Priscilla, Field, Scott E., Gair, Jonathan, Raymond, Vivien, Smith, Rory, and Tiglio, Manuel, “Accelerated Gravitational Wave Parameter Estimation with Reduced Order Modeling”, Phys. Rev. Lett., 114, 071104, (Feb 2015). [DOI].
* [34] Canizares, Priscilla, Field, Scott E., Gair, Jonathan R., and Tiglio, Manuel, “Gravitational wave parameter estimation with compressed likelihood evaluations”, Phys. Rev., D87, 124005, (2013). [DOI].
* [35] Cannon, K., Chapman, A., Hanna, C., Keppel, D., Searle, Antony C., and Weinstein, Alan J., “Singular value decomposition applied to compact binary coalescence gravitational-wave signals”, Phys. Rev. D, 82, 044025, (Aug 2010). [DOI].
* [36] Caudill, Sarah, Field, Scott E., Galley, Chad R., Herrmann, Frank, and Tiglio, Manuel, “Reduced Basis representations of multi-mode black hole ringdown gravitational waves”, Class. Quant. Grav., 29, 095016, (2012).
* [37] Ceberio, Martine, and Kreinovich, Vladik, “Greedy Algorithms for Optimizing Multivariate Horner Schemes”, ACM SIGSAM Bulletin, 38, (11 2003). [DOI].
* [38] Chan, Tony F, “Rank revealing QR-factorizations”, Linear Algebra and Its Applications, 88, 67–82, (1987). [DOI].
* [39] Chaturantabut, S., and Sorensen, D.C., “Discrete Empirical Interpolation for nonlinear model reduction”, in Decision and Control, 2009 held jointly with the 2009 28th Chinese Control Conference. CDC/CCC 2009. Proceedings of the 48th IEEE Conference on, pp. 4316 –4321, (dec. 2009). [DOI].
* [40] Chaturantabut, Saifon, and Sorensen, Danny, “Nonlinear Model Reduction via Discrete Empirical Interpolation”, SIAM J. Scientific Computing, 32, 2737–2764, (01 2010). [DOI].
* [41] Chua, Alvin J. K., Galley, Chad R., and Vallisneri, Michele, “Reduced-Order Modeling with Artificial Neurons for Gravitational-Wave Inference”, Phys. Rev. Lett., 122, 211101, (May 2019). [DOI].
* [42] Civril, Ali, and Magdon-Ismail, Malik, “Deterministic sparse column based matrix reconstruction via greedy approximation of SVD”, in Algorithms and Computation, pp. 414–423. Springer, (2008). [DOI].
* [43] Cotesta, Roberto, Buonanno, Alessandra, Bohé, Alejandro, Taracchini, Andrea, Hinder, Ian, and Ossokine, Serguei, “Enriching the symphony of gravitational waves from binary black holes by tuning higher harmonics”, Phys. Rev. D, 98, 084028, (Oct 2018). [DOI].
* [44] Csizmadia, Peter, Debreczeni, Gergely, Racz, Istvan, and Vasuth, Matyas, “Gravitational waves from spinning eccentric binaries”, Class. Quant. Grav., 29, 245002, (2012). [DOI].
* [45] Da Silva Costa, C. F., Tiwari, S., Klimenko, S., and Salemi, F., “Detection of (2,2) quasinormal mode from a population of black holes with a constructive summation method”, Phys. Rev. D, 98, 024052, (Jul 2018). [DOI].
* [46] Demmel, James, and Koev, Plamen, “Accurate SVDs of polynomial Vandermonde matrices involving orthonormal polynomials”, Linear Algebra and its Applications, 417(2), 382 – 396, (2006). [DOI]. Special Issue in honor of Friedrich Ludwig Bauer.
* [47] Deshpande, A., and Rademacher, L., “Efficient Volume Sampling for Row/Column Subset Selection”, in Foundations of Computer Science (FOCS), 2010 51st Annual IEEE Symposium on, pp. 329–338, (2010). [DOI].
* [48] DeVore, Ronald, Petrova, Guergana, and Wojtaszczyk, Przemyslaw, “Greedy Algorithms for Reduced Bases in Banach Spaces”, Constructive Approximation, 37(3), 455–466, (2013). [DOI].
* [49] Doctor, Zoheyr, Farr, Ben, Holz, Daniel E., and Pürrer, Michael, “Statistical Gravitational Waveform Models: What to Simulate Next?”, Phys. Rev. D, 96(12), 123011, (2017). [DOI].
* [50] Dreyer, Olaf, Kelly, Bernard J., Krishnan, Badri, Finn, Lee Samuel, Garrison, David, et al., “Black hole spectroscopy: Testing general relativity through gravitational wave observations”, Class. Quantum Grav., 21, 787–804, (2004). [DOI].
* [51] Eftang, Jens L., and Stamm, Benjamin, “Parameter multi-domain empirical interpolation”, International Journal for Numerical Methods in Engineering, 90(4), 412–428, (2012). [DOI].
* [52] Epperson, James F., “On the Runge Example”, Am. Math. Monthly, 94(4), 329–341, (April 1987). [DOI].
* [53] Field, Scott E., Galley, Chad R., Herrmann, Frank, Hesthaven, Jan S., Ochsner, Evan, and Tiglio, Manuel, “Reduced basis catalogs for gravitational wave templates”, Phys. Rev. Lett., 106, 221102, (2011). [DOI].
* [54] Field, Scott E., Galley, Chad R., Hesthaven, Jan S., Kaye, Jason, and Tiglio, Manuel, “Fast Prediction and Evaluation of Gravitational Waveforms Using Surrogate Models”, Phys. Rev. X, 4, 031006, (Jul 2014). [DOI].
* [55] Field, Scott E., Galley, Chad R., and Ochsner, Evan, “Towards beating the curse of dimensionality for gravitational waves using Reduced Basis”, Phys. Rev., D86, 084046, (2012). [DOI].
* [56] Funaro, D., Polynomial Approximation of Differential Equations, number v. 8 in Lecture notes in physics: Monographs, (Springer-Verlag, 1992). [DOI].
* [57] Galley, Chad R., Herrmann, Frank, Silberholz, John, Tiglio, Manuel, and Guerberoff, Gustavo, “Statistical constraints on binary black hole inspiral dynamics”, Class. Quantum Grav., 27, 245007, (2010). [DOI].
* [58] Gasca, Mariano, and Sauer, Thomas, “Polynomial interpolation in several variables”, Advances in Computational Mathematics, 12(4), 377–410, (2000). [DOI].
* [59] Gautschi, Walter, “The condition of Vandermonde-like matrices involving orthogonal polynomials”, Linear Algebra and its Applications, 52-53, 293 – 300, (1983). [DOI].
* [60] Gautschi, Walter, “Optimally scaled and optimally conditioned Vandermonde and Vandermonde-like matrices”, BIT Numerical Mathematics, 51, 103–125, (2011). [DOI].
* [61] Gautschi, Walter, “How (Un) stable Are Vandermonde Systems?”, in Asymptotic and Computational Analysis, (2012). [DOI].
* [62] Gerosa, Davide, Hébert, François, and Stein, Leo C., “Black-hole kicks from numerical-relativity surrogate models”, Phys. Rev. D, 97(10), 104049, (2018). [DOI].
* [63] Giraud, Luc, Langou, Julien, Rozloznik, Miroslav, and Eshof, Jasper van den, “Rounding error analysis of the classical Gram-Schmidt orthogonalization process”, Numerische Mathematik, 101(1), 87–100, (2005). [DOI].
* [64] Guruswami, Venkatesan, and Sinop, Ali Kemal, “Optimal column-based low-rank matrix reconstruction”, in Proceedings of the Twenty-Third Annual ACM-SIAM Symposium on Discrete Algorithms, pp. 1207–1214. SIAM, (2012). [DOI].
* [65] Hannam, Mark, Schmidt, Patricia, Bohé, Alejandro, Haegel, Leïla, Husa, Sascha, Ohme, Frank, Pratten, Geraint, and Pürrer, Michael, “Simple Model of Complete Precessing Black-Hole-Binary Gravitational Waveforms”, Phys. Rev. Lett., 113, 151101, (Oct 2014). [DOI].
* [66] Harbrecht, Helmut, Peters, Michael, and Schneider, Reinhold, “On the low-rank approximation by the pivoted Cholesky decomposition”, Applied Numerical Mathematics, 62(4), 428 – 440, (2012). [DOI]. Third Chilean Workshop on Numerical Analysis of Partial Differential Equations (WONAPDE 2010).
* [67] Healy, James, Lousto, Carlos O., Lange, Jacob, O’Shaughnessy, Richard, Zlochower, Yosef, and Campanelli, Manuela, “Second RIT binary black hole simulations catalog and its application to gravitational waves parameter estimation”, Phys. Rev. D, 100, 024021, (Jul 2019). [DOI].
* [68] Herrmann, Frank, Silberholz, John, Bellone, Matias, Guerberoff, Gustavo, and Tiglio, Manuel, “Integrating Post-Newtonian Equations on Graphics Processing Units”, Class. Quantum Grav., 27, 032001, (2010). [DOI].
* [69] Hesthaven, J., and Ubbiali, S., “Non-intrusive reduced order modeling of nonlinear problems using neural networks”, J. Comput. Phys., 363, 55–78, (2018).
* [70] Hesthaven, Jan S, Rozza, Gianluigi, and Stamm, Benjamin, Certified Reduced Basis Methods for Parametrized Partial Differential Equations, Springer Briefs in Mathematics, (Springer, Switzerland, 2015), 1 edition. [DOI].
* [71] Hesthaven, J.S., Gottlieb, S., and Gottlieb, D., Spectral Methods for Time-Dependent Problems, Cambridge Monographs on Applied and Computational Mathematics, (Cambridge University Press, 2007). [DOI].
* [72] Higham, Nicholas J., “Numerical Conditioning”, (2013).
* [73] Hong, Y. P., and Pan, C.-T., “Rank-Revealing QR Factorizations and the Singular Value Decomposition”, Mathematics of Computation, 58(197), 213–232, (1992). [DOI].
* [74] Isi, Maximiliano, Giesler, Matthew, Farr, Will M., Scheel, Mark A., and Teukolsky, Saul A., “Testing the No-Hair Theorem with GW150914”, Phys. Rev. Lett., 123, 111102, (Sep 2019). [DOI].
* [75] Jaranowski, Piotr, and Krolak, Andrzej, Analysis of Gravitational-Wave Data, Cambridge Monographs on Particle Physics, Nuclear Physics and Cosmology, (Cambridge University Press, 2009). [DOI].
* [76] Jolliffe, I.T., Principal Component Analysis, (Springer, New York, 2002). [DOI].
* [77] Kedlaya, Kiran S., and Umans, Christopher, “Fast Polynomial Factorization and Modular Composition”, SIAM J. Comput., 40, 1767–1802, (2008). [DOI].
* [78] Kuian, Mykhailo, Reichel, Lothar, and Shiyanovskii, Sergij V., “Optimally Conditioned Vandermonde-Like Matrices”, SIAM J. Matrix Analysis Applications, 40, 1399–1424, (2019). [DOI].
* [79] Lackey, Benjamin D., Bernuzzi, Sebastiano, Galley, Chad R., Meidam, Jeroen, and Van Den Broeck, Chris, “Effective-one-body waveforms for binary neutron stars using surrogate models”, Phys. Rev. D, 95(10), 104036, (2017). [DOI].
* [80] Lackey, Benjamin D., Pürrer, Michael, Taracchini, Andrea, and Marsat, Sylvain, “Surrogate model for an aligned-spin effective-one-body waveform model of binary neutron star inspirals using Gaussian process regression”, Phys. Rev. D, 100, 024002, (Jul 2019). [DOI].
* [81] Leiserson, Charles E., Li, Liyun, Maza, Marc Moreno, and Xie, Yuzhen, “Efficient Evaluation of Large Polynomials”, in Fukuda, Komei, Hoeven, Joris van der, Joswig, Michael, and Takayama, Nobuki, eds., Mathematical Software – ICMS 2010, pp. 342–353, (Springer Berlin Heidelberg, Berlin, Heidelberg, 2010). [DOI].
* [82] LIGO Scientific Collaboration, Advanced LIGO anticipated sensitivity curves, (2010). LIGO-T0900288-v3.
* [83] LIGO Scientific Collaboration, Virgo Collaboration, Aasi, J., Abadie, J., Abbott, B. P., Abbott, R., Abbott, T. D., Abernathy, M., Accadia, T., Acernese, F., and et al., “Prospects for Localization of Gravitational Wave Transients by the Advanced LIGO and Advanced Virgo Observatories”, ArXiv e-prints, (April 2013). [ADS], [arXiv:1304.0670 [gr-qc]].
* [84] Lodha, Suresh K., and Goldman, Ron, “A unified approach to evaluation algorithms for multivariate polynomials”, Math. Comp, 1521–1553, (1997).
* [85] London, L., and Fauchon-Jones, E., “On modeling for Kerr black holes: Basis learning, QNM frequencies, and spherical-spheroidal mixing coefficients”, Class. Quant. Grav., 36(23), 235015, (2019). [DOI].
* [86] London, L. T., “Modeling ringdown. II. Aligned-spin binary black holes, implications for data analysis and fundamental theory”, Phys. Rev. D, 102(8), 084052, (2020). [DOI].
* [87] London, Lionel, “On gravitational wave modeling: numerical relativity data analysis, the excitation of Kerr quasinormal modes, and the unsupervised machine learning of waveform morphology”, (2015).
* [88] London, Lionel, Shoemaker, Deirdre, and Healy, James, “Modeling ringdown: Beyond the fundamental quasinormal modes”, Physical Review D, 90(12), (Dec 2014). [DOI].
* [89] Lousto, Carlos O., Nakano, Hiroyuki, Zlochower, Yosef, and Campanelli, Manuela, “Statistical studies of Spinning Black-Hole Binaries”, Phys. Rev., D81, 084023, (2010). [DOI].
* [90] Maday, Y., Nguyen, N. .C, Patera, A. T., and Pau, S. H., “A general multipurpose interpolation procedure: the magic points”, Communications on Pure and Applied Analysis, 8, 383–404, (2009). [DOI].
* [91] Magaril-Il’yaev, Georgii G, Osipenko, Konstantin Yu, and Tikhomirov, Vladimir M, “On Exact Values of n-Widths in a Hilbert Space”, Journal of Approximation Theory, 108(1), 97 – 117, (2001). [DOI].
* [92] Manca, Gian Mario, and Vallisneri, Michele, “Cover art: Issues in the metric-guided and metric-less placement of random and stochastic template banks”, Phys. Rev., D81, 024004, (2010). [DOI].
* [93] Mandel, Ilya, Berry, Christopher Philip Luke, Ohme, Frank, Fairhurst, Stephen, and Farr, Will M., “Parameter estimation on compact binary coalescences with abruptly terminating gravitational waveforms”, Classical and Quantum Gravity, 31(15), (August 7, 2014). [DOI]. Copyright: Copyright 2014 Elsevier B.V., All rights reserved.
* [94] Meidam, Jeroen, Tsang, Ka Wa, Goldstein, Janna, Agathos, Michalis, Ghosh, Archisman, Haster, Carl-Johan, Raymond, Vivien, Samajdar, Anuradha, Schmidt, Patricia, Smith, Rory, Blackburn, Kent, Del Pozzo, Walter, Field, Scott E., Li, Tjonnie, Pürrer, Michael, Van Den Broeck, Chris, Veitch, John, and Vitale, Salvatore, “Parametrized tests of the strong-field dynamics of general relativity using gravitational wave signals from coalescing binary black holes: Fast likelihood calculations and sensitivity of the method”, Phys. Rev. D, 97, 044033, (Feb 2018). [DOI].
* [95] Messenger, C., Prix, R., and Papa, M.A., “Random template banks and relaxed lattice coverings”, Phys. Rev., D79, 104017, (2009). [DOI].
* [96] Miller, B., O’Shaughnessy, R., Littenberg, T. B., and Farr, B., “Rapid gravitational wave parameter estimation with a single spin: Systematic uncertainties in parameter estimation with the SpinTaylorF2 approximation”, Phys. Rev. D, 92, 044056, (Aug 2015). [DOI].
* [97] Morisaki, Soichiro, and Raymond, Vivien, “Prompt and accurate sky localization of gravitational-wave sources”, Journal of Physics: Conference Series, 1468, 012219, (feb 2020). [DOI].
* [98] Moroz, Guillaume, Fast polynomial evaluation and composition, (2013)URL:
https://hal.archives-ouvertes.fr/hal-00846961.
* [99] Narayan, Akil, and Xiu, Dongbin, “Stochastic Collocation Methods on Unstructured Grids in High Dimensions via Interpolation”, SIAM J. Scientific Computing, 34(3), (2012). [DOI].
* [100] Nguyen, Van Kien, “Greedy algorithms and Kolmogorov widths in Banach spaces”, Journal of Approximation Theory, 251, 105344, (2020). [DOI].
* [101] O’Brien, Brendan D., Da Silva Costa, Carlos F., and Klimenko, Sergey, “Measurement of sub-dominant harmonic modes for gravitational wave emission from a population of binary black holes”, (1 2019). [arXiv:1901.09072 [gr-qc]].
* [102] Owen, Benjamin J., “Search templates for gravitational waves from inspiraling binaries: Choice of template spacing”, Phys. Rev., D53, 6749–6761, (1996). [DOI].
* [103] Pan, V. Ya, “Methods of Computing Values of Polynomials”, Russian Mathematical Surveys, 21(1), 105–136, (February 1966). [DOI].
* [104] Pan, Victor Y., “How Bad Are Vandermonde Matrices?”, SIAM J. Matrix Analysis Applications, 37, 676–694, (2015). [DOI].
* [105] Pan, Yi, Buonanno, Alessandra, Taracchini, Andrea, Kidder, Lawrence E., Mroué, Abdul H., Pfeiffer, Harald P., Scheel, Mark A., and Szilágyi, Béla, “Inspiral-merger-ringdown waveforms of spinning, precessing black-hole binaries in the effective-one-body formalism”, Phys. Rev. D, 89, 084006, (Apr 2014). [DOI].
* [106] Peña, J., and Sauer, Tomas, “On the multivariate Horner scheme”, SIAM Journal on Numerical Analysis, 37, (06 1998). [DOI].
* [107] Pinkus, Allan, N-widths in approximation theory, (Springer, Amsterdam, 1985). [DOI].
* [108] Press, W. H., Flannery, B. P., Teukolsky, Saul A., and Vetterling, W. T., Numerical Recipes, (Cambridge University Press, New York, 1992), 2nd edition. [DOI].
* [109] Pürrer, Michael, “Frequency domain reduced order models for gravitational waves from aligned-spin compact binaries”, Class. Quant. Grav., 31(19), 195010, (2014). [DOI].
* [110] Pürrer, Michael, “Frequency domain reduced order model of aligned-spin effective-one-body waveforms with generic mass-ratios and spins”, Phys. Rev. D, 93(6), 064041, (2016). [DOI].
* [111] Quarteroni, A., Manzoni, A., and Negri, F., Reduced Basis Methods for Partial Differential Equations: An Introduction, UNITEXT, (Springer International Publishing, 2015). [DOI].
* [112] Quarteroni, Alfio, Rozza, Gianluigi, and Manzoni, Andrea, “Certified Reduced Basis Approximation for Parametrized Partial Differential Equations and Applications”, Journal of Mathematics in Industry, 1, (01 2011). [DOI].
* [113] Quarteroni, Alfio, Sacco, Riccardo, and Saleri, Fausto, Numerical Mathematics, (Springer, Berlin, 2010). [DOI].
* [114] Radice, David, Bernuzzi, Sebastiano, Del Pozzo, Walter, Roberts, Luke F., and Ott, Christian D., “Probing Extreme-Density Matter with Gravitational Wave Observations of Binary Neutron Star Merger Remnants”, Astrophys. J., 842(2), L10, (2017). [DOI].
* [115] Reichel, Lothar, and Opfer, Gerhard, “Chebyshev-Vandermonde systems”. American Mathematical Society, (1991). [DOI].
* [116] Rifat, Nur E. M., Field, Scott E., Khanna, Gaurav, and Varma, Vijay, “Surrogate model for gravitational wave signals from comparable and large-mass-ratio black hole binaries”, Phys. Rev. D, 101, 081502, (Apr 2020). [DOI].
* [117] Ruhe, Axel, “Numerical aspects of gram-schmidt orthogonalization of vectors”, Linear Algebra and its Applications, 52(0), 591 – 601, (1983). [DOI].
* [118] Runge, C., “Ueber die numerische Auflösung von Differentialgleichungen”, Math. Ann., 46, 167, (1895).
* [119] Salomon, David, and Motta, Giovanni, Handbook of Data Compression, (Springer-Verlag, 2010), 5th edition. [DOI].
* [120] Sarbach, Olivier, and Tiglio, Manuel, “Continuum and Discrete Initial-Boundary Value Problems and Einstein’s Field Equations”, Living Reviews in Relativity, 15(1), 9, (Aug 27, 2012). [DOI].
* [121] Sathyaprakash, B.S., and Schutz, Bernard F., “Physics, Astrophysics and Cosmology with Gravitational Waves”, Living Reviews in Relativity, 12(2), (2009). [DOI].
* [122] Schmidt, Michael, and Lipson, Hod, “Distilling Free-Form Natural Laws from Experimental Data”, Science, 324(5923), 81–85, (2009). [DOI].
* [123] Schmidt, Michael, and Lipson, Hod, “Symbolic Regression of Implicit Equations”, in Riolo, Rick, O’Reilly, Una-May, and McConaghy, Trent, eds., Genetic Programming Theory and Practice VII, pp. 73–85, (Springer US, Boston, MA, 2010). [DOI].
* [124] Setyawati, Yoshinta, Pürrer, Michael, and Ohme, Frank, “Regression methods in waveform modeling: a comparative study”, Classical and Quantum Gravity, 37(7), 075012, (mar 2020). [DOI].
* [125] Singer, Leo P., et al., “The First Two Years of Electromagnetic Follow-Up with Advanced LIGO and Virgo”, Astrophys. J., 795(2), 105, (2014). [DOI].
* [126] Smith, R. J. E., Cannon, K., Hanna, C., Keppel, D., and Mandel, I., “Towards rapid parameter estimation on gravitational waves from compact binaries using interpolated waveforms”, Phys. Rev. D, 87, 122002, (Jun 2013). [DOI].
* [127] Smith, Ralph C., Uncertainty Quantification: Theory, Implementation, and Applications, (Society for Industrial and Applied Mathematics, USA, 2013).
* [128] Smith, Rory, Field, Scott E., Blackburn, Kent, Haster, Carl-Johan, Pürrer, Michael, Raymond, Vivien, and Schmidt, Patricia, “Fast and accurate inference on gravitational waves from precessing compact binaries”, Phys. Rev. D, 94, 044031, (Aug 2016). [DOI].
* [129] Stewart, G.W., Afternotes on Numerical Analysis, Other Titles in Applied Mathematics, (Society for Industrial and Applied Mathematics, 1996). [DOI].
* [130] Stewart, G.W., Afternotes Goes to Graduate School: Lectures on Advanced Numerical Analysis, Other Titles in Applied Mathematics, (Society for Industrial and Applied Mathematics, 1998). [DOI].
* [131] Stewart, G.W., “The Gram-Schmidt Algorithm and Its Variations”, Technical Report from UMIACS, (2006)URL:
http://hdl.handle.net/1903/3036.
* [132] Stoer, J., and Bulirsch, R., Introduction to Numerical Analysis, (Springer-Verlag, Berlin and New York, 1980).
* [133] Talukder, Dipongkar, Bose, Sukanta, Caudill, Sarah, and Baker, Paul T., “Improved coincident and coherent detection statistics for searches for gravitational wave ringdown signals”, Phys. Rev. D, 88, 122002, (Dec 2013). [DOI].
* [134] Taracchini, Andrea, Buonanno, Alessandra, Pan, Yi, Hinderer, Tanja, Boyle, Michael, Hemberger, Daniel A., Kidder, Lawrence E., Lovelace, Geoffrey, Mroué, Abdul H., Pfeiffer, Harald P., Scheel, Mark A., Szilágyi, Béla, Taylor, Nicholas W., and Zenginoglu, Anil, “Effective-one-body model for black-hole binaries with generic mass ratios and spins”, Phys. Rev. D, 89, 061502, (Mar 2014). [DOI].
* [135] Taracchini, Andrea, Pan, Yi, Buonanno, Alessandra, Barausse, Enrico, Boyle, Michael, Chu, Tony, Lovelace, Geoffrey, Pfeiffer, Harald P., and Scheel, Mark A., “Prototype effective-one-body model for nonprecessing spinning inspiral-merger-ringdown waveforms”, Phys. Rev. D, 86, 024011, (Jul 2012). [DOI].
* [136] Taylor, J. M., “The condition of gram matrices and related problems”, Proceedings of the Royal Society of Edinburgh: Section A Mathematics, 80(1-2), 45–56, (1978). [DOI].
* [137] Tiglio, Manuel, and Villanueva, Aarón, “On ab initio closed-form expressions for gravitational waves”, (2019).
* [138] Tiglio, Manuel, and Villanueva, Aarón, “On the stability and accuracy of the Empirical Interpolation Method and Gravitational Wave Surrogates”, (9 2020)URL:
https://arxiv.org/abs/2009.06151.
* [139] Trefethen, L.N., Spectral Methods in MATLAB, Software, Environments, and Tools, (Society for Industrial and Applied Mathematics (SIAM, 3600 Market Street, Floor 6, Philadelphia, PA 19104), 2000). [DOI].
* [140] Trefethen, L.N., and Bau, D., Numerical Linear Algebra, (SIAM: Society for Industrial and Applied Mathematics, 1997).
* [141] Umans, Christopher, “Fast Polynomial Factorization and Modular Composition in Small Characteristic”, in Proceedings of the Fortieth Annual ACM Symposium on Theory of Computing, STOC ’08, p. 481–490, (Association for Computing Machinery, New York, NY, USA, 2008). [DOI].
* [142] van der Hoeven, Joris, and Lecerf, Grégoire, “Fast multivariate multi-point evaluation revisited”, Journal of Complexity, 56, 101405, (2020). [DOI].
* [143] Varma, Vijay, Field, Scott E., Scheel, Mark A., Blackman, Jonathan, Gerosa, Davide, Stein, Leo C., Kidder, Lawrence E., and Pfeiffer, Harald P., “Surrogate models for precessing binary black hole simulations with unequal masses”, Phys. Rev. Research., 1, 033015, (2019). [DOI].
* [144] Varma, Vijay, Field, Scott E., Scheel, Mark A., Blackman, Jonathan, Kidder, Lawrence E., and Pfeiffer, Harald P., “Surrogate model of hybridized numerical relativity binary black hole waveforms”, Phys. Rev. D, 99, 064045, (Mar 2019). [DOI].
* [145] Varma, Vijay, Gerosa, Davide, Stein, Leo C., Hébert, François, and Zhang, Hao, “High-accuracy mass, spin, and recoil predictions of generic black-hole merger remnants”, Phys. Rev. Lett., 122(1), 011101, (2019). [DOI].
* [146] Wang, Haiyong, and Xiang, Shuhuang, “On the convergence rates of Legendre approximation”, Math. Comput., 81, 861–877, (2011). [DOI].
* [147] Wang, Qian, Hesthaven, Jan S., and Ray, Deep, “Non-intrusive reduced order modeling of unsteady flows using artificial neural networks with application to a combustion problem”, Journal of Computational Physics, 384, 289 – 307, (2019). [DOI].
* [148] Wojtaszczyk, P., “On greedy algorithm approximating Kolmogorov widths in Banach spaces”, Journal of Mathematical Analysis and Applications, 424(1), 685 – 695, (2015). [DOI].
* [149] Xiu, Dongbin, Numerical Methods for Stochastic Computations: A Spectral Method Approach, (Princeton University Press, 2010).
* [150] Zackay, Barak, Venumadhav, Tejaswi, Roulet, Javier, Dai, Liang, and Zaldarriaga, Matias, “Detecting Gravitational Waves in Data with Non-Gaussian Noise”, (8 2019). [arXiv:1908.05644 [astro-ph.IM]].
* [151] Zhao, Junjie, Shao, Lijing, Cao, Zhoujian, and Ma, Bo-Qiang, “Reduced-order surrogate models for scalar-tensor gravity in the strong field regime and applications to binary pulsars and GW170817”, Phys. Rev. D, 100(6), 064034, (2019). [DOI].
|
[type=editor, auid=000,bioid=1, prefix=, role=, orcid=]
[role=, suffix=, ]
# Steady-State Models of STATCOM and UPFC using Flexible Holomorphic Embedding
Pradeep Singh<EMAIL_ADDRESS>Department of Electrical Engineering, Indian
Institute of Technology, Delhi, 30332 INDIA Nilanjan Senroy
<EMAIL_ADDRESS>
###### Abstract
To investigate the effect and ability of FACTS devices using Fast and Flexible
Holomorphic Embedding technique (FFHE), it is necessary to develop an embedded
system for these devices. Therefore, this paper presents FFHE based embedded
system for STATCOM and UPFC. The embedded system is also proposed for their
controlling modes.
The introduced embedded system for STATCOM and UPFC is flexible which allows
to take any state as an initial guess instead of fixed state, which leads
towards the reduced runtime and decrease the required number of terms, as
compared to standard Holomorphic Embedded Load-Flow method (HELM). To
demonstrate the effectiveness and practicability, the proposed models of
STATCOM and UPFC have been tested for several cases. Further, the developed
recursive formulas for power balance equations, devices’ physical constraints
and their controlling modes are thoroughly investigated and examined. From
several tests, it is found that the proposed embedded system requires less
execution time and reduce the error at higher rate.
###### keywords:
FACTS Holomorphic Embedding Power-Flow STATCOM UPFC Voltage Source Converter
## 1 Introduction
FACTS controllers play vital roles in diversified domains of power system
operation and control. These devices improve voltage profile, transient
stability, voltage stability, available transfer capability etc., thereby
improving the overall performance of the power system. But their optimal
installation and performance analysis is indispensable to harness maximum
benefit-cost ratio. Therefore, their accurate steady-state mathematical models
are required to emulate their functionality and performance. These models are
integrated with non-linear static load-flow equations of the power systems,
which increases the complexity of load-flow problem [1, 2, 3, 4, 5].
The conventional steady-state models of FACTS controllers are based on
numerical iterative techniques e.g. Gauss-Seidal (GS), Newton-Raphson (NR),
Fast Decoupled (FD) etc. These iterative techniques are primarily constrained
by the requirement of a good initial guess and they offer slow convergence or
even divergence in some cases. Additionally, the solution provided by these
methods is indecisive, because sometimes these techniques diverge due to a bad
initial guess (although the solution exists) and sometimes converge to a
spurious solution (even when the solution does not exists). These limitations
were overcome by the Holomorphic Embedding Load-Flow Method (HELM) developed
in [6].
The seminal work in development of HELM framework has been presented in [6, 7,
8, 9, 10, 11, 12]; and advanced applications of HELM in power system analysis
can be found in [13, 14, 15, 16, 17, 18, 19, 20]. The main advantage of HELM
are deterministic initial guess and assured convergence if the solution
exists. However, it has been reported that HELM requires 10 to 20 times more
execution time than NR method [21]. Therefore, Fast and Flexible Holomorphic
Embedded (FFHE) method to reduce the execution time has been proposed in [22].
FFHE provides the flexibility to set any arbitrary state as an initial guess.
This property circumvents the need to restart the load-flow program when bus-
type switching takes place. Therefore, FFHE requires less number of terms to
converge if initial guesses are promising. Hence, FFHE is fast and more
suitable for load flow analysis than conventional iterative methods.
HELM framework comprising of HE models of thyristor-based FACTS controllers
have been proposed in [23]. Unified Power Flow Controller (UPFC) is considered
as the superior controller among all FACTS controllers [2, 3, 4]. UPFC is a
VSC-based FACTS controller which are more advanced and sophisticated as
compared to Thyristor-based FACTS controller because they inject less
harmonics, provide independent control, fast dynamic response and higher
flexibility. In [24], a VSC-based FACTS controller, namely Static Synchronous
Compensator (STATCOM), has been developed as a variable voltage source using
standard HELM. However, the presented model of STATCOM requires larger runtime
and also doesn’t offers the flexibility to use any state as the starting
point. Therefore, the main aim of this work is to develop the FFHE based
models of VSC-based FACTS controllers. In this research work, FFHE based
models of two VSC-based FACTS controllers, namely Static Synchronous
Compensator (STATCOM) and Unified Power Flow controller (UPFC) have been
developed along-with their controlling modes.
The remaining part of this paper is organised as follows: Section 2 deals with
the methodology used to develop the FFHE based models of STATCOM and UPFC
along-with their controlling modes. In Section 3, the key findings and
numerical results of the introduced embedded system have been discussed.
Section 4, presents the conclusion of this study.
## 2 Proposed Holomorphic Embedding Formulations
This section addresses the procedure adopted to develop a flexible embedded
system for STATCOM and UPFC. During development of steady-state model, it is
assumed that the system and stated devices are 3-phase balanced and the
harmonics generated by them are negligible. Firstly, the embedded system and
recursive relationships for the STATCOM and UPFC have been developed and
finally, numerical values of all unknown variables along-with their operation
bounds are investigated.
### 2.1 STATCOM Modeling
A STATCOM is a shunt connected FACTS device acts as a controllable voltage
source. In steady-state, it can inject or absorb reactive power by regulating
the injected voltage source magnitude and phase angle. Generally, it consists
of a voltage source converter, a capacitor and a coupling transformer. The
shunt connected FACTS devices are mainly used to regulate the bus voltage
magnitude, although these devices can also be used to control other parameters
of the system as discussed in [25]. Figure 1 shows the topological structure
of a STATCOM.
Figure 1: Equivalent circuit of a STATCOM
In Figure 1, $y_{SH}$, $I_{SH}$ and $V_{SH}$ are the coupling transformer
leakage impedance, complex current and voltage injected by the STATCOM at bus
$i$ respectively. The required equations to represent the STATCOM are the
power balance equation (PBE) at the connected bus and two equations from
device’s physical constraint: real power injection criteria and control mode.
Equations (1), (2), and (3e) provide the PBE at $i^{th}$ bus, real power
injection criterion and various control modes for STATCOM. Equations (3a),
(3b), (3c), (3d), and (3e) represent the control constraint equations for the
bus voltage magnitude, injected voltage phasor magnitude, injected reactive
power, reactive power-flow between buses and equivalent imaginary admittance
respectively.
$V^{*}_{i}\sum_{k=1}^{N}Y_{ik}V_{k}+V^{*}_{i}(V_{i}+V_{SH})y_{SH}=S^{*}_{i}$
(1) $\Re\big{[}V_{SH}(V^{*}_{i}+V^{*}_{SH})y^{*}_{SH}\big{]}=0$ (2)
$V_{i}V^{*}_{i}=|V^{SP}_{i}|^{2}$ (3a) $V_{SH}V^{*}_{SH}=|V^{SP}_{SH}|^{2}$
(3b) $\Im\big{[}V_{SH}(V^{*}_{i}+V^{*}_{SH})y^{*}_{SH}\big{]}=Q^{SP}_{SH}$
(3c) $\Im\big{[}V_{i}(V^{*}_{i}-V^{*}_{j})y^{*}_{ij}\big{]}=Q^{SP}_{ij}$ (3d)
$\Im\bigg{[}\frac{(V_{i}+V_{SH})}{V_{SH}}y_{SH}\bigg{]}=b^{SP}_{eq(SH)}$ (3e)
where, $V_{i}$, $S_{i}$ and $Y_{ik}$ are the voltage phasor of bus $i$,
injected complex power at bus $i$ and $(i,k)^{th}$ element of the bus
admittance matrix respectively. The quantities with suffix $SP$ are the
specified or targeted values. The symbols $N$, $N_{PV}$, and $N_{PQ}$ denotes
the set of total, generator and load buses respectively. The operators
$\Re(\bullet)$, $\Im(\bullet)$ and ${}^{\prime}*^{\prime}$ represents real,
imaginary and conjugation operation respectively.
The injected voltage phasor $V_{SH}$ is treated as a free variable function of
$\alpha$. The proposed holomorphic embedded system for (1), (2), and (3e) are
as follows:
$V^{*}_{i}(\alpha^{*})\sum_{k=1}^{N}Y_{ik}V_{k}(\alpha)+V^{*}_{i}(\alpha^{*})\big{\\{}V_{i}(\alpha)+V_{SH}(\alpha)\big{\\}}y_{SH}=\alpha\bigg{[}S^{*}_{i}-C^{*}_{i}\sum_{k=1}^{N}Y_{ik}C_{k}-C^{*}_{i}(C_{i}+C_{SH})y_{SH}\bigg{]}+C^{*}_{i}\sum_{k=1}^{N}Y_{ik}C_{k}\\\
+C^{*}_{i}(C_{i}+C_{SH})y_{SH}$ (4)
$\Re\big{[}V_{SH}(\alpha)\big{\\{}V^{*}_{i}(\alpha^{*})+V^{*}_{SH}(\alpha^{*})\big{\\}}y^{*}_{SH}\big{]}=-\alpha\Re\big{[}C_{SH}(C^{*}_{i}+C^{*}_{SH})y^{*}_{SH}\big{]}+\Re[C_{SH}(C^{*}_{i}+C^{*}_{SH})y^{*}_{SH}]$
(5)
$V_{i}(\alpha)V^{*}_{i}(\alpha^{*})=C_{i}C^{*}_{i}+\alpha\big{[}|V^{SP}_{i}|^{2}-C_{i}C^{*}_{i}\big{]}$
(6a)
$V_{SH}(\alpha)V^{*}_{SH}(\alpha^{*})=C_{SH}C^{*}_{SH}+\alpha\big{[}|V^{SP}_{SH}|^{2}-C_{SH}C^{*}_{SH}\big{]}$
(6b)
$\Im\big{[}V_{SH}(\alpha)\big{\\{}V^{*}_{i}(\alpha^{*})+V^{*}_{SH}(\alpha^{*})\big{\\}}y^{*}_{SH}\big{]}=\Im\big{\\{}C_{SH}(C^{*}_{i}+C^{*}_{SH})y^{*}_{SH}\big{\\}}+\alpha\big{[}Q^{SP}_{SH}-\Im\big{\\{}C_{SH}(C^{*}_{i}+C^{*}_{SH})y^{*}_{SH}\big{\\}}\big{]}$
(6c)
$\Im\big{[}V_{i}(\alpha)\big{\\{}V^{*}_{i}(\alpha^{*})-V^{*}_{j}(\alpha^{*})\big{\\}}y^{*}_{ij}\big{]}=\Im\big{\\{}C_{i}(C^{*}_{i}-C^{*}_{j})y^{*}_{ij}\big{\\}}+\alpha\big{[}Q^{SP}_{ij}-\Im\big{\\{}C_{i}(C^{*}_{i}-C^{*}_{j})y^{*}_{ij}\big{\\}}\big{]}$
(6d)
$\Im\bigg{[}\frac{\\{V_{i}(\alpha)+V_{SH}(\alpha)\\}}{V_{SH}(\alpha)}y_{SH}\bigg{]}=\Im\bigg{\\{}\frac{(C_{i}+C_{SH})y_{SH}}{C_{SH}}\bigg{\\}}+\alpha\bigg{[}b^{SP}_{eq(SH)}-\Im\bigg{\\{}\frac{(C_{i}+C_{SH})y_{SH}}{C_{SH}}\bigg{\\}}\bigg{]}$
(6e)
Equations (4), (5) and (6e) correspond to (1), (2), and (3e) respectively and
fulfil the requirement of embedding at the reference state $\alpha_{0}=0$ and
the target state $\alpha_{1}=1$. The constants
$C_{SH}\in\mathbb{C}\setminus\\{0\\}$ and $C_{k}\in\mathbb{C}\setminus\\{0\\}$
are adjustable and can be of any pre-specified values. The constant $C_{SH}$
is used to represent the initial value of injected voltage source $V_{SH}$,
whereas constant $C_{k}$ represents the initial value of bus voltages. At
$\alpha=0$, the solution of (4) and (5) gives $V_{SH}[0]=C_{SH}$ and
$V_{k}[0]=C_{k}$. A STATCOM can’t inject active power into the system but it
can inject reactive power. Therefore, $n_{ST}$ STATCOMs can independently
control $2n_{ST}-1$ parameters. Therefore, this configuration can control only
one quantity of the power system independently.
$[A^{ST}]=\begin{bmatrix}\begin{array}[]{cccccccccc}1&0&0&0&0&0&0&0&0&0\\\
0&1&0&0&0&0&0&0&0&0\\\
\mu_{\mathcal{GF}}&\xi_{\mathcal{GF}}&\mu_{\mathcal{GG}}&\xi_{\mathcal{GG}}&\mu_{\mathcal{GL}}&\xi_{\mathcal{GL}}&\mu_{\mathcal{G}i}&\xi_{\mathcal{G}i}&0&0\\\
0&0&C_{\mathcal{G}re}&C_{\mathcal{G}im}&0&0&0&0&0&0\\\
\mu_{\mathcal{LF}}&\xi_{\mathcal{LF}}&\mu_{\mathcal{LG}}&\xi_{\mathcal{LG}}&\mu_{\mathcal{LL}}&\xi_{\mathcal{LL}}&\mu_{\mathcal{L}i}&\xi_{\mathcal{L}i}&0&0\\\
-\xi_{\mathcal{LF}}&\mu_{\mathcal{LF}}&-\xi_{\mathcal{LG}}&\mu_{\mathcal{LG}}&\mu^{\bigstar}_{\mathcal{LL}}&\xi^{\bigstar}_{\mathcal{LL}}&-\xi_{\mathcal{L}i}&\mu_{\mathcal{L}i}&0&0\\\
\mu_{i\mathcal{F}}&\xi_{i\mathcal{F}}&\mu_{i\mathcal{G}}&\xi_{i\mathcal{G}}&\mu_{i\mathcal{L}}&\xi_{i\mathcal{L}}&\mu_{ii}+2C_{ire}g_{SH}+C_{SHre}g_{SH}-C_{SHim}b_{SH}&\xi_{ii}+2C_{iim}g_{SH}+C_{SHim}g_{SH}+C_{SHre}b_{SH}&C_{ire}g_{SH}+C_{iim}b_{SH}&C_{iim}g_{SH}-C_{ire}b_{SH}\\\
-\xi_{i\mathcal{F}}&\mu_{i\mathcal{F}}&-\xi_{i\mathcal{G}}&\mu_{i\mathcal{G}}&-\xi_{i\mathcal{L}}&\mu_{i\mathcal{L}}&\mu^{\bigstar}_{ii}+2C_{ire}b_{SH}+C_{SHre}b_{SH}+C_{SHim}g_{SH}&\xi^{\bigstar}_{ii}+2C_{iim}b_{SH}+C_{SHim}b_{SH}-C_{SHre}g_{SH}&C_{ire}b_{SH}-C_{iim}g_{SH}&C_{iim}b_{SH}+C_{ire}g_{SH}\\\
0&0&0&0&0&0&C_{SHre}g_{SH}+C_{SHim}b_{SH}&C_{SHim}g_{SH}-C_{SHre}b_{SH}&C_{ire}g_{SH}-C_{iim}b_{SH}+2C_{SHre}g_{SH}&C_{iim}g_{SH}+C_{ire}b_{SH}+2C_{SHim}g_{SH}\\\
\lx@intercol\hfil\cdots
Select~{}row~{}vector~{}from~{}equation~{}(\ref{ST_Modes_Mat})~{}corresponding~{}to~{}desired~{}control~{}mode\cdots\hfil\lx@intercol\\\
\end{array}\end{bmatrix}$
(7)
$\begin{bmatrix}\begin{array}[]{cccccccccc}0&0&0&0&0&0&C_{ire}&C_{iim}&0&0\\\
0&0&0&0&0&0&0&0&C_{SHre}&C_{SHim}\\\
0&0&0&0&0&0&C_{SEim}g_{SH}-C_{SHre}b_{SH}&-C_{SHre}g_{SH}-C_{SHim}b_{SH}&-C_{iim}g_{SH}-C_{ire}b_{SH}-2C_{SHre}b_{SH}&C_{ire}g_{SH}-C_{iim}b_{SH}-2C_{SHim}b_{SH}\\\
0&0&0&0&C_{ire}b_{i\mathcal{L}}-C_{iim}g_{i\mathcal{L}}&C_{iim}b_{i\mathcal{L}}+C_{ire}g_{i\mathcal{L}}&C_{\mathcal{L}re}b_{i\mathcal{L}}+C_{\mathcal{L}im}g_{i\mathcal{L}}-2C_{ire}b_{i\mathcal{L}}&C_{\mathcal{L}im}b_{i\mathcal{L}}-C_{\mathcal{L}re}g_{i\mathcal{L}}-2C_{iim}b_{i\mathcal{L}}&0&0\\\
0&0&0&0&0&0&\Im(\frac{y_{SH}}{C_{SH}})&\Re(\frac{y_{SH}}{C_{SH}})&-\Im(\frac{y_{SH}C_{i}}{C^{2}_{SH}})&-\Re(\frac{y_{SH}C_{i}}{C^{2}_{SH}})\\\
\end{array}\end{bmatrix}\begin{bmatrix}X^{ST}\end{bmatrix}=~{}\begin{bmatrix}\begin{array}[]{c}\Gamma_{M1}[n-1]\\\
\Gamma_{M2}[n-1]\\\ \Gamma_{M3}[n-1]\\\ \Gamma_{M4}[n-1]\\\
\Gamma_{M5}[n-1]\\\ \end{array}\end{bmatrix}$
(8)
$[X^{ST}]=\begin{bmatrix}\begin{array}[]{cccccccccc}V_{\mathcal{F}re}[n]&V_{\mathcal{F}im}[n]&V_{\mathcal{G}re}[n]&V_{\mathcal{G}im}[n]&V_{\mathcal{L}re}[n]&V_{\mathcal{L}im}[n]&V_{ire}[n]&V_{iim}[n]&V_{SHre}[n]&V_{SHim}[n]\end{array}\end{bmatrix}^{{}^{\prime}}$
(9)
$[B^{ST}]=\begin{bmatrix}\begin{array}[]{cccccccccc}\Re[\Gamma_{\mathcal{F}}[n-1]]&\Im[\Gamma_{\mathcal{F}}[n-1]]&\Gamma_{\mathcal{G}}[n-1]&\Gamma_{\mathcal{GV}}[n-1]&\Re[\Gamma_{\mathcal{L}}[n-1]]&\Im[\Gamma_{\mathcal{L}}[n-1]]&\Re[\Gamma^{ST}_{i}[n-1]]&\Im[\Gamma^{ST}_{i}[n-1]]&\Gamma^{ST}_{PBE}[n-1]&\Gamma^{ST}_{Mi}[n-1]\end{array}\end{bmatrix}^{{}^{\prime}}$
(10)
The general recurrence relationships for $n\geq 1$ are obtained by comparing
the coefficient of $\alpha^{n}$ and the system of linear equations for STATCOM
$[A^{ST}]_{(\Upsilon\times\Upsilon)}[X^{ST}]_{(\Upsilon\times
1)}=[B^{ST}]_{(\Upsilon\times 1)}$ has been derived, where
$\Upsilon=2(N+n_{ST})$. The derived coefficient matrix, unknown vector and
known vector can be expressed as in (7), (9) and (10) respectively (shown at
the top of this page). The general recurrence formula for stated control modes
are also formulated and shown in (8).
The some entries of known vector $[B^{ST}]$ can be expressed as follows:
$\Gamma^{ST}_{i}[n-1]=\eta_{n1}\bigg{[}S^{*}_{i}-C^{*}_{i}\sum^{N}_{k=1}Y_{ik}C_{k}-C^{*}_{i}(C_{i}+C_{SH})y_{SH}\bigg{]}-\sum^{n-1}_{d=1}\bigg{\\{}V^{*}_{i}[d]V_{i}[n-d]-V^{*}_{i}[d]V_{SH}[n-d]\bigg{\\}}y_{SH}\\\
-\sum^{N}_{k=1}\sum^{n-1}_{d=1}Y_{ik}V^{*}_{i}[d]V_{k}[n-d]$ (11)
$\Gamma^{ST}_{PBE}[n-1]=-\eta_{n1}\Re\big{[}C_{SH}(C^{*}_{i}+C^{*}_{SH})y^{*}_{SH}\big{]}-\Re\Bigg{[}\sum^{n-1}_{d=1}\bigg{\\{}V_{SH}[d]V^{*}_{i}[n-d]+V_{SH}[d]V^{*}_{SH}[n-d]\bigg{\\}}y^{*}_{SH}\Bigg{]}$
(12)
$\Gamma^{ST}_{M1}[n-1]=\frac{\eta_{n1}}{2}\big{[}(V^{SP}_{i})^{2}-C_{i}C^{*}_{i}\big{]}-\frac{1}{2}\sum^{n-1}_{d=1}V_{i}[d]V^{*}_{i}[n-d]$
(13)
The embedding of (3a) and (3b) are similar to each other, therefore, the
general recurrence relationship for $\Gamma^{ST}_{M2}[n-1]$ can be obtained by
changing the subscript $i$ to $SH$ in (13).
$\Gamma^{ST}_{M3}[n-1]=\eta_{n1}\big{[}Q^{SP}_{SH}-\Im\big{\\{}C_{SH}(C^{*}_{i}+C^{*}_{SH})y^{*}_{SH}\big{\\}}\big{]}-\Im\Bigg{[}\sum^{n-1}_{d=1}\bigg{\\{}V_{SH}[d]V^{*}_{i}[n-d]+V_{SH}[d]V^{*}_{SH}[n-d]\bigg{\\}}y^{*}_{SH}\Bigg{]}$
(14)
$\Gamma^{ST}_{M4}[n-1]=\eta_{n1}\big{[}Q^{SP}_{ij}-\Im\big{\\{}C_{i}(C^{*}_{i}-C^{*}_{j})y^{*}_{ij}\big{\\}}\big{]}-\Im\Bigg{[}\sum^{n-1}_{d=1}\bigg{\\{}V_{i}[d]V^{*}_{i}[n-d]-V_{i}[d]V^{*}_{j}[n-d]\bigg{\\}}y^{*}_{ij}\Bigg{]}$
(15)
$\Gamma^{ST}_{M5}[n-1]=\eta_{n1}\Bigg{[}b^{SP}_{eq(SH)}-\Im\bigg{\\{}\frac{(C_{i}+C_{SH})y_{SH}}{C_{SH}}\bigg{\\}}\Bigg{]}-\Im\Bigg{[}\sum^{n-1}_{d=1}y_{SH}\bigg{\\{}-W_{SH}[d]V_{SH}[n-d]\frac{C_{i}}{C_{SH}}\\\
+V_{i}[d]W_{SH}[n-d]\bigg{\\}}\Bigg{]}$ (16)
### 2.2 UPFC Modeling
Figure 2: Equivalent circuit of a UPFC
A UPFC is a shunt-series connected FACTS device. Figure 2 shows the
topological structure of a UPFC. It consists of two voltage source converters,
two coupling transformers and a common capacitor. The shunt VSC is coupled to
a local bus $i$ through a shunt connected transformer and the series VSC is
coupled to a transmission line through a series connected transformer. The
shunt converter can inject the reactive power into the system and it allows
the exchange of the real power to the series converter through a common DC
link to satisfy the device’s physical constraint. The UPFC with $n_{UP}$
branches (i.e. $2n_{UP}$ free variables) can independently control $2n_{UP}-1$
parameters because one variable is responsible for real power balance of the
device. Therefore, the circuit configuration shown in Figure 2 can
simultaneously control three quantities of the system independently. The real
power can be exchanged amongst the shunt and series converters via a common DC
link, but the sum of active power exchanged must be equal to zero.
Due to incorporation of the UPFC in the system, two new complex variable
$V_{SH}$ and $I_{SE}$ have been introduced, which further add four unknown
variables. Therefore, to find out the unique solution of unknown variables,
four additional equations are required. To model the UPFC into load-flow
algorithms, one equation from device’s physical constraint and three equations
from controlling modes of the device along with the modification of PBE at the
local bus $i$ and receiving end bus $m$ are required. The mathematical
formulation of the shunt and series converter controlling modes are similar to
the controlling modes of STATCOM and SSSC [26] respectively. The PBE at bus
$i$, $m$ and real power power exchange constraints can be expressed as
follows:
$V^{*}_{i}\sum_{k=1}^{N}Y_{ik}V_{k}+V^{*}_{i}I_{SE}+V^{*}_{i}(V_{i}+V_{SH})y_{SH}=S^{*}_{i}$
(17) $V^{*}_{m}\sum_{k=1}^{N}Y_{mk}V_{k}-V^{*}_{m}I_{SE}=S^{*}_{m}$ (18)
$\Re\big{[}(V_{m}-V_{i})I^{*}_{SE}\big{]}+\Re\big{[}V_{SH}(V^{*}_{i}+V^{*}_{SH})y^{*}_{SH}\big{]}=0$
(19)
The following new embedded system is introduced for (17)-(19):
$V^{*}_{i}(\alpha^{*})\sum_{k=1}^{N}Y_{ik}V_{k}(\alpha)+V^{*}_{i}(\alpha^{*})\big{\\{}(V_{i}(\alpha)+V_{SH}(\alpha))y_{SH}+I_{SE}(\alpha)\big{\\}}=C^{*}_{i}\sum_{k=1}^{N}Y_{ik}C_{k}+C^{*}_{i}\big{\\{}(C_{i}+C_{SH})y_{SH}+D_{SE}\big{\\}}\\\
+\alpha\Bigg{[}S^{*}_{i}-C^{*}_{i}\sum_{k=1}^{N}Y_{ik}C_{k}-C^{*}_{i}\big{\\{}(C_{i}+C_{SH})y_{SH}+D_{SE}\big{\\}}\Bigg{]}$
(20)
$V^{*}_{m}(\alpha^{*})\sum_{k=1}^{N}Y_{mk}V_{k}(\alpha)-V^{*}_{m}(\alpha^{*})I_{SE}(\alpha)=C^{*}_{m}\sum_{k=1}^{N}Y_{mk}C_{k}-C^{*}_{m}D_{SE}+\alpha\Bigg{[}S^{*}_{m}-C^{*}_{m}\sum_{k=1}^{N}Y_{mk}C_{k}+C^{*}_{m}D_{SE}\Bigg{]}$
(21)
$\Re\bigg{[}\big{\\{}V_{m}(\alpha)-V_{i}(\alpha)\big{\\}}I^{*}_{SE}(\alpha^{*})+V_{SH}(\alpha)y^{*}_{SH}\big{\\{}V^{*}_{i}(\alpha^{*})+V^{*}_{SH}(\alpha^{*})\big{\\}}\bigg{]}=\Re\big{[}\\{C_{m}-C_{i}\\}D^{*}_{SE}+C_{SH}(C^{*}_{i}+C^{*}_{SH})y^{*}_{SH}\big{]}\\\
-\alpha\Re\big{[}\\{C_{m}-C_{i}\\}D^{*}_{SE}+C_{SH}(C^{*}_{i}+C^{*}_{SH})y^{*}_{SH}\big{]}$
(22)
$[A^{UP}]=\begin{bmatrix}\begin{array}[]{cccccccccccccc}1&0&0&0&0&0&0&0&0&0&0&0&0&0\\\
0&1&0&0&0&0&0&0&0&0&0&0&0&0\\\
\mu_{\mathcal{GF}}&\xi_{\mathcal{GF}}&\mu_{\mathcal{GG}}&\xi_{\mathcal{GG}}&\mu_{\mathcal{GL}}&\xi_{\mathcal{GL}}&\mu_{\mathcal{G}i}&\xi_{\mathcal{G}i}&\mu_{\mathcal{G}m}&\xi_{\mathcal{G}m}&0&0&0&0\\\
0&0&C_{\mathcal{G}re}&C_{\mathcal{G}im}&0&0&0&0&0&0&0&0&0&0\\\
\mu_{\mathcal{LF}}&\xi_{\mathcal{LF}}&\mu_{\mathcal{LG}}&\xi_{\mathcal{LG}}&\mu_{\mathcal{LL}}&\xi_{\mathcal{LL}}&\mu_{\mathcal{L}i}&\xi_{\mathcal{L}i}&\mu_{\mathcal{L}m}&\xi_{\mathcal{L}m}&0&0&0&0\\\
-\xi_{\mathcal{LF}}&\mu_{\mathcal{LF}}&-\xi_{\mathcal{LG}}&\mu_{\mathcal{LG}}&\mu^{\bigstar}_{\mathcal{LL}}&\xi^{\bigstar}_{\mathcal{LL}}&-\xi_{\mathcal{L}i}&\mu_{\mathcal{L}i}&-\xi_{\mathcal{L}m}&\mu_{\mathcal{L}m}&0&0&0&0\\\
\mu_{i\mathcal{F}}&\xi_{i\mathcal{F}}&\mu_{i\mathcal{G}}&\xi_{i\mathcal{G}}&\mu_{i\mathcal{L}}&\xi_{i\mathcal{L}}&\mu_{ii}+D_{SE1re}+D_{SEre}+2C_{ire}g_{SH}+C_{SHre}g_{SH}-C_{SHim}b_{SH}&\xi_{ii}+D_{SEim}+2C_{iim}g_{SH}+C_{SHim}g_{SH}+C_{SHre}b_{SH}&\mu{im}&\xi_{im}&C_{ire}g_{SH}+C_{iim}b_{SH}&C_{iim}g_{SH}-C_{ire}b_{SH}&C_{ire}&C_{iim}\\\
-\xi_{i\mathcal{F}}&\mu_{i\mathcal{F}}&-\xi_{i\mathcal{G}}&\mu_{i\mathcal{G}}&-\xi_{i\mathcal{L}}&\mu_{i\mathcal{L}}&\mu^{\bigstar}_{ii}+D_{SEim}+2C_{ire}b_{SH}+C_{SHre}b_{SH}+C_{SHim}g_{SH}&\xi^{\bigstar}_{ii}-D_{SEre}+2C_{iim}b_{SH}+C_{SHim}b_{SH}-C_{SHre}g_{SH}&-\xi_{im}&\mu_{im}&C_{ire}b_{SH}-C_{iim}g_{SH}&C_{ire}g_{SH}+C_{iim}b_{SH}&-C_{iim}&C_{ire}\\\
\mu_{m\mathcal{F}}&\xi_{m\mathcal{F}}&\mu_{m\mathcal{G}}&\xi_{m\mathcal{G}}&\mu_{m\mathcal{L}}&\xi_{m\mathcal{L}}&\mu_{mi}&\xi_{mi}&\mu_{mm}-D_{SEre}&\xi_{mm}-D_{SEim}&0&0&-C_{mre}&-C_{mim}\\\
-\xi_{m\mathcal{F}}&\mu_{m\mathcal{F}}&-\xi_{m\mathcal{G}}&\mu_{m\mathcal{G}}&-\xi_{m\mathcal{L}}&\mu_{m\mathcal{L}}&-\xi_{mi}&\mu_{mi}&\mu^{\bigstar}_{mm}-D_{SEim}&\xi^{\bigstar}_{mm}+D_{SEim}&0&0&C_{mim}&-C_{mre}\\\
0&0&0&0&0&0&-D_{SE1re}-D_{SE2re}&-D_{SE1im}-D_{SE2im}&D_{SE1re}&D_{SE2im}&C_{mre}-C_{ire}&C_{mim}-C_{iim}&C_{tre}-C_{ire}&C_{tim}-C_{iim}\\\
\lx@intercol\hfil\cdots
Select~{}3~{}different~{}row~{}vector~{}from~{}equation~{}(21)~{}of~{}[26]~{}and~{}(\ref{ST_Modes_Mat})~{}corresponding~{}to~{}desired~{}control~{}mode\cdots\hfil\lx@intercol\\\
\end{array}\end{bmatrix}$
(23)
$[X^{UP}]=\begin{bmatrix}\begin{array}[]{cccccccccccccc}V_{\mathcal{F}re}[n]&V_{\mathcal{F}im}[n]&V_{\mathcal{G}re}[n]&V_{\mathcal{G}im}[n]&V_{\mathcal{L}re}[n]&V_{\mathcal{L}im}[n]&V_{ire}[n]&V_{iim}[n]&V_{mre}[n]&V_{mim}[n]&V_{SHre}[n]&V_{SHim}[n]&I_{SEre}[n]&I_{SEim}[n]\end{array}\end{bmatrix}^{{}^{\prime}}$
(24)
$[B^{UP}]=\begin{bmatrix}\begin{array}[]{cccccccccccccc}\Re[\Gamma_{\mathcal{F}}[n-1]]&\Im[\Gamma_{\mathcal{F}}[n-1]]&\Gamma_{\mathcal{G}}[n-1]&\Gamma_{\mathcal{GV}}[n-1]&\Re[\Gamma_{\mathcal{L}}[n-1]]&\Im[\Gamma_{\mathcal{L}}[n-1]]&\Re[\Gamma^{UP}_{i}[n-1]]&\Im[\Gamma^{UP}_{i}[n-1]]&\Re[\Gamma^{UP}_{m}[n-1]]&\Im[\Gamma^{UP}_{m}[n-1]]&\Gamma^{UP}_{PBE}[n-1]&\Gamma^{UP}_{Mi}[n-1]&\Gamma^{UP}_{Mi}[n-1]&\Gamma^{UP}_{Mi}[n-1]\end{array}\end{bmatrix}^{{}^{\prime}}$
(25)
The similar procedure has been adopted to obtain the system of linear
equations as discussed in Section 2.1. The system of linear equations for UPFC
$[A^{UP}]_{(\Upsilon\times\Upsilon)}[X^{UP}]{(\Upsilon\times
1)}=[B^{UP}]{(\Upsilon\times 1)}$ is given in (23), (24) and (25) respectively
(shown at the bottom of this page), where $\Upsilon=2(N+2n_{UP})$. For purpose
of numerical studies, the different controlling modes for the shunt and series
converters have been adopted from (8) and [26] respectively. Some entries of
known vector $[B^{UP}]$ are directly taken from [26] (e.g. entries of slack,
load and generator buses) and shown in Appendix A; and the remaining are as
follows:
$\Gamma^{UP}_{i}[n-1]=\eta_{n1}\bigg{[}S^{*}_{i}-C^{*}_{i}\sum^{N}_{k=1}Y_{ik}C_{k}-C^{*}_{i}(C_{i}+C_{SH})y_{SH}-C^{*}_{i}D_{SE}\bigg{]}-\sum^{n-1}_{d=1}\bigg{\\{}V^{*}_{i}[d]V_{i}[n-d]+V^{*}_{i}[d]V_{SH}[n-d]\bigg{\\}}y_{SH}\\\
-\sum^{N}_{k=1}\sum^{n-1}_{d=1}Y_{ik}V^{*}_{i}[d]V_{k}[n-d]-\sum^{n-1}_{d=1}V^{*}_{i}[d]I_{SE}[n-d]$
(26)
$\Gamma^{UP}_{m}[n-1]=\eta_{n1}\bigg{[}S^{*}_{m}-C^{*}_{m}\sum^{N}_{k=1}Y_{mk}C_{k}+C^{*}_{m}D_{SE}\bigg{]}-\sum^{N}_{k=1}\sum^{n-1}_{d=1}Y_{mk}V^{*}_{m}[d]V_{k}[n-d]+\sum^{n-1}_{d=1}V^{*}_{m}[d]I_{SE}[n-d]$
(27)
$\Gamma^{UP}_{PBE}[n-1]=-\eta_{n1}\Re\big{[}C_{SH}(C^{*}_{i}+C^{*}_{SH})y^{*}_{SH}+(C_{m}-C_{i})D^{*}_{SE}\big{]}-\Re\Bigg{[}\sum^{n-1}_{d=1}\bigg{\\{}V_{m}[d]I^{*}_{SE}[n-d]-V_{i}[d]I^{*}_{SE}[n-d]\bigg{\\}}\Bigg{]}\\\
-\Re\Bigg{[}\sum^{n-1}_{d=1}\bigg{\\{}V_{SH}[d]V^{*}_{i}[n-d]+V_{SH}[d]V^{*}_{SH}[n-d]\bigg{\\}}y^{*}_{SH}\Bigg{]}$
(28)
## 3 Results and Discussions
The proposed models of STATCOM and UPFC; and their multi-control capabilities
have been tested on IEEE 118-bus test systems [27]. But for demonstration
purpose, only selected numerical results on IEEE 118-bus test system have been
presented in this paper. To validate the proposed models, investigation has
been done by incorporating the different devices at different locations and by
changing the target values of quantities. All the quantities are in p.u. and
the base power being is 100 MVA. The proposed FFHE based models have been
tested and examined for all the control modes as discussed in Section 2. The
proposed FFHE and NR based models of STATCOM and UPFC were modelled in MATLAB
environment and simulated on Intel(R) Core(TM) i3-4150 CPU 3.50 GHz processor
with 4-GB RAM. For all the investigations, the mismatch tolerance of $10^{-8}$
is considered for maximum bus power mismatch. The values of unknown variables
obtained after 3 iterations of NR method have been used as initial guesses for
the FFHE load-flow method throughout this paper. The numerical results of IEEE
118-bus test system in the absence of FACTS devices are presented in Table 1.
The similar procedure as discussed in [26] has been adopted to calculate the
percentage reduction in mismatch and runtime.
Table 1: Numerical results of IEEE 118-bus test system without any FACTS device $V_{16}=0.9839\angle-17.80\degree$ | $S_{16-17}=-0.1752-j0.0370$
---|---
$V_{20}=0.9581\angle-17.81\degree$ | $S_{20-21}=-0.2869+j0.0496$
$V_{75}=0.9675\angle-7.07\degree$ | $S_{75-74}=0.5236+j0.0604$
$V_{114}=0.9607\angle-15.27\degree$ | $S_{114-115}=0.0134+j0.0088$
Table 2: Numerical results when STATCOM is incoporated at bus no. 16 in the
IEEE 118-bus test system
| | Case 1 | Case 2 | Case 3 | Case 4 | Case 5
---|---|---|---|---|---|---
Specified Parameters | $V^{SP}_{16}=1.1$ | $V^{SP}_{SH}=1$ | $Q^{SP}_{SH}=-0.5$ | $Q^{SP}_{16-17}=0$ | $b^{SP}_{eq(SH)}=-0.8$
Power-flow results | $V_{16}$ | $\textbf{1.1}\angle-20.28\degree$ | $0.9977\angle-17.99\degree$ | $0.9519\angle-17.45\degree$ | $0.9867\angle-17.84\degree$ | $1.0324\angle-18.61\degree$
$V_{SH}$ | $1.1197\angle 158.68\degree$ | $\textbf{1}\angle 161.87\degree$ | $0.9466\angle 162.87\degree$ | $0.9872\angle 162.13\degree$ | $1.0407\angle 160.93\degree$
$S_{16-17}$ | $-0.2121+j0.6672$ | $-0.1769+j0.0578$ | $-0.1736-j0.1739$ | $-0.1754+\textbf{j0}$ | $-0.1843+j0.2506$
$S_{SH}$ | $0+j2.2210$ | $0+j0.2352$ | $0-\textbf{j0.5}$ | $0+j0.0470$ | $0+j0.8664$
$b_{eq(SH)}$ | $-1.7717$ | $-0.2352$ | $0.5580$ | $-0.0482$ | -0.8
$\%\vartriangle E$ | $25.91$ | $24.24$ | $27.92$ | $28.95$ | $33.87$
$\%T$ | $11.63$ | $9.25$ | $8.10$ | $10.97$ | $9.23$
Table 3: Numerical results when STATCOM is incoporated at bus no. 114 in the
IEEE 118-bus test system
| | Case 1 | Case 2 | Case 3 | Case 4 | Case 5
---|---|---|---|---|---|---
Specified Parameters | $V^{SP}_{114}=0.9$ | $V^{SP}_{SH}=1$ | $Q^{SP}_{SH}=-0.1$ | $Q^{SP}_{114-115}=0$ | $b^{SP}_{eq(SH)}=0.2$
Power-flow results | $V_{114}$ | $\textbf{0.9}\angle-15.03\degree$ | $0.9927\angle-15.76\degree$ | $0.9569\angle-15.23\degree$ | $0.9598\angle-15.26\degree$ | $0.9539\angle-15.19\degree$
$V_{SH}$ | $0.8843\angle 165.96\degree$ | $\textbf{1}\angle 163.81\degree$ | $0.9559\angle 164.83\degree$ | $0.9595\angle 164.75\degree$ | $0.9520\angle 164.92\degree$
$S_{114-115}$ | $0.0095-j0.6308$ | $0.0083+j0.3854$ | $0.0135-j0.0319$ | $0.0134-\textbf{j0}$ | $0.0135-j0.0662$
$S_{SH}$ | $0-j1.3773$ | $0+j0.7374$ | $0-\textbf{j0.1}$ | $0-j0.0241$ | $0-j0.1813$
$b_{eq(SH)}$ | $1.7613$ | $-0.7374$ | $0.1094$ | $0.0262$ | 0.2
$\%\vartriangle E$ | $22.35$ | $23.16$ | $28.48$ | $24.23$ | $29.18$
$\%T$ | $12.43$ | $9.47$ | $7.69$ | $9.78$ | $10.89$
To validate the proposed embedding for STATCOM and its controlling modes,
several test have been carried out on IEEE 118-bus test system. For
demonstration purpose, only two location results have been presented in Table
2. For examining cases 1-5, STATCOM is assumed to be connected at bus 16. The
specified reference values; and calculated values of STATCOM’s variables and
selected system variables are given in row 2 and 3 respectively. In case 1,
the voltage magnitude of STATCOM connected bus was chosen as a control
variable. In this case, STATCOM is able to set the voltage of bus 16 to the
specified reference 1.1 p.u. by providing the reactive power support to that
bus. The examination of case 2 from this table proves that the proposed
formulation also enables the model of STATCOM to control its injected voltage
to a specified reference 1 p.u. Case 3 demonstrates the reactive power
absorption capability of the device and it can be observed that the absorbed
reactive power is equal to the desired value. As per literature, STATCOM also
has the capability to control reactive power-flow through the line, case 4
verifies the same. In this case, STATCOM controls the reactive power-flow of
line 16-17 from $-$0.0370 p.u. to target value 0 p.u. So, by forcing the
reactive power-flow through the line to zero, the active power transferring
capability of the line can be increased. This particular mode may be
attractive in context of the electricity market environment. The proposed
formulation also enables the model to operate like a controllable admittance
and it can be verified from case 5 as shown in Table 2.
Now, a STATCOM is installed at bus 114 and the desired values of control
parameters, calculated values of system variables, STATCOM’s parameters for
all cases are presented in Table 3. From this table, it can be observed that
the all the desired values (shown in bold letters) have been achieved. From
Tables 2 and 3, it can also be verified that the NR based model of STATCOM
takes more runtime as compared to NR-assisted FFHE based model of STATCOM and
the proposed model also reduces the error at faster rate. In the literature
only STATCOM model based on standard HELM has been developed in [24],
therefore, the performance of NR-assisted FFHE and basic HELM based model of
STATCOM is also investigated. From the various test cases, it has been
observed that the FFHE based model of STATCOM takes 80-90% less time and
requires 40-70% less terms to converge as compared to basic HELM based model
of STATCOM.
Table 4: Numerical results when UPFC in the IEEE 118-bus test system (location: $75-74$) | | Case 1 | Case 2
---|---|---|---
Specified Parameters | $V^{SP}_{75}=1$ | $V^{SP}_{SH}=1$
$P^{SP}_{75-74}=0.75$ | $P^{SP}_{75-74}=0.2$
$Q^{SP}_{75-74}=0$ | $Q^{SP}_{75-74}=0.1$
Power-flow results | $V_{75}$ | $\textbf{1}\angle-7.99\degree$ | $0.9937\angle-6.80\degree$
$V_{SH}$ | $1.0087\angle 171.52\degree$ | $\textbf{1}\angle 172.83\degree$
$S_{SH}$ | $0.0129+j0.8636$ | $-0.002+j0.6315$
$b_{eq(SH)}$ | $-0.8488$ | $-0.6315$
$S_{75-74}$ | 0.75+j0 | 0.2+j0.1
$V_{SE}$ | $0.0725\angle 95.78\degree$ | $0.0855\angle-117.30\degree$
$I_{SE}$ | $0.75\angle-7.99\degree$ | $0.2250\angle-33.37\degree$
$S_{SE}$ | $-0.0129+j0.0528$ | $0.002-j0.0191$
$X_{eq(SE)}$ | $0.0938$ | $-0.3778$
$\%\vartriangle E$ | $26.58$ | $23.37$
$\%T$ | $5.19$ | $6.45$
Further, the proposed embedded system for UPFC has been also tested for two
cases. Firstly, it is assumed that the UPFC is connected between the buses 75
and 74 at the location of bus 75. In case 1, the shunt control of UPFC is to
control the voltage at bus 75 and the series controls of UPFC are to control
active and reactive power flowing through the transmission line 75-74.
Similarly, in case 2, the shunt control of UPFC is to control injected voltage
and series control responsibilities of UPFC are the same except the desired
values. The rows 2, 3, 4, 5 of Table 4 presents the specified parameters,
calculated values of UPFC’s and system variables, percentage reduction in
error and runtime respectively. The active power exchange between the shunt
and series converters is also presented in this table. In Table 4, numerically
it is shown that the all target values (shown in bold letters) have been met
for both cases. The decrement in error ranges from 15% to 27% and reduction in
runtime ranges from 3% to 7%.
Table 5 presents the results for two cases when the UPFC is assumed to be
connected between buses 20 and 21. The UPFC parameters i.e. magnitudes and
phase angles of the UPFC’s injected voltages $V_{SE}$, $V_{SH}$ are also given
in this table, which enforce the system operating point to set as per
specified reference quantities. The calculated values of specified references
are shown in bold letters and it can be verified that all the desired values
have been achieved. From Tables 5, it can be observed that the error reduction
is higher and runtime is lesser for FFHE based model of UPFC. Note that in all
the cases, convergence has been achieved; and the proposed models exhibited
very good convergence and also converge at faster rate as compared to models
based on standard NR method.
Table 5: Numerical results when UPFC in the IEEE 118-bus test system (location: $20-21$) | | Case 1 | Case 2
---|---|---|---
Specified Parameters | $V^{SP}_{20}=1$ | $V^{SP}_{SH}=1$
$P^{SP}_{20-21}=0.6$ | $P^{SP}_{20-21}=-0.4$
$Q^{SP}_{20-21}=0$ | $Q^{SP}_{20-21}=0$
Power-flow results | $V_{20}$ | $\textbf{1}\angle-28.95\degree$ | $0.9981\angle-17.48\degree$
$V_{SH}$ | $1.0051\angle 150.77\degree$ | $\textbf{1}\angle 162.38\degree$
$S_{SH}$ | $0.0093+j0.5087$ | $-0.0290+j0.2182$
$b_{eq(SH)}$ | $-0.5035$ | $-0.2182$
$S_{20-21}$ | 0.6+j0 | $-$0.4+j0
$V_{SE}$ | $0.5545\angle 62.65\degree$ | $0.0866\angle-164.27\degree$
$I_{SE}$ | $0.6\angle-28.95\degree$ | $0.4008\angle 162.53\degree$
$S_{SE}$ | $-0.0093+j0.3326$ | $0.0290+j0.0190$
$X_{eq(SE)}$ | $0.9238$ | $0.1183$
$\%\vartriangle E$ | $24.31$ | $25.20$
$\%T$ | $7.54$ | $6.01$
Table 6: Power-flow results when device limit constraints are imposed Device | Specified Parameter (p.u.) | Limit Violated | Power-flow results
---|---|---|---
STATCOM | $V^{SP}_{16}=1.1$ | Yes | $V_{16}={\color[rgb]{0,0,1}1.0830}\angle-19.81\degree$
$V_{SH}=\textbf{1.1}\angle 159.28\degree$
$S_{16-17}=-0.2038+j0.5580$
$S_{SH}=0+j1.8806$
$b_{eq(SH)}=-1.5542$
$\%\vartriangle E=23.85$
$\%T=6.38$
UPFC | | Yes | $V_{20}=\textbf{1}\angle-23.77\degree$
| $V_{SE}=\textbf{0.3}\angle 72.46\degree$
| $I_{SE}=0.1813\angle-23.77\degree$
| $S_{20-121}={\color[rgb]{0,0,1}0.1813}+j\textbf{0}$
$V^{SP}_{20}=1$ | $S_{SE}=-0.0059+j0.0541$
$P^{SP}_{20-21}=0.6$ | $X_{eq(SE)}=1.6446$
$Q^{SP}_{20-21}=0$ | $V_{SH}=1.0038\angle 156.63$
| $S_{SH}=0.0059+j0.3724$
| $b_{eq(SH)}=-0.3696$
| $\%\vartriangle E=17.37$
| $\%T=3.94$
The introduced embedded system has also been investigated with device limit
constraints. To demonstrate the handling of constraints, one case from Table 2
and another case from Table 5 were chosen. The maximum value of injected
voltage by SSSC and STATCOM were selected as 0.3 p.u. and 1.1 p.u.
respectively. From Table 2, it can be observed that when the target voltage of
bus 16 is chosen as 1.1 p.u. and no limits were set on the injected voltage,
the STATCOM is able to control the voltage magnitude of bus 16 to the
specified reference. But when the limits are imposed, STATCOM is not able to
control the same because higher $V_{SH}$ is required to maintain the same.
Therefore, whenever the limit constraints are violated, current mode is
relaxed; and STATCOM will acts as a constant voltage source (i.e. injected
voltage magnitude mode is activated) and reference value is set equal to the
value of violated limit. From Table 6, it can be observed that the voltage
magnitude of bus 16 is not equal to the reference value but it is equal to the
1.0830 p.u., while the injected voltage is maintained to 1.1 p.u. Table 6 also
presents the results when the limit constraints for UPFC are imposed and
similar results are achieved. Therefore, it can be concluded that the proposed
FFHE based models of STATCOM and UPFC are also suitable when limit constraints
of the devices are imposed.
The NR and FFHE based models have been tested on same platform for comparison,
although the derived inferences may not be general due to various differences
in simulation structure, programming skills etc. It is observed that the
percentage reduction in error and runtime are different for each case. Because
there is no particular choice of the initial guess for the proposed models to
ensure fast convergence in all cases. The selection of promising initial guess
still remains an issue. Briefly, the runtime and rate of convergence of the
proposed models will vary for different initial guess. Although, the
performance of proposed FFHE based FACTS devices model are sensitive to the
initial guess, but the proposed model perform better than basic HELM based
models. Moreover, for most of cases, FFHE based models outperforms NR based
models in context of error reduction. A common conclusion for all the stated
devices and their controlling modes is that the proposed embedding for the
devices are reliable and have good convergence characteristics. Further, the
proposed formulation reduces the mismatch error at faster rate and therefore,
converges slightly faster than the NR method based FACTS devices models.
Therefore, the proposed FFHE based models offers a good alternative to the NR
based models when promising initial guesses are available.
## 4 Conclusion
The purpose of this research work is to develop flexible HELM models of
STATCOM, and UPFC. To do the same, the power balance equations and devices’
controlling modes are embedded with a complex variable $\alpha$ in such a
manner that the embedded equations satisfy the requirement of embedding.
Afterwards, the unknown variables were represented by power series and
recursive formulas have been obtained for $n\geq 1$. Lastly, at $\alpha=1$
numerical values of unknown variables are calculated using determinant method.
The proposed embedded system of stated devices provides flexibility because
any state can serve as an initial guess.
From the numerical results, it is observed that the results of FFHE based
models are similar to results obtained by NR based model of devices in
perspective of the final calculated values of devices parameters, system
variables and operational bounds. The comparison of runtime for FFHE based
models with NR based models showed that later one took more time. Although the
performance of FFHE based models is sensitive to the initial guess but the NR
assisted FFHE based models outperform the NR based models in error reduction.
The proposed model of STATCOM takes very less execution time as compared to
the basic HELM model of STATCOM. The numerous results showed that the FFHE
based models represents a step forward compared to the NR based models. So,
the proposed models may be very useful at the planning stage of the power
system. Therefore, the proposed FFHE based models offers a good alternative to
the NR based models when promising initial guesses are available.
In this paper, no sophisticated technique is used during implementation of
FFHE based models, therefore, possibility of reducing runtime further is still
large via parallel computing, code optimization etc.
## Appendix A Entries of Coefficient matrix and Known Vector
The entries of coefficient matrix $[A^{ST}]$ and $[A^{UP}]$ are as follows:
$\mu_{ik}=G_{ik}C_{ire}+B_{ik}C_{iim}~{}~{};~{}k\neq i$ (29)
$\xi_{ik}=G_{ik}C_{iim}-B_{ik}C_{ire}~{}~{};~{}k\neq i$ (30)
$\mu_{ii}=G_{ii}C_{ire}+B_{ii}C_{iim}+\sum^{N}_{k=1}\big{(}G_{ik}C_{kre}-B_{ik}C_{kim}\big{)}$
(31)
$\xi_{ii}=G_{ii}C_{iim}-B_{ii}C_{ire}+\sum^{N}_{k=1}\big{(}G_{ik}C_{kim}+B_{ik}C_{kre}\big{)}$
(32)
$\mu^{\bigstar}_{ii}=-G_{ii}C_{iim}+B_{ii}C_{ire}+\sum^{N}_{k=1}\big{(}G_{ik}C_{kim}+B_{ik}C_{kre}\big{)}$
(33)
$\xi^{\bigstar}_{ii}=G_{ii}C_{ire}+B_{ii}C_{iim}+\sum^{N}_{k=1}\big{(}B_{ik}C_{kim}-G_{ik}C_{kre}\big{)}$
(34)
The entries of known vector $[B^{ST}]$ and $[B^{UP}]$ are as follows:
$\Gamma_{\mathcal{F}}=\eta_{n1}(V^{SP}_{\mathcal{F}}-C_{\mathcal{F}})$ (35)
$\Gamma_{\mathcal{G}}=\frac{\eta_{n1}}{2}\Bigg{[}2P_{\mathcal{G}}-\Re\bigg{\\{}C_{\mathcal{G}}\sum^{N}_{k=1}Y^{*}_{\mathcal{G}k}C^{*}_{k}\bigg{\\}}\Bigg{]}-\Re\Bigg{[}\sum^{N}_{k=1}\sum^{n-1}_{d=1}Y^{*}_{\mathcal{G}k}V^{*}_{\mathcal{G}}[d]V_{k}[n-d]\Bigg{]}$
(36)
$\Gamma_{\mathcal{GV}}=\frac{\eta_{n1}}{2}\big{[}(V^{SP}_{\mathcal{G}})^{2}-C_{\mathcal{G}}C^{*}_{\mathcal{G}}\big{]}-\frac{1}{2}\sum^{n-1}_{d=1}V_{\mathcal{G}}[d]V^{*}_{\mathcal{G}}[n-d]$
(37)
$\Gamma_{\mathcal{L}}=\eta_{n1}\bigg{[}S^{*}_{\mathcal{L}}-C^{*}_{\mathcal{L}}\sum^{N}_{k=1}Y_{\mathcal{L}k}C_{k}\bigg{]}-\sum^{N}_{k=1}\sum^{n-1}_{d=1}Y_{\mathcal{L}k}V^{*}_{\mathcal{L}}[d]V_{k}[n-d]$
(38)
$\Gamma^{SC}_{M1}=\eta_{n1}\big{[}P^{SP}_{im}-\Re(C_{i}D^{*}_{SE})\big{]}-\Re\Bigg{[}\sum^{n-1}_{d=1}V_{i}[d]I^{*}_{SE}[n-d]\Bigg{]}$
(39)
$\Gamma^{SC}_{M2}=\eta_{n1}\big{[}Q^{SP}_{im}-\Im(C_{i}D^{*}_{SE})\big{]}-\Im\Bigg{[}\sum^{n-1}_{d=1}V_{i}[d]I^{*}_{SE}[n-d]\Bigg{]}$
(40)
As, $F_{SE}(\alpha)$ and $|I_{SE}(\alpha)|$ are the inverse and magnitude of
$I_{SE}(\alpha)$, so, the general recurrence relationship between these can be
expressed as given in (41) and (42).
$F_{SE}[n]=\frac{-1}{D_{SE}}\sum^{n-1}_{d=0}F_{SE}[d]I_{SE}[n-d]$ (41)
$\big{|}I_{SE}[n]\big{|}=\frac{1}{2\big{|}D_{SE}\big{|}}\bigg{\\{}\sum^{n}_{d=0}I_{SE}[d]I^{*}_{SE}[n-d]-\sum^{n-1}_{d=1}\big{|}I_{SE}[d]\big{|}\big{|}I_{SE}[n-d]\big{|}\bigg{\\}}$
(42)
## References
* [1] X.-P. Zhang, Advanced modeling of the multicontrol functional static synchronous series compensator (sssc) in newton power flow, IEEE Transactions on Power Systems 18 (4) (2003) 1410–1416.
* [2] X. P. Zhang, Modelling of the interline power flow controller and the generalised unified power flow controller in newton power flow, IEE Proceedings-Generation, Transmission and Distribution 150 (3) (2003) 268–274.
* [3] X. Wei, J. H. Chow, B. Fardanesh, A. A. Edris, A common modeling framework of voltage-sourced converters for load flow, sensitivity, and dispatch analysis, IEEE Transactions on Power Systems 19 (2) (2004) 934–941.
* [4] X. Jiang, X. Fang, J. H. Chow, A. A. Edris, E. Uzunovic, M. Parisi, L. Hopkins, A novel approach for modeling voltage-sourced converter-based facts controllers, IEEE Transactions on Power Delivery 23 (4) (2008) 2591–2598.
* [5] P. Singh, R. Tiwari, Amalgam power flow controller: A novel flexible, reliable and cost effective solution to control power flow, IEEE Transactions on Power Systems (2017).
* [6] A. Trias, The holomorphic embedding load flow method, in: Power and Energy Society General Meeting, 2012 IEEE, IEEE, 2012, pp. 1–8.
* [7] S. Rao, Y. Feng, D. J. Tylavsky, M. K. Subramanian, The holomorphic embedding method applied to the power-flow problem, IEEE Transactions on Power Systems 31 (5) (2016) 3816–3828.
* [8] S. D. Rao, D. J. Tylavsky, Y. Feng, Estimating the saddle-node bifurcation point of static power systems using the holomorphic embedding method, International Journal of Electrical Power & Energy Systems 84 (2017) 1–12.
* [9] I. Wallace, D. Roberts, A. Grothey, K. McKinnon, Alternative pv bus modelling with the holomorphic embedding load flow method, arXiv preprint arXiv:1607.00163 (2016).
* [10] M. K. Subramanian, Y. Feng, D. Tylavsky, Pv bus modeling in a holomorphically embedded power-flow formulation, in: North American Power Symposium (NAPS), 2013, IEEE, 2013, pp. 1–6.
* [11] A. Trias, J. L. Marín, The holomorphic embedding loadflow method for dc power systems and nonlinear dc circuits, IEEE Transactions on Circuits and Systems I: Regular Papers 63 (2) (2016) 322–333.
* [12] C. Liu, B. Wang, F. Hu, K. Sun, C. L. Bak, Real-time voltage stability assessment for load areas based on the holomorphic embedding method, arXiv preprint arXiv:1702.01464 (2017).
* [13] R. Yao, K. Sun, D. Shi, X. Zhang, Voltage stability analysis of power systems with induction motors based on holomorphic embedding, IEEE Transactions on Power Systems 34 (2) (2018) 1278–1288.
* [14] B. V. Rao, F. Kupzog, M. Kozek, Three-phase unbalanced optimal power flow using holomorphic embedding load flow method, Sustainability 11 (6) (2019) 1774.
* [15] R. Yao, Y. Liu, K. Sun, F. Qiu, J. Wang, Efficient and robust dynamic simulation of power systems with holomorphic embedding, IEEE Transactions on Power Systems 35 (2) (2019) 938–949.
* [16] D. Wu, B. Wang, Holomorphic embedding based continuation method for identifying multiple power flow solutions, IEEE access 7 (2019) 86843–86853.
* [17] P. Singh, R. Tiwari, Extended holomorphic embedded load-flow method and voltage stability assessment of power systems, Electric Power Systems Research 185 (2020) 106381.
* [18] D. K. Asl, M. Mohammadi, A. R. Seifi, Holomorphic embedding load flow for unbalanced radial distribution networks with dfig and tap-changer modelling, IET Generation, Transmission & Distribution 13 (19) (2019) 4263–4273.
* [19] F. D. Freitas, A. C. Santos Jr, L. F. J. Fernandes, Y. G. Acle, Restarted holomorphic embedding load-flow model based on low-order padé approximant and estimated bus power injection, International Journal of Electrical Power & Energy Systems 112 (2019) 326–338.
* [20] Q. Li, Numerical Performance of the Holomorphic Embedding Method, Arizona State University, 2018.
* [21] P. S. Sauter, C. A. Braun, M. Kluwe, S. Hohmann, Comparison of the holomorphic embedding load flow method with established power flow algorithms and a new hybrid approach, in: 2017 Ninth Annual IEEE Green Technologies Conference (GreenTech), IEEE, 2017, pp. 203–210.
* [22] H. D. Chiang, T. Wang, H. Sheng, A novel fast and flexible holomorphic embedding method for power flow problems, IEEE Transactions on Power Systems (2017).
* [23] M. B. Kejani, E. Gholipour, Holomorphic embedding load-flow modeling of thyristor-based facts controllers, IEEE Transactions on Power Systems (2017).
* [24] P. Singh, R. Tiwari, Statcom model using holomorphic embedding, IEEE Access 7 (2019) 33075–33086.
* [25] X.-P. Zhang, E. Handschin, M. Yao, Multi-control functional static synchronous compensator (statcom) in power system steady-state operations, Electric power systems research 72 (3) (2004) 269–278.
* [26] P. Singh, N. Senroy, Steady-state model of vsc based facts devices using flexible holomorphic embedding: (sssc and ipfc), Under Review (2020).
* [27] R. D. Zimmerman, C. E. Murillo-Sánchez, R. J. Thomas, Matpower: Steady-state operations, planning, and analysis tools for power systems research and education, IEEE Transactions on power systems 26 (1) (2010) 12–19.
|
# Many-body effects in the excitations and dynamics of trapped Bose-Einstein
condensates
Ofir E. Alon<EMAIL_ADDRESS>Department of Mathematics, University
of Haifa, Haifa 3498838, Israel Haifa Research Center for Theoretical Physics
and Astrophysics, University of Haifa, Haifa 3498838, Israel Raphael Beinke
<EMAIL_ADDRESS>Theoretische Chemie, Physikalisch-
Chemisches Institut, Universität Heidelberg, Im Neuenheimer Feld 229, D-69120
Heidelberg, Germany Lorenz S. Cederbaum<EMAIL_ADDRESS>heidelberg.de Theoretische Chemie, Physikalisch-Chemisches Institut,
Universität Heidelberg, Im Neuenheimer Feld 229, D-69120 Heidelberg, Germany
###### Abstract
This review explores the dynamics and the low-energy excitation spectra of
Bose-Einstein condensates (BECs) of interacting bosons in external potential
traps putting particular emphasis on the emerging many-body effects beyond
mean-field descriptions. To do so, methods have to be used that, in principle,
can provide numerically exact results for both the dynamics and the excitation
spectra in a systematic manner. Numerically exact results for the dynamics are
presented employing the well-established multicongurational time-dependent
Hartree for bosons (MCTDHB) method. The respective excitation spectra are
calculated utilizing the more recently introduced linear-response theory atop
it (LR-MCTDHB). The latter theory gives rise to an, in general, non-hermitian
eigenvalue problem. The theory and its newly developed implementation are
described in detail and benchmarked towards the exactly-solvable harmonic-
interaction model. Several applications to BECs in one- and two-dimensional
potential traps are discussed. With respect to dynamics, it is shown that both
the out-of-equilibrium tunneling dynamics and the dynamics of trapped vortices
are of many-body nature. Furthermore, many-body effects in the excitation
spectra are presented for BECs in different trap geometries. It is
demonstrated that even for essentially-condensed systems, the spectrum of the
lowest-in-energy excitations computed at the many-body level can differ
substantially from the standard mean-field description. In general, it is
shown that bosons carrying angular momentum are more sensitive to many-body
effects than bosons without. These effects are present in both the dynamics
and the excitation spectrum.
###### Contents
1. 1 Introduction
2. 2 General concepts of many-boson physics
1. 2.1 Many-body Hamiltonian and Schrödinger equation
2. 2.2 Second quantization
1. 2.2.1 Fock states and commutation relations
2. 2.2.2 Bosonic field operator
3. 2.2.3 One- and two-body operators
4. 2.2.4 Hamiltonian in second quantization
3. 2.3 Dirac-Frenkel and least action variational principles
4. 2.4 Definition of Bose-Einstein condensation in traps
3. 3 Many-body theory
1. 3.1 Multiconfigurational time-dependent Hartree method for bosons (MCTDHB)
2. 3.2 Many-body linear response
1. 3.2.1 The connection between linear response and excitation spectra
2. 3.2.2 Linear-response MCTDHB (LR-MCTDHB)
4. 4 Numerical implementation of many-body linear response theory
1. 4.1 Code structure and challenges
2. 4.2 Construction of the LR matrix
3. 4.3 Calculation of eigenvalues
1. 4.3.1 The Arnoldi algorithm
2. 4.3.2 The QR-iteration
3. 4.3.3 Polynomial filtering
4. 4.3.4 The implicitly restarted Arnoldi method (IRAM)
5. 5 Comparison with an exactly-solvable model
1. 5.1 The harmonic-interaction model (HIM)
2. 5.2 Ground state and dynamics
3. 5.3 Excited states
6. 6 Applications to many-body dynamics and excitations of BECs
1. 6.1 Tunneling dynamics in traps
1. 6.1.1 Exact tunneling dynamics in a bosonic Josephson junction
2. 6.1.2 Many-body tunneling dynamics in a two-dimensional radial double well
2. 6.2 Dynamical fragmentation
1. 6.2.1 Tunneling to open space
2. 6.2.2 Phantom vortices
3. 6.3 Excitation spectra
1. 6.3.1 One-dimensional harmonic and double-well systems
2. 6.3.2 Triple wells and larger lattices
3. 6.3.3 Rotating Bose-Einstein condensates in an anharmonic trap
7. 7 Summary and Conclusions
8. Appendix A Mean-field theory
1. A.1 Single-orbital approaches
1. A.1.1 The GP equation
2. A.1.2 The BdG equations
2. A.2 Multi-orbital approaches
1. A.2.1 Time-dependent multi-orbital mean field (TDMF)
2. A.2.2 Linear-response best-mean-field (LR-BMF)
9. Appendix B LR-MCTDHB in block-diagonal form
10. Appendix C Further benchmarks of LR-MCTDHB
11. Appendix D Many-particle variance: A sensitive quantity to many-body effects
## 1 Introduction
In 1924, Albert Einstein discovered that the spectral energy distribution in
certain systems of massive particles is different from the Maxwell-Boltzmann
statistics. He was inspired by a work of Satayendra N. Bose who derived the
thermal energy distribution of black-body radiation solely based on
assumptions from quantum theory [1], which is slightly different to the
previous semi-classical approach of Max Planck. Einstein realized the far-
reaching consequences of this, namely that a composite system of massive
particles at sufficiently low temperatures could behave as a matter wave due
to all particles sharing the same single-particle state. Thus, he applied
Bose’s idea to a single-component ideal gas [2, 3]. This paved the way towards
what is known as Bose-Einstein condensation. Consequently, all particles that
obey the statistical distribution found by Bose and Einstein are termed
bosons.
It took approximately 70 years until the first gaseous Bose-Einstein
condensate (BEC) was realized in an experiment. The group of Eric A. Cornell
and Carl E. Wieman produced a repulsive BEC of rubidium atoms [4], shortly
followed by the group of Wolfgang Ketterle using repulsive sodium atoms in a
magneto-optical trap [5]. The latter three scientists obtained the Nobel prize
in physics for this achievement in 2001. The first attractive BEC was realized
by the group of Randall G. Hulet using lithium atoms [6]. In more recent
experiments, condensation was observed in a large variety of systems, e.g.,
for exciton-polariton systems [7, 8], magnons at room temperature [9], or,
rather astonishingly because of the lack of a rest mass and a vanishing
chemical potential, even for photons in a cavity [10].
Over the past two decades, the experimental control of BECs has grown
remarkably. This includes both the interaction of the constituent particles by
employing Feshbach resonances [11, 12, 13] as well as the trap geometry and
dimensionality [14]. This qualifies BECs for testing the fundamental
principles of quantum mechanics, or very recently even to simulate certain
processes in astrophysics [15, 16, 17].
Of particular interest is the understanding of the dynamics as well as the
spectrum of excited states of trapped BECs made of interacting bosons.
Theoretical descriptions for these were carried out many years before they
were finally realized experimentally. The most prominent approach to obtain
the ground state as well as the dynamical motion of condensates is the Gross-
Pitaevskii (GP) equation [18, 19, 20, 21]. In this theory it is assumed that
actually all bosons in the system occupy the same single-particle state,
making it a mean-field theory. The GP equation has been successfully applied
in many cases [22] and can be seen as the adequate starting point to solve
many physical problems of BECs. Similarly, the linear-response (LR) theory
atop the GP equation, resulting in the famous Bogoliubov-de Gennes (BdG)
equations [23, 24], represents the commonly utilized approach to describe
excitations of trapped ultracold bosons, e.g., for dark-bright soliton pairs
[25], in rotating BECs with dipole-dipole interaction [26], or in a 2D
hexagonal lattice with superimposed harmonic confinement [27], to name only a
few recent applications. Furthermore, the BdG equations were extensively used
for the description of Cooper pairs in superconductors [28] or for the
understanding of the BCS-BEC crossover connecting superconductivity and
superfluidity in correlated fermionic systems [29].
A crucial limitation of the GP mean-field theory is the fact that it ignores
correlations between the bosons, and thus does not account for condensate
depletion, i.e., the fraction of particles outside the condensed mode due to
correlations and interactions, even at zero temperature. In a highly-cited
review article from 1999, the condensate depletion is found to be “very small
(less than $1\%$) in the presently available experimental conditions” [30],
and thus the usage of mean-field methods like the BdG approach has been
considered justified. However, the many-body methods developed in the
following years were capable to demonstrate the remarkable impact of depletion
and even fragmentation of a BEC onto its dynamics and its excitation spectrum.
One of these many-body approaches is the multiconfigurational time-dependent
Hartree method for bosons [MCTDHB, or MCTDHB($M$) with $M$ being the number of
utilized single-particle states, called orbitals], which has been introduced
[31, 32], benchmarked [33], and applied successfully to various systems in
recent years. Being derived from a variational principle, its results can be
improved systematically by increasing the number of orbitals $M$.
Additionally, its LR theory, termed LR-MCTDHB($M$), has been introduced
recently [34, 35]. The latter represents a full many-body description of
excitations in trapped BECs.
This review has two main objectives. At first, the newly developed numerical
implementation of LR-MCTDHB, capable of treating large systems also in $D>1$
spatial dimensions which has not been realized before, is introduced and
benchmarked against an exactly-solvable model. Secondly, various applications
of both MCTDHB and LR-MCTDHB to trapped BECs are discussed in which the
depletion has a strong impact on the dynamics and on the spectrum of excited
states. With regard to dynamics, it is shown that fragmentation can develop
over time although a system is initially almost entirely condensed. With
regard to excitation spectra, it is demonstrated that substantial many-body
effects appear even if the ground-state depletion is of the order of the above
mentioned $1\%$ or lower. Even for the case of marginal depletion where one
might expect mean-field theory to accurately describe the physics, it is
instructive to compare the results obtained from the many-body approaches
described above to the ones obtained from the GP and BdG theories. It is
important to note that for only a single orbital, $M=1$, MCTDHB reduces
exactly to the GP equation. Correspondingly, LR-MCTDHB($1$) yields the BdG
equations. Hence, the commonly used mean-field approaches are included in the
utilized many-body theories as their simplest limiting cases.
To address these goals, the review is structured as follows. First, in Section
2, the main theoretical concepts of many-boson physics are presented. This
includes the many-body Schrödinger equation, the framework of second
quantization which is particularly useful for the many-body theory described
in this work, the introduction of the Dirac-Frenkel variational principle from
which the central equations of motion can be derived, as well as the
definition of Bose-Einstein condensation. Second, in Section 3, a detailed
derivation of MCTDHB is given. The essential ingredients like the ansatz for
the many-boson wave function are discussed. Then, its LR theory is derived,
which represents the central theory utilized in this work to calculate low-
energy excitations of trapped bosonic systems on the many-body level. Third,
in Section 4, the newly developed numerical implementation of LR-MCTDHB is
described in detail. Therefore, general technical challenges and their
solutions are explained. Furthermore, a detailed description of the numerical
algorithms utilized is presented. Subsequently, in Section 5, the used
implementations of MCTDHB and LR-MCTDHB are benchmarked against the exactly-
solvable harmonic-interaction model, whose analytic solutions of the energies
of the ground and excited states are presented beforehand. Applications of
both implementations are then shown and their physics discussed in Section 6.
Concerning the dynamics, the focus is twofold. First, applications dealing
with the phenomenon of tunneling of BECs between wells inside potential traps
in one and two spatial dimensions are discussed, demonstrating the many-body
nature of quantum tunneling of ultracold BECs. Second, several cases of
dynamical fragmentation of initially coherent bosonic systems are presented.
For instance, the tunneling of a BEC to open space is analyzed. Furthermore,
the connection between angular momentum, especially of quantized vortices, and
the development of fragmentation is investigated. Concerning applications of
LR-MCTDHB, the many-body effects in the low-energy spectra of BECs in one-
dimensional multi-well traps and lattices are shown. Moreover, as a 2D
application, the impact of angular momentum on the lowest-in-energy
excitations of a rotating BEC is examined. Essential physical results and
implications are provided. Finally, in Section 7, a summary and an outlook of
possible constituent future research as well as of further enrichment of the
implementation of LR-MCTDHB is provided.
Additionally, the review contains four appendices which broaden its scope and
provide additional insight. In Appendix A, the commonly-used mean-field
theories, the GP theory and the constituent LR theory leading to the BdG
equations, are described. A further mean-field approach for both the statics
and dynamics of trapped BECs that has been developed prior to the many-body
theories of Section 3 is presented, called the time-dependent multi-orbital
mean field (TDMF) and the linear-response best-mean-field (LR-BMF). The latter
can be seen as intermediate theories between the standard single-orbital mean-
field approaches and the full many-body treatment and have their own merits.
Appendix B demonstrates that LR-MCTDHB can be written in block-diagonal form,
which, in some cases, can be beneficial. Advantages and possible weaknesses
are discussed. In Appendix C, additional results of the LR-MCTDHB benchmark
are provided. Finally, in Appendix D, the many-body variance, a quantity that
turns out to be very sensitive to many-body effects, is introduced and derived
for the example of the center-of-mass (c.m.) position of a BEC. Its relevance
is discussed in several examples.
## 2 General concepts of many-boson physics
In this section, the general concepts of describing trapped ultracold bosons
are introduced. This includes the representation of the many-body Hamiltonian,
the bosonic field as well as properties of the wave function for a system of
identical bosons. Moreover, the definition of Bose-Einstein condensation and
depletion which is used in the subsequent sections is presented.
### 2.1 Many-body Hamiltonian and Schrödinger equation
The time-dependent Schrödinger equation for a generic system of $N$ identical
spinless bosons in $D\leq 3$ dimensions is given by
$\hat{H}\Psi(\mathbf{r}_{1},\mathbf{r}_{2},...,\mathbf{r}_{N};t)=i\frac{\partial}{\partial
t}\Psi(\mathbf{r}_{1},\mathbf{r}_{2},...,\mathbf{r}_{N};t)$ (1)
with the full many-body wave function
$\Psi(\mathbf{r}_{1},\mathbf{r}_{2},...,\mathbf{r}_{N};t)$ depending on the
spatial coordinates $\\{\mathbf{r}_{i}\\}$ of all $N$ bosons and the time $t$.
The Hamiltonian $\hat{H}$ reads
$\hat{H}=\sum_{i=1}^{N}\hat{h}(\mathbf{r}_{i})+\lambda_{0}\sum_{j>i}^{N}\hat{W}(\mathbf{r}_{i},\mathbf{r}_{j})$
(2)
with the single-particle term
$\displaystyle\hat{h}(\mathbf{r}_{i})$
$\displaystyle=\hat{T}(\mathbf{r}_{i})+V(\mathbf{r}_{i})$ (3)
$\displaystyle\hat{T}(\mathbf{r}_{i})$
$\displaystyle=-\frac{1}{2}\Delta_{i}=-\frac{1}{2}\frac{\partial^{2}}{\partial\mathbf{r}_{i}^{\,2}}$
(4)
comprising the contributions of the kinetic energy $\hat{T}$ and the external
potential $V$ which is assumed to be real. For simplicity, the terms are given
in dimensionless units with $\hbar=m=1$. The two-body operator
$\hat{W}(\mathbf{r_{i}},\mathbf{r_{j}})$ represents the interaction potential
of strength $\lambda_{0}$. In this work, interaction potentials which solely
depend on the distance $r_{ij}=|\mathbf{r_{i}}-\mathbf{r_{j}}|$ between two
bosons are considered. In general, both $V$ and $\hat{W}$ can be time-
dependent.
### 2.2 Second quantization
In this section, the formalism of second quantization is briefly explained,
including the Fock state representation of an $N$-boson state, the symmetry
properties of the wave function, and the Hamiltonian in second quantization.
#### 2.2.1 Fock states and commutation relations
As a consequence of the spin-statistics theorem [36, 37], the wave function
$\Psi(\mathbf{r}_{1},...,\mathbf{r}_{N})$ of $N$ identical bosons is a fully
symmetrized $N$-particle state, i.e., it is symmetric under the permutation of
two bosons with coordinates $\mathbf{r}_{i}$ and $\mathbf{r}_{j}$, i.e.,
$\Psi(...,\mathbf{r}_{i},...,\mathbf{r}_{j},...)=+\Psi(...,\mathbf{r}_{j},...,\mathbf{r}_{i},...).$
(5)
In second quantization, this is achieved by using a Fock state (or number
state) basis of the $N$-particle Hilbert space $\mathcal{H}^{(N)}$ for the
wave function $\Psi$. This basis incorporates the previously mentioned
symmetrization requirements as well as the commutation relations of identical
bosons shown below.
For a complete set of orthonormalized single-particle states, also referred to
as orbitals, $\\{\phi_{i}:i=1,...,\infty\\}$, a general Fock state reads
$|n_{1},n_{2},...\rangle=\prod_{i}\left[\frac{1}{\sqrt{n_{i}!}}\left(\hat{b}_{i}^{\dagger}\right)^{n_{i}}\right]|\text{vac}\rangle$
(6)
where the integers $\\{n_{i}\\}$ denote the occupation numbers of the
individual orbitals and $|\text{vac}\rangle$ the particle vacuum. In Eq. (6),
the creation and annihilation operators $\hat{b}_{i}^{\dagger}$ and
$\hat{b}_{i}$, which create or annihilate a particle in the state $\phi_{i}$,
were introduced. The action of these operators on a Fock state is given by
$\displaystyle\hat{b}_{i}\,|n_{1},n_{2},...,n_{i},...\rangle=$
$\displaystyle\sqrt{n_{i}}\,|n_{1},n_{2},...,n_{i}-1,...\rangle$
$\displaystyle\hat{b}_{i}^{\dagger}\,|n_{1},n_{2},...,n_{i},...\rangle=$
$\displaystyle\sqrt{n_{i}+1}\,|n_{1},n_{2},...,n_{i}+1,...\rangle$
and thus the number operator for any single-particle state $\phi_{i}$ reads
$\hat{n}_{i}=\hat{b}_{i}^{\dagger}\hat{b}_{i}.$ (7)
For a system with a fixed amount of bosons $N$, the occupation numbers of the
orbitals sum up to the total number of particles $N$, i.e.,
$N=\sum_{i}\,n_{i}.$ (8)
The creation and annihilation operators fulfill the conventional bosonic
commutation relations
$\left[\hat{b}_{i},\hat{b}_{k}\right]=0,\quad\quad\left[\hat{b}^{\dagger}_{i},\hat{b}^{\dagger}_{k}\right]=0,\quad\quad\left[\hat{b}_{i},\hat{b}_{k}^{\dagger}\right]=\delta_{ik}$
(9)
which ensure the exchange symmetry of the Fock state basis in Eq. (6) and thus
of the total wave function $\Psi$ as required in Eq. (5).
#### 2.2.2 Bosonic field operator
Utilizing the creation and annihilation operators of bosons for single-
particle states from the previous section, one can define operators that
create or annihilate a boson at position $\mathbf{r}$ at time $t$. These
operators are the so-called field operators,
$\displaystyle\hat{\Psi}^{\dagger}(\mathbf{r})=\sum_{i}\hat{b}^{\dagger}_{i}(t)\phi_{i}^{\ast}(\mathbf{r},t)$
(10)
$\displaystyle\hat{\Psi}(\mathbf{r})=\sum_{i}\hat{b}_{i}(t)\phi_{i}(\mathbf{r},t),$
(11)
where it is taken into account that the orbitals as well as the creation and
annihilation operators can in general be time-dependent. Thus,
$\hat{b}^{\dagger}(t)$ and $\hat{b}(t)$ can be expressed in terms of the field
operators,
$\hat{b}_{i}^{\dagger}(t)=\int\phi_{i}(\mathbf{r},t)\hat{\Psi}^{\dagger}(\mathbf{r})d\mathbf{r},\quad\quad\hat{b}_{i}(t)=\int\phi_{i}^{\ast}(\mathbf{r},t)\hat{\Psi}(\mathbf{r})d\mathbf{r}$
(12)
where it is assumed that the set $\\{\phi_{i}(\mathbf{r},t)\\}$ consists of
orthonormalized functions. In the following, the time argument in the above
quantities is suppressed. The field operators in Eqs. (10) and (11) fulfill
the bosonic commutation relations
$\left[\hat{\Psi}(\mathbf{r}),\hat{\Psi}(\mathbf{r}^{\,\prime})\right]=0,\quad\quad\left[\hat{\Psi}^{\dagger}(\mathbf{r}),\hat{\Psi}^{\dagger}(\mathbf{r}^{\,\prime})\right]=0,\quad\quad\left[\hat{\Psi}(\mathbf{r}),\hat{\Psi}^{\dagger}(\mathbf{r}^{\,\prime})\right]=\delta\left(\mathbf{r}-\mathbf{r}^{\,\prime}\right).$
(13)
The density operator is defined as
$\hat{\rho}(\mathbf{r})=\hat{\Psi}^{\dagger}(\mathbf{r})\hat{\Psi}(\mathbf{r})$
(14)
such that the operator of the total number of bosons in the system is given by
$\hat{N}=\int\,d\mathbf{r}\,\hat{\rho}(\mathbf{r}).$ (15)
#### 2.2.3 One- and two-body operators
In second quantization, a general one-body operator $\hat{A}^{(1)}$, i.e., an
operator acting on a single particle, reduces to the form
$\hat{A}^{(1)}=\sum_{i,j}a_{ij}\,\hat{b}_{i}^{\dagger}\hat{b}_{j}$ (16)
where the matrix elements $a_{ij}$ are given by
$a_{ij}=\int\phi_{i}^{\ast}(\mathbf{r})\hat{A}^{(1)}\,\phi_{j}(\mathbf{r})\,d\mathbf{r}$
(17)
or, equivalently in bra-ket notation, $a_{ij}=\langle
i|\hat{A}^{(1)}|j\rangle$. Similarly, the general form of a two-body operator,
i.e., an operator acting on two particles, is given by
$\hat{A}^{(2)}=\frac{1}{2}\sum_{i,j,k,l}a_{ijkl}\,\hat{b}_{i}^{\dagger}\hat{b}^{\dagger}_{j}\hat{b}_{k}\hat{b}_{l}$
(18)
with
$a_{ijkl}=\iint\phi_{i}^{\ast}(\mathbf{r})\phi_{j}^{\ast}(\mathbf{r}^{\,\prime})\hat{A}^{(2)}\,\phi_{k}(\mathbf{r})\phi_{l}(\mathbf{r}^{\,\prime})\,d\mathbf{r}d\mathbf{r}^{\,\prime},$
(19)
or $a_{ijkl}=\langle i,j|\hat{A}^{(2)}|k,l\rangle$ in short. The factor $1/2$
in Eq. (18) accounts for double counting of the individual terms.
#### 2.2.4 Hamiltonian in second quantization
The Hamiltonian in Eq. (2) can now be expressed within the framework of second
quantization. Using the bosonic field operators in Eqs. (10) and (11) as well
as the general expressions for one- and two-body operators, Eqs. (16) and
(18), one obtains
$\displaystyle\hat{H}$
$\displaystyle=\sum_{i=1}^{N}\hat{h}(\mathbf{r}_{i})+\lambda_{0}\sum_{j>i}^{N}\hat{W}(\mathbf{r}_{i},\mathbf{r}_{j})\Rightarrow$
$\displaystyle\hat{H}$ $\displaystyle=\int
d\mathbf{r}\left(\hat{\Psi}^{\dagger}(\mathbf{r})\hat{h}(\mathbf{r})\hat{\Psi}(\mathbf{r})+\frac{\lambda_{0}}{2}\int
d\mathbf{r}^{\,\prime}\hat{\Psi}^{\dagger}(\mathbf{r})\hat{\Psi}^{\dagger}(\mathbf{r}^{\,\prime})\hat{W}(\mathbf{r},\mathbf{r}^{\,\prime})\hat{\Psi}(\mathbf{r})\hat{\Psi}(\mathbf{r}^{\,\prime})\right)$
$\displaystyle=\sum_{i,j}h_{ij}\,\hat{b}_{i}^{\dagger}\hat{b}_{j}+\frac{\lambda_{0}}{2}\sum_{i,j,k,l}W_{ijkl}\,\hat{b}_{i}^{\dagger}\hat{b}_{j}^{\dagger}\hat{b}_{k}\hat{b}_{l}$
(20)
with the matrix elements
$\displaystyle h_{ij}$
$\displaystyle=\int\phi_{i}^{\ast}(\mathbf{r})\,\hat{h}(\mathbf{r})\,\phi_{j}(\mathbf{r})\,d\mathbf{r}$
(21) $\displaystyle W_{ijkl}$
$\displaystyle=\iint\,\phi_{i}^{\ast}(\mathbf{r})\phi_{j}^{\ast}(\mathbf{r}^{\,\prime})\hat{W}(\mathbf{r},\mathbf{r}^{\,\prime})\,\phi_{k}(\mathbf{r})\phi_{l}(\mathbf{r}^{\,\prime})\,d\mathbf{r}\,d\mathbf{r}^{\,\prime}.$
(22)
To keep the notation simple, the time argument of the quantities in Eqs.
(20)-(22) is suppressed. The above representation of $\hat{H}$ from Eq. (20)
will be used in the many-body formalism presented in Section 3.
### 2.3 Dirac-Frenkel and least action variational principles
The working equations of all utilized many-body methods in this work can be
derived from the Dirac-Frenkel variational principle (DFVP) [38, 39, 40]. Like
for any other variational method, it yields a condition for the dynamical
evolution for any parametrized ansatz of the wave function $\Psi$. Within the
DFVP, the equation that determines the dynamics is given by
$\left\langle\delta\Psi(t)\left|\hat{H}-i\hbar\frac{\partial}{\partial\,t}\right|\Psi(t)\right\rangle=0$
(23)
which basically means that any allowed variation $|\delta\Psi(t)\rangle$ is
orthogonal to the residual wave function
$\left(\hat{H}-i\hbar\frac{\partial}{\partial\,t}\right)|\Psi(t)\rangle$. It
is important to note that under certain conditions, the DFVP is fully
equivalent to two other commonly used time-dependent variational principles,
which are the variational principle due to MacLachlan [41] and the least
action principle. See in this context Ref. [42]. The least action principle is
based on the assumption that the wave function minimizes the action
$S=\int dt\,\left\langle\Psi\left|\hat{H}-i\hbar\frac{\partial}{\partial
t}\right|\Psi\right\rangle,$ (24)
meaning that its variation with respect to $\Psi$ should vanish, i.e.,
$\delta S=0.$ (25)
Eqs. (24) and (25), in combination with normalization constraints, are used to
derive the equations of motion for MCTDHB. The full derivation of the latter
theory is carried out in detail in Section 3.1. Furthermore, the least action
principle can also be employed to derive the standard mean-field equation for
the ground state and dynamics of a trapped BEC, i.e., the GP equation. Its
derivation is presented in Appendix A.1.1.
### 2.4 Definition of Bose-Einstein condensation in traps
In 1956, Penrose and Onsager developed a theoretical criterion which is
utilized in the definition of Bose-Einstein condensation and is based on
standard quantities from statistical quantum mechanics [43]. It makes use of
the one-body reduced density matrix (RDM) of an $N$-boson system which can be
defined via the first order correlation function
$\displaystyle\mathbf{\rho}^{(1)}(\mathbf{r}|\mathbf{r}^{\,\prime})$
$\displaystyle=\left\langle\Psi\left|\hat{\Psi}^{\dagger}(\mathbf{r}^{\,\prime})\hat{\Psi}(\mathbf{r})\right|\Psi\right\rangle$
$\displaystyle=\sum_{i,j}\,\rho_{ij}\,\phi_{i}^{\ast}(\mathbf{r}^{\,\prime})\phi_{j}(\mathbf{r})$
(26)
with
$\rho_{ij}=\langle\Psi|\hat{\rho}_{ij}|\Psi\rangle,\quad\hat{\rho}_{ij}=\hat{b}_{i}^{\dagger}\hat{b}_{j}$
(27)
where the time argument is suppressed in all quantities. More details on RDMs
are given in Appendix D. For $\mathbf{r}=\mathbf{r}^{\,\prime}$, Eq. (2.4)
denotes the one-particle density. Diagonalizing the one-body RDM yields
$\mathbf{\rho}^{(1)}(\mathbf{r}|\mathbf{r}^{\,\prime})=\sum_{i}n_{i}^{(1)}\,\alpha_{i}^{\ast}(\mathbf{r}^{\,\prime})\alpha_{i}(\mathbf{r})$
(28)
where the eigenvalues $\left\\{n_{i}^{(1)}|n_{1}^{(1)}\geq
n_{2}^{(1)}\geq...\right\\}$ and the eigenfunctions
$\\{\alpha_{i}(\mathbf{r})\\}$ are the natural occupation numbers and the
natural orbitals, respectively. Whenever possible, the superscript ’$(1)$’ of
the natural occupation numbers will be omitted throughout this work. According
to Ref. [43], any $N$-boson system where only the largest eigenvalue is
macroscopic, i.e., $n_{1}=\mathcal{O}(N)$, is said to be condensed. However,
if more than one eigenvalue of the one-body RDM are macroscopic, the system is
said to be depleted/fragmented [44, 45]. In this work, the degree of depletion
(or fragmentation) is quantified by
$f=\frac{1}{N}\sum_{i>1}n_{i}$ (29)
which is the aggregated occupation of all but the first natural orbital.
Another classification that uses the above relation distinguishes the terms
depletion and fragmentation more clearly and was used in Ref. [46].
In Section 6, the dynamics and excited states of BECs in different trap
geometries in one and two spatial dimensions are studied. There, the degree of
depletion/fragmentation will be an essential quantity to distinguish the many-
body results obtained from the methods described in the subsequent Section 3
from the mean-field results obtained by using the theories described in the
Appendices A.1.1 and A.1.2.
## 3 Many-body theory
In this section, the many-body approach utilized in this work to compute the
ground state and the dynamics of trapped BECs as well as their low-energy
excitation spectra is presented. First, an introduction to MCTDHB is made in
Section 3.1, followed by the application of LR theory atop it, termed LR-
MCTDHB, in Section 3.2.2.
### 3.1 Multiconfigurational time-dependent Hartree method for bosons
(MCTDHB)
In this section, a detailed derivation of the MCTDHB [or MCTDHB($M$)] theory
which has been originally introduced in Refs. [31, 32], is presented. It
originates from the closely related MCTDH method [47, 48, 49] which is a
widely used technique to compute the dynamics and excited states of nuclei in
molecules. Further extensions of MCTDHB to the case of type conversion of
particles, to mixtures of bosons and fermions, to Hubbard-type Hamiltonians of
BECs on lattices, as well as to BECs with internal degrees of freedom can be
found in Refs. [50, 51, 52, 53]. Other many-body theories for the ground state
and excitations of BECs were utilized as well. One example is the density
matrix renormalization group (DMRG) theory [54]. It is particularly designed
for applications in the field of one-dimensional condensed matter physics, but
was extended beyond this to quantum chemistry and quantum information. An
early review can be found in [55]. Furthermore, DMRG has been extended and
applied to two-dimensional systems as well [56]. However, this review deals
with MCTDHB and its LR theory.
The ansatz for the many-body wave function to solve the many-boson Schrödinger
equation, Eq. (1), is given by
$|\Psi(t)\rangle=\sum_{\mathbf{n}}C_{\mathbf{n}}(t)\,|\mathbf{n};t\rangle,$
(30)
which is a superposition of permanents $\\{|\mathbf{n};t\rangle\\}$ comprised
of $M$ single-particle orbitals $\\{\phi_{j}(\mathbf{r},t):1\leq j\leq M\\}$
and expansion coefficients $\\{C_{\mathbf{n}}(t)\\}$ where
$\mathbf{n}=(n_{1},\ldots,n_{M})^{t}$ is a vector carrying the individual
occupation numbers of the orbitals for a given permanent. The summation in Eq.
(30) includes all $N_{\text{conf}}=\binom{N+M-1}{N}$ possibilities to
distribute $N$ bosons onto $M$ single-particle orbitals. It is important to
note that both the permanents and the coefficients are time-dependent, which
is the main difference compared to the frequently used many-body approach of
the full configuration interaction (FCI) method. In the latter theory, the
amount of configurations for a given number $M$ of single-particle basis
functions is the same as for MCTDHB. However, these basis states are fixed in
shape, i.e., they remain unchanged and only the coefficients are time-
dependent. For MCTDHB, the single-particle orbitals are time-adaptive and, as
shown below, the corresponding set of equations of motion (EOMs) is coupled to
the set of EOMs of the coefficients. In terms of numerical convergence,
utilizing a time-adaptive instead of a shape-fixed basis dramatically improves
the convergence towards exact results for both the ground state and the
dynamics of trapped BECs, as discussed in Section 5.
The working equations of MCTDHB are determined by the least action principle
of Eqs. (24) and (25), which explicitly reads
$0=\delta S=\delta\left(\int dt\,L(t)\right)$ (31)
with
$L(t)=\left\langle\Psi(t)\left|\hat{H}-i\frac{\partial}{\partial
t}\right|\Psi(t)\right\rangle-\sum_{i,j}\mu_{ij}(t)[\langle\phi_{i}|\phi_{j}\rangle-\delta_{ij}]$
(32)
and $\hbar=1$. The time-dependent Langrange multipliers $\\{\mu_{ij}(t)\\}$
account for the orthonormalization of the orbitals in time. Whenever possible,
the time-argument of all quantities is suppressed in the following. Plugging
in the ansatz for the wave function, Eq. (30), and the expression for the
Hamiltonian, Eq. (20), into Eq. (32) results in
$\displaystyle L(t)$
$\displaystyle=\sum_{k,q}^{M}\rho_{kq}\left[h_{kq}-\left(i\frac{\partial}{\partial
t}\right)_{kq}\right]+\frac{1}{2}\sum_{k,s,q,l}^{M}\rho_{ksql}W_{ksql}$
$\displaystyle-i\sum_{\mathbf{n}}C_{\mathbf{n}}^{\ast}(t)\frac{\partial
C_{\mathbf{n}}(t)}{\partial
t}-\sum_{k,q}\mu_{kq}[\langle\phi_{k}|\phi_{q}\rangle-\delta_{kq}]$ (33)
where the elements of the one- and two-body RDMs in this case read
$\displaystyle\rho_{kq}$
$\displaystyle=\sum_{\mathbf{n},\mathbf{n}^{\,\prime}}C_{\mathbf{n}}^{\ast}(t)C_{\mathbf{n}^{\,\prime}}(t)\left\langle\mathbf{n};t\left|\hat{b}_{k}^{\dagger}\hat{b}_{q}\right|\mathbf{n}^{\,\prime};t\right\rangle$
(34) $\displaystyle\rho_{ksql}$
$\displaystyle=\sum_{\mathbf{n},\mathbf{n}^{\,\prime}}C_{\mathbf{n}}^{\ast}(t)C_{\mathbf{n}^{\,\prime}}(t)\left\langle\mathbf{n};t\left|\hat{b}_{k}^{\dagger}\hat{b}_{s}^{\dagger}\hat{b}_{q}\hat{b}_{l}\right|\mathbf{n}^{\,\prime};t\right\rangle.$
(35)
The variation of $S$ with respect to $\phi_{k}^{\ast}(\mathbf{r},t)$ yields
$\frac{\delta
S}{\delta\phi_{k}^{\ast}(\mathbf{r},t)}=\sum_{q=1}^{M}\left[\rho_{kq}\left(\hat{h}-i\frac{\partial}{\partial
t}\right)-\mu_{kq}+\sum_{s,l=1}^{M}\rho_{ksql}\hat{W}_{sl}\right]|\phi_{q}\rangle$
(36)
with
$\hat{W}_{sl}(\mathbf{r},t)=\int\phi_{s}^{\ast}(\mathbf{r}^{\,\prime},t)\,\hat{W}(\mathbf{r},\mathbf{r}^{\,\prime})\,\phi_{l}(\mathbf{r}^{\,\prime},t)\,d\mathbf{r}^{\,\prime}.$
(37)
According to the DFVP, this variation should vanish, which determines the EOMs
of the orbitals given by
$\sum_{q=1}^{M}\left[\rho_{kq}\hat{h}-\mu_{kq}+\sum_{s,l=1}^{M}\rho_{ksql}\hat{W}_{sl}\right]|\phi_{q}\rangle=i\sum_{q=1}^{M}\rho_{kq}\frac{\partial}{\partial
t}|\phi_{q}\rangle.$ (38)
Taking the scalar product with $\langle\phi_{i}|$ in Eq. (38) results in the
expression for the Lagrange multipliers,
$\mu_{ki}(t)=\sum_{q=1}^{M}\left(\rho_{kq}\left[h_{iq}-\left(i\frac{\partial}{\partial
t}\right)_{iq}\right]+\sum_{s,l=1}^{M}\rho_{ksql}W_{isql}\right).$ (39)
By reinserting them into Eq. (38), multiplying by the inverse of the one-body
RDM $\\{\boldsymbol{\rho}^{-1}\\}_{ik}$ from the left-hand side (LHS), and
summing over $k$, one arrives at
$i\hat{\mathbf{P}}\frac{\partial}{\partial
t}|\phi_{i}\rangle=\hat{\mathbf{P}}\left[\hat{h}|\phi_{i}\rangle+\sum_{k,s,l,q=1}^{M}\\{\boldsymbol{\rho}^{-1}\\}_{ik}\,\rho_{ksql}\hat{W}_{sl}|\phi_{q}\rangle\right]$
(40)
with
$\hat{\mathbf{P}}=\mathbb{1}-\sum_{j^{\prime}=1}^{M}|\phi_{j^{\prime}}\rangle\langle\phi_{j^{\prime}}|$
(41)
being a projection operator on the tangential space of
$\text{span}\left(\phi_{1},...,\phi_{M}\right)$. Due to invariance properties
of the ansatz for the wave function in Eq. (30), the orbitals can be chosen
without loss of generality to be orthogonal to their corresponding time-
derivatives [47, 48], i.e.,
$\langle\phi_{i}|\dot{\phi_{j}}\rangle=0\quad\forall i,j=1,...,M.$ (42)
The time evolution of the orbitals is therefore fully orthogonal and ensures
the orthonormality of the orbitals in time. Hence, one can further simplify
Eq. (40) to read
$i\frac{\partial}{\partial
t}|\phi_{i}\rangle=\hat{\mathbf{P}}\left[\hat{h}|\phi_{i}\rangle+\sum_{k,s,l,q=1}^{M}\\{\boldsymbol{\rho}^{-1}\\}_{ik}\,\rho_{ksql}\hat{W}_{sl}|\phi_{q}\rangle\right],$
(43)
and Eq. (43) is referred to as the orbitals’ EOM in the following.
With respect to the coefficients, the least action principle
$\displaystyle\frac{\delta S}{\delta C_{\mathbf{n}}^{\ast}(t)}=0$ (44)
leads to
$\sum_{\mathbf{n}^{\,\prime}}\left\langle\mathbf{n};t\left|\hat{H}-i\frac{\partial}{\partial
t}\right|\mathbf{n}^{\,\prime};t\right\rangle
C_{\mathbf{n}^{\,\prime}}(t)=i\frac{\partial}{\partial t}C_{\mathbf{n}}(t).$
(45)
Thus, the EOM of the coefficients is given by
$\mathbf{\mathcal{H}}(t)\mathbf{C}(t)=i\frac{\partial}{\partial
t}\mathbf{C}(t)$ (46)
with the matrix elements
$\mathcal{H}_{\mathbf{n}\mathbf{n}^{\,\prime}}=\left\langle\mathbf{n};t\left|\hat{H}-i\frac{\partial}{\partial
t}\right|\mathbf{n}^{\,\prime};t\right\rangle$ (47)
and $\mathbf{C}(t)$ being a vector whose entries are the $N_{\text{conf}}$
coefficients $\\{C_{\mathbf{n}}(t)\\}$. It is worth noting that
$\mathbf{\mathcal{H}}(t)$ is hermitian. Eq. (46) is referred to as the
coefficients’ EOM in the following.
One observes that the sets of working equations for the orbitals and
coefficients are coupled to each other. With regard to the orbitals’ EOM, the
coefficients appear in the expressions for the two-body RDM as well as in the
inverse of the one-body RDM, see Eqs. (34), (35) and (43), respectively. With
regard to the coefficients’ EOM, the implicit dependence on the orbitals is
incorporated in the matrix elements of Eq. (47) since the permanents appearing
therein are assembled by the orbitals.
By employing imaginary time-propagation, i.e., by setting $it\rightarrow\tau$
in the working equations (43) and (46), the MCTDHB theory reduces to the
multiconfigurational Hartree for bosons (MCHB) method which has been
introduced in Ref. [57]. In that way, the ground state of a given system can
be computed very efficiently, as will be shown in Section 5.2.
It can be inferred from Eq. (46) that the evolution of $\mathbf{C}(t)$ is
unitary and therefore its normalization is conserved in time. However, in
comparison to the orbitals’ EOMs in Eq. (43), the coefficient vector
$\mathbf{C}(t)$ does not automatically propagate in an orthogonal manner. The
latter can be achieved by introducing a time-dependent phase of the form
$\mathbf{C}(t)\rightarrow\mathbf{C}(t)e^{-i\int^{t}dt^{\prime}\mathbf{C^{\dagger}}(t^{\prime})\mathbf{\mathcal{H}}(t^{\prime})\mathbf{C}(t^{\prime})}$
(48)
which, by inserting it into Eq. (46), yields
$i\frac{\partial}{\partial
t}\mathbf{C}(t)=\mathbf{\mathcal{P}_{c}}\mathbf{\mathcal{H}}(t)\mathbf{C}(t)$
(49)
with the projector that maps onto the orthogonal space of the coefficients is
given by
$\mathbf{\mathcal{P}_{c}}=\mathbb{1}-\mathbf{C}(t)\mathbf{C^{\dagger}}(t).$
(50)
Including the projector $\mathbf{\mathcal{P}_{c}}$ is however redundant in the
sense that the evolution of the coefficient vector is unitary already without
it, see Eq. (46). On the contrary, this is not the case for the orbitals,
where the projector $\hat{\mathbf{P}}$ that stems from the Lagrange
multipliers in Eq. (31) ensures that the orbitals stay normalized in time.
Henceforth, only the latter projector is considered in the subsequent
derivation of LR-MCTDHB (Section 3.2.2), whereas the coefficients’ projector
$\mathbf{\mathcal{P}_{c}}$ is set to unity in the following without any loss
of generality.
It is important to note that for a single orbital, i.e., $M=1$, the MCTDHB
theory reduces to the GP mean-field equation [18, 19, 20, 21], meaning that
GP$\equiv$MCTDHB(1). The derivation of the latter theory using the variational
principle is described in Appendix A.1.1.
### 3.2 Many-body linear response
#### 3.2.1 The connection between linear response and excitation spectra
Before carrying out the derivation of the many-body LR theory atop MCTDHB, a
rather general perspective of applying LR to the time-dependent Schrödinger
equation, Eq. (1), in order to calculate the excitation spectrum of a trapped
BEC is discussed. The goal is to demonstrate that in general, a LR analysis
atop an exact eigenstate of the unperturbed Hamiltonian $\hat{H}^{0}$,
typically the ground state, yields the exact excitation spectrum of the many-
particle system. The following derivation is closely related to the one in
Ref. [35].
One considers the time-dependent Hamiltonian
$\displaystyle\hat{H}(t)$ $\displaystyle=\hat{H}^{0}+\hat{H}_{\text{ext}}(t),$
(51) $\displaystyle\hat{H}_{\text{ext}}(t)$
$\displaystyle=\hat{f}^{+}(\mathbf{r})e^{-i\omega
t}+\hat{f}^{-}(\mathbf{r})e^{i\omega t}$ (52)
where a weak time-dependent external field with amplitudes $\hat{f}^{+}$ and
$\hat{f}^{-}$ and oscillation frequency $\omega$ is added to the stationary
hermitian Hamiltonian $\hat{H}^{0}$. The latter has eigenenergies
$E_{k},\,k\in\mathbb{N}$, and the ground-state energy is denoted by
$\varepsilon^{0}$. The projected time-dependent Schrödinger equation with
$\hbar=1$,
$\hat{\mathbf{P}}_{\Psi}\hat{H}(t)\Psi(t)=i\dot{\Psi}(t),\quad\hat{\mathbf{P}}_{\Psi}=\mathbb{1}-|\Psi(t)\rangle\langle\Psi(t)|,$
(53)
which can be obtained by introducing the phase
$\Psi(t)\rightarrow\Psi(t)e^{-i\int^{t}dt^{\prime}\langle\Psi(t^{\prime})|\hat{H}(t^{\prime})\rangle\Psi(t^{\prime})}$
(54)
to the system’s wave function $\Psi(t)$ and inserting it into Eq. (1), ensures
that the time evolution is orthogonal in the sense that
$\langle\Psi(t)|\dot{\Psi}(t)\rangle=0$ (55)
at any time $t$. To solve Eq. (53), the ansatz
$\Psi(\mathbf{r},t)=e^{-\varepsilon^{0}t}\left(\Psi_{0}(\mathbf{r})+u(\mathbf{r})\,e^{-i\omega
t}+v^{\ast}(\mathbf{r})\,e^{i\omega t}\right)$ (56)
with is employed. The ground state $\Psi_{0}$ obtained by
$\hat{H}^{0}\Psi_{0}=\varepsilon^{0}\Psi_{0}$ as well as the correction
amplitudes $u$ and $v$ are assumed to be stationary. The attached time-
dependent oscillations have the same frequency $\omega$ as the external
perturbing field. Utilizing the orthogonality condition, Eq. (55), one obtains
$\displaystyle\langle\Psi_{0}|u\rangle=0\quad\text{and}\quad\langle\Psi_{0}|v^{\ast}\rangle=0,$
(57)
meaning that the correction amplitudes are orthogonal to the ground-state wave
function and thus
$\displaystyle\hat{\mathbf{P}}_{\Psi_{0}}|u\rangle=|u\rangle\quad\text{and}\quad\hat{\mathbf{P}}_{\Psi_{0}}|v^{\ast}\rangle=|v^{\ast}\rangle$
(58)
for the operator projecting onto the tangential space of $\Psi_{0}$, i.e.,
$\hat{\mathbf{P}}_{\Psi_{0}}=\mathbb{1}-|\Psi_{0}\rangle\langle\Psi_{0}|$.
Inserting the ansatz (56) into Eq. (53), one arrives at
$\displaystyle\hat{\mathbf{P}}_{\Psi_{0}}\hat{H}^{0}(u\,e^{-i\omega
t}+v^{\ast}\,e^{i\omega t})+\hat{\mathbf{P}}_{\Psi_{0}}(\hat{f}^{+}e^{-i\omega
t}+\hat{f}^{-}e^{i\omega t})\Psi_{0}$
$\displaystyle=(\omega+\varepsilon^{0})(u\,e^{-i\omega t}-v^{\ast}\,e^{i\omega
t})$ (59)
where only terms linear to $u$ and $v$ are kept. This can be written in matrix
form by ordering terms proportional to $e^{-i\omega t}$ and $e^{i\omega t}$,
respectively. The result reads
$\left[\begin{pmatrix}\hat{\mathbf{P}}_{\Psi_{0}}(\hat{H}^{0}-\varepsilon^{0})&0\\\
0&-\hat{\mathbf{P}}^{\ast}_{\Psi_{0}}(\hat{H}^{0,\ast}-\varepsilon^{0})\end{pmatrix}-\omega\right]\begin{pmatrix}u\\\
v\end{pmatrix}=\begin{pmatrix}-\hat{\mathbf{P}}_{\Psi_{0}}\hat{f}^{+}\,\Psi_{0}\\\
\hat{\mathbf{P}}^{\ast}_{\Psi_{0}}\hat{f}^{-,\ast}\,\Psi_{0}\end{pmatrix}$
(60)
where complex conjugation to the lower equation has been applied. Defining the
LR matrix $\mathcal{L}$ as
$\mathcal{L}=\begin{pmatrix}\hat{\mathbf{P}}_{\Psi_{0}}(\hat{H}^{0}-\varepsilon^{0})\hat{\mathbf{P}}_{\Psi_{0}}&0\\\
0&-\hat{\mathbf{P}}^{\ast}_{\Psi_{0}}(\hat{H}^{0,\ast}-\varepsilon^{0})\hat{\mathbf{P}}^{\ast}_{\Psi_{0}}\end{pmatrix}$
(61)
where redundantly the projectors $\hat{\mathbf{P}}_{\Psi_{0}}$ and
$\hat{\mathbf{P}}^{\ast}_{\Psi_{0}}$ have been inserted from the right-hand
side (RHS) to the upper and lower blocks, the homogeneous version of Eq. (60)
is given by
$\mathcal{L}\begin{pmatrix}u\\\ v\end{pmatrix}=\omega\begin{pmatrix}u\\\
v\end{pmatrix}.$ (62)
For the upper block, one can immediately deduce that the spectrum of
corresponding eigenvalues is given by $\omega_{k}=E_{k}-\varepsilon^{0}$,
meaning that one obtains the eigenvalues of the unperturbed Hamiltonian
$\hat{H}^{0}$. Moreover, one obtains $\omega_{k}=-(E_{k}-\varepsilon^{0})$ for
the lower block.
This is an important result because it means that the LR due to an external
perturbation, as described in Eq. (52), yields, if applied to an exact
eigenstate of $\hat{H}^{0}$ (not necessarily the ground state), the full
spectrum of exact excitation energies. It is worth stressing the generality of
the above derivation because no physical properties of the system were
specified, i.e., whether $\hat{H}^{0}$ describes a single- or multi-particle
system, bosons or fermions (or even mixtures), or identical or distinguishable
particles. Thus, LR represents a generic method for obtaining the spectrum of
excited states of a quantum system. Section 3.2.2, explicitly deals with the
many-body linear response of systems consisting of identical bosons.
This section is closed by referring to systems where no exact eigenstate of
$\hat{H}^{0}$, e.g., the ground state, is known. This is naturally the most
common case since the number of systems where analytic solutions are available
is very small. Hence, in order to obtain accurate results for the excitation
energies from a LR analysis, one is in need of a very accurate approximation
of the system’s ground state. In other words, increasing the quality of the
underlying ground-state approximation will increase the accuracy of the
spectrum of excited states. From a numerical point of view, Section 5 shows
that MCTDHB turns out to be a very powerful approach to calculate the ground
state of a trapped BEC to very high accuracy, and that it is clearly superior
to other widely used methods like the GP mean-field equation or the FCI
method. As a result, LR-MCTDHB leads to highly accurate excitation spectra,
for which numerical evidence is presented in Section 5.3.
#### 3.2.2 Linear-response MCTDHB (LR-MCTDHB)
In this section, LR-MCTDHB [or LR-MCTDHB($M$)] is derived. To this end, a
trapped $N$-boson system is considered and its ground-state orbitals and
coefficients, $\\{\phi_{k}^{0}:1\leq k\leq M\\}$ and $\mathbf{C}^{0}$, are
calculated by using imaginary time-propagation of the MCTDHB working equations
[57], Eqs. (43) and (46). Afterwards, the same time-dependent external
perturbation as given in Eq. (52) is added to the Hamiltonian, such that the
latter is finally given by Eqs. (51). The EOMs are linearized according to
this. The subsequent derivation of the LR equations follows Refs. [34, 35].
The ansätze for the response orbitals and coefficients due to the external
perturbation are
$\displaystyle\phi_{k}(\mathbf{r},t)$
$\displaystyle\approx\phi_{k}^{0}(\mathbf{r})+\delta\phi_{k}(\mathbf{r},t)$
(63) $\displaystyle\delta\phi_{k}(\mathbf{r},t)$
$\displaystyle=u_{k}(\mathbf{r})e^{-i\omega
t}+v_{k}^{\ast}(\mathbf{r})e^{i\omega t}$ (64) $\displaystyle\mathbf{C}(t)$
$\displaystyle\approx e^{-i\epsilon^{0}t}[\mathbf{C}^{0}+\delta\mathbf{C}(t)]$
(65) $\displaystyle\delta\mathbf{C}(t)$
$\displaystyle=\mathbf{C}_{u}e^{-i\omega t}+\mathbf{C}_{v}^{\ast}e^{i\omega
t}$ (66)
with $1\leq k\leq M$. The $k$-th response orbital is thus given by the zeroth-
order orbital, typically the time-independent ground-state orbital
$\phi_{k}^{0}(\mathbf{r})$, plus a weak, time-dependent perturbation
$\delta\phi_{k}(\mathbf{r},t)$ with oscillation frequency $\omega$ and
stationary response amplitudes $u_{k}(\mathbf{r})$ and $v_{k}(\mathbf{r})$.
The response coefficients consist of similar parts and are additionally
multiplied with a time-dependent phase $e^{-i\varepsilon^{0}t}$ to which it is
referred below. The time-dependence is therefore essentially incorporated in
the perturbation part. In the following, the linearization of the orbitals’
and coefficients’ EOMs are discussed separately.
For linearizing the orbitals’ EOMs, it is instructive to utilize the version
in Eq. (38) given by
$i\sum_{q=1}^{M}\rho_{kq}|\dot{\phi}_{q}(t)\rangle=\sum_{q=1}^{M}[\hat{Z}_{kq}-\mu_{kq}(t)]|\phi_{q}(t)\rangle,$
(67)
with the replacement
$\hat{Z}_{kq}=\rho_{kq}[\hat{h}+\hat{H}_{\text{ext}}(t)]+\sum_{s,l=1}^{M}\rho_{ksql}\hat{W}_{sl}.$
(68)
Inserting Eq. (63) into Eq. (67), one can group the resulting expression in
terms of perturbation orders. In zeroth order, one obtains
$0=\sum_{q}[\hat{Z}_{kq}^{0}-\mu^{0}_{kq}]|\phi_{q}^{0}\rangle$ (69)
with
$\hat{Z}^{0}_{kq}=\rho^{0}_{kq}\hat{h}+\sum_{s,l=1}^{M}\rho^{0}_{ksql}\hat{W}^{0}_{sl}.$
(70)
Here and in the following, quantities with the superscript ’0’ denote that
they only contain zeroth-order orbitals and coefficients. Eq. (69) represents
the orbital part of the MCHB working equations, see Ref. [57] in this context.
For the first order, the LHS of Eq. (67) yields
$i\sum_{q}(\delta\rho_{kq}\underbrace{|\dot{\phi}_{q}^{0}\rangle}_{=0}+\rho_{kq}^{0}|\dot{\delta\phi_{q}}(t)\rangle)=i\sum_{q}\rho_{kq}^{0}|\dot{\delta\phi_{q}}(t)\rangle.$
(71)
The RHS is given by
$\sum_{q}[\hat{Z}_{kq}^{0}-\mu_{kq}^{0}]\,|\delta\phi_{q}(t)\rangle+\sum_{q}[\delta\hat{Z}_{kq}-\delta\mu_{kq}(t)]\,|\phi_{q}^{0}\rangle$
(72)
with
$\displaystyle\delta\hat{Z}_{kq}$
$\displaystyle=\delta\left(\rho_{kq}\hat{h}+\sum_{s,l}\rho_{ksql}\hat{W}_{sl}\right)+\rho_{kq}^{0}\hat{H}_{\text{ext}}(t)$
$\displaystyle=\delta\rho_{kq}\hat{h}+\sum_{s,l}\rho_{ksql}^{0}\delta\hat{W}_{sl}+\sum_{s,l}\delta\rho_{ksql}\hat{W}_{sl}^{0}+\rho_{kq}^{0}\hat{H}_{\text{ext}}(t).$
(73)
To compute the RHS further, the variation of the chemical potential
$\delta\mu_{kq}(t)$, given by
$\displaystyle\delta\mu_{kq}(t)$
$\displaystyle=\left\langle\delta\phi_{q}\left|\sum_{j}\left[\hat{Z}_{kj}^{0}-i\rho_{kj}\frac{\partial}{\partial
t}\right]\,\right|\phi_{j}^{0}\right\rangle+\left\langle\phi_{q}^{0}\left|\sum_{j}\delta\left[\hat{Z}_{kj}^{0}-i\rho_{kj}\frac{\partial}{\partial
t}\,\right|\phi_{j}\right\rangle\right]$
$\displaystyle\overset{\text{Eq.}(\ref{LR-
MCTDHB_zero})}{=}\sum_{j}\mu_{kj}^{0}\left\langle\delta\phi_{q}|\phi_{j}^{0}\right\rangle+\left\langle\phi_{q}^{0}\left|\sum_{j}\rho_{kj}^{0}\hat{H}_{\text{ext}}(t)\right|\phi_{j}^{0}\right\rangle$
$\displaystyle+\left\langle\phi_{q}^{0}\left|\left(\sum_{j}\delta\rho_{kj}\hat{h}+\sum_{j,s,l}(\rho_{ksjl}^{0}\delta\hat{W}_{sl}+\delta\rho_{ksjl}\hat{W}_{sl}^{0})\right)\right|\phi_{j}^{0}\right\rangle$
$\displaystyle+\left\langle\phi_{q}^{0}\left|\sum_{j}\left(\hat{Z}_{kj}^{0}-i\rho_{kj}^{0}\frac{\partial}{\partial
t}\right)\right|\delta\phi_{j}\right\rangle\,,$ (74)
is employed. Thus, Eq. (72) can be written as
$\displaystyle\sum_{q}[\hat{Z}_{kq}^{0}-\mu_{kq}^{0}]\,|\delta\phi_{q}\rangle+\sum_{q}\left\\{\delta\rho_{kq}\hat{h}+\rho_{kq}^{0}\hat{H}_{\text{ext}}(t)+\sum_{s,l}(\rho_{ksql}\,\delta\hat{W}_{sl}+\delta\rho_{ksql}\,\hat{W}_{sl}^{0})\right\\}|\phi_{q}^{0}\rangle$
$\displaystyle-\sum_{q}\left\\{\sum_{j}(-\mu_{kj}^{0})\left\langle\phi_{q}^{0}|\delta\phi_{j}\right\rangle+\left\langle\phi_{q}^{0}\left|\sum_{j}\rho_{kj}^{0}\hat{H}_{\text{ext}}(t)\right|\phi_{j}^{0}\right\rangle\right.$
$\displaystyle+\left.\left\langle\phi_{q}^{0}\left|\left(\sum_{j}\delta\rho_{kj}\hat{h}+\sum_{j,s,l}(\rho_{ksjl}^{0}\delta\hat{W}_{sl}+\delta\rho_{ksjl}\hat{W}_{sl}^{0})\right)\right|\phi_{j}^{0}\right\rangle\right.$
$\displaystyle+\left.\left\langle\phi_{q}^{0}\left|\sum_{j}\left(\hat{Z}_{kj}^{0}-i\rho_{kj}^{0}\frac{\partial}{\partial
t}\right)\right|\delta\phi_{j}\right\rangle\ \right\\}|\phi_{q}^{0}\rangle$
(75)
which, combined with the LHS, Eq. (71), leads to
$\displaystyle\sum_{q}\hat{\mathbf{P}}\left\\{\left(\hat{Z}_{kq}^{0}-\mu_{kq}^{0}\right)\,|\delta\phi_{q}\rangle+\left(\delta\rho_{kq}\hat{h}+\rho_{kq}^{0}\hat{H}_{\text{ext}}(t)\right.\right.$
$\displaystyle+\sum_{s,l}\delta\rho_{ksql}\,\hat{W}_{sl}^{0}+\sum_{s,l}\left.\left.\rho_{ksql}^{0}\,\delta\hat{W}_{sl}\right)|\phi_{q}^{0}\rangle\right\\}$
$\displaystyle=i\hat{\mathbf{P}}\sum_{q}\rho_{kq}^{0}|\delta\dot{\phi}_{q}\rangle$
(76)
where the projection operator
$\hat{\mathbf{P}}=\mathbb{1}-\sum_{j^{\prime}}|\phi_{j^{\prime}}^{0}\rangle\langle\phi_{j^{\prime}}^{0}|$
projects on the tangential space of the stationary orbitals. From the
invariance condition, Eq. (42), one can deduce
$\displaystyle 0$
$\displaystyle=\delta\langle\phi_{k}|\dot{\phi}_{q}\rangle=\langle\delta\phi_{k}\underbrace{|\dot{\phi}_{q}^{0}\rangle}_{=0}+\langle\phi_{k}^{0}|\delta\dot{\phi}_{q}\rangle$
$\displaystyle=-i\omega\langle\phi_{k}^{0}|u_{q}\rangle e^{-i\omega
t}+i\omega\langle\phi_{k}^{0}|v^{\ast}_{q}\rangle e^{i\omega t}$
$\displaystyle\Rightarrow\langle\phi_{k}^{0}|u_{q}\rangle=\langle\phi_{k}^{0}|v^{\ast}_{q}\rangle=0\quad\quad\forall\,1\leq
k,q\leq M$ (77)
such that
$\hat{\mathbf{P}}|\delta\dot{\phi}_{q}\rangle=\left(\mathbb{1}-\sum_{j^{\prime}=1}^{M}|\phi^{0}_{j^{\prime}}\rangle\langle\phi^{0}_{j^{\prime}}|\right)|\delta\dot{\phi}_{q}\rangle=|\delta\dot{\phi}_{q}\rangle$
(78)
which makes the projection operator on the RHS of Eq. (3.2.2) redundant. One
further utilizes
$\displaystyle\delta\hat{W}_{sl}\left[u_{l},u_{s}^{\ast},v_{l}^{\ast},v_{s}\right]$
$\displaystyle=\int
d\mathbf{r}^{\,\prime}\left\\{\delta\phi_{s}^{\ast}(\mathbf{r}^{\,\prime},t)\,\hat{W}(\mathbf{r},\mathbf{r}^{\,\prime})\,\phi_{l}^{0}(\mathbf{r}^{\,\prime})\right.$
$\displaystyle+\left.\phi^{0,\ast}_{s}(\mathbf{r}^{\,\prime})\,\hat{W}(\mathbf{r},\mathbf{r}^{\,\prime})\,\delta\phi_{l}(\mathbf{r}^{\,\prime},t)\right\\}$
(79)
$\displaystyle\delta\rho_{kq}\left[\mathbf{C}_{u},\mathbf{C}_{v},\mathbf{C}_{u}^{\ast},\mathbf{C}_{v}^{\ast}\right]$
$\displaystyle=\langle\delta\mathbf{C}(t)|\hat{b}^{\dagger}_{k}\hat{b}_{q}|\mathbf{C}^{0}\rangle+\langle\mathbf{C}^{0}|\hat{b}^{\dagger}_{k}\hat{b}_{q}|\delta\mathbf{C}(t)\rangle$
(80)
$\displaystyle\delta\rho_{ksql}\left[\mathbf{C}_{u},\mathbf{C}_{v},\mathbf{C}_{u}^{\ast},\mathbf{C}_{v}^{\ast}\right]$
$\displaystyle=\langle\delta\mathbf{C}(t)|\hat{b}^{\dagger}_{k}\hat{b}^{\dagger}_{s}\hat{b}_{q}\hat{b}_{l}|\mathbf{C}^{0}\rangle+\langle\mathbf{C}^{0}|\hat{b}^{\dagger}_{k}\hat{b}^{\dagger}_{s}\hat{b}_{q}\hat{b}_{l}|\delta\mathbf{C}(t)\rangle$
(81)
where the action of, e.g., $\hat{b}^{\dagger}_{k}\hat{b}_{q}$ on
$|\mathbf{C}\rangle$ results in a new vector of coefficients for the same Fock
states, i.e.,
$\hat{b}^{\dagger}_{k}\hat{b}_{q}|\mathbf{C}\rangle=\sum_{\mathbf{n}}C_{\mathbf{n}}(t)\hat{b}^{\dagger}_{k}\hat{b}_{q}|\mathbf{n};t\rangle=\sum_{\mathbf{n}^{\prime}}C_{\mathbf{n}^{\prime}}(t)|\mathbf{n}^{\prime};t\rangle\equiv|\mathbf{C}^{\prime}\rangle,$
(82)
and inserts the above equations, together with Eqs. (52), (64), and (66), into
Eq. (3.2.2). Collecting all terms proportional to $e^{-i\omega t}$ results in
$\displaystyle\sum_{q}\hat{\mathbf{P}}\left\\{\left(\hat{Z}_{kq}^{0}-\mu_{kq}^{0}\right)\,|u_{q}\rangle+\left(\delta\rho_{kq}\left[\mathbf{C}_{u},\mathbf{C}_{v}\right]\hat{h}+\rho_{kq}^{0}\,\hat{f}^{+}\right.\right.$
$\displaystyle+\sum_{s,l}\delta\rho_{ksql}\left[\mathbf{C}_{u},\mathbf{C}_{v}\right]\,\hat{W}_{sl}^{0}+\sum_{s,l}\left.\left.\rho_{ksql}^{0}\,\delta\hat{W}_{sl}[u_{l},v_{s}]\right)|\phi_{q}^{0}\rangle\right\\}$
$\displaystyle=\omega\sum_{q}\rho_{kq}^{0}|u_{q}\rangle.$ (83)
Similarly, collecting all terms proportional to $e^{i\omega t}$ yields
$\displaystyle\sum_{q}\hat{\mathbf{P}}\left\\{\left(\hat{Z}_{kq}^{0}-\mu_{kq}^{0}\right)\,|v^{\ast}_{q}\rangle+\left(\delta\rho_{kq}\left[\mathbf{C}_{u}^{\ast},\mathbf{C}_{v}^{\ast}\right]\hat{h}+\rho_{kq}^{0}\,\hat{f}^{-}\right.\right.$
$\displaystyle+\sum_{s,l}\delta\rho_{ksql}\left[\mathbf{C}_{u}^{\ast},\mathbf{C}_{v}^{\ast}\right]\,\hat{W}_{sl}^{0}+\sum_{s,l}\left.\left.\rho_{ksql}^{0}\,\delta\hat{W}_{sl}[u_{s}^{\ast},v_{l}^{\ast}]\right)|\phi_{q}^{0}\rangle\right\\}$
$\displaystyle=-\omega\sum_{q}\rho_{kq}^{0}|v_{q}^{\ast}\rangle$
$\displaystyle\overset{\text{C.C.}}{\iff}$
$\displaystyle\sum_{q}\hat{\mathbf{P}}^{\ast}\left\\{\left(\hat{Z}_{kq}^{0,\ast}-\mu_{qk}^{0}\right)\,|v_{q}\rangle+\left(\delta\rho_{qk}\left[\mathbf{C}_{u},\mathbf{C}_{v}\right]\hat{h}^{\ast}+\rho_{qk}^{0}\,\hat{f}^{-,\ast}\right.\right.$
$\displaystyle+\sum_{s,l}\delta\rho_{qlks}\left[\mathbf{C}_{u},\mathbf{C}_{v}\right]\,\hat{W}_{ls}^{0}+\sum_{s,l}\left.\left.\rho_{qlks}^{0}\,\delta\hat{W}_{ls}[u_{s},v_{l}]\right)|\phi_{q}^{0,\ast}\rangle\right\\}$
$\displaystyle=-\omega\sum_{q}\rho_{qk}^{0}|v_{q}\rangle$ (84)
where ’C.C.’ denotes complex conjugation of the entire equation. Here, it is
used that the matrix $\\{\mu_{ij}\\}$ of the Lagrange multipliers is
hermitian. Furthermore, the two-body interaction potential is assumed to be
real throughout this work.
For linearizing the coefficients’ EOMs, one obtains in zeroth order the
stationary equation
$\boldsymbol{\mathcal{H}^{0}}\mathbf{C}^{0}=\varepsilon^{0}\,\mathbf{C}^{0}$
(85)
with the matrix elements
$\\{\mathcal{H}^{0}_{\mathbf{n}\mathbf{n}^{\,\prime}}\\}=\langle\mathbf{n}|\hat{H}^{0}|\mathbf{n}^{\,\prime}\rangle$.
Eq. (85) is the working equation for the coefficients within the MCHB theory.
In first order, one arrives at
$i\frac{\partial}{\partial
t}\delta\mathbf{C}(t)=\delta\boldsymbol{\mathcal{H}}\mathbf{C}^{0}+\left(\boldsymbol{\mathcal{H}^{0}}-\varepsilon^{0}\right)\delta\mathbf{C}(t)$
(86)
with
$\delta\boldsymbol{\mathcal{H}}=\sum_{k,q}\delta
h_{kq}\,\hat{b}^{\dagger}_{k}\hat{b}_{q}+\frac{1}{2}\sum_{k,s,q,l}\delta
W_{ksql}\,\hat{b}^{\dagger}_{k}\hat{b}^{\dagger}_{s}\hat{b}_{q}\hat{b}_{l}$
(87)
and
$\displaystyle\delta h_{kq}$
$\displaystyle=\int\,d\mathbf{r}\,\delta\phi^{\ast}_{k}(\mathbf{r},t)\hat{h}\phi_{q}^{0}(\mathbf{r})+\int\,d\mathbf{r}\,\phi^{0,\ast}_{k}(\mathbf{r})\hat{h}\delta\phi_{q}(\mathbf{r},t)$
$\displaystyle+\underbrace{\int\,d\mathbf{r}\,\phi^{0,\ast}_{k}(\mathbf{r})\hat{H}_{\text{ext}}(\mathbf{r},t)\phi_{q}^{0}(\mathbf{r})}_{\equiv(H_{\text{ext}})_{kq}}$
(88) $\displaystyle\delta W_{ksql}$
$\displaystyle=\iint\,d\mathbf{r}d\mathbf{r}^{\,\prime}\,\delta\phi^{\ast}_{k}(\mathbf{r},t)\phi_{s}^{0,\ast}(\mathbf{r}^{\,\prime})\hat{W}(\mathbf{r},\mathbf{r}^{\,\prime})\phi_{q}^{0}(\mathbf{r})\phi_{l}^{0}(\mathbf{r}^{\,\prime})$
$\displaystyle+\iint\,d\mathbf{r}d\mathbf{r}^{\,\prime}\,\phi^{0,\ast}_{k}(\mathbf{r})\delta\phi_{s}^{\ast}(\mathbf{r}^{\,\prime},t)\hat{W}(\mathbf{r},\mathbf{r}^{\,\prime})\phi_{q}^{0}(\mathbf{r})\phi_{l}^{0}(\mathbf{r}^{\,\prime})$
$\displaystyle+\iint\,d\mathbf{r}d\mathbf{r}^{\,\prime}\,\phi^{0,\ast}_{k}(\mathbf{r})\phi_{s}^{0,\ast}(\mathbf{r}^{\,\prime})\hat{W}(\mathbf{r},\mathbf{r}^{\,\prime})\delta\phi_{q}(\mathbf{r},t)\phi_{l}^{0}(\mathbf{r}^{\,\prime})$
$\displaystyle+\iint\,d\mathbf{r}d\mathbf{r}^{\,\prime}\,\phi^{0,\ast}_{k}(\mathbf{r})\phi_{s}^{0,\ast}(\mathbf{r}^{\,\prime})\hat{W}(\mathbf{r},\mathbf{r}^{\,\prime})\phi_{q}(\mathbf{r})\delta\phi_{l}(\mathbf{r}^{\,\prime},t).$
(89)
Collecting all terms proportional to $e^{-i\omega t}$ yields
$\displaystyle\omega\mathbf{C}_{u}$
$\displaystyle=(\boldsymbol{\mathcal{H}^{0}}-\varepsilon^{0})\mathbf{C}_{u}+\left[\sum_{k,q}\delta
h_{k,q}[v_{k},u_{q},\hat{f}^{+}]\hat{b}^{\dagger}_{k}\hat{b}_{q}\right.$
$\displaystyle+\left.\frac{1}{2}\sum_{k,s,q,l}\delta
W_{ksql}[v_{k},v_{s},u_{q},u_{l}]\hat{b}^{\dagger}_{k}\hat{b}^{\dagger}_{s}\hat{b}_{q}\hat{b}_{l}\right]\mathbf{C}^{0},$
(90)
whereas collecting all terms proportional to $e^{+i\omega t}$ results in
$\displaystyle-\omega\mathbf{C}_{v}^{\ast}$
$\displaystyle=(\boldsymbol{\mathcal{H}^{0}}-\varepsilon^{0})\mathbf{C}_{v}^{\ast}+\left[\sum_{k,q}\delta
h_{k,q}[u^{\ast}_{k},v^{\ast}_{q},\hat{f}^{-}]\hat{b}^{\dagger}_{k}\hat{b}_{q}\right.$
$\displaystyle+\left.\frac{1}{2}\sum_{k,s,q,l}\delta
W_{ksql}[u^{\ast}_{k},u^{\ast}_{s},v^{\ast}_{q},v^{\ast}_{l}]\hat{b}^{\dagger}_{k}\hat{b}^{\dagger}_{s}\hat{b}_{q}\hat{b}_{l}\right]\mathbf{C}^{0}$
$\displaystyle\overset{\text{C.C.}}{\iff}-\omega\mathbf{C}_{v}$
$\displaystyle=(\boldsymbol{\mathcal{H}^{0,\ast}}-\varepsilon^{0})\mathbf{C}_{v}+\left[\sum_{k,q}\delta
h_{q,k}[u_{k},v_{q},\hat{f}^{-,\ast}]\left(\hat{b}^{\dagger}_{k}\hat{b}_{q}\right)^{\ast}\right.$
$\displaystyle+\left.\frac{1}{2}\sum_{q,l,k,s}\delta
W_{qlks}[u_{k},u_{s},v_{q},v_{l}]\left(\hat{b}^{\dagger}_{k}\hat{b}^{\dagger}_{s}\hat{b}_{q}\hat{b}_{l}\right)^{\ast}\right]\mathbf{C}^{0,\ast}.$
(91)
After the linearization of both the orbitals’ and coefficients’ EOMs, one can
combine the sets of LR equations by casting Eqs. (3.2.2), (3.2.2), (3.2.2) and
(3.2.2) into matrix form, leading to
$\left(\mathbf{\mathcal{P}}\mathcal{L}-\mathcal{M}\omega\right)\begin{pmatrix}\mathbf{u}\\\
\mathbf{v}\\\ \mathbf{C}_{u}\\\
\mathbf{C}_{v}\end{pmatrix}=\mathbf{\mathcal{M}}\,\mathbf{\mathcal{P}}\,\mathbf{\mathcal{R}}[\hat{f}^{+},\hat{f}^{-}]$
(92)
with $\mathbf{u}=(u_{1},...,u_{M})^{t}$, $\mathbf{v}=(v_{1},...,v_{M})^{t}$
and the LR matrix
$\mathcal{L}=\begin{pmatrix}\mathbf{\mathcal{L}}_{oo}&\mathbf{\mathcal{L}}_{oc}\\\
\mathbf{\mathcal{L}}_{co}&\mathbf{\mathcal{L}}_{cc}\end{pmatrix}$ (93)
where the submatrices refer to the couplings between the orbitals and the
coefficients. In Eq. (92), $\mathbf{\mathcal{P}}$ is a projection matrix,
$\mathbf{\mathcal{M}}$ is a metric which contains the one-body RDM, and the
RHS $\mathbf{\mathcal{R}}$ contains the amplitudes of the external
perturbation, $\hat{f}^{+}$ and $\hat{f}^{-}$. In the following, the
submatrices in Eq. (93) are explicitly derived.
The $(2M)$-dimensional square orbital matrix $\mathcal{L}_{oo}$ consists of
four blocks and reads
$\mathcal{L}_{oo}=\begin{pmatrix}\mathcal{L}^{u}_{oo}&\mathcal{L}^{v}_{oo}\\\
-\left(\mathcal{L}^{v}_{oo}\right)^{\ast}&-\left(\mathcal{L}^{u}_{oo}\right)^{\ast}\end{pmatrix}.$
(94)
The upper part can be obtained from Eq. (3.2.2) by collecting all terms
proportional to $\mathbf{u}$, yielding
$\mathcal{L}^{u}_{oo}=\boldsymbol{\rho^{0}}\hat{h}-\boldsymbol{\mu^{0}}+\mathbf{\Omega^{0}}+\boldsymbol{\kappa}^{1}$
(95)
with $\boldsymbol{\rho^{0}}=\\{\rho^{0}_{kq}\\}$,
$\boldsymbol{\mu^{0}}=\\{\mu^{0}_{kq}\\}$,
$\mathbf{\Omega^{0}}=\\{\Omega^{0}_{kq}\\}=\sum_{s,l}\rho^{0}_{ksql}\hat{W}^{0}_{sl}$
and
$\boldsymbol{\kappa}^{1}=\\{\kappa^{1}_{kq}\\}=\sum_{s,l}\rho^{0}_{ksql}\hat{K}^{0}_{sl}$.
Therein, the exchange interaction operator is defined as
$\hat{K}_{sl}=\int
d\mathbf{r}^{\,\prime}\phi_{s}^{\ast}(\mathbf{r}^{\,\prime})\hat{W}(\mathbf{r},\mathbf{r}^{\,\prime})\hat{\mathcal{P}}_{\mathbf{r}\mathbf{r}^{\,\prime}}\phi_{l}(\mathbf{r}^{\,\prime})$
(96)
where $\hat{\mathcal{P}}_{\mathbf{r}\mathbf{r}^{\,\prime}}$ permutes the
coordinates $\mathbf{r}$ and $\mathbf{r}^{\,\prime}$, i.e.,
$\hat{\mathcal{P}}_{\mathbf{r}\mathbf{r}^{\,\prime}}f(\mathbf{r})=f(\mathbf{r}^{\,\prime})$
(97)
for any function $f(\mathbf{r})$. The action of $\hat{K}_{sl}$ on an arbitrary
function $f(\mathbf{r})$ therefore reads
$\hat{K}_{sl}\,f(\mathbf{r})=\hat{W}_{sf}\,\phi_{l}(\mathbf{r}).$ (98)
Furthermore, collecting all terms proportional to $\mathbf{v}$ yields
$\mathcal{L}^{v}_{oo}=\boldsymbol{\kappa}^{2},\quad\boldsymbol{\kappa}^{2}=\\{\kappa^{2}_{kq}\\}=\sum_{s,l}\rho_{kqsl}^{0}\hat{K}_{l^{\ast}s}$
(99)
for the upper right block of $\mathcal{L}_{oo}$. The lower two submatrices can
be obtained from Eq. (3.2.2) and turn out to be the negative complex conjugate
of the upper submatrices.
The upper right block of $\mathcal{L}$, i.e., the $2(M\times
N_{\text{conf}})$-matrix $\mathcal{L}_{oc}$, has a similar structure than
$\mathcal{L}_{oo}$, meaning it consists of four separate blocks given by
$\mathcal{L}_{oc}=\begin{pmatrix}\mathcal{L}^{u}_{oc}&\mathcal{L}^{v}_{oc}\\\
-\left(\mathcal{L}^{v}_{oc}\right)^{\ast}&-\left(\mathcal{L}^{u}_{oc}\right)^{\ast}\end{pmatrix}.$
(100)
The upper two submatrices can again be extracted from Eq. (3.2.2) and yield,
collecting all terms proportional to $\mathbf{C_{u}}$,
$\mathcal{L}^{u}_{oc}=\sum_{q}\left\\{\hat{h}\left(\hat{\rho}_{qk}\mathbf{C^{0}}\right)^{\dagger}+\sum_{s,l}\hat{W}_{sl}^{0}\left(\hat{\rho}_{qlks}\mathbf{C^{0}}\right)^{\dagger}\right\\}\phi_{q}^{0}\,,$
(101)
whereas collecting all terms proportional to $\mathbf{C_{v}}$ results in
$\mathcal{L}^{v}_{oc}=\sum_{q}\left\\{\hat{h}\left(\hat{\rho}_{kq}\mathbf{C^{0}}\right)^{t}+\sum_{s,l}\hat{W}_{sl}^{0}\left(\hat{\rho}_{ksql}\mathbf{C^{0}}\right)^{t}\right\\}\phi_{q}^{0}\,.$
(102)
The lower two submatrices can be extracted from Eq. (3.2.2), see Ref. [35].
With respect to the $2(N_{\text{conf}}\times M)$-matrix $\mathcal{L}_{co}$,
one finds
$\mathcal{L}_{co}=\begin{pmatrix}\mathcal{L}^{u}_{co}&\mathcal{L}^{v}_{co}\\\
-\left(\mathcal{L}^{v}_{co}\right)^{\ast}&-\left(\mathcal{L}^{u}_{co}\right)^{\ast}\end{pmatrix}$
(103)
with
$\mathcal{L}_{co}^{u}=\sum_{k}\phi_{k}^{0,\ast}\left\\{\left(\hat{\rho}_{kq}\mathbf{C^{0}}\right)\hat{h}+\sum_{sl}\left(\hat{\rho}_{klqs}\mathbf{C^{0}}\right)\hat{W}_{ls}^{0}\right\\}$
(104)
from collecting all terms proportional to $u_{q}$ in Eq. (3.2.2) and
$\mathcal{L}_{co}^{v}=\sum_{k}\phi_{k}^{0}\left\\{\left(\hat{\rho}_{qk}\mathbf{C^{0}}\right)\hat{h}^{\ast}+\sum_{sl}\left(\hat{\rho}_{qskl}\mathbf{C^{0}}\right)\hat{W}_{sl}^{0}\right\\}$
(105)
from collecting all terms proportional to $v_{q}$ in Eq. (3.2.2). It can thus
be seen that
$\left(\mathcal{L}_{oc}^{u}\right)^{\dagger}=\mathcal{L}_{co}^{u},\quad\left(\mathcal{L}_{oc}^{v}\right)^{t}=\mathcal{L}_{co}^{v}.$
(106)
Furthermore, one can extract the $(2N_{\text{conf}})$-dimensional square
matrix $\mathcal{L}_{cc}$ from Eqs. (3.2.2) and (3.2.2), which gives
$\mathcal{L}_{cc}=\begin{pmatrix}\boldsymbol{\mathcal{H}^{0}}-\varepsilon^{0}&\mathbf{0_{cc}}\\\
\mathbf{0_{cc}}&-\left(\boldsymbol{\mathcal{H}^{0,\ast}}-\varepsilon^{0}\right)\end{pmatrix}$
(107)
where $\mathbf{0_{cc}}$ is the ($N_{\text{conf}}$)-dimensional zero matrix.
The projection operator $\mathbf{\mathcal{P}}$ is given by
$\mathbf{\mathcal{P}}=\begin{pmatrix}\boldsymbol{\mathcal{P}}_{oo}&\mathbf{0_{oc}}\\\
\mathbf{0_{co}}&\boldsymbol{\mathbb{1}_{cc}}\end{pmatrix}$ (108)
with
$\boldsymbol{\mathcal{P}}_{oo}=\begin{pmatrix}\hat{\mathbf{P}}&0\\\
0&\mathbf{\hat{P}^{\ast}}\end{pmatrix}$ (109)
and $\mathbf{\mathbb{1}_{cc}}$ being the $(2N_{\text{conf}})$-dimensional unit
matrix. The matrices $\mathbf{0_{oc}}$ and $\mathbf{0_{co}}$ denote the
$2(M\times N_{\text{conf}})$\- and $2(N_{\text{conf}}\times M)$-dimensional
zero matrices, respectively.
The metric $\mathbf{\mathcal{M}}$ reads
$\mathbf{\mathcal{M}}=\begin{pmatrix}\boldsymbol{\rho}_{oo}&\mathbf{0_{oc}}\\\
\mathbf{0_{co}}&\mathbf{\mathbb{1}_{cc}}\end{pmatrix}$ (110)
where $\mathbf{\rho}_{oo}$ is given by
$\boldsymbol{\rho}_{oo}=\begin{pmatrix}\boldsymbol{\rho^{0}}&\mathbf{0_{o}}\\\
\mathbf{0_{o}}&\boldsymbol{\rho^{0,\ast}}\end{pmatrix}$ (111)
and $\mathbf{0_{o}}$ being the ($M\times M$)-dimensional zero matrix.
In the following, the response amplitudes and coefficients are written in bra-
ket notation for reasons that become clear below. In order to render Eq. (92)
an eigenvalue equation, it is helpful to notice that only the term
proportional to the external frequency $\omega$ has no projector, which
implies
$\mathbf{\mathcal{P}}\mathbf{\mathcal{M}}\begin{pmatrix}|\mathbf{u}\rangle\\\
|\mathbf{v}\rangle\\\ |\mathbf{C}_{u}\rangle\\\
|\mathbf{C}_{v}\rangle\end{pmatrix}=\mathbf{\mathcal{M}}\begin{pmatrix}|\mathbf{u}\rangle\\\
|\mathbf{v}\rangle\\\ |\mathbf{C}_{u}\rangle\\\
|\mathbf{C}_{v}\rangle\end{pmatrix}$ (112)
and thus one may add a redundant projector
$\mathbf{\mathcal{P}}\mathcal{L}\Rightarrow\mathbf{\mathcal{P}}\mathcal{L}\mathbf{\mathcal{P}}$
to the left term of the LHS of Eq. (92). Multiplying now with
$\mathbf{\mathcal{M}^{-1/2}}$ from the left yields
$\left(\mathbf{\mathcal{M}^{-1/2}}\mathbf{\mathcal{P}}\mathcal{L}\mathbf{\mathcal{P}}\mathbf{\mathcal{M}^{-1/2}}-\omega\right)\mathbf{\mathcal{M}^{+1/2}}\begin{pmatrix}|\mathbf{u}\rangle\\\
|\mathbf{v}\rangle\\\ |\mathbf{C}_{u}\rangle\\\
|\mathbf{C}_{v}\rangle\end{pmatrix}=\mathbf{\mathcal{M}^{+1/2}}\mathbf{\mathcal{P}}\mathbf{\mathcal{R}}[\hat{f}^{+},\hat{f}^{-}]$
(113)
or, in a more compact notation,
$\left(\bar{\mathcal{L}}-\omega\right)\begin{pmatrix}|\mathbf{\overline{u}}\rangle\\\
|\mathbf{\overline{v}}\rangle\\\ |\overline{\mathbf{C}_{u}}\rangle\\\
|\overline{\mathbf{C}_{v}}\rangle\end{pmatrix}=\mathbf{\overline{\mathcal{R}}}[\hat{f}^{+},\hat{f}^{-}]$
(114)
where
$\displaystyle\bar{\mathcal{L}}$
$\displaystyle=\mathbf{\mathcal{M}^{-1/2}}\mathbf{\mathcal{P}}\mathcal{L}\mathbf{\mathcal{P}}\mathbf{\mathcal{M}^{-1/2}}$
$\displaystyle=\begin{pmatrix}\boldsymbol{\rho_{oo}}^{-1/2}\,\boldsymbol{\mathcal{P}}_{oo}\,\mathcal{L}_{oo}\,\boldsymbol{\mathcal{P}}_{oo}\,\boldsymbol{\rho_{oo}}^{-1/2}&\quad\boldsymbol{\rho_{oo}}^{-1/2}\,\boldsymbol{\mathcal{P}}_{oo}\,\mathcal{L}_{oc}\\\
\mathcal{L}_{co}\,\boldsymbol{\mathcal{P}}_{oo}\,\boldsymbol{\rho_{oo}}^{-1/2}&\mathcal{L}_{cc}\end{pmatrix},$
(115)
as well as
$\left(\mathbf{\overline{u}},\mathbf{\overline{v}},\overline{\mathbf{C}_{u}},\overline{\mathbf{C}_{v}}\right)^{t}=\mathbf{\mathcal{M}^{+1/2}}\left(\mathbf{u},\mathbf{v},\mathbf{C}_{u},\mathbf{C}_{v}\right)^{t}$
and
$\mathbf{\overline{\mathcal{R}}}=\mathbf{\mathcal{M}^{+1/2}}\mathbf{\mathcal{P}}\mathbf{\mathcal{R}}$
were introduced. To keep the notation simple, the bar over the individual
quantities in Eq. (114) will be omitted in the following. For the homogeneous
case, i.e., with $\mathbf{\mathcal{R}}=0$, one arrives at the central equation
of this work, given by
$(\mathcal{L}-\omega)\begin{pmatrix}|\mathbf{u}\rangle\\\
|\mathbf{v}\rangle\\\ |\mathbf{C}_{u}\rangle\\\
|\mathbf{C}_{v}\rangle\end{pmatrix}=0$ (116)
which is a standard eigenvalue problem of the LR matrix $\mathcal{L}$. Its
solutions, i.e., the set of eigenvalues
$\\{\omega_{k}=E_{k}-\varepsilon^{0}\\}$ and eigenvectors
$\\{(|\mathbf{u}^{k}\rangle,|\mathbf{v}^{k}\rangle,|{\mathbf{C}_{u}}^{k}\rangle,|{\mathbf{C}_{v}}^{k}\rangle)^{t}\\}$
where $k\in\mathbb{N}$ labels the excitations, can be used to solve the
inhomogeneous problem of Eq. (114) by employing the ansätze
$\begin{pmatrix}|\mathbf{u}\rangle\\\ |\mathbf{v}\rangle\\\
|\mathbf{C}_{u}\rangle\\\
|\mathbf{C}_{v}\rangle\end{pmatrix}=\sum_{k}c_{k}\begin{pmatrix}|\mathbf{u}^{k}\rangle\\\
|\mathbf{v}^{k}\rangle\\\ |{\mathbf{C}_{u}}^{k}\rangle\\\
|{\mathbf{C}_{v}}^{k}\rangle\end{pmatrix}$ (117)
and
$\mathbf{\mathcal{R}}=-\sum_{k}\gamma_{k}\begin{pmatrix}|\mathbf{u}^{k}\rangle\\\
|\mathbf{v}^{k}\rangle\\\ |{\mathbf{C}_{u}}^{k}\rangle\\\
|{\mathbf{C}_{v}}^{k}\rangle\end{pmatrix}$ (118)
where the minus sign is chosen arbitrarily. One finally obtains
$\sum_{k}c_{k}(\omega_{k}-\omega)\begin{pmatrix}|\mathbf{u}^{k}\rangle\\\
|\mathbf{v}^{k}\rangle\\\ |{\mathbf{C}_{u}}^{k}\rangle\\\
|{\mathbf{C}_{v}}^{k}\rangle\end{pmatrix}=-\sum_{k}\gamma_{k}\begin{pmatrix}|\mathbf{u}^{k}\rangle\\\
|\mathbf{v}^{k}\rangle\\\ |{\mathbf{C}_{u}}^{k}\rangle\\\
|{\mathbf{C}_{v}}^{k}\rangle\end{pmatrix}$ (119)
which means that the expansion coefficients $c_{k}$ in Eq. (117) are given by
$c_{k}=\frac{\gamma_{k}}{\omega-\omega_{k}}\,.$ (120)
It is stressed that the response weights $\\{\gamma_{k}\\}$ are the only
quantities that depend on the perturbing fields $\hat{f}^{+}$ and
$\hat{f}^{-}$ and thus on the actual shape of the perturbation. All other
quantities like the LR matrix $\mathcal{L}$, the eigenenergies
$\\{\omega_{k}\\}$ and the eigenvectors
$\\{(|\mathbf{u}^{k}\rangle,|\mathbf{v}^{k}\rangle,|{\mathbf{C}_{u}}^{k}\rangle,|{\mathbf{C}_{v}}^{k}\rangle)^{t}\\}$
are obtained from the homogeneous eigenvalue problem of Eq. (116) where the
amplitudes $\hat{f}^{+}$ and $\hat{f}^{-}$ of the perturbing field do not
appear.
Multiplying Eq. (118) by
$\left(\langle\mathbf{u}^{k}|,-\langle\mathbf{v}^{k}|,\langle{\mathbf{C}_{u}}^{k}|,-\langle{\mathbf{C}_{v}}^{k}|\right)$
from the left, taking into account the orthonormality conditions for the
perturbed amplitudes and coefficients given by
$\displaystyle\langle\mathbf{u}^{k}|\mathbf{u}^{k^{\prime}}\rangle-\langle\mathbf{v}^{k}|\mathbf{v}^{k^{\prime}}\rangle+\langle{\mathbf{C}_{u}}^{k}|{\mathbf{C}_{u}}^{k^{\prime}}\rangle-\langle{\mathbf{C}_{v}}^{k}|{\mathbf{C}_{v}}^{k^{\prime}}\rangle$
$\displaystyle=\delta_{kk^{\prime}}$ (121)
$\displaystyle\langle\mathbf{v}^{k}|\mathbf{u}^{k^{\prime},\ast}\rangle-\langle\mathbf{u}^{k}|\mathbf{v}^{k^{\prime},\ast}\rangle+\langle{\mathbf{C}_{v}}^{k}|{\mathbf{C}_{u}}^{k^{\prime},\ast}\rangle-\langle{\mathbf{C}_{u}}^{k}|{\mathbf{C}_{v}}^{k^{\prime},\ast}\rangle$
$\displaystyle=0,$ (122)
leads to the response weights
$\displaystyle\gamma_{k}$
$\displaystyle=-\left(\langle\mathbf{u}^{k}|,-\langle\mathbf{v}^{k}|,\langle{\mathbf{C}_{u}}^{k}|,-\langle{\mathbf{C}_{v}}^{k}|\right)\,\mathbf{\mathcal{R}}$
$\displaystyle=-\left(\langle\mathbf{u}^{k}|,-\langle\mathbf{v}^{k}|,\langle{\mathbf{C}_{u}}^{k}|,-\langle{\mathbf{C}_{v}}^{k}|\right)\begin{pmatrix}-(\boldsymbol{\rho^{0}})^{1/2}\,\hat{\mathbf{P}}\hat{f}^{+}|\boldsymbol{\phi^{0}}\rangle\\\
(\boldsymbol{\rho^{0,\ast}})^{1/2}\,\hat{\mathbf{P^{\ast}}}\hat{f}^{-,\ast}|\boldsymbol{\phi^{0,\ast}}\rangle\\\
-\sum_{i,j}\langle\phi_{i}|\hat{f}^{+}|\phi_{j}\rangle\,\hat{b}_{i}^{\dagger}\hat{b}_{j}\,|\mathbf{C^{0}}\rangle\\\
\sum_{i,j}\langle\phi_{j}|\hat{f}^{-,\ast}|\phi_{i}\rangle\,\left(\hat{b}_{i}^{\dagger}\hat{b}_{j}\right)^{\ast}\,|\mathbf{C^{0,\ast}}\rangle\end{pmatrix}$
$\displaystyle=\langle\mathbf{u}^{k}|\hat{f}^{+}(\boldsymbol{\rho^{0}})^{1/2}|\boldsymbol{\phi^{0}}\rangle+\langle\mathbf{v}^{k}|\hat{f}^{-,\ast}(\boldsymbol{\rho^{0,\ast}})^{1/2}|\boldsymbol{\phi^{0,\ast}}\rangle$
$\displaystyle+\sum_{i,j}\langle\phi_{i}|\hat{f}^{+}|\phi_{j}\rangle\langle{\mathbf{C}_{u}}^{k}|\hat{b}_{i}^{\dagger}\hat{b}_{j}\,|\mathbf{C^{0}}\rangle+\sum_{i,j}\langle\phi_{j}|\hat{f}^{-,\ast}|\phi_{i}\rangle\left\langle{\mathbf{C}_{v}}^{k}\left|\left(\hat{b}_{i}^{\dagger}\hat{b}_{j}\right)^{\ast}\right|\mathbf{C^{0,\ast}}\right\rangle$
(123)
where the vector
$\boldsymbol{|\phi^{0}\rangle}=\left(|\phi_{1}^{0}\rangle,...,|\phi_{M}^{0}\rangle\right)^{t}$
collects the stationary ground-state orbitals. Each response weight
$\gamma_{k}$ quantifies how strong the $k$-th excitation contributes to the
overall response of the system to an external perturbation with amplitudes
$\hat{f}^{+}$ and $\hat{f}^{-}$. For instance, it may happen that for a chosen
shape of the perturbing external fields, some excited states do not respond at
all and thus appear to be transparent.
The response density of the $k$-th excited state due to the perturbation can
in general be expressed as
$\Delta\rho^{k}(\mathbf{r})=\Delta\rho_{o}^{k}(\mathbf{r})+\Delta\rho_{c}^{k}(\mathbf{r})$
(124)
where $\Delta\rho_{o}^{k}(\mathbf{r})$ and $\Delta\rho_{c}^{k}(\mathbf{r})$
denote the orbitals’ and coefficients’ contribution, respectively. Those are
given by
$\displaystyle\Delta\rho_{o}^{k}$
$\displaystyle=\langle\boldsymbol{\phi^{0}}|(\boldsymbol{\rho^{0}})^{1/2}\left(\mathbf{|u^{k}\rangle}+\mathbf{|v^{k,\ast}\rangle}\right)$
$\displaystyle+\left(\mathbf{|u^{k}\rangle}+\mathbf{|v^{k,\ast}\rangle}\right)^{\dagger}(\boldsymbol{\rho^{0,\ast}})^{-1/2}\boldsymbol{\rho^{0}}\,\boldsymbol{|\phi^{0}\rangle}$
(125)
and
$\Delta\rho_{c}^{k}=\sum_{i,j=1}^{M}\,\phi_{i}^{0,\ast}\phi_{j}^{0}\left(\langle\mathbf{C}^{0}|\hat{b}_{i}^{\dagger}\hat{b}_{j}|{\mathbf{C}_{u}}^{k}\rangle+\langle{\mathbf{C}_{v}}^{k,\ast}|\hat{b}_{i}^{\dagger}\hat{b}_{j}|\mathbf{C}^{0}\rangle\right).$
(126)
To summarize, the many-body LR theory based on the MCTDHB working equations,
termed LR-MCTDHB, was introduced and derived. For a given $N$-bosons system in
a potential trap $V$ and a two-body interaction $\hat{W}$, the central
equation in order to obtain the excitation spectrum is Eq. (116). Although
this equation appears to be rather straightforward to solve, it in general
represents a demanding task because firstly, $\mathcal{L}$ is not hermitian
and secondly, it can become extremely large in size when the number of bosons
$N$ and/or the number of orbitals $M$ grows. In Appendix B, it is shown that
the LR matrix can be brought to block-diagonal form by using a complex
transformation, which halves the dimensionality of the eigenvalue problem.
However, it involves matrix-matrix products of the individual blocks of
$\mathcal{L}$, which can be numerically very expensive. It hence remains a
system-dependent question whether it is beneficial to perform this
transformation or not.
In the next section, Section 4, details on the numerical implementation of LR-
MCTDHB are presented. First, general rather technical challenges of the
eigenvalue problem in Eq. (116) are discussed before explaining the code
structure and the numerical methods utilized. Finally, it is once again
stressed that for $M=1$, LR-MCTDHB reduces to the (particle-conserving) BdG
equations, i.e., BdG$\equiv$LR-GP$\equiv$LR-MCTDHB(1). The derivation of the
BdG theory is described in Appendix A.1.2. Since the BdG is commonly used to
calculate excitations of trapped BECs, its results for certain applications
(see Section 6) are compared to the many-body results for $M>1$.
## 4 Numerical implementation of many-body linear response theory
In this section, information about the LR-MCTDHB implementation are presented.
Before dealing with the technical and numerical details in Sections 4.2 and
4.3, a brief overview about the general structure of the code is given in
Section 4.1, together with a description of technical challenges that were to
be solved.
### 4.1 Code structure and challenges
As a starting point, the structure of the numerical implementation, consisting
of several thousand lines of code, is explained. The entire task of
calculating the energies of excited states and the corresponding correction
amplitudes of a trapped interacting BEC, i.e., solving the eigenvalue problem
of Eq. (116), is separated into two major parts. Those are (i) the
construction of the LR matrix $\bar{\mathcal{L}}$ (simply denoted by
$\mathcal{L}$ in the following) as described in Eq. (3.2.2) and (ii) finding a
desired amount of eigenpairs, i.e., eigenvalues and the corresponding
eigenvectors, of the low-energy spectrum. To this end, two running modes were
implemented. Mode 1 is the construction mode. As an input, it requires a
stationary state of a system of $N$ bosons in a trap $V$ with interaction
$\hat{W}$, where the bosons can occupy $M$ single-particle orbitals.
Typically, this stationary state is the ground state of the system, and it can
therefore be obtained from a previous MCTDHB calculation using imaginary time
propagation. The output of mode 1 is the LR matrix $\mathcal{L}$. Afterwards,
the code can be restarted in mode 2, the diagonalization mode, which takes
$\mathcal{L}$ as an input and calculates a desired amount of eigenpairs in the
low-energy part of the excitation spectrum. Furthermore, the response weights
and response densities, described in Section 3.2.2, are calculated.
The main challenge of both the construction and the eigenvalue calculation is
the memory consumption of $\mathcal{L}$, and, especially in mode 1, the size
of intermediate matrices that appear during the construction, like the
projector $\mathbf{\mathcal{P}}$ or the individual blocks
$\mathcal{L}_{oo},\,\mathcal{L}_{oc},\,\mathcal{L}_{co}$ and
$\mathcal{L}_{cc}$. Although the memory consumption might be a minor issue for
systems in 1D with small values of $N$ and $M$, it can become a massive
problem for larger systems. To give an example, one can consider a system with
$N=100$ particles and $M=4$ orbitals. In such a case, already the coefficient
matrix $\mathcal{L}_{cc}$ has a dimensionality of $353702$, which means it
would need approximately $4$ terabytes of memory to store all matrix elements,
assuming those are in general complex numbers of 32 bytes. However, the
available working memory (RAM) of compute nodes on modern high performance
computation clusters is typically much smaller, being of the order of $128$ or
$256$ gigabytes. One possible solution of this problem is to store all (large)
matrices that appear during the construction of $\mathcal{L}$, as well as
$\mathcal{L}$ itself, in a sparse matrix storage format. There are multiple
possibilities to do that, e.g., the compressed-sparse-row (CSR) or compressed-
sparse-column (CSC) format (see, e.g., Ref. [58] for additional formats).
To obtain eigenpairs of $\mathcal{L}$ by running the code in mode 2, one is in
need of a numerical method that is capable of dealing with sparse and large
non-hermitian matrices. The method of choice is the implicitly restarted
Arnoldi method (IRAM), implemented in the ARPACK numerical library [59]. In
the current implementation, the parallel version, termed PARPACK, is utilized
[60]. The IRAM is presented in detail in Section 4.3. The PARPACK routines
have a significant advantage over eigenvalue routines from other libraries
like ScaLAPACK [61] because they do not require the matrix that one wishes to
diagonalize as an input, not even in any sparse format. It is only required to
hand over the result of matrix-vector multiplications (matvecs) as input, and
the actual calculation of these matvecs takes place outside of any PARPACK
routine, meaning that the implementation of the matvecs is completely
independent. It can be carried out separately, utilizing routines from other
numerical libraries, e.g., the Intel Math Kernel Library (MKL) [62], that
calculate matvecs of large and sparse matrices in a very efficient manner.
Moreover, calculating matvecs it is a process that can be easily parallelized
which further improves the code performance.
Due to the large amount of operations in both running modes, especially for
large $N$ and $M$ in two- or three-dimensional space, sequential runs of the
code would require very long runtimes. Therefore, a fully MPI-parallelized
implementation of the code was carried out for both running modes. Whereas
this was a comparatively straightforward task for mode 2 – since only the
matvecs needed to be parallelized – it turned out to be a demanding challenge
with respect to mode 1. A parallelization scheme that is a compromise between
taking care of the least possible memory consumption on the one hand, and
distributing the workload among all participating MPI processes (PEs) as equal
as possible on the other hand, is employed. In Section 4.2, the overall
construction process is explained in more detail.
### 4.2 Construction of the LR matrix
In this section, details on the implementation of the LR matrix construction
are presented. Figure 1 shows the structure of the code and its main steps. As
input, the code requires the stationary state to which the LR analysis should
be applied, usually the ground state of a trapped BEC. In particular, one
needs to hand over the set of orbitals $\\{\phi_{i}^{0}:1\leq i\leq M\\}$ as
well as the corresponding coefficients $\mathbf{C^{0}}$. Afterwards,
intermediate matrices are calculated which are essential to built up the
individual blocks of $\mathcal{L}$ as described in Eq. (93). This includes the
kinetic energy operator $\hat{T}_{\text{kin}}$, the direct and exchange
interaction operators $\hat{W}_{sl}$ and $\hat{K}_{sl}$, as well as the
projection operator $\mathbf{\mathcal{P}}$. All of these calculations are done
in parallel, and the final result is stored in a sparse storage format on the
master process PE 0.
Then, the matrices $\mathcal{L}_{oo}$ and $\mathcal{L}_{oc}$ are calculated.
Whereas $\mathcal{L}_{oo}$ is a $2(M\cdot N_{\text{grid}})$ square matrix with
$N_{\text{grid}}$ denoting the number of grid points, $\mathcal{L}_{oc}$ is a
$2(M\cdot N_{\text{grid}}\times N_{\text{conf}})$ matrix. It is stressed that
constructing these two matrices is usually the most time-consuming part of the
entire calculation of $\mathcal{L}$ due to the large amount of operations
involved. On the one hand, for problems where the grid is large (e.g., in 2D
and 3D) and $M\geq 7$, the construction of $\mathcal{L}_{oo}$ represents the
essential part of the calculations. On the other hand, if the number of
particles is large such that $N_{\text{conf}}=\mathcal{O}(10^{5})$ or higher,
the construction of $\mathcal{L}_{oc}$ becomes the most demanding task.
Additionally, both matrices need to be multiplied by
$\boldsymbol{\rho_{oo}}^{-1/2}\boldsymbol{\mathcal{P}}_{oo}$ from the left,
which further increases the amount of operations significantly. Because of
this, both the construction of $\mathcal{L}_{oo}$ and $\mathcal{L}_{oc}$ are
MPI-parallelized, as well as the subsequent multiplication with the metric and
projector matrices. As for the preliminary calculation described above, the
final results of
$(\boldsymbol{\rho_{oo}}^{-1/2}\boldsymbol{\mathcal{P}}_{oo}\mathcal{L}_{oo})$
and
$(\boldsymbol{\rho_{oo}}^{-1/2}\boldsymbol{\mathcal{P}}_{oo}\mathcal{L}_{oc})$
are stored on PE 0 in a sparse storage format.
Input: $\\{\phi_{i}^{0}\\},\,\mathbf{C^{0}}$ Preliminary calculations:
$\hat{T}_{\text{kin}}$ $\hat{W}_{sl}$ $\hat{K}_{sl}$ $\mathbf{\mathcal{P}}$ •
MPI-parallelized • On exit stored on PE 0 Calculate $\mathcal{L}_{oo}$
$\mathcal{L}_{oc}$ to do matrix multiplications
$\boldsymbol{\rho_{oo}}^{-1/2}\boldsymbol{\mathcal{P}}_{oo}\mathcal{L}_{oo}$
$\boldsymbol{\rho_{oo}}^{-1/2}\boldsymbol{\mathcal{P}}_{oo}\mathcal{L}_{oc}$ •
MPI-parallelized • On exit stored on PE 0 Row-by-row generation of upper
blocks of $\mathcal{L}$ • Each PE generates different row • Written to file by
PE 0 Row-by-row generation of lower blocks of $\mathcal{L}$ • For lower left
block: Each PE generates different row • For $\mathcal{L}_{cc}$: All PEs share
work of each row • Written to file by PE 0 Output: $\mathcal{L}$
Figure 1: Schematic diagram of the MPI-parallelized implementation for the
construction of $\mathcal{L}$. The parallelization scheme shares the amount of
calculations equally between all PEs. Before $\mathcal{L}$ is generated,
preliminary calculations of important intermediate matrices are performed. To
minimize the memory consumption, those are only stored on PE 0 in a sparse
storage format. For the row-by-row construction of the upper and lower blocks
of $\mathcal{L}$, PE 0 distributes all necessary information to the other PEs,
collects the results afterwards and writes them to a file. See text for
details.
Next, the generation of $\mathcal{L}$ starts. Therefore, the problem is
separated into the construction of the upper and lower part. With respect to
the upper blocks, this means that only the upper left block needs to be
multiplied by $\boldsymbol{\mathcal{P}}_{oo}\,\boldsymbol{\rho_{oo}}^{-1/2}$
from the right, whereas the upper right block is already fully calculated and
can be written to the output file directly. To efficiently calculate the
upper-left block, a row-by-row parallelization scheme is employed where each
PE calculates a different row at a time. Therefore, PE 0 distributes all
necessary matrix elements of
$(\boldsymbol{\rho_{oo}}^{-1/2}\boldsymbol{\mathcal{P}}_{oo}\mathcal{L}_{oo})$
and $\mathbf{\mathcal{P}}$ to the other PEs. Finally, all PEs send back their
calculated rows (i.e., the non-zero elements and their coordinates) to PE 0,
which writes the results to the output file.
Concerning the construction of the lower part, again a row-by-row
parallelization scheme is utilized. For the lower-left block, each PE
generates a different row at a time. However, in the case of
$\mathcal{L}_{cc}$, all PEs work on the same row at a time. Upon finalization,
the results are collected by PE 0 and written to the output file, again in a
sparse storage format.
This section is ended by briefly discussing the issue of small but non-zero
matrix elements. As already stated, the matrices produced are typically
sparse, i.e., most elements are zero. Moreover, many other elements can be
very small but nonzero. Their impact on the subsequent eigenvalue calculation,
which is described in the next sections, is very often negligible. However,
for a problem where the dimensionality of $\mathcal{L}$ is very large, those
elements can become challenging in terms of memory consumption. Thus, it is
important to define a tolerance value $\tau$ that is used to minimize the
amount of non-zero elements which need to be stored. In the current
implementation, a matrix element is treated as zero and is therefore not
stored if its magnitude is smaller than $\tau$. For very large matrices, the
value of $\tau$ has to be defined such that on the one hand not too many
matrix elements are set to zero in order to obtain numerically accurate
results for the eigenvalues and eigenvectors, but, on the other hand, it has
to be large enough because too many very small elements can significantly
decrease the performance of the numerical method for calculating eigenpairs of
$\mathcal{L}$. The latter numerical procedure will be explained in the next
section.
### 4.3 Calculation of eigenvalues
In order to calculate the lowest eigenvalues of a given LR matrix, the IRAM is
employed. In the following, the main ideas of this approach are presented by
first discussing the basic ingredients (the Arnoldi method, the QR-iteration,
and polynomial filtering) before dealing with the IRAM explicitly. In short,
it generates a Krylov subspace of a given matrix whose basis vectors are used
to build a much smaller matrix. The eigenvalues of this matrix are also
approximate eigenvalues of the original matrix. After each iteration of the
IRAM, the convergence of a set of wanted eigenvalues is checked, and if it is
still not converged, the entire process is restarted with an initial vector
where the directions of unwanted eigenvalues are projected out. Thus, the
convergence with respect to the wanted eigenvalues is gradually enhanced.
Below, the individual steps of the IRAM are explained in more detail.
#### 4.3.1 The Arnoldi algorithm
The first step is to calculate an orthogonal projection of a sparse, complex,
in general non-hermitian matrix $A\in\mathbb{C}^{n\times n}$. Therefore, the
Arnoldi algorithm is used which generates an orthonormal basis (ONB) of the
$k$-dimensional subspace given by
$\mathcal{K}^{k}(A,v)=\text{span}(v,Av,...,A^{k-1}v)$ (127)
where $v\in\mathbb{C}^{n}$ is an initial vector and $k\leq n$, typically even
$k\ll n$. $\mathcal{K}^{k}$ is called the $k$-dimensional Krylov subspace of
$A$. The entire Arnoldi algorithm is given by Algorithm 1. The ONB is obtained
by performing a Gram-Schmidt-orthonormalization at each Arnoldi step, i.e.,
making the vector $v_{j}$ in the $j$-th step orthogonal to all previous basis
vectors $v_{i},\,i\leq j,$ and normalizing it afterwards. The elements of the
ONB are called the Arnoldi vectors. If the resulting Arnoldi vector $v_{j+1}$
of the $j$-th step is non-zero, its norm is stored in the matrix element
$h_{j+1,j}$ of the matrix $\tilde{H}_{k+1}\in\mathbb{C}^{(k+1)\times k}$. If
$j\leq k$, the next Arnoldi step is initialized. After the step for $j=k$, the
iteration stops and returns the $k$-dimensional ONB of $\mathcal{K}^{k}$, plus
an additional orthonormal vector $v_{k+1}\in\mathbb{C}^{n}$, together with the
full matrix $\tilde{H}_{k+1}$. It is possible that for a certain step $j<k$,
the norm of the Arnoldi vector $v_{j+1}$ is zero, meaning that one has found
an ONB of an invariant Krylov subspace of $A$ with dimensionality $j$. In this
case, the Arnoldi algorithm finishes immediately.
1: $v_{1}\leftarrow v,\,||v||_{2}=1$
2: for $j=1,\,j\leq k$ do
3: $z\leftarrow Av_{j}$
4: for $i=1,\,i\leq j$ do
5: $h_{ij}\leftarrow\langle v_{i},z\rangle$
6: $z\leftarrow z-h_{ij}v_{i}$
7: end for
8: $h_{j+1,j}\leftarrow||z||_{2}$
9: if $h_{j+1,j}=0$ then
10: quit
11: else
12: $v_{j+1}\leftarrow z/h_{j+1,j}$
13: end if
14: end for
Algorithm 1 The Arnoldi algorithm
As an essential result, one obtains the Arnoldi decomposition of $A$ given by
$AV_{k}=V_{k}H_{k}+h_{k+1,k}\,v_{k+1}\,e_{k}^{t}$ (128)
where $V_{k}\in\mathbb{C}^{n\times k}$ contains the Arnoldi basis as columns
and $H_{k}\in\mathbb{C}^{k\times k}$ is an upper Hessenberg matrix, i.e.,
$\\{(H_{k})_{ij}\\}=0$ if $i>j+1$, built by the first $k$ rows of
$\tilde{H}_{k+1}$. The second term in Eq. (128), which contains the matrix
element $h_{k+1,k}$, the additional Arnoldi vector $v_{k+1}$ and the transpose
of the $k$-th unit vector, i.e., $e_{k}=(0,0,...,0,1)^{t}\in\mathbb{C}^{k}$,
takes care of the fact that the orthonormal projection of $A$,
$A\approx V_{k}H_{k}V_{k}^{\dagger},$ (129)
is in general not exact. However, for the case of having found an invariant
Krylov subspace of $A$ of lower dimensionality $j\leq k$, it becomes exact. It
is straightforward to realize that the eigenvalues of $H_{k}$ are approximate,
or, in the case of an invariant subspace, exact eigenvalues of A. This is due
to the cyclic permutation symmetry of the determinant that defines the
characteristic polynomial of $A$, i.e.,
$\displaystyle 0$ $\displaystyle=\det(A-\lambda
I)\overset{\text{Eq.\,}(\ref{Arnoldi_orth_proj})}{\approx}\det(V_{k}H_{k}V_{k}^{\dagger}-\lambda
I)$ $\displaystyle=\det(V_{k}[H_{k}-\lambda I]V_{k}^{\dagger})$
$\displaystyle=\det(\underbrace{V_{k}^{\dagger}V_{k}}_{=I}[H_{k}-\lambda I])$
$\displaystyle=\underbrace{\det(I)}_{=1}\,\det(H_{k}-\lambda I)$
$\displaystyle=\det(H_{k}-\lambda I)$ (130)
with $\lambda\in\mathbb{C}$ and $I$ denoting the identity matrix.
If $u_{i}\in\mathbb{C}^{k}$ is an eigenvector of $H_{k}$ and
$\theta_{i}\in\mathbb{C}$ the corresponding eigenvalue, the pair
$(\theta_{i},V_{k}\,u_{i})$ is called a Ritz pair of $A$ with Ritz value
$\theta_{i}$ and the associated Ritz vector $V_{k}\,u_{i}\in\mathbb{C}^{n}$.
In order to estimate how accurate this Ritz pair approximates an exact
eigenpair of $A$, one can calculate the residual norm $r$ defined as
$\displaystyle r$
$\displaystyle\equiv||A(V_{k}u_{i})-\theta_{i}(V_{k}u_{i})||_{2}\overset{\text{Eq.\,}(\ref{Arnoldi_decomp})}{=}||(V_{k}H_{k}+h_{k+1,k}\,v_{k+1}\,e_{k}^{t})u_{i}-\theta_{i}(V_{k}u_{i})||_{2}$
$\displaystyle=||\theta_{i}(V_{k}u_{i})+h_{k+1,k}\,v_{k+1}\,e_{k}^{t}\,u_{i}-\theta_{i}(V_{k}u_{i})||_{2}$
$\displaystyle=|h_{k+1,k}|\cdot\underbrace{|v_{k+1}|}_{=1}\cdot|e_{k}^{t}u_{i}|$
$\displaystyle=|h_{k+1,k}|\cdot|e_{k}^{t}u_{i}|,$ (131)
i.e., $r$ is the product of the magnitudes of the matrix element $h_{k+1,k}$
of $\tilde{H}_{k+1}$ and the $k$-th element of $u_{i}$. The smaller it is, the
better does the Ritz pair $(\theta_{i},V_{k}u_{i})$ approximate the exact
$i$-th eigenpair of $A$.
To summarize, the eigenvalue problem of a typically large matrix $A$ can be
mapped onto an eigenvalue problem of a much smaller matrix $H_{k}$ whose
eigenvalues are believed to be good approximations of the eigenvalues of $A$.
Due to the Hessenberg form of $H_{k}$, its eigenvalues can efficiently be
calculated by utilizing the QR-iteration, which is explained in the next
section. Additional details on the Arnoldi algorithm can be found in Ref.
[63].
#### 4.3.2 The QR-iteration
The QR-iteration is a technique to calculate the eigenvalues and eigenvectors
of a given matrix $A\in\mathbb{C}^{n\times n}$. It is based on the so-called
QR-decomposition which reads
$A=QR,\quad Q,R\in\mathbb{C}^{n\times n}$ (132)
where $Q$ is a unitary matrix, i.e., $QQ^{\dagger}=I$, and $R$ is upper
triangular. This decomposition exists for any matrix [63]. Plugging the ansatz
of Eq. (132) into the characteristic polynomial gives
$\displaystyle 0$ $\displaystyle=\det(A-\lambda I)$
$\displaystyle=\det(QR-\lambda I)$
$\displaystyle=\det(QQ^{\dagger}(QR)QQ^{\dagger}-Q\lambda IQ^{\dagger})$
$\displaystyle=\det(Q(RQ-\lambda I)Q^{\dagger})$
$\displaystyle=\det(RQ-\lambda I)$ (133)
which means that the eigenvalues of $A$ are the same as the eigenvalues of the
matrix $RQ$. Furthermore,
$A=QR\iff Q^{\dagger}AQ=RQ,$ (134)
i.e., $RQ$ is just a unitary transformation of $A$.
The QR-iteration is described in Algorithm 2. At each step $k$, the QR-
decomposition of the current matrix $A_{k}$ is used to define the matrix of
the next step as $A_{k+1}=R_{k}Q_{k}=Q_{k}^{\dagger}A_{k}Q_{k}$. It can be
shown that this sequence converges to a triangular matrix, i.e.,
$\lim_{k\rightarrow\infty}A_{k}=\tilde{A}$ with $\\{\tilde{A}_{ij}\\}=0$ if
$i>j$, and the eigenvalues of a triangular matrix are simply given by its
diagonal elements. One can show that the eigenvalues of $A_{k}$ are the same
as the ones of $A$ by considering again the characteristic polynomial
$\displaystyle 0$ $\displaystyle=\det(A_{k}-\lambda I)$
$\displaystyle=\det(Q_{k-1}^{\dagger}A_{k-1}Q_{k-1}-\lambda I)$
$\displaystyle=\det(Q_{k-1}^{\dagger}Q_{k-2}^{\dagger}A_{k-2}Q_{k-2}Q_{k-1}-\lambda
I)$
$\displaystyle=...=\det(Q_{k-1}^{\dagger}Q_{k-2}^{\dagger}...Q_{0}^{\dagger}\underbrace{A_{0}}_{=A}Q_{0}...Q_{k-2}Q_{k-1}-\lambda
I)$ $\displaystyle=\det(A-\lambda I)$ (135)
where the cyclic permutation symmetry of the determinant is used in the last
step.
Although the QR-iteration possibly requires a large amount of operations per
step for a general matrix $A$, it turns out to be of very low computational
cost with respect to upper Hessenberg matrices like the matrix $H_{k}$
obtained from a $k$-step Arnoldi process described in Algorithm 1. In such a
case, one can apply $(k-1)$ Givens rotations $G_{i}$, $i\in\\{1,...,k-1\\}$,
which are all unitary operations such that the overall QR-decomposition of
$H_{k}$ at each QR-step $j$ is given by
$(H_{k})_{j}=Q_{j}R_{j}=(G_{k-1})_{j}...(G_{1})_{j}R_{j}.$ (136)
Further details on Gives rotations can be found in Ref. [64]. With respect to
the IRAM, the QR-iteration turns out to be important not only because it is
used to calculate the eigenvalues of $H_{k}$, but also in terms of restarting
the Arnoldi iteration with an initial vector that incorporates shifts that
filter out the direction of unwanted eigenvectors of $H_{k}$. Those shifts are
obtained by utilizing QR-steps, which are explained in detail in Section
4.3.4. In Ref. [63], further details on the QR-iteration are presented.
1: $A_{0}\leftarrow A$
2: $\text{Calculate the}\,\text{QR-decomposition}\,A_{0}=Q_{0}R_{0}$
3: for $k=1\,\text{until convergence}$ do
4: $A_{k}\leftarrow R_{k-1}Q_{k-1}$
5: if $A_{k}$ ”triangular enough” then
6: quit
7: else
8: $\text{Calculate the}\,\text{QR-decomposition}\,A_{k}=Q_{k}R_{k}$
9: end if
10: end for
Algorithm 2 The QR-iteration
#### 4.3.3 Polynomial filtering
Before dealing with the IRAM in particular, a brief discussion on polynomial
filtering techniques of the initial vector $v$ of a Krylov subspace iteration,
like the Arnoldi algorithm of Section 4.3.1, is realized. In general, the
target of applying a filter to a vector $v$ is to make the contribution of
unwanted directions very small, or even zero.
To understand how polynomial filtering works in general, a square matrix
$A\in\mathbb{C}^{n\times n}$ whose eigenpairs are given by the set
$\\{(\lambda_{i},v_{i})|1\leq i\leq n\\}$ with $\lambda_{i}\in\mathbb{C}$ and
$v_{i}\in\mathbb{C}^{n}$ is considered. Suppose the eigenvectors form a basis.
Each arbitrary vector $v\in\mathbb{C}^{n}$ can then be written as
$v=\sum_{j=1}^{n}c_{j}v_{j}$ with complex expansion coefficients
$\\{c_{j}\\}$. Assuming that one wants to filter out the contribution of the
$k$-th eigenvector $v_{k}$, one can define
$p(t)\equiv t-\lambda_{k}$ (137)
and apply this very simple polynomial onto $v$:
$\displaystyle p(A)v$
$\displaystyle=\sum_{j=1}^{n}c_{j}p(A)v_{j}=\sum_{j=1}^{n}c_{j}(A-\lambda_{k})v_{j}$
$\displaystyle=\sum_{j=1}^{n}c_{j}(\lambda_{j}-\lambda_{k})v_{j}$
$\displaystyle=\sum_{j\neq k}^{n}c_{j}(\lambda_{j}-\lambda_{k})v_{j}.$ (138)
If $A$ is hermitian, the overlap with $v_{k}$ vanishes, i.e.,
$\langle v_{k}|p(A)v\rangle=\sum_{j\neq
k}^{n}c_{j}(\lambda_{j}-\lambda_{k})\underbrace{\langle
v_{k}|v_{j}\rangle}_{=\delta_{kj}}=0,$ (139)
because in that case the eigenvectors of $A$ would be orthogonal. Thus, the
vector $p(A)v$ is orthogonal to $v_{k}$, and any additional power iteration of
this vector with respect to $A$ preserves this property, meaning that
$\langle v_{k}|A^{n}p(A)v\rangle=\sum_{j\neq
k}^{n}c_{j}\lambda_{j}^{n}(\lambda_{j}-\lambda_{k})\underbrace{\langle
v_{k}|v_{j}\rangle}_{=\delta_{kj}}=0.$ (140)
Therefore, the Arnoldi iteration, applied to an initial vector $p(A)v$, yields
a Krylov basis which is orthogonal to $v_{k}$, and also the orthogonal
projection $H_{k}$ would not contain $\lambda_{k}$ in its spectrum of
eigenvalues. If $A$ is non-hermitian, the overlap of $p(A)v$ and $v_{k}$ is
not strictly zero since the eigenvectors $\\{v_{i}\\}$ can be linearly
dependent. However, within the IRAM, the overall contribution of $v_{k}$
gradually decreases upon increasing the number of restarts with properly
shifted initial vectors. Nevertheless, this shows that the LR-matrix
$\mathcal{L}$, which is in general non-hermitian, leads to additional
numerical effort which would be absent in the case of a hermitian matrix.
It is also possible to do multiple shifts at once by using a filter given by
$p(t)=\prod_{i=1}^{N}\left(t-\lambda_{\pi(i)}\right)$ (141)
where $N\leq n$ is the number of shifts and
$\pi(i)\in\\{1,...,n\\}\,\forall\,i$. In this case, the directions of $N$
different eigenvectors of $A$ are filtered out. The above choices of $p(t)$ do
only reflect the simplest possibilities of filter polynomials, but also more
sophisticated ones like the Chebyshev polynomials can be chosen [64].
#### 4.3.4 The implicitly restarted Arnoldi method (IRAM)
1: Take an initial vector $v_{1}\leftarrow v$, $||v||_{2}=1$
2: Perform a $k$-step Arnoldi iteration to get $\quad\quad
AV_{k}=V_{k}H_{k}+f_{k}\,e_{k}^{t}$ with $f_{k}:=h_{k+1,k}\,v_{k+1}$
3: Calculate the eigenvalues $\theta_{1},...,\theta_{k}$ of $H_{k}$ by using
the QR-iteration
4: while $m$ wanted eigenvalues of $H_{k}$ not converged do
5: Select $p=k-m$ unwanted eigenvalues $\theta_{1},...,\theta_{p}$
6: Perform $p$ QR-steps with unwanted eigenvalues as shifts:
$\quad\quad[H^{(p)}_{k},Q]:=QR[H_{k},\theta_{1},...,\theta_{p}],\quad
V^{(p)}_{k}:=V_{k}Q$
7: Set $f_{m}:=h_{k+1,k}\,Q_{kp}\,v_{k+1}+h^{(p)}_{m+1,m}\,v^{(p)}_{m+1}$
8: Define $H_{m}:=H^{(p)}_{k}(1:m,1:m)$, $V_{m}:=V^{(p)}_{k}(1:n,1:m)$
9: Perform $p$ additional Arnoldi steps onto $V_{m},\,H_{m}$ and $f_{m}$ to
obtain a $k$-step Arnoldi decomposition
10: Calculate the eigenvalues of the new $H_{k}$ by using the QR-iteration
11: end while
Algorithm 3 The implicitly restarted Arnoldi method (IRAM)
After the introduction of the three essential ingredients of the IRAM, namely
the Arnoldi algorithm, the QR-iteration, and polynomial filtering, one can
finally discuss the IRAM itself in detail. The full algorithm is given in
Algorithm 3. At first, a $k$-step Arnoldi iteration is performed, yielding an
Arnoldi decomposition of $A$ as described in Eq. (128). Then, the eigenvalues
$(\theta_{1},...,\theta_{k})$ are calculated utilizing the QR-iteration. From
this set, $p$ unwanted eigenvalues $(\theta_{1},...,\theta_{p})$ are selected.
If one for example seeks the eigenvalues of $A$ that have the smallest
magnitude, an appropriate selection would be the $p$ eigenvalues of $H_{k}$
that have the largest magnitude. Afterwards, a sequence of $p$ QR-shifts with
the unwanted eigenvalues is performed. For the first shift, the Arnoldi
decomposition reads
$\displaystyle(A-\theta_{1}I)V_{k}$
$\displaystyle=V_{k}(H_{k}-\theta_{1}I)+h_{k+1,k}\,v_{k+1}\,e_{k}^{t}$ (142)
$\displaystyle\iff(A-\theta_{1}I)V_{k}$
$\displaystyle=V_{k}Q_{1}R_{1}+h_{k+1,k}\,v_{k+1}\,e_{k}^{t}$ (143)
$\displaystyle\iff AV_{k}Q_{1}$
$\displaystyle=V_{k}Q_{1}(R_{1}Q_{1}+\theta_{1}I)+h_{k+1,k}\,v_{k+1}\,e_{k}^{t}\,Q_{1}$
(144)
where the QR-decomposition $H_{k}-\theta_{1}I=Q_{1}R_{1}$ is used in Eq.
(143). Setting
$\displaystyle H_{k}^{(1)}:=R_{1}Q_{1}+\theta_{1}I,\quad
V_{k}^{(1)}:=V_{k}Q_{1},\quad\left(e_{k}^{(1)}\right)^{t}:=e_{k}^{t}Q_{1}$
(145)
the next QR-shift can be applied:
$\displaystyle(A-\theta_{2}I)V_{k}^{(1)}$
$\displaystyle=V_{k}^{(1)}(H^{(1)}_{k}-\theta_{2}I)+h_{k+1,k}\,v_{k+1}\,\left(e_{k}^{(1)}\right)^{t}$
(146)
$\displaystyle\overset{H^{(1)}_{k}-\theta_{2}I=Q_{2}R_{2}}{\iff}AV_{k}^{(1)}Q_{2}$
$\displaystyle=V_{k}^{(1)}Q_{2}(R_{2}Q_{2}+\theta_{2}I)+h_{k+1,k}\,v_{k+1}\,\left(e_{k}^{(1)}\right)^{t}\,Q_{2}$
(147)
where one can define, similar to Eq. (145),
$\displaystyle H_{k}^{(2)}:=R_{2}Q_{2}+\theta_{2}I,\quad
V_{k}^{(2)}:=V_{k}^{(1)}Q_{2},\quad\left(e_{k}^{(2)}\right)^{t}:=\left(e_{k}^{(1)}\right)^{t}Q_{2}.$
(148)
This procedure is continued until all $p$ QR-shifts are applied.
Multiplying Eq. (143) by $e_{1}$ yields
$\displaystyle(A-\theta_{1}I)V_{k}e_{1}$
$\displaystyle=V_{k}Q_{1}R_{1}e_{1}+h_{k+1,k}\,v_{k+1}\,e_{k}^{t}e_{1}$
$\displaystyle\iff(A-\theta_{1}I)v_{1}$ $\displaystyle=V_{k}Q_{1}r_{11}e_{1}$
$\displaystyle\iff(A-\theta_{1}I)v_{1}$ $\displaystyle=v^{(1)}_{1}r_{11}$
(149)
where $r_{11}$ is the upper left element of $R_{1}$ and $v_{1}^{(1)}$ is the
first column of $V_{k}Q_{1}$. This in turn means that the first vector of the
new basis, i.e., $v^{(1)}_{1}$, is proportional to $(A-\theta_{1}I)v_{1}$. As
discussed in the previous section, this polynomial filter reduces the
contribution of the unwanted eigenvector corresponding to $\theta_{1}$ in
$v^{(1)}_{1}$, or, in the optimal case, makes its contribution even vanish
completely. This holds for all successive vectors in $V_{k}^{(1)}$. After
having applied the second QR-shift, one can show that
$v_{1}^{(2)}\sim(A-\theta_{2}I)(A-\theta_{1}I)v_{1}$. Continuing in the same
manner until the $p$-th QR-shift is applied yields
$v_{1}^{(p)}\sim\prod_{i=1}^{p}(A-\theta_{i}I)v_{1}$. Then, by taking the
first $m=k-p$ columns of $V_{k}^{(p)}$ as the new Arnoldi basis $V_{m}$, as
well as taking the leading principal $(m\times m)$-matrix of $H_{k}^{(p)}$ as
the new orthogonal projection $H_{m}$ of $A$, one obtains an $m$-step Arnoldi
decomposition $AV_{m}=V_{m}H_{m}+f_{m}\,e_{m}^{t}$ with an error term given by
$f_{m}=h_{k+1,k}\,Q_{kp}\,v_{k+1}+h^{(p)}_{m+1,m}\,v^{(p)}_{m+1}$ (150)
where it is used that the unitary matrix $Q$ is of Hessenberg form. Based on
this, one performs $p$ additional Arnoldi steps yielding again a $k$-step
Arnoldi decomposition. The essential difference to the previous $k$-step
Arnoldi decomposition is that now all basis vectors in $V_{k}$ are
approximately or even fully orthogonal to the eigenvectors of the associated
unwanted eigenvalues. The accuracy of the eigenvalues and eigenvectors of the
newly computed matrix $H_{k}$ can be estimated with the residual norm $r$
introduced in Section 4.3.1, Eq. (4.3.1). If it is below a predefined
tolerance for a desired amount of wanted eigenpairs, the algorithm finishes.
Otherwise, the set of eigenvalues of $H_{k}$ is again separated into wanted
and unwanted eigenvalues, and the procedure of QR-shifting and refining the
given Arnoldi basis restarts.
By employing the above described shift strategy without restarting the Arnoldi
iteration from the very beginning, one obtains the same result as if an
explicit restart of the entire $k$-step Arnoldi process with the initial
vector $v_{1}^{(p)}$ had been performed. Because of this similarity, the
overall algorithm is said to be implicitly restarted. The computational effort
in terms of matvecs with $A$ is given by $(k+p)$ operations in both cases,
i.e., for explicit and implicit restarting. However, implicit restarting
turned out to be numerically more stable, and has additional advantages which
are beyond the scope of this work.
Figure 2 shows the general structure of the code for the matrix
diagonalization utilizing the implementation of IRAM in the PARPACK library
[60]. Essential for the performance is the efficient parallel implementation
of the matvecs which are done outside of the PARPACK routine pznaupd. The
latter carries out the IRAM iterations in parallel [60]. It is important to
note that the IRAM, although it is particularly efficient for sparse matrices,
would also work for dense matrices. Further details are presented in Refs.
[64, 59, 63].
To summarize, the newly developed implementation of LR-MCTDHB was presented in
this section and the code was explained in detail. Furthermore, the IRAM, a
very powerful method for computing eigenpairs of very large and sparse
matrices, was described. In the next section, the code is benchmarked against
an exactly-solvable model.
Input: $\mathcal{L}$ Distribution of $\mathcal{L}$ • Each PE gets the same
amount of rows from the upper and lower blocks Define initial vector $v_{0}$
Arnoldi iteration • Call pznaupd (PARPACK routine that performs IRAM
iterations) Eigenvalues converged? Matvec $z:=Av_{i}$ Calculate eigenvectors,
response weights and response densities Outputnoyes
Figure 2: Schematic diagram of the MPI-parallelized implementation for the
diagonalization of $\mathcal{L}$. Different chunks of non-zero elements of
$\mathcal{L}$ are distributed among the PEs. The Arnoldi iteration is carried
out by the routine pznaupd of Parallel ARPACK (PARPACK) [60], which either
finishes upon convergence of a desired amount of eigenvalues or expects the
user to perform a matvec and return the result to it. Finally, the
eigenvalues, eigenvectors, as well as response weights and response densities
are calculated and written to output files. See text for details.
## 5 Comparison with an exactly-solvable model
After the presentation of the theoretical and numerical frameworks utilized,
along with the structure of the implemented code, the newly developed
implementation of LR-MCTDHB is benchmarked in the following. Furthermore,
since it builds the basis for the utilized LR theory, results from a benchmark
of the MCTDHB implementation used in this work [65], which makes use of a
sophisticated mapping of bosonic operators in Fock space [66], are presented
as well. This implementation is now also available with a recently developed
graphical user interface and works on all common operating systems [67]. Other
codes became available during the last years, e.g., the MCTDH-X software
package [68] which contains a separate graphical user interface. Furthermore,
the multi-layer (ML-)MCTDHB approach [69, 70, 71], which is particularly
suited for the description of Bose-Bose or Bose-Fermi mixtures, is now
implemented.
As a reference system, the exactly-solvable harmonic-interaction model (HIM)
is used. In the latter model, both the single-particle and interaction
potentials are of harmonic type. This model system has been used recently to
benchmark the fermionic counterpart of MCTDHB, termed MCTDHF [221], as well as
the time-dependent restricted-active-space self-consistent-field theory for
bosons (TD-RASSCF-B) [222] and its variant for bosonic mixtures [223]. In the
first part of this section, an introduction to the HIM and the analytic
results of its energy spectrum are given (Section 5.1). Afterwards, an
overview of the MCTDHB benchmark against the HIM, which has been carried out
in great detail in Ref. [33], is presented (Section 5.2). This includes both
the ground state and the out-of-equilibrium quantum dynamics of a bosonic
many-particle system. Finally, a comparison of the numerical results of the
low-energy spectrum obtained from LR-MCTDHB and the analytic values, both for
the 1D and 2D cases, is made (Section 5.3).
### 5.1 The harmonic-interaction model (HIM)
The HIM Hamiltonian contains the harmonic trap potential
$V(\mathbf{r})=\frac{1}{2}\left(\Omega_{x}^{2}x^{2}+\Omega_{y}^{2}y^{2}+\Omega_{z}^{2}z^{2}\right),\quad\mathbf{r}=(x,y,z)^{t}$
(151)
where the boson mass $m$ is set to unity. In $D>1$ dimensions, one
distinguishes between the isotropic and anisotropic HIM, where the former
refers to the case where all frequencies $\\{\Omega_{i}\\}$ are equal, whereas
the latter refers to the case where at least one frequency differs from the
others.
The interaction between the $i$-th and $j$-th boson is described as
$\hat{W}(\mathbf{r}_{i},\mathbf{r}_{j})=\lambda_{0}|\mathbf{r}_{i}-\mathbf{r}_{j}|^{2}$
(152)
where a positive (negative) interaction strength $\lambda_{0}$ denotes
attraction (repulsion) between the bosons.
As described in Ref. [72], the following coordinate transformation yields the
separation of the relative and center-of-mass (c.m.) coordinates,
$\mathbf{Q}_{k}=\frac{1}{\sqrt{k(k+1)}}\sum_{i=1}^{k}(\mathbf{r}_{k+1}-\mathbf{r}_{i}),\quad
1\leq k\leq N-1$ (153)
and
$\mathbf{Q}_{N}=\frac{1}{\sqrt{N}}\sum_{i=1}^{k}\mathbf{r}_{i},$ (154)
where $N$ is the number of bosons in the system. This leads to the Hamiltonian
in the c.m. frame given by
$\displaystyle\hat{H}$ $\displaystyle=\hat{H}_{rel}+\hat{H}_{c.m.}$ (155)
$\displaystyle\hat{H}_{rel}$
$\displaystyle=\frac{1}{2}\sum_{j=x,y,z}\sum_{k=1}^{N-1}(p_{j,k}^{2}+\delta_{j}^{2}Q_{j,k}^{2})$
(156) $\displaystyle\hat{H}_{c.m.}$
$\displaystyle=\frac{1}{2}\sum_{j=x,y,z}(p_{j,N}^{2}+\Omega_{j}^{2}Q_{j,N}^{2})$
(157)
with momenta $p_{j,N}=(1/i)\partial_{Q_{j,N}}$ of the c.m. coordinates,
$p_{j,k}=(1/i)\partial_{Q_{j,k}}$ of the relative coordinates, and the
parameters $\delta_{j}^{2}=\Omega_{j}^{2}\pm 2N|\lambda_{0}|$ where again the
plus (minus) sign denotes attraction (repulsion) between the bosons. For
simplicity $\hbar=1$ is assumed. It is instructive to interpret the
Hamiltonian as a composite model system of $(N-1)$ identical non-interacting
particles in a harmonic confinement with trap frequencies
$\mathbf{\delta}=(\delta_{x},\delta_{y},\delta_{z})^{t}$ and the c.m. particle
which moves in the harmonic confinement with the original trap frequencies
$\\{\Omega_{i}\\}$.
The ground-state wave function in the c.m. frame takes the form
$\Psi_{0}(\mathbf{Q}_{1},...,\mathbf{Q}_{N})=\prod_{j=x,y,z}\left(\frac{\delta_{j}}{\pi}\right)^{\frac{N-1}{4}}\left(\frac{\Omega_{j}}{\pi}\right)^{\frac{1}{4}}\exp\left(-\frac{1}{2\delta_{j}}\sum_{k=1}^{N-1}Q_{j,k}^{2}-\frac{\Omega_{j}}{2}Q_{j,N}^{2}\right),$
(158)
i.e., it is basically a product of $N$ harmonic oscillator (HO) ground-state
wave functions in each spatial direction [72]. However, in the laboratory
frame, the representation of the ground state is much more involved. Already
for $N=2$ bosons in 1D, it is an infinite sum of Hartree products of HO
eigenfunctions instead of a single product of two HO ground states in the c.m.
frame [33]. Numerical convergence towards this state is therefore a very
demanding challenge in the laboratory frame.
The exact solution for the energy levels reads
$\displaystyle E\left(\\{n_{j}\\},\\{m_{j}\\}\right)$
$\displaystyle=\sum_{j=x,y,z}\left\\{\left(n_{j}+\frac{N-1}{2}\right)\delta_{j}+\left(m_{j}+\frac{1}{2}\right)\Omega_{j}\right\\}$
(159)
where the $\\{n_{j}\\}$ denote the quantum numbers of excitations of the
relative coordinates and the $\\{m_{j}\\}$ denote the quantum numbers of c.m.
excitations in the $x$-, $y$\- and $z$-directions. In particular, this means
that the ground-state energy is given by
$E_{0}\equiv\varepsilon^{0}=\sum_{j=x,y,z}\left\\{\frac{(N-1)}{2}\delta_{j}+\frac{1}{2}\Omega_{j}\right\\}$
(160)
whereas the energy distance between any excited state and the ground state
reads
$\omega({\\{n_{j}\\},\\{m_{j}\\}})=\sum_{j=x,y,z}\left\\{n_{j}\delta_{j}+\Omega_{j}m_{j}\right\\}.$
(161)
Eqs. (160) and (161) will be used for the comparison to the numerical results
obtained from (LR-)MCTDHB. The degeneracies of excited states is discussed in
[73]. It is worth noting that for any number of bosons $N$, there is no
solution for an excitation corresponding to $\sum_{j=x,y,z}n_{j}=1$, i.e., a
single-particle excitation with respect to the relative coordinates. This is
due to the even permutation symmetry for identical bosons, Eq. (5), and it can
be shown by mathematical induction as follows. For the sake of simplicity, the
proof is restricted to the 1D case, i.e., $\mathbf{Q}_{i}=Q_{i}$, which does
not limit the generality since in higher dimensions, the Hamiltonian decouples
into the $x$-, $y$\- and $z$-components, see Eqs. (156) and (157). The wave
function corresponding to a first order excitation of the relative
coordinates, denoted in the following by $\Psi^{\prime}$, obeys
$\Psi^{\prime}(Q_{1},...,Q_{N})\sim\frac{\partial}{\partial
Q_{k}}\Psi_{0}(Q_{1},...,Q_{N}),\quad 1\leq k\leq N-1.$ (162)
For $N=2$, this yields
$\displaystyle\Psi^{\prime}(Q_{1},Q_{2})\sim
Q_{1}\Psi_{0}\sim(x_{2}-x_{1})\Psi_{0},$ (163)
see Eq. (158). Because the wave function needs to be fully symmetric under the
exchange $x_{1}\leftrightarrow x_{2}$, one finds
$\displaystyle\Psi^{\prime}\sim\left\\{(x_{2}-x_{1})+(x_{1}-x_{2})\right\\}\Psi_{0}=0,$
(164)
meaning that there is no first order relative excited state for two bosons.
Assuming that this is also true for $N$ bosons, i.e.,
$\Psi^{\prime}(Q_{1},...,Q_{N})\sim\frac{\partial}{\partial
Q_{k}}\Psi_{0}(Q_{1},...,Q_{N})=0\quad\forall\,1\leq k\leq N-1,$ (165)
one obtains for $(N+1)$ bosons and $K\leq N$
$\displaystyle\Psi^{\prime}(Q_{1},...,Q_{N+1})$
$\displaystyle\sim\frac{\partial}{\partial Q_{K}}\Psi_{0}(Q_{1},...,Q_{N+1})$
$\displaystyle\sim\sum_{i=1}^{K}(x_{K+1}-x_{i})\Psi_{0}$
$\displaystyle=\left(Kx_{K+1}-\sum_{i=1}^{K}x_{i}\right)\Psi_{0}$
$\displaystyle=\left(Kx_{K+1}-x_{K}-\sum_{i=1}^{K-1}x_{i}\right)\Psi_{0}$
$\displaystyle=\left(Kx_{K+1}-x_{K}-(K-1)x_{K}+\underbrace{(K-1)x_{K}-\sum_{i=1}^{K-1}x_{i}}_{\sim
Q_{K-1}}\right)\Psi_{0}$ $\displaystyle\sim
K\left(x_{K+1}-x_{K}\right)\Psi_{0}+Q_{K-1}\Psi_{0}$ (166)
where the argument of $\Psi_{0}$ is suppressed from the second line onward.
For the second term in Eq. (5.1), one can make use of the assumption from Eq.
(165) that the contribution of $Q_{K-1}\Psi_{0}$ to $\Psi^{\prime}$ vanishes
after symmetrization of the wave function. Moreover, the first term is
proportional to the relative coordinate of only two bosons, for which it was
already shown that its contribution to $\Psi^{\prime}$ vanishes as well after
symmetrization. Thus, one infers that also for $(N+1)$ bosons
$\Psi^{\prime}(Q_{1},...,Q_{N+1})=0$ (167)
which ends the proof.
### 5.2 Ground state and dynamics
At first, numerical results obtained from MCTDHB for the ground state of the
1D HIM for different particle numbers and the exact values from Eq. (160) are
compared. The index ’$x$’ is suppressed in the following for simplicity. The
trap frequency is set to $\Omega=1$ and the two-particle interaction strength
$\lambda_{0}$ is adjusted such that the mean-field interaction parameter
$\Lambda=\lambda_{0}(N-1)$ is kept fixed at $\Lambda=0.5$. Results are shown
in Table 1, which originally appeared in Ref. [33] where it has been
demonstrated that the MCTDHB method is much more efficient and accurate for
the description of the ground state and dynamics of a many-boson system than
the widely used FCI method. One observes that MCTDHB gives highly accurate
results for the ground-state energy for all considered particle numbers.
Furthermore, one needs less self-consistent orbitals $M$ in order to converge
to the exact ground-state energy when $N$ increases. This is anticipated,
since in the Hartree limit where $N\rightarrow\infty$ for fixed $\Lambda$, the
energy per particle converges to the GP value. What is however surprising is
the degree of accuracy. As an example, for $N=1000$ the relative error
$(\varepsilon^{0}_{M=3}-\varepsilon^{0}_{\text{exact}})/\varepsilon^{0}_{\text{exact}}$
is lower than $10^{-10}\,\%$. The fact that MCTDHB can very accurately
assemble the ground state of a large many-boson system with only very few
orbitals has far-reaching consequences for the accuracy of excitation energies
obtained by LR-MCTDHB. As discussed in Section 3.2.1, the latter will be
closer to the exact values if the ground state to which LR is applied is very
accurately described. Further examples for the ground-state energies in the 1D
and 2D HIM will be given in Section 5.3 and in Appendix C, where in particular
the convergence of excitation energies is discussed.
With regard to the dynamics, comparisons of numerical and exact results were
carried out in Ref. [33] for (i) the out-of-equilibrium quench dynamics and
(ii) the dynamics due to time-dependent driving of the trapping frequency and
the interaction strength. Concerning (i), Fig. 3 compares the exact, FCI and
MCTDHB results for the oscillations of the density $\rho$ of $N=2$ bosons
after a sudden quench of the interaction parameter from $\Lambda=0$ to
$\Lambda=0.5$. The initial state $\Psi_{0}$ is therefore the ground state of
the non-interacting system. In principle, the time-evolution is given by
$\displaystyle|\Psi(t)\rangle$
$\displaystyle=\sum_{n=0}^{\infty}c_{j}\,e^{-iE_{j}t}|\alpha_{j}\rangle,$
(168) $\displaystyle c_{j}$ $\displaystyle=\langle\alpha_{j}|\Psi(0)\rangle$
(169)
where the $\\{|\alpha_{j}\rangle\\}$ are an arbitrary basis of the two-
particle Hilbert space and the $\\{E_{j}\\}$ the corresponding eigenenergies.
One possible choice of this basis are the solutions of the time-independent
problem. It remains to compute the overlap integrals in Eq. (169). In
practice, the infinite sum in Eq. (168) has to be truncated, and for the case
described in Fig. 3, numerical convergence at each time step of the expansion
in Eq. (168) is ensured by including the $60$ lowest energy eigenstates of the
time-independent problem. However, utilizing only $M=4$ time-adaptive
orbitals, the MCTDHB calculations lead to a density oscillation that can no
longer be distinguished from the exact result. Already for $M=3$ the results
are highly accurate. If a fixed set of orbitals is used (FCI), even 8 orbitals
do not lead to the same accuracy than MCTDHB(4). This remarkably shows the
numerical benefit of using a time-adaptive, self-consistent basis set.
$M$ | $N=10$ | $N=100$ | $N=1000$
---|---|---|---
1 | 7.071067811865483 | 70.71067811865483 | 707.1067811865483
2 | 7.038769026303168 | 70.68016951747168 | 707.0764334257315
3 | 7.038350652406389 | 70.68012541218675 | 707.0764289871865
4 | 7.038348424909910 | 70.68012539174549 |
5 | 7.038348415349058 | 70.68012539173762 |
6 | 7.038348415311494 | |
7 | 7.038348415311018 | |
$\varepsilon^{0}_{exact}$ | 7.038348415311011 | 70.68012539173752 | 707.0764289869851
Table 1: Ground-state energies of $N=10,100,$ and $1000$ bosons in the 1D HIM
with interaction parameter $\Lambda=0.5$ and trap frequency $\Omega=1.0$.
Numerical results are shown for different numbers of orbitals $M$. Exact
results are obtained from Eq. (160). The table is adapted from Ref. [33].
Figure 3: Out-of-equilibrium dynamics of the one-particle density $\rho(x=0)$
for $N=2$ bosons after an interaction quench from $\Lambda=0$ to
$\Lambda=0.5$. Results are shown for different numbers of orbitals $M$. The
trap frequency is $\Omega=1.0$. For the time scale shown, the MCTDHB(4)
dynamics cannot be distinguished from the exact dynamics. The FCI method is
less accurate even for $8$ orbitals. All quantities are dimensionless. See
text for details. The figure is taken from Ref. [33]. Figure 4: Dynamics in
the HIM due to a time-dependent driving $f_{2}(t)$ of the trapping frequency
$\Omega(t)=\Omega_{0}\sqrt{1+f_{2}(t)}$ and the interaction strength
$\lambda(t)=\lambda_{0}\left[1-\frac{\Omega_{0}^{2}}{2N\lambda_{0}}f_{2}(t)\right]$.
The upper panel shows the complicated time-dependent profile of the driving
frequency $f_{2}(t)$. The lower panel shows the evolution of the c.m. energy
$\varepsilon_{2}(t)$ for different levels of MCTDHB($M$). For $N=10$ bosons,
the curve of MCTDHB(7) can hardly be distinguished from the exact result for
the shown time interval. The same holds true for $M=6$ and $N=50$ bosons. Even
for less orbitals, MCTDHB yields an accurate description of the evolution of
$\varepsilon_{2}(t)$ up to $t\approx 35$ for $N=10,\,M=6$ and $t\approx 25$
for $N=50,\,M=5$. On the contrary, the GP results do only follow the exact
curves for very short times and clearly fail to describe the long-time
behavior, even qualitatively. All quantities are dimensionless. See text for
details. The figure is taken from Ref. [33].
Fig. 4 shows the time-evolution of the c.m. energy $\epsilon_{2}(t)$ when both
the trapping frequency and the two-body interaction strength are time-
dependent and given by $\Omega(t)=\Omega_{0}\sqrt{1+f_{2}(t)}$ with a constant
$\Omega_{0}$ and
$\lambda(t)=\lambda_{0}\left[1-\frac{\Omega_{0}^{2}}{2N\lambda_{0}}f_{2}(t)\right]$.
The function $f_{2}(t)$ denotes a time-dependent driving frequency. In that
way, the trapping frequency of the $(N-1)$ relative coordinates remains time-
independent, i.e.,
$\delta=\sqrt{\Omega_{0}^{2}[1-f_{2}(t)]+2N\lambda_{0}\left[1+\frac{\Omega_{0}^{2}}{2N\lambda_{0}}f_{2}(t)\right]}=\sqrt{\Omega_{0}^{2}+2N\lambda_{0}}.$
(170)
Thus, although in the laboratory frame the bosons move in a time-dependent
trap and interact via a time-dependent potential, the same system in the c.m.
frame appears to be much less complicated because only the trapping potential
of the c.m. particle is time-dependent due to $f_{2}(t)$. Because the c.m. and
relative motions separate, the time-dependent energy of the system is given by
$E=\frac{N-1}{2}\delta+\varepsilon_{2}(t)$ where $\varepsilon_{2}(t)$ is
determined by the energy expectation value
$\varepsilon_{2}(t)=\langle\Psi_{N}(t)|\hat{H}_{c.m.}(t)|\Psi_{N}(t)\rangle$
(171)
of the c.m. particle. The state $|\Psi_{N}(t)\rangle$ is obtained from the
effective single-particle Schrödinger equation
$i\partial_{t}|\Psi_{N}(t)\rangle=\hat{H}_{c.m.}(t)|\Psi_{N}(t)\rangle$. From
Fig. 4, one can deduce that MCTDHB is capable to account for the dynamical
evolution of $\varepsilon_{2}(t)$ very accurately, even for longer times.
Furthermore, one observes that for a higher number of particles $N$ a smaller
number of time-adaptive orbitals $M$ is required to obtain the same level of
numerical accuracy. In contrast to that, the GP evolution of
$\varepsilon_{2}(t)$ only follows the exact curves of $N=10$ and $50$ bosons
for a very short time up to $t\approx 1$, and becomes remarkably inaccurate
afterwards. For the long-time behavior, it cannot even reproduce qualitatively
the profile of $\varepsilon_{2}(t)$.
To sum up, the previous examples clearly indicate that the MCTDHB theory
qualifies for describing the ground state, the dynamics due to a sudden quench
of system parameters as well as the dynamics due to time-dependent trapping
and interaction potentials, and that it is clearly superior to the FCI many-
body and GP mean-field methods. The full benchmark, together with additional
valuable details, can be found in Ref. [33].
### 5.3 Excited states
In this section, the applicability of LR-MCTDHB to obtain highly accurate
results for the energies of excited states of trapped multi-boson systems is
discussed. As for the benchmark concerning the ground state and dynamics, the
analytic results of the HIM from Eq. (161) are used to benchmark the numerical
results. In the following, LR-MCTDHB is first benchmarked against the 1D HIM
for repulsive bosons, i.e., with $\lambda_{0}<0$. Then, for the 2D case, the
isotropic HIM with $\Omega_{x}=\Omega_{y}$ is considered. In Appendix C,
further benchmarks against the 1D HIM with attractive bosons, against the
anisotropic case in 2D, as well as against the isotropic HIM in the rotating
frame of reference are presented.
| $M=1$ | $M=3$ | $M=4$ | $(m_{x},n_{x})$ | Exact
---|---|---|---|---|---
$\varepsilon^{0}$ | 4.527693 | 4.524922 | 4.524922 | $(0,0)$ | 4.524922
$\omega_{1}$ | 1.000000 | 1.000000 | 1.000000 | $(1,0)$ | 1.000000
$\omega_{2}$ | 1.811077 | 1.788855 | 1.788854 | $(0,2)$ | 1.788854
$\omega_{3}$ | n/a | 2.000005 | 2.000000 | $(2,0)$ | 2.000000
$\omega_{4}$ | 2.716616 | 2.683282 | 2.683282 | $(0,3)$ | 2.683282
$\omega_{5}$ | n/a | 2.788858 | 2.788854 | $(1,2)$ | 2.788854
$\omega_{6}$ | n/a | 3.000009 | 3.000000 | $(3,0)$ | 3.000000
$\omega_{7}$ | 3.622154 | 3.577961 | 3.573164 | $(0,4)$ | 3.577709
$\omega_{8}$ | n/a | 3.580111 | 3.577712 | $(0,4)$ | 3.577709
$\omega_{9}$ | n/a | 3.685038 | 3.681347 | $(1,3)$ | 3.683282
$\omega_{10}$ | n/a | 3.790078 | 3.788483 | $(2,2)$ | 3.788854
$\omega_{11}$ | n/a | 4.002097 | 3.999997 | $(4,0)$ | 4.000000
Table 2: Benchmark of the LR-MCTDHB implementation to the repulsive 1D HIM
with $N=10$ bosons. Shown are the ground-state energy $\varepsilon^{0}$ and
the energies $\omega_{i}$ of the first few excitations for different values of
$M$. The trapping frequency is $\Omega=1.0$ and the interaction strength is
$\lambda_{0}=-0.01$, yielding a marginal ground-state depletion of $f\approx
0.03\%$ for $M=4$. Whereas the BdG ($M=1$) spectrum misses many states, the
many-body spectra for $M=3$ and $4$ contain all low-lying excitations. Even
the two-fold degenerate state $(0,4)$, as proposed in Ref. [73], is obtained
(see $\omega_{7}$ and $\omega_{8}$). The assignment of the corresponding
single GP state to $\omega_{7}$ is arbitrary, and one would need to analyse
for instance the response density to identify whether it correlates with the
7th or 8th many-body excitation. The overall accuracy of excitation energies
increases with the number of orbitals. Underlined digits denote deviations
from the exact values from Eqs. (160) and (161). All quantities are
dimensionless. See text for more details.
For the benchmark against the 1D repulsive HIM, the system parameters are set
to $\Omega=1$ for the trap frequency and $\lambda_{0}=-0.01$ for the
interaction strength, which yields a slight depletion of $f\approx 0.03\%$ for
$N=10$ bosons and $M=4$ orbitals. Calculations were carried out on a grid from
$[-9,9)$ with 128 grid points. Results for the excitation energies relative to
the ground-state energy, $\omega_{i}=E_{i}-\varepsilon^{0}$, are shown in
Table 2. There are several comments to be made. At first, it can be observed
that the mean-field BdG calculation ($M=1$) misses several excited states in
the low-energy spectrum. This is due to the fact that the BdG theory by
construction does not account for excitations were several particles at a time
are excited from the ground-state manifold, so-called multi-particle
excitations. For further details see Appendix A.1.1. The second observation is
that the numerical accuracy is enhanced if more self-consistent orbitals are
included into the ground-state description. Also the two-fold degenerate
excitation $(0,4)$, as proposed in Ref. [73], is obtained. The accuracy of the
BdG spectrum, apart from the c.m. excitation $(1,0)$, is clearly lower than
for the many-body results of $M=3$ and $4$ orbitals. In addition, the
convergence towards the exact ground-state energy is observed.
Table 3 shows excitation energies obtained for the isotropic 2D HIM with trap
frequencies $\Omega_{x}=\Omega_{y}=1$ and $N=100$ repulsive bosons with
interaction parameter $\lambda_{0}=-0.001$. All computations were carried out
on a $[-9,9)\times[-9,9)$ grid with $64\times 64$ grid points. At first, one
sees that the BdG approach yields the energy of the lowest c.m. excited state,
which is two-fold degenerate, to perfect accuracy. All other states obtained
are pure relative excitations. Higher c.m. excited states, as well as
combinations of those with relative excitations, are missing in the BdG
spectrum.
$\omega_{i}(\gamma)$ | $M=1$ | $M=2$ | $M=3$ | $(m_{x},m_{y},n_{x}$+$n_{y})$ | Exact
---|---|---|---|---|---
$\varepsilon^{0}(1)$ | 89.554453 | 89.551373 | 89.548293 | $(0,0,0)$ | 89.548292
$\omega_{1}(2)$ | 1.000000 | 1.000000 | 1.000000 | $(1,0,0)$ | 1.000000
| 1.000000 | 0.999792 | 1.000000 | $(0,1,0)$ | 1.000000
$\omega_{2}(3)$ | 1.791089 | 1.788859 | 1.788859 | $(0,0,2)$ | 1.788854
| 1.791089 | 1.791089 | 1.788861 | $(0,0,2)$ | 1.788854
| 1.791089 | 1.791143 | 1.788859 | $(0,0,2)$ | 1.788854
$\omega_{3}(3)$ | n/a | 2.000438 | 2.000439 | $(2,0,0)$ | 2.000000
| n/a | n/a | 2.000657 | $(1,1,0)$ | 2.000000
| n/a | n/a | 2.000439 | $(0,2,0)$ | 2.000000
$\omega_{4}(4)$ | 2.686634 | 2.683292 | 2.683226 | $(0,0,3)$ | 2.683282
| 2.686634 | 2.684452 | 2.683226 | $(0,0,3)$ | 2.683282
| 2.686634 | 2.685519 | 2.683312 | $(0,0,3)$ | 2.683282
| 2.686634 | 2.686634 | 2.683312 | $(0,0,3)$ | 2.683282
$\omega_{5}(6)$ | n/a | 2.789191 | 2.787088 | $(1,0,2)$ | 2.788854
| n/a | 2.791314 | 2.787088 | $(1,0,2)$ | 2.788854
| n/a | 2.793711 | 2.789270 | $(1,0,2)$ | 2.788854
| n/a | n/a | 2.789270 | $(0,1,2)$ | 2.788854
| n/a | n/a | 2.791536 | $(0,1,2)$ | 2.788854
| n/a | n/a | 2.791535 | $(0,1,2)$ | 2.788854
Table 3: Benchmark of the LR-MCTDHB implementation to the isotropic HIM in 2D.
Results are presented for $N=100$ bosons and different numbers of orbitals
$M$. The trapping frequencies are $\Omega_{x}=\Omega_{y}=1.0$, and the
interaction strength is $\lambda_{0}=-0.001$. The depletion is marginal with
$f\approx 0.003\%$ for $M=2$ and $f\approx 0.006\%$ for $M=3$. Shown are the
energies of the ground state $\varepsilon^{0}$ and the first $5$ excited
states, $\omega_{i}=E_{i}-\varepsilon^{0}$, together with their multiplicities
$\gamma$. In the BdG case, $M=1$, only the first pure c.m. excitation is
found, together with all pure excitations of relative coordinates. For LR-
MCTDHB(2), one obtains more c.m. excited states in the $x$-direction, together
with combinations of excitations in the relative coordinates. For LR-
MCTDHB(3), one obtains the missing c.m. excited states in the $y$-direction
and their combinations with excitations of the relative coordinates. The
overall accuracy increases with $M$, with few exceptions in the spectrum for
$M=2$ which are discussed in the text. Underlined digits denote deviations
from the exact values from Eqs. (160) and (161). All quantities are
dimensionless. See text for more details.
In contrast to the BdG approach, LR-MCTDHB(2) gives access to more excited
states. Before discussing those, it is worth examining the two underlying
ground-state orbitals. The first one is a Gaussian in the trap center, whereas
the second one resembles a $p_{x}$-orbital. The degree of depletion is
marginal with $f\approx 0.003\%$. In addition to the states obtained for
$M=1$, the spectrum contains the second-order c.m. excitation in the
$x$-direction, as well as three out of six states corresponding to
$\omega_{5}$, which are excited states combining a first-order c.m. excitation
in the $x$-direction and relative excitations. The accuracy of levels that
were already accessible within the BdG approach has increased, e.g., for the
first state of $\omega_{2}$ or the states of $\omega_{4}$. Nevertheless, some
states have lost accuracy. The most remarkable example is the second first-
order c.m. excitation of $\omega_{1}$, which is a c.m. excitation in the
$y$-direction. Naturally the question arises why this is the case. The reason
for this is that the second ground-state orbital, as discussed above,
resembles a $p_{x}$-orbital. This means that the ground-state description for
$M=2$ orbitals shows a preferred direction, which in this case is the
direction along the $x$-axis. Apparently, the consequence of this is that
excitations in the $y$-direction either lose in accuracy or remain
inaccessible. A possible reason for the loss in accuracy is that, compared to
the BdG case, a small amount of probability weight of the first orbital has
been transferred to the second orbital that effectively only covers the
$x$-direction. It is stressed that rotating the second orbital by an arbitrary
angle, e.g., $\pi/4$ or $\pi/2$, leads to exactly the same spectrum, which
means that it does not matter for the numerical results which direction is
preferred.
Including $M=3$ orbitals significantly enhances the obtained spectrum. One
observes that both problems of the spectrum for $M=2$, i.e., the loss in
accuracy for certain states as well as the absence of excitations in the low-
energy spectrum, are solved. The third ground-state orbital resembles a
$p_{y}$-orbital, i.e., it is perpendicular to the preferred direction of the
$M=2$ ground state. One can deduce from this observations a fundamental
difference between the computation of the ground state of a given $N$-boson
system and its excited states. Whereas increasing the number of orbitals
always leads to a lower ground state energy, it can happen that the spectrum
for ($M+1$) orbitals contains states that are less accurate than for only $M$
orbitals. This is in particular likely to happen in 2D and 3D because
preferred directions may occur. As a result, in order to obtain accurate
excitation spectra, one has to pay much more attention to the symmetry of the
problem in higher spatial dimensions.
To summarize, it was demonstrated for the HIM that both the MCTDHB and LR-
MCTDHB implementations utilized in this work are well-suited to obtain
accurate results for the spectrum and dynamics of trapped interacting bosons.
Furthermore, the many-body approaches clearly improve the GP and BdG mean-
field methods. As mentioned above, further benchmarks are given in Appendix C.
## 6 Applications to many-body dynamics and excitations of BECs
Since the advent of MCTDHB in 2007, the amount of scientific publications
where it is utilized grew substantially. As one of the first applications, the
splitting of 1D repulsive, fully-coherent BECs by a potential barrier has been
investigated [74, 75], showing that it can lead to fragmented condensates.
Similar observations were made by splitting a radially-symmetric BEC in 2D
with a ring-shaped barrier [76]. Also the reverse is possible, meaning that
two initially independent and thus fragmented BECs with no overlap in space
can collide, interfere and build up coherence [77]. Also for attractive BECs,
the splitting due to a Gaussian-shaped barrier results in the formation of a
superposition of two distinct bosonic clouds, a so-called caton, which is not
describable at the GP level [78]. Moreover, it has been shown for the first
time that coherent attractive BECs in 1D can fragment even without a potential
barrier, namely when the energy exceeds a threshold value [79]. Fragmentation
persists and may even intensify when more orbitals are included in the
calculation [80]. In the attractive case full convergence is difficult to
achieve as it requires more orbitals, see also [81]. When an attractive BEC
propagates towards a potential barrier and scatters from it, the system also
evolves to be highly fragmented although it was initially condensed [82, 83].
The degree of fragmentation depends on the degree of transmission through the
barrier. Even bright-soliton trains, i.e., stable multi-hump matter waves of
attractive bosons, whose dynamics were believed to be fully describable at the
mean-field level, are shown to lose their initial coherence [84].
The splitting of BECs has been further investigated with respect to the
generation of number- or phase-squeezed states applying optimal control theory
[85, 86, 87, 89, 88]. Optimal control has also been utilized in 1D bosonic
Josephson junctions (BJJ), on the one hand to steer the system parameters such
that an enhancement of the Shapiro effect was observed [90], and on the other
hand to drive a BEC from an initial state to a target state at the quantum
speed limit [91]. Moreover, it has been demonstrated that Mach-Zehnder
interferometry with ultracold bosons is relatively robust against the
nonlinear interaction between the particles [92].
A remarkable correspondence between the onset of wave chaos at the GP level
and the onset of fragmentation at the many-body level has been found for BECs
that scatter from either shallow periodic or disordered potential landscapes
[93, 94], but the result holds also for other potentials. Thus, the
development of wave chaos, i.e., the exponential separation of initially close
states, can be seen as an indication that a many-body treatment is necessary.
The distance between states is measured utilizing the $L^{2}$ norm between two
Hilbert space vectors.
Of potentially high relevance is the demonstration how the positions of
individual particles can be constructed from the many-body wave function by
simulating single shots [95]. The latter are connected to experiments on
trapped BECs where typically an image of the density represents a histogram of
a single shot.
Over the last years, MCTDHB has been applied to a growing number of systems,
e.g., to study the breathing dynamics in 1D harmonic traps due to a quantum
quench [96] or the many-body effects in solitonic excitations of ultracold
BECs [97, 98]. Other examples are the correlated dynamics of a single atom
coupled to a trapped BEC by collisions [99] or the examination of the
structure of mesoscopic molecular ions [100], which contributes to the
currently very active research in this field.
Recently, MCTDHB has also been applied to dipolar BECs in a double well [101]
and in 1D lattices [102, 103]. In addition, the coupling of one- and two-
component BECs in a cavity to a radiation field has been of interest [104,
105, 106]. One of the main results of the latter research was the finding of a
fragmented superradiant phase when the pump power of the laser exceeds a
critical threshold. This fragmented superradiance cannot be explained by the
Dicke model which assumes the BEC to be a simple two-level system.
In the remaining part of this section, the focus is laid on several
applications where the tunneling dynamics in traps (Section 6.1) and the
dynamical fragmentation of initially coherent BECs (Section 6.2) is
investigated, both in 1D and 2D systems. Afterwards, all applications of LR-
MCTDHB that exist up to now are presented in detail (Section 6.3). Special
emphasis for the 2D systems under consideration is laid on the impact of
angular momentum on the appearance of many-body effects in the dynamics and
excitation spectra.
### 6.1 Tunneling dynamics in traps
Subsequently, the tunneling dynamics of BECs held in double-well traps in both
one and two spatial dimensions are discussed in this section. The main purpose
is to apply MCTDHB to solve the many-boson Schrödinger equation and to
discover novel phenomena that are not captured by the standard methods.
#### 6.1.1 Exact tunneling dynamics in a bosonic Josephson junction
The quantum dynamics of a tunneling BEC in a 1D BJJ have constituted a large
research field in the past decades. Before the first experimental realizations
[107, 108], theoretical predictions concerning Josephson oscillations and
self-trapping, i.e., the suppression of tunneling between the wells, were
obtained by a (multi-mode) GP description [109, 110, 111, 112, 113, 114], a
path-integral formulation [115], a two-mode Bose-Hubbard (BH) approach [116]
or a Fock-space WKB method [117]. A review of various theoretical and
experimental results is given in [118]. A more recent study dealt with the
tunneling dynamics in a controlled BJJ, where the spin state of an ionic
impurity controls the tunneling between the wells [119]. The latter extends
the findings on the dynamics of trapped ultracold BECs due to the presence of
an ion [120, 121]. Furthermore, the orbital Josephson effect due to a time-
dependent driving potential was investigated numerically by employing MCTDHB,
not only in a double-well system [122].
The main objective of this section is however to study the out-of-equilibrium
tunneling dynamics of a single-component BEC in a BJJ from a many-body
perspective, and the method of choice is MCTDHB. As described below, the
latter approach allows for the observation of fundamentally new physics
dealing with, e.g., reduced self-trapping and dynamical fragmentation of
initially fully-coherent bosonic clouds. A comparison to other popular
theoretical tools like the GP approach and the BH model is made. The findings
presented below were recently published in Refs. [123, 124, 125, 126], where
further details can be found. Recently, results of a comparable study where
both the BH model and MCTDHB were used to analyze the tunneling of few-boson
systems in asymmetric double-wells, investigating the interaction blockade
that isolates the motion of a single particle in the vicinity of others,
became available [127].
The trapping confinement considered here reflects two harmonic wells of the
form
$V_{\pm}(x)=\frac{1}{2}(x\pm 2)^{2}$ (172)
which are connected by a cubic spline in the interval $|x|\leq 0.5$ that
generates a potential barrier between the wells. The plus sign refers to the
left well, whereas the minus sign refers to the right well. The repulsion
between the bosons is modeled by the contact interaction potential,
$\lambda_{0}\hat{W}(x,x^{\prime})=\lambda_{0}\delta(x-x^{\prime})$. The
repulsion strength is given by the mean-field parameter
$\Lambda=\lambda_{0}(N-1)$. An important quantity with respect to the
tunneling dynamics is the occupation probability of the left and right wells,
where the former is given by
$P_{L}(t)=\frac{1}{N}\int_{-\infty}^{0}\rho(x,t)dx$ (173)
where $\rho(x,t)$ describes the time-dependent one-body density, see Eq.
(2.4), and $N$ refers to the number of particles in the trap.
Since the MCTDHB results are compared to the BH model below, it is instructive
to define the left- and right-well orbitals, $\phi_{L}$ and $\phi_{R}$, with
the ground and the first excited state of the trap, denoted by $\phi_{g}$ and
$\phi_{u}$ due to their gerade and ungerade symmetry. These two states are
energetically well below the barrier. The orbitals $\phi_{L}$ and $\phi_{R}$
are defined via
$\phi_{L,R}(x)=\frac{\phi_{g}(x)\pm\phi_{u}(x)}{\sqrt{2}}.$ (174)
The BH parameters for on-site interaction $U$ and hopping $J$ are derived from
$\phi_{L}$ and $\phi_{R}$ by
$U=\lambda_{0}\int|\phi_{L}(x)|^{4}\,dx,\quad
J=-\int\phi_{L}^{\ast}(x)\hat{h}(x)\phi_{R}(x)\,dx$ (175)
and utilized to define the BH interaction parameter
$\Lambda_{\text{BH}}=UN/(2J)$. The two-site BH Hamiltonian then reads
$\hat{H}_{\text{BH}}=-J(\hat{b}_{L}^{\dagger}\hat{b}_{R}+\hat{b}_{R}^{\dagger}\hat{b}_{L})+\frac{U}{2}(\hat{b}_{L}^{\dagger}\hat{b}_{L}^{\dagger}\hat{b}_{L}\hat{b}_{L}+\hat{b}_{R}^{\dagger}\hat{b}_{R}^{\dagger}\hat{b}_{R}\hat{b}_{R})$
(176)
where the operators $\hat{b}_{L,R}^{(\dagger)}$ annihilate (create) a boson in
the left and right well, respectively. Within the two-mode BH approach, the
eigenvalues and eigenfunctions of the one-body RDM
$\rho^{(1)}=\begin{pmatrix}\rho_{LL}&\rho_{LR}\\\
\rho_{RL}&\rho_{RR}\end{pmatrix}$ (177)
with $\rho_{LL}=\langle
N,0|e^{+i\hat{H}_{\text{BH}}t}\,\hat{b}^{\dagger}_{L}\hat{b}_{L}\,e^{-i\hat{H}_{\text{BH}}t}|N,0\rangle$
(with $\hbar=1$) and the other elements defined analogously, represent the
natural occupations and natural orbitals of the BH method. Therefore, the
population of the second natural orbital can be used as a measure for
fragmentation.
In the following, the tunneling dynamics when the BEC is initially trapped in
the left well, i.e., $P_{L}(0)=1$, will be explored for different repulsion
strengths. Most importantly, for a two-mode GP description of the system,
self-trapping of the bosons is predicted for
$\Lambda_{\text{BH}}>\Lambda_{c}=2$ when the magnitude of the initial
imbalance between the occupations of the two wells, denoted by
$z=\frac{N_{L}-N_{R}}{N}$, is unity [109, 110, 111].
Figure 5: Short-time dynamics of the occupation probability of the left well,
$P_{L}(t)$, for (a) $N=20$, $\Lambda=0.152$, (b) $N=100$, $\Lambda=0.152$
($\Lambda_{\text{BH}}<\Lambda_{c}$) (c) $N=20$, $\Lambda=0.245$, and (d)
$N=100$, $\Lambda=0.245$ ($\Lambda_{\text{BH}}>\Lambda_{c}$). The GP [dotted
black], BH [dashed magenta] and numerically exact results from MCTDHB($M$)
[solid blue] are compared. For the weak interaction in panels (a) and (b), the
GP theory predicts Rabi oscillations of the full density. At the many-body
level, a collapse of these oscillations occurs after approximately three
cycles. For stronger interaction in panels (c) and (d), GP still predicts that
all bosons tunnel between the wells, whereas at the many-body level tunneling
is clearly suppressed. However, MCTDHB and the BH model yield different
dynamics in all cases. The deviations become more pronounced for stronger
repulsion. The inset shows the numerical convergence with respect to $M$ for
2, 4, 6 and 8 orbitals [(a) and (c)] and 2 and 4 orbitals [(b) and (d)]. All
quantities are dimensionless. See text for details. The figure is taken from
Ref. [123]. Figure 6: Upper panel: Dynamics of $P_{L}(t)$ for very strong
repulsion. The interaction parameter $\Lambda_{\text{BH}}=47.8$ ($43.4$) for
$N=10$ [solid blue] ($N=100$, solid green) is clearly above the critical value
of $\Lambda_{c}=2$. While the BH predicts full self-trapping of the bosons
[dashed magenta], the exact results show tunneling between the wells. Lower
panel: Evolution of the natural occupations $n_{i}^{(1)}(t)$. At the BH level,
the system remains condensed throughout the propagation time, while it
fragments at the exact many-body level where essentially four orbitals become
macroscopically occupied. All quantities are dimensionless. See text for
details. The figure is adapted from Ref. [123]. Figure 7: BH versus exact
dynamics for both attractive and repulsive BECs with system parameters $N=20$,
$\frac{|U|}{J}=0.226$ and $|\lambda_{0}|=0.0129$. Left panel: Occupation
probability $P_{L}(t)$ of the left well. While the BH predicts the same
evolution for attraction and repulsion, differences can be observed at the
exact many-body level. Inset: Time-evolution of the ratio of $P_{L}(t)$ for
attraction and repulsion. Right panel: The same for the first two natural
occupation numbers $n_{i}^{(1)}(t)$. Again, the BH model predicts equivalent
dynamics for attraction and repulsion, whereas the exact results show clear
differences. All quantities are dimensionless. See text for details. The
figures are taken from Ref. [124].
Fig. 5 shows a comparison between the short-time dynamics of $P_{L}(t)$ for
BECs with $N=20$ ($100$) bosons obtained from MCTDHB (solid blue), BH (dashed
magenta) and GP (dotted black) for two different interaction strengths where
one is below and one is above the critical value $\Lambda_{c}$. One observes
that for the case of $20$ bosons with weak interaction parameter
$\Lambda_{\text{BH}}<\Lambda_{c}$ [panel (a)], the GP prediction is totally
different than the BH and numerically exact MCTDHB predictions. According to
GP, Rabi oscillations of the entire cloud between the wells occur, whereas the
BH model and MCTDHB predict a density collapse after approximately three Rabi
cycles, resulting in roughly $50\%$ of the bosons in each well. The time for
one Rabi cycle is close to the theoretical prediction of
$t_{\text{Rabi}}=\pi/J$. Although the BH curve is clearly superior to the GP
curve, it also deviates from the numerically exact result after half a Rabi
cycle. In the case of $100$ bosons [panel (b)], similar observations can be
made. However, the density collapse occurs after a longer time.
As mentioned above, for stronger repulsion with
$\Lambda_{\text{BH}}>\Lambda_{c}$, tunneling should be entirely suppressed due
to self-trapping according to a two-mode GP description. It can be deduced
from the exact dynamics of $20$ bosons [panel (c)] that indeed less bosons
tunnel between the wells, meaning that the oscillation amplitude is clearly
smaller than for the weaker repulsion. As predicted by MCTDHB, at most $50$
bosons tunnel into the right well after half a Rabi cycle, and even less in
the BH dynamics. Nevertheless, tunneling is not completely suppressed. This
means that the effect of self-trapping, as predicted by the two-mode GP
theory, is obviously reduced at the accurate many-body level. The deviations
between the MCTDHB and BH results are more pronounced for stronger than for
weaker repulsion. The dynamics of $N=100$ bosons yields similar observations
[panel (d)]. In contrast to that, the GP theory predicts tunneling of the
whole cloud.
The reason why GP fails to describe the dynamics correctly is the development
of fragmentation, already on the short time scales shown. For the case of weak
repulsion in the upper panels of Fig. 5, the initially condensed BEC with
$f=10^{-4}$ ($10^{-5}$) for $N=20$ ($100$) evolves to be fragmented by $33\%$
($26\%$) after three Rabi cycles. Interestingly, for stronger repulsion, the
degree of fragmentation of the again initially fully-coherent BEC is less
after three Rabi cycles than for the case of weaker repulsion, yielding
approximately $28\%$ and $18\%$ for $20$ and $100$ bosons, respectively. The
predicted GP dynamics differ both quantitatively and qualitatively from the
exact results. Most important, for all system parameters used in Fig. 5, GP
predicts that essentially all bosons tunnel back and forth between the wells.
This is clearly not the case at the many-body level, neither for BH nor for
MCTDHB. It is stressed that although the chosen system parameters are within
the expected regime of validity for both the GP and BH approaches, both
theories fail to describe the short-time dynamics of the BECs.
To further illustrate the effect of reduced self-trapping as well as the
degree of fragmentation, Fig. 6 shows the evolution of $P_{L}(t)$ and of the
occupation of the first few natural orbitals $n_{i}^{(1)}(t)$, both for MCTDHB
and the BH model. The utilized repulsion parameter exceeds the critical value
significantly with $\Lambda_{\text{BH}}=47.8\,(43.4)$ for $N=10\,(100)$. From
the top panel, it can be inferred that the BH model predicts complete self-
trapping for both values of $N$, meaning that all bosons remain in the left
well throughout the time-evolution. On the contrary, the predictions of MCTDHB
show that indeed bosons tunnel from left to right, and one can anticipate that
$P_{L}(t)$ approaches the long-time average of $0.5$. With respect to the
natural occupations, one observes that the BEC remains fully-condensed in time
at the BH level, while the MCTDHB results show that fragmentation sets in
already during the first Rabi cycle. Moreover, more than ten orbitals are
necessary to accurately describe the dynamics. The different results are
associated with a quick loss of coherence between the bosons that is only
described by MCTDHB and not at the BH level. In this context, see Ref. [123]
for more details.
Another interesting aspect is the symmetry of the two-mode BH Hamiltonian in
Eq. (176) under the unitary transformation
$\hat{R}=\\{\hat{b}_{L}\rightarrow\hat{b}_{L},\hat{b}_{R}\rightarrow-\hat{b}_{R}\\}$,
leading to $\hat{R}\hat{H}_{\text{BH}}(U)\hat{R}=-\hat{H}_{\text{BH}}(-U)$. A
consequence of this is the equivalence of the occupation of the wells for
attractive and repulsive interactions, i.e.,
$P_{L,R}(t;U)=P_{L,R}(t;-U).$ (178)
Moreover, it can be shown that the eigenvalues of the BH one-body RDM in Eq.
(177) do not depend on the sign of the interaction. Thus, $P_{L}(t)$ and the
occupations $n_{i}^{(1)}(t)$ will be the same for attraction and repulsion of
the same magnitude. Naturally, the full many-body Hamiltonian of the system
does not possess such a symmetry, and one can therefore expect different
dynamics for attraction and repulsion between the bosons.
The left panel of Fig. 7 shows the short-time dynamics of $p_{L}(t)$ for a BEC
with $N=20$ bosons with interaction parameter $|U|/J=0.226$, corresponding to
$|\lambda_{0}|=0.0129$. As anticipated, the evolution at the BH level is
identical for attraction and repulsion. However, the results obtained by
MCTDHB show differences between the cases of positive and negative
$\lambda_{0}$. On the time scale shown, the BH model mainly underestimates
tunneling for the repulsive case, whereas it overestimates tunneling for the
attractive case. The right panel of Fig. 7 shows the evolution of the first
few natural occupation numbers for the same system as in the left panel, both
at the BH and full many-body level. As expected, the BH model predicts the
same evolution for $n_{1}^{(1)}(t)$ and $n_{2}^{(1)}(t)$, irrespective of the
sign of the interaction parameter. In contrast to that, the results obtained
from MCTDHB show that the evolution of the different occupations are distinct
for the attractive and repulsive cases. Compared to the exact results, the BH
model mainly overestimates the fragmentation of the attractive BEC, whereas it
mainly underestimates it for the repulsive BEC.
Figure 8: Universality of the degree of fragmentation. Upper panel: Evolution
of $P_{L}(t)$ for different numbers of bosons ($N=100$ in blue, $N=1000$ in
red and $N=10000$ in orange) where the interaction parameter
$\Lambda=\lambda_{0}(N-1)=0.152$ is kept fixed. Intermediate particle numbers
($N=200,\,500,\,2000$ and $5000$) are shown in gray. The less particles are
contained in the BEC, the faster the density collapses. Lower panel: Evolution
of the first two natural occupations. Irrespective of the number of bosons,
the degree of fragmentation upon the density collapse is the same. All
quantities are dimensionless. See text for details. The figure is taken from
[126].
The analysis of the dynamics in a 1D BJJ is closed by discussing the
compelling feature of universal fragmentation. The latter means that all BECs
with identical interaction parameter $\Lambda$ fragment to the same value
after the density has collapsed. Fig. 8 shows exactly this behavior for
$\Lambda=0.152$ and BECs with $N=100$, $1000$, and $10000$ particles where
initially all bosons are kept in the left well, i.e., $P_{L}(0)=1$. From the
upper panel, one sees that the density oscillations between the wells collapse
after a few Rabi cycles, leading to the equal distribution of $50\%$ of the
bosons in each well. In general, the less bosons are contained in the BEC, the
faster the density collapses. Coming along with the damping of the
oscillations, the BEC starts to fragment, reaching its maximal value after the
collapse. Again, the less bosons are in the system, the faster the system
fragments. Most importantly, irrespective of $N$, the final degree of
fragmentation is the same. The fact that the degree of fragmentation appears
to be universal is unexpected, since usually fragmentation strongly depends on
the number of particles in the BEC. Moreover, the degree of fragmentation upon
the density collapse, denoted by $f_{\text{col}}$, depends highly on the
initial imbalance $z$ of bosons in the left and right wells. The larger $z$,
the stronger does the system fragment in time. This holds true even in the
limit of very weak interactions, $\Lambda\ll 1$, meaning that there is no
weak-interaction regime where GP yields the correct dynamics. It is important
to note that the universality is not an artifact of any approximation, since
the result is obtained by solving the full many-boson Schrödinger equation
numerically exact. Interestingly, the two-mode BH approach can be used to
derive an analytic prediction of $f_{\text{col}}$ yielding results close to
the exact ones obtained by MCTDHB. Additional details on the analytic
expression can be found in Ref. [126].
To summarize, it was demonstrated that the out-of-equilibrium tunneling
dynamics of BECs in a 1D BJJ show significant many-body features that are not
accounted for at the GP mean-field level. This includes the appearance of a
density collapse in the oscillations between the wells as well as dynamical
fragmentation which the GP theory cannot capture by construction. The degree
of fragmentation is universal for fixed $\Lambda=\lambda_{0}(N-1)$ for a wide
range of initial states. Applying the commonly used BH model, it was shown
that it is superior in certain aspects in comparison to the GP mean-field
approach. However, it does also not predict the numerically exact tunneling
dynamics as obtained by MCTDHB, e.g., on the short-time dynamics where the
interaction parameter is in the vicinity of its critical value or for very
strong repulsion where it predicts complete self-trapping of the bosonic
cloud. However, the BH model can be utilized to derive an analytic expression
for the universal degree of fragmentation that is in good agreement with the
exact results.
#### 6.1.2 Many-body tunneling dynamics in a two-dimensional radial double
well
After having shown the many-body nature of the tunneling dynamics of BECs in a
1D BJJ, the question of how a repulsive BEC behaves in a radially symmetric 2D
system with a double-well structure is addressed. In particular, as for the
previous section, the out-of-equilibrium tunneling dynamics are of interest,
and it is studied whether many-body effects can be observed. The results
presented in this section were published in Ref. [128], and additional details
can be found therein.
Figure 9: (a) Schematic plot of the radial double well. The IN region denotes
the trap center, the OUT region denotes the outer annulus. Both are separated
by the central ring-shaped barrier. (b) Ground-state energy per particle of a
BEC with $N=100$ bosons and total angular momentum $L=0$ for different
positions $R_{B}$ of the radial barrier. The results are calculated at the
mean-field GP level ($M=1$). The quantities $E_{0}^{\text{IN}}$ (short-dashed
red) and $E_{0}^{\text{OUT}}$ (solid green) denote the energies for the IN and
OUT regions for interaction strength $\Lambda=0$, whereas $E_{2}^{\text{IN}}$
(dash-double-dotted blue) and $E_{2}^{\text{OUT}}$ (dash-dotted magenta)
denote the energies for $\Lambda=2$. The crossing points, i.e., the radii at
which the energies for IN and OUT are alike, are located at $R_{1}=3.271$
($\Lambda=0$) and $R_{2}=3.412$ ($\Lambda=2$), see the vertical black lines.
Larger repulsion between the bosons shifts the crossing point to larger radii.
(c) Same analysis as in (b) but for total angular momentum $L=N$, meaning that
each particle carries on average one unit of angular momentum. The location of
the crossing points are $R_{3}=4.089$ ($\Lambda=0$) and $R_{4}=4.093$
($\Lambda=0.2$). The angular momentum has a significant effect on the position
of the crossing points. Increasing $L$ shifts them to larger radii, which is
qualitatively similar as for increasing $\Lambda$. However, one would need to
highly increase the interaction strength in order to shift the crossing points
as much as adding one unit of angular momentum per particle does. All
quantities are dimensionless. See text for details. Figs. (b) and (c) are
adapted from Ref. [128].
To study tunneling phenomena in 2D geometries has been of high interest in
recent years, especially for trapped vortices, i.e., macroscopic quantum
states with definite angular momentum, characterized by a density node at its
core and phase discontinuities around it. Research concerning tunneling of
vortices has been carried out for example in 2D superfluids [129]. Others
dealt with vortices tunneling through a Gaussian-shaped barrier [130], in
pinning potentials [131, 132, 133], or between two Gaussian wells [134], to
mention only a few. However, there has been no investigation of the tunneling
dynamics of vortices in a radially-symmetric double well, in particular not at
the many-body level. The latter geometry is very interesting to study many-
body dynamics since the trap symmetry preserves the angular momentum of the
bosonic cloud. It has been utilized recently to study many-body vortices which
are separated in space due to the central barrier [135]. The following
results, recently published in Ref. [128], especially deal with the impact of
angular momentum on the tunneling dynamics.
Accurate many-body dynamics are obtained by solving the full many-body
Schrödinger equation utilizing MCTDHB and compared to the corresponding mean-
field dynamics from the GP theory.
For the dynamics, the radial double well in which a BEC with $N=100$ spinless
bosons is confined reads
$V(r)=\begin{cases}B\,e^{-2(r-R_{B})^{4}}+C\,e^{-0.5\,(r-R_{C})^{4}}\,\,&\mbox{if
}r=(x^{2}+y^{2})^{1/2}\leq R_{C}\\\ C\,\,&\mbox{if }r>R_{C}\end{cases}$ (179)
where $B$ is the height of the central ring-shaped barrier located at the
radius $R_{B}$, and $C$ is the height of the crater wall at radius $R_{C}$. A
schematic illustration of $V$ is shown in Fig. 9(a). The region enclosed by
the barrier is denoted by IN, the external rim, i.e., the region between the
barrier and the crater wall, is denoted by OUT. For the numerical
calculations, $C=200.0$ and $B=1.0$ are set which ensures on the one hand that
the energy per particle is sufficiently small such that the BEC stays trapped
in the crater, and on the other hand allows for tunneling between IN and OUT,
as described below. The radius of the crater is set to $R_{C}=9.0$.
The short-range repulsion between the particles is modeled by a Gaussian that
has been employed recently [76, 136],
$\lambda_{0}\hat{W}(\mathbf{r}_{i},\mathbf{r}_{j})=\frac{\lambda_{0}}{2\pi\sigma^{2}}\,e^{-|\mathbf{r}_{i}-\mathbf{r}_{j}|^{2}/2\sigma^{2}},$
(180)
which circumvents the regularization problems of the contact interaction
[137]. The repulsion strength will again be expressed in terms of the mean-
field parameter $\Lambda=\lambda_{0}(N-1)$ in the following. The effect of the
range of the interaction on the degree of fragmentation has been investigated
in detail for repulsive BECs in 1D and 2D single wells [138], essentially
showing that it becomes increasingly dependent on the density the more long-
ranged it is.
The analysis starts with the ground-state energy of two distinct cases, namely
for (i) having all particles in the IN region with closed external rim and
(ii) having all particles in the OUT region with closed trap center. The
notation $E^{\text{IN}}_{\Lambda}$ and $E^{\text{OUT}}_{\Lambda}$, where the
superscript denotes whether the BEC is held in the IN or OUT region and the
subscript denotes the repulsion strength $\Lambda$, is utilized. The ground-
state energies for different radial positions of the barrier $R_{B}$ are
calculated using the GP equation. The corresponding results are shown in Fig.
9(b) for total angular momentum $L=0$ and Fig. 9(c) for $L=N$, i.e., where
each boson on average carries one unit of angular momentum. One observes that
for all shown parameters ($\Lambda=0,\,2.0$ for $L=0$ and $\Lambda=0,\,0.2$
for $L=N$) the curves for $E^{\text{IN}}_{\Lambda}$ and
$E^{\text{OUT}}_{\Lambda}$ intersect at certain radii which are referred to as
the crossing points or simply the crossings in the following. These are
defined as the positions where the ground-state energies of the BEC in the IN
and OUT regions are identical. Putting the barrier onto the crossing point is
the 2D analogue of the symmetric double well in 1D. However, the IN and OUT
regions are only energetically equivalent, whereas their geometries are
totally different. It is found that both the interaction strength $\Lambda$
and the angular momentum $L$ push the crossing point to larger radii once they
are increased. Though, the effect of $L$ is significantly stronger since
already for $\Lambda=0$, the transition from $L=0$ to $L/N=1$ leads to a shift
of the crossing point from $R_{B}=3.271$ to $R_{B}=4.089$. In order to obtain
a similar shift just by increasing $\Lambda$, the repulsion needs to be
largely increased, and thus one would certainly leave the weakly-interacting
regime. The depletion of the ground states with $M=4$ orbitals are $f\approx
10^{-3}$ for $L=0$ and $\Lambda=2.0$ and $f\approx 10^{-9}$ for $L=N$ and
$\Lambda=0.2$, respectively. This means that in both cases, the systems’
ground states are essentially condensed.
(a)
(b)
(c)
(d)
Figure 10: (a) Time evolution of the occupation probability of the external
rim, $P_{\text{OUT}}(t)$, for $N=100$ bosons with $L=0$ and $\Lambda=2.0$. For
$M=4$, initially almost all bosons tunnel. The oscillation amplitude is damped
over time and effectively collapses after 12 Rabi cycles with approximately
half of the bosons in each part (solid red). The system evolves from being
almost fully condensed to two-fold fragmented (see solid green and dark blue
curves for $n_{1}$ and $n_{2}$). The occupations of the third and fourth
orbital are identical and negligible (solid light blue). On the contrary, the
GP theory predicts undamped oscillations (dotted gray). (b) Same as in (a) but
for $L=N$ and $\Lambda=0.2$. Again, the system evolves from being fully
coherent to two-fold fragmented, and the particles finally occupy both the IN
and OUT parts by roughly 50$\%$. Mind the different time scales
($\tau_{1}>\tau_{2}$). Inset: Detailed view between $150\tau_{2}\leq t\leq
155\tau_{2}$. The GP dynamics again do not show the density collapse. (c)
Real, imaginary, and absolute value of the first two natural orbitals
$\alpha_{1}$ and $\alpha_{2}$ at $t=18.4\,\tau_{1}$ where the density is
already collapsed. The GP case is shown in the top panels. For $\alpha_{1}$
and $\alpha_{2}$, the density occupies both the IN and OUT regions. At the
same time, the GP orbital only covers the external part. (d) Same as in (c)
but for the dynamics described in (b). At the many-body level, the nodal
structure of the vortex state is clearly visible in the trap center. All
quantities are dimensionless. See text for details. All figures are adapted
from Ref. [128].
To study the out-of-equilibrium dynamics for different values of $\Lambda$ and
$L$, the focus is laid on the situation where the barrier is located on the
corresponding crossing points. First, the case of zero angular momentum, i.e.,
$L=0$, is analyzed. As an initial state, the ground state of trapping the BEC
in the OUT region with closed central part is chosen. To trigger the dynamics,
the system is quenched by suddenly opening the trap center, such that the
entire potential is given by Eq. (179). It was found numerically that there is
no difference in starting from either the IN or OUT region when the barrier is
located at the crossing point. Fig. 10(a) shows the dynamical evolution of the
occupation probability of the external rim,
$P_{\text{OUT}}(t)=1-P_{\text{IN}}(t)=\frac{1}{N}\int_{R_{B}<r\leq
R_{C}}\rho(\mathbf{r};t)d\mathbf{r},$ (181)
where $\rho(\mathbf{r},t)$ denotes the single-particle density. For the first
two Rabi cycles of $\tau_{1}=110.23$, nearly the whole cloud tunnels between
IN and OUT. Then, the occupation of the second natural orbital starts to grow,
accompanied by a damping of the tunneling oscillations. After approximately 12
cycles, the density oscillations have collapsed and the particles are almost
equally distributed between IN and OUT. The BEC is now highly (two-fold)
fragmented, with occupations of 56.7% and 41.1% of the first two natural
orbitals $\alpha_{1}$ and $\alpha_{2}$, respectively. In contrast to that, the
GP dynamics do not show any damping of the oscillations, not even for the
long-time behavior (not shown). Fig. 10(c) shows the real and imaginary parts
as well as the density of the GP orbital (top panels) and of the first two
natural orbitals in the many-body case (middle and lower panels) at an instant
in time at which the density oscillations have already collapsed. Whereas
there are no bosons in the trap center for the GP case, there is significant
population of the IN region at the many-body level.
Similar observations can be made for $L=N$ and $\Lambda=0.2$, presented in
Fig. 10(b). The many-body dynamics show that the density collapses and that
approximately 50 bosons are occupying the IN and OUT regions finally. The BEC
is two-fold fragmented, the occupation of higher natural orbitals is
negligible. The GP dynamics (see inset) do not account for the damped
oscillations. With regard to the density of the GP and natural orbitals in
Fig. 10(d), it can be seen that $\alpha_{1}$ and $\alpha_{2}$ clearly show the
characteristic nodal vortex structure in the center, whereas at the GP level
there are no bosons in the IN region at the chosen instant of time
$t=183.3\,\tau_{2}$ with $\tau_{2}=66.79$.
It was further found that many-body effects like the density collapse or the
onset of fragmentation are more likely to occur when the particles carry
angular momentum. For $L=N$ and $\Lambda=2.0$, the system quickly becomes
four-fold fragmented, i.e., at least four orbitals are necessary to account
for the many-body dynamics. For $L=0$, the dynamics are essentially governed
by only two orbitals, as can be seen from Fig. 10(c). Thus, the weakly-
interacting regime is apparently smaller for the case of $L>0$, meaning that
vortices are more sensitive to many-body effects. In Section 6.3.3, this will
be discussed again when excitations in rotating BECs are investigated.
To summarize, the many-body nature of the tunneling dynamics of BECs in a 2D
radial double well was demonstrated . The GP mean-field theory does not
account for the collapse of the density oscillations between IN and OUT, and
therefore fails to give an accurate description of the dynamics. Furthermore,
the initially condensed BECs fragment. It was found that bosons carrying
angular momentum are more sensitive to many-body effects than non-rotating
ones because fragmentation sets in already for weaker interaction strengths.
In agreement with these findings, a recent work demonstrates analytically that
a many-body approach is necessary for describing the conservation of angular
momentum in a 2D radially-symmetric system [139].
### 6.2 Dynamical fragmentation
#### 6.2.1 Tunneling to open space
The obtained findings on the many-body tunneling process of initially trapped
bosons to open space in 1D are summarized in this section. The tunneling
process of a single particle was understood already at the beginning of the
20th century [140, 141, 142, 143] and is still of current interest, e.g., with
respect to the exit time and momentum of an electron in the ionization process
[144, 145]. Tunneling of a many-boson system, which has also been studied
[146, 147, 148, 149], is naturally more complicated due to the interactions
and correlations between the constituent particles. The following results,
published in Refs. [150, 151, 152, 153, 154], explain the tunneling mechanism
to open space from a many-body point-of-view utilizing MCTDHB. The main points
that are addressed deal with the questions whether the particles tunnel one by
one or if several particles tunnel simultaneously, and whether the entire
process is indeed of many-body nature.
To study the dynamics of the tunneling to open space, a system of $N$ bosons
confined in a harmonic trap of the form $V(x,t=0)=\frac{1}{2}x^{2}$ is
considered. For $t>0$, the trap is quenched such that it becomes open to one
side. The region where the bosons are trapped initially and the open space are
separated by a potential barrier. The full potential reads
$V(x,t>0)=\theta(x_{c1}-x)\frac{1}{2}x^{2}+\theta(x-x_{c1})\theta(x_{c2}-x)P(x)+\theta(x-x_{c2})T$
(182)
where $\theta(x)$ denotes the Heaviside step function, $P(x)$ denotes a third
order polynomial that ensures a smooth connection between the harmonic trap,
the potential barrier and the threshold potential $T$ whose role will be
discussed below. The coordinates $x_{c1}$ and $x_{c2}$ refer to connection
points, where the former connects the harmonic trap and the barrier and the
latter connects the barrier and the threshold. The bosons repel each other via
the contact interaction potential, and its strength is measured via the mean-
field parameter $\Lambda=\lambda_{0}(N-1)$.
Fig. 11 shows a schematic plot of the trapping potential for $t>0$ (solid
black). The maximum height of the barrier separates the IN and OUT regions,
denoted by the vertical red line. Moreover, the colored horizontal lines mark
the total energies of states with different numbers of bosons $N$ in the trap.
Depending on the threshold value $T$, certain states can become bound, meaning
that no additional particles can escape from the trap via tunneling through
the barrier. How the threshold can be utilized to control the tunneling-to-
open-space dynamics will be discussed below.
Figure 11: Schematic plot of the potential $V$ for $t>0$ (solid black). The
initial Gaussian-shaped density $\rho(x,t=0)$ is shown in solid blue. The IN
and OUT regions are separated at the position of the maximum of the potential
barrier (see solid red lines and labels). Colored horizontal lines denote the
energies of states with different numbers of bosons $N$ in the trap. The
energy differences between those states are denoted by the chemical potentials
$\mu_{i}$. The value $T$ denotes the height of the threshold potential in the
OUT region. All quantities are dimensionless. See text for details. The figure
is taken from Ref. [153].
Important quantities for the analysis are the density $\rho(x,t)$ [see Eq.
(2.4)], as well as the density in momentum space $\rho(k,t)$, or simply the
momentum distribution, which is obtained by performing a Fourier
transformation (FT) of the one-particle reduced density in real space.
Furthermore, the degree of fragmentation $f$, defined in Eq. (29), and
especially its evolution in time, is of major interest. The above quantities
are computed at the many-body level employing MCTDHB.
The analysis starts by looking onto the shape of the momentum distribution in
general. Fig. 12 shows $\rho(k,t)$ for a system of $N=101$ bosons with
interaction parameter $\Lambda=0.3$ at four different instants in time
$t_{1}<t_{2}<t_{3}<t_{4}$. It consists of two parts, a Gaussian background and
a peak structure. The former refers to the particles that did not escape from
the trap yet, since the initial real-space density is also a Gaussian. On the
contrary, the peaks refer to the momenta of the already escaped particles. The
heights of the peaks are time-dependent, which can be seen from the growth of
the dominant peak. The position of the peaks can be understood from a static
model introduced below.
Figure 12: The momentum distribution $\rho(k,t)$ for $N=101$ bosons with
repulsion strength $\Lambda=0.3$ at four instants in time with
$t_{1}<t_{2}<t_{3}<t_{4}$. The Gaussian background refers to the bosons
remaining in the trap, whereas the peaks refer to the momenta of the emitted
bosons. The height of the dominant peak is clearly growing in time. All
quantities are dimensionless. See text for details. The figure is adapted from
Ref. [152].
The minimal total energy $E_{\text{TOT}}$ of the system is comprised of the
energy of the bosons that remain in the IN region, denoted by
$E_{\text{HO}}(\lambda_{0},N_{\text{IN}})$ indicating its dependence on the
two-body repulsion strength $\lambda_{0}$ and the number of particles in the
trap $N_{\text{IN}}$, as well as on the minimal energy of the $N_{\text{OUT}}$
bosons that tunneled through the barrier into open space. Their minimal energy
is the height of the threshold $T$. One thus obtains
$E_{\text{TOT}}=E_{\text{HO}}(\lambda_{0},N_{\text{IN}})+N_{\text{OUT}}\,T.$
(183)
Assuming that the interaction among the emitted bosons is negligible, the
kinetic energy of the $i$-th escaped boson is given by
$E_{\text{kin}}(T,\mu_{i})=\mu_{i}-T$ (184)
where the $\\{\mu_{i}\\}$ are chemical potentials that refer to the energy to
bring an additional boson into the IN region with already $N-i$ bosons. The
corresponding momenta read
$k_{i}^{T}=\sqrt{2(\mu_{i}-T)}$ (185)
where $m=\hbar=1$ is assumed. It is stressed that this model describes the
tunneling of a single boson.
Figure 13: The tunneling to open space of $N=2$ bosons. Left panel: Static
consideration of the energy dependence of the possible final states
$|N_{\text{IN}},N_{\text{OUT}}\rangle$ on the threshold $T$. For $T\lesssim
0.5$, both particles are expected to tunnel, whereas for $0.5\lesssim
T\lesssim 0.8$, only one boson is expected to tunnel. For even larger $T$, the
bosons should remain trapped. Right panel: The obtained peaks in the momentum
distribution $\rho(k,t=600,T)$ for different $T$. For $T\leq 0.4$, two peaks
appear at $k_{1,2}^{T}$, for $T=0.5$ and $0.6$, only one peak appears. This is
in agreement with the expected number of ejected bosons from the trap. For
larger $T$, the peaks are shifted to smaller momenta since the bosons have to
overcome a larger threshold. All quantities are dimensionless. See text for
details. The figures are adapted from Ref. [153].
These model assumptions are tested against a system with $N=2$ bosons and two-
body repulsion strength $\lambda_{0}=1$. The left panel of Fig. 13 shows the
dependence of $E_{\text{TOT}}$ on the threshold value $T$ for the three
possible final states $|N_{\text{IN}},N_{\text{OUT}}\rangle$, given by
$|2,0\rangle$, $|1,1\rangle$ and $|0,2\rangle$. The utilized notation
$|N_{\text{IN}},N_{\text{OUT}}\rangle$ should not be confused with a Fock
state and is just used to count the particles in the two regions IN and OUT.
For $T\lesssim 0.5$, the lowest-in-energy state is $|0,2\rangle$, i.e., the
final state where both bosons have tunneled to open space. For $0.5\lesssim
T\lesssim 0.8$, the state $|1,1\rangle$ is the lowest, meaning that only one
boson is expected to tunnel in the long-time dynamics. For even higher $T$,
tunneling is suppressed completely because the state $|2,0\rangle$ corresponds
to the lowest energy. This is in full agreement with the results shown in the
right panel of Fig. 13. There, the peaks in the momentum distribution
$\rho(k,t=600)$, corresponding to the momenta $k_{1}^{T}$ and $k_{2}^{T}$ of
the emitted particles, are presented for different values of the threshold
$T$. The $k_{2}^{T}$-peaks are plotted upside-down for better visibility. The
first observation is that increasing $T$ shifts the peaks to smaller momenta,
which is due to the higher energy that is necessary to overcome the threshold.
Secondly, two peaks appear up to $T=0.4$, whereas for $T=0.5$ and $0.6$ only
one peak appears. This can be understood in terms of the static considerations
from the left panel, where it was observed that for $T\lesssim 0.5$ both
particles are expected to tunnel through the barrier, whereas only one boson
is expected to tunnel for $0.5\lesssim T\lesssim 0.8$. Thus, the threshold can
be utilized to control the number of tunneling bosons. The arrows in the
bottom of the right panel mark the calculated positions of the peaks with
respect to the model of Eq. (185). The good agreement with the positions of
the peaks of the momentum distribution suggests that the tunneling of single
particles can be well described by the above model.
Figure 14: Evolution of the natural occupations of the first few natural
orbitals. Left panel: The case of zero threshold, $T=0$, utilizing $M=4$
orbitals. The interaction strength is $\lambda_{0}=0.3$. For all particle
numbers shown, the system evolves from being condensed to being multi-fold
fragmented. Right panel: The occupation of the first two natural orbitals for
$N=2$ particles and $\lambda_{0}=1$, but for different heights of the
threshold $T$. In general, the higher the threshold the later the development
of fragmentation sets in. Moreover, the system with larger $T$ fragment less
than the ones with smaller $T$. All quantities are dimensionless. See text for
details. The figures are taken from Refs. [152, 153].
It remains the question whether the bosons do only tunnel one by one, or if it
is also possible that several bosons tunnel at once. To answer this question,
one has to consider that the peaks for $T\lesssim 0.5$ in Fig. 13 (right
panel) appear sequentially, i.e., first the dominant peak develops, and then
the second, smaller peak appears. The fact that the height of the peaks are
time-dependent suggests that the two processes, namely the separate ejection
of the first and second boson, are not independent. If however a multi-boson
ejection happened, peaks at higher momenta than the ones observed would
appear. For instance, in the limit of large $N$, the chemical potential
corresponding to a two-boson ejection would be $\mu^{2b}\approx 2\mu_{1}$,
which, for the case of $T=0$, yields $k^{2b,0}\approx\sqrt{2}k_{1}^{0}$. This
is simply not observed in the Fourier spectrum of $\rho(k,t)$. Thus, it can be
deduced that the bosons tunnel to open space one at a time. Nevertheless, as
mentioned above, the individual processes of the ejection of single bosons are
not independent, meaning that they quantum-mechanically interfere.
Finally, it is discussed whether the overall process of tunneling to open
space is indeed a many-body process or whether it can also be described at the
mean-field level. To this end, the time-evolution of the occupations of the
first few natural orbitals for systems with different numbers of bosons and
different heights of the threshold is studied in Fig 14. In the weakly-
interacting regime considered here, all initial states of the bosonic clouds
are essentially condensed. For the case of zero threshold (left panel), one
observes that for all numbers of bosons considered the system becomes multi-
fold fragmented. Especially for $N=4$ bosons, at least four orbitals become
macroscopically occupied. With respect to the case of $N=2$ particles and non-
zero thresholds, the behavior is similar, meaning that the initially coherent
BEC becomes highly fragmented. In general, the onset of fragmentation happens
at later times when $T$ is larger. Furthermore, the degree of fragmentation is
largest in the absence of any threshold.
To summarize, it has been found that the tunneling-to-open-space dynamics of
repulsive bosons from a harmonic trap are of many-body nature. With time, the
system evolves from being coherent to become highly fragmented, and thus a
mean-field description is inapplicable even in the weakly-interacting regime.
The bosons tunnel one by one, meaning that no multi-boson tunneling occurs
where more than a single boson at the same time is emitted jointly from the
trap. However, the individual tunneling processes are not independent, they
quantum-mechanically interfere. The threshold in the open space can be
utilized to control the number of emitted particles. In particular, it can be
used to suppress tunneling completely. A detailed analysis on the coherence of
the system, showing that the emitted bosons lose coherence with the bosons in
the trap and among themselves, can be found in Refs. [150, 151, 152, 153,
154], as well as additional details.
#### 6.2.2 Phantom vortices
A recently found vortex type in a rotating and repulsive 2D single-component
BEC, termed phantom vortex, is presented in this section. Different to the
common understanding of a quantum vortex that, among other features, has a
density node (or core) and a phase discontinuity surrounding it, a phantom
vortex cannot be detected in the condensate density. Although they were not
found in single-component condensates so far, coreless vortices have been
predicted earlier for multi-component condensates, e.g., spinor condensates
where the core of one species, i.e., bosons with a certain spin, is filled by
another species [155, 156]. Since the experimental realization of gaseous BECs
[4, 6, 5, 157, 158], the nucleation of vortices in rotating BECs was of
interest and it was found that there is a critical rotation frequency below
which no quantized vortices appear [161, 159, 162, 163, 164, 160]. In contrast
to that, phantom vortices appear also below this critical rotation frequency.
Instead of a node in the density, they rather manifest themselves in nodes of
the densities of underlying natural orbitals. Thus, phantom vortices are a
many-body object that cannot be described at the mean-field level. Moreover,
their occurrence is accompanied by the onset of fragmentation. The following
results were published in Ref. [165]. Earlier results on phantom vortices in a
stirred BEC where, e.g., the fragmentation entropy is analyzed, can be found
in Ref. [166].
The system under consideration is a BEC of $N=100$ spinless bosons. They
interact via the same Gaussian repulsion as in Eq. (180) with interaction
parameter $\Lambda=\lambda_{0}(N-1)=17.1$. They are confined in a time-
dependent trap that is varied stepwise. It reads
$V(\mathbf{r},t)=\frac{1}{2}\left(x(t)^{2}+y(t)^{2}\right)+\frac{1}{2}\eta(t)\left(x(t)^{2}-y(t)^{2}\right)$
(186)
where $\eta$ is the parameter of the anisotropy given by the second term in
Eq. (186). The $x-$ and $y-$coordinates are varied in time according to
$\begin{pmatrix}x(t)\\\ y(t)\end{pmatrix}=\begin{pmatrix}\cos(\omega
t)&\sin(\omega t)\\\ -\sin(\omega t)&\cos(\omega t)\end{pmatrix}$ (187)
where the frequency $\omega$ is set to $0.78$. Initially, the trapping
potential is isotropic and harmonic, meaning that $\eta(t=0)=0$, and the
corresponding ground state is computed both at the mean-field level using the
GP equation as well as at the many-body level using imaginary time-propagation
of the MCTDHB equations. Afterwards, a slight anisotropy is added by
increasing $\eta$ from zero to its maximal value of $0.1$ until $t=80$. The
resulting elliptic trap is kept between $80<t\leq 300$, before $\eta$ is
ramped down to zero again between $300<t\leq 380$. For the last interval until
$t=500$, the isotropic harmonic trap is kept constant. The entire procedure is
shown in the upper panel of Fig. 15, together with representative densities of
the BEC for the individual intervals. These densities already suggest that at
no point in time there is a vortex contained in it, as discussed in more
detail below.
Figure 15: Top panel: Time-evolution of the anisotropy parameter $\eta(t)$
(red line) for the 4 distinct intervals discussed in the text, together with
representative densities (blue). The dynamic protocol is divided into a period
of ’ramp up’, ’maximal anisotropy’, ’ramp down’ and ’isotropic’. Second panel:
Time-evolution of the first 4 natural occupation numbers
$\rho_{i}^{(\text{NO})}(t)$. Fragmentation sets in at the end of the ramp up
of the anisotropy. Third panel: Time-evolution of the orbital angular momenta
$(L_{z})_{ii}(t)$ and the total angular momentum per particle, $L_{z}(t)/N$.
Fluctuations are maximal when fragmentation sets in. Bottom panel: Comparison
between the time-evolution of the total energy $E$ relative to the initial
energy $E_{0}$ and $L_{z}(t)/N$. The GP theory underestimates both quantities,
especially the angular momentum is approximately three times less than at the
many-body level. All quantities are dimensionless. See text for details. The
figure is taken from Ref. [165].
Of particular interest for the analysis will be, apart from the density
$\rho(\mathbf{r})$, the natural occupation numbers, denoted here by
$\rho_{i}^{(\text{NO})}$ but with the same meaning as the $n_{i}$ in Eq. (28),
and the densities $|\phi_{i}|^{2}$ of the natural orbitals (mind the different
notation as compared to Section 3 where the $\phi_{i}$ denote the working
orbitals of MCTDHB). Furthermore, the time-evolution of the expectation values
of the orbital angular momenta, given by
$(L_{z})_{ii}=\langle\phi_{i}|\hat{L}_{z}|\phi_{i}\rangle$, and of the total
angular momentum per particle,
$L_{z}/N=\frac{1}{N}\sum_{i,j=1}^{M}\rho_{ij}\,(L_{z})_{ij}$, are analyzed.
The time-argument is suppressed in the above mentioned quantities.
Computations are carried out for orbital numbers $M=4$ (many-body) and $M=1$
(GP). Additionally, signatures of phantom vortices can also be seen in the
phase of the first-order correlation function
$g^{(1)}(\mathbf{r}|\mathbf{r}^{\prime};t)$ [see Eq. (D.6) in Appendix D]
defined by
$S_{g}(\mathbf{r}|\mathbf{r}^{\prime};t)=\text{arg}[g^{(1)}(\mathbf{r}|\mathbf{r}^{\prime};t)]$
(188)
as well as in the phases of the natural orbitals given by
$S_{i}(\mathbf{r},t)=\text{arg}[\phi_{i}(\mathbf{r};t)].$ (189)
The discussion of the results is split into two parts. At first, the
appearance of phantom vortices is analyzed at an instant in time when the trap
is already isotropic again. Afterwards, the degree of fragmentation in the
system is discussed.
Figure 16: Top panels: Total density $\rho(\mathbf{r})$ (a) and densities of
the natural orbitals (b)-(e) at $t=450$, i.e., when the trap is already
isotropic again. Although no vortex is observed in the total density, phantom
vortices in the orbitals $\phi_{1}$, $\phi_{3}$ and $\phi_{4}$ appear. Bottom
panels: Phases of the first-order correlation function (f) and of the natural
orbitals (g)-(j). Phase discontinuities are observed for all phantom vortices.
All quantities are dimensionless. See text for details. The figure is taken
from Ref. [165].
With respect to the first part, the total density and the densities of the
natural orbitals at $t=450$, i.e., when the anisotropy is ramped down to zero
again, are examined in Fig. 16. One observes that the total density [panel
(a)] does not exhibit any vortex. On the contrary, several phantom vortices
appear in the densities of the natural orbitals. The densities of the first
and fourth natural orbitals contain a charge-1 and a charge-3 phantom vortex,
respectively [panels (b) and (e)]. The density of the third orbital contains a
pair of charge-1 phantom vortices [panel (d)]. The mechanism responsible for
the nucleation of the phantom vortices in $\phi_{1}$ and $\phi_{4}$ is
different to the one responsible for the pair in $\phi_{3}$. The former
mechanism is called node mutation, and is explained for the charge-1 phantom
vortex in $\phi_{1}$. This phantom vortex originates from the node of the
initial second orbital $\phi_{2}$, which resembles a 2$p_{x}$ orbital that has
an angular node. When angular momentum is added to the system due to the
anisotropy, the lobes of the $2p_{x}$ orbital spread out and tend to close
this angular node, leaving an elliptic density node in the center. After the
trap is isotropic again, a phantom vortex is left at the center, which now
appears in the density of $\phi_{1}$ since the first and second orbital have
switched labels at $t\approx 220$ because of their changing occupation numbers
[see second panel from the top in Fig. 15]. Thus, the remaining phantom vortex
was nucleated from an initial node. A similar process happens for the charge-3
phantom vortex in the density of $\phi_{4}$. It originates from
$\phi_{3}(t=0)$ which resembles a $2p_{y}$ orbital. There, the angular node
mutated first to three single phantom vortices that finally merged into a
larger charge-3 phantom vortex. The fact that it finally appears in the fourth
natural orbital is due to the label switching of $\phi_{3}$ and $\phi_{4}$
during the ramp-up of the anisotropy.
Figure 17: Time-evolution of the fragmentation for different numbers of
particles $N$. The same dynamical protocol of the trapping potential as in
Fig. 15 is used. For all considered boson numbers, the occupation
$\rho_{1}^{(\text{NO})}$ decreases clearly. In general, fragmentation sets in
later for higher particle numbers. Results are computed using $M=2$ orbitals.
The interaction parameter is kept fixed at $\Lambda=17.1$. All quantities are
dimensionless. See text for details. The figure is taken from the
supplementary material of Ref. [165].
The mechanism responsible for the phantom-vortex pair in $\phi_{3}$ [panel
(d)] is called slow orbital-orbital vorticity transfer because of the
following observation. At $t\approx 220$, the pair of co-rotating phantom
vortices appeared in the density of $\phi_{2}$. When time progressed, the two
phantom vortices started to move towards the density edge and disappeared
finally at $t\approx 275$. At the same time, two phantom vortices were build
at the density edge of $\phi_{3}$, which afterwards continued to move towards
the center. At $t\approx 340$, the separation between this pair is the same as
before in the density of $\phi_{2}$. Thus, vorticity has been transferred
between the two natural orbitals over a period much longer than the rotation
period of $\tau=2\pi/\omega\approx 8$. This process between the second and
third orbital can also be seen from the correlated orbital angular momenta
$(L_{z})_{22}$ and $(L_{z})_{33}$ in Fig. 15 for $t\geq 250$. The position of
the phase discontinuities shown in Figs. 16(g), (i) and (j) coincide with the
corresponding positions of the phantom vortex cores. Especially in panel (j)
one can see from the merging three discontinuity branches in the center that
the corresponding phantom vortex in panel (e) is of charge 3. Comparing the
density in panel (a) and the phase $S_{g}(\mathbf{r}|0;450)$ in panel (f), one
observes the occurrence of several so-called ghost vortices at the density
edge. The latter are distinct from phantom vortices because they do not enter
the bulk density of the condensate and do also not contribute much to the
total angular momentum.
The dynamical fragmentation that occurs during the time-evolution, as well as
its consequences for the angular momenta and the energy of the system [see
again Fig. 15] is now analyzed. The initial state is essentially condensed
($\rho_{1}^{(\text{NO})}\approx 99.7\%$), and it stays condensed for almost
the entire period in which the anisotropy is ramped up to its maximal value.
In this time frame, the total angular momentum and energy computed at the
many-body and mean-field levels coincide as long as the BEC remains coherent.
Once fragmentation sets in, the latter quantities start to deviate from each
other. During the first part of the time interval in which the trap is kept
maximally anisotropic ($80<t\leq 150$), $\rho_{1}^{(\text{NO})}$ decreases
significantly to roughly $40\%$, and the second, third and fourth natural
orbitals become macroscopically occupied. At the same time, the orbital
angular momenta vary the most and several of them reach their maximum values.
Then, the occupations and angular momenta of $\phi_{1}$ and $\phi_{4}$ remain
more or less constant. For the orbitals $\phi_{2}$ and $\phi_{3}$, both
quantities change slightly in a correlated manner, i.e., when
$\rho_{2}^{(\text{NO})}$ increases $\rho_{3}^{(\text{NO})}$ decreases by the
same amount. This again reflects the slow orbital-orbital vorticity transfer
between $\phi_{2}$ and $\phi_{3}$ discussed above. With regard to the total
angular momentum and energy, both quantities vary strongly during the periods
of maximal anisotropy and ramp down, and saturate only afterwards. The
corresponding GP values are substantially different since they underestimate
the system’s gain of energy and angular momentum. In particular for the latter
quantity, it stays below $0.5$, well below the threshold for the occurrence of
vortices at the mean-field level at unit angular momentum per particle.
Interestingly, $L_{z}/N$ for $M=4$ orbitals is greater than this threshold,
but still no vortices are observed in the density $\rho(\mathbf{r})$. This can
be rationalized because one can show analytically that a vortex in the density
of a fragmented BEC, i.e., a fragmented vortex, requires more than unit
angular momentum per particle [see the supplementary information of Ref.
[165]]. The reason is that a fragmented vortex originates from coincident
phantom vortices, meaning that all natural orbitals must have a phantom vortex
at the same position, and all of these phantom vortices need to have different
charges such that the natural orbitals are orthogonal. As seen from panels
(b)-(d) in Fig. 16, there are no coincident phantom vortices, and thus the
fragmented system does not show any vortices in the density.
The onset of fragmentation was also observed for different numbers of bosons.
Fig. 17 shows the evolution of $\rho_{1}^{(\text{NO})}$ for different $N$
between $10$ and $10^{4}$, utilizing $M=2$ time-adaptive orbitals. The
interaction parameter $\Lambda=17.1$ is kept fixed. One generally observes
that fragmentation sets in for all considered particle numbers, and that
furthermore the onset of fragmentation occurs at later times for higher
particle numbers.
In conclusion, the nucleation of phantom vortices has been investigated for a
rotating BEC that acquires angular momentum via a time-dependent modification
of the trap. Two different mechanisms for the creation of phantom vortices
were found, namely node mutation as well as slow orbital-orbital vorticity
transfer. Phantom vortices, by definition many-body objects, cannot be
described at the mean-field level. The initially coherent system highly
fragments, also for different numbers of particles in the condensate. Further
information can be found in Ref. [165], in particular in the supplementary
material containing videos that show, e.g., the evolution of all quantities
shown in Fig. 16.
### 6.3 Excitation spectra
In this section, details on current applications of LR-MCTDHB are presented.
The amount of scientific publications where it is utilized is still rather
small because the theory itself has been developed only a few years ago [34,
35]. In addition to that, the efficient MPI-parallelized numerical
implementation of LR-MCTDHB was a very demanding task, and all the technical
challenges going along with it, extensively described in Section 4, were
solved very recently. It is stressed again that without such an
implementation, a vast majority of problems cannot be treated at all,
especially in $D>1$ spatial dimensions. Before developing the sophisticated
and complex parallel implementation, there was a first implementation
available that could treat small 1D systems. The results of Refs. [34] and
[167] were obtained with this code [see below and Section 6.3.1]. In more
recent applications, discussed in Secs. 6.3.2 and 6.3.3, the new
implementation was utilized since it was impossible to obtain the same results
with the former code.
(a)
(b)
Figure 18: (a) Numerical convergence of the first three center-of-mass excited
states (labeled by 1, $2^{\prime}$ and $3^{\prime}$) with respect to the
number of orbitals $M$ for $N=10$ bosons in a harmonic trap with trap
frequency $\omega_{\text{ho}}=\sqrt{2}$. The strength of the contact repulsion
is $\lambda_{0}N=1$. For $M=5$ orbitals, the energies obtained from LR-MCTDHB
coincide with the analytic predictions. The numerical implementation used is
the precursor of the implementation described and benchmarked in Sections 4
and 5. (b) Comparison between the excitation spectra obtained from BdG
(bottom) and LR-MCTDHB($2$) (top) for $N=10$ bosons in a shallow double well.
The interaction strength is the same as in panel (a). Excitations labeled with
primes refer to states solely obtained at the many-body level. The intensity
of the response to a linear perturbation ($f^{+}=f^{-}=x$) is denoted by red
triangles, whereas the intensity of the response to a quadratic perturbation
($f^{+}=f^{-}=x^{2}$) is denoted by green squares. The insets show the trap
potential and the ground-state densities. One observes that the many-body low-
energy spectrum contains a lot more states than the mean-field spectrum. All
quantities are dimensionless. See the main text and Ref. [34], from which both
figures are taken, for further details, especially for the discussion of the
excitation type shown at the top of panel (b).
The main purpose of Ref. [34] is to introduce LR-MCTDHB by giving a derivation
of its main equations, presented in Section 3.2.2 of this work. In particular,
the equations for the submatrices as described in Eq. (93) are developed for
the case of contact interaction between the bosons. As an application,
excitations for BECs in harmonic and shallow double-well traps are analyzed.
Fig. 18(a) shows the numerical convergence for the first three c.m. excited
states (labeled by 1, $2^{\prime}$ and $3^{\prime}$) of $N=10$ bosons in a
harmonic trap with trap frequency $\omega_{\text{ho}}=\sqrt{2}$ in terms of
the number of orbitals $M$. The strength of the repulsion is given by
$\lambda_{0}N=1$. One can observe that the c.m. excitations are accurately
obtained for $M\geq 5$ orbitals. The BdG theory only yields the first c.m.
excitation. In Fig. 18(b), the low-energy spectrum as well as the response
intensities for linear and quadratic perturbations ($f^{+}=f^{-}=x,\,x^{2}$)
are presented for $N=10$ bosons in a shallow double well of the form
$V(x)=b/2\cos\left(\pi x/3\right)+\omega_{\text{ho}}^{2}x^{2}/2$ with barrier
height $b=5$. The most important observation by comparing the number of
obtained states for BdG and LR-MCTDHB($2$) is that the many-body spectrum
contains a lot more states. This in turn means that although the degree of
depletion is only about $0.2\%$, a many-body description for the lowest-in-
energy excitation spectrum is unavoidable in order to obtain accurate results.
In the subsequent Section 6.3.1, the excitation spectrum for the case of BECs
trapped in harmonic and double-well potentials is elaborated in more detail,
especially with respect to the question of how many-body excitations can be
triggered in the out-of-equilibrium dynamics of the condensates.
Before discussing further applications of LR-MCTDHB in the following, a remark
on how mean-field and many-body excitations are defined in this work is made
in order to avoid misconceptions of these important terms. Excitations
calculated from Eq. (116), i.e., from LR-MCTDHB ($M>1$), are termed many-body
excitations since they stem from a many-body theory. On the contrary, all
excitations calculated with the BdG equations in Eq. (A.16) of Appendix A.1.2
are referred to as mean-field excitations since they stem from a mean-field
theory. Moreover, excitations where only a single boson is excited from the
condensed mode are called single-particle excitation, whereas excitations
where more than a single boson is excited are called multi-particle
excitations. These definitions are utilized consistently in the applications
below.
#### 6.3.1 One-dimensional harmonic and double-well systems
As a first application of LR-MCTDHB, the many-body nature of excitations in a
1D harmonic trap as well as in shallow and deep symmetric double wells is
analyzed. The low-energy spectra obtained from the BdG equation at the mean-
field level and from LR-MCTDHB at the many-body level are compared.
Furthermore, it is investigated which excitations are involved in the dynamics
due to certain quench scenarios. The following results where published in Ref.
[167].
Excited states are studied in a harmonic trapping potential with a Gaussian
barrier given by
$V(x)=ax^{2}+b\,\exp\left(-cx^{2}\right)$ (190)
where $a$ denotes the harmonic trap frequency and $B$ denotes the barrier
height. For simplicity, $\hbar=m=1$ is assumed. Furthermore, the trapping
parameters are set to $a=1/2$ and $c=1$. The bosons interact via the zero-
ranged contact interaction potential. The interaction strength is again
measured by the mean-field parameter $\Lambda=\lambda_{0}(N-1)$ where
$\lambda_{0}$ is chosen to be positive to account for repulsion. Fig. 19 shows
$V(x)$ for different barrier heights, together with the corresponding ground-
state densities. For the purely harmonic trap ($b=0$), the BEC is essentially
condensed with $n_{1}=99.9\%$. On the contrary, for the shallow ($b=5$) and
deep ($b=10$) double wells, the ground state is already fragmented by
approximately $5\%$ and $40\%$, respectively. One can further observe that for
the deep well, the overlap between the densities in the left and right wells
is marginal, whereas there is still a clear overlap observable in the shallow
case.
Since the trap is reflection invariant, the excitations can be categorized in
gerade ($g$) and ungerade ($u$) symmetry. Fig. 20 shows the low-energy
spectrum of $N=10$ bosons for different barrier heights $0\leq b\leq 10$ and
$\Lambda=1.0$. Excitation energies are given relative to the ground-state
energy, i.e., $\Delta E=E-E_{0}$. One readily observes the formation of bands,
which is even more pronounced for lower interaction strengths [see Fig. 2 in
Ref. [167]]. The bands are labeled by integer numbers, where the zeroth band
denotes the one that is lowest in energy. Comparing the mean-field and many-
body approaches, one can see that the BdG spectrum only contains 5 different
excitations, whereas the spectrum computed with LR-MCTDHB shows a much richer
structure where a lot more excited states appear. In terms of numerical
convergence, $M=2$ orbitals are sufficient to describe the spectrum up to
$\Delta E\sim 2$ for not too low barrier heights, i.e., $b\geq 4$. For states
higher in energy, at least four orbitals are necessary.
Figure 19: Trap potentials (solid red) and ground-state densities (gray) for
different barrier heights $b$. The ground states are computed with
MCTDHB($2$). The BEC consists of $N=10$ bosons and the repulsion strength is
$\Lambda=1.0$. The occupation number $n_{1}$ of the first natural orbital
shows that increasing the barrier height leads to fragmentation. Vertical
lines refer to the lowest-in-energy excited states computed with LR-
MCTDHB($2$). The green potentials and the blue eigenvalues refer to the
harmonic approximation of the right well. All quantities are dimensionless.
See text for details. The figure is taken from Ref. [167]. Figure 20: Low-
energy excitation spectrum of $N=10$ bosons with interaction parameter
$\Lambda=1.0$ for different barrier heights $b$. Many-body results computed
with LR-MCTDHB($2$) are denoted by solid lines (gerade excitations in green,
ungerade excitations in red). Results for $M=4$ orbitals are given by dotted
lines, and the corresponding BdG energies are given by open black circles. All
states up to $\Delta E\sim 2$ are converged with two orbitals. The BdG
approach misses many states in the shown energy range. All quantities are
dimensionless. See text for details. The figure is taken from Ref. [167].
In the following, different shift and quench protocols are applied to trigger
excitations from the low-energy spectrum. For gerade excitations, a quench of
the harmonic confinement from $a=0.4\rightarrow 0.5$ is performed, whereas
ungerade excitations are addressed by a shift of the trap position, meaning
that $x\rightarrow x+x_{\text{shift}}$ with $x_{\text{shift}}=0.1$. The gerade
excitations are referred to as breathing modes, and the ungerade excitations
are referred to as dipole modes. The dynamics of the bosonic clouds for the
two cases are shown in Fig. 21 on a short time scale. In the left panel,
dipole oscillations for a BEC in the deep double well are depicted, which
means that in each well, the bosons oscillate around the equilibrium position
given by the minima of the wells. On the RHS, the breathing dynamics for
$N=10$ bosons in the purely harmonic trap is shown. One can observe that the
density spreads to the left and right in a symmetric manner, and returns to
the minimum of the well. The Fourier spectra of the time-evolution of certain
quantities is used to extract the spectrum of involved excitations in the
dynamics. This comprises $(a)$ the expectation value $\langle x\rangle$, $(b)$
the variance per particle $\frac{1}{N}\Delta^{2}_{\hat{X}}$, and $(c)$ the
density $\rho$ at the position $x=1$. The many-body variance is discussed in
detail in Appendix D. In Ref. [167], the Fourier spectra for $b=0\,,5$ and
$10$ are investigated. Here, the discussion is restricted to the latter. To
remind the reader, the system is highly fragmented by approximately $40\%$ in
this case.
Figure 21: Many-body dynamics of the density of $N=10$ bosons with interaction
parameter $\Lambda=1.0$ for a shift of the trap position in the deep double
well (left panel) and a quench of the harmonic confinement (right panel). In
the former scenario, one observes dipole oscillations of the bosonic clouds in
the two wells, whereas one observes breathing oscillations for the dynamics
due to the quench of the confinement. All quantities are dimensionless. See
text for details. The figure is taken from Ref. [167].
Table 4 shows excitation and de-excitation energies where the obtained values
from the BdG and LR-MCTDHB($2$) theories (columns 2 and 3) are compared to the
obtained excitation energies of the FT spectra from the shift and quench
scenarios (columns 4-7). Moreover, several de-excitations, i.e., transitions
between a higher and a lower excited state, are presented (last 4 rows). Those
refer to higher-order effects and are in principle not obtained directly by
LR. However, since LR yields the exact excitation spectrum once it is applied
to an exact eigenstate of the full many-body Hamiltonian, the de-excitations
can be extracted from differences between the excitation energies of the LR
spectrum. The excitation labels are shown in the first column. Many-body
excitations, i.e., excitation that only appear in the many-body spectrum, are
denoted by primed labels.
| $M=1$ | $M=2$ | Shift GP | Shift MB | Quench GP | Quench MB
---|---|---|---|---|---|---
$0u$ | 0.035 (0.46) | 0.063 (0.19) | 0.034 (0.02a) | | |
$0g^{\prime}$ | | 0.064 | | | | 0.066 (0.36b)
$1g$ | 2.085 (0.91) | 2.092 (0.81) | | | 2.085 (1.0b,0.93c) | 2.092 (0.33b,0.89c)
$1u$ | 2.102 (0.46) | 2.095 (0.57) | 2.102 (0.94a,0.66c) | 2.095 (0.93a,0.7c) | |
$1u^{\prime}$ | | 2.158 (0.08) | | 2.163 (0.02a,0.02c) | |
$1g^{\prime}$ | | 2.161 (0.09) | | | | 2.161 (0.12b)
$2g$ | 3.912 (0.04) | 3.934 (0.03) | 3.914 (0.04c) | | 3.914 (0.07c) | 3.934 (0.07b,0.1c)
$2u$ | 3.997 (0.07) | 3.943 (0.07) | 3.997 (0.04a,0.23c) | 3.943 (0.03a,0.15c) | |
$2u^{\prime}$ | | 4.029 (0.05) | | 4.032 (0.02a,0.07c) | |
$2g^{\prime}$ | | 4.039 | | | |
$1g^{\prime}\leftrightarrow 1g$ | | 0.069 | | | |
$2g\leftrightarrow 1g$ | 1.827 | 1.841 | | | | 1.893 (0.01b)
$2u\leftrightarrow 1u$ | 1.894 | 1.848 | 1.895 (0.01c) | 1.849 (0.02c) | |
$1g\leftrightarrow 0g^{\prime}$ | | 2.028 | | | | 2.032 (0.02b)
Table 4: Table of excitation and de-excitation energies for $N=10$ bosons with
$\Lambda=1.0$ in the deep double well. Listed are results from BdG and LR-
MCTDHB($2$), as well as from the Fourier spectra of the expectation value of
the position $\langle x\rangle$ (superscript $a$), the variance per particle
$\frac{1}{N}\Delta^{2}_{\hat{X}}$ (superscript $b$) and the real-space density
$\rho(x=1)$ (superscript $c$), for both the shift and quench scenarios. GP
denotes the mean-field and MB the many-body case. The intensities are given in
brackets. All quantities are dimensionless. See text for more details. The
table is adapted from Ref. [167].
Up to $\Delta E\sim 4$, the many-body approach yields ten excited states,
whereas on the contrary the BdG spectrum only consists of five excitations.
The energies of states that appear in both spectra differ slightly from each
other for $1g$, $1u$, $2g$ and $2u$, but significantly for the lowest
excitation $0u$. That demonstrates that strong deviations between the mean-
field and many-body excitation energies can already occur for the lowest
excited states. Also the intensities of several states differ clearly from
each other, especially for the lowest states $0u$, $1g$ and $1u$. The
intensities from the LR calculations are identical to the response weights
described in Eq. (3.2.2), with $f^{+}=f^{-}=x$ for the ungerade and
$f^{+}=f^{-}=x^{2}$ for the gerade states.
Figure 22: Fourier spectra of $\langle x\rangle$,
$\frac{1}{N}\Delta^{2}_{\hat{X}}$, and $\rho(x=1)$ from the shift and quench
dynamics in the deep double well. The GP spectra are denoted by solid black
and the MCTDHB($2$) spectra by dashed red lines. The intensities are
normalized by the sum of the intensities of all constituent peaks, and
excitations with less than $1\%$ intensity are omitted. Especially the
spectrum of the many-body variance consists of several peaks that are not
observed in the corresponding GP spectrum. All quantities are dimensionless.
See text for details. The figure is adapted from Ref. [167].
Columns 4 and 5 show the obtained peaks in the Fourier analysis of $\langle
x\rangle$ and $\rho(x=1)$ for the shift scenario, both for the GP and many-
body cases. The Fourier spectra are shown in the left panels of Fig. 22.
Surprisingly, the shift also triggered a gerade excitation ($2g$) at the mean-
field level, whereas at the many-body level only ungerade states are observed.
Moreover, no peak is observed for $0u$ in the many-body dynamics, presumably
because its intensity is too weak to be distinguished from the background
noise. The obtained peak positions coincide with the predictions from BdG and
LR-MCTDHB($2$), respectively. With regard to de-excitations, the mean-field
and many-body approaches yield a very small peak for the $2u\leftrightarrow
1u$ transition in the density spectrum.
The Fourier analysis of the quench dynamics of
$\frac{1}{N}\Delta^{2}_{\hat{X}}$ and $\rho(x=1)$ are shown in column 6 (GP)
and column 7 [MCTDHB($2$)]. Only peaks corresponding to gerade excitations are
observed. Both approaches yield peaks for $1g$ and $2g$. At the many-body
level, two additional peaks referring to $0g^{\prime}$ and $1g^{\prime}$ are
obtained. The peak positions again accurately match the predictions from the
static analysis. Most importantly, the intensities of these excitations in the
variance spectrum are comparatively large, and therefore $0g^{\prime}$ and
$1g^{\prime}$ are the most promising many-body excited states to be detected
in an experiment. Furthermore, no de-excitation peaks are found in the GP
spectra, while the variance spectrum shows small peaks for the
$2g\leftrightarrow 1g$ and $1g\leftrightarrow 0g^{\prime}$ transitions at the
many-body level, coinciding with the values from LR-MCTDHB($2$).
In general, the sensitivity of the utilized operators to the detection of
excited states can be seen as follows. The spectrum of the local operator
$\rho(x=1)$ is in particular useful to detect ungerade excitations in the
shift scenario. The intensities of the peaks of many-body excitations are
however weak. To detect gerade excitation, nonlocal quantities, especially the
variance, turn out to be more sensitive than the density. In addition, the
variance spectrum yields intense peaks also for many-body excitations.
To summarize, it has been demonstrated that from a methodological point of
view, utilizing LR-MCTDHB in order to obtain the many-body excitation spectra
of repulsive BECs in 1D harmonic and double-well trap geometries is a
promising and highly accurate technique. Its predictions coincide with the
Fourier spectra of the time-evolution of different quantities like the local
density or the position variance. From a physical point of view, some of the
obtained excited states corresponding to many-body states not included in
standard mean-field approaches yield intense responses especially in the
variance spectra and are therefore likely to be measured in experiments. More
details on the presented results can be obtained from Ref. [167].
#### 6.3.2 Triple wells and larger lattices
In this section, many-body effects in the excitation spectra of weakly-
interacting BECs in 1D lattices are investigated. The main objective is to
demonstrate with numerically converged results that a marginal depletion of
the order of $1\%$ is sufficient in order to observe clear differences between
the lowest-in-energy excitations computed at the mean-field and many-body
levels. The results presented below were recently published in Ref. [168].
Studying dynamics and excitations of ultracold bosons in 1D optical lattices
was of high interest in the past, and the literature is extensive in this
research field. Examples are the dynamics in the superfluid phase [169],
nonlinear dynamics in periodic potentials [170] or in fluctuation-driven
binary condensates [171]. A review of earlier works can be found in Ref.
[172]. Of particular relevance for the findings discussed in this section are
studies were the BdG theory does not agree with numerical and experimental
observations. It has been found that the oscillation frequency of a superfluid
BEC in a large optical lattice with a superimposed harmonic confinement
deviates clearly from the BdG prediction [173, 174]. In addition, the strong
depletion of a gaseous BEC in optical lattices cannot be fully explained by
the BdG theory, not only in 1D [175].
Optical lattices, in particular triple-well systems, have also been analyzed
recently utilizing MCTDHB. This includes the out-of-equilibrium dynamics of
rather small BECs consisting of a few particles only, induced by quenching the
interaction and the lattice depth [176, 177, 178] and periodic driving of the
system [179, 180]. Of particular interest was the analysis of cradle modes,
dipole oscillations corresponding to the transport of bosons over the barrier
between adjacent wells. Furthermore, the many-body dynamics of BECs initially
prepared in Mott-insulating states have been of interest [181]. The different
phases of the BEC, i.e., the superfluid, the Mott-insulating, and the
fermionized gas phase, can be detected from the many-body entropy and the
Glauber correlation functions [182]. It was even shown for a BEC in a tilted
triple well with hard-wall boundary conditions how the correlations between
wells can be controlled by the well depth, the interaction strength and the
tilt [183]. Many-body effects in the ground state of a bosonic mixture in a
more complex 2D lattice where studied using the multi-layer version of MCTDHB
[184].
Many-body excitations in optical lattices were addressed as well, either by
making a Fourier analysis of certain quantities [176] or by doing an exact
diagonalization. There, the system’s wave function was expressed using a many-
body ansatz based on a number state expansion including localized Wannier
functions [179, 180, 177, 178]. The latter approach of exact diagonalization
is possible for small systems with a few bosons only.
(a)
(b)
(c)
(d)
Figure 23: Low-energy spectra of $N=10$ bosons in a triple well computed with
LR-MCTDHB($7$). (a) Shallow case with lattice depth $V_{0}=1.01\,E_{R}$. For
$\Lambda=0$, the excitations form levels of degenerate states. Degeneracies
are lifted for $\Lambda>0$. The BdG results only contain the two doublets
denoted by BdG(1) and BdG(2). (b) Same as in (a) but in the deep lattice with
$V_{0}=10.13\,E_{R}$. The BEC for $\Lambda=2.0$ is fragmented by $8.4\%$. (c)
Comparison between BdG(1) and its multiples (open circles) and the many-body
results (colored closed symbols) for the system in (a) with $\Lambda=4.0$
where only the states $(+1)$ and $(-1)$ of the first single-particle band are
occupied by $N^{(1)}$ bosons. The difference grows with $N^{(1)}$. Inset:
Evolution of the relative error as defined in the main text. (d) Same as in
(c), but for the system in (b) with $\Lambda=2.0$. All quantities are
dimensionless. See text for details. The figures are adapted from Ref. [168].
Below, the lowest-in-energy excitations of BECs in 1D lattice potentials with
periodic boundary conditions are investigated using LR-MCTDHB. This opens the
possibility to study also larger systems with more wells and particles. The
utilized trap potential reads
$V(x)=V_{0}\,\cos^{2}\left(\frac{\pi}{l}\,x\right)$ (191)
where $V_{0}$ denotes the lattice depth and $l$ denotes the distance between
two neighboring lattice sites. The depth of the lattice is commonly expressed
in terms of the recoil energy $E_{R}=\frac{\hbar^{2}k_{0}^{2}}{2m}$ with the
lattice momentum $k_{0}=\frac{\pi}{l}$. For simplicity, $\hbar=m=l=1$ is
assumed. The bosons repel each other via the contact interaction potential and
its strength is again expressed in terms of the mean-field parameter
$\Lambda=\lambda_{0}(N-1)$.
(a)
(b)
Figure 24: (a) Same as in Fig. 23(a) but in a lattice with 10 sites. The
depletion grows faster with $\Lambda$ than in the triple well, meaning that
the system is even more sensitive for many-body effects. (b) Same as in Fig.
23(c) but for 10 sites and $\Lambda=1.0$. The relative error (inset) reaches
similar values as for systems in the shallow triple well with comparable
depletion. All quantities are dimensionless. See text for details. The figures
are adapted from [168].
As a start, $N=10$ bosons confined in a triple-well potential with periodic
boundary conditions are considered. The low-energy excitation spectra for both
a shallow ($V_{0}=5.0\approx 1.01\,E_{R}$) and a deep well ($V_{0}=50.0\approx
10.13\,E_{R}$) are discussed. Results are presented in Figs. 23(a) and (b). In
both wells, the spectra for the non-interacting BEC are structured in
sequences of levels with different degeneracies. The states and levels can be
associated with appropriate quantum numbers, envisioning that for $\Lambda=0$
eigenstates of a single particle in a periodic lattice are exact quasimomentum
eigenstates. The eigenstates appear in a band structure that is more
pronounced for deep lattices. For the shallow lattice in panel (a), the first
level consisting of two degenerate states denotes the excitations of putting
one boson in either a state with quasimomentum $(+1)$ or with quasimomentum
$(-1)$. Similarly, the states of level 5 denote the excitations where one
boson is put to either $(+1)$ or $(-1)$ from the second single-particle band.
To this end, the notation
$\left(n_{+1}^{(1)},n_{-1}^{(1)};n_{+1}^{(2)},n_{-1}^{(2)}\right)$ is
introduced, where the first two entries in brackets denote the occupation of
$(+1)$ and $(-1)$ from the first single-particle band and the last two entries
denote the occupation of $(+1)$ and $(-1)$ from the second single-particle
band. As another example, level 2 contains the states $(2,0;0,0)$, $(0,2;0,0)$
and $(1,1;0,0)$. It is stressed that the BdG spectrum only contains the
excitations of levels 1 and 5, and misses all multi-particle excitations in
the low-energy part. However, for $\Lambda=0$, the missing levels can be
anticipated by taking sums and multiples of the energies of the first and
fifth levels. As soon as there are correlations between the bosons, i.e., for
$\Lambda>0$, this is no longer possible, as discussed below. In panel (b),
only the first single-particle band appears in the shown energy range because,
as shown in Fig. 25, the first and second band are strongly separated from
each other.
Figure 25: Proof of numerical convergence for the systems discussed in Figs.
23 and 24. LR-MCTDHB($7$) clearly suffices in all cases. For the BEC in the
deep triple well [middle panel], all 65 single- and multi-particle excitations
below the bottom of the second band [denoted by BdG(2)] are obtained. Already
for $M=3$, the results are highly accurate which is due to the large
separation of the first two single-particle bands. All quantities are
dimensionless. See text for details. The figures are adapted from Ref. [168].
For non-zero repulsion in the shallow well in panel (a), the BECs are slightly
depleted ($f=1.1\%$ for $\Lambda=4.0$), whereas for the deep well in panel (b)
the BEC for $\Lambda=2.0$ is already fragmented by $8.4\%$. One can observe
that all excited states grow in energy once the bosons interact, and that
splittings of several states from the same level occur although they were
degenerate for $\Lambda=0$. Thus, the clear level structure of the non-
interacting case vanishes as $\Lambda$ increases, which makes it more
difficult to identify the individual states with regard to the categorization
in terms of quasimomentum states. The corresponding mean-field results contain
only two doublets for the shallow and one doublet for the deep well, denoted
by BdG(1) and BdG(2). All other states, representing multi-particle
excitations where more than one boson at a time is excited from the ground
state, are absent.
Figs. 23(c) and (d) demonstrate in how far the BdG energies can be utilized to
anticipate the positions of the missing multi-particle states. As mentioned
above, this yields the exact energies for the missing states in the non-
interacting system. In particular, multi-particle excitations of putting in
total $N^{(1)}$ bosons into the states $(+1)$ and $(-1)$ of the first single-
particle band are discussed. With regard to the BEC with $\Lambda=4.0$ in the
shallow triple well [panel (c)], one observes that BdG overestimates the
excitation energies for the states of $1\leq N^{(1)}\leq 5$, i.e., the
accurate many-body values are lower. The discrepancy between mean-field and
many-body results grows with $N^{(1)}$. Especially the relative error, defined
as
$E_{\text{rel}}=\left|\frac{\omega_{\text{BdG}}\left(N^{(1)}\right)-\omega_{\text{MB}}\left(N^{(1)}\right)}{\omega_{\text{MB}}\left(N^{(1)}\right)}\right|$
where $\omega_{\text{MB}}\left(N^{(1)}\right)$ denotes the many-body energy
with the largest distance to
$\omega_{\text{BdG}}\left(N^{(1)}\right)=N^{(1)}\cdot\text{BdG(1)}$, grows
quickly to more than $10\%$. It is emphasized again that the BEC is only
slightly depleted by approximately $1\%$. For the fragmented BEC with
$\Lambda=2.0$ in the deep triple well [panel (d)], the differences become even
more substantial because already for the single-particle excitation
($N^{(1)}=1$) the relative error is about $5\%$. This means that in this case
BdG does not only miss all multi-particle excited states in the low-energy
spectrum, but it is also very inaccurate for the single-particle excitations.
Fig. 24 shows a similar analysis for $N=10$ bosons in a larger shallow lattice
with 10 sites. In principle, the described many-body effects observed in the
triple well also appear in this extended lattice. However, with respect to the
degree of fragmentation, which is $f=0.84\%$ for $\Lambda=1.0$, one can deduce
that these many-body effects set in even earlier than in the triple well
because a comparable depletion is achieved already for weaker repulsion.
Hence, an accurate many-body description is of even larger importance if the
lattice size is increased.
Figure 26: (a) Energies of the zero-quasimomentum modes (ZQMs) for the system
in Fig. 23(a) with $\Lambda=4.0$, up to the top of the second single-particle
band [denoted by BdG(3)]. All shown states are not obtained by the BdG mean-
field approach. The ZQMs are categorized by the symmetry of their density
responses, which is either gerade or ungerade on each lattice site. Mind the
different index $n_{z}$, which enumerates only the ZQMs compared to the index
$n$ of the upper figures which enumerates all excitations. (b) and (c): Real
part of the density responses of the states $(3,0;0,0)$ and $(0,3;0,0)$ [panel
(c)] as well as $(4,1;0,0)$ and $(1,4;0,0)$ [panel (d)]. The vertical lines
separate the lattice sites from each other. Notice the different scales. All
quantities are dimensionless. See text for more details. The figures are taken
from Ref. [168].
The results shown in Fig. 25 demonstrate numerical convergence with respect to
the number of self-consistent orbitals $M$ utilized. In all cases, $M=7$
orbitals clearly suffice for the low-energy excitations. In particular, for
the middle panel describing a BEC in the deep triple well, one can see that
firstly, $M=3$ orbitals already give very accurate values for the excitation
energies, and, secondly, all multi-particle states build up by the states
$(+1)$ and $(-1)$ from the first single-particle band are obtained (65 in
total). To remind the reader, it is stressed again that the BEC is fragmented
in this case, and nevertheless LR-MCTDHB is capable to give converged results.
So far, it was shown that the BdG energies for a BEC in a 1D lattice can be
largely inaccurate, even for weakly-depleted condensates with $f\approx 1\%$.
Though it has been shown analytically that for a BEC with sufficiently weak
and long-ranged repulsion in the Hartree limit, i.e., where
$N\rightarrow\infty$ with fixed $\Lambda$, the BdG equation yields the exact
energies of all single-particle excitations, and the energies of the multi-
particle states are given by multiples and sums of the BdG energies [185]
(extending the findings of a previous work on homogeneous systems [186]).
Whether this result is applicable also for BECs with contact interaction
remains unclear. However, naturally the question arises how fast, if at all,
the excitation spectrum of a BEC with arbitrary shape of the repulsion
approaches the BdG energies upon increasing the number of bosons. This
question is discussed further in the subsequent section, Section 6.3.3, where
excited states in rotating BECs are analyzed.
Figure 27: Low-energy Fourier spectrum of the time evolution of the position
variance per particle in $x$-direction, $\frac{1}{N}\Delta^{2}_{\hat{X}}$, due
to a quantum quench of the lattice depth from $V_{0}=4.9$ to $V_{0}=5.0$. The
system parameters are the same as in Fig. 23(a) with $\Lambda=4.0$. The many-
body spectrum (dashed red), obtained from the evolution with $M=5$ orbitals,
shows two sharp peaks referring to two-particle excitations and one less
intense peak referring to a three-particle excitation. The first intense peak
refers to the two-particle excitation $(1,1;0,0)$, whereas the second intense
peak refers to $(1,0;0,1)$ [or $(0,1;1,0)$]. The less intense peak refers to
$(3,0;0,0)$ [or $(0,3;0,0)$]. All obtained peaks are ZQMs with gerade
symmetry. The GP spectrum (solid black) in the shown energy range is
completely flat. The depletion of the BEC is $f\sim 1.1\%$ during the entire
propagation time. All quantities are dimensionless. See text for more details.
Finally, the zero-quasimomentum modes (ZQMs) are briefly discussed. The latter
are defined in this work as the excitations where $\text{MOD}(P,L)=0$, with
$P$ denoting the total quasimomentum up to the second single-particle band
given by
$P=n_{+1}^{(1)}+n_{+1}^{(2)}-n_{-1}^{(1)}-n_{-1}^{(2)},$ (192)
and $L$ denoting the number of lattice sites. The spectrum of the ZQMs up to
BdG(3), which marks the top of the second single-particle band, is shown in
Fig. 26(a). The system parameters are the same as in Fig. 23(a) with
$\Lambda=4.0$. It is worth noting that all of the shown ZQMs are multi-
particle excitations, meaning that BdG does not account for a single ZQM in
the low-energy spectrum although there exist more than 20. By analyzing the
response densities of the ZQMs, it was found that they can be categorized by
their symmetry with respect to the individual wells, which is either gerade or
ungerade. Specific examples are given in Figs. 26(b) and (c). The ZQMs are
most likely the easiest states to excite in an experiment where the BEC is
subject to either a quench of the potential or the interaction strength. For
example, a simple quench scenario to excite the gerade ZQMs would be a sudden
change of the lattice depth. To this end, the ground state of a BEC with
$N=10$ bosons in the triple well with interaction parameter $\Lambda=4.0$ and
lattice depth $V_{0}=4.9$ is calculated. Then, the depth is suddenly quenched
to $V_{0}=5.0$, and the subsequent time-evolution is investigated. Fig. 27
shows the Fourier spectrum of the dynamical evolution of the position variance
per particle, $\frac{1}{N}\Delta^{2}_{\hat{X}}$, obtained from both
MCTDHB($5$) and GP. Shown is the energy range $5.0\leq\omega\leq 15.5$. At the
many-body level, 3 distinct peaks are observed, whose positions correspond to
the first, second, and fourth gerade excitations of the spectrum in Fig.
26(a). They can thus be labeled according to the occupation of (+1) and (-1)
from the first and second single-particle band. The peaks corresponding to the
two-particle excitations at $\omega\sim 6.35$ and $\omega\sim 13.48$ are more
intense than the peak corresponding to the three-particle excitation at
$\omega\sim 9.02$. Excitations where higher particle numbers are involved are
not intense enough to be distinguished from the background noise. At the GP
level, the spectrum is flat, which means that none of the ZQMs obtained at the
many-body level is included in the mean-field dynamics, which was anticipated
due to the multi-particle nature of the ZQMs.
To summarize, it was demonstrated that the excitation spectrum of a weakly-
depleted BEC in a 1D optical lattice shows clear many-body effects which are
not accounted for at the mean-field level. A depletion of $f\sim 1\%$ is
sufficient to observe these effects. In larger lattices, the sensitivity to
many-body features even increases. It was further shown that a special type of
many-body excited states (ZQMs) are easily accessible by performing a simple
quantum quench to the system. A promising quantity to detect these states is
the position variance, as already seen in the previous section.
#### 6.3.3 Rotating Bose-Einstein condensates in an anharmonic trap
After the discussion of many-body excited states in 1D trapped BECs, an
example in 2D, namely the case of rotating BECs in an anharmonic, radially-
symmetric potential crater, is examined. Of particular interest is the weakly-
interacting regime in which the ground-state depletion is marginal and one
would naively expect mean-field approaches like the BdG theory to be valid and
accurate. The results below were recently published in Ref. [187] and it is
referred to this work for additional details.
Rotating BECs were of high interest in the past. Research dealt with the
occurrence of quantized vortices [161, 188, 162, 163, 164] or with the
similarity to the fractional quantum Hall effect for charged particles in the
vicinity of a magnetic field [189, 191, 190, 192]. Moreover, low-lying
excitations in such systems were addressed, for example in vortex lattices
[26, 193, 194, 195]. Furthermore, investigations were made on the decay of the
counter-rotating quadrupole mode [196], on Tkachenko modes [197, 198], on the
twiston spectrum [199], or on excitations in anharmonic traps [200, 201]. In
the latter works, the mean-field excitations were calculated by utilizing the
BdG theory. Computations employing a many-body description arerather rare. An
examples is the analysis of the yrast spectra in a harmonic confinement
obtained by exact diagonalization using the lowest Landau level approximation
[203, 204, 202, 205].
As mentioned above, accurate many-body results for the low-energy spectrum of
rotating BECs in an anharmonic, radially-symmetric trapping potential are
presented. Especially, the effect of the angular momentum in the BECs’ ground
states on the excitation energies is discussed. As for the examples in the
earlier sections, the BdG and many-body results are compared to each other.
Figure 28: (a)-(e) Ground-state densities of rotating BECs with $N=10$ bosons
and interaction parameter $\Lambda=0.5$ for different vorticities $l$. Results
are shown for MCTDHB($7$), the corresponding GP densities look alike (not
shown). For non-zero vorticity, the ground state is a single vortex whose core
size grows with its charge $l$. (f) Low-energy excitation spectra atop the
ground states from the upper panels. Many-body results obtained by LR-
MCTDHB($7$) are denoted by red and blue squares, the BdG results are denoted
by black and blue lines. Once the ground state has non-zero vorticity,
deviations between the mean-field and many-body excitation energies are
observed, and they grow with $l$. Moreover, the density of (many-body) excited
states grows strongly. All ground states are only slightly depleted, with a
maximum of $f\approx 0.38\%$ for $l=4$. Symbols colored in blue refer to the
(+1) excitation energies described in the main text. All quantities are
dimensionless. See text for details. The figure is adapted from Ref. [187].
The crater in which the BEC is confined reads
$V(r)=\begin{cases}C\,e^{-0.5\,(r-R_{C})^{4}}\,\,&\mbox{if }r\leq R_{C}\\\
C\,\,&\mbox{if }r>R_{C}\end{cases}$ (193)
which is essentially the same as in Eq. (179) but without the central ring-
shaped barrier. Here, the radius of the crater is set to $R_{C}=3.0$ and the
height of the wall is kept at $C=200.0$. Furthermore, the same Gaussian
repulsion as in Eq. (180) is employed.
Excitations in this system are analyzed in the rotating frame of reference,
where the Hamiltonian $\hat{H}$ in Eq. (2) contains an additional contribution
accounting for the rotation around the $z$-axis. The whole Hamiltonian thus
reads
$\hat{H}_{\text{rot}}=\hat{H}-\Omega_{\text{rot}}\hat{L}_{z}$ (194)
where $\Omega_{\text{rot}}$ denotes the angular velocity around the $z$-axis
and $\hat{L}_{z}$ denotes the operator of the angular momentum in the
$z$-direction.
A LR analysis is applied atop the ground states of $N=10$ bosons with weak
repulsion strength $\Lambda=0.5$ for different values of
$\Omega_{\text{rot}}$, and thus of the angular momentum per particle $l$,
which is referred to as the vorticity or the charge below. The degree of
condensation, $N(1-f)$, ranges from $9.999$ ($l=0$) to $9.962$ ($l=4$) out of
10 particles, meaning that the BEC is highly condensed for all considered
parameter values. The upper panels in Fig. 28 show the ground-state densities
calculated with MCTDHB($7$) for vorticities $0\leq l\leq 4$. Whereas for
$l=0$, the density resembles a Gaussian in the center of the crater, the shape
of the densities for $l>0$ indicate that the ground state is a vortex with a
density node at its core. One can further see that this core is growing with
$l$. The corresponding GP densities look alike (not shown).
The lower panel in Fig. 28 compares the low-energy excitation spectrum
calculated atop these ground states, both for the mean-field BdG case (black
and blue lines) and for LR-MCTDHB($7$) (red and blue squares). The excitations
denoted by blue symbols are explained below. At first, the spectrum for $l=0$
is examined. Counting the number of excited states, one obtains 2 states from
the BdG approach and 3 states from the many-body approach. Moreover, the
energies of the two mean-field excitations coincide with the corresponding
many-body results. Since the BdG theory only accounts for single-particle
excitations, the third excited stated obtained from LR-MCTDHB is an excitation
where more than one particle is excited from the ground state. In particular,
it represents the state where two bosons are put into the orbital
corresponding to the first excitation. The latter single-particle excitation
is referred to as (+1) since it describes the case of putting a boson from the
ground-state mode with angular momentum $l_{z}=0$ to an orbital with
$l_{z}=1$. Examining the spectra for $l>0$, one observes two major changes. At
first, the density of excitations in the low-energy spectrum clearly grows
with $l$, which can be deduced from the large number of many-body excited
states entering. The number of mean-field excitations also grows (from 2
states for $l=0$ to 4 states for $l=4$), but very slightly compared to the
number of many-body states. Therefore, the amount of excited states that BdG
misses is strongly growing with the vorticity of the BEC’s ground state. The
second observation is that the values of the single-particle energies from BdG
and LR-MCTDHB do not coincide any longer, as it was the case for $l=0$. In
fact, the entire BdG spectrum for vorticity $l=4$ appears to be completely
unrelated to the corresponding many-body spectrum. The increasing gap between
the (+1) results of the two approaches is clearly visible.
Fig. 29 addresses the evolution of the energy of (+1), denoted by
$\omega_{(+1)}$, with respect to the angular velocity $\Omega_{\text{rot}}$.
Shown are the results obtained from the BdG theory, as well as many-body
results for different numbers of bosons ($N=10,\,100$ and $1000$). In the
$l=0$ sector, all energies coincide, indicating that for this case the BdG
theory is sufficient. Henceforth, this obviously changes. With growing vortex
charge of the ground state, the mean-field and many-body lines for (+1)
separate further from each other. Most importantly, increasing the number of
bosons in the condensate only marginally changes this, meaning that the many-
body and mean-field results for large $N$ approach each other very slowly (see
inset). As mentioned in the previous section, it was proven analytically that
for a trapped BEC with sufficiently weak and long-ranged repulsion, the
excitation energies in the Hartree limit converge towards the BdG energies
[185]. The same is believed to happen for a rotating BEC under certain
conditions [206].
Figure 29: Excitation energies $\omega_{(+1)}$ for different angular
velocities $\Omega_{\text{rot}}$. Compared are the BdG energies and the many-
body results for different numbers of bosons $N$, obtained by LR-MCTDHB. The
interaction strength is $\Lambda=0.5$. The transition from ground-state
vorticity $l$ to $l+1$ between two adjacent analyzed values of
$\Omega_{\text{rot}}$ are denoted by dotted vertical lines (mean-field in
black and many-body in red). The higher the ground-state vorticity $l$, the
larger is the separation between the BdG and many-body energies. Inset:
Magnified view for $\Omega_{\text{rot}}=3.8$. All quantities are
dimensionless. See text for details. The figure is adapted from Ref. [187].
Figure 30: Evolution of the energetic difference $\Delta_{N}$ between the
mean-field and many-body results for $\omega_{(+1)}$ for different particle
numbers $N$. The differences are shown relative to $\Delta_{10}$, i.e., the
gap for $N=10$ bosons. Solid lines denote exponential fits. The gap becomes
larger as $l$ grows. Even for $l=1$ and $N=1000$, it is approximately $97\%$
of $\Delta_{10}$, and the slopes of the fit curves do not indicate a sharp
decent towards zero if $N$ is increased further. All quantities are
dimensionless. See text for details. The figure is taken from Ref. [187].
In Fig. 30, the dependence of the difference $\Delta_{N}$ between the mean-
field and many-body energies of (+1) on the number of bosons in the BEC is
analyzed in more detail. The results are given relative to the energy for
$N=10$ bosons, denoted by $\Delta_{10}$. For all shown vorticities, the gap
becomes smaller with increasing $N$. However, it is obvious that even for BECs
with an experimentally relevant number of bosons ($N=1000$) one is still far
away from the predictions of the BdG theory. For $l=1$, one obtains that the
gap size $\Delta_{1000}$ is still about 97$\%$ of $\Delta_{10}$, and
correspondingly higher for $l>1$. Additionally, the shape of the exponential
fit curves to the numerical data does not show any sign of a steep descend
when $N$ is increased further. It can firstly be deduced that the excitation
energies approach the BdG predictions for $l>0$ very slowly, and, secondly,
one apparently needs a very high number of particles in order to enter the
regime where BdG yields approximately the correct results. However, it is
important to note that the convergence towards the BdG excitation energies in
the Hartree limit is not proven for the system under consideration, since Ref.
[206] only deals with the case of a rotating BEC with broken radial symmetry.
The latter situation appears, e.g., for a BEC in a harmonic confinement when
the rotation is fast enough to generate multiple quantized vortices in the
ground state (see, e.g., Ref. [207]). Although the numerical results cannot
give a definite answer for the Hartree limit, they at least suggest that the
regime in which one is clearly in need of an accurate many-body theory for the
excitation energies covers the experimentally relevant regime of rotating
BECs, with $N$ being of the order of $10^{3}-10^{6}$ bosons. A further
discussion on the applicability of the GP and BdG theories in the Hartree
limit can be found in Appendix A.
Figure 31: Low-energy Fourier spectrum of the time evolution of the position
variance in $x$-direction, $\frac{1}{N}\Delta^{2}_{\hat{X}}$, due to a quantum
quench of the radius $R_{C}$ from $3.01$ to $3.0$. The number of bosons is
$N=10$. The many-body spectrum (red), obtained from the evolution with $M=7$
orbitals, shows two sharp peaks in the considered energy range, whereas the GP
spectrum (black) is flat. The first peak refers to the two-particle excitation
$(1_{+1},1_{-1})$, whereas the second peak indicates a de-excitation between
two higher-lying states (see right inset). The left inset shows the evolution
of $\frac{1}{N}\Delta^{2}_{\hat{X}}$ from $0\leq t\leq 30$. The depletion of
the BEC stays below $0.4\%$ during the entire propagation. All quantities are
dimensionless. See text for details.
It is finally suggested, similar to the 1D systems discussed in Section 6.3.1,
how a quench can be utilized in order to excite breathing modes. To this end,
the ground state of a BEC with $N=10$ bosons and angular velocity
$\Omega_{\text{rot}}=3.2$ in the same trap as described in Eq. (193), but with
a slightly larger radius of $R_{C}=3.01$, is calculated. The obtained ground
state has the same symmetry and vorticity $l=3$ as the one for $R_{C}=3.0$ and
$\Omega_{\text{rot}}=3.2$. Then, the crater is suddenly shrinked back to
$R_{C}=3.0$ and the dynamical evolution is computed with MCTDHB($7$), which is
sufficient in order to obtain converged dynamics for the position variance. As
could be seen in the previous sections, the latter quantity is very sensitive
for many-body effects, and therefore, a spectral analysis of its dynamics by
computing the Fourier spectrum of $\frac{1}{N}\Delta^{2}_{\hat{X}}(t)$ is
carried out. Due to the radial symmetry, one could as well analyze the
position variance in the $y$-direction. The result are shown in Fig. 31. The
many-body spectrum shows two distinct peaks, where the first corresponds to
the $(1_{+1},1_{-1})$ two-particle excitation, where one boson is excited to
an orbital with $l=4$ and one is excited to an orbital with $l=2$, such that
the total angular momentum is unchanged. The energy that LR-MCTDHB($7$)
predicts for this state is $\omega=0.605$ which coincides with the peak
position of the Fourier analysis. The second peak at $\omega\approx 1.12$
denotes a de-excitation between two states that are much higher in energy
($\omega\approx 10.6\rightarrow 9.47$, see right inset in Fig. 31). The
occurrence of de-excitations is an indication that the quench described above
already leads to second-order effects, meaning that one has left the LR
regime. With regard to the mean-field spectrum from the GP dynamics, one
observes that in the shown energy range, the Fourier spectrum is completely
flat. The absence of the first peak is anticipated because the GP theory does
not account for two-particle excitations by construction. The absence of the
second peak can be understood from the right inset where the Fourier spectrum
in the interval $9\leq\omega\leq 11$ is shown. One sees that in the GP case,
there is only one peak at $\omega\approx 10.6$, but no second one at
$\omega\approx 9.47$ as in the many-body spectrum. Thus, very different
spectra from the dynamics of the variance at the mean-field and many-body
levels are obtained, and it is shown that the suggested quench leads to an
excitation of a many-body breathing mode. It is stressed that the depletion
stays below $0.4\%$ for the entire propagation, which means that the BEC is
essentially condensed.
To summarize, the low-energy spectrum of a rotating BEC in an anharmonic,
radially-symmetric crater shows significant many-body effects, even for
weakly-interacting, highly-condensed systems. These effects can be observed
already for the lowest-in-energy, single-particle excited states, and remain
even for mesoscopic and large BECs. Furthermore, a simple quench triggers a
many-body breathing mode, which can be observed, e.g., by analyzing the
Fourier spectrum of the position variance.
## 7 Summary and Conclusions
This review presents recently developed many-body approaches capable to
describe and compute accurately the ground state and dynamics as well as the
energy spectrum of trapped ultracold bosons. It is relevant to employ methods
which are on the same footing. The dynamics is calculated by the well-known
MCTDHB method, the ground state is obtained by imaginary-time propagation
using the same method, and the energy spectrum of the system is then
determined by applying the linear-response theory to the ground state computed
via MCTDHB. The resulting theory is named LR-MCTDHB.
The MCTDHB theory is briefly discussed and more attention is paid to the less
spread LR-MCTDHB theory. As the implementation of the MCTDHB is well
documented in the literature, only references are given here, while the newly
developed LR-MCTDHB implementation which makes the theory widely applicable,
is presented here in some detail for the first time. The latter posed a major
challenge due to both the complexity of the underlying theory and the
accompanying technical difficulties like the vast memory consumption and the
necessity of parallelization. The code structure is explained in detail both
with respect to the construction of the linear-response matrix and its
subsequent partial diagonalization which provides the desired excitation
energies. A comprehensive description of the procedure used to calculate the
lowest eigenvalues is also provided. The corresponding benchmarks of the LR-
MCTDHB code (see also Appendix C) with an analytically solvable many-boson
non-trivial model demonstrate its correct functioning and show that the theory
provides numerically exact results.
The second main goal of this review is to present applications to many-body
dynamics and excitations of BECs in interesting scenarios, and, of course, to
discuss the underlying physics. Several systems of trapped BECs in one and two
dimensions are addressed in detail and the main attention is drawn to the
detection of many-body effects that are not describable by applying mean-field
theory. Particularly emphasized are the tunneling dynamics in double-well
systems and the development of fragmentation as a function of time in
initially coherent BECs. The applications shown in this work have demonstrated
that a full many-body treatment of the dynamics of trapped interacting BECs is
inevitable. For instance, it is exhibited that the out-of-equilibrium
tunneling dynamics of initially coherent BECs in 1D and 2D double wells are of
many-body nature. After a few Rabi cycles, tunneling between the wells becomes
suppressed at the many-body level, whereas unperturbed oscillations are
predicted at the mean-field level. Surprisingly, the degree of fragmentation
in 1D is universal for constant values of the interaction parameter. Going to
2D, the sensitivity of the tunneling dynamics to many-body effects is even
enhanced as soon as the bosons carry angular momentum. This can also be
deduced from the larger amount of orbitals necessary in the limit of weak
repulsion as compared to the case of bosons with zero angular momentum.
Further applications deal with the nucleation of phantom vortices, which are
pure many-body objects not observable in the condensate density, as well as
with the tunneling of bosons to open space. In both cases, the initially
coherent systems become fragmented and are, therefore, in need of an accurate
many-body description.
With respect to excitations, it is observed that the ground-state depletion
has a strong impact on the low-energy spectrum. Even for marginal depletion
where one might expect mean-field approaches to give accurate results,
substantial many-body effects like the shift of excitation energies and the
appearance of complicated many-boson excitations occur. Applications in 1D
explores the many-body effects in double- and triple-well systems as well as
in larger lattices. In 2D, it is found that once the bosons carry angular
momentum and the ground state becomes a vortex, the deviations between the
mean-field and many-body excitation spectra grow strongly. Moreover, this
discrepancy tends to persist when the number of particles in the condensate
increases up to an experimentally relevant size. This has a potentially large
impact on experiments on ultracold bosons since the dynamics are mostly
determined by the lowest-in-energy excitations. Hopefully, these findings and
the new insights into the physics of ultracold bosons will stimulate future
experiments.
Before closing, we would like to add several remarks concerning further
general and technical developments of the methods which can further broaden
their range of applicability. Until now the applications of the many-body
linear-response theory were for excitations on top of the system’s ground
state. We emphasize that the LR-MCTDHB can also be applied to any other state
obtained via MCTDHB. In other words, it is possible to compute excitations on
top of any other state. This will help to address excitations which are
difficult to reach from the ground state. Another general possibility lies in
the formulation of LR-MCTDHB in block-diagonal form, which is potentially
advantageous in calculating the low-energy spectrum, as discussed in Appendix
B.
Apart from these, further general extensions and developments of both the LR
theory itself as well as of its numerical implementation are worth of
additional investigation. With regard to theoretical developments, an
extension of LR-MCTDHB to BEC mixtures, i.e., compound systems of different
types of bosons, is highly desirable as their popularity has substantially
grown [208, 184, 209, 210, 211, 212, 213]. The underlying MCTDHB theory for
bosonic mixtures, also including the possibility of internal degrees of
freedom, has been formulated [53, 70, 71], but a linear-response theory is
completely missing. It turns out that the so-called multi-layer MCTDHB,
introduced and described in Ref. [70], is particularly advantageous
numerically. To this end, an interesting goal would be to apply the linear-
response theory to multi-layer MCTDHB. Moreover, mixtures of bosons and
fermions might become even more relevant [71, 214, 218, 215, 216, 217, 219],
meaning that also the equations of motion of the fermionic counterpart of
MCTDHB, called MCTDHF (see, e.g., Ref. [220] and references therein, and the
recently developed implementation described in Ref. [221]), should be utilized
to develop the respective linear-response theory.
Further technical developments which accelerate the computations are, of
course, also welcome. One interesting possible extension lies in the
application of linear-response theory to the time-dependent restricted-active-
space self-consistent-field (TD-RASSCF) method, introduced in Ref. [222]. The
latter theory is similar to MCTDHB, but is an approximation to it. It is,
however, capable of reducing systematically the amount of necessary
coefficients significantly. This is achieved by restricting the possible
excitations in the active space of the single-particle orbitals according to
certain conditions or symmetries of the physical problem at hand. Thus, the
dimensionalities of the linear-response matrices which would result when
constructed based on the linear-response theory on top of TD-RASSCF, could
also be greatly reduced. Such an approach will substantially simplify the
computation of the excitation spectrum and, moreover, enable the treatment of
even larger systems, i.e., with more particles and if necessary higher numbers
of orbitals.
With regard to the technical development of the current implementation, the
application of additional techniques to diagonalize the often huge linear-
response matrix can be beneficial. A potentially very powerful method may be
the FEAST algorithm [224], which is a contour integration and density matrix-
based method. It is available in a free software package [225] and also allows
for high-level parallelization. In particular, it enables the specification of
an energy interval in which the desired eigenvalues should be found. Hence, it
might become easier to explore not just the lowest but also intermediate
eigenvalues.
The above outlined avenues for future research demonstrate the large variety
of perspectives, which this work will hopefully stimulate.
## Appendix A Mean-field theory
In this Appendix, several mean-field approaches to describe the ground state
and dynamics as well as the excitation spectra of trapped BECs are presented.
At first, the GP single-orbital mean-field theory (Section A.1.1) and the LR
theory atop it, yielding the famous BdG equations (Section A.1.2), are
described. Those approaches are the most common ones in the literature, and
are expected to accurately describe the physics of coherent BECs. They are
also used in this work to compare mean-field and many-body results.
Afterwards, a time-dependent multi-orbital mean-field approach, called TDMF
(Section A.2.1), and its corresponding LR theory LR-BMF (Section. A.2.2), are
discussed. The latter two theories are suitable for describing fragmented BECs
at the mean-field level. They can be understood as bridging theories between
the GP/BdG theories and the full many-body descriptions given by MCTDHB and
LR-MCTDHB, respectively.
### A.1 Single-orbital approaches
#### A.1.1 The GP equation
The GP equation is commonly used to obtain the ground state and the time
evolution of a coherent BEC, i.e., where all $N$ bosons occupy the same
single-particle state $\phi(\mathbf{r},t)$. The many-boson wave function
$\Psi$ is thus given by a single Hartree product,
$\Psi(\mathbf{r}_{1},...,\mathbf{r}_{N};t)=\phi(\mathbf{r}_{1},t)\,...\,\phi(\mathbf{r}_{N},t)\equiv|N;t\rangle.$
(A.1)
Originally, the GP equation is derived assuming contact interaction between
the bosons. However, to be more general, the two-body interaction potential is
taken to be of general form $\hat{W}(\mathbf{r},\mathbf{r}^{\,\prime})$.
The GP equation can be derived from the least action principle defined in Eq.
(25). The expectation value that appears in the action functional, one obtains
$\left\langle
N;t\left|\hat{H}-i\frac{\partial}{\partial\,t}\right|N;t\right\rangle=Nh_{\text{\tiny{GP}}}+\frac{\lambda_{0}N(N-1)}{2}W_{\text{\tiny{GP}}}-iN\left(\frac{\partial}{\partial
t}\right)_{\text{\tiny{GP}}}$ (A.2)
where the quantities $h_{\text{\tiny{GP}}}$ and $W_{\text{\tiny{GP}}}$ are
defined in Eqs. (21) and (22) utilizing $\phi(\mathbf{r},t)$. The time
derivative is given by $\left(\frac{\partial}{\partial
t}\right)_{\text{\tiny{GP}}}=\int\phi^{\ast}(\mathbf{r},t)\,\frac{\partial}{\partial
t}\,\phi(\mathbf{r},t)\,d\mathbf{r}$. One arrives at
$\displaystyle 0$ $\displaystyle=\frac{\delta
S}{\delta\phi^{\ast}(\mathbf{r},t)}\left[Nh_{\text{\tiny{GP}}}+\frac{\lambda_{0}N(N-1)}{2}W_{\text{\tiny{GP}}}-iN\left(\frac{\partial}{\partial
t}\right)_{\text{\tiny{GP}}}\right]$
$\displaystyle=N\left[\hat{h}+\lambda_{0}(N-1)\int\hat{W}(\mathbf{r},\mathbf{r}^{\,\prime})|\phi(\mathbf{r}^{\,\prime},t)|^{2}d\mathbf{r}^{\,\prime}-i\frac{\partial}{\partial
t}\right]\phi(\mathbf{r},t).$ (A.3)
Thus, the time-dependent GP equation used in this work is thus given by
$\left[\hat{h}+\Lambda\int\hat{W}(\mathbf{r},\mathbf{r}^{\,\prime})\,|\phi(\mathbf{r}^{\,\prime},t)|^{2}\,d\mathbf{r}^{\,\prime}\right]\phi(\mathbf{r},t)=i\frac{\partial}{\partial
t}\phi(\mathbf{r},t)$ (A.4)
with the mean-field interaction parameter $\Lambda=\lambda_{0}(N-1)$. As
mentioned above, for the contact interaction potential it takes the original
form as derived in [18, 19] which can also be found in standard textbooks [20,
21]. The time-independent GP equation reads
$\left[\hat{h}+\Lambda\int\hat{W}(\mathbf{r},\mathbf{r}^{\,\prime})\,|\phi^{0}(\mathbf{r}^{\,\prime})|^{2}\,d\mathbf{r}^{\,\prime}\right]\phi^{0}(\mathbf{r})=\mu\phi^{0}(\mathbf{r})$
(A.5)
where $\mu$ is the chemical potential. Whereas Eq. (A.4) describes the time-
evolution of a fully-condensed BEC, Eq. (A.5) is used to determine the GP
ground-state orbital $\phi^{0}(\mathbf{r})$.
This section is closed with a brief discussion of the applicability of the GP
theory in the Hartree limit of repulsive trapped BECs, meaning for
$N\rightarrow\infty$ with fixed interaction parameter $\Lambda$. It has been
proven analytically that in this limit, the GP theory provides the exact
energy and density per particle of the system’s ground state [226]. This,
however, does not necessarily mean that the corresponding GP wave function,
Eq. (A.1), coincides with the exact wave function in this limit. To this end,
it has been demonstrated recently that the overlap of the exact and GP wave
function in the Hartree limit can be clearly less than unity or even be
vanishingly small [227]. Moreover, a subsequent study has shown the many-body
character of the exact wave function in the Hartree limit by applying many-
body perturbation theory to the mean-field Hamiltonian [228], which
demonstrates that the mean-field and exact wave functions can be very
different from each other. Moreover, as a further interesting result, it was
found that the condensate depletion in the Hartree limit becomes a constant.
These findings potentially have severe consequences for the excitation
spectrum of a BEC in or close to the Hartree limit, which is discussed at the
end of Section A.1.2.
#### A.1.2 The BdG equations
The LR theory atop the time-dependent GP equation, Eq. (A.4), is derived
below. Therefore, a weak time-dependent perturbation is added to the
Hamiltonian, i.e.,
$\hat{H}(\mathbf{r})\rightarrow\hat{H}(\mathbf{r})+\hat{H}_{\text{ext}}(\mathbf{r},t)$.
The perturbation $\hat{H}_{\text{ext}}(\mathbf{r},t)$ explicitly reads
$\hat{H}_{\text{ext}}(\mathbf{r},t)=f^{+}(\mathbf{r})e^{-i\omega
t}+f^{-}(\mathbf{r})e^{i\omega t}$ (A.6)
with the real amplitudes $f^{+}(\mathbf{r})$ and $f^{-}(\mathbf{r})$ and the
frequency $\omega$. The response of the system is calculated around the
stationary GP ground-state orbital $\phi^{0}(\mathbf{r})$ obtained from Eq.
(A.5). For the perturbed orbital $\phi(\mathbf{r},t)$, the ansatz
$\sqrt{N}\phi(\mathbf{r},t)=e^{-i\mu
t}\left(\sqrt{N}\phi^{0}(\mathbf{r})+u(\mathbf{r})e^{-i\omega
t}+v^{\ast}(\mathbf{r})e^{i\omega t}\right)$ (A.7)
is utilized. In Eq. (A.7), the quantities $u(\mathbf{r})$ and $v(\mathbf{r})$
denote small and time-independent response amplitudes. Plugging in Eqs. (A.6)
and (A.7) into the time-dependent GP equation, Eq. (A.4), yields in zeroth
order the time-independent GP equation, Eq. (A.5). In first order, collecting
all terms proportional to $e^{-i\omega t}$ leads to
$\displaystyle(\omega+\mu)u(\mathbf{r})$
$\displaystyle=\sqrt{N}f^{+}(\mathbf{r})\phi^{0}(\mathbf{r})+\hat{H}_{\text{\tiny{GP}}}u(\mathbf{r})+\Lambda\int\hat{W}(\mathbf{r},\mathbf{r}^{\,\prime})\phi^{0}(\mathbf{r}^{\,\prime})v(\mathbf{r}^{\,\prime})\phi^{0}(\mathbf{r})d\mathbf{r}^{\,\prime}$
$\displaystyle+\Lambda\int\hat{W}(\mathbf{r},\mathbf{r}^{\,\prime})\phi^{0,\ast}(\mathbf{r}^{\,\prime})u(\mathbf{r}^{\,\prime})\phi^{0}(\mathbf{r})d\mathbf{r}^{\,\prime}$
(A.8)
where the GP Hamiltonian $\hat{H}{\text{\tiny{GP}}}$ is defined as
$\hat{H}{\text{\tiny{GP}}}=\hat{h}+\Lambda\int\hat{W}(\mathbf{r},\mathbf{r}^{\,\prime})\,|\phi^{0}(\mathbf{r}^{\,\prime},t)|^{2}\,d\mathbf{r}^{\,\prime}.$
(A.9)
Collecting all terms proportional to $e^{i\omega t}$ results in
$\displaystyle(\mu-\omega)v^{\ast}(\mathbf{r})$
$\displaystyle=\sqrt{N}f^{-}(\mathbf{r})\phi^{0}(\mathbf{r})+\hat{H}_{\text{\tiny{GP}}}v^{\ast}(\mathbf{r})+\Lambda\int\hat{W}(\mathbf{r},\mathbf{r}^{\,\prime})\phi^{0}(\mathbf{r}^{\,\prime})u^{\ast}(\mathbf{r}^{\,\prime})\phi^{0}(\mathbf{r})\,d\mathbf{r}^{\,\prime}$
$\displaystyle+\Lambda\int\hat{W}(\mathbf{r},\mathbf{r}^{\,\prime})\phi^{0,\ast}(\mathbf{r}^{\,\prime})v^{\ast}(\mathbf{r}^{\,\prime})\phi^{0}(\mathbf{r})\,d\mathbf{r}^{\,\prime}.$
(A.10)
Complex conjugation and multiplication by $(-1)$ of Eq. (A.1.2) opens the
possibility to cast Eqs. (A.1.2) and (A.1.2) in matrix form. The result reads
$\left(\mathcal{L}_{\text{BdG}}-\omega\right)\begin{pmatrix}u(\mathbf{r})\\\
v(\mathbf{r})\end{pmatrix}=\begin{pmatrix}-\sqrt{N}f^{+}(\mathbf{r})\phi^{0}(\mathbf{r})\\\
\sqrt{N}f^{-,\ast}(\mathbf{r})\phi^{0,\ast}(\mathbf{r})\end{pmatrix}$ (A.11)
with the BdG matrix
$\mathcal{L}_{\text{BdG}}=\begin{pmatrix}A&B\\\
-B^{\ast}&-A^{\ast}\end{pmatrix}$ (A.12)
where the quantities $A$ and $B$ are given by
$A=\hat{H}_{\text{\tiny{GP}}}-\mu+\Lambda\int
d\mathbf{r}^{\,\prime}\hat{W}(\mathbf{r},\mathbf{r}^{\,\prime})\phi^{0,\ast}(\mathbf{r}^{\,\prime})\phi^{0}(\mathbf{r})\hat{\mathcal{P}}_{\mathbf{r}\mathbf{r}^{\,\prime}}$
(A.13)
and
$B=\Lambda\int
d\mathbf{r}^{\,\prime}\hat{W}(\mathbf{r},\mathbf{r}^{\,\prime})\phi^{0}(\mathbf{r}^{\,\prime})\phi^{0}(\mathbf{r})\hat{\mathcal{P}}_{\mathbf{r}\mathbf{r}^{\,\prime}}\,.$
(A.14)
where the operator $\hat{\mathcal{P}}_{\mathbf{r}\mathbf{r}^{\,\prime}}$ was
introduced in Eq. (96). In the literature, Eq. (A.11) with contact interaction
$\hat{W}(\mathbf{r},\mathbf{r}^{\,\prime})=\delta(\mathbf{r}-\mathbf{r}^{\,\prime})$
and zero RHS, i.e.,
$\left(\mathcal{L}_{\text{BdG}}-\omega\right)\begin{pmatrix}u(\mathbf{r})\\\
v(\mathbf{r})\end{pmatrix}=0,$ (A.15)
is referred to as the BdG equations [23, 24]. In this work, however, the
interaction potential is in general different from the contact interaction
potential. Moreover, the particle-conserving BdG equations [229, 230, 231,
232], given by
$\mathbf{\mathcal{P}}\mathcal{L}_{\text{BdG}}\mathbf{\mathcal{P}}\begin{pmatrix}u(\mathbf{r})\\\
v(\mathbf{r})\end{pmatrix}=\omega\begin{pmatrix}u(\mathbf{r})\\\
v(\mathbf{r})\end{pmatrix}$ (A.16)
where the projector
$\mathbf{\mathcal{P}}=\mathbb{1}-|\phi^{0}\rangle\langle\phi^{0}|$ removes any
contribution of the ground-state orbital $\phi^{0}(\mathbf{r})$ from the
response amplitudes $u(\mathbf{r})$ and $v(\mathbf{r})$, are used to calculate
the low-energy excitation spectra at the mean-field level.
From the above derivation one can readily see the close connection of the GP
orbital and the BdG equations. As discussed in the previous section on the GP
theory, the wave function of the system in the Hartree limit can be very
different from the GP wave function, although the latter gives the exact
energy and density per particle. Thus, it can be questioned whether the LR
theory atop the GP wave function, resulting in the BdG equations, is in
general reliable in the Hartree limit. Scientific works that prove the
exactness of the BdG spectrum in this limit under certain conditions were
discussed in Section 6.3.3. Nevertheless, the example of a rotating BEC in an
anharmonic trap, which does not fulfill the assumptions made in Ref. [206],
serves as an example where at least in the experimentally relevant regime of
$N\approx 10^{3}-10^{6}$ bosons in the condensate the lowest-in-energy
excitations substantially deviate from the BdG predictions.
### A.2 Multi-orbital approaches
#### A.2.1 Time-dependent multi-orbital mean field (TDMF)
The TDMF theory, introduced in Ref. [233], is briefly described. It can be
seen as an intermediate step between the GP theory and MCTDHB because it
allows for more than one single-particle orbital that the bosons can occupy,
and is thus applicable to fragmented condensates. However, as a major
difference to the full many-body description in Section 3, the occupation of
these single-particle orbitals is fixed. The ansatz for the $N$-boson wave
function is given by
$\Psi(\mathbf{r}_{1},...,\mathbf{r}_{N};t)=\hat{\mathcal{S}}\phi_{1}(\mathbf{r}_{1},t)\phi_{2}(\mathbf{r}_{2},t)\,...\,\phi_{N}(\mathbf{r}_{N},t)$
(A.17)
where the operator $\hat{\mathcal{S}}$ accounts for the symmetrization of the
wave function. In general, not all single-particle orbitals $\phi_{i}$ need to
be different. The only constraint is that the number of particles is fixed
such that $N=\sum_{i=1}^{M}n_{i}$ where $M$ is the number of different
orbitals in Eq. (A.17) and the $n_{i}$ are their fixed occupation numbers.
Applying the least action principle to
$S=\int dt\left(\langle\Psi|\hat{H}-i\frac{\partial}{\partial
t}|\Psi\rangle-\sum_{i,j=1}^{M}\mu_{ij}(t)[\langle\phi_{i}|\phi_{j}\rangle-\delta_{ij}]\right)$
(A.18)
with the time-dependent Lagrange multipliers $\mu_{ij}(t)$ yields
$\displaystyle 0=\frac{\delta S}{\delta\phi_{k}^{\ast}(\mathbf{r},t)}$
$\displaystyle=n_{k}\left(\hat{h}-i\frac{\partial}{\partial
t}+\lambda_{0}(n_{k}-1)\hat{J}_{k}+\sum_{l\neq
k}^{M}\lambda_{0}n_{l}(\hat{J}_{l}+\hat{K}_{l})\right)|\phi_{k}\rangle$
$\displaystyle=\sum_{j=1}^{M}\mu_{kj}(t)|\phi_{j}(t)\rangle\quad\forall\,k=1,...,M$
(A.19)
where
$\displaystyle\hat{J}_{l}(\mathbf{r},t)$ $\displaystyle=\int
d\mathbf{r}^{\,\prime}\phi_{l}^{\ast}(\mathbf{r}^{\,\prime},t)\hat{W}(\mathbf{r},\mathbf{r}^{\,\prime})\phi_{l}(\mathbf{r}^{\,\prime},t)$
(A.20)
and
$\displaystyle\hat{K}_{l}(\mathbf{r},t)$ $\displaystyle=\int
d\mathbf{r}^{\,\prime}\phi_{l}^{\ast}(\mathbf{r}^{\,\prime},t)\hat{W}(\mathbf{r},\mathbf{r}^{\,\prime})\hat{\mathcal{P}}_{\mathbf{r}\mathbf{r}^{\,\prime}}\phi_{l}(\mathbf{r}^{\,\prime},t)$
(A.21)
are the direct and exchange interaction potentials, respectively. Multiplying
Eq. (A.19) by $\langle\phi_{l}|$ gives
$\mu_{kl}(t)=n_{k}\left(h_{lk}-\left(\frac{\partial}{\partial
t}\right)_{lk}+\lambda_{0}(n_{k}-1)\hat{J}_{k}W_{lkkk}+\sum_{j\neq
k}^{M}\lambda_{0}n_{j}(W_{ljjk}+W_{ljkj})\right)$ (A.22)
for the Lagrange multipliers. Plugging this into Eq. (A.19) results in
$\displaystyle i\mathbf{\hat{P}}\frac{\partial}{\partial
t}|\phi_{k}\rangle=\mathbf{\hat{P}}\left(\hat{h}+\lambda_{0}(n_{k}-1)\hat{J}_{k}+\sum_{l\neq
k}^{M}\lambda_{0}n_{l}(\hat{J}_{l}+\hat{K}_{l})\right)|\phi_{k}\rangle\quad\forall\,k=1,...,M$
(A.23)
where the projector
$\mathbf{\hat{P}}=\mathbb{1}-\sum_{j=1}^{M}|\phi_{j}\rangle\langle\phi_{j}|$
projects onto the tangential space of $\text{span}(\phi_{1},...,\phi_{M})$. To
simplify the above EOMs for the orbitals, the condition
$\langle\phi_{i}|\dot{\phi}_{j}\rangle=0\quad\forall\,i,j=1,...,M$ (A.24)
is invoked. This leads to the conservation of energy, and the projector on the
LHS of Eq. (A.23) can be omitted, i.e.,
$i\frac{\partial}{\partial
t}|\phi_{k}\rangle=\mathbf{\hat{P}}\left(\hat{h}+\lambda_{0}(n_{k}-1)\hat{J}_{k}+\sum_{l\neq
k}^{M}\lambda_{0}n_{l}(\hat{J}_{l}+\hat{K}_{l})\right)|\phi_{k}\rangle\quad\forall\,k=1,...,M.$
(A.25)
It is important to note that the condition in Eq. (A.24) is a manually-invoked
restriction to the orbitals within the TDMF theory, whereas the same condition
in MCTDHB appears naturally and rigorously. Furthermore, one observes that for
$M=1$, i.e., when only a single orbital is available, Eq. (A.25) reduces to
the (projected) time-dependent GP equation, Eq. (A.4). The TDMF theory has
been extended to describe spinor condensates and Bose-Bose or Bose-Fermi
mixtures as well, see in this context Ref. [234]. By imaginary-time
propagation of the TDMF EOMs, one obtains the corresponding theory for the
stationary equations, called the best mean field for condensates [235].
Applications can be found in, e.g., Refs. [236, 237, 238, 239].
#### A.2.2 Linear-response best-mean-field (LR-BMF)
A possible LR approach based on a multi-orbital mean-field description is the
LR-BMF theory which has been introduced in Ref. [240]. To derive it, the EOMs
of the previously described TDMF theory, Eq. (A.19), are linearized for the
case of adding a weak time-dependent external field to the Hamiltonian, i.e.,
$\hat{H}\rightarrow\hat{H}+\hat{H}_{\text{ext}}(t)$. The perturbation
$\hat{H}_{\text{ext}}(t)$ is assumed to be of the same form as in Eq. (A.6).
Plugging this into Eq. (A.19) yields
$\displaystyle\left(\hat{Z}_{i}-i\frac{\partial}{\partial
t}\right)|\phi_{i}\rangle-\sum_{j=1}^{M}\frac{\mu_{ij}(t)}{n_{i}}|\phi_{j}\rangle=-\hat{H}_{\text{ext}}(t)|\phi_{i}\rangle$
(A.26)
with the replacement
$\hat{Z}_{i}=\hat{h}+\lambda_{0}(n_{i}-1)\hat{J}_{i}+\sum_{l\neq
i}^{M}\lambda_{0}n_{l}(\hat{J}_{l}+\hat{K}_{l}).$ (A.27)
The Lagrange multipliers, after redefining them as
$\mu_{ij}(t)/n_{i}\rightarrow\mu_{ij}(t)$, can be expressed by
$\mu_{ij}(t)=\langle\phi_{j}|\hat{Z}_{i}+\hat{H}_{\text{ext}}(t)|\phi_{i}\rangle$
(A.28)
where the condition of Eq. (A.24) was used. The ansatz for the perturbed
orbitals is given by
$\phi_{i}(\mathbf{r},t)\approx\phi_{i}^{0}(\mathbf{r})+\delta\phi_{i}(\mathbf{r},t)$
(A.29)
Plugging Eq. (A.29) into Eq. (A.26) leads to
$\displaystyle-\hat{H}_{\text{ext}}(t)|\phi_{i}^{0}\rangle$
$\displaystyle=\left(\hat{Z}_{i}^{0}-i\frac{\partial}{\partial
t}\right)|\delta\phi_{i}\rangle-\sum_{j=1}^{M}\mu_{ij}^{0}(t)|\delta\phi_{j}\rangle$
$\displaystyle+\lambda_{0}(n_{i}-1)\delta\hat{J}_{i}|\phi_{i}^{0}\rangle+\sum_{j\neq
i}^{M}\lambda_{0}n_{j}(\delta\hat{J}_{j}+\delta\hat{K}_{j})|\phi_{i}^{0}\rangle$
(A.30) $\displaystyle-\sum_{j=1}^{M}\delta\mu_{ij}(t)|\phi_{j}^{0}\rangle$
with
$\displaystyle\delta\hat{J}_{i}(\mathbf{r},t)$ $\displaystyle=\int
d\mathbf{r}^{\,\prime}\delta\phi_{i}^{\ast}(\mathbf{r}^{\,\prime},t)\hat{W}(\mathbf{r},\mathbf{r}^{\,\prime})\phi_{i}^{0}(\mathbf{r}^{\,\prime})+\int
d\mathbf{r}^{\,\prime}\phi_{i}^{0,\ast}(\mathbf{r}^{\,\prime})\hat{W}(\mathbf{r},\mathbf{r}^{\,\prime})\delta\phi_{i}^{0}(\mathbf{r}^{\,\prime},t)$
(A.31)
and
$\displaystyle\delta\hat{K}_{i}(\mathbf{r},t)$ $\displaystyle=\int
d\mathbf{r}^{\,\prime}\delta\phi_{i}^{\ast}(\mathbf{r}^{\,\prime},t)\hat{W}(\mathbf{r},\mathbf{r}^{\,\prime})\hat{\mathcal{P}}_{\mathbf{r}\mathbf{r}^{\,\prime}}\phi_{i}^{0}(\mathbf{r}^{\,\prime})+\int
d\mathbf{r}^{\,\prime}\phi_{i}^{0,\ast}(\mathbf{r}^{\,\prime})\hat{W}(\mathbf{r},\mathbf{r}^{\,\prime})\hat{\mathcal{P}}_{\mathbf{r}\mathbf{r}^{\,\prime}}\delta\phi_{i}^{0}(\mathbf{r}^{\,\prime},t).$
(A.32)
The variation of the Lagrange multipliers can be expressed as
$\displaystyle\delta\mu_{ij}(t)=-\sum_{l=1}^{M}\mu_{il}^{0}\langle\phi_{j}^{0}|\delta\phi_{l}\rangle+\langle\phi_{j}^{0}|\delta(\hat{Z}_{i}|\phi_{i}\rangle)+\langle\phi_{j}^{0}|\hat{H}_{\text{ext}}(t)|\phi_{i}^{0}\rangle$
(A.33)
where the identity
$\hat{Z}_{i}^{0}|\phi_{i}^{0}\rangle=\sum_{l=1}^{M}\mu_{il}^{0}|\phi_{l}^{0}\rangle$
and integration by parts was used in the first term. One finally obtains
$\displaystyle\mathbf{\hat{P}}\left[\hat{Z}_{i}^{0}|\delta\phi_{i}\rangle-\sum_{j=1}^{M}\mu_{ij}^{0}|\delta\phi_{j}\rangle+\left(\lambda_{0}(n_{i}-1)\delta\hat{J}_{i}+\sum_{j\neq
i}^{M}\lambda_{0}n_{j}(\delta\hat{J}_{j}+\delta\hat{K}_{j})\right)|\phi_{i}^{0}\rangle\right]$
$\displaystyle=-\mathbf{\hat{P}}\hat{H}_{\text{ext}}(t)|\phi_{i}^{0}\rangle+i\frac{\partial}{\partial
t}|\delta\phi_{i}\rangle$ (A.34)
with the projector
$\mathbf{\hat{P}}=\mathbb{1}-\sum_{k}|\phi_{k}^{0}\rangle\langle\phi_{k}^{0}|$.
By first making the ansatz
$\delta\phi_{i}(\mathbf{r},t)=\frac{1}{\sqrt{n_{i}}}\left(u_{i}(\mathbf{r})e^{-i\omega
t}+v_{i}^{\ast}(\mathbf{r})e^{i\omega t}\right)$ (A.35)
with stationary response amplitudes $u_{i}(\mathbf{r})$ and
$v_{i}(\mathbf{r})$ and plugging this into Eq. (A.2.2) afterwards, it is again
possible, similar to derivation of the BdG equations in Section A.1.2, to
group the terms of the resulting equations with respect to their linearity to
either $e^{-i\omega t}$ or $e^{i\omega t}$. As a result, one arrives at an
eigenvalue equation of the form
$(\mathbf{\mathcal{P}}\mathbf{\mathcal{L}}-\omega)\begin{pmatrix}|\mathbf{u}\rangle\\\
|\mathbf{v}\rangle\end{pmatrix}=\mathbf{\mathcal{P}}\begin{pmatrix}-f^{+}(\mathbf{r})|\boldsymbol{\phi_{n}^{0}}\rangle\\\
f^{-,\ast}(\mathbf{r})|\boldsymbol{\phi_{n}^{0,\ast}}\rangle\end{pmatrix}$
(A.36)
with the vector
$|\boldsymbol{\phi_{n}^{0}}\rangle=|\sqrt{n_{1}}\phi_{1}^{0},...,\sqrt{n_{M}}\phi_{M}^{0}\rangle$,
the $2(M\times M)$ square matrix
$\mathbf{\mathcal{L}}=\begin{pmatrix}\mathbf{Z^{0}}-\boldsymbol{\mu^{0}}+\mathbf{A}&\mathbf{B}\\\
-\mathbf{B}^{\ast}&-(\mathbf{Z^{0}}-\boldsymbol{\mu^{0}}+\mathbf{A})^{\ast}\end{pmatrix},$
(A.37)
and the projection matrix
$\mathbf{\mathcal{P}}=\\{\mathcal{P}_{ij}\\}=\begin{cases}\mathbf{\hat{P}}&i=j\leq
M\\\ \mathbf{\hat{P}}^{\ast}&i=j>M\\\ 0&i\neq j\end{cases}$ (A.38)
of the same size. The matrix
$\mathbf{Z^{0}}=\text{diag}(\hat{Z}_{1}^{0},...,\hat{Z}_{M}^{0})$ is a
diagonal matrix, whereas $\boldsymbol{\mu^{0}}=\\{\mu_{ij}^{0}\\}$ contains
the Lagrange multipliers. The matrices $\mathbf{A}$ and $\mathbf{B}$, for the
case of the contact interaction potential, read
$\mathbf{A}=\\{A_{ij}\\}=\begin{cases}\lambda_{0}(n_{i}-1)|\phi_{i}^{0}|^{2}&i=j\\\
2\lambda_{0}\sqrt{n_{i}n_{j}}\phi_{i}^{0}\phi_{j}^{0,\ast}&i\neq j\end{cases}$
(A.39)
and
$\mathbf{B}=\\{B_{ij}\\}=\begin{cases}\lambda_{0}(n_{i}-1)(\phi_{i}^{0})^{2}&i=j\\\
2\lambda_{0}\sqrt{n_{i}n_{j}}\phi_{i}^{0}\phi_{j}^{0}&i\neq j.\end{cases}$
(A.40)
Since the term proportional to $\omega$ is the only one without a projector in
Eq. (A.36), a redundant projector can be added to the RHS of
$\mathbf{\mathcal{P}}\mathbf{\mathcal{L}}$, i.e.,
$\mathbf{\mathcal{P}}\mathbf{\mathcal{L}}\rightarrow\mathbf{\mathcal{P}}\mathbf{\mathcal{L}}\mathbf{\mathcal{P}}$.
The resulting homogeneous eigenvalue equation
$\mathbf{\mathcal{P}}\mathbf{\mathcal{L}}\mathbf{\mathcal{P}}\begin{pmatrix}|\mathbf{u}\rangle\\\
|\mathbf{v}\rangle\end{pmatrix}=\omega\begin{pmatrix}|\mathbf{u}\rangle\\\
|\mathbf{v}\rangle\end{pmatrix}$ (A.41)
is the central equation in the LR-BMF theory. A comparison to the BdG matrix
$\mathcal{L}_{\text{BdG}}$ in Eqs. (A.12)-(A.14) shows that for $M=1$, LR-BMF
reduces to the BdG theory, i.e., LR-BMF($M=1$)$\equiv$BdG. If one however
compares the LR matrix of Eq. (A.37) to the one of LR-MCTDHB in Section 3.2.2,
it can be seen that the latter is clearly more complicated because it also
includes the couplings between orbitals and coefficients. Again, in both mean-
field approaches presented in this Appendix, there is only one possible
configuration, i.e., the configuration $|N;t\rangle$ for the BdG case and the
configuration $|n_{1},...,n_{M};t\rangle$ for the multi-orbital mean-field
case. The occupations of the underlying single-particle orbitals are fixed. In
fact, the LR-BMF matrix
$\mathbf{\mathcal{P}}\mathbf{\mathcal{L}}\mathbf{\mathcal{P}}$ can be seen as
the upper-left block $\mathcal{L}_{oo}$ of the LR matrix in Eq. (93). Further
details on the LR-BMF theory, together with an application to a BEC in a
double well, can be found in Ref. [240].
## Appendix B LR-MCTDHB in block-diagonal form
In this Appendix, a complex transformation that essentially halves the
dimensionality of the homogeneous eigenvalue problem of Eq. (116) is
introduced. However, as shown below, it involves matrix-matrix multiplications
of individual blocks of the LR matrix $\mathcal{L}$. The latter products might
be numerically expensive. It is therefore not in general beneficial to perform
the transformation described below, which was first discussed in Ref. [241].
One considers the transformation matrix
$Q=\frac{1}{\sqrt{2}}\begin{pmatrix}1&1\theta\\\ 1&-1\theta\end{pmatrix},\quad
Q^{-1}=\frac{1}{\sqrt{2}}\begin{pmatrix}1&1\\\
1\theta&-1\theta\end{pmatrix},\quad QQ^{-1}=Q^{-1}Q=\mathbb{1}$ (B.1)
where the operation $\theta$ is defined via its action on any operator
$\hat{O}$ given by $\theta\hat{O}=\hat{O}^{\ast}$. A matrix of the form
$\mathcal{A}=\begin{pmatrix}A&B\\\ -B^{\ast}&-A^{\ast}\end{pmatrix}$ (B.2)
behaves under the transformation in Eq. (B.1) as
$\displaystyle Q\mathcal{A}Q^{-1}$
$\displaystyle=\frac{1}{2}\begin{pmatrix}1&1\\\
1\theta&-1\theta\end{pmatrix}\begin{pmatrix}A&B\\\
-B^{\ast}&-A^{\ast}\end{pmatrix}\begin{pmatrix}1&1\theta\\\
1&-1\theta\end{pmatrix}$ $\displaystyle=\begin{pmatrix}0&A-B\theta\\\
A+B\theta&0\end{pmatrix}.$ (B.3)
Multiplying the result of Eq. (B) with itself yields the block-diagonal matrix
$\left(Q\mathcal{A}Q^{-1}\right)\left(Q\mathcal{A}Q^{-1}\right)=\begin{pmatrix}(A-B\theta)(A+B\theta)&0\\\
0&(A+B\theta)(A-B\theta)\end{pmatrix}.$ (B.4)
This transforms the eigenvalue problem $\mathcal{A}\alpha=\lambda\alpha$ into
$\displaystyle\lambda^{2}(Q\alpha)$
$\displaystyle=\left(Q\mathcal{A}Q^{-1}\right)\left(Q\mathcal{A}Q^{-1}\right)(Q\alpha)$
$\displaystyle\Leftrightarrow\,\lambda^{2}\frac{1}{\sqrt{2}}\begin{pmatrix}u+v^{\ast}\\\
u-v^{\ast}\end{pmatrix}$
$\displaystyle=\begin{pmatrix}(A-B\theta)(A+B\theta)&0\\\
0&(A+B\theta)(A-B\theta)\end{pmatrix}\frac{1}{\sqrt{2}}\begin{pmatrix}u+v^{\ast}\\\
u-v^{\ast}\end{pmatrix}$ (B.5)
with $\alpha=\frac{1}{\sqrt{2}}\begin{pmatrix}u\\\ v\end{pmatrix}$. Thus, the
original eigenvalue problem of $\mathcal{A}$ can be split up into two smaller
eigenvalue problems of $\mathcal{L}^{(1)}\equiv(A-B\theta)(A+B\theta)$ and
$\mathcal{L}^{(2)}\equiv(A+B\theta)(A-B\theta)$ with
$\text{dim}(\mathcal{L}^{(1)})=\text{dim}(\mathcal{L}^{(2)})=\text{dim}(\mathcal{A})/2$.
Moreover, both $\mathcal{L}^{(1)}$ and $\mathcal{L}^{(2)}$ have the same
eigenvalues, which are the squared eigenvalues of $\mathcal{A}$.
Interchanging the second and third row as well as the second and third column
of the LR matrix in Eq. (3.2.2) which explicitly reads
$\displaystyle\mathcal{L}=\begin{pmatrix}\boldsymbol{\rho}^{-\frac{1}{2}}\hat{\mathbf{P}}\mathcal{L}_{oo}^{u}\hat{\mathbf{P}}\boldsymbol{\rho}^{-\frac{1}{2}}&\boldsymbol{\rho}^{-\frac{1}{2}}\hat{\mathbf{P}}\mathcal{L}_{oo}^{v}\hat{\mathbf{P}}^{\ast}(\boldsymbol{\rho^{\ast}})^{-\frac{1}{2}}&\boldsymbol{\rho}^{-\frac{1}{2}}\hat{\mathbf{P}}\mathcal{L}_{oc}^{u}&\boldsymbol{\rho}^{-\frac{1}{2}}\hat{\mathbf{P}}\mathcal{L}_{oc}^{v}\\\
-(\boldsymbol{\rho^{\ast}})^{-\frac{1}{2}}\hat{\mathbf{P}}^{\ast}\mathcal{L}_{oo}^{v,\ast}\hat{\mathbf{P}}\boldsymbol{\rho}^{-\frac{1}{2}}&-(\boldsymbol{\rho^{\ast}})^{-\frac{1}{2}}\hat{\mathbf{P}}^{\ast}\mathcal{L}_{oo}^{u,\ast}\hat{\mathbf{P}}^{\ast}(\boldsymbol{\rho^{\ast}})^{-\frac{1}{2}}&-(\boldsymbol{\rho^{\ast}})^{-\frac{1}{2}}\hat{\mathbf{P}}^{\ast}\mathcal{L}_{oc}^{v,\ast}&-(\boldsymbol{\rho^{\ast}})^{-\frac{1}{2}}\hat{\mathbf{P}}^{\ast}\mathcal{L}_{oc}^{u,\ast}\\\
\mathcal{L}_{co}^{u}\hat{\mathbf{P}}\boldsymbol{\rho}^{-\frac{1}{2}}&\mathcal{L}_{co}^{v}\hat{\mathbf{P}}^{\ast}(\boldsymbol{\rho^{\ast}})^{-\frac{1}{2}}&\mathbf{\mathcal{H}}-\varepsilon^{0}&0\\\
-\mathcal{L}_{co}^{v,\ast}\hat{\mathbf{P}}\boldsymbol{\rho}^{-\frac{1}{2}}&-\mathcal{L}_{co}^{u,\ast}\hat{\mathbf{P}}^{\ast}(\boldsymbol{\rho^{\ast}})^{-\frac{1}{2}}&0&-(\mathbf{\mathcal{H}}-\varepsilon^{0})^{\ast}\end{pmatrix}$
(B.6)
yields a matrix of the form given in Eq. B.2. Its submatrices are
$A=\begin{pmatrix}\boldsymbol{\rho}^{-\frac{1}{2}}\hat{\mathbf{P}}\mathcal{L}_{oo}^{u}\hat{\mathbf{P}}\boldsymbol{\rho}^{-\frac{1}{2}}&\boldsymbol{\rho}^{-\frac{1}{2}}\hat{\mathbf{P}}\mathcal{L}_{oc}^{u}\\\
\mathcal{L}_{co}^{u}\hat{\mathbf{P}}\boldsymbol{\rho}^{-\frac{1}{2}}&\mathbf{\mathcal{H}}-\varepsilon^{0}\end{pmatrix}$
(B.7)
and
$B=\begin{pmatrix}\boldsymbol{\rho}^{-\frac{1}{2}}\hat{\mathbf{P}}\mathcal{L}_{oo}^{v}\hat{\mathbf{P}}^{\ast}(\boldsymbol{\rho^{\ast}})^{-\frac{1}{2}}&\boldsymbol{\rho}^{-\frac{1}{2}}\hat{\mathbf{P}}\mathcal{L}_{oc}^{v}\\\
\mathcal{L}_{co}^{v}\hat{\mathbf{P}}^{\ast}(\boldsymbol{\rho^{\ast}})^{-\frac{1}{2}}&0\end{pmatrix}.$
(B.8)
Mind the suppressed superscript ’0’ for the density matrices as compared to,
e.g., Eq. (111). It is hence possible to diagonalize either
$\mathcal{L}^{(1)}$ or $\mathcal{L}^{(2)}$ with $A$ and $B$ defined in Eqs.
(B.7) and (B.8) to obtain the squared eigenvalues of the original LR matrix
$\mathcal{L}$. The advantage is, as mentioned above, that the dimensionality
of $\mathcal{L}^{(1)}$ and $\mathcal{L}^{(2)}$ is only half of the
dimensionality of $\mathcal{L}$. Moreover, for two distinct eigenvalues
$\omega_{1}$ and $\omega_{2}$ of $\mathcal{L}$ with
$\omega_{2}\geq\omega_{1}>0.5$ and separation
$\Delta\equiv\omega_{2}-\omega_{1}\geq 0$, the separation $\Delta^{(2)}$ in
the spectrum of the squared eigenvalues is given by
$\displaystyle\Delta^{(2)}$
$\displaystyle\equiv\omega_{2}^{2}-\omega_{1}^{2}=(\omega_{1}+\Delta)^{2}-\omega_{1}^{2}$
$\displaystyle=2\omega_{1}\Delta+\Delta^{2}=\Delta(2\omega_{1}+\Delta)$
$\displaystyle\geq 2\omega_{1}\Delta$ $\displaystyle>\Delta.$ (B.9)
Thus, the separation grows in the spectra of $\mathcal{L}^{(1)}$ and
$\mathcal{L}^{(2)}$. Since in general, the IRAM converges in less iterations
when the eigenvalues are more separated from each other, eigenvalues of larger
magnitude than $0.5$ might be obtained with less iterations by diagonalizing
either $\mathcal{L}^{(1)}$ or $\mathcal{L}^{(2)}$ instead of $\mathcal{L}$. On
the contrary, for eigenvalues with $0\leq\omega_{1}\leq\omega_{2}\leq 0.5$,
one finds that
$\displaystyle\Delta^{(2)}$ $\displaystyle=\Delta(2\omega_{1}+\Delta)$
$\displaystyle=\Delta\underbrace{(\omega_{2}+\omega_{1})}_{\leq 1}$
$\displaystyle\leq\Delta,$ (B.10)
which means that the separation of these eigenvalues is decreased in the
spectra of $\mathcal{L}^{(1)}$ and $\mathcal{L}^{(2)}$. Therefore, although
each iteration of the IRAM is less expensive for the smaller matrices, more
iterations might be necessary to obtain converged eigenvalues with smaller
magnitude than 0.5. A possible solution to this problem is to shift
$\mathcal{L}\rightarrow\mathcal{L}+0.5\cdot\mathbb{1}$, where $\mathbb{1}$ is
the $2(M+N_{\text{conf}})$-dimensional unit matrix, before applying the
transformation in Eq. (B.1).
It therefore remains a system-dependent question whether it is beneficial to
use the full LR matrix $\mathcal{L}$ or one of the smaller matrices
$\mathcal{L}^{(1)}$ and $\mathcal{L}^{(2)}$ to obtain the low-energy spectrum.
One needs to take into account the computational cost of building the latter
two matrices. Moreover, it is not in general true that, assuming the blocks
$A$ and $B$ are sparse matrices, their product is also sparse. If one is in
the unfortunate situation that the matrices $\mathcal{L}^{(1)}$ and
$\mathcal{L}^{(2)}$ are much denser than the original matrix $\mathcal{L}$, it
is potentially also much more demanding to converge to their lowest
eigenvalues, although they are much smaller. It is therefore a priori not
possible to decide which approach is more efficient, and in principle, one
would always need to first calculate the (shifted) matrices
$\mathcal{L}^{(1)}$ and $\mathcal{L}^{(2)}$ and analyze how dense they are
before making an appropriate decision. All applications of LR-MCTDHB discussed
in this work, however, avoided this preliminary analysis, and used the full LR
matrix $\mathcal{L}$ to obtain the low-energy spectra. For additional details,
it is referred to Ref. [241].
## Appendix C Further benchmarks of LR-MCTDHB
Further results on the benchmark of the newly developed implementation of LR-
MCTDHB against the HIM are presented in this Appendix. In particular,
comparisons of the exact and numerical results for attractive bosons in 1D as
well as for repulsive bosons in the anisotropic HIM in 2D are made. In
addition, the numerical convergence in the rotating frame of reference, which
is important for the validity of the results presented in Section 6.3.3, is
demonstrated.
| $M=1$ | $M=4$ | $M=6$ | $(m_{x},n_{x})$ | Exact
---|---|---|---|---|---
$\varepsilon^{0}$ | 9.137833 | 9.038151 | 9.038150 | $(0,0)$ | 9.038150
$\omega_{1}$ | 1.000000 | 1.000000 | 1.000000 | $(1,0)$ | 1.000000
$\omega_{2}$ | n/a | 2.000222 | 2.000000 | $(2,0)$ | 2.000000
$\omega_{3}$ | n/a | 3.000432 | 3.000000 | $(3,0)$ | 3.000000
$\omega_{4}$ | 3.655133 | 3.794752 | 3.794733 | $(0,2)$ | 3.794733
$\omega_{5}$ | n/a | 4.011839 | 4.000012 | $(4,0)$ | 4.000000
$\omega_{6}$ | n/a | 4.794870 | 4.794733 | $(1,2)$ | 4.794733
$\omega_{7}$ | n/a | 5.022646 | 5.000028 | $(5,0)$ | 5.000000
$\omega_{8}$ | 5.482700 | 5.692143 | 5.692100 | $(0,3)$ | 5.692100
$\omega_{9}$ | n/a | 5.797912 | 5.794737 | $(2,2)$ | 5.794733
$\omega_{10}$ | n/a | 6.142955 | 6.000676 | $(6,0)$ | 6.000000
Table C.1: Benchmark of LR-MCTDHB to the attractive 1D HIM. Shown are the
ground-state energy $\varepsilon^{0}$ and the energies $\omega_{i}$ of the
first ten excitations for $N=10$ bosons and different numbers of orbitals $M$.
The trapping frequency is $\Omega=1.0$ and the interaction strength is
$\lambda_{0}=0.13$, yielding a ground-state depletion of $f=0.94\%$. This is
comparable to the average depletion of the 1D lattice systems discussed in
Section 6.3.2. Underlined digits denote deviations from the exact values from
Eqs. (160) and (161). All quantities are dimensionless. See text for more
details.
For the case of attractive bosons in the 1D HIM, the system parameters are
chosen as $\Omega=1.0$ for the trapping frequency and $\lambda_{0}=0.13$ for
the two-body interaction strength. The number of bosons is $N=10$. The
calculations were carried out on a grid with 128 grid points in the interval
$[-9,9)$. Results for the excitation energies relative to the ground-state
energy, $\omega_{i}=E_{i}-\varepsilon^{0}$, are shown in Table C.1. With
respect to $\varepsilon^{0}$, one observes that $M=4$ and $M=6$ lead to highly
accurate results. With respect to the excited states, it can be observed that
the BdG approach, i.e., for $M=1$, misses several excitations in the low-
energy spectrum. In contrast to that, the many-body calculations yield all
states. The numerical accuracy clearly improves upon increasing the number of
self-consistent orbitals. For $M=6$, even the higher c.m. excited states
$\omega_{5},\,\omega_{7}$, and $\omega_{10}$ are obtained to very high
accuracy. Together with the benchmark from Table 2, it can be deduced that the
implementation of LR-MCTDHB is capable of accurately describing the low-energy
spectrum of both repulsive and attractive bosons.
| $M=1$ | $M=2$ | $M=3$ | $(m_{x},m_{y}$ $n_{x},n_{y})$ | Exact
---|---|---|---|---|---
$\varepsilon^{0}$ | 110.007587 | 110.004507 | 110.003452 | $(0,0,0,0)$ | 110.003452
$\omega_{1}$ | 1.000000 | 0.999999 | 1.000000 | $(1,0,0,0)$ | 1.000000
$\omega_{2}$ | 1.378405 | 1.378280 | 1.378405 | $(0,1,0,0)$ | 1.378405
$\omega_{3}$ | 1.791089 | 1.788859 | 1.788860 | $(0,0,2,0)$ | 1.788854
$\omega_{4}$ | n/a | 2.000438 | 2.000477 | $(2,0,0,0)$ | 2.000000
$\omega_{5}$ | 2.200152 | 2.200520 | 2.198271 | $(0,0,1,1)$ | 2.198268
$\omega_{6}$ | n/a | 1.940430 | 2.378704 | $(1,1,0,0)$ | 2.378405
$\omega_{7}$ | 2.609215 | 2.609214 | 2.607684 | $(0,0,0,2)$ | 2.607681
$\omega_{8}$ | 2.686634 | 2.683292 | 2.683294 | $(0,0,3,0)$ | 2.683282
$\omega_{9}$ | n/a | n/a | 2.757078 | $(0,2,0,0)$ | 2.756810
$\omega_{10}$ | n/a | 2.789191 | 2.789227 | $(1,0,2,0)$ | 2.788854
$\omega_{11}$ | n/a | 3.000758 | 3.000816 | $(3,0,0,0)$ | 3.000000
Table C.2: Benchmark of LR-MCTDHB to the anisotropic 2D HIM with $N=100$
bosons. The trapping frequencies are $\Omega_{x}=1.0$ and
$\Omega_{y}=\sqrt{1.9}$ in $x$\- and $y$-directions, and the two-body
interaction strength id $\lambda=-0.001$. All previously obtained degeneracies
in the isotropic case (see Table 3) are lifted. The accuracy of the energies
increases with the number of orbitals. Underlined digits denote deviations
from the exact values from Eqs. (160) and (161). All quantities are
dimensionless. See text for more details.
Table C.2 shows the numerical results obtained for the anisotropic HIM in 2D
with trap frequencies $\Omega_{x}=1.0$ and $\Omega_{y}=\sqrt{1.9}$. The
remaining system parameters are the same as in the isotropic case from Table
3. One first observes that the anisotropy lifts all previous degeneracies of
the isotropic case. Apart from that, the numerical results are qualitatively
similar. In the BdG case, only the first c.m. excitation is obtained
numerically exact, all other c.m. excited states are inaccessible. The pure
relative excitations are also found to a good accuracy. For $M=2$, one gets
access to higher c.m. excited states in the $x$-direction, together with
combined excitations of that type with excitations of relative coordinates. As
in the isotropic case, all excited states obtained in the $y$-coordinate
deteriorate in accuracy compared to the BdG case, see, e.g., $\omega_{2}$ and
$\omega_{5}$. The reason is the same as for the isotropic case, namely that
the additional orbital resembles a $p_{x}$-orbital and leads to preferred
direction in the description. Including a third orbital that resembles a
$p_{y}$-orbital solves this problem. It additionally improves the overall
accuracy of the low-energy excitations, and moreover ensures that all low-
lying excited states are obtained.
| $M=1$ | $M=3$ | $(n,m,l_{z})$ | Exact
---|---|---|---|---
$\varepsilon^{0}$ | 89.554453 | 89.548293 | (0,0,0) | 89.548292
$\omega_{1}$ | 0.900000 | 0.900000 | (0,1,1) | 0.900000
$\omega_{2}$ | 1.100000 | 1.100000 | (0,1,-1) | 1.100000
$\omega_{3}$ | 1.591089 | 1.588859 | (2,0,2) | 1.588854
$\omega_{4}$ | 1.791089 | 1.788862 | (2,0,0) | 1.788854
$\omega_{5}$ | n/a | 1.800439 | (0,2,2) | 1.800000
$\omega_{6}$ | 1.991089 | 1.988859 | (2,0,-2) | 1.988854
$\omega_{7}$ | n/a | 2.000657 | (0,2,0) | 2.000000
$\omega_{8}$ | n/a | 2.200438 | (0,2,-2) | 2.200000
$\omega_{9}$ | 2.386633 | 2.383226 | (3,0,3) | 2.383282
$\omega_{10}$ | n/a | 2.487088 | (2,1,3) | 2.488854
$\omega_{11}$ | 2.586634 | 2.583318 | (3,0,1) | 2.583282
$\omega_{12}$ | n/a | 2.689270 | (2,1,1) | 2.688854
$\omega_{13}$ | n/a | 2.691535 | (2,1,1) | 2.688854
$\omega_{14}$ | n/a | 2.700528 | (0,3,3) | 2.700000
$\omega_{15}$ | 2.786634 | 2.783318 | (3,0,-1) | 2.783282
Table C.3: Benchmark of LR-MCTDHB to the isotropic 2D HIM in the rotating
frame. The angular velocity around the $z$-axis is $\Omega_{\text{rot}}=0.1$.
Results are presented for $N=100$ bosons and different numbers of orbitals
$M$. The trapping frequencies are $\Omega_{x}=\Omega_{y}=1.0$, whereas the
strength of the repulsive interaction is $\lambda_{0}=-0.001$. Shown are the
energies of the ground state, $\varepsilon^{0}$, and of the first $15$ excited
states, $\omega_{i}=E_{i}-\varepsilon^{0}$. Underlined digits denote
deviations from the exact values from Eqs. (160), (161) and (C.1). The quantum
numbers $n$ and $m$ refer to the relative and c.m. coordinates, whereas
$l_{z}$ refers to the angular momentum in the $z$-direction. All quantities
are dimensionless. See text for more details.
Finally, excitations in the rotating frame of reference are discussed. The
energies of the ground and excited states, given by Eqs. (160) and (161),
become shifted by the term
$E_{\text{rot}}=-\Omega_{\text{rot}}\,l_{z}$ (C.1)
where $\Omega_{\text{rot}}$ denotes the angular velocity around the $z$-axis
and $l_{z}$ denotes the angular momentum of the exited state in the
$z$-direction. In Table 3, all obtained excitations are labeled by
corresponding quantum numbers for relative and c.m. excitations, $n$ and $m$,
as well as by $l_{z}$. The increased numerical accuracy of LR-MCTDHB($3$) in
comparison to the BdG results becomes clearly obvious. Moreover, all states
are obtained for $M=3$, whereas the BdG spectrum misses several states. As a
conclusion, the utilized implementation of LR-MCTDHB is also capable of
accurately describing the low-energy excitation spectra in the rotating frame.
## Appendix D Many-particle variance: A sensitive quantity to many-body
effects
In this Appendix, a brief introduction of the many-body variance, which is
employed in the main text to study many-body excitations involved in the
dynamics of trapped BECs, is given. The many-body variance has been of
interest in recent studies on correlations between bosons in the ground state
[242] and for out-of-equilibrium BECs [243, 244], in particular in a 1D
bosonic Josephson junction [245] or in anharmonic and anisotropic trapping
potentials [246, 81]. It has also been found that the variance is a sensitive
measure for the numerical convergence of MCTDHB [80]. The focus is laid on the
position variance in the $x$-direction in the following.
The position operator of $N$ bosons is given by
$\hat{X}=\sum_{i=1}^{N}\hat{x}_{i}$ (D.1)
where the operator $\hat{x}_{i}$ denotes the position operator of the $i$-th
particle. To compute its variance per particle in the state
$\Psi(\mathbf{r}_{1},...,\mathbf{r}_{N})$, which reads
$\frac{1}{N}\Delta^{2}_{\hat{X}}=\frac{1}{N}\left(\langle\Psi|\hat{X}^{2}|\Psi\rangle-\langle\Psi|\hat{X}|\Psi\rangle^{2}\right),$
(D.2)
one also needs to evaluate the expectation value of the square of the position
operator. The latter is given by
$\hat{X}^{2}=\sum_{i=1}^{N}\hat{x}_{i}^{2}+\sum_{i<k}2\,\hat{x}_{i}\hat{x}_{k},$
(D.3)
i.e., it consists of a one-body and a two-body term. For the expectation value
of $\hat{X}$ one obtains
$\displaystyle\frac{1}{N}\langle\Psi|\hat{X}|\Psi\rangle$
$\displaystyle=\frac{1}{N}\sum_{i=1}^{N}\int
d\mathbf{r}_{1}...d\mathbf{r}_{N}\,\Psi^{\ast}(\mathbf{r}_{1},...,\mathbf{r}_{N})\,x_{i}\,\Psi(\mathbf{r}_{1},...,\mathbf{r}_{N})$
$\displaystyle=\frac{1}{N}\sum_{i=1}^{N}\int
d\mathbf{r}_{i}\,\frac{\rho^{(1)}(\mathbf{r}_{i}|\mathbf{r}_{i})}{N}x_{i}$
$\displaystyle=\int d\mathbf{r}\,\frac{\rho(\mathbf{r})}{N}x.$ (D.4)
In the second step, the definition of the density
$\rho(\mathbf{r})\equiv\rho^{(1)}(\mathbf{r}|\mathbf{r})$, employing the
diagonal of the one-body RDM, was used. In general, the time-dependent $p$-th
order RDM of a given $N$-boson system in the state
$\Psi(\mathbf{r}_{1},...,\mathbf{r}_{N};t)$ is defined as
$\displaystyle\rho^{(p)}(\mathbf{r}_{1},...,\mathbf{r}_{p}|\mathbf{r}_{1}^{\prime},...,\mathbf{r}_{p}^{\prime};t)$
$\displaystyle=\frac{N!}{(N-p)!}\int
d\mathbf{r}_{p+1}...d\mathbf{r}_{N}\,\Psi(\mathbf{r}_{1},...,\mathbf{r}_{p},\mathbf{r}_{p+1},...,\mathbf{r}_{N};t)$
$\displaystyle\times\Psi^{\ast}(\mathbf{r}_{1}^{\prime},...,\mathbf{r}_{p}^{\prime},\mathbf{r}_{p+1},...,\mathbf{r}_{N};t).$
(D.5)
From Eq. (D), one can define the $p$-th order correlation function $g^{(p)}$
as
$\displaystyle
g^{(p)}(\mathbf{r}_{1}^{\prime},...,\mathbf{r}_{p}^{\prime}|\mathbf{r}_{1},...,\mathbf{r}_{p};t)=\frac{\rho^{(p)}(\mathbf{r}_{1}^{\prime},...,\mathbf{r}_{p}^{\prime}|\mathbf{r}_{1},...,\mathbf{r}_{p};t)}{\sqrt{\Pi_{i=1}^{p}\,\rho^{(1)}(\mathbf{r}_{i}|\mathbf{r}_{i};t)\,\rho^{(1)}(\mathbf{r}_{i}^{\prime}|\mathbf{r}_{i}^{\prime};t)}}.$
(D.6)
Whereas the wave function $\Psi$ is assumed to be normalized to unity, i.e.,
$\langle\Psi(t)|\Psi(t)\rangle=1$, the $p$-th order RDM is not.
One further obtains
$\displaystyle\frac{1}{N}\langle\Psi|\hat{X}^{2}|\Psi\rangle$
$\displaystyle=\int d\mathbf{r}\,\frac{\rho(\mathbf{r})}{N}x^{2}$
$\displaystyle+\frac{2}{N}\sum_{i<k}\int
d\mathbf{r}_{1}...d\mathbf{r}_{N}\,\Psi^{\ast}(\mathbf{r}_{1},...,\mathbf{r}_{N})\,x_{i}x_{k}\,\Psi(\mathbf{r}_{1},...,\mathbf{r}_{N})$
(D.7)
where the second term can be simplified due to
$\displaystyle\sum_{i<k}\int
d\mathbf{r}_{1}...d\mathbf{r}_{N}\,\Psi^{\ast}(\mathbf{r}_{1},...,\mathbf{r}_{N})\,x_{i}x_{k}\,\Psi(\mathbf{r}_{1},...,\mathbf{r}_{N})$
$\displaystyle=\sum_{i<k}\int
d\mathbf{r}_{i}d\mathbf{r}_{k}\,\frac{\rho^{(2)}(\mathbf{r}_{i},\mathbf{r}_{k},\mathbf{r}_{i},\mathbf{r}_{k})}{N(N-1)}x_{i}x_{k}$
$\displaystyle=\frac{N(N-1)}{2}\int
d\mathbf{r}_{1}d\mathbf{r}_{2}\,\frac{\rho^{(2)}(\mathbf{r}_{1},\mathbf{r}_{2},\mathbf{r}_{1},\mathbf{r}_{2})}{N(N-1)}x_{1}x_{2}$
(D.8)
where the definition of the two-body RDM $\rho^{(2)}$ was utilized. The choice
of the coordinates of the first and second particle in the last step is
arbitrary. This yields
$\displaystyle\frac{1}{N}\langle\Psi|\hat{X}^{2}|\Psi\rangle$
$\displaystyle=\int d\mathbf{r}\,\frac{\rho(\mathbf{r})}{N}x^{2}$
$\displaystyle+\int
d\mathbf{r}_{1}d\mathbf{r}_{2}\,\frac{\rho^{(2)}(\mathbf{r}_{1},\mathbf{r}_{2},\mathbf{r}_{1},\mathbf{r}_{2})}{N}x_{1}x_{2}$
(D.9)
such that one obtains
$\displaystyle\frac{1}{N}\Delta^{2}_{\hat{X}}$ $\displaystyle=\int
d\mathbf{r}\,\frac{\rho(\mathbf{r})}{N}x^{2}-\frac{1}{N}\left[\int
d\mathbf{r}\,\rho(\mathbf{r})\,x\right]^{2}$ $\displaystyle+\int
d\mathbf{r}_{1}d\mathbf{r}_{2}\,\frac{\rho^{(2)}(\mathbf{r}_{1},\mathbf{r}_{2},\mathbf{r}_{1},\mathbf{r}_{2})}{N}x_{1}x_{2}$
(D.10)
for the variance of the position operator $\hat{X}$.
To arrive at an expression in the Hartree limit, we first expand $\rho^{(2)}$
by the natural orbitals $\\{\alpha_{i}\\}$, which yields
$\rho^{(2)}(\mathbf{r}_{1},\mathbf{r}_{2},\mathbf{r}_{1},\mathbf{r}_{2})=\sum_{i,j,k,l}\,\rho_{ijkl}\,\alpha^{\ast}_{i}(\mathbf{r}_{1})\alpha^{\ast}_{j}(\mathbf{r}_{2})\alpha_{k}(\mathbf{r}_{1})\alpha_{l}(\mathbf{r}_{2}).$
(D.11)
Then, we can perform the limit which gives
$\displaystyle\lim_{N\rightarrow\infty}\frac{1}{N}\Delta_{\hat{X}}^{2}$
$\displaystyle=\Delta_{\text{GP}}^{2}+\Delta_{\text{correlations}}^{2},$
(D.12) $\displaystyle\Delta_{\text{GP}}^{2}$ $\displaystyle=\int
d\mathbf{r}\,|\phi^{0}(\mathbf{r})|^{2}\,x^{2}-\left[\int
d\mathbf{r}\,|\phi^{0}(\mathbf{r})|^{2}\,x\right]^{2}$ (D.13)
$\displaystyle\Delta_{\text{correlations}}^{2}$
$\displaystyle=\lim_{N\rightarrow\infty}\sum_{i,j,k,l\neq 1111}\int
d\mathbf{r}_{1}d\mathbf{r}_{2}\,\frac{\rho_{ijkl}}{N}\alpha^{\ast}_{i}(\mathbf{r}_{1})\alpha^{\ast}_{j}(\mathbf{r}_{2})\,x_{1}x_{2}\,\alpha_{k}(\mathbf{r}_{1})\alpha_{l}(\mathbf{r}_{2}).$
(D.14)
Eq. (D.13) expresses the fact that, in the Hartree limit, the GP orbital
$\phi^{0}$, obtained from Eq. (A.5), exactly describes the ground-state
density per particle, i.e.,
$\lim_{N\rightarrow\infty}\frac{\rho(\mathbf{r})}{N}=|\phi^{0}(\mathbf{r})|^{2}$.
To remind the reader, only one natural orbital exists in the GP case. In
obtaining Eq. (D.13) one makes use of the relation
$\lim_{N\rightarrow\infty}\frac{\rho_{1111}}{N(N-1)}=1$ as well. This means
that many-body contributions to the variance per particle, Eq. (D.14), can
only occur due to the occupation of higher natural orbitals, as is expressed
in the excluded summation $i,j,k,l\neq 1111$ in Eq. (D.14). It has been shown
that even a small occupation of the higher natural orbitals, i.e., the number
of depleted particles, can have a sizable effect at the infinite particle
limit. For further details, also including the variance of other quantities
like the momentum and angular momentum operators, see Refs. [242, 139].
Applications to exact many-body excitations from the dynamics of the variance
were discussed in section 6.3. Here, we show an example that demonstrates the
numerical convergence of the ground states of the rotating BECs of Fig. 28
with $M=7$ orbitals, the position variance in $x$-direction depending on
different values of the vorticity $l$ is shown in Fig. 32. One readily
observes that already for $M=5$ orbitals the ground state variances are
essentially condensed, and the deviations are less than $\mathcal{O}(10^{-3})$
of the exact values (see inset).
Figure 32: The position variance per particle in $x$-direction for the ground
states of the rotating BECs treated in Fig. 28. For all vorticities $l$ shown,
essentially $M=5$ self-consistent orbitals are sufficient to reach numerical
convergence because the results lie atop the ones for higher values of $M$.
Inset: Enlarged view for the case of $l=4$. All quantities are dimensionless.
See text for details. The figure is taken from Ref. [187].
## Acknowledgement
We thank Alexej I. Streltsov and Shachar Klaiman for many discussions.
Computation time on the Cray XC40 cluster Hazel Hen at the High Performance
Computing Center Stuttgart (HLRS) and the BwForCluster is acknowledged. RB
acknowledges financial support by the IMPRS-QD (International Max Plack
Research School for Quantum Dynamics), the Landesgraduiertenförderung Baden-
Württemberg, and the Minerva Foundation. OEA acknowledges funding by the
Israel Science Foundation (Grant No. 600/15).
## Bibliography
* [1] S. N. Bose, Plancks Gesetz und Lichtquantenhypothese., Z. Phys. 26, 178 (1924).
* [2] A. Einstein, Quantentheorie des einatomigen idealen Gases., Sitzungsber. Preuss. Akad. Wiss., Bericht 22, 261 (1924).
* [3] A. Einstein, Quantentheorie des einatomigen idealen Gases. Zweite Abhandlung., Sitzungsber. Preuss. Akad. Wiss., Bericht 1, 3 (1925).
* [4] M. H. Anderson, J. R. Ensher, M. R. Matthews, C. E. Wieman, and E. Cornell, Observation of Bose-Einstein condensation in a dilute atomic vapor, Science 269, 198 (1995).
* [5] K. B. Davis, M.-O. Mewes, M. R. Andrews, N. J. van Druten, D. S. Durfee, D. M. Kurn, and W. Ketterle, Bose-Einstein condensation in a gas of sodium atoms, Phys. Rev. Lett. 75, 3969 (1995).
* [6] C. C. Bradley, C. A. Sackett, J. J. Tollett, and R. G. Hulet, Evidence of Bose-Einstein Condensation in an Atomic Gas with Attractive Interactions, Phys. Rev. Lett. 75, 1687 (1995).
* [7] J. Kasprzak, M. Richard, S. Kundermann, A. Baas, P. Jeambrun, J. M. J. Keeling, F. M. Marchetti, M. H. Szymańska, R. André, J. L. Staehli, V. Savona, P. B. Littlewood, B. Deveaud, and L. S. Dang, Bose-Einstein condensation of exciton polaritons, Nature 443, 409 (2006).
* [8] H. Deng, H. Haug, and Y. Yamamoto, Exciton-polariton Bose-Einstein condensation, Rev. Mod. Phys. 82, 1489 (2010).
* [9] S. O. Demokritov, V. E. Demidov, O. Dzyapko, G. A. Melkov, A. A. Serga, B. Hillebrands, and A. N. Slavin, Bose-Einstein condensation of quasi-equilibrium magnons at room temperature under pumping, Nature volume 443, 430 (2006).
* [10] J. Klaers, J. Schmitt, F. Vewinger, and M. Weitz, Bose-Einstein condensation of photons in an optical microcavity, Nature 468, 545 (2010).
* [11] P. Courteille, R. S. Freeland, D. J. Heinzen, F. A. van Abeelen, and B. J. Verhaar, Observation of a Feshbach Resonance in Cold Atom Scattering, Phys. Rev. Lett. 81, 69 (1998).
* [12] S. Inouye, M. R. Andrews, J. Stenger, H.-J. Miesner, D. M. Stamper-Kurn, and W. Ketterle, Observation of Feshbach resonances in a Bose-Einstein condensate, Nature 392, 151 (1998).
* [13] C. Chin , R. Grimm, P. Julienne , and E. Tiesinga, Feshbach resonances in ultracold gases, Rev. Mod. Phys. 82, 1225 (2010).
* [14] A. Görlitz, J. M. Vogels, A. E. Leanhardt, C. Raman, T. L. Gustavson, J. R. Abo-Shaeer, A. P. Chikkatur, S. Gupta, S. Inouye, T. Rosenband, and W. Ketterle, Realization of Bose-Einstein Condensates in Lower Dimensions, Phys. Rev. Lett. 87, 130402 (2001).
* [15] O. Lahav, A. Itah, A. Blumkin, C. Gordon, S. Rinott, A. Zayats, and J. Steinhauer, Realization of a Sonic Black Hole Analog in a Bose-Einstein Condensate, Phys. Rev. Lett. 105, 240401 (2010).
* [16] J. Steinhauer, Observation of self-amplifying Hawking radiation in an analogue black-hole laser, Nat. Phys. 10, 864 (2014).
* [17] J. Steinhauer, Observation of quantum Hawking radiation and its entanglement in an analogue black hole, Nat. Phys. 12, 959 (2016).
* [18] E. P. Gross, Structure of a quantized vortex in boson systems, Nuovo Cimento 20, 454 (1961).
* [19] L. P. Pitaevskii, Vortex lines in an imperfect Bose gas, Zh. Eksp. Teor. Fiz. 40, 646 (1961) [Sov. Phys. JETP 13, 451 (1961)].
* [20] L. P. Pitaevskii and S. Stringari, Bose-Einstein Condensation, Oxford University Press, New York, 2003.
* [21] C. J. Pethick and H. Smith, Bose-Einstein Condensation in Dilute Gases, 2nd edition, Cambridge University Press, New York, 2008.
* [22] A. J. Leggett, Bose-Einstein condensation in the alkali gases: Some fundamental concepts, Rev. Mod. Phys. 73, 307 (2003).
* [23] N. N. Bogoliubov, On the theory of superfluidity, J. Phys. 11 (USSR), 23 (1947).
* [24] P.-G. de Gennes, Superconductivity of metals and alloys, Benjamin, New York, 1966.
* [25] G. C. Katsimiga, P. G. Kevrekidis, B. Prinari, G. Biondini, and P. Schmelcher, Dark-bright soliton pairs: Bifurcations and collisions, Phys. Rev A 97, 043623 (2018).
* [26] L. Jia, A.-B. Wang, and S. Yi, Low-lying excitations of vortex lattices in condensates with anisotropic dipole-dipole interaction, Phys. Rev. A 97, 043614 (2018).
* [27] B. Galilo, D. K. K. Lee, and R. Barnett, Topological Edge-State Manifestation of Interacting 2D Condensed Boson-Lattice Systems in a Harmonic Trap, Phys. Rev. Lett. 119, 203204 (2017).
* [28] J.-X. Zhu, Bogoliubov-de Gennes Method and Its Applications, Springer, Heidelberg, 2016.
* [29] A. Spuntarelli, P. Pieri, G. C. Strinati, Solution of the Bogoliubov-de Gennes equations at zero temperature throughout the BCS-BEC crossover: Josephson and related effects, Phys. Rep. 488, 111 (2010).
* [30] F. Dalfovo, S. Giorgini, L. P. Pitaevskii, and S. Stringari, Theory of Bose-Einstein condensation in trapped gases, Rev. Mod. Phys. 71, 463 (1999).
* [31] O. E. Alon, A. I. Streltsov, and L. S. Cederbaum, Unified view on multiconfigurational time-propagation for systems consisting of identical particles, J. Chem. Phys. 127, 154103 (2007).
* [32] O. E. Alon, A. I. Streltsov, and L. S. Cederbaum, Multiconfigurational time-dependent Hartree method for bosons: Many-body dynamics of bosonic systems, Phys. Rev. A 77, 033613 (2008).
* [33] A. U. J. Lode, K. Sakmann, O. E. Alon, L. S. Cederbaum, and A. I. Streltsov, Numerically exact quantum dynamics of bosons with time-dependent interactions of harmonic type, Phys. Rev. A 86, 063606 (2012).
* [34] J. Grond, A. I. Streltsov, A. U. J. Lode, K. Sakmann, L. S. Cederbaum, and O. E. Alon, Excitation spectra of many-body systems by linear response: General theory and applications to trapped condensates, Phys. Rev. A 88, 023606 (2013).
* [35] O. E. Alon, A. I. Streltsov, and L. S. Cederbaum, Unified view on linear-response of interacting identical and distinguishable particles from multiconfigurational time-dependent Hartree methods, J. Chem. Phys 140, 034108 (2014).
* [36] Markus Fierz, Über die relativistische Theorie kräftefreier Teilchen mit beliebigem Spin, Helvetica Physica Acta 12, 3 (1939).
* [37] Wolfgang Pauli, The Connection Between Spin and Statistics, Phys. Rev. 58, 716 (1940).
* [38] P. Kramer and M. Sarenco, Geometry of the Time-Dependent Variational Principle, Springer, Heidelberg, 1981.
* [39] P. A. M. Dirac, Note on Exchange Phenomena in the Thomas Atom, Proc. Cambridge Philos. Soc. 26, 376 (1930).
* [40] J. Frenkel, Wave Mechanics: advanced general theory, Oxford, Clarendon Press (1934).
* [41] A. D. McLachlan, A variational solution of the time-dependent Schrödinger equation, Mol. Phys. 8, 39 (1964).
* [42] J. Kucar, H.-D. Meyer, and L. S. Cederbaum, Time-dependent rotated Hartree approach, Chem. Phys. Lett. 140, 525 (1987).
* [43] O. Penrose and L. Onsager, Bose-Einstein Condensation and Liquid Helium, Phys. Rev. 104, 576 (1956).
* [44] P. Nozières, D. Saint James, Particle vs. pair condensation in attractive Bose liquids, J. Phys. (France) 43, 1133 (1982).
* [45] A. J. Leggett, Quantum Liquids: Bose condensation and Cooper pairing in condensed matter systems, Oxford University Press, Oxford (2006).
* [46] K. Sakmann, Numerically exact dynamics of the interacting many-body Schrödinger equation for Bose-Einstein condensates: comparison to Bose-Hubbard and Gross-Pitaevskii theory, PhD thesis, Heidelberg University, 2010.
* [47] H.-D. Meyer, U. Manthe, and L. S. Cederbaum, The Multi-Configurational Time-Dependent Hartree Approach, Chem. Phys. Lett. 165, 73 (1990).
* [48] U. Manthe, H.-D. Meyer, and L. S. Cederbaum, Wave-packet dynamics within the multiconfiguration Hartree framework: General aspects and application to NOCl, J. Chem. Phys. 97, 3199 (1992).
* [49] H.–D. Meyer, F. Gatti, and G. A. Worth, Multidimensional Quantum Dynamics: MCTDH Theory and Applications, Wiley-VCH, 2009.
* [50] O. E. Alon, A. I. Streltsov, and L. S. Cederbaum, Many-body theory for systems with particle conversion: Extending the multiconfigurational time-dependent Hartree method, Phys. Rev. A 79, 022503 (2009).
* [51] O. E. Alon, A. I. Streltsov, and L. S. Cederbaum, Multiconfigurational time-dependent Hartree method for mixtures consisting of two types of identical particles, Phys. Rev. A. 76, 062501 (2007).
* [52] A. U. J. Lode and C. Bruder, Dynamics of Hubbard Hamiltonians with the multiconfigurational time-dependent Hartree method for indistinguishable particles, Phys. Rev. A 94, 013616 (2016).
* [53] A. U. J. Lode, Multiconfigurational time-dependent Hartree method for bosons with internal degrees of freedom: Theory and composite fragmentation of multicomponent Bose-Einstein condensates, Phys. Rev. A 93, 063601 (2016).
* [54] S. R. White, Density matrix formulation for quantum renormalization groups, Phys. Rev. Lett. 69, 2363 (1992).
* [55] K. A. Hallberg, New trends in density matrix renormalization, Adv. Phys. 55, 477 (2006).
* [56] E. M. Stoudenmire and S. R. White, Studying Two-Dimensional Systems with the Density Matrix Renormalization Group, Annu. Rev. Cond. Matt. Phys. 3, 111 (2012).
* [57] A. I. Streltsov, O. E. Alon, L. S. Cederbaum, General variational many-body theory with complete self-consistency for trapped bosonic systems, Phys. Rev. A 73, 063626 (2006).
* [58] Y. Saad, Iterative Methods for Sparse Linear Systems, 2nd edition, Society for Industrial and Applied Mathematics, Philadelphia, 2003.
* [59] http://www.caam.rice.edu/software/ARPACK/.
* [60] http://www.caam.rice.edu/~kristyn/parpack_home.html.
* [61] http://www.netlib.org/scalapack/.
* [62] https://software.intel.com/en-us/mkl.
* [63] Steffen Börm and Christian Mehl, Numerical Methods for Eigenvalue Problems, De Gruyter, Berlin/Boston, 2012.
* [64] Y. Saad, Numerical Methods for Large Eigenvalue Problems, Halstead Press, New York, 1992.
* [65] A. I. Streltsov, L. S. Cederbaum, O. E. Alon, K. Sakmann, A. U. J. Lode, J. Grond, O. I. Streltsova, S. Klaiman, and R. Beinke, The Multiconfigurational Time-Dependent Hartree for Bosons Package, version 3.x, http://mctdhb.org, Heidelberg/Kassel (2006-present).
* [66] A. I. Streltsov, O. E. Alon, and L. S. Cederbaum, General mapping for bosonic and fermionic operators in Fock space, Phys. Rev. A 81, 022124 (2010).
* [67] A. I. Streltsov and O. I. Streltsova, MCTDHB-Lab, version 1.5, 2015, http://www.mctdhb-lab.com.
* [68] http://www.ultracold.org.
* [69] S. Krönke, L. Cao, O. Vendrell, and P. Schmelcher, Non-equilibrium quantum dynamics of ultra-cold bosonic mixtures: the multi-layermulti-configuration time-dependent Hartree method for bosons, New J. Phys. 15, 063018 (2013).
* [70] L. Cao, S. Krönke, O. Vendrell, and P. Schmelcher, The multi-layer multi-configuration time-dependent Hartree method for bosons: Theory, implementation, and applications, J. Chem. Phys. 139, 134103 (2013).
* [71] L. Cao, V. Bolsinger, S. I. Mistakidis, G. M. Koutentakis, S. Krönke, J. M. Schurer, and P. Schmelcher, A unified ab initio approach to the correlated quantum dynamics of ultracold fermionic and bosonic mixtures, J. Chem. Phys. 147, 044106 (2017).
* [72] L. Cohen and C. Lee, Exact reduced density matrices for a model problem, J. Math. Phys. 26, 3105 (1985).
* [73] J. Yan, Harmonic Interaction Model and Its Applications in Bose-Einstein Condensation, J. Stat. Phys. 113, 314 (2003).
* [74] A. I. Streltsov, O. E. Alon, and L. S. Cederbaum, Role of Excited States in the Splitting of a Trapped Interacting Bose-Einstein Condensate by a Time-Dependent Barrier, Phys. Rev. Lett. 99, 030402 (2007).
* [75] K. Sakmann, A. I. Streltsov, O. E. Alon, and L. S. Cederbaum, Reduced density matrices and coherence of trapped interacting bosons, Phys. Rev. A 78, 023615 (2008).
* [76] S. Klaiman, A. U. J. Lode, A. I. Streltsov, L. S. Cederbaum, and O. E. Alon, Breaking the resilience of a two-dimensional Bose-Einstein condensate to fragmentation, Phys. Rev. A 90, 043620 (2014).
* [77] O. E. Alon, A. I. Streltsov, and L. S. Cederbaum, Build-up of coherence between initially-independent subsystems: The case of Bose-Einstein condensates, Phys. Lett. A 373, 9 (2009).
* [78] A. I. Streltsov, O. E. Alon, and L. S. Cederbaum, Efficient generation and properties of mesoscopic quantum superposition states in an attractive Bose-Einstein condensate threaded by a potential barrier, J. Phys. B: At. Mol. Opt. Phys. 42, 091004 (2009).
* [79] A. I. Streltsov, O. E. Alon, and L. S. Cederbaum, Formation and Dynamics of Many-Boson Fragmented States in One-Dimensional Attractive Ultracold Gases, Phys. Rev. Lett. 100, 130401 (2008).
* [80] J. G. Cosme, C. Weiss, and J. Brand, Center-of-mass motion as a sensitive convergence test for variational multimode quantum dynamics, Phys. Rev. A 94, 043603 (2016).
* [81] O. E. Alon and L. S. Cederbaum, Attractive Bose-Einstein condensates in anharmonic traps: Accurate numerical treatment and the intriguing physics of the variance, Chem. Phys. 515, 287 (2018).
* [82] C. Weiss and Y. Castin, Creation and Detection of a Mesoscopic Gas in a Nonlocal Quantum Superposition, Phys Rev. Lett. 102, 010403 (2009).
* [83] A. I. Streltsov, O. E. Alon, and L. S. Cederbaum, Scattering of an attractive Bose-Einstein condensate from a barrier: Formation of quantum superposition states, Phys. Rev. A 80, 043616 (2009).
* [84] A. I. Streltsov, O. E. Alon, and L. S. Cederbaum, Swift Loss of Coherence of Soliton Trains in Attractive Bose-Einstein Condensates, Phys. Rev. Lett. 106, 240401 (2011).
* [85] J. Grond, J. Schmiedmayer, and U. Hohenester, Optimizing number squeezing when splitting a mesoscopic condensate, Phys. Rev. A 79, 021603(R) (2009).
* [86] J. Grond, G. von Winckel, J. Schmiedmayer, and U. Hohenester, Optimal control of number squeezing in trapped Bose-Einstein condensates, Phys. Rev. A 80, 053625 (2009).
* [87] J. Grond, J. Schmiedmayer, and U. Hohenester, Shaking the condensates: Optimal number squeezing in the dynamic splitting of a Bose-Einstein condensate, Physica E 42, 432 (2010).
* [88] G. Jäger, T. Berrada, J. Schmiedmayer, T. Schumm, and U. Hohenester, Parametric-squeezing amplification of Bose-Einstein condensates, Phys. Rev. A 92, 053632 (2015).
* [89] J. Grond, U. Hohenester, I. Mazets, and J. Schmiedmayer, Atom interferometry with trapped Bose-Einstein condensates: impact of atom-atom interactions, New J. Phys. 12, 065036 (2010).
* [90] J. Grond, T. Betz, U. Hohenester, N. J. Mauser, J. Schmiedmayer, and T. Schumm, The Shapiro effect in atomchip-based bosonic Josephson junctions, New J. Phys. 13 065026 (2011).
* [91] I. Brouzos, A. I. Streltsov, A. Negretti, R. S. Said, T. Caneva, S. Montangero, and T. Calarco, Quantum speed limit and optimal control of many-boson dynamics, Phys. Rev. A 92, 062110 (2015).
* [92] J. Grond, U. Hohenester, J. Schmiedmayer, and A. Smerzi, Mach-Zehnder interferometry with interacting trapped Bose-Einstein condensates, Phys. Rev. A 84, 023619 (2011).
* [93] I. Březinová, A. U. J. Lode, A. I. Streltsov, O. E. Alon, L. S. Cederbaum, and J. Burgdörfer, Wave chaos as signature for depletion of a Bose-Einstein condensate, Phys. Rev. A 86, 013630 (2012).
* [94] I. Březinová, J. Burgdörfer, A. U. J. Lode, A. I. Streltsov, L. S. Cederbaum, O. E. Alon, L. A. Collins, and B. I. Schneider, Elastic scattering of a Bose-Einstein condensate at a potential landscape, J. Phys.: Conf. Ser. 488, 012032 (2014).
* [95] K. Sakmann and M. Kasevich, Single-shot simulations of dynamic quantum many-body systems, Nat. Phys. 12, 451 (2016).
* [96] R. Schmitz, S. Krönke, L. Cao, and P. Schmelcher, Quantum breathing dynamics of ultracold bosons in one-dimensional harmonic traps: Unraveling the pathway from few- to many-body systems, Phys. Rev. A 88, 043601 (2013).
* [97] S. Krönke and P. Schmelcher, Many-body processes in black and gray matter-wave solitons, Phys. Rev. A 91, 053614 (2015).
* [98] G. C. Katsimiga, S. I. Mistakidis, G. M. Koutentakis, P. G. Kevrekidis, and P. Schmelcher, Many-body quantum dynamics in the decay of bent dark solitons of Bose-Einstein condensates, New J. Phys. 19, 123012 (2017).
* [99] S. Krönke, J. Knörzer, and P. Schmelcher, Correlated quantum dynamics of a single atom collisionally coupled to an ultracold finite bosonic ensemble, New J. Phys. 17, 053001 (2015).
* [100] J. M. Schurer, A. Negretti, and P. Schmelcher, Unraveling the Structure of Ultracold Mesoscopic Collinear Molecular Ions, Phys. Rev. Lett. 119, 063001 (2017).
* [101] Q.-S. Tan, H.-Y. Lu, and S. Yi, Spin squeezing of a dipolar Bose gas in a double-well potential, Phys. Rev. A 93, 013606 (2016).
* [102] B. Chatterjee and A. U. J. Lode, Order parameter and detection for a finite ensemble of crystallized one-dimensional dipolar bosons in optical lattices, Phys. Rev A 98, 053624 (2018).
* [103] B. Chatterjee, M. C. Tsatsos, and A. U. J. Lode, Correlations of strongly interacting one-dimensional ultracold dipolar few-boson system in optical lattices, arXiv:1806.05048 [cond-mat.quant-gas] (2018), accepted at New J. Phys.
* [104] A. U. J. Lode and C. Bruder, Fragmented Superradiance of a Bose-Einstein Condensate in an Optical Cavity, Phys. Rev. Lett. 118, 013603 (2017).
* [105] P. Molignini, L. Papariello, A. U. J. Lode, and R. Chitra, Superlattice switching from parametric instabilities in a driven-dissipative Bose-Einstein condensate in a cavity, Phys. Rev A 98, 053620 (2018).
* [106] A. U. J. Lode, F. S. Diorico, R. Wu, P. Molignini, L. Papariello, R. Lin, C. Lévêque, L. Exl, M. C. Tsatsos, R. Chitra, and N. J. Mauser, Many-body physics in two-component Bose-Einstein condensates in a cavity: fragmented superradiance and polarization, New J. Phys. 20, 055006 (2018).
* [107] M. Albiez, R. Gati, J. Fölling, S. Hunsmann, M. Cristiani, and M. K. Oberthaler, Direct Observation of Tunneling and Nonlinear Self-Trapping in a Single Bosonic Josephson Junction, Phys. Rev. Lett. 95, 010402 (2005).
* [108] S. Levy, E. Lahoud, I. Shomroni, and J. Steinhauer, The a.c. and d.c. Josephson effects in a Bose-Einstein condensate, Nature (London) 449, 579 (2007).
* [109] G. J. Milburn J. Corney, E. M. Wright, and D. F. Walls, Quantum dynamics of an atomic Bose-Einstein condensate in a double-well potential, Phys. Rev. A 55, 4318 (1997).
* [110] A. Smerzi, S. Fantoni, S. Giovanazzi, and S. R. Shenoy, Quantum Coherent Atomic Tunneling between Two Trapped Bose-Einstein Condensates, Phys. Rev. Lett. 79 4950 (1997).
* [111] S. Raghavan, A. Smerzi, S. Fantoni, and S. R. Shenoy, Coherent oscillations between two weakly coupled Bose-Einstein condensates: Josephson effects, $\pi$ oscillations, and macroscopic quantum self-trapping, Phys. Rev. A 59, 620 (1999).
* [112] E. A. Ostrovskaya, Y. S. Kivshar, M. Lisak, B. Hall, F. Cattani, and D. Anderson, Coupled-mode theory for Bose-Einstein condensates, Phys. Rev. A 61, 031601(R).
* [113] D. Ananikian and T. Bergeman, Gross-Pitaevskii equation for Bose particles in a double-well potential: Two-mode models and beyond, Phys. Rev. A 73, 013604 (2006).
* [114] X. Y. Jia, W. D. Li, and J. Q. Liang, Nonlinear correction to the boson Josephson-junction model, Phys. Rev. A 78, 023613 (2008).
* [115] Y. Zhou, H. Zhai, R. Lü, Z. Xu, and L. Chang, Quantum coherence of double-well Bose-Einstein condensates: An SU(2) coherent-state path-integral approach, Phys. Rev. A 67, 043606 (2003).
* [116] G. Ferrini, A. Minguzzi, and F. W. J. Hekking, Number squeezing, quantum fluctuations, and oscillations in mesoscopic Bose Josephson junctions, Phys. Rev. A 78, 023606 (2008).
* [117] V. S. Shchesnovich and M. Trippenbach, Fock-space WKB method for the boson Josephson model describing a Bose-Einstein condensate trapped in a double-well potential, Phys. Rev. A 78 023611 (2008).
* [118] R. Gati and M. K. Oberthaler, A bosonic Josephson junction, J. Phys. B 40, R61 (2007).
* [119] J. M. Schurer, A. Negretti, and P. Schmelcher, Capture dynamics of ultracold atoms in the presence of an impurity ion, New J. Phys. 17 083024 (2015).
* [120] J. M. Schurer, P. Schmelcher, and A. Negretti, Ground-state properties of ultracold trapped bosons with an immersed ionic impurity, Phys. Rev. A 90, 033601 (2014).
* [121] J. M. Schurer, R. Gerritsma, P. Schmelcher, and A. Negretti, Impact of many-body correlations on the dynamics of an ion-controlled bosonic Josephson junction, Phys. Rev. A 93, 063602 (2016).
* [122] M. Heimsoth, D. Hochstuhl, C. E. Creffield, L. D. Carr, and F. Sols, Effective Josephson dynamics in resonantly driven Bose-Einstein condensates, New J. Phys. 15, 103006 (2013).
* [123] K. Sakmann, A. I. Streltsov, O. E. Alon, and L. S. Cederbaum, Exact Quantum Dynamics of a Bosonic Josephson Junction, Phys. Rev. Lett. 103, 220601 (2009).
* [124] K. Sakmann, A. I. Streltsov, O. E. Alon, and L. S. Cederbaum, Quantum dynamics of attractive versus repulsive bosonic Josephson junctions: Bose-Hubbard and full-Hamiltonian results, Phys. Rev. A 82, 013620 (2010).
* [125] K. Sakmann, Many-Body Schrödinger Dynamics of Bose-Einstein Condensates, Springer, Heidelberg, 2011.
* [126] K. Sakmann, A. I. Streltsov, O. E. Alon, and L. S. Cederbaum, Universality of fragmentation in the Schrödinger dynamics of bosonic Josephson junctions, Phys. Rev. A 89, 023602 (2014).
* [127] J. G. Cosme, M. F. Andersen, and J. Brand, Interaction blockade for bosons in an asymmetric double well, Phys. Rev. A 96, 013616 (2017).
* [128] R. Beinke, S. Klaiman, L. S. Cederbaum, A. I. Streltsov, and O. E. Alon, Many-body tunneling dynamics of Bose-Einstein condensates and vortex states in two spatial dimensions, Phys. Rev. A 92, 043627 (2015).
* [129] D. P. Arovas and A. Auerbach, Quantum tunneling of vortices in two-dimensional superfluids, Phys. Rev. B 78, 094508 (2008).
* [130] A. M. Martin, R. G. Scott, and T. M. Fromhold, Transmission and reflection of Bose-Einstein condensates incident on a Gaussian tunnel barrier, Phys. Rev. A 75, 065602 (2007).
* [131] O. Fialko, A. S. Bradley, and J. Brand, Quantum tunneling of a vortex between two pinning potentials, Phys. Rev. Lett. 108, 015301 (2012).
* [132] P. Ao and D. J. Thouless, Tunneling of a quantized vortex: Roles of pinning and dissipation, Phys. Rev. Lett. 72, 132 (1994).
* [133] A. Auerbach, D. P. Arovas, and S. Ghosh, Quantum tunneling of vortices in two-dimensional condensates, Phys. Rev. B 74, 064511 (2006).
* [134] J. R. Salgueiro, M. Zacarés, H. Michinel, and A. Ferrando, Vortex replication in Bose-Einstein condensates trapped in double-well potentials, Phys. Rev. A 79, 033625 (2009).
* [135] S. Klaiman and O. E. Alon, Spatially-partitioned many-body vortices, J. Phys.: Conf. Ser. 691, 012015 (2016).
* [136] J. Christensson, C. Forssén, S. Åberg, and S. M. Reimann, Effective-interaction approach to the many-boson problem, Phys. Rev. A 79, 012707 (2009).
* [137] R. A. Doganov, S. Klaiman, O. E. Alon, A. I. Streltsov, and L. S. Cederbaum, Two trapped particles interacting by a finite-range two-body potential in two spatial dimensions, Phys. Rev. A 87, 033631 (2013).
* [138] U. R. Fischer, A. U. J. Lode, and B. Chatterjee, Condensate fragmentation as a sensitive measure of the quantum many-body behavior of bosons with long-range interactions, Phys. Rev. A 91, 063621 (2015).
* [139] K. Sakmann and J. Schmiedmayer, Conserving symmetries in Bose-Einstein condensate dynamics requires many-body theory, arXiv:1802.03746v2 [cond-mat.quant-gas] (2018).
* [140] M. Razavy, Quantum Theory of Tunneling, World Scientific, Singapore, 2003.
* [141] H. A. Kramers, Wave mechanics and half-integral quantization, Z. Phys. A 39, 828 (1926).
* [142] R. W. Gurney and E. U. Condon, Wave mechanics and radioactive disintegration, Naure (London) 122, 439 (1928).
* [143] R. W. Gurney and E. U. Condon, Quantum mechanics and radioactive disintegration, Phys. Rev. 33, 127 (1929).
* [144] N. Teeny, E. Yakaboylu, H. Bauke, and C. H. Keitel, Ionization Time and Exit Momentum in Strong-Field Tunnel Ionization, Phys. Rev. Lett. 116, 063003 (2016).
* [145] N. Teeny, C. H. Keitel, and H. Bauke, Virtual-detector approach to tunnel ionization and tunneling times, Phys. Rev. A 94, 022104 (2016).
* [146] L. D. Carr, M. J. Holland, and B. A. Malomed, Macroscopic quantum tunneling of Bose-Einstein condensates in a finite potential well, J. Phys. B 38, 3217 (2005).
* [147] N. Moiseyev and L. S. Cederbaum, Resonance solutions of the nonlinear Schrödinger equation: Tunneling lifetime and fragmentation of trapped condensates, Phys. Rev. A 72, 033605 (2005).
* [148] S. Kim and J. Brand, Decay modes of two repulsively interacting bosons, J. Phys. B 44, 195301 (2011).
* [149] S. Hunn, K. Zimmermann, M. Hiller, and A. Buchleitner, Tunneling decay of two interacting bosons in an asymmetric double-well potential: A spectral approach, Phys. Rev. A 87, 043626 (2013).
* [150] A. U. J. Lode, A. I. Streltsov, O. E. Alon, H.-D. Meyer, and L. S. Cederbaum, Exact decay and tunnelling dynamics of interacting few-boson systems, J. Phys. B: At. Mol. Opt. Phys. 42, 044018 (2009).
* [151] A. U. J. Lode, A. I. Streltsov, O. E. Alon, H.-D. Meyer, and L. S. Cederbaum, Corrigendum: Exact decay and tunnelling dynamics of interacting few-boson systems, J. Phys. B: At. Mol. Opt. Phys. 43, 029802 (2010).
* [152] Axel U. J. Lode, A. I. Streltsov, K. Sakmann, O. E. Alon, and L. S. Cederbaum, How an interacting many-body system tunnels through a potential barrier to open space, Proc. Natl. Acad. Sci. USA 109, 13521 (2012).
* [153] Axel U. J. Lode, S. Klaiman, O. E. Alon, A. I. Streltsov, L. S. Cederbaum, Controlling the velocities and the number of emitted particles in the tunneling to open space dynamics, Phys. Rev. A 89, 053620 (2014).
* [154] Axel U. J. Lode, Tunneling Dynamics in Open Ultracold Bosonic Systems, Springer, Heidelberg, 2015.
* [155] K. Kasamatsu, M. Tsubota, and M. Ueda, Vortex Phase Diagram in Rotating Two-Component Bose-Einstein Condensates, Phys. Rev. Lett. 91, 150406 (2003).
* [156] J. Lovegrove, M. O. Borgh, and J. Ruostekoski, Energetic Stability of Coreless Vortices in Spin-1 Bose-Einstein Condensates with Conserved Magnetization, Phys. Rev. Lett. 112, 075301 (2014).
* [157] E. A. Cornell and C. E. Wieman, Nobel Lecture: Bose-Einstein condensation in a dilute gas, the first 70 years and some recent experiments, Rev. Mod. Phys. 74, 875 (2002).
* [158] W. Ketterle, Nobel lecture: When atoms behave as waves: Bose-Einstein condensation and the atom laser, Rev. Mod. Phys. 74, 1131 (2002).
* [159] J. R. Abo-Shaeer, C. Rama, J. M. Vogels, and W. Ketterle, Observation of vortex lattices in Bose-Einstein condensates, Science 292, 476 (2001).
* [160] D. Dagnino, N. Barberán, M. Lewenstein, and J. Dalibard, Vortex nucleation as a case study of symmetry breaking in quantum systems, Nat. Phys. 5, 431 (2009).
* [161] D. A. Butts and D. S. Rokhsar, Predicted signatures of rotating Bose-Einstein condensates, Nature 397, 327 (1999).
* [162] K. W. Madison, F. Chevy, W. Wohlleben, and J. Dalibard, Vortex Formation in a Stirred Bose-Einstein Condensate, Phys. Rev. Lett. 84, 806 (2000).
* [163] F. Chevy, K. W. Madison, and J. Dalibard, Measurement of the Angular Momentum of a Rotating Bose-Einstein Condensate, Phys. Rev. Lett. 85, 2223 (2000).
* [164] K. W. Madison, F. Chevy, V. Bretin, and J. Dalibard, Stationary States of a Rotating Bose-Einstein Condensate: Routes to Vortex Nucleation, Phys. Rev. Lett. 86, 4443 (2001).
* [165] S. E. Weiner, M. C. Tsatsos, L. S. Cederbaum, and A. U. J. Lode, Phantom vortices: hidden angular momentum in ultracold dilute Bose-Einstein condensates, Sci. Rep. 7, 40122 (2017). License for use of content: https://creativecommons.org/licenses/by/4.0/legalcode.
* [166] M. C. Tsatsos and A. U. J. Lode, Resonances and Dynamical Fragmentation in a Stirred Bose-Einstein Condensate, J. Low Temp. Phys. 181, 171 (2015).
* [167] M. Theisen and A. I. Streltsov, Many-body excitations and deexcitations in trapped ultracold bosonic clouds, Phys. Rev. A 94, 053622 (2016).
* [168] R. Beinke, S. Klaiman, L. S. Cederbaum, A. I. Streltsov, and O. E. Alon, Many-body effects in the excitation spectrum of weakly interacting Bose-Einstein condensates in one-dimensional optical lattices, Phys. Rev. A 95, 063602 (2017).
* [169] S. Burger, F. S. Cataliotti, C. Fort, F. Minardi, M. Inguscio, M. L. Chiofalo, and M. P. Tosi, Superfluid and Dissipative Dynamics of a Bose-Einstein Condensate in a Periodic Optical Potential, Phys. Rev. Lett. 86, 4447 (2001).
* [170] I. Bloch, Ultracold quantum gases in optical lattices, Nature Physics 1, 23 (2005).
* [171] K. Suthar, A. Roy, and D. Angom, Fluctuation-driven topological transition of binary condensates in optical lattices, Phys. Rev. A 91, 043615 (2015).
* [172] O. Morsch and M. Oberthaler, Dynamics of Bose-Einstein condensates in optical lattices, Rev. Mod. Phys. 78, 179 (2006).
* [173] A. K. Tuchman, C. Orzel, A. Polkovnikov, and M. A. Kasevich, Nonequilibrium coherence dynamics of a soft boson lattice, Phys. Rev. A 74, 051601(R) (2006).
* [174] A. K. Tuchman, W. Li, S. Dettmer, and M. A. Kasevich, Localization and anomalous transport in a 1D soft boson optical lattice, New J. Phys. 8, 311 (2006).
* [175] K. Xu, Y. Liu, D. E. Miller, J. K. Chin, W. Setiawan, and W. Ketterle, Observation of Strong Quantum Depletion in a Gaseous Bose-Einstein Condensate, Phys. Rev. Lett. 96, 180405 (2006).
* [176] S. I. Mistakidis, L. Cao, and P. Schmelcher, Negative-quench-induced excitation dynamics for ultracold bosons in one-dimensional lattices , Phys. Rev. A 91, 033611 (2015).
* [177] G. M. Koutentakis, S. I. Mistakidis, and P. Schmelcher, Quench-induced resonant tunneling mechanisms of bosons in an optical lattice with harmonic confinement, Phys. Rev. A 95, 013617 (2017).
* [178] J. Neuhaus-Steinmetz, S. I. Mistakidis, and P. Schmelcher, Quantum dynamical response of ultracold few-boson ensembles in finite optical lattices to multiple interaction quenches, Phys. Rev. A 95, 053610 (2017).
* [179] S. I. Mistakidis, T. Wulf, A. Negretti, and P. Schmelcher, Resonant quantum dynamics of few ultracold bosons in periodically driven finite lattices, J. Phys. B: At. Mol. Opt. Phys. 48, 244004 (2015).
* [180] S. I. Mistakidis and P. Schmelcher, Mode coupling of interaction quenched ultracold few-boson ensembles in periodically driven lattices, Phys. Rev. A 95, 013625 (2017).
* [181] A. I. Streltsov, K. Sakmann, O. E. Alon, and L. S. Cederbaum, Accurate multi-boson long-time dynamics in triple-well periodic traps, Phys. Rev. A 83, 043604 (2011).
* [182] R. Roy, A. Gammal, M. C. Tsatsos, B. Chatterjee, B. Chakrabarti, and A. U. J. Lode, Phases, many-body entropy measures, and coherence of interacting bosons in optical lattices, Phys. Rev. A 97, 043625 (2018).
* [183] S. Dutta, M. C. Tsatsos, S. Basu, and A. U. J. Lode, Management of the Correlations of Ultracold Bosons in Triple Wells, arXiv:1802.02407 [cond-mat.quant-gas] (2018).
* [184] L. Cao, S. Krönke, J. Stockhofe, J. Simonet, K. Sengstock, D.-S. Lühmann, and P. Schmelcher, Beyond-mean-field study of a binary bosonic mixture in a state-dependent honeycomb lattice, Phys. Rev. A 91, 043639 (2015).
* [185] P. Grech and R. Seiringer, The Excitation Spectrum for Weakly Interacting Bosons in a Trap, Commun. Math. Phys. 322, 559 (2013).
* [186] R. Seiringer, The Excitation Spectrum for weakly interacting Bosons, Commun. Math. Phys. 306, 565 (2011).
* [187] R. Beinke, L. S. Cederbaum, and O. E. Alon, Enhanced many-body effects in the excitation spectrum of a weakly-interacting rotating Bose-Einstein condensate, arXiv:1803.04762 [cond-mat.quant-gas] (2018).
* [188] M. R. Matthews, B. P. Anderson, P. C. Haljan, D. S. Hall, C. E. Wieman, and E. A. Cornell, Vortices in a Bose-Einstein Condensate, Phys. Rev. Lett. 83, 2498 (1999).
* [189] N. Regnault and T. Jolicoeur, Quantum Hall Fractions in Rotating Bose-Einstein Condensates, Phys. Rev. Lett. 91, 030402 (2003).
* [190] S. Stock, B. Battelier, V. Bretin, Z. Hadzibabic, and J. Dalibard, Bose-Einstein condensates in fast rotation, Laser Phys. Lett. 2, 275 (2005).
* [191] C.-C. Chang, N. Regnault, T. Jolicoeur, and J. K. Jain, Composite fermionization of bosons in rapidly rotating atomic traps, Phys. Rev. A 72 013611 (2005).
* [192] M. Roncaglia, M. Rizzi, and J. Dalibard, From rotating atomic rings to quantum Hall states, Sci. Rep. 1, 43 (2011).
* [193] I. Coddington, P. Engels, V. Schweikhard, and E. A. Cornell, Observation of Tkachenko Oscillations in Rapidly Rotating Bose-Einstein Condensates, Phys. Rev. Lett. 91, 100402 (2003).
* [194] V. Schweikhard, I. Coddington, P. Engels, V. P. Mogendorff, and E. A. Cornell, Rapidly Rotating Bose-Einstein Condensates in and near the Lowest Landau Level, Phys. Rev. Lett. 92, 040404 (2004).
* [195] A. Aftalion, X. Blanc, and J. Dalibard, Vortex patterns in a fast rotating Bose-Einstein condensate, Phys. Rev. A 71, 023611 (2005).
* [196] T. Mizushima, M. Ichioka, and K. Machida, Beliaev Damping and Kelvin Mode Spectroscopy of a Bose-Einstein Condensate in the Presence of a Vortex Line, Phys. Rev. Lett. 90, 180401 (2003).
* [197] T. P. Simula and K. Machida, Kelvin-Tkachenko waves of few-vortex arrays in trapped Bose-Einstein condensates , Phys. Rev. A 82, 063627 (2010).
* [198] L. O. Baksmaty, S. J. Woo, S. Choi, and N. P. Bigelow, Tkachenko Waves in Rapidly Rotating Bose-Einstein Condensates, Phys. Rev. Lett. 92, 160405 (2004).
* [199] F. Chevy, Low-lying twisting and acoustic modes of a rotating Bose-Einstein condensate, Phys. Rev. A 73, 041604(R) (2006).
* [200] A. Collin, Collective modes and the broken symmetry of a rotating attractive Bose gas in an anharmonic trap, Phys. Rev. A 73, 013611 (2006).
* [201] F. Ancilotto and F. Toigo, Dipolar Bose gas in a highly anharmonic trap, Phys. Rev. A 89, 023617 (2014).
* [202] S. Viefers, T. H. Hansson, and S. M. Reimann, Bose condensates at high angular momenta , Phys. Rev. A 62, 053604 (2000).
* [203] S. M. Reimann, M. Koskinen, Y. Yu, and M. Manninen, Rotating quantum liquids crystallize , New J. Phys. 8, 59 (2006).
* [204] J. C. Cremon, G. M. Kavoulakis, B. R. Mottelson, and S. M. Reimann, Vortices in Bose-Einstein condensates: Finite-size effects and the thermodynamic limit, Phys. Rev. A 87, 053615 (2013).
* [205] J. C. Cremon, A. D. Jackson, E. Ö. Karabulut, G. M. Kavoulakis, B. R. Mottelson, and S. M. Reimann, Rotating Bose-Einstein condensates: Closing the gap between exact and mean-field solutions, Phys. Rev. A 91, 033632 (2015).
* [206] P. T. Nam and R. Seiringer, Collective Excitations of Bose Gases in the Mean-Field Regime, Arch. Rational Mech. Anal. 215, 381 (2015).
* [207] A. L. Fetter, Rotating trapped Bose-Einstein condensates, Rev. Mod. Phys. 81, 647 (2009).
* [208] O. E. Alon, Solvable model of a generic trapped mixture of interacting bosons: reduced density matrices and proof of Bose-Einstein condensation, J. Phys. A 50, 295002 (2017).
* [209] S. I. Mistakidis, G. C. Katsimiga, P. G. Kevrekidis, and P. Schmelcher, Correlation effects in the quench-induced phase separation dynamics of a two species ultracold quantum gas, New J. Phys. 20, 043052 (2018).
* [210] M. Pyzh, S. Krönke, C. Weitenberg, and P. Schmelcher, Spectral properties and breathing dynamics of a few-body Bose-Bose mixture in a 1D harmonic trap, New J. Phys. 20, 015006 (2018).
* [211] V. Cikojević, L. Vranješ Markić, and J. Boronat, Harmonically trapped Bose-Bose mixtures: a quantum Monte Carlo study, New J. Phys. 20 085002 (2018).
* [212] O. I. Utesov, M. I. Baglay, and S. V. Andreev, Effective interactions in a quantum Bose-Bose mixture, Phys. Rev. A 97, 053617 (2018).
* [213] P. Cheiney, C. R. Cabrera, J. Sanz, B. Naylor, L. Tanzi, and L. Tarruell, Bright Soliton to Quantum Droplet Transition in a Mixture of Bose-Einstein Condensates, Phys. Rev. Lett. 120, 135301 (2018).
* [214] Y. Asano, M. Narushima, S. Watabe, and T. Nikuni, Collective excitations in Bose-Fermi mixtures, arXiv:1807.06747 [cond-mat.quant-gas] (2018).
* [215] J. Chen, J. M. Schurer, and P. Schmelcher, Bunching-antibunching crossover in harmonically trapped few-body Bose-Fermi mixtures, Phys. Rev. A 98, 023602 (2018).
* [216] J. Chen, J. M. Schurer, and P. Schmelcher, Entanglement Induced Interactions in Binary Mixtures, Phys. Rev. Lett. 121, 043401 (2018).
* [217] P. Siegl, S. I. Mistakidis, and P. Schmelcher, Many-body expansion dynamics of a Bose-Fermi mixture confined in an optical lattice, Phys. Rev. A 97, 053626 (2018).
* [218] R. S. Lous, I. Fritsche, M. Jag, F. Lehmann, E. Kirilov, B. Huang, and R. Grimm, Probing the Interface of a Phase-Separated State in a Repulsive Bose-Fermi Mixture, Phys. Rev. Lett. 120, 243403 (2018).
* [219] S. Häfner, J. Ulmanis, E. D. Kuhnle, Y. Wang, C. H. Greene, and M. Weidemüller, Role of the intraspecies scattering length in the Efimov scenario with large mass difference, Phys. Rev. A 95, 062708 (2017).
* [220] O. E. Alon, A. I. Streltsov, and L. S. Cederbaum, Unified view on multiconfigurational time propagation for systems consisting of identical particles, J. Chem. Phys. 127, 154103 (2007).
* [221] E. Fasshauer and A. U. J. Lode, Multiconfigurational time-dependent Hartree method for fermions: Implementation, exactness, and few-fermion tunneling to open space, Phys. Rev. A 93, 033635 (2016).
* [222] C. Lévêque and L. B. Madsen, Time-dependent restricted-active-space self-consistent-field theory for bosonic many-body systems, New. J. Phys. 19, 043007 (2013).
* [223] C. Lévêque and L. B. Madsen, Multispecies time-dependent restricted-active-space self-consistent-field theory for ultracold atomic and molecular gases, J. Phys. B 51, 155302 (2018).
* [224] E. Polizzi, Density-Matrix-Based Algorithms for Solving Eigenvalue Problems, Phys. Rev. B. Vol. 79, 115112 (2009).
* [225] http://www.feast-solver.org.
* [226] E. H. Lieb and R. Seiringer, Proof of Bose-Einstein Condensation for Dilute Trapped Gases, Phys. Rev. Lett. 88, 17 (2002).
* [227] S. Klaiman and L. S. Cederbaum, Overlap of exact and Gross-Pitaevskii wave functions in Bose-Einstein condensates of dilute gases, Phys. Rev. A 94, 063648 (2016).
* [228] L. S. Cederbaum, Exact many-body wave function and properties of trapped bosons in the infinite-particle limit, Phys. Rev. A 96, 013615 (2017).
* [229] P. A. Ruprecht, M. Edwards, K. Burnett, and C. W. Clark, Probing the linear and nonlinear excitations of Bose-condensed neutral atoms in a trap, Phys. Rev. A 54, 4178 (1996).
* [230] Y. Castin, and R. Dum, Instability and Depletion of an Excited Bose-Einstein Condensate in a Trap, Phys. Rev. Lett. 79, 3553 (1997).
* [231] C. W. Gardiner, Particle-number-conserving Bogoliubov method which demonstrates the validity of the time-dependent Gross-Pitaevskii equation for a highly condensed Bose gas, Phys. Rev A 56, 1414 (1997).
* [232] Y. Castin, and R. Dum, Low-temperature Bose-Einstein condensates in time-dependent traps: Beyond the U(1) symmetry-breaking approach, Phys. Rev. A 57, 3008 (1998).
* [233] O. E. Alon, A. I. Streltsov, and L.S. Cederbaum, Time-dependent multi-orbital mean-field for fragmented Bose-Einstein condensates, Phys. Lett. A 362, 453 (2007).
* [234] O. E. Alon, A. I. Streltsov, and L. S. Cederbaum, Multiorbital mean-field approach for bosons, spinor bosons, and Bose-Bose and Bose-Fermi mixtures in real-space optical lattices, Phys. Rev. A 76, 013611 (2007).
* [235] L. S. Cederbaum and A. I. Streltsov, Best mean-field for condensates, Phys. Lett. A 318, 564 (2003).
* [236] L. S. Cederbaum and A. I. Streltsov, Self-consistent fragmented excited states of trapped condensates, Phys. Rev. A 70, 023610 (2004).
* [237] O. E. Alon and L. S. Cederbaum, Pathway from Condensation via Fragmentation to Fermionization of Cold Bosonic Systems, Phys. Rev. Lett. 95, 140402 (2005).
* [238] O. E. Alon, A. I. Streltsov, and L. S. Cederbaum, Zoo of Quantum Phases and Excitations of Cold Bosonic Atoms in Optical Lattices, Phys. Rev. Lett. 95, 030405 (2005).
* [239] L. S. Cederbaum, A. I. Streltsov, and O. E. Alon, Fragmented Metastable States Exist in an Attractive Bose-Einstein Condensate for Atom Numbers Well above the Critical Number of the Gross-Pitaevskii Theory, Phys. Rev. Lett. 100, 040402 (2008).
* [240] J. Grond, A. I. Streltsov, and L. S. Cederbaum, Excitation spectra of fragmented condensates by linear-response: General theory and application to a condensate in a double-well, Phys. Rev. A 86, 063607 (2012).
* [241] O. E. Alon, Many-body excitation spectra of trapped bosons with general interaction by linear response, J. Phys.: Conf. Ser. 594, 012039 (2015).
* [242] S. Klaiman and O. E. Alon, Variance as a sensitive probe of correlations, Phys. Rev. A 91, 063613 (2015).
* [243] S. Klaiman, A. I. Streltsov, and O. E. Alon, Uncertainty product of an out-of-equilibrium many-particle system, Phys. Rev. A 93, 023605 (2016).
* [244] S. Klaiman, A. I. Streltsov, and O. E. Alon, Uncertainty product of an out-of-equilibrium Bose-Einstein condensate, J. Phys.: Conf. Ser. 826 012020 (2017).
* [245] S. K. Haldar and O. E. Alon, Impact of the range of the interaction on the quantum dynamics of a bosonic Josephson junction, Chem. Phys. 509, 72 (2018).
* [246] S. Klaiman, R. Beinke, L. S. Cederbaum, A. I. Streltsov, and O. E. Alon, Variance of an anisotropic Bose-Einstein condensate, Chem. Phys. 509, 45 (2018).
|
# Near-Field Millimeter Wave Vector Measurements - Experimental Design &
Measurement Interpretation
Laurent Chusseau1, Thibaut Auriac1, Jérémy Raoult1 1IES, Université de
Montpellier, CNRS, Montpellier, France
###### Abstract
Near-field imaging experiments exist both in optics and microwaves with often
different methods and theoretical supports. For millimeter waves or THz waves,
techniques from both fields can be merged to identify materials at the micron
scale on the surface or in near-surface volumes. The principle of such near-
field vector imaging at the frequency of 60 GHz is discussed in detail here.
We develop techniques for extracting vector voltages and methods for
extracting the normalized near-field vector reflection on the sample. In
particular, the subharmonic IQ mixer imbalance, which produced corrupted
outputs either due to amplitude or phase differences, must be taken into
account and compensated for to avoid any systematic errors. We provide a
method to fully characterize these imperfections and to isolate the only
contribution of the near-field interaction between the probe and the sample.
The effects of the mechanical modulation waveform and harmonic rank used for
signal acquisition are also discussed.
###### Index Terms:
Near-field measurement, mm-waves, microwave probes, vector local
characterization, subharmonic IQ mixer.
## I Introduction
Near-field imaging is now an essential technique for the characterization of
devices and materials. Depending on the targeted application, reflection or
transmission modes are used in the infrared [1], microwave [2] or THz [3], and
for the last two take advantage of the transparency of many materials for non-
destructive testing through the surface [4, 5, 6, 7].
If high spatial resolution is desired in respect to the measurement
wavelength, a probe is required to concentrate the field and overcome the Abbe
diffraction limit by means of evanescent waves. At GHz frequencies, the
scanning microwave microscope (SMM) frequently uses coaxial resonant probes
whose quality factor and resonance can be related to the interaction with the
sample [8, 9, 10], or AFM tips [11, 12, 13] that can be modeled
electromagnetically in 3D [14, 15, 16]. The latter technique is also used in
the millimeter wave (mmW) region, which is the realm where it is possible to
combine optical and microwave techniques, for example the joint use of
propagating waves [17] and guided waves coupled to specific probes including
small antennae capable of concentrating the electric or magnetic field with a
very high resolution [18, 19, 20].
Although most of near-field experiments in the mmW have focused on intensity
measurements [21, 22, 6, 23, 24, 17], local vector characterization is
necessary to distinguish material changes from loss changes. Many such
experiments have recently been proposed in scattering-type scanning near-field
optical microscopy (s-SNOM) applied to the THz domain [25, 26, 27, 28], but
only the latter seems to be easy to translate to mmW. We do the same
transformation on our 60 GHz test bench [24] and chose to include a
subharmonic IQ mixer to simultaneously detect the two quadratures of the
reflected field, but we kept a complete waveguide configuration. This paper
details the new experiment and elaborate the required processing of
experimental signals in order to determine without systematic error and with
reasonable accuracy the vector voltages and reflection coefficients related to
the near-field probe-sample interaction. We show how the rapid phase rotation
due to the path length in mmW can be used to achieve these values, and how
neutralize the IQ imbalance of the mixer, either in amplitude or phase, by its
theoretical account in data processing. In the end, this gives us an overall
figure of merit for the subharmonic IQ mixer and the entire detection system.
The paper is organized as follows. In §II we describe the experimental setup
and its basic operation. §III details the processing to be applied to
experimental data, namely the vector voltage determination and how mixer non-
ideal behavior transforms the observed vector voltage. We then develop the
method for tracing the true reflection coefficient, provided that a quality
reference measurement is available. In §IV, we deal with the effect of the
harmonic rank of the detection in relation to the modulation waveform applied
to the mechanical modulation of the distance between the probe and the sample,
which is a peculiarity of our experiment as compared to s-SNOM. Finally, in §V
we qualify the frequency response of our experiment and the subharmonic mixer
imbalance.
## II Measurement Setup
The scheme of the new experiment is given in Fig. 1 with its central element
for the vector output: the subharmonic IQ mixer which operates at zero
intermediate frequency in the band 55-65 GHz. Compared to our previous
intensity reflectometer [24], the whole mmW transmitting/detecting system has
changed and now uses an HF synthesizer followed by multiplication chains for
the detection and LO signal. In the measuring path, an active $\times 4$
multiplier delivering 10 dBm is used in the front of a 3 dB coupler which
returns the reflected signal at the input of the mixer. In between, we take
care of the mmW signal level entering the subharmonic IQ mixer by
incorporating a variable attenuator to validate the -10 dBm requirement for
linear operation. The LO path of the mixer is fed by an active doubler
delivering 10 dBm. The $\times 4$ and $\times 2$ active multipliers are not
really perfectly flat in frequency, unlike the reference synthesizer. As a
result, the subharmonic IQ mixer’s conversion loss variations, which are
typically 15 to 18 dB, may experience additional ripple due to these power
variations of the mmW sources. The whole system was provided by LTEQ Microwave
according to our specifications.
As in [24], probes are homemade bow-ties attached to a WR15 open end. Metal
triangles used for the bow-tie were produced using fs-laser cut on a 10 µm
thick tungsten sheet. The Fig. 2 illustrates this realization and the probe
positioning 10 µm above a surface. The lateral and vertical positioning over
the sample is ensured by a motorized step-by-step system (Newport XPS-C8) and
is optically monitored with two cameras at right angle as shown in the
picture. When set at their highest magnification, cameras equipped with
Navitar UltraZoom lenses can observe a displacement of 1 µm, which corresponds
to one motor step.
Figure 1: Scheme of the mmW near-field experimental setup. All mmW links are
made using WR15 waveguides up to the subharmonic IQ mixer.
The modulation of the probe-to-sample distance is carried out by the
piezoelectric actuator (PI P-611.Z Precision Z-stage) driven by a signal
generator whose frequency is set to 40 Hz due to the limited weight capacity
of this actuator. Subharmonic IQ mixer outputs are thus modulated at this low-
frequency that is also the reference of the two lock-in (AMETEK 7265 Dual
Phase DSP). This allows the filtering of the near-field spatial component
ideally split in its real and imaginary parts. In practice both lock-in
outputs are complex and refer to the phase of the low-frequency modulation.
Correct usage thus requires to set the same phase reference on the two lock-
in.
Figure 2: Photos of the experiment. Left: emission/detection board in place
with the waveguide probe and the two cameras. Right up: tungsten bow-tie probe
view from above with the microscope. Right down: Probe as observed by the
camera 10 µm above the gold mirror, the reflection is used to set the minimum
distance to $h=10$ µm.
At first glance, the real part in channel I (respectively, the imaginary part
in channel Q) of the complex near-field reflected voltage $\widetilde{v}$ is
expected to have the module $m_{I}=\sqrt{X_{I}^{2}+Y_{I}^{2}}$ (respectively,
$m_{Q}=\sqrt{X_{Q}^{2}+Y_{Q}^{2}}$). If these values are obtained
straightforwardly, their respective sign must be determined to fully locate
$\widetilde{v}$ in the entire complex plane.
Our first test sample was chosen to obtain the highest possible near-field
reflectivity. It is a gold optical mirror from Edmund Optics with an ultra-
flat substrate ($\lambda/20$) and a thin (a few nm) dielectric protective
layer. A far-field reflectivity of $\approx 97\%$ is certified in the mid-
infrared, a value also expected in the far infrared and in the mmW range since
gold is homogeneous and well known [29]. This sample is intended to be a
reference for reflectivity, to which any other sample can be compared.
When scanning the input frequency in a tiny range of 200 MHz, the four outputs
of the lock-ins oscillate rapidly, as shown in Fig. 3. This is due to the
total electrical length of the system here estimated at about $1.25$ m using
the $240\pm 2$ MHz period deduced for the $X_{i}$ and the $Y_{i}$. Since the
probe senses the reflection on a flat and homogeneous material, it is not
anticipated that it will contribute to a rapidly changing $\widetilde{v}$ with
frequency, but rather that it can be regarded as constant. Consequently, we
assume in what follows that the almost sinusoidal evolution of the $X_{i}$ and
$Y_{i}$ is solely attributed to the electrical length.
Figure 3: Lock-in outputs as measured $h=10$ µm above a gold mirror with a
sinusoidal vibration amplitude $\delta h=20$ µm while scanning the input
frequency.
In addition, a close inspection of Fig. 3 shows a deviation from the expected
ideal subharmonic IQ mixer: at 60.765 GHz channel I cancels ($X_{I}=Y_{I}=0$)
but channel Q is not maximum, which occurs at 60.785 GHz. Conversely, the same
occurs at 60.720 GHz for channel Q and 60.705 GHz for channel I. This is quite
unexpected for a perfect IQ mixer and shows the mixture of channel signals
because of its imperfection. As a result, the measurement of near-field
reflection at a single frequency would inevitably be flawed by a systematic
error regarding the contribution of the material placed just under the probe.
We therefore prefer to carry out frequency sweeps like the one shown in Fig. 3
and process the data to eliminate these errors by assuming that the near-field
contribution is constant over the low measurement frequency band.
## III Measurement Data Processing
### III-A Vector Voltage Determination
Proceeding from a measurement like the one shown in Fig. 3, the modulus of the
reflected near-field voltage is directly obtained by
$\left\lvert{\widetilde{v}}\right\rvert=\sqrt{m_{I}^{2}+m_{Q}^{2}}.$ (1)
Accessing the phase is more tedious since the rough estimate
$\angle\widetilde{v}=\arctan\frac{m_{Q}}{m_{I}}$ is inherently limited to
$[0,\frac{\pi}{2}]$ because of the positive $\arctan$ argument. Going further
is possible using the original lock-in outputs, which include the sign changes
for channels I and Q when $m_{I}$ and $m_{Q}$ go through zero. Noting that
these outputs always present opposite $X$ and $Y$ values, the sign are
straightforwardly obtained from
$\operatorname{sgn}(I)=\operatorname{sgn}(X_{I}-Y_{I})$ and
$\operatorname{sgn}(Q)=\operatorname{sgn}(X_{Q}-Y_{Q})$. Keeping track of all
these sign changes, we define the phase of the reflected voltage using the
heuristic formula
$\angle\widetilde{v}=\arctan\frac{\operatorname{sgn}(Q)m_{Q}}{\operatorname{sgn}(I)m_{I}}.$
(2)
We have applied these formulas to the experimental case of the gold mirror in
Fig. 3 and also for an unintentionally doped GaAs wafer (ITME, Warsaw). The
same geometrical conditions were applied for the measurement, _i.e._ $h=10$
µm, $\delta h=20$ µm and a sinusoidal modulation. The results are shown in
Fig. 4 with the modulus and phase of $\widetilde{v}$ and their locations in
the complex plane. As expected, $\angle\widetilde{v}$ covers the whole range
$[-\pi,\pi]$ which leads to elliptic representations in the complex plane.
Figure 4: Evolutions of the complex reflected near-field voltages
$\widetilde{v}$ for a gold mirror and a GaAs undoped wafer. Left up: modulus,
left down: phase, right: complex representation. Dots are for gold and open
triangles for GaAs.
Although the two materials can be easily distinguished on these curves,
especially by their average values $\left\lvert{\widetilde{v}}\right\rvert$
which are clearly different and by their shifted $\angle\widetilde{v}$
evolutions, it is impossible at this stage to characterize a material with a
single complex value whereas this was hoped by assuming a constant influence
of the material on the limited frequency range considered. In this framework,
centered ellipses are distorted views of circles usually obtained in microwave
by measurements using sliding loads. We attribute this distortion to the
imbalance of the IQ mixer and model it in the sequel to refine the analysis.
### III-B Normalized Vector Reflection
Let us consider the ideally reflected complex voltage $v$ in the near-field of
the tested surface. As said it should be constant over the tiny frequency
range considered because our materials are far from any resonance. The voltage
$v$ thus undergoes only a phase shift $\exp j\phi$ due to the electrical
length before being transformed by the subharmonic IQ mixer to produce
$\widetilde{v}$. We assume that mixer imbalance is twofold: first, it affects
the gain by a factor $\gamma$ possibly different from unity that we apply only
to the imaginary part, and second, the imperfect quadrature is accounted for
by an additional angle $\theta$ added to the $\pi/2$ phase shift of the Q
channel. This is written
$\widetilde{v}=\Re(v\exp j\phi)+j\gamma\Im(v\exp j(\phi+\theta)).$ (3)
If we now divide the vector voltage measured on the sample
$\widetilde{v}_{S}$, for instance the GaAs wafer of Fig. 4 by the vector
voltage measured on a reference $\widetilde{v}_{R}$, namely the gold mirror in
the same figure, which is highly conductive at mmW and supposed to produce the
highest possible reflection, we obtain
$\frac{\widetilde{v}_{S}}{\widetilde{v}_{R}}=\rho\frac{\cos(\phi+\varphi)+j\gamma\sin(\phi+\varphi+\theta)}{\cos\phi+j\gamma\sin(\phi+\theta)},$
(4)
where $\rho=\left\lvert{v_{S}}\right\rvert/\left\lvert{v_{R}}\right\rvert$ and
$\varphi=\angle v_{S}-\angle v_{R}$. As in reflectometry, the normalized
vector reflection $\rho\exp j\varphi$ is thus intended to be the near-field
mmW reflection coefficient $\Gamma$ of the tested material, provided that the
reference is a perfect metal.
Eq. 4 is easily identified as a circle in the complex plane parametrized by
$\phi$. It is completely described if $\phi\in[0,\pi]$ and collapses to a
single point with a perfect IQ mixer ($\gamma=1$ and $\theta=0$). Solving a
set of 3 equations for 3 different points such as $\phi=0$, $\phi=\pi/2$ and
$\phi=-\theta$ which must belong to the same circle
$(X-X_{c})^{2}+(Y-Y_{c})^{2}-R^{2}=0$, provides the link with our parameters
in Eq. (4). After some lengthly developments, we end up with the definition of
the circle
$\displaystyle X_{c}$ $\displaystyle=\rho\cos\varphi$ (5a) $\displaystyle
Y_{c}$ $\displaystyle=\frac{\rho(1+\gamma^{2})\sin\varphi}{2\gamma\cos\theta}$
(5b) $\displaystyle R$ $\displaystyle=\zeta Y_{c}$ (5c)
with
$\zeta=\sqrt{1-\frac{4\gamma^{2}\cos^{2}\theta}{\left(1+\gamma^{2}\right)^{2}}}.$
(6)
Interestingly the term $\zeta$ is completely independent of both the sample
and reference. It is thus some kind of measure of the IQ mixer ideality within
our experiment. The gain imbalance $\gamma$ and the quadrature error $\theta$
are mixed in $\zeta$ and cannot be solved separately because we are missing a
fourth variable. Order of magnitude of $\zeta$-values are given in Fig. 5 as a
function of $\gamma$ and $\theta$. For gain errors lower than 3 dB and phase
errors lower than 30° it does not exceed 0.6 which is nonetheless significant
compared to its limit value of 1 obtained when $\gamma$ or $1/\gamma\to\infty$
or $\theta\to\pm\pi/2$. We must keep in mind this limit value of 1 which
cannot be exceeded under any circumstances in our experiments. It is a
criterion of suitability of the measurements and their exploitation that will
be implemented in the following. Values on the order of 0.4 and less, however,
seem acceptable.
Figure 5: Ideality factor $\zeta$ as a function of gain mismatch $\gamma$ and
quadrature phase mismatch $\theta$.
Go back to measurements to calculate and plot the ratio
$\widetilde{v}_{S}/\widetilde{v}_{R}$ from data of Fig. 4. As shown in Fig. 6
we end up with a nearly perfect circle, just like predicted above. We have
limited here the frequency range previously considered to $[60.690-60.815]$
GHz in order to describe the circle without overlap. Since the points are not
evenly distributed on the circle, there is no simple criteria to immediately
access the position of its center, so a least-squares fit was used to
determine the triplet $(X_{c},Y_{c},R)$. A secondary result of the fit that
minimizes the distance between the measured points and the circle is the error
distance accumulated for measurement points to the circle. It is used in the
sequel to estimate the uncertainties on each of the inferred values using
Student’s t-test with 95% confidence level and classical propagation of errors
at the first level on the equations. The error is only $\approx 6\,10^{-4}$
per point in this example, as could be expected given the excellent agreement
between the best circle, represented by a solid line in the Fig. 6, with the
measurement points. The position of the center is noted by a filled square.
The module and phase of the near-field reflection are then straightforwardly
calculated from Eqs. (5) by
$\displaystyle\rho$ $\displaystyle=\sqrt{X_{c}^{2}+Y_{c}^{2}-R^{2}}$ (7a)
$\displaystyle\varphi$ $\displaystyle=\arccos\frac{X_{c}}{\rho}.$ (7b)
Figure 6: Ratio $\widetilde{v}_{S}/\widetilde{v}_{R}$ in the complex plane for
the GaAs sample and the gold mirror reference of Fig. 4. Dots are
measurements, the solid line is the corresponding best circle fit whose center
is the filled square, and the filled diamond is the deduced $\Gamma=\rho\exp
j\varphi$ normalized vector reflection.
In the case of Fig. 6 this deduced point is represented by a diamond. It is
obviously different from the center, even if according to Eq. (5a) their real
parts are identical. In essence, we end up here with the normalized vectorial
reflection $\Gamma=v_{S}/v_{R}=(0.482+j0.225)\pm 0.004$ with an ideality
factor $\zeta=0.436\pm 0.005$, which illustrate the effectiveness of the
entire measurement operation process, including error estimates. In the ideal
case where our reference is perfect, it is expected that $\Gamma$ is the true
reflection coefficient in near-field at mmW.
## IV Influence of the mechanical modulation
In scattering-type scanning near-field optical microscopy (s-SNOM) the
mechanical probe-to-sample distance modulation is always sinusoidal since it
is provided by a quartz resonance [25, 12, 17, 28]. Unconventional modulation
waveforms are possible in our experiment since the mechanical displacement is
performed by a piezoelectric actuator controlled by a low-frequency generator.
The actuator is optimized for 40 Hz but has a cut-off frequency of about 100
Hz, so we can use any arbitrary waveform validating these conditions.
Measurements have been performed as test cases with our two samples: the GaAs
wafer and the gold mirror.
We try successively sinusoid, square and triangle modulations at constant 40
Hz fundamental frequency. In each case a measure is done on the reference and
reproduce in exactly the same conditions for position, modulation and
detection for the GaAs sample. Initial position of the probe is always the
minimum distance $h=10$ µm above the surface before applying the modulation
whose amplitude is intended to be $\delta h=20$ µm, as calibrated for
sinusoid. Depending on the modulation waveform, the effective $h$ may be
smaller, especially with square modulation where the transfer function of the
piezoelectric actuator exhibits an overshoot. Although the displacement
approximately follows the command, the rise and fall times are not
instantaneous and the effective distance between the probe and the sample is
distorted. This occurs also to a lesser extent with triangular modulation.
TABLE I: Deduced normalized reflection coefficient $\Gamma$ of GaAs versus the modulation waveform and the harmonic rank. Rank | Sinus Modulation | Square Modulation | Triangle Modulation
---|---|---|---
$\Gamma$ | $\zeta$ | $\Gamma$ | $\zeta$ | $\Gamma$ | $\zeta$
1 | $(0.482+j0.225)\pm 0.004$ | $0.436$ | $(0.489+j0.226)\pm 0.004$ | $0.435$ | $(0.487+j0.226)\pm 0.005$ | $0.436$
2 | $(0.655+j0.280)\pm 0.006$ | $0.408$ | $(0.679+j0.270)\pm 0.009$ | $0.440$ | $(0.687+j0.283)\pm 0.006$ | $0.410$
3 | — | $5.6$ | $(0.511+j0.231)\pm 0.005$ | $0.436$ | $(0.528+j0.234)\pm 0.004$ | $0.431$
4 | — | $1.00$ | $(0.728+j0.289)\pm 0.005$ | $0.414$ | $(0.735+j0.283)\pm 0.021$ | $0.418$
5 | | | — | $2.9$ | — | $2.9$
6 | | | $(0.760+j0.263)\pm 0.017$ | $0.418$ | |
7 | | | $(0.411+j0.173)\pm 0.010$ | $0.452$ | |
8 | | | $(0.779+j0.276)\pm 0.026$ | $0.428$ | |
9 | | | $(0.355+j0.146)\pm 0.021$ | $0.47$ | |
The experimental procedure and the determination of the vectorial voltage
described in Fig. 4 were reproduced for the three modulations with increasing
harmonic ranks selected on the lock-ins. This was done as long as the detected
signal emerged from the noise. As can be seen from the average
$\left\lvert{\widetilde{v}}\right\rvert$ voltages plotted in Fig. 7, the
detected signal decreases rapidly with harmonic rank, but it also depends a
lot on the waveform of the modulation. Nevertheless,
$\left\lvert{\widetilde{v}}\right\rvert$ is very similar for all waveforms at
the first two harmonics. Beyond that, the decrease in intensity is huge for
sinusoidal modulation, which generates those high harmonics only by the
nonlinear probe-to-sample interaction caused by the rapid decrease of
intensity in the near-field. This peculiarity of probe-to-sample interaction
is often used to improve the spatial resolution [30, 31, 24]. For other
modulation waveforms, higher harmonics are less attenuated, especially in the
case of square modulation. Actually they tend to follow the intensities that
could be computed by the Fourier transform of the initial waveforms, in
particular the high attenuation of even harmonics. For such waveforms we can
no more argue for an improvement in resolution by detecting high order
harmonics.
Figure 7: Average module of the voltage detected
$\left\lvert{\widetilde{v}}\right\rvert$ for $f\in[60.690,60.815]$ GHz versus
harmonic rank and modulation waveform: $\bullet$ are for sinus, $\blacksquare$
for square and $\blacklozenge$ for triangle modulations.
The results were interpreted using Eqs. (7) applied to the best fitted circles
on the experiments. We used $\zeta$ as a determinant: if its value exceeds 1,
which according to the discussion in §III-B is impossible in the framework of
our subharmonic IQ mixer model, then we rejected the measure as a whole. The
practical reasons for such a mismatch can be multiple, such as non-
reproducible positioning over our test surfaces, weak or noisy signals… The
factor $\zeta$ is therefore an effective quality criterion for the control of
our measurements. In the case of triangle modulation, the Fig. 8 illustrates
the measurement points with the 4 circles obtained for the first 4 harmonics.
Although the measurement points are all well located on circles, it is obvious
that the even and odd harmonics are grouped together, but the two groups do
not merge even if they overlap. As a result the extracted reflection
coefficient with even harmonics shows real and imaginary parts which both have
higher values.
Figure 8: Normalized reflection measured for the four first harmonic ranks
with a triangle modulation waveform. Dots are measurements and circles their
corresponding best fits.
The complete results for all harmonics of all waveforms are given in the Table
I. Rejected cases occurring because $\zeta\geq 1$ are represented by a dash
and this occurs especially for all harmonics 5. We believe that this is
because this frequency of 200 Hz is the only one that is also a harmonic of
the power supply frequency, which inevitably produces a large increase of
noise in the detection circuit. As soon as
$\left\lvert{\widetilde{v}}\right\rvert$ is high enough in all other cases,
except for the higher harmonics of the sinusoidal modulation, we obtain
$\zeta\approx 0.4$ and the results do not seem to be biased.
The data of Table I are plotted in Fig. 9. The superposition of the results
for all waveforms highlights the experimental independence in relation to this
setting. On the contrary, the alternation of the values between even and odd
harmonics is conspicuous. While the real parts have values that increase with
the rank of even harmonics, the opposite occurs for odd harmonics, and the
same trend is practically true for the imaginary part. Nevertheless the most
notable thing is the dichotomy between even and odd harmonics. The most
surprising fact of these results is that for a given harmonic rank the
normalized reflection value determined is the same regardless of the
modulation shape, and this is verified within the limit of the error bars for
the first two harmonics and still very close for next two one. Note that
increasing the harmonic rank rapidly decreases the signal and therefore
mechanically increases the error.
Figure 9: Real (left) and imaginary parts (right) of the normalized reflection
$\Gamma$ versus harmonic rank and modulation waveform: $\bullet$ are for
sinus, $\blacksquare$ for square and $\blacklozenge$ for triangle modulations.
Obtaining different $\Gamma$ values from one rank to another is deeply
counter-intuitive, as is the fact that even and odd harmonics seem to behave
identically. Actually, if only the material tested in the near-field matters,
the reflection should be expressed from the dielectric constant $\epsilon$ of
the material as predicted by the Fresnel coefficients [32],
$\Gamma=(\epsilon-1)/(\epsilon+1)$. For instance with sinusoidal modulation at
rank 1, it leads to $\epsilon=2.863+j1.581$, _i.e._ a refractive index of
$1.751+j0.451$, which is far from the tabulated value of $3.606+j0.0012$ at
300 GHz [29].
Therefore, the quantities determined here must not be regarded as reflection
coefficients in the usual sense of microwaves and it is not realistic to
extract any $\epsilon$ from them. Although we use the oscillation of the
probe-to-sample distance, which concentrates the useful signal on the near-
field, we had already shown by approach curves with these bow-tie probes that
a longer distance interaction also exists [24], which certainly enters here in
the final quantitative result. The probable origin of this interaction was
identified and attributed to the mandatory openings between the two triangles
constituting the bow-tie, and we had also noticed that such a long-distance
contribution may be favored if a large dielectric support is used to support
the two triangles. For the present probe all precautions had been taken to
minimize these effects. Nonetheless, the quantities extracted are not
reflection coefficients even if we used the usual $\Gamma$ notation of
microwaves. They seem however characteristic of the materials placed in the
near-field of our bow-tie probe and are obtained with a very good accuracy.
## V Measurement of subharmonic IQ mixer ideality
Since the $\zeta$ factor is a measure of the ideality of our instrumentation,
it is worth characterizing it over the entire available bandwidth. Under
conditions of sinusoidal modulation with $h=10$ µm and $\delta h=20$ µm and
detection on the fundamental frequency, we therefore measured both our gold
mirror reference and our GaAs sample from 55 to 65 GHz with a 10 MHz stepsize.
We then proceed to the extraction defined in §III to obtain $\zeta$ and
$\Gamma$ using a 120 MHz sliding window which allows the application of Eqs.
(5-7).
Figure 10 shows the evolution of $\zeta$ as a function of frequency over the
entire operating band of the subharmonic IQ mixer. Obviously, $\zeta$ varies a
lot in the whole band with really high values above 61 GHz which exceed unity
around 64GHz. This is in full agreement with the observation of a drastic drop
of the measured $\widetilde{v}$ at such frequencies for both gold and GaAs. It
is attributed to the conjunction of a power drop of the two multipliers,
especially the one dedicated to the LO which provides 3dBm less than below,
and an increase in the conversion losses of the subharmonic IQ mixer, leading
to an almost $\widetilde{v}$ cancellation at lock-in outputs. Figure 10 shows
that the best frequency range for our system is around 59 GHz, but if we
assume a maximum value of $\approx 0.4$ it is still quite usable from about 56
GHz to 61 GHz. We set out this criterion here because although the frequency
around 60.75 GHz used in §III and §IV is at the limit of this useful band, we
saw that the results were quite satisfactory and accurate.
Figure 10: Ideality factor $\zeta$ versus frequency.
It should be noted here that due to a handling error, the spacing between the
probe tips has been modified since the recordings in §III. If this spacing was
previously 11 µm, it is now 15 µm due to a too abrupt touch that exceeded the
tungsten elasticity limits. Nevertheless the probe has been repaired and kept
its global shape with only this new spacing parameter changed. A fundamental
consequence of this probe crash and repair is that $\zeta$ is independent of
this spacing since we obtain here in Fig. 10 $\zeta=0.433\pm 0.024$ at 60.75
GHz whereas it was determined in §IV at $\zeta=0.436\pm 0.005$ (see Table I).
The parameter $\zeta$ is therefore a characteristic that is essentially
associated with the mixer imbalance but which nevertheless includes any
possible variations in LO power that affect its operating point.
The determination of $\zeta$ over the whole frequency band also yields
$\Gamma$ thanks to Eqs. (7). It is plotted in Fig. 11 where frequencies
leading to $\zeta\geq 1$, for instance around 64 GHz, have been rejected. If
$\Gamma$ really represented the reflection coefficient only related to the
interaction of a dipole with a homogeneous material, we should expect a value
corresponding to the Fresnel reflection coefficient,
$\Gamma=(\epsilon-1)/(\epsilon+1)$, which should be constant since both gold
and GaAs are not dispersive at mmW. This is not the case here, even if in the
whole range 56-60 GHz the values vary very little. Since, on the other hand,
$\zeta$ encompasses all the imperfections of the experiment involving the
subharmonic mixer and its LO, this variation can only be attributed here to
the interaction of the probe and the sample which does not seem to follow a
simple model. In truth this is confirmed by the fact that we have here at
60.75 GHz $\Gamma=(0.400+j0.152)\pm 0.014$ instead of
$\Gamma=(0.482+j0.225)\pm 0.004$ found previously before the probe was
modified.
Figure 11: Real and imaginary parts of the normalized reflection $\Gamma$
versus frequency.
We can therefore conclude that our near-field measurement remains dependent on
the probe used, even if the uniqueness of $\zeta$ has shown that all other
sources of variability related to the experiment have been compensated for.
This is clearly a drawback as we cannot trace back to universal material
characteristics but it is probably still possible to distinguish near-field
materials after learning with each probe used. In future work, we will try to
demonstrate and illustrate this behavior with other probes and materials.
## VI Conclusion
We presented the realization of a new measurement bench for near-field
reflectometry in mmW. A detailed analysis of the operation of this bench has
been carried out, making it possible to take advantage of the electrical
length that exists between the test and measurement paths. We illustrate its
use to extract the near-field contribution that is independent of the mmW
board from measurements distorted by systematic errors. A spin-off of this
theoretical analysis of the operation of the experiment is the definition of a
coefficient of ideality $\zeta$ to which the IQ mixer imbalance and the gain
variations of the mmW sources mainly contribute.
We applied this technique to the near-field measurement of a GaAs substrate
and a gold mirror. At fixed frequency the study shows a result that is
substantially independent of the waveform of the mechanical modulation but not
of the order of the harmonic used for detection. Curiously the even and odd
harmonics lead to two groups of results that are quite consistent with each
other but remain unexplained.
A study of our measurement system over the entire 55-65 GHz band has allowed
us to qualify $\zeta$. For reasons related to our equipment, the latter one
turns out to be rather bad in high frequency but allows very reliable
measurements between 56 and 61 GHz. Moreover, $\zeta$ appears independent of
the probe, which is not the case of the normalized vectorial voltage $\Gamma$
measured in the near-field. It is clear that our probe does not have a trivial
interaction with the sample and is responsible for this unexpected variation.
In the past, we had already observed similar effects with a long range
detection effect from the lateral radiation of our probes [24]. However, here
we have taken great care to minimize it by eliminating any protruding
dielectric material used for mounting the probe. The values obtained for
$\Gamma$ and its evolution with frequency cannot be explained with a simple
model of the probe-sample interaction. Linking $\Gamma$ to the material
parameters of the sample under test in the near-field will therefore require a
further study, which can certainly only be carried out through full 3D
electromagnetic modeling of the coupled probe-sample system.
Experimentally, the deduced $\Gamma$ values remain reproducible and therefore
characteristic of the material for a given probe. In future work, they will be
used to distinguish materials in the near field, which probably implies prior
calibration and learning procedures. The sub-wavelength spatial resolution
that has been demonstrated for intensity measurement should also be studied in
this context.
## References
* [1] X. Chen, D. Hu, R. Mescall, G. You, D. Basov, Q. Dai, and M. Liu, “Modern scattering-type scanning near-field optical microscopy for advanced material research,” Advanced Mater., vol. 31, no. 24, p. 1804774, 2019.
* [2] A. Imtiaz, T. M. Wallis, and P. Kabos, “Near-field scanning microwave microscopy,” IEEE Microw. Mag., pp. 52–64, 2014.
* [3] A. J. L. Adam, “Review of near-field terahertz measurement methods and their applications,” Int. J. Infrared Millim. Waves, vol. 32, pp. 976–1019, 2011.
* [4] S. Kharkovsky and R. Zoughi, “Microwave and millimeter wave nondestructive testing and evaluation–Overview and recent advances,” IEEE Instrumentation & Measurement Magazine, vol. 10, no. 2, pp. 26–38, 2007.
* [5] C. Plassard, E. Bourillot, J. Rossignol, Y. Lacroute, E. Lepleux, L. Pacheco, and E. Lesniewska, “Detection of defects buried in metallic samples by scanning microwave microscopy,” Phys. Rev. B, vol. 83, no. 12, p. 121409, 2011.
* [6] K. Haddadi, D. Glay, and T. Lasri, “A 60 GHz scanning near-field microscope with high spatial resolution sub-surface imaging,” IEEE Microw. Wirel. Compon. Lett., vol. 21, no. 11, pp. 625–627, 2011.
* [7] K. Moon, H. Park, J. Kim, Y. Do, S. Lee, G. Lee, H. Kang, and H. Han, “Sub-surface nanoimaging by broadband terahertz pulse near-field microscopy,” Nano Lett., vol. 15, pp. 549–552, 2014.
* [8] C. Gao and X.-D. Xiang, “Quantitative microwave near-field microscopy of dielectric properties,” Review Scientific Instruments, vol. 69, no. 11, pp. 3846–3851, 1998.
* [9] A. Imtiaz, T. M. Wallis, S.-H. Lim, H. Tanbakuchi, H.-P. Huber, A. Hornung, P. Hinterdorfer, J. Smoliner, F. Kienberger, and P. Kabos, “Frequency-selective contrast on variably doped p-type silicon with a scanning microwave microscope,” J. Appl. Phys., vol. 111, no. 9, p. 093727, 2012.
* [10] T. Monti, O. B. Udoudo, K. A. Sperin, C. Dodds, S. W. Kingman, and T. J. Jackson, “Statistical description of inhomogeneous samples by scanning microwave microscopy,” IEEE Transactions on Microw. Theory Techniques, vol. 65, pp. 2162–2170, June 2017.
* [11] H. Huber, I. Humer, M. Hochleitner, M. Fenner, M. Moertelmaier, C. Rankl, A. Imtiaz, T. Wallis, H. Tanbakuchi, P. Hinterdorfer, et al., “Calibrated nanoscale dopant profiling using a scanning microwave microscope,” J. Appl. Phys., vol. 111, no. 1, p. 014301, 2012.
* [12] S. Berweger, J. C. Weber, J. John, J. M. Velazquez, A. Pieterick, N. A. Sanford, A. V. Davydov, B. Brunschwig, N. S. Lewis, T. M. Wallis, et al., “Microwave near-field imaging of two-dimensional semiconductors,” Nano Letters, vol. 15, no. 2, pp. 1122–1127, 2015.
* [13] M. Horibe, S. Kon, and I. Hirano, “Measurement capability of scanning microwave microscopy: Measurement sensitivity versus accuracy,” IEEE Transactions on Instrumentation Measurement, vol. 68, pp. 1774–1780, June 2019.
* [14] A. Oladipo, M. Kasper, S. Lavdas, G. Gramse, F. Kienberger, and N. Panoiu, “Three-dimensional finite-element simulations of a scanning microwave microscope cantilever for imaging at the nanoscale,” Appl. Phys. Letters, vol. 103, no. 21, p. 213106, 2013\.
* [15] Z. Wei, Y. Cui, E. Y. Ma, S. Johnston, Y. Yang, R. Chen, M. Kelly, Z. Shen, and X. Chen, “Quantitative theory for probe-sample interaction with inhomogeneous perturbation in near-field scanning microwave microscopy,” IEEE Transactions on Microw. Theory Techniques, vol. 64, pp. 1402–1408, May 2016.
* [16] B.-Y. Wu, X.-Q. Sheng, R. Fabregas, and Y. Hao, “Full-wave modeling of broadband near field scanning microwave microscopy,” Scientific Reports, vol. 7, no. 1, p. 16064, 2017.
* [17] G. Dai, G. Geng, X. Zhang, J. Wang, T. Chang, and H. Cui, “W-band near-field microscope,” IEEE Access, vol. 7, pp. 48060–48067, 2019.
* [18] B. Rosner, J. Peck, and D. van der Weide, “Near-field antennas integrated with scanning probes for thz to visible microscopy: scale modeling and limitations on performance,” IEEE Transactions on Antennas Propag., vol. 50, pp. 670–675, May 2002.
* [19] L. Novotny and N. van Hulst, “Antennas for light,” Nat. Photonics, vol. 5, pp. 83–90, 2011.
* [20] T. Grosjean, M. Mivelle, F. I. Baida, G. W. Burr, and U. C. Fischer, “Diabolo nanoantenna for enhancing and confining the magnetic optical field,” Nano Letters, vol. 11, no. 3, pp. 1009–1013, 2011.
* [21] T. Nozokido, J. Bae, and K. Mizuno, “Scanning near-field millimeter-wave microscopy using a metal slit as a scanning probe,” IEEE Transactions on Microw. Theory Techniques, vol. 49, pp. 491–499, March 2001.
* [22] J.-P. Guillet, L. Chusseau, R. Adam, T. Grosjean, A. Penarier, F. Baida, and D. Charraut, “Continuous-wave scanning terahertz near-field microscope,” Microw. Optical Technology Letters, vol. 53, pp. 580–582, 2011.
* [23] R. Omarouayache, P. Payet, J. Raoult, and L. Chusseau, “Millimeter-wave near-field imaging with bow-tie antennas,” Opt. Express, vol. 23, pp. 12144–12151, May 2015.
* [24] L. Chusseau, P. Payet, and J. Raoult, “Optimization of near-field image capture with millimeter-wave bow-tie probes,” IEEE Transactions on Instrumentation Measurement, vol. 66, pp. 61–68, Jan 2017\.
* [25] H. G. von Ribbeck, M. Brehm, D. W. van der Weide, S. Winnerl, O. Drachenko, M. Helm, and F. Keilmann, “Spectroscopic THz near-field microscope,” Opt. Express, vol. 16, no. 5, pp. 3430–3438, 2008.
* [26] F. Keilmann, A. J. Huber, and R. Hillenbrand, “Nanoscale conductivity contrast by scattering-type near-field optical microscopy in the visible, infrared and thz domains,” J. Infrared, Millimeter, Terahertz Waves, vol. 30, no. 12, pp. 1255–1268, 2009.
* [27] M. C. Giordano, S. Mastel, C. Liewald, L. L. Columbo, M. Brambilla, L. Viti, A. Politano, K. Zhang, L. Li, A. G. Davies, E. H. Linfield, R. Hillenbrand, F. Keilmann, G. Scamarcio, and M. S. Vitiello, “Phase-resolved terahertz self-detection near-field microscopy,” Opt. Express, vol. 26, pp. 18423–18435, Jul 2018.
* [28] C. Liewald, S. Mastel, J. Hesler, A. J. Huber, R. Hillenbrand, and F. Keilmann, “All-electronic terahertz nanoscopy,” Optica, vol. 5, no. 2, pp. 159–163, 2018.
* [29] E. D. Palik, Handbook of optical constants of solids, vol. 3. Academic press, 1998.
* [30] P. Adam, J. Bijeon, G. Viardot, and P. Royer, “Analysis of the influence of the tip vibration in the formation of images in apertureless scanning near-field optical microscopy,” Opt. Commun., vol. 174, no. 1–4, pp. 91–98, 2000.
* [31] B. Knoll and F. Keilmann, “Enhanced dielectric contrast in scattering-type scanning near-field optical microscopy,” Opt. Commun., vol. 182, no. 4, pp. 321–328, 2000.
* [32] L. Novotny and B. Hecht, Principles of Nano-Optics. Cambridge: Cambridge University Press, 2006.
|
# Dimensional Reduction
and (Anti) de Sitter Bounds
###### Abstract
Dimensional reduction has proven to be a surprisingly powerful tool for
delineating the boundary between the string landscape and the swampland.
Bounds from the Weak Gravity Conjecture and the Repulsive Force Conjecture,
for instance, are exactly preserved under dimensional reduction. Motivated by
its success in these cases, we apply a similar dimensional reduction analysis
to bounds on the gradient of the scalar field potential $V$ and the mass scale
$m$ of a tower of light particles in terms of the cosmological constant
$\Lambda$, which ideally may pin down ambiguous $O(1)$ constants appearing in
the de Sitter Conjecture and the (Anti) de Sitter Distance Conjecture,
respectively. We find that this analysis distinguishes the bounds $|\nabla
V|/V\geq\sqrt{4/(d-2)}$, $m\lesssim|\Lambda|^{1/2}$, and
$m\lesssim|\Lambda|^{1/d}$ in $d$-dimensional Planck units. The first of these
bounds precludes accelerated expansion of the universe in Einstein-dilaton
gravity and is almost certainly violated in our universe, though it may apply
in asymptotic limits of scalar field space. The second bound cannot be
satisfied in our universe, though it is saturated in supersymmetric AdS vacua
with well-understood uplifts to 10d/11d supergravity. The third bound likely
has a limited range of validity in quantum gravity as well, so it may or may
not apply to our universe. However, if it does apply, it suggests a possible
relation between the cosmological constant and the neutrino mass, which (by
the see-saw mechanism) may further provide a relation between the cosmological
constant problem and the hierarchy problem. We also work out the conditions
for eternal inflation in general spacetime dimensions, and we comment on the
behavior of these conditions under dimensional reduction.
###### Contents
1. 1 Introduction
2. 2 Review: Dimensional Reduction and Weak Gravity Conjectures
3. 3 Light Particles and the Cosmological Constant
4. 4 Interlude: Eternal Inflation in Higher Dimensions
5. 5 Derivatives of Scalar Field Potentials
6. 6 Conclusions
## 1 Introduction
The search for universal features of quantum gravity–also known as the
swampland program [1, 2]–has seen a resurgence in recent years. Strong
evidence has been given in favor of certain conjectured properties of quantum
gravities (so-called “swampland conjectures”), some longstanding conjectures
have been discarded as counterexamples have emerged, and many seemingly-
unrelated aspects of physics and mathematics have been connected through an
ever-growing swampland web.
But the swampland program also faces some serious difficulties. First and
foremost is the infamous swampland tradeoff: conjectures that tend to have
more evidence in their favor tend to have less interesting applications to
cosmology and phenomenology, whereas conjectures that have significant
implications for observable physics tend to be more speculative. Related to
this is a lack of precision: many swampland statements are plagued by
squiggles ($\sim$, $\lesssim$, $\gtrsim$) and $O(1)$ constants to be
determined later. These squiggles make the conjectures difficult to test, and
they suggest that we may not yet understand the underlying physics responsible
for such conjectures. Without this, it is difficult to say when exactly the
conjectures may be expected to hold, and when they may cease to be valid.
The goal of this paper is to sharpen some of these more speculative swampland
conjectures, primarily through a mechanism that has successfully sharpened
swampland conjectures in the past: dimensional reduction. As we will see in
Section 2 below, $O(1)$ factors in the Weak Gravity Conjecture (WGC) bound and
the Repulsive Force Conjecture (RFC) bound can be fixed precisely by demanding
that these bounds should be exactly preserved under dimensional reduction from
$D$ to $d=D-1$ dimensions: the bound will be saturated in $d$ dimensions if
and only if it is saturated in $D$ dimensions. In these cases, dimensional
reduction succeeds in selecting bounds that are both physically meaningful and
seemingly-universal: all known quantum gravities satisfy both the WGC and the
RFC, and compelling evidence has been provided that these conjectures should
hold more generally.
Of course, this does not prove that _any_ bound that is preserved under
dimensional reduction is necessarily a universal constraint on quantum
gravity. But the success of dimensional reduction in the above cases suggests
that it may serve as a useful diagnostic for distinguishing certain $O(1)$
values in conjectured bounds when previous arguments have left us with
ambiguity.
Relatedly, this paper will not address the ultimate question of whether
certain proposed conjectures are universally true in quantum gravity. It seems
very likely that many of these conjectures are true within a limited domain of
validity. For instance, the relationship between the masses of a tower of
light particles and the cosmological constant proposed in the AdS Distance
Conjecture [3] is likely true in infinite families of quantum gravities, each
member of which is distinguished by some discrete flux parameter. The lower
bound on the gradient of a scalar field potential proposed in the de Sitter
Conjecture [4] is likely true in asymptotic regions of scalar field space, and
relatedly eternal inflation is likely forbidden in such regions [5]. The
bounds we will obtain in this paper do not necessarily apply to all quantum
gravities–in fact, some of them are violated even in our own universe–but they
may point us to interesting, universal behavior in certain limits of quantum
gravity. At the least, they give us a mark to aim for and to explore further.
The main results and structure of our paper is as follows: in Section 2, we
will review how dimensional reduction distinguishes the WGC and RFC bounds,
fixing their $O(1)$ factors.
In Section 3, we will apply a similar dimensional reduction analysis to bounds
on the masses of light particles in terms of the cosmological constant of the
form
$m\lesssim|\Lambda_{d}|^{\alpha_{d}}M_{d}^{1-d\alpha_{d}}\,,$ (1.1)
where $\Lambda_{d}$ and $M_{d}$ are the cosmological constant and the Planck
mass in $d$ dimensions, respectively. In [3], it was conjectured that a tower
of light particles should satisfy this bound in any AdS (and perhaps dS)
vacuum. We will see that dimensional reduction picks out $\alpha_{d}=1/2$ and
$\alpha_{d}=1/d$ as special values of $\alpha_{d}$, depending on whether or
not we include the contribution from 1-loop Casimir effects. The former value
is saturated in known supersymmetric compactifications [3] and was previously
singled out in [6], whereas the latter is interesting from the perspective of
the standard model, since neutrinos have a mass of roughly
$m_{\nu}\sim\Lambda_{4}^{1/4}$. If this mass comes from the see-saw mechanism,
$m_{\nu}\sim y^{2}v^{2}/\Lambda_{\text{UV}}$, where $y$ is the Yukawa
coupling, $v$ is the Higgs vev, and $\Lambda_{\text{UV}}$ is some UV scale,
this bound further becomes
$v^{2}\lesssim\Lambda_{4}^{1/4}\Lambda_{\text{UV}}/y\,,$ (1.2)
which bounds the Higgs vev in terms of the cosmological constant and therefore
relates the hierarchy problem to the cosmological constant problem. In
particular, if $\Lambda_{\text{UV}}\sim M_{\text{GUT}}$, $y\sim 0.1$, then
(1.2) becomes $v\lesssim 100$ GeV, which is not far from the measured value
$v=246$ GeV. (Note that we have swept a number of $O(1)$ factors here under
the rug.)
In Section 4, we will derive constraints for eternal inflation in $d\geq 4$
spacetime dimensions by solving the associated Fokker-Planck equation
analytically, generalizing the $d=4$ result of [5]. We will find that eternal
stochastic inflation occurs only if the following conditions are satisfied on
the first and second derivatives of the potential:
$\displaystyle\frac{|\nabla V|^{2}}{V^{(d+2)/2}}$
$\displaystyle<\frac{2(d-1)^{3}}{\pi\Omega_{d-2}}\left(\frac{2}{(d-1)(d-2)M_{d}^{d-2}}\right)^{(d+2)/2}\,,$
(1.3) $\displaystyle\frac{V^{\prime\prime}}{V}$
$\displaystyle>-\frac{2(d-1)}{d-2}\frac{1}{M_{d}^{d-2}}\,,$ (1.4)
where $\Omega_{d-2}={2\pi^{(d-1)/2}/\Gamma(\frac{d-1}{2})}$. In multifield
landscapes, $V^{\prime\prime}$ above should be replaced by a sum over the
negative eigenvalues of the Hessian of the potential.
In Section 5, we will consider bounds of the form
$\frac{|\nabla V_{d}|}{V_{d}^{\gamma_{d}}}\geq
c_{d}M_{d}^{d(1-\gamma_{d})-(d-2)/2}\,.$ (1.5)
Preservation under dimensional reduction distinguishes $\gamma_{d}=1$ as a
special value of $\gamma_{d}$, in agreement with the de Sitter Conjecture
(dSC) bound of [4] and the Trans-Planckian Censorship Conjecture (TCC) bound.
It further selects $c_{d}=\sqrt{4/(d-2)}$ as a special value of $c_{d}$, in
which case (1.5) corresponds exactly to the condition required to avoid
accelerated expansion of the universe. Since this particular value of $c_{d}$
is larger than the coefficient $c_{d}=2/\sqrt{(d-1)(d-2)}$ that appears in the
TCC bound, we learn that a theory satisfying the TCC bound will still satisfy
the TCC bound after dimensional reduction. Likewise, since our value
$\gamma_{d}=1$ is smaller than the value $\gamma_{d}=(d+2)/4$ appearing in the
eternal inflation bound (1.3), we see that the dimensionally reduced theory
will not lead to eternal inflation in asymptotic regions of scalar field
space; eternal inflation can occur only if the radion is stabilized.
We end in Section 6 with conclusions and directions for future research.
## 2 Review: Dimensional Reduction and Weak Gravity Conjectures
In this section, we review how the Weak Gravity Conjecture (WGC) [7] bound and
the Repulsive Force Conjecture (RFC) [8] bounds are preserved under
dimensional reduction.
Our starting point is an Einstein-Maxwell-dilaton action for a $P$-form gauge
field $A_{\mu_{1}\ldots\mu_{P}}$ in $D=d+1$ dimensions:
$S=\frac{1}{2\kappa_{D}^{2}}\int\mathrm{d}^{D}x\sqrt{-g}\left({\cal
R}_{D}-\frac{1}{2}(\nabla\phi)^{2}\right)-\frac{1}{2e_{P;D}^{2}}\int\mathrm{d}^{D}x\sqrt{-g}\mathrm{e}^{-\alpha_{P;D}\phi}F_{P+1}^{2}\,.$
(2.1)
Here, $F_{P+1}=\mathrm{d}A_{P}$ is the field strength for $P$-form gauge
field, and
$F_{q}^{2}\mathrel{:=}\frac{1}{q!}F_{\mu_{1}\ldots\mu_{q}}F^{\mu_{1}\ldots\mu_{q}}\,.$
(2.2)
Note that the action (2.1) is rather special and does not describe the low-
energy theory of a generic quantum gravity, but it will suffice for our
purposes to restrict our attention to this case.
We define
$\frac{1}{\kappa_{D}^{2}}\mathrel{:=}M_{D}^{D-2},$ (2.3)
where $M_{D}$ is the reduced Planck mass in $D$ dimensions. With this
convention, the WGC bound for a $(P-1)$-brane of quantized charge $q$ and
tension $T_{P}$ is given by:
$e_{P;D}^{2}q^{2}M_{D}^{D-2}\geq\left[\frac{\alpha_{P;D}^{2}}{2}+\frac{P(D-P-2)}{D-2}\right]T_{P}^{2}\mathrel{:=}\gamma^{-1}_{P;D}(\alpha)T_{P}^{2}\,.$
(2.4)
The RFC bound in this theory is given by
$e_{P;D}^{2}q^{2}M_{D}^{D-2}\geq
2(\partial_{\phi}T_{P})^{2}+\frac{P(D-P-2)}{D-2}T_{P}^{2}\,,$ (2.5)
where $\partial_{\phi}T_{P}\mathrel{:=}\partial T_{P}/\partial\phi$ is the
partial derivative of the brane tension with respect to the dilaton $\phi$,
holding the Planck scale fixed. A $(P-1)$-brane that satisfies the WGC bound
is said to be superextremal: its charge-to-mass ratio is greater than that of
an extremal black brane. A $(P-1)$-brane that satisfies the RFC bound is self-
repulsive: two such parallel branes separated by a parametrically large
distance will feel a repulsive force between them.
We now show how these bounds are preserved under dimensional reduction. We
consider a dimensional reduction ansatz of the form,
$\mathrm{d}s^{2}=\mathrm{e}^{\frac{\lambda(x)}{d-2}}\mathrm{d}{\hat{s}}^{2}(x)+\mathrm{e}^{-\lambda(x)}\mathrm{d}y^{2},$
(2.6)
where $y\sim y+2\pi R$. This ansatz is chosen so that the dimensionally
reduced action is in Einstein frame, i.e., there is no kinetic mixing between
$\lambda$ and the $d$-dimensional metric. For simplicity, we do not include a
Kaluza-Klein photon, but we do include a massless radion $\lambda(x)$, which
controls the radius of the circle. Under such a dimensional reduction, the
Einstein-Hilbert term reduces as
$\frac{1}{2\kappa_{D}^{2}}\int\mathrm{d}^{D}x\sqrt{-g}{\cal
R}_{D}\rightarrow\frac{1}{2\kappa_{d}^{2}}\int\mathrm{d}^{d}x\sqrt{-\hat{g}}{\cal
R}_{d}-\frac{1}{2}\int\mathrm{d}^{d}x\sqrt{-\hat{g}}\,G_{\lambda\lambda}(\nabla\lambda)^{2}\,,$
(2.7)
where
$\displaystyle\frac{1}{\kappa_{d}^{2}}$ $\displaystyle=M_{d}^{d-2}=(2\pi
R)M_{D}^{D-2}\,,$ (2.8) $\displaystyle G^{(d)}_{\lambda\lambda}$
$\displaystyle=\frac{(d-1)}{4\kappa_{d}^{2}(d-2)}=M_{d}^{d-2}\frac{d-1}{4(d-2)}\,.$
(2.9)
The $P$-form in $D$ dimensions gives both a $P$-form and a
$p\mathrel{:=}P-1$-form in $d$ dimensions. The former comes from taking all of
the legs of the $P$-form to lie along noncompact directions, while the latter
comes from taking one of the legs of the $P$-form to wrap the compact circle.
Similarly, a $(P-1)$-brane charged under the $P$-form descends to both a
$(P-1)$-brane and a $(p-1)$-brane, charged under the respective forms. For
simplicity, we present only the former here: the latter computation is
analogous.
Consider then the $P$-form gauge field in $d$ dimensions. The associated gauge
coupling is given by
$e_{P;d}^{2}=\frac{1}{2\pi R}e_{P;D}^{2}\mathrm{e}^{\frac{P\lambda}{d-2}}\,.$
(2.10)
The tension of a $(P-1)$-brane transverse to the compact circle is given by
$T_{P}^{(d)}=\mathrm{e}^{\frac{P\lambda}{2(d-2)}}T_{P}^{(D)}\,.$ (2.11)
Upon reduction, the radion $\lambda$ and dilaton $\phi$ each couple
exponentially to the Maxwell term in the action, as in (2.1). We may then
redefine the dilaton as the linear combination of $\lambda$ and $\phi$ that
couples to the Maxwell term. This effectively shifts the coupling $\alpha$
appearing in the WGC bound to
$\alpha_{P;d}^{2}=\alpha_{P;D}^{2}+\frac{2P^{2}}{(d-1)(d-2)}.$ (2.12)
Plugging this into the WGC bound (2.4) with $D\rightarrow d$, we find
$\displaystyle
e_{P;d}^{2}q^{2}M_{d}^{d-2}\geq\left[\frac{\alpha_{P;d}^{2}}{2}+\frac{P(d-P-2)}{d-2}\right]T_{P}^{2}=\left[\frac{\alpha_{P;D}^{2}}{2}+\frac{P(D-P-2)}{D-2}\right]T_{P}^{2}.$
(2.13)
The right-hand side of this bound matches that of (2.4). This shows that the
WGC bound has been exactly preserved by the dimensional reduction process. An
analogous result holds for the case of decreasing $P\rightarrow P-1$ and for
Kaluza-Klein modes when $P=1$ provided we include a Kaluza-Klein photon in the
dimensional reduction ansatz of (2.6) [9]: in all such cases, the WGC bound is
exactly preserved under reduction.
The RFC bound in $d$ dimensions is then given by
$e_{P;d}^{2}q^{2}\geq\frac{1}{2\pi
R}(\partial_{\phi}T_{P}^{(d)})^{2}+(G^{(d)})^{\lambda\lambda}(\partial_{\lambda}T_{P}^{(d)})^{2}-\frac{P(d-P-2)}{d-2}\frac{\big{(}T_{P}^{(d)}\big{)}^{2}}{M_{d}^{d-2}}.$
(2.14)
Using (2.9) and (2.11), we have
$(G^{(d)})^{\lambda\lambda}(\partial_{\lambda}T_{P}^{(d)})^{2}=\frac{4(d-2)}{M_{d}^{d-2}(d-1)}\left(\frac{P}{2(d-2)}\right)^{2}\big{(}T_{P}^{(d)}\big{)}^{2}=\frac{P^{2}}{(d-1)(d-2)}\frac{\big{(}T_{P}^{(d)}\big{)}^{2}}{M_{d}^{d-2}}\,.$
(2.15)
This combines with the last term in (2.14) to give
$\left[\frac{P^{2}}{(d-1)(d-2)}+\frac{P(d-P-2)}{d-2}\right]\frac{\big{(}T_{P}^{(d)}\big{)}^{2}}{M_{d}^{d-2}}=\frac{P(D-P-2)}{D-2}\frac{\big{(}T_{P}^{(d)}\big{)}^{2}}{M_{d}^{d-2}}\,.$
(2.16)
With this, we see that the RFC bound in $d$ dimensions (2.14) is exactly the
same as the RFC bound in $D$ dimensions (2.5). This result may be generalized
to multiple scalar fields, multiple gauge fields, and general scalar field
couplings. It holds also for the case of $P$ reduced to $p=P-1$ and for
Kaluza-Klein modes when $P=1$ [10]: in all cases, the RFC bound is exactly
preserved under dimensional reduction.
Within the string landscape, there are no known counterexamples to either the
WGC or the RFC. Many examples have been confirmed to satisfy these conjectures
[7, 10, 11, 12, 13, 14, 15, 16, 17, 18], and a number of arguments suggest
that they should hold more generally [19, 20, 21, 22, 23, 24].
The takeaway lesson here is that preservation under dimensional reduction can
be a useful tool for identifying universal, physically meaningful behavior of
quantum gravities. The value of $\gamma_{P;D}$ in (2.4) is distinguished in
that it dictates the extremality bound for charged black branes in the theory
of (2.1), and it is also distinguished in that it is preserved under
dimensional reduction. Similarly, the bound (2.5) is distinguished by the fact
that it dictates the self-repulsiveness of a brane at long distances, and
simultaneously it is distinguished in that it is preserved under dimensional
reduction. Both of these bounds are further distinguished by the fact that
they seem to be satisfied in all quantum gravity theories. We see that, at
least in these two cases, dimensional reduction picks out universal,
physically meaningful constraints on quantum gravities.
## 3 Light Particles and the Cosmological Constant
Suppose that a family of AdS or dS quantum gravities contain a tower of light
particles whose masses satisfy
$m_{n}\lesssim
n|\Lambda_{D}|^{\alpha_{D}}M_{D}^{1-D\alpha_{D}}\,,~{}~{}~{}~{}~{}n\in\mathbb{Z}_{>0}.$
(3.1)
Such towers of light particles occur, for instance, in all known AdS vacua of
string theory, typically arising as Kaluza-Klein modes of some compactified
dimensions. Such towers were discussed in detail in [3], where it was further
conjectured that there exists a universal, $O(1)$ value of $\alpha_{D}$ that
is obeyed in all (A)dS vacua of string theory. This conjecture was called the
(A)dS Distance Conjecture ((A)dSDC).
We now apply a dimensional reduction analysis to the bound (3.1). Our starting
point is the $D$-dimensional action:
$S=\int\mathrm{d}^{D}x\sqrt{-g}\left[\frac{M_{D}^{D-2}}{2}{\cal
R}_{D}-\Lambda_{D}-\frac{1}{2}\sum_{n}\left((\nabla\phi_{n})^{2}+m_{n}^{2}\phi_{n}^{2}\right)\right]\,,$
(3.2)
where $\phi_{n}$ represents the $n$th particle in the tower. We have written
the action as if these particles are scalar fields, but this assumption is not
necessary.
We now perform a dimensional reduction to $d=D-1$ dimensions using the same
ansatz as above:
$\mathrm{d}s^{2}=\mathrm{e}^{\frac{\lambda(x)}{d-2}}\mathrm{d}{\hat{s}}^{2}(x)+\mathrm{e}^{-\lambda(x)}\mathrm{d}y^{2},$
(3.3)
The resulting action in $d$ dimensions takes effectively the same form as
(3.2):
$S=\int\mathrm{d}^{d}x\sqrt{-\hat{g}}\left[\frac{M_{d}^{d-2}}{2}\left({\cal
R}_{d}-\frac{d-1}{4(d-2)}(\nabla\lambda)^{2}\right)-\Lambda_{d}-\frac{1}{2}\sum_{n}\left((\nabla\phi_{n})^{2}+(m_{n}^{(d)})^{2}\phi_{n}^{2}\right)\right]\,,$
(3.4)
At a classical level, the parameters are related by
$M_{d}^{d-2}=(2\pi R)M_{D}^{D-2}\,,~{}~{}~{}~{}\Lambda_{d}=(2\pi
R)\Lambda_{D}\mathrm{e}^{\frac{\lambda}{d-2}}\,,~{}~{}~{}~{}m_{n}^{(d)}=m_{n}^{(D)}\mathrm{e}^{\frac{\lambda}{2(d-2)}}\,.$
(3.5)
There will also be Kaluza-Klein modes, which we have ignored. For simplicity,
we assume here that $\langle\lambda\rangle=0$: we can ensure this by shifting
$\lambda\rightarrow\lambda-\langle\lambda\rangle$, if necessary.
From this, we see that if the masses $m_{n}^{(D)}$ satisfy (3.1) in $D$
dimensions, then the masses $m_{n}^{(d)}$ will satisfy
$m_{n}\lesssim
n|\Lambda_{d}|^{\alpha_{d}}M_{d}^{1-2\alpha_{d}}\,,~{}~{}~{}~{}~{}n\in\mathbb{Z}_{>0}\,,$
(3.6)
provided
$|\Lambda_{d}|^{\alpha_{d}}M_{d}^{1-d\alpha_{d}}\geq|\Lambda_{D}|^{\alpha_{D}}M_{D}^{1-D\alpha_{D}}=|\Lambda_{d}|^{\alpha_{D}}M_{d}^{\frac{d-2}{d-1}(1-D\alpha_{D})}(2\pi
R)^{\frac{1}{D-2}(2\alpha_{D}-1)}\,.$ (3.7)
From this, we see that the value $\alpha_{d}=1/2$ is distinguished by
dimensional reduction: for this particular value, the $R$ dependence cancels,
and the bound (3.1) is exactly preserved under dimensional reduction. As
pointed out in [3], the value $\alpha_{d}=1/2$ is also distinguished
physically in that it is saturated in all known string/M-theory backgrounds
that can be understood as 10 or 11-dimensional solutions. The authors further
conjecture that the bound (3.1) should be satisfied with $\alpha_{d}=1/2$ for
all supersymmetric AdS vacua. This value of $\alpha_{d}$ was also singled out
it [6], where it was connected to the dSC, the TCC, and the Swampland Distance
Conjecture (SDC) [2]. Once again, dimensional reduction has distinguished a
bound of physical interest.
It is worth pointing out that this bound cannot be satisfied in our own
universe, however. A tower of particles of this mass would run in loops,
correct the graviton propagator, and lead gravity to become strongly coupled
at the “species bound” scale $E_{\text{QG}}$ of order
$E_{\text{QG}}\sim
M_{4}/\sqrt{N(E_{\text{QG}})}\,,~{}~{}~{}N(E_{\text{QG}})\sim
E_{\text{QG}}M_{4}/\Lambda_{4}^{1/2}\,.$ (3.8)
This in turn implies
$E_{\text{QG}}\sim\Lambda_{4}^{1/6}M_{4}^{1/3}\sim 0.02\text{ GeV}\,,$ (3.9)
which conflicts with the experimental fact that gravity remains weakly coupled
at energies accessible at the LHC.
Our above analysis takes place entirely at a classical level. As shown in
(3.5), the cosmological constant $\Lambda_{d}$ acquires an exponential
dependence on the radion $\lambda$. If the radion is not stabilized by quantum
effects, the dimensionally reduced theory does not have a vacuum. Suppose that
we now insist that the theory should in fact have a vacuum, so the radion must
be stabilized. This can be done most simply by including the 1-loop Casimir
energy contributions to the radion potential from the light particles. Upon
dimensional reduction to $d$ dimensions, the contribution from a particle of
mass $m$ takes a very simple form in the massless limit ($mR\ll 1$) [25]:
$V_{C}(\lambda)=\mp\frac{2}{(2\pi
R)^{d}\Omega_{d}}\zeta(d+1)\mathrm{e}^{\frac{d(d-1)}{2(d-2)}\lambda}\,,~{}~{}~{}\Omega_{d}=\frac{2\pi^{(d+1)/2}}{\Gamma(\frac{d+1}{2})}\,.$
(3.10)
Here, $\zeta(x)$ is the Riemann zeta function, $\Omega_{d}$ is the volume of
the unit $d$-sphere, and the $+$ sign is for bosons or fermions with
antiperiodic boundary conditions, while the $-$ sign is for fermions with
periodic boundary conditions. The 1-loop Casimir energy for a particle much
heavier than $1/R$ is exponentially suppressed as $\mathrm{e}^{-2\pi mR}$, so
for our purposes, it suffices to consider contributions only from light
particles with $m\lesssim 1/R$. For simplicity, we approximate all such
particles to be massless.
The full potential at 1-loop, including both the classical contribution and
the 1-loop Casimir energy from light particles, is then given by
$V(\lambda)=V_{\Lambda}(\lambda)+V_{C}(\lambda)=(2\pi
R)\Lambda_{D}\mathrm{e}^{\frac{\lambda}{d-2}}-\sum_{n|m_{n}<1/R}(-)^{F_{n}}\frac{2}{(2\pi
R)^{d}\Omega_{d}}\zeta(d+1)\mathrm{e}^{\frac{d(d-1)}{2(d-2)}\lambda}\,.$
(3.11)
We differentiate with respect to $\lambda$ and set this to 0 to find a
critical point:
$0=\partial_{\lambda}V(\lambda)=\frac{1}{d-2}(2\pi
R)\Lambda_{D}\mathrm{e}^{\frac{\lambda}{d-2}}-\frac{d(d-1)}{2(d-2)}\sum_{n|m_{n}<1/R}(-)^{F_{n}}\frac{2}{(2\pi
R)^{d}\Omega_{d}}\zeta(d+1)\mathrm{e}^{\frac{d(d-1)}{2(d-2)}\lambda}\,.$
(3.12)
The exponent of the second term is larger than that of the first term. This
means that in order to find a minimum, the second term must be positive, while
the first is negative. For $\Lambda_{D}<0$, we may therefore obtain a minimum
in $d$ dimensions provided that fermions with periodic boundary conditions
dominate the Casimir energy. For $\Lambda_{D}>0$, on the other hand, we may
find a maximum in $d$ dimensions provided that bosons or fermions with
antiperiodic boundary conditions dominate the Casimir energy. Creating a de
Sitter minimum in this way would require a delicate interplay between bosons
and fermions, and it would require us to go beyond the massless particle
approximation we have employed here.
Let us now suppose that we have a tower of particles of equal spin $F$, with
masses
$m_{n}=nm\,,~{}~{}~{}~{}n\in\mathbb{Z}_{>0}\,,$ (3.13)
so that the sum in (3.12) becomes
$\sum_{n|m_{n}<1/R}(-)^{F_{n}}=\sum_{n=1}^{1/(mR)}(-)^{F}=(-)^{F}\frac{1}{mR}\,.$
(3.14)
In order for the two terms in (3.12) to balance, we must therefore have
(ignoring $O(1)$ factors):
$\frac{1}{mR^{D}}\sim|\Lambda_{d}|\,.$ (3.15)
We further impose that the bound (3.1) must be saturated both before and after
reduction:
$m\sim|\Lambda_{D}|^{\alpha_{D}}M_{D}^{1-2\alpha_{D}}\sim|\Lambda_{d}|^{\alpha_{d}}M_{d}^{1-2\alpha_{d}}\,.$
(3.16)
Putting these together, we may eliminate $R$ and $m$ to find a relation
between $\Lambda_{D}$ and $\Lambda_{d}$, which gives the following relation
between $\alpha_{D}$ and $\alpha_{d}$:
$\alpha_{D}=\alpha_{d}-\frac{2\alpha_{d}^{2}+\alpha_{d}-1}{2\alpha_{d}-d^{2}+3}\,.$
(3.17)
This is a recursive relation between $\alpha_{d}$ and $\alpha_{d+1}$, which
has a 1-parameter family of solutions. One such solution is the value
$\alpha_{d}=1/2$, which we found above from a strictly classical analysis.
Another solution is $\alpha_{d}=1/d$: this value is special in that it is the
smallest possible value of $\alpha_{d}$ consistent with the recursion relation
(3.17) that remains non-negative in the limit $d\rightarrow\infty$. Note that
it is also the unique value of $\alpha_{d}$ for which the Planck scale $M_{d}$
drops out of (3.1), and in the reduction studied above it leads to $m\sim
1/R\sim|\Lambda_{d}|^{1/d}\sim|\Lambda_{D}|^{1/D}$, so the sum over $n$ in
(3.14) runs over an $O(1)$ number of terms.
The bound $m\lesssim|\Lambda_{d}|^{1/d}$ is rather tantalizing, as the current
upper bound on the sum of the masses of the three standard model neutrinos is
$0.3$ eV, coming from a combination of CMB measurements, galaxy surveys, and
Lyman-alpha forest data. The lightest neutrino may in principle be massless,
though there are lower bounds on the differences in the squared-masses of the
neutrinos from observations of neutrino oscillations. This requires that at
least one neutrino must have a mass above about $0.05$ eV. The cosmological
constant in our universe is measured to be
$\Lambda_{4}^{1/4}\simeq 10^{-30}M_{4}\simeq 0.002\text{ eV}\,,$ (3.18)
which is quite close to the neutrino mass scale of roughly $0.01-0.1$ eV. As
discussed in Section 1, the relationship between the neutrino mass and the
cosmological constant implies a relationship between the Higgs vev and the
cosmological constant, if we assume that the neutrino mass comes from the see-
saw mechanism, thereby relating the cosmological constant problem to the
hierarchy problem. This argument is reminiscent of the work of [26, 27, 28].
This application to our own universe comes with an important caveat, however:
our 1-loop calculation produced minima of the radion potential only in the
case of a negative cosmological constant $\Lambda_{d}<0$, whereas our universe
has $\Lambda_{d}>0$, for which our simple analysis yields maxima of the
potential. This could be remedied by moving beyond the massless limit and
considering both fermions and bosons (or fermions with antiperiodic boundary
conditions). Indeed, balancing the 1-loop Casimir energies of particles in the
standard model leads to a vacuum in three dimensions upon dimensional
reduction [25].
## 4 Interlude: Eternal Inflation in Higher Dimensions
We now take a short break from our dimensional reduction analysis to work out
the conditions for eternal inflation in general spacetime dimension $d$ by
solving the Fokker-Planck equation, generalizing the computations of [5] (see
also [29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39]).aaaWe are very thankful to
Liam McAllister for discussions on the computations in this section. In
Section 5 below, we will comment briefly on the behavior of the conditions for
avoiding eternal inflation under dimensional reduction.
We begin from the $d$-dimensional metric
$ds^{2}=-dt^{2}+a(t)^{2}d\vec{x}^{2}.$ (4.1)
and the action
$S=\int
d^{d}x\sqrt{-g}\left[\frac{M_{d}^{d-2}}{2}\mathcal{R}-\frac{1}{2}(\nabla\phi)^{2}-V(\phi)\right].$
(4.2)
For simplicity, we take $\phi$ to be canonically normalized, and at 0th order
we take it to be homogenous so that its spatial derivatives vanish. The
equation of motion for the scalar field is thus
$\ddot{\phi}+(d-1)H\dot{\phi}=-V^{\prime}(\phi)\,,$ (4.3)
where $H=\dot{a}/a$. The stress-energy tensor is given by
$\displaystyle T^{\mu}_{\nu}$
$\displaystyle=-g^{\mu\delta}\frac{2}{\sqrt{-g}}\frac{\delta S}{\delta
g^{\delta\nu}}=\text{diag}\left(-\frac{1}{2}\dot{\phi}^{2}-V(\phi),\frac{1}{2}\dot{\phi}^{2}-V(\phi),...,\frac{1}{2}\dot{\phi}^{2}-V(\phi)\right).$
(4.4)
This is the stress-energy tensor of a perfect fluid of density $\rho$,
pressure $p$, with
$\rho=\frac{1}{2}\dot{\phi}^{2}+V(\phi)\,,~{}~{}~{}p=\frac{1}{2}\dot{\phi}^{2}-V(\phi)\,.$
(4.5)
Einstein’s equations are given by
$G_{\mu\nu}=\kappa_{d}^{2}T_{\mu\nu}=\frac{1}{M_{d}^{d-2}}T_{\mu\nu}\,.$ (4.6)
Here, we have
$\displaystyle G_{00}=\frac{1}{2}(d-1)(d-2)H^{2}=\frac{\rho}{M_{d}^{d-2}}\,,$
(4.7)
$\displaystyle\frac{1}{a^{2}}G_{11}=...=\frac{1}{a^{2}}G_{d-1,d-1}=-(d-2){\ddot{a}\over
a}-\frac{1}{2}(d-2)(d-3)H^{2}=\frac{p}{M_{d}^{d-2}}\,.$ (4.8)
This gives the Friedmann equations:
$H^{2}=\frac{2\rho}{(d-1)(d-2)M_{d}^{d-2}}\,,~{}~{}~{}\frac{\ddot{a}}{a}=-\frac{p}{(d-2)M_{d}^{d-2}}-\frac{(d-3)H^{2}}{2}\,.$
(4.9)
Inflation takes place in the slow-roll regime,
$\dot{\phi}^{2}\ll V(\phi)\,,~{}~{}~{}~{}~{}|\ddot{\phi}|\ll
H|\dot{\phi}|,|V^{\prime}(\phi)|\,,$ (4.10)
in which case the equation of motion for $\phi$ and the first Friedmann
equation become
$(d-1)H\dot{\phi}=-V^{\prime}(\phi)\,,~{}~{}~{}~{}H^{2}=\frac{2V(\phi)}{(d-1)(d-2)M_{d}^{d-2}}\,.$
(4.11)
To study eternal inflation, we further need to incorporate backreaction from
quantum fluctuations of the scalar field. The scalar 2-point function in dSd
takes the form [40]
$\langle\phi^{2}\rangle=\frac{H^{d-1}}{\pi\Omega_{d-2}}t+\ldots\,,~{}~{}~{}~{}~{}~{}\Omega_{d-2}=\frac{2\pi^{(d-1)/2}}{\Gamma(\frac{d-1}{2})}\,.$
(4.12)
where $\ldots$ represents terms that are not linear in $t$. This linear term
encodes the effect of Gaussian quantum fluctuations that exit the horizon,
decohere, and backreact on the classical slow-roll equation of motion (4.11).
This backreaction takes the form
$(d-1)H\dot{\phi}+V^{\prime}(\phi)=N(t)\,,$ (4.13)
where $N(t)$ is a Gaussian noise term, which induces a random walk of the
field $\phi$ in the potential $V(\phi)$. In other words, in an infinitesimal
time $\delta t$, $\phi$ will vary according to
$\delta\phi=-\frac{1}{(d-1)H}V^{\prime}(\phi)\delta t+\delta\phi_{q}(\delta
t)\,,~{}~{}~{}~{}~{}\delta\phi_{q}(\delta
t)\sim\mathcal{N}(0,\frac{H^{d-1}}{\pi\Omega_{d-2}}\delta t)\,,$ (4.14)
where the variance of the normal distribution comes from the coefficient of
the linear term in (4.12).
If we further approximate $H$ as a constant, which is well-justified in the
slow-roll regime, the evolution of the probability distribution $P[\phi,t]$ of
$\phi$ as a function of $t$ is then described by a Fokker-Planck equation [29,
30, 31]:
$\dot{P}[\phi,t]=\frac{1}{2}\left(\frac{H^{d-1}}{\pi\Omega_{d-2}}\right)\partial_{\phi}^{2}P[\phi,t]+\frac{1}{(d-1)H}\partial_{\phi}\Big{(}(\partial_{\phi}V(\phi))P[\phi,t]\Big{)}\,.$
(4.15)
This equation is difficult to solve for a general potential $V(\phi)$, but the
solution takes a simple, analytic form when the potential is linear or
quadratic in $\phi$. In both cases, the solution takes the form of a Gaussian
distribution:
$P[\phi,t]=\frac{1}{\sigma(t)\sqrt{2\pi}}\exp\left[-\frac{(\phi-\mu(t))^{2}}{2\sigma^{2}(t)}\right]\,.$
(4.16)
For a linear potential $V=V_{0}-\alpha\phi$, the parameters $\mu(t)$,
$\sigma^{2}(t)$ are given by
$\mu(t)=\frac{\alpha}{(d-1)H}\,t\,,~{}~{}~{}~{}\sigma^{2}(t)=\frac{H^{d-1}}{\pi\Omega_{d-2}}\,t\,.$
(4.17)
For a quadratic hilltop potential $V=V_{0}-\frac{1}{2}m^{2}\phi^{2}$, the
parameters are instead given by
$\mu(t)=0\,,~{}~{}~{}~{}\sigma^{2}(t)=\frac{(d-1)H^{d}}{2\pi\Omega_{d-2}m^{2}}\left[-1+\exp\left(\frac{2m^{2}}{(d-1)H}\,t\right)\right]\,.$
(4.18)
For the linear potential, inflation occurs if $\phi<\phi_{c}$, where
$\phi_{c}$ is some critical value whose precise value will be unimportant to
us. The probability that $\phi<\phi_{c}$ at time $t$ is given by
$\text{Pr}[\phi>\phi_{c},t]=\int_{-\infty}^{\phi_{c}}d\phi\,P[\phi,t],$ (4.19)
where $P[\phi,t]$ is the probability density function for a Gaussian
distribution, given in (4.16), with mean $\mu(t)$ and variance $\sigma^{2}(t)$
given by (4.17). The result is
$\text{Pr}[\phi>\phi_{c},t]=\frac{1}{2}\text{erfc}\left[\frac{\mu(t)-\phi_{c}}{\sigma(t)\sqrt{2}}\right]=\frac{1}{2}\text{erfc}\left[\frac{\frac{\alpha}{(d-1)H}t-\phi_{c}}{\left(\frac{2H^{d-1}t}{\pi\Omega_{d-2}}\right)^{1/2}}\right],$
(4.20)
with erfc the error function. For large $t$, this error function can be
approximated to leading order as an exponential,
$\text{Pr}[\phi>\phi_{c},t]\sim\exp\left[-\left(\frac{\frac{\alpha}{(d-1)H}t-\phi_{c}}{\left(\frac{2H^{d-1}t}{\pi\Omega_{d-2}}\right)^{1/2}}\right)^{2}\,\right]\sim\exp\left[-\frac{\pi\Omega_{d-2}\alpha^{2}}{2(d-1)^{2}H^{d+1}}t\right],$
(4.21)
This means that the probability of inflation occurring at time $t$ for a fixed
comoving observer decays exponentially with time. On the other hand, the
volume of the inflating region grows exponential in time as $\exp((d-1)Ht)$.
Eternal inflation occurs if this exponential growth beats the exponential
decay, i.e., if
$(d-1)H>\frac{\pi\Omega_{d-2}\alpha^{2}}{2(d-1)^{2}H^{d+1}}\,.$ (4.22)
Substituting $V$ for $H$ using (4.11) and setting $\alpha=\nabla V(\phi)$,
this becomes
$\frac{|\nabla
V|^{2}}{V^{(d+2)/2}}<\frac{2(d-1)^{3}}{\pi\Omega_{d-2}}\left(\frac{2}{(d-1)(d-2)M_{d}^{d-2}}\right)^{(d+2)/2}\,.$
(4.23)
A similar analysis applies to the quadratic hilltop potential. Now, inflation
occurs if $|\phi|<\phi_{c}$, and the probability of this occurring at time $t$
is given by
$\text{Pr}[|\phi|<\phi_{c},t]=\int_{-\phi_{c}}^{\phi_{c}}d\phi\,P[\phi,t]\,,$
(4.24)
The probability density function is a Gaussian with mean and variance given by
(4.18). This gives
$\displaystyle\text{Pr}[|\phi|<\phi_{c},t]=\text{erf}\left[\frac{\phi_{c}-\mu(t)}{\sigma(t)\sqrt{2}}\right]$
$\displaystyle=\text{erf}\left[\left(\frac{\pi\Omega_{d-2}m^{2}\phi_{c}^{2}}{(d-1)H^{d}\left(-1+\exp\left(\frac{2m^{2}}{(d-1)H}\,t\right)\right)}\right)^{1/2}\right]$
$\displaystyle\sim\exp\left[-\frac{m^{2}}{(d-1)H}\,t\right]\,,$ (4.25)
where in the last line we have Taylor-expanded the error function at the
origin. As before, eternal inflation occurs if the exponential expansion of
the universe dominates the exponential decay of (4.25), which is equivalent to
the condition
$(d-1)H>\frac{m^{2}}{(d-1)H}\,.$ (4.26)
Substituting $V$ for $H$ using (4.11) and setting $m^{2}=-V^{\prime\prime}$,
this becomes
$\frac{V^{\prime\prime}}{V}>-\frac{2(d-1)}{d-2}\frac{1}{M_{d}^{d-2}}\,.$
(4.27)
As in [5], this analysis may be generalized straightforwardly to theories with
multiple scalar fields: as long as the potential $V(\phi^{i})$ separates into
a sum $\sum_{i}V_{i}(\phi^{i})$, where each $V_{i}$ is linear or quadratic and
depends only on $\phi^{i}$, the solution to the Fokker-Planck equation will
separate into a product of Gaussian wavepackets. The main upshot of this is
that, at a critical point of a multifield potential, one should replace
$V^{\prime\prime}$ in (4.27) with a sum over the negative eigenvalues of the
Hessian.
To conclude this section, let us compare the bound (4.27) to the analogous
bound in the Refined de Sitter Conjecture (RdSC) [41, 42, 43]:
$\frac{\text{min}(\nabla_{i}\nabla_{j}V)}{V}\leq-\frac{c_{d}^{\prime}}{M_{d}^{d-2}}\,,$
(4.28)
where $c_{d}^{\prime}$ is some $O(1)$ constant, and
$\text{min}(\nabla_{i}\nabla_{j}V)$ is the minimum eigenvalue of the Hessian.
We see here that the RdSC is incompatible with eternal inflation provided
$c_{d}^{\prime}>2(d-1)/(d-2)$. The RdSC bound with this value of
$c_{d}^{\prime}$ seems to be violated in some 4d examples with 10d
supergravity uplifts [44, 45], but these solutions do not lie in the classical
regime of string theory, so they may be modified by stringy effects [46].
Furthermore, even if the RdSC bound with this value of $c_{d}^{\prime}$ is
violated for the smallest eigenvalue of the Hessian, one might still violate
the necessary conditions for eternal inflation upon summing over the other
negative eigenvalues of the Hessian.
## 5 Derivatives of Scalar Field Potentials
Next, we consider a bound on scalar field potentials of the form
$\frac{|\nabla V_{D}|}{V_{D}^{\gamma_{D}}}\geq
c_{D}M_{D}^{D(1-\gamma_{D})-(D-2)/2}\,,$ (5.1)
for $V_{D}>0$, where $\gamma_{D}$ and $c_{D}$ are $O(1)$ constants. Here,
$|\nabla V_{D}|^{2}=G^{ij}\partial_{i}V_{D}\partial_{j}V_{D}$, and the action
takes the form
$\displaystyle S=\int\mathrm{d}^{D}x\sqrt{-g}\left[\frac{M_{D}^{D-2}}{2}{\cal
R}_{D}-\frac{1}{2}G_{ij}(\phi)\nabla\phi_{i}\nabla\phi_{j}-V_{D}(\phi)\right]\,.$
(5.2)
We want to understand how this behaves under dimensional reduction. We take
our usual dimensional reduction ansatz:
$ds^{2}=e^{\lambda/(d-2)}d\hat{s}^{2}+e^{-\lambda}dy^{2}\,.$ (5.3)
The resulting action in $d$ dimensions is given by
$S=\int\mathrm{d}^{d}x\sqrt{-\hat{g}}\left[\frac{M_{d}^{d-2}}{2}\left({\cal
R}_{d}-\frac{d-1}{4(d-2)}(\nabla\lambda)^{2}\right)-\frac{1}{2}G_{ij}^{(d)}(\phi)\nabla\phi_{i}\nabla\phi_{j}-V_{d}(\phi)\right]\,,$
(5.4)
where we have
$\displaystyle M_{d}^{d-2}=(2\pi R)M_{D}^{D-2}\,,~{}~{}~{}G_{ij}^{(d)}=(2\pi
R)G_{ij}^{(D)}\,,~{}~{}~{}V_{d}=V_{D}(2\pi R)e^{\lambda/(d-2)}\,.$ (5.5)
There is also a contribution to $V_{d}$ from the Casimir energy, but for large
$R$, this will be parametrically subdominant to the classical term, and we
will neglect it here.
Thus we have
$\displaystyle|\nabla V_{d}|^{2}$
$\displaystyle=(G^{(d)})^{ij}\partial_{i}V_{d}\partial_{j}V_{d}+\frac{4(d-2)}{d-1}\frac{1}{M_{d}^{d-2}}(\partial_{\lambda}V_{d})^{2}\,,$
$\displaystyle=(2\pi
R)(G^{(D)})^{ij}\partial_{i}V_{D}\partial_{j}V_{D}+\frac{4}{(d-1)(d-2)}\frac{2\pi
R}{M_{D}^{D-2}}V_{D}^{2}\,,$ (5.6) $\displaystyle=(2\pi R)|\nabla
V_{D}|^{2}+\frac{4}{(d-1)(d-2)}\frac{2\pi R}{M_{D}^{D-2}}V_{D}^{2}\,,$
where we have set $\langle\lambda\rangle=0$ after taking the $\lambda$
derivative. This yields
$\frac{|\nabla
V_{d}|^{2}}{V_{d}^{2\gamma_{d}}}M_{d}^{d-2-2d(1-\gamma_{d})}=(2\pi
R)^{\frac{4}{d-2}(\gamma_{d}-1)}M_{D}^{\frac{d-1}{d-2}(2d\gamma_{d}-d-2)}\frac{|\nabla
V_{D}|^{2}+\frac{4}{(d-1)(d-2)}M_{D}^{2-D}V_{D}^{2}}{V_{D}^{2\gamma_{d}}}\,.$
(5.7)
From this, we see that the value $\gamma_{D}=\gamma_{d}=1$ is distinguished by
dimensional reduction: for this particular value, the $R$-dependence cancels,
and if we ignore the $\lambda$ dependence of $V_{d}$, the bound (5.1) is
exactly preserved under dimensional reduction. With $\gamma_{D}=1$, this bound
matches the “de Sitter Conjecture” bound of [4].
Setting $\gamma_{D}=\gamma_{d}=1$, we may further fix the constant $c_{D}$ by
including the $\lambda$ dependence of $V_{d}$. In particular, assuming that
the bound (5.1) is exactly preserved under dimensional reduction, so that
$V_{D}$ saturates the bound in $d$ dimensions precisely when $V_{d}$ saturates
the bound in $d$ dimensions, we have
$c_{d}^{2}=\frac{|\nabla
V_{d}|^{2}}{V_{d}^{2}}M_{d}^{d-2}=\frac{c_{D}^{2}V_{D}^{2}+\frac{4}{(d-1)(d-2)}V_{D}^{2}}{V_{D}^{2}}\,,$
(5.8)
where in the first equality we have assumed that the bound is saturated $d$
dimensions, and in the second we have used (5.6) and assumed that it is
saturated in $D$ dimensions. This gives a recursive relation for $c_{D}$:
$c_{d}^{2}=c_{D}^{2}+\frac{4}{(d-1)(d-2)}\,,$ (5.9)
which is solved by
$c_{d}^{2}=\beta+\frac{4}{d-2}\,,$ (5.10)
where $\beta$ is a free parameter. This leads to the bound
$\frac{|\nabla
V_{d}|}{V_{d}}\geq\sqrt{\beta+\frac{4}{d-2}}\frac{1}{M_{d}^{(d-2)/2}}\,.$
(5.11)
Our dimensional reduction argument leaves $\beta$ unfixed, but the value
$\beta=0$ is distinguished for two reasons: first of all, it is the smallest
value of $\beta$ such that $c_{d}$ remains non-negative in the limit
$d\rightarrow\infty$. Secondly, the bound with $\beta=0$,
$\frac{|\nabla
V_{d}|}{V_{d}}\geq\sqrt{\frac{4}{d-2}}\frac{1}{M_{d}^{(d-2)/2}}\,,$ (5.12)
has physical meaning: it is exactly the condition required for a theory of
Einstein gravity coupled to a dilaton field to avoid accelerated expansion of
the universe in $d$ dimensions [4].
Observational constraints rule out single-field quintessence models satisfying
(5.12) [47]. Even if one entertains the possibility of fine-tuned, multi-field
quintessence models [48, 49, 50], it is very hard to imagine that (5.12) could
be satisfied at the maximum of the Higgs potential [51]. However, this bound
may yet be satisfied in asymptotic limits of scalar field space, as is
evidenced by many examples in string theory. For instance, toroidal
compactifications of $O(16)\times O(16)$ heterotic string theory to $d$
dimensions satisfy the bound [4]:
$\frac{|\nabla
V_{d}|}{V_{d}}\geq\text{min}\left(2\sqrt{\frac{3d-5}{d-2}},\frac{4\sqrt{2}}{\sqrt{(10-d)(d-2)}}\right)\frac{1}{M_{d}^{(d-2)/2}}\,.$
(5.13)
This bound implies (5.12), and it implies (5.11) in $d\geq 3$ provided
$\beta\leq 4/7$.
Next, we may consider the KKLT scenario [52] and the LVS scenario [53]. Both
of these models have de Sitter critical points, so they violate (5.12) in the
interior of scalar field space. However, as noted in [54], the KKLT potential
behaves asymptotically as
$V(\phi)\sim\exp\left(-\sqrt{6}\frac{\phi}{M_{4}}\right)\,,$ (5.14)
and the LVS potential behaves asymptotically as
$V(\phi)\sim\exp\left(-3\sqrt{\frac{3}{2}}\frac{\phi}{M_{4}}\right)\,.$ (5.15)
Thus, in asymptotic limits of field space, both of these satisfy (5.12), and
more generally the former satisfies (5.11) provided $\beta\leq 4$, whereas the
latter requires $\beta\leq 23/2$. More examples of 4d theories satisfying
(5.12) in asymptotic regions of scalar field space can be found in [4, 6, 45,
55, 56].
However, some of the string compactifications considered in [55, 56, 6]
naively seem to violate (5.12). For instance, in Type IIA compactifications in
the presence of O4-planes but no D4-brane sources, [55] derived a lower bound
on $|\nabla V|M_{4}/V$ of $\sqrt{2/3}$, which if saturated would violate
(5.12) while saturating the Transplackian Censorship Conjecture (TCC) bound
[54]:
$\frac{|V^{\prime}|}{V}\geq\frac{2}{\sqrt{(d-1)(d-2)}}\frac{1}{M_{d}^{(d-2)/2}}\,.$
(5.16)
However, there is an important subtlety here: the lower bound in this example
comes from considering only the volume modulus and the dilaton, so it is
plausible that including other scalar fields could lead to consistency with
(5.12) in asymptotic limits of scalar field space. Likewise, the lower bounds
on $|V^{\prime}|M_{4}/V$ listed in Table 2 of [6] come from considering the
derivative $V^{\prime}$ of the potential with respect to the geodesic distance
along certain geodesics in scalar field space, not from considering the
gradient $\nabla V$ of the potential with respect to _all_ scalar fields in
the theory.bbbThe fact that the derivative with respect to geodesic distance
$V^{\prime}$ may saturate the TCC bound (5.16) is crucial for the connection
between the TCC and the Swampland Distance Conjecture (SDC) proposed in [6].
Once contributions to the gradient $\nabla V$ from additional scalar fields
are included, it is plausible that our bound (5.12) may yet be satisfied.cccWe
are very thankful to David Andriot for discussions on these points.
It may sound like wishful thinking to hope that contributions to the gradient
from additional scalar fields will always come to the rescue, ensuring
consistency with (5.12). However, we already know of at least one simple
example where this is precisely what happens: in heterotic string theory
compactified to four dimensions, one has
$|\partial_{\rho}V|M_{4}/V=\sqrt{2/3}$, where $\rho$ is the volume modulus of
the compactification manifold (see Table 2 of [6]). This saturates the TCC
bound (5.16) along a geodesic in the $\rho$ direction and naively violates our
bound (5.12). However, the potential also depends asymptotically on the
dilaton, $|\partial_{\tau}V|M_{4}/V=\sqrt{2}$. This additional contribution to
the gradient leads to consistency with (5.12).
Note, however, that while this example is a useful proof of principle, it does
not represent the typical scenario. In general, it is quite difficult to
determine the dependence of the potential on _every_ scalar field in the
theory, so falsifying (5.12) in asymptotic regions of scalar field space is
not easy to do.††footnotemark:
Finally, note that (5.12) is incompatible with the condition (4.23) needed for
eternal inflation, since $\gamma_{d}=1<\gamma_{d}^{\text{EI}}=(d+2)/4$. This
is rather unsurprising from a physical perspective: (5.12) forbids accelerated
expansion of the universe, which is obviously a prerequisite for eternal
inflation. Dimensional reduction induces an exponential potential for the
radion $\lambda$, $V_{d}=V_{D}\exp(\lambda/(d-2))$, so if the radion is not
stabilized, this exponential potential alone will suffice to violate the
condition (4.23) required for eternal inflation. Radion stabilization will not
occur in asymptotic regions of scalar field space [57], so dimensional
reduction will not lead to eternal inflation in such regions.
## 6 Conclusions
We have seen that dimensional reduction distinguishes particular $O(1)$
constants appearing in the WGC, RFC, (A)dSDC, and dSC. In the case of the
(A)dSDC, one mass scale that emerges is suggestively close to the neutrino
mass scale, which may be related to the Higgs vev via the see-saw mechanism.
It might be worthwhile to search for other swampland-related reasons for a
connection between the neutrino mass/Higgs vev and the cosmological constant.
In the case of the dSC, the bound we obtain precisely forbids accelerated
expansion of the universe in Einstein-dilaton gravity. This is interesting in
that it offers an alternative physical explanation to Trans-Planckian
Censorship [54] as to why the dSC seems to hold universally in asymptotic
limits of scalar field space: perhaps quantum gravity simply forbids
accelerated expansion in such regions. This possibility merits further
investigation.
We have also worked out conditions for eternal inflation in $d\geq 4$
spacetime dimensions, generalizing previous bounds obtained in $d=4$. Models
of inflation in more than four dimensions have not received very much
attention, due to the obvious fact that such models are of little experimental
interest. But given the recent, renewed interest in scalar field potentials
and de Sitter vacua in quantum gravity, it may be worthwhile to temporarily
ignore the question of experimental relevance and instead explore more general
theories of inflation and cosmology in diverse dimensions, as well as their
possible embeddings in string theory. It would be interesting to study de
Sitter critical points of scalar field potentials in string compactifications
to $d>4$ dimensions, to see if these critical points may support eternal
inflation.
We have not answered the most important question: to what extent are the
bounds we have obtained actually satisfied in quantum gravities? It is clear
that many string compactifications satisfy them, but it is also quite possible
that these results suffer from a lamppost effect. Understanding the domain of
validity of these conjectured bounds is a crucial task for the swampland
program, though it may not be an easy one.
## Acknowledgements
We thank David Andriot, Ben Heidenreich, Juan Maldacena, Liam McAllister,
Georges Obied, Eran Palti, Matthew Reece, and Cumrun Vafa for useful
discussions. We thank David Andriot, Ben Heidenreich, Liam McAllister, and
Matthew Reece for comments on a draft of this paper. We acknowledge
hospitality from the 2019 Simons Summer Workshop at the Simons Center for
Geometry and Physics at Stony Brook University, where some of these
discussions took place. The work of TR was supported by NSF grant PHY1820912,
the Simons Foundation, and the Berkeley Center for Theoretical Physics.
## References
* [1] C. Vafa, “The String landscape and the swampland,” arXiv:hep-th/0509212 [hep-th].
* [2] H. Ooguri and C. Vafa, “On the Geometry of the String Landscape and the Swampland,” Nucl. Phys. B766 (2007) 21–33, arXiv:hep-th/0605264 [hep-th].
* [3] D. Lüst, E. Palti, and C. Vafa, “AdS and the Swampland,” Phys. Lett. B 797 (2019) 134867, arXiv:1906.05225 [hep-th].
* [4] G. Obied, H. Ooguri, L. Spodyneiko, and C. Vafa, “De Sitter Space and the Swampland,” arXiv:1806.08362 [hep-th].
* [5] T. Rudelius, “Conditions for (No) Eternal Inflation,” JCAP 08 (2019) 009, arXiv:1905.05198 [hep-th].
* [6] D. Andriot, N. Cribiori, and D. Erkinger, “The web of swampland conjectures and the TCC bound,” JHEP 07 (2020) 162, arXiv:2004.00030 [hep-th].
* [7] N. Arkani-Hamed, L. Motl, A. Nicolis, and C. Vafa, “The String landscape, black holes and gravity as the weakest force,” JHEP 0706 (2007) 060, arXiv:hep-th/0601001 [hep-th].
* [8] E. Palti, “The Weak Gravity Conjecture and Scalar Fields,” JHEP 08 (2017) 034, arXiv:1705.04328 [hep-th].
* [9] B. Heidenreich, M. Reece, and T. Rudelius, “Sharpening the Weak Gravity Conjecture with Dimensional Reduction,” JHEP 02 (2016) 140, arXiv:1509.06374 [hep-th].
* [10] B. Heidenreich, M. Reece, and T. Rudelius, “Repulsive Forces and the Weak Gravity Conjecture,” JHEP 10 (2019) 055, arXiv:1906.02206 [hep-th].
* [11] B. Heidenreich, M. Reece, and T. Rudelius, “Evidence for a sublattice weak gravity conjecture,” JHEP 08 (2017) 025, arXiv:1606.08437 [hep-th].
* [12] S.-J. Lee, W. Lerche, and T. Weigand, “Tensionless Strings and the Weak Gravity Conjecture,” JHEP 10 (2018) 164, arXiv:1808.05958 [hep-th].
* [13] D. Klaewer, S.-J. Lee, T. Weigand, and M. Wiesner, “Quantum Corrections in 4d N=1 Infinite Distance Limits and the Weak Gravity Conjecture,” arXiv:2011.00024 [hep-th].
* [14] S.-J. Lee, W. Lerche, and T. Weigand, “Modular Fluxes, Elliptic Genera, and Weak Gravity Conjectures in Four Dimensions,” arXiv:1901.08065 [hep-th].
* [15] S.-J. Lee, W. Lerche, and T. Weigand, “A Stringy Test of the Scalar Weak Gravity Conjecture,” Nucl. Phys. B938 (2019) 321–350, arXiv:1810.05169 [hep-th].
* [16] T. W. Grimm and D. van de Heisteeg, “Infinite Distances and the Axion Weak Gravity Conjecture,” arXiv:1905.00901 [hep-th].
* [17] N. Gendler and I. Valenzuela, “Merging the Weak Gravity and Distance Conjectures Using BPS Extremal Black Holes,” arXiv:2004.10768 [hep-th].
* [18] C. F. Cota, A. Klemm, and T. Schimannek, “State counting on fibered CY-3 folds and the non-Abelian Weak Gravity Conjecture,” arXiv:2012.09836 [hep-th].
* [19] C. Cheung and G. N. Remmen, “Infrared Consistency and the Weak Gravity Conjecture,” JHEP 12 (2014) 087, arXiv:1407.7865 [hep-th].
* [20] S. Hod, “A proof of the weak gravity conjecture,” Int. J. Mod. Phys. D26 (2017) 1742004, arXiv:1705.06287 [gr-qc].
* [21] B. Heidenreich, M. Reece, and T. Rudelius, “The Weak Gravity Conjecture and Emergence from an Ultraviolet Cutoff,” Eur. Phys. J. C78 no. 4, (2018) 337, arXiv:1712.01868 [hep-th].
* [22] C. Cheung, J. Liu, and G. N. Remmen, “Proof of the Weak Gravity Conjecture from Black Hole Entropy,” JHEP 10 (2018) 004, arXiv:1801.08546 [hep-th].
* [23] A. Urbano, “Towards a proof of the Weak Gravity Conjecture,” arXiv:1810.05621 [hep-th].
* [24] M. Montero, “A Holographic Derivation of the Weak Gravity Conjecture,” JHEP 03 (2019) 157, arXiv:1812.03978 [hep-th].
* [25] N. Arkani-Hamed, S. Dubovsky, A. Nicolis, and G. Villadoro, “Quantum Horizons of the Standard Model Landscape,” JHEP 06 (2007) 078, arXiv:hep-th/0703067.
* [26] H. Ooguri and C. Vafa, “Non-supersymmetric AdS and the Swampland,” Adv. Theor. Math. Phys. 21 (2017) 1787–1801, arXiv:1610.01533 [hep-th].
* [27] L. E. Ibanez, V. Martin-Lozano, and I. Valenzuela, “Constraining the EW Hierarchy from the Weak Gravity Conjecture,” arXiv:1707.05811 [hep-th].
* [28] L. E. Ibanez, V. Martin-Lozano, and I. Valenzuela, “Constraining Neutrino Masses, the Cosmological Constant and BSM Physics from the Weak Gravity Conjecture,” JHEP 11 (2017) 066, arXiv:1706.05392 [hep-th].
* [29] A. A. Starobinsky, “STOCHASTIC DE SITTER (INFLATIONARY) STAGE IN THE EARLY UNIVERSE,” Lect. Notes Phys. 246 (1986) 107–126.
* [30] S.-J. Rey, “Dynamics of Inflationary Phase Transition,” Nucl. Phys. B284 (1987) 706–728.
* [31] A. D. Linde, “Hard art of the universe creation (stochastic approach to tunneling and baby universe formation),” Nucl. Phys. B372 (1992) 421–442, arXiv:hep-th/9110037 [hep-th].
* [32] P. Creminelli, S. Dubovsky, A. Nicolis, L. Senatore, and M. Zaldarriaga, “The Phase Transition to Slow-roll Eternal Inflation,” JHEP 09 (2008) 036, arXiv:0802.1067 [hep-th].
* [33] G. Barenboim, W.-I. Park, and W. H. Kinney, “Eternal Hilltop Inflation,” JCAP 1605 no. 05, (2016) 030, arXiv:1601.08140 [astro-ph.CO].
* [34] W. H. Kinney, “Eternal Inflation and the Refined Swampland Conjecture,” arXiv:1811.11698 [astro-ph.CO].
* [35] S. Brahma and S. Shandera, “Stochastic eternal inflation is in the swampland,” arXiv:1904.10979 [hep-th].
* [36] M.-S. Seo, “Eternal inflation in light of Wheeler-DeWitt equation,” JCAP 11 (2020) 007, arXiv:2007.03866 [hep-th].
* [37] A. Bedroya, M. Montero, C. Vafa, and I. Valenzuela, “de Sitter Bubbles and the Swampland,” arXiv:2008.07555 [hep-th].
* [38] R. Blumenhagen, C. Kneissl, and A. Makridou, “De Sitter Quantum Breaking, Swampland Conjectures and Thermal Strings,” arXiv:2011.13956 [hep-th].
* [39] J. Chojnacki, J. Krajecka, J. H. Kwapisz, O. Slowik, and A. Strag, “Is asymptotically safe inflation eternal?,” arXiv:2101.00866 [gr-qc].
* [40] S. Dubovsky, L. Senatore, and G. Villadoro, “Universality of the Volume Bound in Slow-Roll Eternal Inflation,” JHEP 05 (2012) 035, arXiv:1111.1725 [hep-th].
* [41] S. K. Garg and C. Krishnan, “Bounds on Slow Roll and the de Sitter Swampland,” arXiv:1807.05193 [hep-th].
* [42] H. Ooguri, E. Palti, G. Shiu, and C. Vafa, “Distance and de Sitter Conjectures on the Swampland,” arXiv:1810.05506 [hep-th].
* [43] D. Andriot and C. Roupec, “Further refining the de Sitter swampland conjecture,” Fortsch. Phys. 67 no. 1-2, (2019) 1800105, arXiv:1811.08889 [hep-th].
* [44] D. Andriot, P. Marconnet, and T. Wrase, “New de Sitter solutions of 10d type IIB supergravity,” JHEP 08 (2020) 076, arXiv:2005.12930 [hep-th].
* [45] D. Andriot, “Tachyonic de Sitter solutions of 10d type II supergravities,” arXiv:2101.06251 [hep-th].
* [46] D. Andriot, P. Marconnet, and T. Wrase, “Intricacies of classical de Sitter string backgrounds,” Phys. Lett. B 812 (2021) 136015, arXiv:2006.01848 [hep-th].
* [47] P. Agrawal, G. Obied, P. J. Steinhardt, and C. Vafa, “On the Cosmological Implications of the String Swampland,” Phys. Lett. B784 (2018) 271–276, arXiv:1806.09718 [hep-th].
* [48] M. Cicoli, G. Dibitetto, and F. G. Pedro, “Out of the Swampland with Multifield Quintessence?,” JHEP 10 (2020) 035, arXiv:2007.11011 [hep-th].
* [49] Y. Akrami, M. Sasaki, A. R. Solomon, and V. Vardanyan, “Multi-field dark energy: cosmic acceleration on a steep potential,” arXiv:2008.13660 [astro-ph.CO].
* [50] F. Farakos, “On the F-term problem and quintessence supersymmetry breaking,” arXiv:2101.05759 [hep-th].
* [51] F. Denef, A. Hebecker, and T. Wrase, “de Sitter swampland conjecture and the Higgs potential,” Phys. Rev. D98 no. 8, (2018) 086004, arXiv:1807.06581 [hep-th].
* [52] S. Kachru, R. Kallosh, A. D. Linde, and S. P. Trivedi, “De Sitter vacua in string theory,” Phys. Rev. D68 (2003) 046005, arXiv:hep-th/0301240 [hep-th].
* [53] J. P. Conlon, F. Quevedo, and K. Suruliz, “Large-volume flux compactifications: Moduli spectrum and D3/D7 soft supersymmetry breaking,” JHEP 08 (2005) 007, arXiv:hep-th/0505076 [hep-th].
* [54] A. Bedroya and C. Vafa, “Trans-Planckian Censorship and the Swampland,” JHEP 09 (2020) 123, arXiv:1909.11063 [hep-th].
* [55] T. Wrase and M. Zagermann, “On Classical de Sitter Vacua in String Theory,” Fortsch. Phys. 58 (2010) 906–910, arXiv:1003.0029 [hep-th].
* [56] D. Andriot, “Open problems on classical de Sitter solutions,” Fortsch. Phys. 67 no. 7, (2019) 1900026, arXiv:1902.10093 [hep-th].
* [57] M. Dine and N. Seiberg, “Is the Superstring Weakly Coupled?,” Phys. Lett. 162B (1985) 299–302.
|
# The RR Lyrae Delay-Time Distribution: A Novel Perspective on Models of Old
Stellar Populations
Sumit K. Sarbadhicary Department of Physics and Astronomy, Michigan State
University, East Lansing, MI 48824, USA Mairead Heiger Pittsburgh Particle
Physics, Astrophysics and Cosmology Center (PITT PACC), University of
Pittsburgh, 3941 O’Hara St, Pittsburgh, PA 15260, USA Carles Badenes
Pittsburgh Particle Physics, Astrophysics and Cosmology Center (PITT PACC),
University of Pittsburgh, 3941 O’Hara St, Pittsburgh, PA 15260, USA Cecilia
Mateu Departamento de Astronomía, Facultad de Ciencias, Universidad de la
República, Iguá 4225, 14000, Montevideo, Uruguay Jeffrey Newman Pittsburgh
Particle Physics, Astrophysics and Cosmology Center (PITT PACC), University of
Pittsburgh, 3941 O’Hara St, Pittsburgh, PA 15260, USA Robin Ciardullo
Department of Astronomy & Astrophysics, The Pennsylvania State University,
University Park, PA 16802, USA Na’ama Hallakoun Department of Particle
Physics and Astrophysics, Weizmann Institute of Science, Rehovot, 7610001,
Israel Dan Maoz School of Physics and Astronomy, Tel-Aviv University, Tel-
Aviv 6997801, Israel Laura Chomiuk Department of Physics and Astronomy,
Michigan State University, East Lansing, MI 48824, USA
###### Abstract
The delay-time distribution (DTD) is the occurrence rate of a class of objects
as a function of time after a hypothetical burst of star formation. DTDs are
mainly used as a statistical test of stellar evolution scenarios for supernova
progenitors, but they can be applied to many other classes of astronomical
objects. We calculate the first DTD for RR Lyrae variables using 29,810 RR
Lyrae from the OGLE-IV survey and a map of the stellar-age distribution (SAD)
in the Large Magellanic Cloud (LMC). We find that $\sim 46\%$ of the OGLE-IV
RR Lyrae are associated with delay-times older than 8 Gyr (main-sequence
progenitor masses less than 1 M⊙), and consistent with existing constraints on
their ages, but surprisingly about $51\%$ of RR Lyrae appear have delay times
$1.2-8$ Gyr (main-sequence masses between $1-2$ M⊙ at LMC metallicity). This
intermediate-age signal also persists outside the Bar-region where crowding is
less of a concern, and we verified that without this signal, the spatial
distribution of the OGLE-IV RR Lyrae is inconsistent with the SAD map of the
LMC. Since an intermediate-age RR Lyrae channel is in tension with the lack of
RR Lyrae in intermediate-age clusters (noting issues with small-number
statistics), and the age-metallicity constraints of LMC stars, our DTD result
possibly indicates that systematic uncertainties may still exist in SAD
measurements of old-stellar populations, perhaps stemming from the
construction methodology or the stellar evolution models used. We described
tests to further investigate this issue.
RR Lyrae variable stars (1410); Large Magellanic Cloud (903); Stellar
populations (1622); Stellar evolution (1599); Stellar evolutionary models
(2046); Horizontal branch (2048); Stellar ages (1581); Stellar pulsations
(1625); Hertzsprung Russell diagram (725);
## 1 Introduction
A detailed understanding of stellar evolution remains one of the most sought-
after goals in astrophysics. Popular stellar evolution codes such as Geneva
(Schaller et al., 1992; Schaerer et al., 1993), Y2 (Kim et al., 2002; Yi et
al., 2003; Demarque et al., 2004), BaSTI (Pietrinferni et al., 2004, 2006;
Hidalgo et al., 2018), Darthmouth (Dotter et al., 2008), PARSEC (Bressan et
al., 2012; Chen et al., 2014) and MESA (Paxton et al., 2011, 2013, 2015, 2018)
are powerful tools for interpreting observations of stellar populations.
However, many essential topics in stellar evolution are still not well
understood, and/or not properly taken into account in even the most state-of-
the-art models. Examples of such topics include convection, mass loss and mass
transfer, common envelope evolution, and binary interaction. Often, these
three-dimensional phenomena are approximated by simplified parametric models
tuned to specific observables and integrated into one-dimensional stellar
evolution codes. These uncertainties limit our understanding of many important
phases of stellar evolution, such as the horizontal branch, the asymptotic
giant branch (AGB) and post-AGB phase, planetary nebulae, and supernovae (see
discussions in Gallart et al., 2005; Conroy et al., 2009; Conroy, 2013).
The delay-time distribution (DTD) is a promising method for testing stellar
evolution models in complex stellar populations (Maoz & Mannucci, 2012; Maoz
et al., 2014). The DTD is defined as the occurrence rate of a class of
astronomical object as a function of time since a hypothetical brief burst of
star formation; it is equivalent to the impulse response, or Green’s function.
The DTD constrains the evolutionary timescale and formation efficiency of the
object’s progenitors, and theoretical DTDs are common predictions of stellar
population synthesis models (Mennekens et al., 2010; Nelemans et al., 2013;
Toonen et al., 2013; Zapartas et al., 2017). Observationally, DTDs can be
derived from surveys of objects, provided that the stellar age distributions
(SADs) of their host galaxies are measured (Gal-Yam & Maoz, 2004; Totani et
al., 2008; Maoz & Sharon, 2010; Maoz et al., 2011, 2012; Graur et al., 2014;
Maoz & Graur, 2017; Friedmann & Maoz, 2018).
More recently it was shown that DTDs can be a useful stellar evolution
diagnostic in Local Group galaxies with high-quality observations of their
resolved stellar populations (Badenes et al., 2010; Maoz & Badenes, 2010;
Badenes et al., 2015). Using this approach, Badenes et al. (2015) measured the
first DTD for planetary nebula, showing evidence of two distinct populations
of planetary nebula progenitors: one with ages of 5–8 Gyrs, and another with
ages of 35–800 Myrs. The key advantage of a DTD is that it constrains the
evolutionary timescales for the progenitors of the entire population of
objects in a galaxy, taking into account the variety of star-formation
histories that these objects have evolved from. It can be a powerful tool for
identifying or ruling out the presence of specific formation channels,
measuring their efficiency, and identifying physical mechanisms that are not
part of canonical progenitor models.
Figure 1: The spatial distribution of LMC RR Lyrae (red dots) in the OGLE-IV
sample overlaid on the spatial cells (blue) from the Harris & Zaritsky (2009)
SAD map of the galaxy. Figure 2: $I$-band luminosity function of the OGLE-IV
RR Lyrae. The cumulative histograms show the fraction of RR Lyrae dimmer than
a certain $I$-band brightness. The luminosity function of the full RR Lyrae
sample from Figure 1 is shown in black, and that of RR Lyrae in the ‘LMC503’
field of the OGLE-IV survey, a region with dense stellar crowding (Udalski et
al., 2015), is shown in blue. The inset shows the location of the LMC503
region on an $r$-band continuum map (as reference) of the LMC Bar from the
Magellanic Cloud Emission-Line Survey (MCELS, Winkler et al., 2005; Pellegrini
et al., 2012). The $I$-band completeness limit of the LMC503 region is shown
with the dashed line.
In this paper, we will use a SAD map of the Large Magellanic Cloud (LMC) to
derive the DTD of RR Lyrae stars— pulsating horizontal branch stars with
periods between 0.2 and 1 day (Smith, 2004). We chose to test the DTD method
on RR Lyrae for several reasons. Firstly, the sample size of RR Lyrae in the
LMC is quite large (see Section 2.1), allowing us to measure a DTD with high
significance. Secondly, there is strong evidence that RR Lyrae are mostly
ancient stars, older than 10 Gyrs, given their pulsational properties (Smith,
2004; Marconi et al., 2015) and abundance in old globular clusters (Clement et
al., 2001; Soszyński et al., 2014, 2016). Measuring an RR Lyrae DTD therefore
provides a rigorous test of the DTD method for the recovery of progenitor age-
distributions, as well as star-formation history measurements of old resolved
stellar populations. Lastly, a DTD analysis provides an opportunity to test
stellar evolution models of RR Lyrae in a completely new way. While there is
consensus on the interpretation of RR Lyrae as ancient stars, the lower-limit
on their ages has been somewhat unconstrained. For example, the absence of RR
Lyrae in the SMC cluster Lindsey 1 ($t\approx 9$ Gyr), compared to their
presence in NGC 121 ($t\approx$10–11 Gyr), is generally cited as evidence of a
lower limit of 10 Gyr for progenitor age of RR Lyrae (Olszewski et al., 1996;
Glatt et al., 2008). However, growing evidence of thin disk RR Lyrae in the
Milky Way raises questions about whether this limit might be lower, and if an
intermediate-age progenitor channel can exist (Layden, 1995; Zinn et al.,
2019; Prudil et al., 2020). Additionally uncertainties may exist in the
evolutionary models of RR Lyrae stars themselves: since RR Lyrae are
horizontal-branch stars, their positions on the color-magnitude diagram depend
on an unknown combination of factors (commonly known as the second-parameter
phenomenon) like metallicity, age, mass-loss on the red-giant branch, stellar
rotation, core structure, and chemical abundance (see Fusi Pecci & Bellazzini,
1997; Catelan, 2009; Dotter, 2013, for reviews).
The paper is organized as follows: Section 2 briefly describes the two
ingredients for calculating the RR Lyrae DTD – the OGLE-IV survey of RR Lyrae,
and the Harris & Zaritsky (2009) SAD map of LMC, Section 3 describes the
measurement of the RR Lyrae DTD from the OGLE-IV survey, Section 4 checks
whether the measured DTD is consistent with the observed spatial distribution
of RR Lyrae in the OGLE-IV survey, incompleteness in the SAD map and RR Lyrae
in star clusters, Section 5 discusses the two possible interpretations of the
DTD result – that RR Lyrae may have a previously-unidentified younger
progenitor channel, or unknown systematics still exist in SAD measurements of
resolved stellar populations older than a Gyr.
## 2 Ingredients for Calculating The DTD
### 2.1 OGLE-IV Sample of LMC RR Lyrae
We use the Soszyński et al. (2016) catalog of 39,082 RR Lyrae stars from the
Optical Gravitational Microlensing Experiment (OGLE-
IV)111http://ogle.astrouw.edu.pl/ survey of the LMC (Udalski et al., 2015).
These RR Lyrae were selected by the OGLE-IV pipeline from the full database of
OGLE $I$-band light curves with periods between 0.2 and 1 day, and then
further classified as fundamental (RRab), first-overtone (RRc), and mixed mode
pulsators based on their periods, Fourier amplitudes, and light curve shapes.
We excluded catalog entries that were flagged as Galactic RR Lyrae or
eclipsing variables, objects with uncertain classification, and sources that
fall outside the SAD maps. This produced a final sample of 29,810 RR Lyrae
(Figure 1). The photometric completeness of the sample is quite high, as
evidenced by the I-band luminosity function of RR Lyrae in the most crowded
OGLE-IV field (LMC503, see Figure 2). The field has a completeness limit of
$I\approx 20.5$ (Udalski et al., 2015), whereas the RR Lyrae sample in the
field has a median magnitude $\bar{I}=18.82$ and standard deviation
$\sigma_{I}=0.4$. The median magnitude of the RR Lyrae is nearly
4.2$\sigma_{I}$ above the completeness limit (almost 99.9$\%$ of the RR Lyrae
have I-band magnitude above the completeness limit). Thus even in the most
crowded region, the RR Lyrae sample can be considered photometrically
complete. Also as seen in Figure 2, the I-band luminosity function of the full
OGLE-IV sample inside the Harris & Zaritsky (2009) region (see Section 2.2) is
also well above the completeness limit of the most crowded region. Such high
completeness is important for measuring unbiased rates and DTDs (Maoz &
Badenes, 2010; Badenes et al., 2010).
### 2.2 Stellar Age Distribution Map of LMC
We use the SAD map of the LMC constructed by Harris & Zaritsky (2009)
(hereafter, HZ09). This map provides the best-fit stellar mass formed as a
function of lookback time in spatial cells resolving the central
$8.5^{\circ}\times 7.5^{\circ}$ of the galaxy (Zaritsky et al., 2004). HZ09
also provides the associated 1$\sigma$ upper and lower limits to the SAD in
each cell, which we will incorporate into the DTD uncertainties in Section
3.2. The SADs were calculated using data from the Magellanic Cloud Photometric
Survey (MCPS) of nearly 4 million stars collected with the 1m Swope telescope,
down to a completeness of V=20–21 mag (Zaritsky et al., 1997, 2004). The MCPS
region was divided into 1376 cells, each measuring $24^{\prime}\times
24^{\prime}$, or $12^{\prime}\times 12^{\prime}$ if the field contained more
than 25,000 stars. The contours of these cells are shown in Figure 1. SADs
were derived using the StarFISH algorithm by fitting the color magnitude
diagram in each cell with a linear combination of isochrones (Harris &
Zaritsky, 2001). After accounting for extinction and photometric errors, each
cell’s SAD was fit using 16 logarithmically-spaced bins spanning the ages
between 4 Myr and 20 Gyr, and four metallicity bins ($Z=0.008$, 0.004, 0.0025,
and 0.001). For ages younger than 100 Myr, a single metallicity of Z=0.008 was
used because the different metallicity isochrones are almost
indistinguishable.
Note that while the lower limit on the age of RR Lyrae stars is generally
quoted as 10 Gyr, the SAD map has a single indivisible age-bin of 8–12 Gyr,
and so we refer to this lower limit as 8 Gyr in the rest of the paper.
## 3 The RR Lyrae Delay-Time Distribution
### 3.1 Method
The RR Lyrae catalog and SAD maps of the LMC are used to estimate the RR Lyrae
DTD using the non-parametric method described in Badenes et al. (2015)
(hereafter, B15), although we improve on some aspects of it. The RR Lyrae DTD
is the number of RR Lyrae formed per unit stellar mass as a function of the
time-delay between star-formation and the RR Lyrae phase. The convolution of
the DTD with the SAD in each spatial cell predicts the number of RR Lyrae that
will be produced by the stellar population in that cell. For each SAD cell
$i$, the number of RR Lyrae predicted ($\lambda_{i}$) is:
$\lambda_{i}=\sum_{j=1}^{N}M_{ij}\left(\Psi T_{vis}\right)_{j}$ (1)
where $M_{ij}$ is the stellar mass formed in age-bin $j$ and cell $i$, $\Psi$
is the RR Lyrae formation rate (RR Lyrae per year per unit M⊙), and Tvis is
the duration of the RR Lyrae phase (both $\Psi$ and Tvis are function of age-
bin). The widths of the age-bins, $j$, are selected using the methodology
described in Appendix A and provide the best compromise between detection
significance and temporal resolution.
In this paper, we retain the notation of B15 and refer to the quantity
$\left(\Psi T_{vis}\right)$ as the DTD (with units of RR Lyrae per unit M⊙).
The DTD ($\Psi T_{vis}$) is determined by minimizing the difference between
the predicted and observed number of RR Lyrae across spatial cells. This is
carried out in a similar manner to B15 using the Markov Chain Monte Carlo
(MCMC) solver emcee (Foreman-Mackey et al., 2013). We denote
$\mathbf{N}=[N_{i}]$ as the vector representing the number of RR Lyrae in each
spatial cell, and $\mathbf{\Psi T_{vis}}=[\left(\Psi T_{vis}\right)_{j}]$ as
the vector of predicted number of RR Lyrae per stellar mass for each age-bin.
The posterior is calculated as
$p\left(\mathbf{\Psi
T_{vis}}|\mathbf{N}\right)\propto\mathcal{L}\left(\mathbf{N}|\mathbf{\Psi
T_{vis}}\right)\pi\left(\mathbf{\Psi T_{vis}}\right)$ (2)
where $\mathcal{L}$ is the likelihood and $\pi$ is the prior. We assume
$\mathcal{L}$ is either a product of Poisson or Gaussian probabilities
depending on Ni,
$\mathcal{L}\left(\mathbf{N}|\mathbf{\Psi
T_{vis}}\right)=\begin{cases}\displaystyle\prod_{i=1}^{K}\frac{\mathrm{e}^{-\lambda_{i}}\lambda_{i}^{N_{i}}}{N_{i}!}&\quad
N_{i}\leq 25\\\
\displaystyle\prod_{i=1}^{K}\frac{1}{2\pi\lambda_{i}}\left(\frac{(\lambda_{i}-N_{i})^{2}}{2\lambda_{i}}\right)&\quad
N_{i}>25\end{cases}$ (3)
where $K$ is the number of SAD cells. Although Poisson distributions converge
to Gaussian for large $N_{i}$, we specify the likelihoods separately to avoid
computational issues. The prior is defined such that the $(\Psi T_{vis})_{j}$
values have uniform probabilities in log-space:
$\pi\left(\mathbf{\Psi T_{vis}}\right)=\begin{cases}\mathbf{\left(\Psi
T_{vis}\right)^{-1}}&\quad(\Psi T_{vis})_{j}>0\\\ 0&\quad(\Psi
T_{vis})_{j}\leq 0\end{cases}$ (4)
### 3.2 Estimating Uncertainties
We propagate the errors on the $M_{ij}$ values in the SAD map into
uncertainties on $\mathbf{\Psi T_{vis}}$. B15 calculated DTDs for the best-fit
SAD, the upper-limit on the SAD, and the lower-limit on the SAD. The
differences between the best-fit and upper/lower limits were treated as the
DTD’s 1$\sigma$ uncertainties.
In this paper, we use an improved method of propagating SAD uncertainties into
the DTD. We randomly generate 100 mock SADs, assuming $M_{ij}$ has the
normally-distributed uncertainties given by HZ09, and calculate a DTD for each
mock SAD. We combine the MCMC posterior chains from these 100 DTDs into a
single chain, and estimate the $95\%$ credible interval on this chain using a
highest posterior density criterion. We define the mode of the distribution
minus the upper and lower-limits of this interval as our $2\sigma_{+}$ and
$2\sigma_{-}$ confidence intervals, respectively. We define a “signal”
detection in each bin $j$ of the DTD as a value of $\Psi T_{vis}$ in that bin
that is $\geq$ $2\sigma_{-}$ above 0. Non-detections are presented as
$2\sigma$ limits on the DTD signal in a particular age bin.
### 3.3 Sample Contribution per Age-bin
We estimate the contribution of each age-bin to the total sample by
multiplying the value of $\Psi T_{vis}$ with the total stellar mass formed in
each age-bin. The contribution percentage will then have uncertainties due to
both the DTD and the total stellar masses formed. We estimate these
uncertainties with a Monte Carlo method. We draw DTD values from the recovered
posterior probability distributions (called $\left(\Psi T_{vis}\right)_{j}$),
multiply by the total stellar mass per age-bin in each randomized SAD map
(called Mj), and get the number of RR Lyrae contributed by age-bin,
$\lambda_{j}$. The contribution percentages (=
$\lambda_{j}/\Sigma\lambda_{j}$) and their 1$\sigma$ uncertainties are listed
in Table 1.
Figure 3: The delay-time distribution, in units of number of RR Lyrae per M⊙,
as a function of delay-time in Myr. The grey histogram represents the DTD for
the full OGLE-IV sample, while the colored histograms are DTDs for RR Lyrae
subtypes RRab and RRc Filled histograms represent bins with a statistically
significant signal from the MCMC analysis. The error bars are 1$\sigma$
uncertainties that include uncertainties from the SAD maps. The arrows
represent 2$\sigma$ upper limits in bins that do not have a statistically
significant signal. At top, the progenitor mass-scales for two different
metallicities, calculated using PARSEC, are shown for reference.
### 3.4 Results
The DTD for the full OGLE-IV RR Lyrae sample and the two main RR Lyrae
subtypes is shown in Figure 3, and the values are tabulated in Table 1. We
detect signal in the DTD at a high significance ($>5\sigma$) for all age-bins
older than 1.3 Gyrs. The detections with the highest significance are found in
the 2–3, 5–8 and $>12$ Gyr bins. Most of this signal is contributed by the
RRab stars, which are the most common subtype in the OGLE-IV sample. The RRc
subclass contributes to the DTD mostly above 2 Gyr. Although RRc’s have been
susceptible to confusion with other variable sources in time domain-surveys
(Kinman & Brown, 2010; Mateu et al., 2012; Drake et al., 2014), Figure 3 shows
that the total DTD is dominated by RRab objects in all age-bins, making it
unlikely that the DTD is biased by sample contamination. The detected signal
of the full DTD is relatively flat above 1.2 Gyr, with about $1-3$ RR Lyrae
produced per 105 M⊙ of stellar mass formed. While there is signal in the DTD
below 0.8 Gyrs, it falls below our 2$\sigma_{-}$ threshold, and we only show
the 2$\sigma_{+}$ upper limit.
Our detection of a DTD signal below 8 Gyrs is a surprising result. About
$46\%$ of the LMC’s RR Lyrae stars are produced from populations older than 10
Gyr, the age range generally inferred for RR Lyrae in star clusters. But about
$51\%$ of the DTD signal comes from progenitors with ages between 1.3 and 8
Gyr, and this result has high ($>5\sigma$) significance. If we re-calculate
the DTD without assuming normal errors on $M_{ij}$ (i.e., using the same
method as B15, see Appendix B), we still detect a strong signal below 8 Gyr,
but with a total contribution of 41$\%$. A comparison of these timescales with
those of the PARSEC222https://people.sissa.it/~sbressan/parsec.html. We use
PARSEC because it is one of the latest stellar evolution codes with both
publicly available main-sequence and horizontal branch tracks. We get similar
results with the MIST evolutionary tracks of Choi et al. (2016). models for
the onset of helium burning leads to the conclusion that RR Lyrae can arise
from main-sequence progenitors as massive as $\sim 2$ M⊙ at LMC metallicity.
Incidentally, this upper limit is similar to the mass at which stars
transition between igniting He under degenerate conditions (the “He-flash”) to
burning helium under stable, non-degenerate conditions (Bildsten et al., 2012;
Mosser et al., 2014). We also tested the dependence of the DTD as a function
of RR Lyrae star period and brightness by sub-dividing the OGLE-IV data by
pulsation time (3 bins with $<0.45$ d, $0.45-0.58$ d, and $>0.58$ d) and
$I$-band magnitude (3 bins with I$<19.2$, $19.2-19.4$ and $>19.4$). By sub-
dividing the sample in this way, we ensured that each period and magnitude bin
contained enough RR Lyrae for a robust measurement of their DTDs. We found
that all the sub-samples have significant ($>2\sigma$) DTD signals in the
range 1 to 8 Gyr, with no discernible trend in the shape of the DTD with
magnitude or period.
Although, in principle, it is possible to derive a metallicity-dependent DTD
with the HZ09 SAD maps, we defer that study to a future work. However, it is
well-known that LMC RR Lyrae are generally metal-poor, with [Fe/H]$<-0.5$ and
a peak in the metallicity distribution function at [Fe/H]$\sim-1.5$ (Skowron
et al., 2016). We can therefore check to see if there is DTD signal below 8
Gyrs under the assumption that OGLE-IV RR Lyrae are only produced by the HZ09
SAD in the two metal-poor bins, i.e., for $Z<0.0025$ or [Fe/H]$\lesssim-1.02$
(Bertelli et al., 1994). Star-formation in this metallicity range dominated
the LMC SAD until $\sim 2$ Gyrs, as seen in Figure 4, and the resulting DTD is
shown in Figure 4. Even if we restrict our analysis to the metal-poor SAD, the
signal below 8 Gyrs persists. This implies that, even if we assume that all
LMC RR Lyrae are produced by metal-poor stars, progenitors younger than 8 Gyrs
are still needed to explain their distribution.
Figure 4: (a) The global (summed over all SAD cells) stellar mass formed
versus age ($M_{ij}$ in Eq (1)) with $Z<0.0025$ (red) and $Z>0.0025$ (blue) in
the LMC as given by the HZ09 SAD. The black histogram shows the sum of the red
and blue histograms. Overlapping regions appear in a darker shade. (b) The DTD
measured from the OGLE-IV survey using only the SAD in the $Z<0.0025$ bin
(grey; see Section 3.4 for details). For comparison, the DTD from Figure 3 is
shown in red.
Before delving into the implications for RR Lyrae studies, we carry out a more
detailed analysis of the recovered DTD to assess the robustness of our
results.
Table 1: The RR Lyrae DTD calculations, with lifetimes, significance of detection and contribution to the OGLE-IV RR Lyrae sample considered in this study. Delay-Times | DTD ($\Psi T_{\rm{vis}}$) | Significance | Contribution
---|---|---|---
(Gyr) | (N/105 M⊙) | ($\sigma$) | ( $\%$)
$<$ 0.8 | $<0.75$ | $-$ | $-$
0.8–1.3 | $2.12^{+0.61}_{-0.78}$ | $3.8$ | $2.9\pm 0.8$
1.3–2.0 | $1.36^{+0.2}_{-0.21}$ | $5.6$ | $8.4\pm 1.2$
2.0–3.2 | $2.05^{+0.11}_{-0.22}$ | $20.5$ | $20.7\pm 1.2$
3.2–5.0 | $1.10^{+0.18}_{-0.23}$ | $5.6$ | $6.0\pm 1.0$
5.0–7.9 | $2.17^{+0.19}_{-0.36}$ | $11.6$ | $15.6\pm 1.3$
7.9–12.6 | $1.08^{+0.17}_{-0.29}$ | $6.8$ | $7.5\pm 1.1$
12.6–20.0 | $1.18^{+0.07}_{-0.1}$ | $18.5$ | $39.0\pm 1.8$
## 4 On the robustness of the recovered RR Lyrae DTD
### 4.1 Mock DTD Test
The HZ09 SAD map of the LMC allows us to test whether the traditional DTD for
RR Lyrae (i.e., one in which the progenitors are always older than 8 Gyr) is
consistent with the spatial distribution of RR Lyrae observed by OGLE. This is
equivalent to inverting the DTD recovery process. To do this, we generate mock
RR Lyrae maps by convolving the SAD map with a DTD that is non-zero only in
the 2 oldest age-bins, 8–12 and 12–20 Gyr. The total stellar mass that formed
in these age-bins was $1.23\times 10^{9}$ M⊙.
Assuming all 29,810 RR Lyrae were produced by progenitors in these age-bins
results in a DTD of the form,
$\Psi T_{vis}=\begin{cases}2.42\times 10^{-5}\mathrm{\ RRL\
M_{\odot}^{-1}}&\quad t\geq 8\ \mathrm{Gyrs}\\\ 0&\quad t<8\
\mathrm{Gyrs}\end{cases}$ (5)
We generate 100 mock RR Lyrae maps using this DTD, where the number of RR
Lyrae per cell, $N_{i}$ is drawn from the Poisson distribution in Equation (3)
with $\lambda_{i}$ given by Equation (1).
Figure 5 shows the DTD recovered from this analysis. Our MCMC solver correctly
measures the input mock DTD with strong detections only in the two oldest age-
bins. The younger bins have 2$\sigma$ upper limits that are almost an order of
magnitude lower than the DTD recovered from the OGLE-IV sample. Moreover, the
difference between populations of RR Lyrae with the OGLE-IV and mock DTDs can
be seen visually in Figure 6. The measured OGLE-IV DTD predicts a distribution
of RR Lyrae stars elongated along the LMC Bar and declining smoothly with
radius (center bottom panel of Figure 6); this is very similar to what is seen
in the OGLE-IV map (left panel). In contrast, our mock distribution from a
uniformly old RR Lyrae progenitor population (Equation (5)) has more structure
and leaves larger residuals in the difference map (top center and right panels
of Figure 6). To identify significant regions of discrepancy in the OGLE-IV
DTD and mock old DTD maps, we generate 104 maps of RR Lyrae, where in each map
the number of RR Lyrae in each cell is drawn from a Poisson distribution with
mean number of RR Lyrae in that cell given by the DTD according to Eq 1. We
then estimate the mean and standard deviation of the number of RR Lyrae per
cell, and identify cells with black squares in Figure 6 where the observed RR
Lyrae is greater than 5 times the standard deviation from the mean. We find
the mock old DTD map has a larger number of cells where the RR Lyrae number
count is discrepant, and these cells are mostly located in the Bar region,
with a few located outside (note that this is only for the purpose of
visualizing the discrepancy; the actual DTD is constrained by the joint
likelihood measured from _all_ the cells as in Eq 3.)
Figure 5: Result of our mock DTD test in Section 4.1. The red histogram shows
the DTD measured with the original OGLE-IV survey in Figure 3. The black
histogram shows the mock input DTD defined in Eq 5. The blue histogram with
arrows shows the DTD recovered from the mock input DTD and the SAD maps of
HZ09.
We conclude that if the OGLE-IV RR Lyrae had indeed been produced exclusively
from the old ($t>8$ Gyr) stars of the HZ09 SAD map, our method would have
recovered the correct DTD, without spurious signals at younger ages. The DTD
signal we recover at ages between 0.8 and 8 Gyr must either be real, or an
artifact produced by systematics in the HZ09 SADs. Next, we investigate the
impact of these SAD systematics on the DTD more thoroughly.
Figure 6: Comparison between the spatial distribution of RR Lyrae in OGLE-IV
(left, large panel) and the distributions predicted by convolving the HZ09 SAD
with the ‘mock’ old DTD (upper middle panel) and our recovered RR Lyrae DTD
(lower middle panel). The right panels show “residuals”, or the difference
between the observed distribution and DTD-predicted distribution of RR Lyrae
for the ‘mock’ old DTD (top right) and measured DTD (bottom right). The black
squares show the cells where the difference between predicted and observed
number of RR Lyrae differ by 5$\sigma$ (see Section 4.1 for details)
.
### 4.2 Effect of incompleteness and crowding in the SAD map
We explored if any incompleteness in the MCPS photometry (from which the SAD
was measured) could be driving the intermediate-age signal. The photometric
completeness limit is around V=21 mag, with the Bar region being almost 1 mag
shallower as a result of the stellar crowding. Because of this, StarFISH could
reliably solve for the number of stars per age-bin only for ages younger than
4 Gyr (although we note that our intermediate-age DTD signal is detected
younger than 4 Gyrs). Populations older than 4 Gyr were traced by their giant
branch stars, and assigned by StarFISH to a single age-bin. To anchor the
‘shape’ of the SAD beyond 4 Gyrs, stellar ages were measured in a few isolated
_HST_ -fields in the LMC Bar (Holtzman et al., 1999; Olsen, 1999; Smecker-Hane
et al., 2002). As shown in Figure 7 of HZ09, the SADs of each _HST_ Bar field
are characterized by star-formation at look-back times of 10 Gyrs and 5 Gyrs,
with a period of quiescence in between. This consistency allowed HZ09 to adopt
a common shape for the SAD in the rest of the Bar region.
We first checked if an underestimation of the total mass formed at old stellar
ages due to any photometric incompleteness could be driving the intermediate-
age signal. Obtaining a complete census of old stellar mass is generally
difficult without including infra-red data (Conroy, 2013), so we studied the
changes to our DTD by manually changing the old SADs. We recalculated the RR
Lyrae DTD with the same SAD map, except with each cell’s 8–12 Gyr and 12–20
Gyr stellar mass multiplied by factors of 2, 4 and 10. We find that regardless
of how much mass is added, the DTDs still find signals for ages between 0.8
and 8 Gyrs with $>5\sigma$ confidence. Since the incompleteness in the SAD map
is dominant in the Bar region, we also tried re-calculating the DTD without
the LMC Bar (i.e., by removing the SAD cells and their RR Lyrae from our
calculations). We show the RR Lyrae DTDs without the “Inner Bar” in Figure 7,
and without both the “Inner” and “Outer Bar” in Figure 8 (the cells of these
regions are defined according to Figure 6 in HZ09). The recovered DTDs outside
the excluded regions are similar to the DTD obtained from the full OGLE-IV
sample. For the case with both the Inner and Outer Bar removed, the number of
RR Lyrae and the number of SAD cells are smaller, leading to a higher upper
limit on the DTD older than $0.8$ Gyr, and a non-detection in the 3–5 Gyr bin.
However, there is still significant signal at ages younger than 8 Gyr, and
particularly below 4 Gyr where the ages are determined by the main-sequence
turnoff in the MCPS photometry. It therefore appears unlikely that missing
stellar masses at ages $>8$ Gyrs due to incompleteness, or crowding in the Bar
region are the only factors driving the intermediate-age signal in the DTD.
Figure 7: RR Lyrae DTD of the OGLE-IV survey, but excluding the SAD cells and
RR Lyrae of the Inner Bar. The excluded cells of HZ09 are shown in the inset
plot, overlaid on the $r$-band continuum map from MCELS. Figure 8: Same as
Figure 7, but now excluding both the Inner and Outer Bar. Note that the
excluded cells for both Inner and Outer Bar are based on Figure 6 of HZ09.
### 4.3 Comparison with RR Lyrae in LMC star clusters
Figure 9: (_Left_): Comparison of ages of the 62 clusters in Baumgardt et al.
(2013) older than 1 Gyr, and their $\Delta n=(n_{o}-n_{b})-n_{p}$, the
difference between the background-subtracted number of RR Lyrae observed per
cluster, and the number predicted by the DTD given the cluster mass and age
(see Eq (6), (7) and Section 4.3 for details). Colors indicate the cluster
mass. The inset zooms-in on clusters between 1–4 Gyr. (_Right_): $\Delta n$ vs
metallicity for a subset of 29 clusters with metallicity information available
in the literature. Colorbar indicates the cluster age. The inset zooms-in on
the 1–4 Gyr clusters, which also happen to cluster in the metallicity range
$-1\lesssim$ [Fe/H] $\lesssim-0.25$.
Our measured DTD suggests that progenitors of RR Lyrae can be as young as
$\sim$1 Gyr, which is much smaller than the lower limit on RR Lyrae age of 10
Gyrs inferred from star clusters (Olszewski et al., 1996). We therefore
examine the number of OGLE-IV RR Lyrae in the LMC star clusters as a function
of age, realizing that cluster membership can only be confirmed with
spectroscopic and proper motion measurements that are beyond the scope of this
work.
To perform this test, we use the catalog of LMC clusters compiled by Baumgardt
et al. (2013) and only include the 296 systems inside the HZ09 area. Baumgardt
et al. (2013) compiled ages and masses for these clusters measured using
either isochrone fitting or broadband spectral energy distribution fitting of
data obtained by previous surveys (Pietrzynski & Udalski, 2000; Hunter et al.,
2003; Mackey & Gilmore, 2003; de Grijs & Anders, 2006; Milone et al., 2009;
Glatt et al., 2010; Popescu et al., 2012). We determine the cluster radii,
$r_{c}$ from their major ($a$) and minor axes ($b$) reported in Bica et al.
(2008) as $r_{c}=(a+b)/4$. Bica et al. (2008) notes that these are ‘apparent’
sizes (measured as far as the background limit in the images), but they most
likely enclose the majority of the cluster mass. We also obtained chemical
abundances for 29 clusters ([Fe/H], shown in Table 2 with references), and as
we show later, this is sufficient for understanding the correlation of cluster
RR Lyrae statistics with metallicity. Although there have been more recent
studies of the LMC star cluster population (e.g. Nayak et al., 2016; Bitsakis
et al., 2017; Piatti, 2017, 2018), the effects of field star contamination and
asterisms on cluster identification in these studies are unclear. Since
Baumgardt et al. (2013) compiles previously identified clusters above 5000 M⊙,
we rely on this catalog for our analysis. Additionally, this catalog (to the
best of the authors’ knowledge) is the only catalog of LMC clusters in the
literature with both age _and_ mass estimates of the clusters—both being
critical to our analysis.
Out of the 296 clusters, we investigate the RR Lyrae statistics of 62 clusters
that have ages above 1 Gyr in order to compare with the DTD. We define $n_{o}$
as the number of OGLE-IV RR Lyrae observed within a circle of radius $r_{c}$
centered on the cluster, and $n_{b}$ as the expected number of RR Lyrae within
the cluster area but unassociated with the cluster (we refer to these as
‘background’ RR Lyrae). Assuming $N$ is the number of RR Lyrae within radii
$2r_{c}$ and $4r_{c}$ of the cluster (we allow a buffer region of $r_{c}$ to
account for uncertainty in the actual extent of the cluster), we define
$n_{b}$ as,
$n_{b}=\frac{N\pi r_{c}^{2}}{\pi(4r_{c})^{2}-\pi(2r_{c})^{2}}=\frac{N}{12}$
(6)
Values for $n_{o}$ and $n_{b}$ are listed in Table 2 and shown in Figure 9.
LMC globular clusters older than 10 Gyr clearly host populations of multiple
RR Lyrae that exceed the background. The richest population is in NGC 1835
with over a 100 RR Lyrae, while NGC 2005, NGC 1928 and NGC 1939 have less than
10 RR Lyrae. In contrast the intermediate-age clusters (1–10 Gyr old) are
generally lacking in RR Lyrae. Out of the 53 intermediate-age clusters, 42
have 0 RR Lyrae and only three clusters (NGC 1978, SL180 and HS190) have more
than 1 RR Lyrae star after subtracting the background numbers (i.e.
$n_{o}-n_{b}$). This is also the case in the Milky Way, where studies of
variable stars in well known intermediate-age systems such as M67 (Pribulla et
al., 2008), NGC 188 (Zhang et al., 2004), NGC 6791 (de Marchi et al., 2007)
and NGC 6253 (Kaluzny et al., 2014) have found no RR Lyrae that are likely to
be cluster members. In contrast, the Milky Way’s globular cluster system
contains almost 2000 of such objects (Clement et al., 2001; Soszyński et al.,
2014).
While this deficit of RR Lyrae in intermediate-age clusters appears
contradictory to the 1–8 Gyr signal in the DTD, we note that the intermediate-
age clusters in the LMC are generally less massive than old globular clusters,
and are therefore less likely to host relatively short-lived objects such as
RR Lyrae stars. We quantify this with $n_{p}$, the expected number of RR Lyrae
per cluster given our measured DTD $\Psi T_{vis}$, the mass $M_{c}$, and age
$t_{c}$ of a cluster:
$n_{p}=M_{c}(\Psi T_{vis})|_{t=t_{c}}$ (7)
For each cluster, we measure $\Delta n=(n_{o}-n_{b})-n_{p}$, i.e. the
difference between the observed background-subtracted number of RR Lyrae
($n_{o}-n_{b}$) and the number predicted by the DTD. We estimate the
uncertainty in $\Delta n$ in the following way: we create 106 random samples
of the 62 clusters, each sample having the same $n_{o}$ per cluster as given
in Table 2 but with $n_{b}$ and $n_{p}$ generated from a Poisson distribution
with mean $n_{b}$ and $n_{p}$ values given in Table 2. For any cluster in
these samples with a negative value of $n_{o}-n_{b}$, we set its
$n_{o}-n_{b}=0$. The mean and standard deviation of this random sampling is
taken to be the value and error of $\Delta n$. These $\Delta n$ are shown as a
function of cluster age and metallicity in Figure 9. We see that $\Delta n$ is
consistent with 0 within 2$\sigma$ for most intermediate-age clusters in the
LMC. This is because the DTD predicts $\sim 1-3$ RR Lyrae per 105 M⊙ of stars,
whereas the intermediate-age clusters have an average mass of $4.3\times
10^{4}$ M⊙, so the majority will have less than 1 RR Lyrae star. This caveat
was also raised by Olszewski et al. (1996) in their study of the SMC: if the
RR-Lyrae-rich cluster NGC 121 ($t\approx 12$ Gyr) were scaled down to the
magnitude of the 9 Gyr system Lindsey 1 (which has no RR Lyrae), it would host
less than 1 RR Lyrae star. The intermediate-age clusters of the Milky Way,
including the ones listed above, have the same issue: all have masses less
than $10^{4}$ M⊙ (Kaluzny & Udalski, 1992; Mermilliod, 2000; Piskunov et al.,
2008; Grundahl et al., 2008; Elsanhoury et al., 2016; Kruijssen et al., 2019).
Thus, the traditional lower limit on the age of the RR Lyrae progenitor
population is subject to small number statistics.
We note that the globular clusters host a diversity of RR Lyrae populations.
NGC 1939, NGC 1916 and NGC 2005 are consistent with $\Delta n=0$ within their
uncertainties, and their production rates $(n_{o}-n_{b})/M_{c}=(1.8-2.5)\times
10^{-5}$ RR Lyrae per M⊙ are similar to the measured DTD. The other old
clusters, however, host RR Lyrae significantly in excess of their DTD-
predicted $n_{p}$ (i.e. $\Delta n\gg 0$), and have an average production rate
$\sim 10^{-4}$ RR Lyrae per M⊙, a factor $\sim 5$ times higher than the DTD.
The high production rate in the old clusters may be an effect of their low
metallicities, which is consistent with the differences in production rates of
RR Lyrae in the halo versus the bulge and disk of our Galaxy (Layden, 1995;
Dékány et al., 2018), although the large differences in $\Delta n$ for the
globular clusters could also be an effect of the second parameter phenomenon
in clusters (see Fusi Pecci & Bellazzini, 1997; Catelan, 2009; Dotter, 2013,
for reviews). A more detailed assessment of the field and cluster DTDs can be
done with a metallicity-dependent DTD that we reserve for future work.
Table 2: LMC Clusters older than 1 Gyr in the Baumgardt et al. (2013) catalog
and inside the Harris & Zaritsky (2009) region.
Name | Age (Gyr) | [Fe/H] | aRef | Log Mass (M⊙) | $n_{o}$ | $n_{b}$ | $n_{p}$
---|---|---|---|---|---|---|---
SL569 | 1.2 $\pm$ 0.03 | -0.32 $\pm$ 0.05 | 6 | 4.29 | 0 | 0 | 0
KMK88-38 | 1.26${}^{+0.15}_{-0.14}$ | – | – | 3.71 | 0 | 0 | 0
BSDL880 | 1.26${}^{+0.15}_{-0.14}$ | – | – | 3.83 | 0 | 0 | 0
SL197 | 1.26${}^{+0.25}_{-0.21}$ | – | – | 3.93 | 0 | 0 | 0
HS102 | 1.26${}^{+0.15}_{-0.14}$ | – | – | 3.87 | 1 | 0 | 0
HS223A | 1.26${}^{+0.15}_{-0.14}$ | – | – | 4.02 | 0 | 0 | 0
NGC1795 | 1.29${}^{+0.16}_{-0.14}$ | -0.23 | 4 | 4.36 | 0 | 0 | 0
NGC1917 | 1.29${}^{+0.16}_{-0.14}$ | -0.21 | 4 | 4.42 | 0 | 1 | 1
KMHK898 | 1.32 $\pm$ 0.09 | – | – | 4.16 | 0 | 0 | 0
NGC1852 | 1.32${}^{+0.16}_{-0.14}$ | – | – | 4.51 | 1 | 0 | 0
SL282 | 1.35${}^{+0.35}_{-0.28}$ | – | – | 3.71 | 1 | 0 | 0
BSDL946 | 1.35${}^{+0.2}_{-0.17}$ | – | – | 4.17 | 0 | 0 | 0
NGC2154 | 1.41${}^{+0.17}_{-0.15}$ | -0.56 | 3 | 4.57 | 1 | 0 | 1
NGC1751 | 1.41${}^{+0.17}_{-0.15}$ | -0.44 $\pm$ 0.05 | 6 | 4.6 | 0 | 0 | 1
SL151 | 1.41${}^{+0.25}_{-0.21}$ | – | – | 4.1 | 0 | 0 | 0
HODGE7 | 1.48${}^{+0.18}_{-0.16}$ | – | – | 4.47 | 0 | 0 | 0
NGC1846 | 1.48${}^{+0.18}_{-0.16}$ | -0.4 | 1 | 5.1 | 1 | 1 | 2
NGC1783 | 1.51${}^{+0.18}_{-0.16}$ | -0.75 | 3 | 5.26 | 0 | 1 | 2
NGC1806 | 1.51${}^{+0.18}_{-0.16}$ | -0.71 $\pm$ 0.23 | 1 | 5.01 | 1 | 1 | 1
SL136 | 1.55${}^{+0.45}_{-0.35}$ | – | – | 4.39 | 0 | 0 | 0
NGC2213 | 1.58${}^{+0.46}_{-0.35}$ | -0.7 $\pm$ 0.1 | 1 | 4.56 | 0 | 0 | 0
SL180 | 1.58${}^{+0.41}_{-0.33}$ | – | – | 4.32 | 3 | 0 | 0
H1 | 1.58${}^{+0.41}_{-0.33}$ | -0.29 | 5 | 4.84 | 1 | 2 | 1
SL357 | 1.58${}^{+0.32}_{-0.27}$ | – | – | 4.75 | 0 | 1 | 1
SL390 | 1.58 $\pm$ 0.11 | -0.4 | 2 | 4.48 | 0 | 1 | 0
HS117 | 1.58${}^{+0.41}_{-0.33}$ | – | – | 4.15 | 2 | 1 | 0
BSDL1102 | 1.62${}^{+0.47}_{-0.36}$ | – | – | 3.77 | 0 | 0 | 0
SL66 | 1.7${}^{+0.7}_{-0.5}$ | – | – | 4.34 | 0 | 0 | 0
NGC1652 | 1.7${}^{+0.21}_{-0.18}$ | -0.46 | 3 | 4.29 | 0 | 0 | 0
BSDL2652 | 1.74${}^{+0.5}_{-0.39}$ | – | – | 3.73 | 0 | 0 | 0
HS87 | 1.78${}^{+0.22}_{-0.19}$ | – | – | 3.98 | 0 | 0 | 0
OGLE-LMC0114 | 1.78${}^{+0.46}_{-0.37}$ | – | – | 3.71 | 0 | 0 | 0
HS37 | 1.78${}^{+1.24}_{-0.73}$ | – | – | 4.02 | 1 | 0 | 0
HS177 | 1.78${}^{+0.22}_{-0.19}$ | – | – | 3.95 | 0 | 0 | 0
H2 | 1.78${}^{+0.46}_{-0.37}$ | -0.38 | 5 | 4.98 | 0 | 1 | 1
BSDL734 | 1.78${}^{+0.22}_{-0.19}$ | – | – | 3.8 | 1 | 0 | 0
KMHK355 | 1.86${}^{+0.96}_{-0.63}$ | – | – | 3.7 | 0 | 0 | 0
OGLE-LMC0531 | 2.0${}^{+0.24}_{-0.22}$ | – | – | 3.91 | 0 | 0 | 0
NGC1978 | 2.0${}^{+0.24}_{-0.22}$ | -0.38 $\pm$ 0.07 | 1 | 5.33 | 2 | 0 | 3
NGC1651 | 2.0${}^{+0.46}_{-0.37}$ | -0.53 $\pm$ 0.03 | 1 | 5.24 | 0 | 0 | 2
SL629 | 2.04${}^{+0.1}_{-0.09}$ | – | – | 4.0 | 0 | 0 | 0
KMHK1112 | 2.4${}^{+0.69}_{-0.54}$ | – | – | 3.78 | 1 | 0 | 0
HS88 | 2.45${}^{+1.09}_{-0.76}$ | – | – | 4.1 | 0 | 0 | 0
SL244 | 2.69 $\pm$ 0.06 | -0.7 $\pm$ 0.2 | 1 | 4.77 | 0 | 0 | 1
SL150 | 2.75 $\pm$ 0.06 | – | – | 4.34 | 0 | 0 | 0
HS190 | 2.75${}^{+3.41}_{-1.52}$ | – | – | 4.51 | 2 | 0 | 1
BSDL2300 | 2.95${}^{+0.21}_{-0.2}$ | – | – | 4.7 | 0 | 0 | 1
KMHK1188 | 3.02${}^{+0.14}_{-0.14}$ | – | – | 4.0 | 0 | 0 | 0
H88-93 | 3.02${}^{+4.06}_{-1.73}$ | – | – | 3.85 | 0 | 0 | 0
BSDL1334 | 3.02${}^{+4.74}_{-1.85}$ | -0.4 | 2 | 3.86 | 0 | 0 | 0
SL663 | 3.24${}^{+0.48}_{-0.42}$ | -0.7 $\pm$ 0.1 | 1 | 5.23 | 0 | 0 | 2
NGC2121 | 3.24${}^{+0.48}_{-0.42}$ | -0.4 $\pm$ 0.1 | 1 | 5.69 | 0 | 1 | 5
NGC2155 | 3.24${}^{+0.48}_{-0.42}$ | -0.7 $\pm$ 0.1 | 1 | 4.9 | 0 | 0 | 1
NGC1939 | 13.49${}^{+3.49}_{-2.77}$ | -2.1 $\pm$ 0.19 | 1 | 5.08 | 4 | 1 | 1
NGC1928 | 13.49${}^{+3.49}_{-2.77}$ | -1.27 $\pm$ 0.14 | 1 | 4.87 | 7 | 1 | 1
NGC1898 | 14.13${}^{+2.47}_{-2.1}$ | -1.37 $\pm$ 0.15 | 1 | 5.88 | 38 | 2 | 9
NGC1786 | 15.14${}^{+0.35}_{-0.34}$ | -2.1 $\pm$ 0.3 | 1 | 5.57 | 47 | 1 | 4
NGC1754 | 15.49${}^{+2.29}_{-2.0}$ | -1.42 $\pm$ 0.15 | 1 | 5.39 | 32 | 0 | 3
NGC1916 | 15.85${}^{+3.65}_{-2.97}$ | -2.08 | 4 | 5.79 | 14 | 3 | 7
NGC2005 | 16.6${}^{+6.31}_{-4.57}$ | -1.35 $\pm$ 0.15 | 1 | 5.49 | 9 | 2 | 4
NGC1835 | 16.6${}^{+2.9}_{-2.47}$ | -1.62 $\pm$ 0.15 | 1 | 5.83 | 105 | 2 | 8
NGC2019 | 17.78${}^{+3.6}_{-2.99}$ | -1.23 $\pm$ 0.15 | 1 | 5.68 | 49 | 1 | 6
Note. — _a_ – References for [Fe/H] measurement. (1) Harris & Zaritsky (2009)
(2) Palma et al. (2016), (3) Girardi & Marigo (2007) (4) Kontizas et al.
(1993) (5) Olszewski et al. (1991) (6) Grocholski et al. (2006)
* - These 2 clusters are listed as having ages $\sim 1.7$ Gyr in Baumgardt et al. (2013), but are more likely to be $\gtrsim 10$ Gyr based on HST WFPC2 observations (Mackey & Gilmore, 2004).
## 5 Implications of the RR Lyrae DTD
If we assume the DTD measured from the OGLE-IV survey is correct, there are
two possible interpretations of this result: 1) RR Lyrae can form from
progenitors younger than 10 Gyr _in addition_ to the conventional route via
older stars, and this result was undetected in previous studies due to various
observational limitations, or 2) all OGLE-IV RR Lyrae are older than 10 Gyr
stars, and our result is a product of significant (though not readily obvious)
systematics in the age derivation of older stellar populations. We discuss
both of these interpretations below.
### 5.1 Can LMC RR Lyrae have an intermediate-age channel?
Since the DTD is generally regarded as a reflection of the progenitor age
distribution (Maoz & Badenes, 2010; Badenes et al., 2015), it is tempting to
consider that RR Lyrae in the LMC are being produced by an as-of-yet
undiscovered intermediate-age progenitor channel between 1-8 Gyrs, in addition
to the usual channel older than 8 Gyrs.
However, looking at currently available evidence, the possibility of such an
undiscovered intermediate-age channel appears to be questionable. Although the
small-number statistics described in Section 4.3 is a factor, it is
nevertheless true that ancient globular clusters host abundant and rich
populations of RR Lyrae compared to the RR Lyrae-poor intermediate-age
clusters, a feature easily explained by an exclusively old channel for RR
Lyrae formation. Evidence that a small fraction of RR Lyrae may arise from
progenitors only a few Gyrs old was recently obtained from thin-disk metal-
rich RR Lyrae observed in the _Gaia_ data (Zinn et al., 2019; Prudil et al.,
2020; Iorio & Belokurov, 2020), and from measurements of companion masses
greater than 1 Msun in wide-orbit RR Lyrae binaries (Kervella et al., 2019a,
b). The thin disk, metal-rich RR Lyrae population was speculated by Iorio &
Belokurov (2020) to be manifestation of binary evolution pulsators (Karczmarek
et al., 2017), which will register as younger stars. However, the existence of
such an intermediate-age channel for LMC RR Lyrae (assuming single-stellar
evolution) would be in tension with the age-metallicity relation of the LMC
constrained by mutiple studies of field and cluster stars(Cole et al., 2005;
Carrera et al., 2008; Rubele et al., 2012; Meschin et al., 2014). According to
these studies, the LMC star-formation history between 2-8 Gyrs was associated
with [Fe/H] between -0.4 and -1. In contrast, the LMC RR Lyrae are
predominantly metal-poor, with [Fe/H]$<-1$ and peaking at [Fe/H]$\sim-1.5$, as
confirmed by spectroscopic studies of field RR Lyrae (Gratton et al., 2004;
Borissova et al., 2006), photometric light curves of RR Lyrae (Haschke et al.,
2012; Wagner-Kaiser & Sarajedini, 2013; Skowron et al., 2016), and RR Lyrae-
hosting globular clusters (Fig 9). Thus the LMC stellar population at ages
below 8 Gyrs, where we measure significant signal in the DTD, is more metal-
rich than the metallicity range measured for LMC RR Lyrae.
### 5.2 Systematic uncertainties in old SADs?
The DTD represents an empirical connection between the RR Lyrae sample and the
SAD map. Since the RR Lyrae sample is highly complete (Section 2.1) and has
strong independent evidence of originating from old stars (Section 5.1), and
since our DTD recovery method would have correctly recovered a purely old RR
Lyrae DTD signal from this SAD map (Section 4.1), it may be possible that the
intermediate-age signal in the DTD is indicative of some systematic
uncertainty in measuring older stellar ages in the LMC. The source and
magnitude of this uncertainty is not readily obvious. The global star-
formation history measured by HZ09 is broadly consistent with the interaction
history of the Magellanic Clouds derived from proper motion modeling (Lin et
al., 1995; Zaritsky & Harris, 2004; Besla et al., 2007) and with the star
formation and chemical enrichment history derived from star clusters
(Chilingarian & Asa’d, 2018). As shown in Section 4.2, we verified that any
incompleteness or statistical uncertainties in the MCPS photometry is unlikely
to be affecting our DTD because: 1) we propagate the reported uncertainties in
the SAD map into our DTDs; 2) the DTD retains signal below 8 Gyrs even when
measured outside the crowded LMC Bar; 3) wiping out the DTD signal of younger
progenitors would require an unreasonably large unseen stellar mass with age
$>$8 Gyr; and 4) we directly detect a signal in the DTD below 4 Gyrs, where
the main-sequence turnoff is detectable above the MCPS completeness limit.
On the other hand, the SAD solutions per region depend on the overall
methodology adopted by the study. An example of this can be found by comparing
the Harris & Zaritsky (2004) SAD map of the SMC, which was derived using
roughly the same methodology used for the LMC, with the Rubele et al. (2018)
SAD map of SMC derived from the deeper VISTA near-infrared survey of the
Magellanic System (VMC, Kerber et al., 2009; Cioni et al., 2011). The Harris &
Zaritsky (2004) 2–3 Gyr SADs has a distinct ring pattern in the SMC, which
they suspected was either the result of photometric incompleteness or
systematic uncertainties in the photometric zero-point of the central SMC
stars. In the VMC SAD, this ring pattern is absent in the 2–3 Gyr stellar
population. Figure 11 of Rubele et al. (2018) shows that the global stellar
mass formed at ages $>1$ Gyr in the SMC differ by more than a factor of 2
between Rubele et al. (2018) and Harris & Zaritsky (2004), and the bimodal
star-formation history at 2.5 and 10 Gyrs found by Harris & Zaritsky (2004) is
replaced with a single broad peak at 5 Gyr in Rubele et al. (2018). These
systematic differences could stem from differences in IMF and distance
assumed, the stellar isochrone models used, and/or age-metallicity binning.
For example, the age-metallicity relation measured by the two studies diverges
for stars older than 4 Gyrs; this is most likely because Rubele et al. (2018)
uses metallicity bins that extend to lower abundances than Harris & Zaritsky
(2004). Similar systematic differences in the LMC SADs may also exist. For
example, the star-formation history of the Northern Void region measured by
Meschin et al. (2014), using $V$\- and $I$-band photometry with the CTIO
Blanco 4 m telescope, differs from that of HZ09 for ages younger than 4 Gyr.
Partial estimates from upcoming SMASH data reveals a well-mixed SAD in the LMC
for ages
Another source of systematic uncertainty in the SADs may be due to the use of
single-stellar evolution isochrone models. According to Moe et al. (2019), the
close ($<10$ AU) binary fraction for Milky Way field stars of LMC metallicity
is $\sim 30\%$ (compared to $\sim 20\%$ for Solar metallicity), so the
influence of binary evolution physics in older populations may be non-
negligible. Stanway & Eldridge (2018) have shown that models of integrated
spectra and photometry of globular clusters and elliptical galaxies that
include binary evolution physics yield age estimates that are different by a
few Gyr compared to single-star models. A well-known observational
manifestation of binary interaction in old stellar populations is the
appearance of blue straggler stars, which are formed from the merger of $\sim
1M_{\odot}$ stars. Blue straggler stars can mimic younger stars in color-
magnitude diagrams as they appear brighter and bluer than the main-sequence
turnoff (Santana et al., 2016), and single stellar population models that
correct for the presence of blue straggler stars yield globular cluster ages
older by a few Gyrs (Fan & de Grijs, 2012). Correcting for blue stragglers
however is non-trivial as the frequency of blue straggler stars is likely a
function of stellar density (Weisz et al., 2014; Santana et al., 2013, 2016),
and the contribution of blue straggler stars at $\sim$Gyr ages have been
difficult to determine in composite stellar populations (Surot et al., 2019).
Leaving binary evolution aside, there are uncertainties even in the physics of
single-star evolution models that can affect age estimates. For example, Tayar
et al. (2017) showed that the mixing length parameter—commonly used to
approximate convection theory in 1D stellar models—appears to be correlated
with metallicity, and if left unaccounted for when estimating ages from the
giant branch (which is the case for ages $>4$ Gyr in the HZ09 maps), it can
lead to age uncertainties up to a factor of 2.
Any number of these reasons could distort the SAD solutions in a subset of the
cells, and therefore the final DTD, which is derived from these data.
### 5.3 Caveats and Future work
From our work, we have shown that DTD provides a new rigorous and quantitative
test of SADs of Local Group galaxies, in addition to its original purpose as a
stellar evolution diagnostic. Although it is possible that unknown sources in
systematic uncertainties in the SAD map may be driving our DTD result, we
would need further tests to verify the authenticity of this issue, which we
will perform in subsequent papers. Estimating the precise form of the HZ09 SAD
map that would be consistent with a purely old RR Lyrae DTD is non-trivial
because of the large number of SAD parameters involved in this exercise (the
stellar masses per age and metallicity bin per cell), and also because RR
Lyrae can only reliably constrain the oldest ages. In addition, the coarse age
and metallicity binning of the HZ09 map may be less than ideal for
constraining the production rate of RR Lyrae if they are coming from a
narrower range of ages and metallicities within each bin (as e.g. indicated in
Section 4.3. We can however check if our DTD result persists when calculated
with SADs derived from deeper photometric data with finer age and metallicity
resolution, such as the upcoming SMASH star-formation histories (Ruiz-Lara et
al., 2020), as well as in other Local Group galaxies like the SMC (Rubele et
al., 2018), M31 (Williams et al., 2017) and Local Group dwarfs (Weisz et al.,
2014). A more constraining test of the SADs at intermediate and younger ages
can also be obtained by calculating DTDs of younger variables stars with well-
constrained ages, such as $\delta$-Scutis (ages 1-3 Gyrs, Petersen &
Christensen-Dalsgaard, 1996) and Classical Cepheids (ages 70-200 Myrs, Bono et
al., 2005). We will pursue these in future papers.
## 6 Conclusions
We have calculated the first delay time-distribution (DTD) of RR Lyrae stars
using the large sample of LMC RR Lyrae from the OGLE-IV survey (Soszyński et
al., 2016) and the LMC’s SAD map from Harris & Zaritsky (2009). Our DTD, shown
in Figure 3 and Table 1, constrains the age-distribution of the full LMC RR
Lyrae population, given the measured SAD of the LMC. The OGLE-IV RR Lyrae
sample which overlaps the SAD map of HZ09 contains 29,810 objects, allowing us
to recover a DTD with an unprecedented balance of age resolution and detection
significance. We determined the DTD signal in each age-bin with an MCMC
solver, and used a randomization technique to propagate uncertainties in the
SAD map into the final DTD.
Our measured RR Lyrae DTD has statistically significant ($>5\sigma$) power in
all age-bins above 1.3 Gyrs, with about 51$\%$ of the RR Lyrae associated with
ages between 1.3 and 8 Gyr, and only 46$\%$ with ages above 8 Gyr (the
conventional lower limit to RR Lyrae age; note that while the lower limit
quoted in the literature is 10 Gyr, the SAD map has a single indivisible age-
bin of 8–12 Gyr, and so we refer to the lower limit as 8 Gyr in this paper).
This would imply that the progenitors of RR Lyrae have zero-age main-sequence
masses $\lesssim 2$ M⊙ at LMC metallicity, in contrast with existing
constraints.
We checked the DTD for possible sources of bias. Completeness of the RR Lyrae
sample is probably not an issue based on their $I$-band luminosity function
and the predominance of RRab (fundamental) pulsators, which are least
susceptible to confusion with other types of variables. We also tested our DTD
algorithm on fake RR Lyrae maps drawn from a DTD that assumes all RR Lyrae are
older than 8 Gyrs, and found that our MCMC algorithm recovers this old DTD
without any outlying detections at younger ages. The spatial distribution of
RR Lyrae from a purely old DTD is also inconsistent with the spatial
distribution of the OGLE-IV RR Lyrae. A possible caveat to our result is the
incomplete photometry in the MCPS data and heavy crowding in the LMC central
region, which limits the reliability of the SAD maps to ages younger than 4
Gyr; information about older populations are all based on _HST_ -derived SADs
in a few narrow fields. However it is not readily obvious how this is
producing an intermediate-age signal since: (1) we recover the DTD signal at
ages younger than 8 Gyr even after excluding the Bar region, (2) we measure a
DTD signal younger than 4 Gyr, and (3) we find that in order to affect the
solution, the estimates of the LMC’s old stellar mass must be more than an
order-of-magnitude greater than current measurements.
The direct interpretation of our result would be that RR Lyrae have an
intermediate-age progenitor channel of 1.3-8 Gyrs stars, in addition to its
conventional route via ancient stars, but this possibility is in tension with
existing constraints. Both in the Milky Way and Magellanic Clouds, RR Lyrae
are abundantly hosted in ancient globular clusters, in contrast with
intermediate-age clusters (although this conclusion is somewhat affected by
small-number statistics due to the lower-masses of intermediate-age clusters).
In addition, the 1.3-8 Gyr RR Lyrae population would likely have [Fe/H]$>$-1
based on existing constraints on the LMC age-metallicity relation, whereas the
metallicity distribution of LMC RR Lyrae measured from spectroscopy and
photometric light curves is in the range of [Fe/H]$<$-1 with a peak at
[Fe/H]$\sim-1.5$.
The other possibility of the intermediate-age DTD result is the presence of
unknown systematics in the LMC SAD map. This is not obvious because the global
star-formation history based on the SAD is consistent with the LMC–SMC–Milky
Way interaction histories and the chemical enrichment history of the LMC
derived from independent studies. However, comparison of the SMC SADs of
Harris & Zaritsky (2004) and Rubele et al. (2018) shows that spatial solutions
of the SAD maps can be influenced by the overall methodology adopted for their
construction (e.g., assumptions about the IMF, the spatial size of cells, the
age and metallicity bin size, and stellar isochrone models). Age estimation of
old stellar populations from color-magnitude diagrams can also be affected by
binary evolution processes, such as mass-transfer and mergers, as well as
approximations in single-star evolution models. Any combination of these
reasons could be skewing the SADs for old stellar populations, and this error
could be propagating into the DTD results. We laid out further tests to
investigate the physical nature of the systematics that are driving the
intermediate-age signal in the DTD, such as revisiting the RR Lyrae DTD once
LMC SADs from deeper photometric studies (e.g., SMASH) become available, and
also measuring DTDs in SMC and dwarf galaxies, which have less crowded star
fields and deep _HST_ -derived SADs. We will also continue to apply this
technique to other types of variable stars, such as Cepheids and Delta Scutis,
to probe other enigmatic phases of stellar evolution.
## 7 Acknowledgements
We are grateful to Horace Smith for reading the manuscript and providing many
helpful comments and insight on RR Lyrae observations and models. We also
acknowledge the detailed feedback from our anonymous referee on the
interpretation of RR Lyrae ages in the LMC and their relation with
metallicity, and Dennis Zaritsky, Knut Olsen, Thomas Matheson, Benjamin
Williams, J.J. Eldridge and Jay Strader for insightful discussions and
feedback on this work.
SKS, CB, and LC are grateful for the suppport of NSF grants AST-1412980 and
AST-1907790. CM acknowledges support from the DGAPA/UNAM PAPIIT program grant
IG100319 and from the ICC University of Barcelona visiting academic grants and
thanks the _Gaia_ -UB team for hosting her during part of this research. CM
also thanks the Polo de Desarrollo Universitario (PDU) en Ciencias Físicas at
CURE-UdelaR (Rocha), for their hospitality. DM acknowledges support by grants
from the Israel Science Foundation, the German Israeli Science Foundation and
the European Research Council (ERC) under the European Union’s FP7 Programme,
Grant No. 833031.
This work made use of the publicly available OGLE-IV variable star catalog
333http://ogledb.astrouw.edu.pl/ ogle/OCVS/. This research has made use of
NASA’s Astrophysics Data System, and the VizieR catalogue access tool, CDS,
Strasbourg, France. The original description of the VizieR service was
published in A&AS 143, 23. This work made use of the IPython package (Pérez &
Granger, 2007), SciPy (Jones et al., 2001–), NumPy (van der Walt et al.,
2011), matplotlib, a Python library for publication quality graphics (Hunter,
2007), and Astropy, a community-developed core Python package for Astronomy
(Astropy Collaboration et al., 2013). The Institute for Gravitation and the
Cosmos is supported by the Eberly College of Science and the Office of the
Senior Vice President for Research at The Pennsylvania State University.
## References
* Astropy Collaboration et al. (2013) Astropy Collaboration, Robitaille, T. P., Tollerud, E. J., et al. 2013, A&A, 558, A33, doi: 10.1051/0004-6361/201322068
* Badenes et al. (2015) Badenes, C., Maoz, D., & Ciardullo, R. 2015, Astrophysical Journal Letters, 804, L25, doi: 10.1088/2041-8205/804/1/L25
* Badenes et al. (2010) Badenes, C., Maoz, D., & Draine, B. T. 2010, Monthly Notices of the Royal Astronomical Society, 407, 1301, doi: 10.1111/j.1365-2966.2010.17023.x
* Baumgardt et al. (2013) Baumgardt, H., Parmentier, G., Anders, P., & Grebel, E. K. 2013, Monthly Notices of the Royal Astronomical Society, 430, 676, doi: 10.1093/Monthly Notices of the Royal Astronomical Society/sts667
* Bertelli et al. (1994) Bertelli, G., Bressan, A., Chiosi, C., Fagotto, F., & Nasi, E. 1994, A&AS, 106, 275
* Besla et al. (2007) Besla, G., Kallivayalil, N., Hernquist, L., et al. 2007, ApJ, 668, 949, doi: 10.1086/521385
* Bica et al. (2008) Bica, E., Bonatto, C., Dutra, C. M., & Santos, J. F. C. 2008, Monthly Notices of the Royal Astronomical Society, 389, 678, doi: 10.1111/j.1365-2966.2008.13612.x
* Bildsten et al. (2012) Bildsten, L., Paxton, B., Moore, K., & Macias, P. J. 2012, ApJ, 744, L6, doi: 10.1088/2041-8205/744/1/L6
* Bitsakis et al. (2017) Bitsakis, T., Bonfini, P., González-Lópezlira, R. A., et al. 2017, ApJ, 845, 56, doi: 10.3847/1538-4357/aa8090
* Bono et al. (2005) Bono, G., Marconi, M., Cassisi, S., et al. 2005, ApJ, 621, 966, doi: 10.1086/427744
* Borissova et al. (2006) Borissova, J., Minniti, D., Rejkuba, M., & Alves, D. 2006, A&A, 460, 459, doi: 10.1051/0004-6361:20054132
* Bressan et al. (2012) Bressan, A., Marigo, P., Girardi, L., et al. 2012, MNRAS, 427, 127, doi: 10.1111/j.1365-2966.2012.21948.x
* Carrera et al. (2008) Carrera, R., Gallart, C., Hardy, E., Aparicio, A., & Zinn, R. 2008, AJ, 135, 836, doi: 10.1088/0004-6256/135/3/836
* Catelan (2009) Catelan, M. 2009, Astrophysics and Space Science, 320, 261, doi: 10.1007/s10509-009-9987-8
* Chen et al. (2014) Chen, Y., Girardi, L., Bressan, A., et al. 2014, MNRAS, 444, 2525, doi: 10.1093/mnras/stu1605
* Chilingarian & Asa’d (2018) Chilingarian, I. V., & Asa’d, R. 2018, ApJ, 858, 63, doi: 10.3847/1538-4357/aaba77
* Choi et al. (2016) Choi, J., Dotter, A., Conroy, C., et al. 2016, Astrophysical Journal, 823, 102, doi: 10.3847/0004-637X/823/2/102
* Cioni et al. (2011) Cioni, M. R. L., Clementini, G., Girardi, L., et al. 2011, A&A, 527, A116, doi: 10.1051/0004-6361/201016137
* Clement et al. (2001) Clement, C. M., Muzzin, A., Dufton, Q., et al. 2001, Astronomical Journal, 122, 2587, doi: 10.1086/323719
* Cole et al. (2005) Cole, A. A., Tolstoy, E., Gallagher, John S., I., & Smecker-Hane, T. A. 2005, AJ, 129, 1465, doi: 10.1086/428007
* Conroy (2013) Conroy, C. 2013, ARA&A, 51, 393, doi: 10.1146/annurev-astro-082812-141017
* Conroy et al. (2009) Conroy, C., Gunn, J. E., & White, M. 2009, ApJ, 699, 486, doi: 10.1088/0004-637X/699/1/486
* de Grijs & Anders (2006) de Grijs, R., & Anders, P. 2006, MNRAS, 366, 295, doi: 10.1111/j.1365-2966.2005.09856.x
* de Marchi et al. (2007) de Marchi, F., Poretti, E., Montalto, M., et al. 2007, A&A, 471, 515, doi: 10.1051/0004-6361:20077386
* Dékány et al. (2018) Dékány, I., Hajdu, G., Grebel, E. K., et al. 2018, ApJ, 857, 54, doi: 10.3847/1538-4357/aab4fa
* Demarque et al. (2004) Demarque, P., Woo, J.-H., Kim, Y.-C., & Yi, S. K. 2004, ApJS, 155, 667, doi: 10.1086/424966
* Dotter (2013) Dotter, A. 2013, Memorie della Societa Astronomica Italiana, 84, 97. https://arxiv.org/abs/1307.5589
* Dotter et al. (2008) Dotter, A., Chaboyer, B., Jevremović, D., et al. 2008, ApJS, 178, 89, doi: 10.1086/589654
* Drake et al. (2014) Drake, A. J., Graham, M. J., Djorgovski, S. G., et al. 2014, The Astrophysical Journal Supplement, 213, 9, doi: 10.1088/0067-0049/213/1/9
* Elsanhoury et al. (2016) Elsanhoury, W. H., Haroon, A. A., Chupina, N. V., et al. 2016, New A, 49, 32, doi: 10.1016/j.newast.2016.06.002
* Fan & de Grijs (2012) Fan, Z., & de Grijs, R. 2012, MNRAS, 424, 2009, doi: 10.1111/j.1365-2966.2012.21346.x
* Foreman-Mackey et al. (2013) Foreman-Mackey, D., Hogg, D. W., Lang, D., & Goodman, J. 2013, Publications of the Astronomical Society of the Pacific, 125, 306, doi: 10.1086/670067
* Friedmann & Maoz (2018) Friedmann, M., & Maoz, D. 2018, MNRAS, 479, 3563, doi: 10.1093/mnras/sty1664
* Fusi Pecci & Bellazzini (1997) Fusi Pecci, F., & Bellazzini, M. 1997, in The Third Conference on Faint Blue Stars, ed. A. G. D. Philip, J. Liebert, R. Saffer, & D. S. Hayes, 255
* Gal-Yam & Maoz (2004) Gal-Yam, A., & Maoz, D. 2004, Monthly Notices of the Royal Astronomical Society, 347, 942, doi: 10.1111/j.1365-2966.2004.07237.x
* Gallart et al. (2005) Gallart, C., Zoccali, M., & Aparicio, A. 2005, ARA&A, 43, 387, doi: 10.1146/annurev.astro.43.072103.150608
* Girardi & Marigo (2007) Girardi, L., & Marigo, P. 2007, A&A, 462, 237, doi: 10.1051/0004-6361:20065249
* Glatt et al. (2010) Glatt, K., Grebel, E. K., & Koch, A. 2010, A&A, 517, A50, doi: 10.1051/0004-6361/201014187
* Glatt et al. (2008) Glatt, K., Gallagher, John S., I., Grebel, E. K., et al. 2008, AJ, 135, 1106, doi: 10.1088/0004-6256/135/4/1106
* Gratton et al. (2004) Gratton, R. G., Bragaglia, A., Clementini, G., et al. 2004, A&A, 421, 937, doi: 10.1051/0004-6361:20035840
* Graur et al. (2014) Graur, O., Rodney, S. A., Maoz, D., et al. 2014, Astrophysical Journal, 783, 28, doi: 10.1088/0004-637X/783/1/28
* Grocholski et al. (2006) Grocholski, A. J., Cole, A. A., Sarajedini, A., Geisler, D., & Smith, V. V. 2006, AJ, 132, 1630, doi: 10.1086/507303
* Grundahl et al. (2008) Grundahl, F., Clausen, J. V., Hardis, S., & Frandsen, S. 2008, A&A, 492, 171, doi: 10.1051/0004-6361:200810749
* Harris & Zaritsky (2001) Harris, J., & Zaritsky, D. 2001, ApJS, 136, 25, doi: 10.1086/321792
* Harris & Zaritsky (2004) —. 2004, AJ, 127, 1531, doi: 10.1086/381953
* Harris & Zaritsky (2009) Harris, J., & Zaritsky, D. 2009, Astronomical Journal, 138, 1243, doi: 10.1088/0004-6256/138/5/1243
* Haschke et al. (2012) Haschke, R., Grebel, E. K., Duffau, S., & Jin, S. 2012, AJ, 143, 48, doi: 10.1088/0004-6256/143/2/48
* Hidalgo et al. (2018) Hidalgo, S. L., Pietrinferni, A., Cassisi, S., et al. 2018, ApJ, 856, 125, doi: 10.3847/1538-4357/aab158
* Holtzman et al. (1999) Holtzman, J. A., Gallagher, III, J. S., Cole, A. A., et al. 1999, Astronomical Journal, 118, 2262, doi: 10.1086/301097
* Hunter et al. (2003) Hunter, D. A., Elmegreen, B. G., Dupuy, T. J., & Mortonson, M. 2003, AJ, 126, 1836, doi: 10.1086/378056
* Hunter (2007) Hunter, J. D. 2007, Computing In Science & Engineering, 9, 90, doi: 10.1109/MCSE.2007.55
* Iorio & Belokurov (2020) Iorio, G., & Belokurov, V. 2020, arXiv e-prints, arXiv:2008.02280. https://arxiv.org/abs/2008.02280
* Jones et al. (2001–) Jones, E., Oliphant, T., Peterson, P., et al. 2001–, SciPy: Open source scientific tools for Python
* Kaluzny et al. (2014) Kaluzny, J., Rozyczka, M., Pych, W., & Thompson, I. B. 2014, Acta Astron., 64, 77. https://arxiv.org/abs/1405.5750
* Kaluzny & Udalski (1992) Kaluzny, J., & Udalski, A. 1992, Acta Astron., 42, 29
* Karczmarek et al. (2017) Karczmarek, P., Wiktorowicz, G., Iłkiewicz, K., et al. 2017, Monthly Notices of the Royal Astronomical Society, 466, 2842, doi: 10.1093/Monthly Notices of the Royal Astronomical Society/stw3286
* Kerber et al. (2009) Kerber, L. O., Girardi, L., Rubele, S., & Cioni, M. R. 2009, A&A, 499, 697, doi: 10.1051/0004-6361/200811118
* Kervella et al. (2019a) Kervella, P., Gallenne, A., Remage Evans, N., et al. 2019a, A&A, 623, A116, doi: 10.1051/0004-6361/201834210
* Kervella et al. (2019b) Kervella, P., Gallenne, A., Evans, N. R., et al. 2019b, A&A, 623, A117, doi: 10.1051/0004-6361/201834211
* Kim et al. (2002) Kim, Y.-C., Demarque, P., Yi, S. K., & Alexander, D. R. 2002, ApJS, 143, 499, doi: 10.1086/343041
* Kinman & Brown (2010) Kinman, T. D., & Brown, W. R. 2010, Astronomical Journal, 139, 2014, doi: 10.1088/0004-6256/139/5/2014
* Kontizas et al. (1993) Kontizas, M., Kontizas, E., & Michalitsianos, A. G. 1993, A&A, 269, 107
* Kruijssen et al. (2019) Kruijssen, J. M. D., Pfeffer, J. L., Reina-Campos, M., Crain, R. A., & Bastian, N. 2019, MNRAS, 486, 3180, doi: 10.1093/mnras/sty1609
* Layden (1995) Layden, A. C. 1995, AJ, 110, 2312, doi: 10.1086/117691
* Lin et al. (1995) Lin, D. N. C., Jones, B. F., & Klemola, A. R. 1995, ApJ, 439, 652, doi: 10.1086/175205
* Mackey & Gilmore (2003) Mackey, A. D., & Gilmore, G. F. 2003, MNRAS, 338, 85, doi: 10.1046/j.1365-8711.2003.06021.x
* Mackey & Gilmore (2004) —. 2004, MNRAS, 352, 153, doi: 10.1111/j.1365-2966.2004.07908.x
* Maoz & Badenes (2010) Maoz, D., & Badenes, C. 2010, Monthly Notices of the Royal Astronomical Society, 407, 1314, doi: 10.1111/j.1365-2966.2010.16988.x
* Maoz & Graur (2017) Maoz, D., & Graur, O. 2017, ApJ, 848, 25, doi: 10.3847/1538-4357/aa8b6e
* Maoz & Mannucci (2012) Maoz, D., & Mannucci, F. 2012, Publications of the Astronomical Society of Australia, 29, 447, doi: 10.1071/AS11052
* Maoz et al. (2012) Maoz, D., Mannucci, F., & Brandt, T. D. 2012, Monthly Notices of the Royal Astronomical Society, 426, 3282, doi: 10.1111/j.1365-2966.2012.21871.x
* Maoz et al. (2011) Maoz, D., Mannucci, F., Li, W., et al. 2011, Monthly Notices of the Royal Astronomical Society, 412, 1508, doi: 10.1111/j.1365-2966.2010.16808.x
* Maoz et al. (2014) Maoz, D., Mannucci, F., & Nelemans, G. 2014, Annual Reviews of Astronomy and Astrophysics, 52, 107, doi: 10.1146/annurev-astro-082812-141031
* Maoz & Sharon (2010) Maoz, D., & Sharon, K. 2010, Astrophysical Journal, 722, 1879, doi: 10.1088/0004-637X/722/2/1879
* Marconi et al. (2015) Marconi, M., Coppola, G., Bono, G., et al. 2015, Astrophysical Journal, 808, 50, doi: 10.1088/0004-637X/808/1/50
* Mateu et al. (2012) Mateu, C., Vivas, A. K., Downes, J. J., et al. 2012, Monthly Notices of the Royal Astronomical Society, 427, 3374, doi: 10.1111/j.1365-2966.2012.21968.x
* Mennekens et al. (2010) Mennekens, N., Vanbeveren, D., De Greve, J. P., & De Donder, E. 2010, Astronomy and Astrophysics, 515, A89, doi: 10.1051/0004-6361/201014115
* Mermilliod (2000) Mermilliod, J. C. 2000, Astronomical Society of the Pacific Conference Series, Vol. 211, Massive Clusters in the Milky Way and Magellanic Clouds, ed. A. Lançon & C. M. Boily, 43
* Meschin et al. (2014) Meschin, I., Gallart, C., Aparicio, A., et al. 2014, MNRAS, 438, 1067, doi: 10.1093/mnras/stt2220
* Milone et al. (2009) Milone, A. P., Bedin, L. R., Piotto, G., & Anderson, J. 2009, A&A, 497, 755, doi: 10.1051/0004-6361/200810870
* Moe et al. (2019) Moe, M., Kratter, K. M., & Badenes, C. 2019, ApJ, 875, 61, doi: 10.3847/1538-4357/ab0d88
* Mosser et al. (2014) Mosser, B., Benomar, O., Belkacem, K., et al. 2014, A&A, 572, L5, doi: 10.1051/0004-6361/201425039
* Nayak et al. (2016) Nayak, P. K., Subramaniam, A., Choudhury, S., Indu, G., & Sagar, R. 2016, MNRAS, 463, 1446, doi: 10.1093/mnras/stw2043
* Nelemans et al. (2013) Nelemans, G., Toonen, S., & Bours, M. 2013, in IAU Symposium, Vol. 281, Binary Paths to Type Ia Supernovae Explosions, ed. R. Di Stefano, M. Orio, & M. Moe, 225–231
* Olsen (1999) Olsen, K. A. G. 1999, Astronomical Journal, 117, 2244, doi: 10.1086/300854
* Olszewski et al. (1991) Olszewski, E. W., Schommer, R. A., Suntzeff, N. B., & Harris, H. C. 1991, AJ, 101, 515, doi: 10.1086/115701
* Olszewski et al. (1996) Olszewski, E. W., Suntzeff, N. B., & Mateo, M. 1996, Annual Reviews of Astronomy and Astrophysics, 34, 511, doi: 10.1146/annurev.astro.34.1.511
* Palma et al. (2016) Palma, T., Gramajo, L. V., Clariá, J. J., et al. 2016, A&A, 586, A41, doi: 10.1051/0004-6361/201527305
* Paxton et al. (2011) Paxton, B., Bildsten, L., Dotter, A., et al. 2011, ApJS, 192, 3, doi: 10.1088/0067-0049/192/1/3
* Paxton et al. (2013) Paxton, B., Cantiello, M., Arras, P., et al. 2013, ApJS, 208, 4, doi: 10.1088/0067-0049/208/1/4
* Paxton et al. (2015) Paxton, B., Marchant, P., Schwab, J., et al. 2015, ApJS, 220, 15, doi: 10.1088/0067-0049/220/1/15
* Paxton et al. (2018) Paxton, B., Schwab, J., Bauer, E. B., et al. 2018, ApJS, 234, 34, doi: 10.3847/1538-4365/aaa5a8
* Pellegrini et al. (2012) Pellegrini, E. W., Oey, M. S., Winkler, P. F., et al. 2012, ApJ, 755, 40, doi: 10.1088/0004-637X/755/1/40
* Pérez & Granger (2007) Pérez, F., & Granger, B. E. 2007, Computing in Science and Engineering, 9, 21, doi: 10.1109/MCSE.2007.53
* Petersen & Christensen-Dalsgaard (1996) Petersen, J. O., & Christensen-Dalsgaard, J. 1996, A&A, 312, 463
* Piatti (2017) Piatti, A. E. 2017, A&A, 606, A21, doi: 10.1051/0004-6361/201731246
* Piatti (2018) —. 2018, MNRAS, 475, 2553, doi: 10.1093/mnras/stx3344
* Pietrinferni et al. (2004) Pietrinferni, A., Cassisi, S., Salaris, M., & Castelli, F. 2004, ApJ, 612, 168, doi: 10.1086/422498
* Pietrinferni et al. (2006) —. 2006, ApJ, 642, 797, doi: 10.1086/501344
* Pietrzynski & Udalski (2000) Pietrzynski, G., & Udalski, A. 2000, Acta Astron., 50, 337. https://arxiv.org/abs/astro-ph/0010360
* Piskunov et al. (2008) Piskunov, A. E., Schilbach, E., Kharchenko, N. V., Röser, S., & Scholz, R. D. 2008, A&A, 477, 165, doi: 10.1051/0004-6361:20078525
* Popescu et al. (2012) Popescu, B., Hanson, M. M., & Elmegreen, B. G. 2012, ApJ, 751, 122, doi: 10.1088/0004-637X/751/2/122
* Pribulla et al. (2008) Pribulla, T., Rucinski, S., Matthews, J. M., et al. 2008, MNRAS, 391, 343, doi: 10.1111/j.1365-2966.2008.13889.x
* Prudil et al. (2020) Prudil, Z., Dékány, I., Grebel, E. K., & Kunder, A. 2020, arXiv e-prints, arXiv:2001.02486. https://arxiv.org/abs/2001.02486
* Rubele et al. (2012) Rubele, S., Kerber, L., Girardi, L., et al. 2012, A&A, 537, A106, doi: 10.1051/0004-6361/201117863
* Rubele et al. (2018) Rubele, S., Pastorelli, G., Girardi, L., et al. 2018, MNRAS, 478, 5017, doi: 10.1093/mnras/sty1279
* Ruiz-Lara et al. (2020) Ruiz-Lara, T., Gallart, C., Monelli, M., et al. 2020, A&A, 639, L3, doi: 10.1051/0004-6361/202038392
* Santana et al. (2013) Santana, F. A., Muñoz, R. R., Geha, M., et al. 2013, ApJ, 774, 106, doi: 10.1088/0004-637X/774/2/106
* Santana et al. (2016) Santana, F. A., Muñoz, R. R., de Boer, T. J. L., et al. 2016, ApJ, 829, 86, doi: 10.3847/0004-637X/829/2/86
* Schaerer et al. (1993) Schaerer, D., Meynet, G., Maeder, A., & Schaller, G. 1993, A&AS, 98, 523
* Schaller et al. (1992) Schaller, G., Schaerer, D., Meynet, G., & Maeder, A. 1992, AAPS, 96, 269
* Skowron et al. (2016) Skowron, D. M., Soszyński, I., Udalski, A., et al. 2016, Acta Astronomica, 66, 269. https://arxiv.org/abs/1608.00013
* Smecker-Hane et al. (2002) Smecker-Hane, T. A., Cole, A. A., Gallagher, III, J. S., & Stetson, P. B. 2002, Astrophysical Journal, 566, 239, doi: 10.1086/337985
* Smith (2004) Smith, H. A. 2004, RR Lyrae Stars, 166
* Soszyński et al. (2014) Soszyński, I., Udalski, A., Szymański, M. K., et al. 2014, Acta Astronomica, 64, 177. https://arxiv.org/abs/1410.1542
* Soszyński et al. (2016) —. 2016, Acta Astronomica, 66, 131. https://arxiv.org/abs/1606.02727
* Stanway & Eldridge (2018) Stanway, E. R., & Eldridge, J. J. 2018, MNRAS, 479, 75, doi: 10.1093/mnras/sty1353
* Surot et al. (2019) Surot, F., Valenti, E., Hidalgo, S. L., et al. 2019, A&A, 623, A168, doi: 10.1051/0004-6361/201833550
* Tayar et al. (2017) Tayar, J., Somers, G., Pinsonneault, M. H., et al. 2017, ApJ, 840, 17, doi: 10.3847/1538-4357/aa6a1e
* Toonen et al. (2013) Toonen, S., Nelemans, G., Bours, M., et al. 2013, 18th European White Dwarf Workshop., 469, 6. https://arxiv.org/abs/1302.0495
* Totani et al. (2008) Totani, T., Morokuma, T., Oda, T., Doi, M., & Yasuda, N. 2008, Publications of the Astronomical Society of Japan, 60, 1327, doi: 10.1093/ Publications of the Astronomical Society of Japan/60.6.1327
* Udalski et al. (2015) Udalski, A., Szymański, M. K., & Szymański, G. 2015, Acta Astronomica, 65, 1. https://arxiv.org/abs/1504.05966
* van der Walt et al. (2011) van der Walt, S., Colbert, S. C., & Varoquaux, G. 2011, Computing in Science & Engineering, 13, 22, doi: 10.1109/MCSE.2011.37
* Wagner-Kaiser & Sarajedini (2013) Wagner-Kaiser, R., & Sarajedini, A. 2013, MNRAS, 431, 1565, doi: 10.1093/mnras/stt277
* Weisz et al. (2014) Weisz, D. R., Dolphin, A. E., Skillman, E. D., et al. 2014, ApJ, 789, 147, doi: 10.1088/0004-637X/789/2/147
* Williams et al. (2017) Williams, B. F., Dolphin, A. E., Dalcanton, J. J., et al. 2017, ApJ, 846, 145, doi: 10.3847/1538-4357/aa862a
* Winkler et al. (2005) Winkler, P. F., Young, A. L., Braziunas, D., et al. 2005, in American Astronomical Society Meeting Abstracts, Vol. 207, American Astronomical Society Meeting Abstracts, 132.03
* Yi et al. (2003) Yi, S. K., Kim, Y.-C., & Demarque, P. 2003, ApJS, 144, 259, doi: 10.1086/345101
* Zapartas et al. (2017) Zapartas, E., de Mink, S. E., Izzard, R. G., et al. 2017, Astronomy & Astrophysics, 601, A29, doi: 10.1051/0004-6361/201629685
* Zaritsky & Harris (2004) Zaritsky, D., & Harris, J. 2004, ApJ, 604, 167, doi: 10.1086/381795
* Zaritsky et al. (1997) Zaritsky, D., Harris, J., & Thompson, I. 1997, AJ, 114, 1002, doi: 10.1086/118531
* Zaritsky et al. (2004) Zaritsky, D., Harris, J., Thompson, I. B., & Grebel, E. K. 2004, AJ, 128, 1606, doi: 10.1086/423910
* Zhang et al. (2004) Zhang, X. B., Deng, L., Zhou, X., & Xin, Y. 2004, MNRAS, 355, 1369, doi: 10.1111/j.1365-2966.2004.08418.x
* Zinn et al. (2019) Zinn, R., Chen, X., Layden, A. C., & Casetti-Dinescu, D. I. 2019, MNRAS, 3209, doi: 10.1093/mnras/stz3580
## Appendix A Choice of Binning
Since we are calculating the DTD non-parametrically, the DTD depends on the
binning we choose for the ages. The SAD comes with a native resolution of 16
age-bins. We can decide on which combination of these bins provide the optimal
DTD through the Bayesian Information Criteria. These are defined as follows,
$BIC=k\ \mathrm{ln}(N)-2\ \mathrm{ln}(\mathcal{L}_{max})$ (A1)
where $k$ is the number of age-bins, $N$ is the number of SAD cells and ln
$\mathcal{L}_{max}$ is the maximum likelihood. The binning scheme that
minimizes BIC (i.e., the information loss) is favored. We calculate DTDs for
different age binnings using just the best-fit SAD (i.e., no randomized SADs)
and the original OGLE-IV sample inside the SAD area.
We show eight example binning schemes for our DTDs in Figure 10, which differ
in how the young and old ages are binned. The four binning schemes in the
bottom row have the smallest bin-sizes for ages $>0.8$ Gyr and varying bin-
sizes for young ages. The schemes in the top row have coarser resolution in
the oldest age-bins. More binning schemes similar to these are possible and
can be tested, but these eight provide a general sense of the impact of
binning. The DTDs measured in the detected bins vary at most by a factor of 2,
and are generally around $10^{-5}$ RR Lyrae per M⊙. Statistical errors in the
DTDs increase with the number of bins, because we are essentially increasing
the number of fitting parameters. Binning schemes that retain the highest
resolution in the oldest bins measure smaller values of BIC, irrespective of
the binning in the younger age-bins. This implies that the RR Lyrae DTD has
the strongest signal in the oldest bins, which is consistent with an older
stellar origin of RR Lyrae.
Apart from BIC, we also show the acceptance fraction of the MCMC solver,
$a_{f}$ in Figure 10. This is the fraction of new steps accepted by the emcee
walkers as the scan the multi-dimensional parameter space. While not a model
selection statistic like BIC, values of $a_{f}$ = 0.2–0.5 indicate that the
emcee algorithm is performing optimally (Foreman-Mackey et al., 2013). Values
of $a_{f}$ which are too low indicate multiple peaks in the posterior space
separated by valleys of “low probability”, while high values imply that the
walkers are simply random-walking, with no regard for the target probability
density. We note that while binning schemes in the second row have similar
values of BIC, they have different $a_{f}$, with the native resolution of the
SAD (last panel) having the smallest $a_{f}$.
Based on these tests, we choose the binning scheme 7 which has the smallest
value of BIC, and the largest value of $a_{f}$ = 0.28 in Figure 10.
Figure 10: RR Lyrae DTDs calculated with different age binning. The scheme
number is shown in the top right corner of each panel. The acceptance fraction
$a_{f}$ of the MCMC solver and the value of the BIC test (BIC) are shown in
each panel. Shaded grey regions mark age-bins with significant detections,
while arrows show 2$\sigma$ upper limits. We use the binning scheme plotted at
bottom right, as it has the smallest information loss, and also has the
highest acceptance fraction.
## Appendix B Uncertainties using the B15 method
When generating the randomized SADs, we assumed that the uncertainties of the
stellar masses were normally distributed. However, this is only an
approximation, and the underlying shape of the probability distribution of
masses is unknown. We therefore also calculated the DTD uncertainties based on
the method in B15, which can be treated as a more conservative estimate of the
uncertainties. In B15, the 1 $\sigma$ uncertainty due to the SAD is equal to
the difference between the DTDs for the best fit SAD, and DTDs for the 68$\%$
upper and lower limits on the SADs. This difference is added in quadrature to
the statistical uncertainties in the best-fit DTD, and the total value is used
for assessing detectability. The uncertainties using this method are larger,
and as a result the DTD (Figure 11) is different from the DTD in Figure 3.
Nevertheless, it still shows significant signal below 10 Gyrs, particularly in
the age range 2-8 Gyrs. The signals in the other age-bins fall below the
2$\sigma$ limit. The total contribution below 10 Gyrs in this case is about
41.6$\%$. Our main science result,$-$ that the DTD has statistically
significant signal below 10 Gyrs $-$ is unchanged, even with our most
conservative estimates of the uncertainties.
Figure 11: Same as Figure 3, but with uncertainties calculated using the
method in Badenes et al. (2015), described in Section B.
|
# Spacetime Quantum Reference Frames and superpositions of proper times
Flaminia Giacomini<EMAIL_ADDRESS>Perimeter Institute for
Theoretical Physics, 31 Caroline St. N, Waterloo, Ontario, N2L 2Y5, Canada
###### Abstract
In general relativity, the description of spacetime relies on idealised rods
and clocks, which identify a reference frame. In any concrete scenario,
reference frames are associated to physical systems, which are ultimately
quantum in nature. A relativistic description of the laws of physics hence
needs to take into account such quantum reference frames (QRFs), through which
spacetime can be given an operational meaning.
Here, we introduce the notion of a spacetime quantum reference frame,
associated to a quantum particle in spacetime. Such formulation has the
advantage of treating space and time on equal footing, and of allowing us to
describe the dynamical evolution of a set of quantum systems from the
perspective of another quantum system, where the parameter in which the rest
of the physical systems evolves coincides with the proper time of the particle
taken as the QRF. Crucially, the proper times in two different QRFs are not
related by a standard transformation, but they might be in a quantum
superposition one with respect to the other.
Concretely, we consider a system of $N$ relativistic quantum particles in a
weak gravitational field, and introduce a timeless formulation in which the
global state of the $N$ particles appears “frozen”, but the dynamical
evolution is recovered in terms of relational quantities. The position and
momentum Hilbert space of the particles is used to fix the QRF via a
transformation to the local frame of the particle such that the metric is
locally inertial at the origin of the QRF. The internal Hilbert space
corresponds to the clock space, which keeps the proper time in the local frame
of the particle. Thanks to this fully relational construction we show how the
remaining particles evolve dynamically in the relational variables from the
perspective of the QRF. The construction proposed here includes the Page-
Wootters mechanism for non interacting clocks when the external degrees of
freedom are neglected. Finally, we find that a quantum superposition of
gravitational redshifts and a quantum superposition of special-relativistic
time dilations can be observed in the QRF.
## 1 Introduction
In both quantum theory and general relativity, spacetime is treated as an
abstract entity. In quantum theory, spacetime is considered as a fixed, non-
dynamical background providing an arena for physical phenomena. In general
relativity, spacetime is a dynamical quantity, which is influenced by a change
in the configuration of massive bodies. In both cases, spacetime coordinates
acquire meaning via rods and clocks, which identify a reference frame.
However, these rods and clocks, and hence reference frames, are typically
treated as idealised classical systems. In this sense, the description of
spacetime is not fully operational.
Reference frames are very useful to specify the point of view from which
observations are carried out. Although measurements are usually made in a
specific reference frame, all physical laws are formulated in a way that is
independent of the reference frame chosen, thanks to the principle of general
covariance. This means that there is no preferred reference frame.
The rods and clocks which specify the reference frames are physical systems,
and ultimately quantum in nature. Hence, an operational formulation of the
laws of physics requires being able to describe physical phenomena from the
point of view of such quantum systems, which can be in a superposition or
entangled with other physical systems. This is the main intuition behind the
notion of Quantum Reference Frames (QRFs).
QRFs have been studied in quantum information, quantum foundations, and
quantum gravity since 1967. In the quantum information literature [1, 2, 3, 4,
5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21], they have been
used to overcome superselection rules, applied to quantum communication
scenarios, and related to invariance properties and symmetries of quantum
systems. In the quantum gravity literature, it has been conjectured that QRFs
are needed to formulate a quantum theory of gravity [22, 23], especially when
taking a relational perspective on spacetime [24]. In particular, a relational
approach to QRFs is common to both the quantum gravity and quantum information
approaches. Recently, Ref. [25] introduced a formalism to describe physics
from the point of view of a specific QRF and transformations to change the
description between different QRFs. Subsequent work developed this approach
further in different contexts. Refs. [26, 27, 28] adopt a symmetry principle
to describe a Galilean system in a fully relational way, and Refs. [29, 30]
extend the formalism to the special relativistic regime, applying it to
concrete problems in Relativistic Quantum Information. Other works [31, 32,
33] have focussed on the group properties of QRFs both in discrete and in
continuous-variable systems. A time version of QRFs, also related to the Page-
Wootters mechanism [34, 35], has been introduced in Ref. [36] for interacting
clocks. Recently, a generalisation of the formalism has been applied to to
quantum systems living on a superposition of gravitational fields, and showed
that the Einstein Equivalence Principle can be extended to QRFs [37].
The work on QRFs is related to research on quantum clocks, which have been
studied with a relational approach, both in the interacting and non-
interacting case, in Refs. [38, 39, 40, 41, 42, 43]. In particular, Ref. [43]
studied relativistic clocks by extending the Page-Wootters approach to a
relativistic particle with internal and external degrees of freedom. A
conceptually similar approach to QRFs, but with some important differences, is
the quantum coordinate systems [44, 45].
All these works have studied QRFs either in space or in time, and separately
considered the role of internal and external variables (such as, respectively,
position or momentum and internal quantities acting as clocks) in identifying
the QRF. So far, no unifying approach to QRFs in spacetime has been
formulated. A spacetime formulation of QRFs is a crucial step towards a fully
relational and covariant description of physics from the point of view of a
quantum system, and could also provide a natural setting for a quantum
approach to time. However, achieving such a formulation faces some important
challenges, such as describing space and time in quantum theory on an equal
footing. This task requires the incorporation of a time operator into the
description of QRFs in space, and is conceptually related to the problem of
time in quantum gravity [46, 47, 48]. The methodology and the scope of the
present work are different to those employed in the canonical approaches to
the problem of time.
In this work, we overcome the difficulty of incorporating a time operator into
QRFs and develop a model for spacetime QRFs. We introduce a timeless and fully
relational formulation, which methodologically adopts techniques from
different approaches to QRFs [25, 26, 27, 36, 37], unifying them by adopting
aspects of Covariant Quantum Mechanics [49, 50, 51] and the Page-Wootters
mechanism [34, 35]. In particular, we consider a set of $N$ relativistic
quantum particles in a weak gravitational field, each of which has a quantum
state living in the tensor product Hibert space of position/momentum and some
internal “clock” Hilbert space (see Refs. [52, 43] for a similar analysis in a
different context). In this timeless formulation, the global state of the $N$
particles is “frozen”, but the dynamical evolution is recovered in terms of
the relational variables to one of the particles, which is chosen as the QRF.
In order to obtain this relational description on the reduced set of $N-1$
particles, we build a transformation to some relational variables (e.g.,
relative positions) of the particle chosen as the QRF, and then fix the origin
of the QRF at the location of this particle. The evolution of the remaining
$N-1$ particles is parametrised in terms of the proper time of the particle
serving as the QRF. Concretely, the proper time of each particle is encoded in
its internal variables, which serve as quantum clocks. Hence, both the
external and the internal variables play an important role in buiding the
spacetime QRF: the former allow us to transform to the local frame of a
quantum particle, whose spacetime position can be in a superposition or
entangled with other systems, and the latter allow us to describe the
dynamical evolution of the remaining particles in terms of the proper time in
each QRF. Notice that, although proper time is a standard evolution parameter
in each QRF, the proper times of two different quantum particles are not in a
classical relation one with respect to the other, but can be in a quantum
superposition. This joint use of the external variables to fix the QRF and the
internal variables as clocks is one of the novel aspects of this work.
In addition, we show that we can always find a transformation to the QRF of
the particle such that the metric is locally inertial at the origin of the
QRF. This is an instance of a Quantum Locally Inertial Frame (QLIF), as
introduced in Ref. [37]. We find the relational dynamics of the remaining
particles from the point of view of the chosen QRF, which coincides, in the
appropriate limit, with the dynamical evolution predicted with different
methods (for instance, see Refs. [53, 54, 55, 56]). The expression of this
relational dynamics has the same functional form in any chosen QRF, and is
thus symmetric under QRF transformations, as defined in Ref. [25].
Finally, we find that this description of QRFs in spacetime allows us to
observe a _superposition of special relativistic time dilations_ and a
_superposition of gravitational redshifts_ , which could be measured
experimentally in the future.
Overall, our model paves the way for a full extension of the formalism of QRFs
to arbitrary spacetimes, and superpositions thereof. This is an important step
to achieve a fully relational formulation of physics on a nonclassical
spacetime, which is a key requirement to formulate a quantum theory of
gravity.
The structure of the paper is as follows. In Section 2, we introduce the model
for spacetime QRFs: we start from a fully constrained description of the
system, which is neutral to the specific perspective, and is encoded in a set
of first-class constraints. We then show how to recover a relational
description from the perspective of one of the particles. In Section 3, we
show that this construction leads to an extension of the spacetime symmetries
to the set of QRF transformations. This result generalises to the spacetime
picture and to quantum clocks the notion of extended symmetry transformation
introduced in Ref. [25]. In Section 4, we introduce a measurement model and
show how the phenomenon of _superposition of special relativistic time
dilation_ and _superposition of gravitational redshift arises_. For clarity,
the technicalities are kept to the essential in the main text, however the
relevant calculations are detailed in the Appendices, which are referred to
where appropriate.
## 2 From a covariant timeless model to relational dynamics in a locally
inertial quantum reference frame
[scale=0.35]GlobalModel
Figure 1: We consider a system of $N$ relativistic quantum particles in a
weak gravitational field produced by a mass $M$. Each particle has a quantum
state representing its position in the spacetime diagram, as well as a clock
state (the hands of the clocks) which keeps the proper time in the particle’s
frame. We introduce a timeless model, in which the global state of the $N$
particles and of the mass $M$ is “frozen”, and is described by an $N$-particle
quantum state in spacetime. Concretely, this means that the quantum state of
the external variables can be, for instance, in a quantum superposition of
coordinate times $x^{0}$ and spatial coordinates $\mathbf{x}$. Classically,
the position in spacetime and the velocity of a particle influence its proper
time due to relativistic time dilation. When the particle is in a quantum
superposition of positions or velocities, the proper time displayed by its
internal clock is also in a quantum superposition from an external
perspective.
While, at the global level, the system does not evolve, the dynamical
evolution is recovered in terms of the relational variables between one of the
particles, which is chosen as the quantum reference frame (QRF), and the rest
of the particles. To transform to the QRF of one of the particles, we first
map the initial spacetime coordinates to the relational spacetime coordinates
from the perspective of the particle chosen as the QRF. We then find a
transformation which makes the metric locally inertial at the origin of the
QRF. Finally, we use the proper time of the particle to parametrise the
dynamics of the remaining particles from the perspective of the chosen QRF. In
such QRF, the dynamical evolution of the remaining particles is described in
terms of a Hamiltonian operator and is parametrised by the clock’s proper
time, which is just a classical parameter in the clock’s rest frame.
In this section, we introduce a fully relational model for $N$ quantum
particles in spacetime, where
* •
Space and time are treated on equal footing, as in special and general
relativity, also when the systems under study are quantum, and not classical;
* •
Time evolution emerges from internal correlations of physical clocks, which
are described as living in the internal Hilbert spaces of the particles. The
way in which the dynamical laws are recovered from the perspective of such
clocks incorporates the Page-Wootters formulation [34, 35];
* •
The formalism is fully relational, i.e., all external spacetime structure is
eliminated.
Furthermore, we show that it is possible to take the perspective of one of
these quantum particles, and how the transformation to such QRF is
constructed.
We consider a system of N particles of mass $m_{I}$, for $I=1,\cdots,N$, each
of which has a quantum state living in the tensor product of an external
Hilbert space, representing its position in spacetime or momentum, and an
internal Hilbert space (the clock), and a mass $M$ which is the source of the
gravitational field, as illustrated in Fig. 1. The ideas we present in this
paper are described in their simplest form in the time and radial components
of a model with spherical symmetry. To simplify the presentation, we work in
$1+1$ dimensions, which allows us to neglect the angular part of the model. We
adopt a timeless formulation [34, 49, 50, 51, 35] describing the full state of
the particles. The intuitive idea behind these formulations is that the
dynamics of particles is an emergent phenomenon stemming from the relational
degrees of freedom. Formally, this means that the state of the $N$ particles
and of the mass $M$ does not evolve, but the dynamics is recovered by
“conditioning” on one of the systems considered. The global state of the $N$
particles and of the mass $M$ — called the _physical state_ — is then a
“frozen” (i.e., non-evolving) state $\ket{\Psi}_{ph}$ satisfying the relation
$\hat{C}\ket{\Psi}_{ph}=0,$ (1)
where $\hat{C}$ is a set of first-class constraints defined on the Hilbert
space of the $N$ particles and of the mass $M$. This fully constrained model
is not reference-frame specific, but corresponds to a neutral perspective. The
set of constraints encodes both the dynamics of the particles in spacetime and
the global symmetries (global translations in space and time) of the model. In
this Section, we discuss in detail the form of the constraints and their
physical meaning.
Differently to other formulations for QRFs, we here describe both the quantum
state associated to the position/momentum of the particle and the clock state,
and they both play a role in the identification of the QRF. In particular,
when one of the particles, say particle $1$, acts as a QRF, the external part
of the quantum state is used to fix the QRF (in our case, the origin of the
coordinate system in spacetime), while the internal part serves as a clock,
ticking according to the proper time in each particle’s frame. We will see
that imposing the constraints corresponds to removing absolute (or external)
variables from the description, and that the relational variables, defined
from the point of view of one of the physical systems considered, evolve
dynamically in a non-trivial way. In particular, the presence of the internal
clocks is crucial to recover the relational dynamics of the rest of the
particles from the point of view of the particle chosen as the QRF.
Throughout the paper, we use the following notation: Greek letters indicate
spacetime labels, i.e., $\mu=0,1$, while capital Latin letters label the
particles from $1$ to $N$. Vectors with no indices are, unless differently
specified, two-vectors in $1+1$ spacetime, i.e., $v=(v^{0},\mathbf{v})$, where
$v^{0}$ is the time component of the vector and $\mathbf{v}$ denotes the
spatial component of the vector. We use the Einstein’s convention on all sums,
unless explicitly stated.
We take the particles to be special-relativistic and in a Newtonian field
generated by a mass $M$. We describe the latter as the weak-field limit of a
more general gravitational field, corresponding to a metric
$\begin{split}&g_{00}=1+\frac{2\Phi(\mathbf{x}-\mathbf{x}_{M})}{c^{2}};\\\
&g_{01}=g_{10}=0;\\\ &g_{11}=-1,\end{split}$ (2)
where $\Phi(\mathbf{x})$ is the Newtonian potential due to the mass $m_{M}$ of
the system $M$ and $|\Phi(\mathbf{x})|/c^{2}\ll 1$ in the spacetime region
considered.
The external Hilbert space of each particle $I=1,\cdots,N$ is
$\mathcal{H}_{I}\simeq L^{2}(\mathbb{R}^{2})$. The mass $M$ is also assigned a
Hilbert space $\mathcal{H}_{M}\simeq L^{2}(\mathbb{R}^{2})$. The motion of any
of the $N$ particles in this external, weak gravitational field sourced by the
mass $M$ can be encoded in the expression
$\hat{C}_{I}=\sqrt{g^{00}(\hat{\mathbf{x}}_{I}-\hat{\mathbf{x}}_{M})}\hat{p}^{I}_{0}-\hat{\omega}_{p}^{I},$
(3)
where
$[\hat{x}^{\mu}_{I},\hat{p}_{\nu}^{I}]=\mathrm{i}\hbar\delta^{\mu}_{\nu}$,
$\hat{\omega}_{p}^{I}=\sqrt{m_{I}^{2}c^{2}+\hat{\mathbf{p}}^{2}_{I}}$ with
$I=1,\cdots,N$, and
$[\hat{x}^{\mu}_{M},\hat{p}_{\nu}^{M}]=\mathrm{i}\hbar\delta^{\mu}_{\nu}$. The
constraints $\hat{C}_{I}$ of Eq. (3) straightforwardly follow from the
quantisation of the (classical) general-relativistic dispersion relation
$g^{\mu\nu}(x)p_{\mu}p_{\nu}-m^{2}c^{2}=0$ thanks to the weak-field
approximation, which is crucial to avoid ordering ambiguities. Notice that,
for simplicity, we assume that the state of the mass $M$ does not undergo any
dynamical evolution in this initial description, hence we do not associate a
dynamical constraint to it. In the following we show that the mass acquires a
dynamical evolution as a result of taking the perspective of one of the
particles.
By applying Eq. (3) to a quantum state $\ket{\psi_{I}}_{ph}$ that solves the
constraint, i.e., $\hat{C}_{I}\ket{\psi_{I}}_{ph}=0$ and conditioning on the
time coordinate $\ket{x^{0}_{I}}_{I}$ via the procedure
$\braket{x^{0}_{I}}{\hat{C}_{I}}{\psi_{I}}_{ph}=0$, it is possible to recover
the Schrödinger equation for a quantum particle in the Newtonian field (see
Appendix A for details). Hence, the dynamical evolution emerges from the
entanglement, at the level of $\ket{\psi_{I}}_{ph}$, between the time and
spatial degrees of freedom.
We are working here in a perturbative regime in which the commutator
$\left[\sqrt{g_{00}(\hat{\mathbf{x}}_{I}-\hat{\mathbf{x}}_{M})},\hat{\omega}_{p}^{I}\right]$
is negligible. This is equivalent to demanding that all terms which are at
least of the order
$\frac{\Phi(\hat{\mathbf{x}}_{I}-\hat{\mathbf{x}}_{M})\hat{\mathbf{p}}^{2}_{I}}{m_{I}^{2}c^{4}}$
are negligible whenever they are applied to the quantum state of each particle
$I$. Notice that we retain terms of the order
$\left[\hat{\mathbf{p}}_{I}/(m_{I}c)\right]^{4}$, corresponding to special-
relativistic corrections in the velocity of the particles.
The constraint $\hat{C}_{I}$ encoding the dynamical evolution of the particle,
which holds for each particle of the set we consider, is not yet cast in
relational terms. The introduction, compared to the standard quantum-
mechanical formalism, of the $\hat{x}_{I}^{0}$ and $\hat{p}_{0}^{I}$
operators, such that
$[\hat{x}_{I}^{0},\hat{p}_{0}^{J}]=\mathrm{i}\hbar\delta_{I}^{J}$ is necessary
to treat space and time on the same footing, as in the case of Covariant
Quantum Mechanics [49, 50]. In order to enforce the relational character of
the formalism, we need to eliminate the external structure. In Appendix B we
review and compare different relational approaches.
In the following, we show how to build a fully relational model for quantum
particles in spacetime and in a weak gravitational field. Methodologically,
this construction adopts the tools of QRFs, and unifies the framework for
spatial QRFs [25, 26, 27] with time reference frames [36], the latter in the
non-interacting case. The resulting model utilises techniques of Covariant
Quantum Mechanics in order to recover a full spacetime covariance, and
incorporates both external and internal degrees of freedom.
In particular, we consider two constraints on top of the $N$ constraints of
Eq. (3), one for each spacetime dimension, which encode the conservation of
the total energy and the total momentum of our $N$-particle model. Thus, we
write
$\begin{split}&\hat{f}^{0}=\sum_{I=1}^{N}\left[\hat{p}^{I}_{0}+\Delta(\hat{\mathbf{x}}_{I}-\hat{\mathbf{x}}_{M},\hat{\mathbf{p}}_{I})\frac{\hat{H}_{I}}{c}\right]+\hat{p}_{0}^{M};\\\
&\hat{f}^{1}=\sum_{I=1}^{N}\hat{\mathbf{p}}_{I}+\hat{\mathbf{p}}_{M},\end{split}$
(4)
where
$\Delta(\hat{\mathbf{x}}_{I}-\hat{\mathbf{x}}_{M},\hat{\mathbf{p}}_{I})=\sqrt{g_{00}(\hat{\mathbf{x}}_{I}-\hat{\mathbf{x}}_{M})}\left(1+\frac{\hat{\mathbf{p}}^{2}_{I}}{m_{I}^{2}c^{2}}\right)^{-1/2}$.
The constraint $\hat{f}^{1}$ corresponds to the global translational
invariance required in Refs. [26, 27]. The constraint $\hat{f}^{0}$
corresponds to the conservation of the total energy. Notice that particle $M$
counts towards the total energy and momentum balance because, although its
state does not obey a dynamical constraint, it is a quantum system in its own
right, with energy and momentum associated to it. Such quantum state in
spacetime should be chosen so that it corresponds to a physical configuration
of the mass $M$ which does not dynamically evolve in the initial coordinates.
The constraint $\hat{f}^{0}$ is derived by summing all the contributions to
the energies (up to a multiplicative factor of $c$) coming from the motion of
the particles in spacetime, from the internal Hamiltonian $\hat{H}_{I}$,
living in the Hilbert space $\mathcal{H}_{C_{I}}\simeq L^{2}(\mathbb{R})$ of
the $N$ particles, and from the system $M$. In particular, the factor in front
of the internal energy Hamiltonian corresponds to the relativistic time
dilation from the rest frame to an arbitrary frame. The relation between the
rest frame and an external frame to the particle is quantum, because the
particles have a quantum state associated to the position/momentum, which can
be in a quantum superposition of positions and velocities. QRF techniques
justify operationally the change between the rest frame of a quantum system
and an arbitrary frame [25, 29, 30]. The relation between the internal
energies in the rest frame and in an arbitrary frame is encoded in the
operator
$\Delta(\hat{\mathbf{x}}_{I}-\hat{\mathbf{x}}_{M},\hat{\mathbf{p}}_{I})$,
which coincides with the worldline element of particle $I$. In Appendix C we
provide an explicit derivation of the operator
$\Delta(\hat{\mathbf{x}}_{I}-\hat{\mathbf{x}}_{M},\hat{\mathbf{p}}_{I})$.
The two constraints $\hat{f}^{0}$ and $\hat{f}^{1}$, being first-class
constraints, generate a gauge transformation on the canonically conjugated
variables. These constraints have the effect of reducing the number of degrees
of freedom of the full system to obtain the dynamics of the relative positions
between the particles. In the following we show that, by picking one of the
particles as the preferred point of view from which we describe the dynamics
of the remaining particles, the resulting dynamics can be seen as the dynamics
of the $N-1$ particles and the mass $M$ from the point of view of the chosen
particle.
We then see that the introduction of the zero component of the coordinate and
momentum operators not only allows us to treat the $N$-particle system in a
covariant way, but allows us to generalise the symmetry principle of Refs.
[26, 27] to all the spacetime components and to show the connection with the
Page-Wootters inspired approach of Ref. [36]. When only the internal state of
the particles is considered, the constraint $\hat{f}^{0}$ reduces to the Page-
Wootters constraint of the clocks. We stress here that the constraint
$\hat{f}^{0}$ can be equivalently seen as arising from a Page-Wootters
construction, which emphasises the relational approach to time, or from a
relational formulation of physics, which eliminates the background structure
by enforcing that the total energy is zero.
As a result of this construction, we have a fully constrained model describing
the $N$ particle system, where the full constraint is
$\hat{C}=\sum_{I=1}^{N}\mathcal{N}_{I}\hat{C}_{I}+z_{\mu}\hat{f}^{\mu},$ (5)
where $\mathcal{N}_{I}$, with $I=1,\cdots,N$ and $z_{\mu}$, with $\mu=0,1$ are
Lagrange multipliers. Notice that, to our order of approximation, all
constraints are first-class, i.e., they at least weakly commute with each
other. In the two cases $\left[\hat{C}_{I},\hat{f}^{1}\right]=0$ and
$\left[\hat{f}^{0},\hat{f}^{1}\right]=0$, the relation holds exactly. For the
commutator $\left[\hat{C}_{I},\hat{f}^{0}\right]=0$, the result can be easily
obtained by remembering that we are working in a regime where
$\left[g_{00}(\hat{\mathbf{x}}_{I}-\hat{\mathbf{x}}_{M}),\hat{\omega}_{p}^{I}\right]=0$
for all $I=1,\cdots,N$. Notice that, thanks to the fact that all the
constraints commute, it is possible, as we will show, to straightforwardly
build a state that satisfies all of them.
Here, we only consider non-interacting clocks in order to maintain the
commutative property of the constraints. Had we allowed for gravitational
interaction between the particles, we would have had non-commutative
constraints (namely, second-class constraints), which require a different
treatment to what is outlined here. We thus leave this question for future
work.
The physical state of Eq. (1) with the constraint $\hat{C}$ of Eq. (5) can
equivalently be written as
$\ket{\Psi}_{ph}\propto\int
d^{N}\mathcal{N}d^{2}ze^{\frac{\mathrm{i}}{\hbar}\mathcal{N}_{I}\hat{C}_{I}}e^{\frac{\mathrm{i}}{\hbar}z_{\mu}\hat{f}^{\mu}}\ket{\phi},$
(6)
where $\ket{\phi}$ is an arbitrary state that can be expanded in a basis as
$\ket{\phi}=\int\Pi_{I}\left[d\mu(x_{I})dE_{I}\right]d^{2}x_{M}\phi(x_{1},\cdots,x_{N},x_{M},E_{1},\cdots,E_{N})\ket{x_{1},\cdots,x_{N},x_{M}}\ket{E_{1},\cdots,E_{N}},$
(7)
with $d\mu(x_{I})=\sqrt{g_{00}(\mathbf{x}_{I}-\mathbf{x}_{M})}d^{2}x_{I}$
being the covariant integration measure and
$\ket{x_{I}}=\ket{x_{I}^{0},\mathbf{x}_{I}}$. The states $\ket{E_{I}}$ are the
eigenstates of the internal hamiltonian of each particle $I$, such that
$\hat{H}_{I}\ket{E_{I}}=E_{I}\ket{E_{I}}$. The state $\ket{\phi}$, in general,
is not a solution of the constraint. However, the state that solves the
constraint $\hat{C}\ket{\Psi}_{ph}=0$ can be obtained via the improper
projection $\hat{P}\ket{\phi}=\ket{\Psi}_{ph}$ defined in Eq. (6), where
$\hat{P}=\int
d^{N}\mathcal{N}d^{2}ze^{\frac{\mathrm{i}}{\hbar}\mathcal{N}_{I}\hat{C}_{I}}e^{\frac{\mathrm{i}}{\hbar}z_{\mu}\hat{f}^{\mu}}$.
The procedure to obtain the state $\ket{\Psi}_{ph}$ involves a technical
subtlety, due to the fact that $\hat{P}^{2}\neq\hat{P}$. For instance, in
situations in which the spectrum of the constraint is continuous around zero,
the solution the set of states satisfying the constraint is empty. However,
this problem can be solved, e.g., by redefining the inner product in order to
renormalise the state $\ket{\Psi}_{ph}$ (for details and different approaches
to the solution see, e.g., [57, 58, 59, 60]).
In order to take the perspective of a specific particle, e.g., particle $1$,
we need to first map the phase-space observables to the relational observables
to particle $1$, and then enforce that particle $1$ is in the origin of the
reference frame via a projection operation. This procedure extends the
technique introduced in Ref. [26, 27] to our model for $N$ relativistic
particles in a weak gravitational field. Overall, we obtain the relational
state from the perspective of particle $1$ to be
$\ket{\psi}^{(1)}={}_{1}\bra{q_{1}=0}\hat{\mathcal{T}}_{1}\ket{\Psi}_{ph},$
(8)
where we choose
$\hat{\mathcal{T}}_{1}=e^{\frac{i}{\hbar}\frac{\log\sqrt{g_{00}(\hat{\mathbf{x}}_{M})}}{2}\sum_{I=1}^{N}(\hat{x}^{0}_{I}\hat{p}_{0}^{I}+\hat{p}_{0}^{I}\hat{x}^{0}_{I})}e^{\frac{\mathrm{i}}{\hbar}\hat{\mathbf{x}}_{1}\left(\hat{f}^{1}-\hat{\mathbf{p}}^{1}\right)}e^{\frac{\mathrm{i}}{\hbar}\hat{x}_{1}^{0}\left(\hat{f}^{0}-\hat{p}^{1}_{0}\right)},$
(9)
and we have defined the positions after the transformation
$\hat{\mathcal{T}}_{1}$ as $q_{I}$.
The operator $\hat{\mathcal{T}}_{1}$ maps the initial positions to the
relative positions to particle $1$. Its construction can be intuitively
explained by distinguishing two parts of the operator. The two rightmost terms
map the spacetime coordinates $\hat{x}_{\ell}$, with $\ell=2,\cdots,N,M$ to
the relative coordinates to particle $1$, i.e.,
$\hat{x}_{\ell}-\hat{x}_{1}\mapsto\hat{x}_{\ell}$. The last term on the left,
which is a function of the metric, sets the metric field to a locally inertial
metric field at the location of particle $1$. In order to achieve this, the
metric at the location of particle $1$ is mapped to the Minkowski metric, and
the motion of the other particles is encoded in a new constraint
$\hat{C}^{\prime}_{i}=\hat{\mathcal{T}}_{1}\hat{C}_{i}\hat{\mathcal{T}}_{1}^{\dagger}=\sqrt{\frac{g_{00}(\hat{\mathbf{q}}_{i}-\hat{\mathbf{q}}_{M})}{g_{00}(\hat{\mathbf{q}}_{M})}}\hat{k}_{0}^{i}-\hat{\omega}_{k}^{i},$
(10)
where $i=2,\cdots,N$, $\hat{q}_{i}$, $\hat{q}_{M}$ being the (spacetime)
position operators of the particles relative to particle $1$ in the new
coordinate system and $\hat{k}_{i}$, $\hat{k}_{M}$ the momentum operators that
are canonically conjugated respectively to $\hat{q}_{i}$ and $\hat{q}_{M}$,
i.e.,
$\left[\hat{q}_{i}^{\mu},\hat{k}_{\nu}^{j}\right]=\mathrm{i}\hbar\,\delta^{\mu}_{\nu}\,\delta_{ij}$
and
$\left[\hat{q}_{M}^{\mu},\hat{k}_{\nu}^{M}\right]=\mathrm{i}\hbar\,\delta^{\mu}_{\nu}$
(see Appendix D for the complete action of the operator
$\hat{\mathcal{T}}_{1}$ on the phase-space operators). We hence find that
$g^{\prime}_{00}(\hat{\mathbf{q}}_{i},\hat{\mathbf{q}}_{M})=\hat{\mathcal{T}}_{1}\frac{g_{00}(\hat{\mathbf{x}}_{i}-\hat{\mathbf{x}}_{M})}{g_{00}(\hat{\mathbf{x}}_{1}-\hat{\mathbf{x}}_{M})}\hat{\mathcal{T}}_{1}^{\dagger},$
(11)
is the new metric field in the perspective of particle $1$. In summary, the
transformation $\hat{\mathcal{T}}_{1}$ realises a transformation to the
Quantum Locally Inertial Frame (QLIF) centred in particle $1$, in the spirit
of Ref. [37]. A different choice of the operator $\hat{\mathcal{T}}_{1}$ is
possible, and would correspond to a different set of relational variables (and
coordinate system).
By performing a lengthy calculation (which is detailed in Appendix E), it is
possible to show that the state of Eq. (8) can be cast as
$\ket{\psi}^{(1)}=\int
d\tau_{1}e^{-\frac{\mathrm{i}}{\hbar}\hat{H}^{(1)}\tau_{1}}\ket{\psi^{(1)}_{0}}\ket{\tau_{1}},$
(12)
where $\ket{\tau_{1}}$ is the state of the internal clock $1$ such that, given
the internal time operator $\hat{T}_{I}$,
$\hat{T}_{I}\ket{\tau_{I}}=\tau_{I}\ket{\tau_{I}}$ and
$\left[\hat{T}_{I},\hat{H}_{J}\right]=\mathrm{i}\hbar\delta_{IJ}$, the
Hamiltonian $\hat{H}^{(1)}$ is
$\hat{H}^{(1)}=\hat{\gamma}_{\Sigma
p,1}\sum_{i}\sqrt{g^{\prime}_{00}(\hat{\mathbf{q}}_{i},\hat{\mathbf{q}}_{M})}\left\\{c\hat{\omega}_{k}^{i}+\hat{\gamma}_{i}^{-1}\hat{H}_{i}\right\\}+c\hat{\gamma}_{\Sigma
k,1}\sqrt{g^{00}(\hat{\mathbf{q}}_{M})}\hat{k}_{0}^{M}+m_{1}c^{2}\hat{\gamma}_{\Sigma
k,1}^{2},$ (13)
with $\hat{\gamma}_{\Sigma
k,1}=\sqrt{1+\frac{\left(\sum_{i}\hat{\mathbf{k}}_{i}+\hat{\mathbf{k}}_{M}\right)^{2}}{m_{1}^{2}c^{2}}}$,
$\hat{\gamma}_{i}=\sqrt{1+\frac{\hat{\mathbf{k}}_{i}^{2}}{m_{i}^{2}c^{2}}}$.
Finally, the state $\ket{\psi^{(1)}_{0}}$ formally plays the role of an
initial state, whose relation with the state $\ket{\phi}$ is given in Appendix
E. The state of Eq. (12) can be interpreted as a “history state”, namely a
state that associates to any arbitrary time eigenstate $\ket{\tau_{1}}$ read
in the QRF of particle $1$ a state
$\ket{\psi^{(1)}_{\tau_{1}}}=e^{-\frac{\mathrm{i}}{\hbar}\hat{H}^{(1)}\tau_{1}}\ket{\psi^{(1)}_{0}}$
of the remaining particles at the time $\tau_{1}$ read by the clock. A further
projection on the state of the clock $1$, i.e.,
$\braket{\tau_{1}}{\psi^{(1)}}$, recovers the standard description of quantum
mechanics.
From the expression in Eq. (13), we see that the Hamiltonian of the particles
$i=2,\cdots,N$ is affected by the presence of particle $1$. On the one hand,
this is because the metric field at the location of the other particles, in
the QLIF of particle $1$, is a quantum operator depending also on the initial
value of the metric field at the location of particle $1$, as it is clear by
looking at Eq. (11). This is an expression of the relational character of
general relativity, extended to when we consider quantum particles as QRFs:
the metric field in each QRF is defined purely in terms of relational
quantities (operators) of the particles and the reference frame. On the other
hand, there is also a contribution from the special-relativistic time-dilation
operator. This contribution is due to the motion of particle $1$ and appears
as the operator $\hat{\gamma}_{\Sigma k,1}$. The interpretation and the
operational consequences of this result are discussed in Section 4.
In this Section, we have outlined the method to reduce to the perspective of
one of the particle in the general model that we have introduced. However,
this model has a well-defined limit to the Galilean free-particle case, the
special relativistic case, and the Newtonian case of slowly-moving particles.
In all these cases, a similar procedure to what is described in this Section
can be defined. All these limiting cases are detailed in Appendix F.
## 3 Extended symmetries and QRF changes
The construction we have introduced in the previous section is compatible with
an extended notion of spacetime symmetries. This means that every quantum
particle can equally serve as a QRF, and that no preferred QRF is singled out.
This is easy to see by noticing that the procedure we have followed to
transform to the QLIF of particle $1$ can be repeated if we take another
particle, i.e., particle $2$, as the QRF. In this case, we obtain the same
result, but with all indices $1$ and $2$ swapped. The “history state” from the
point of view of particle $2$ is then
$\ket{\psi}^{(2)}=\int
d\tau_{2}e^{-\frac{\mathrm{i}}{\hbar}\hat{H}^{(2)}\tau_{2}}\ket{\psi^{(2)}_{0}}\ket{\tau_{2}},$
(14)
where all quantities are defined as in the previous section, and
$\hat{H}^{(2)}=\hat{\gamma}_{\Sigma u,2}\sum_{k\neq
2}\sqrt{g^{\prime}_{00}(\hat{\mathbf{r}}_{k},\hat{\mathbf{r}}_{M})}\left\\{c\hat{\omega}_{u}^{k}+\hat{\gamma}_{u}^{-1}\hat{H}_{k}\right\\}+c\hat{\gamma}_{\Sigma
u,2}\sqrt{g^{00}(\hat{\mathbf{r}}_{M})}\hat{u}_{0}^{M}+m_{2}c^{2}\hat{\gamma}_{\Sigma
u,2}^{2},$ (15)
with $\hat{\gamma}_{\Sigma u,2}=\sqrt{1+\frac{\left(\sum_{k\neq
2}\hat{\mathbf{u}}_{k}+\hat{\mathbf{u}}_{M}\right)^{2}}{m_{2}^{2}c^{2}}}$.
Here, $\hat{r}_{k}$, with $k=1,3,\cdots,N$, and $\hat{r}_{M}$ are the position
operators in the QRF of particle $2$, and $\hat{u}_{k}$, $\hat{u}_{M}$ their
conjugate momenta.
This construction then induces an invertible transformation from the reduced
“history state” of Eq. (8) to the reduced “history state” of Eq. (14) which
amounts to changing the QRF from the particle $1$ to the particle $2$. It is
then clear by construction, and can also be checked by direct computation,
that the state of Eq. (14) is obtained from Eq. (8) via the operation
$\ket{\psi}^{(2)}={}_{2}\bra{r_{2}=0}\hat{\mathcal{T}}_{12}\ket{\psi}^{(1)}\otimes\ket{p_{1}=0}_{1},$
(16)
where
$\hat{\mathcal{T}}_{12}=\hat{\mathcal{T}}_{2}\hat{\mathcal{T}}_{1}^{\dagger}$
(see Appendix D for the action of this operator on the phase space operators).
This construction parallels the one introduced in Ref. [26].
We have thus found that the Hamiltonian from the perspective of particle $1$
is mapped to a Hamiltonian that is form-invariant, but where all labels $1$
and $2$ are swapped, by a reversible QRF transformation. Thus, we have
generalised the notion of an extended symmetry introduced in Ref. [25] to the
full spacetime picture and quantum relativistic particles in a weak
gravitational field.
## 4 Gravitational redshift and relativistic time-dilation from a QRF
a) [scale=0.4]MeasurementSup b) [scale=0.4]SupTime
Figure 2: We depict the particular situation in which particle $2$ ($C_{2}$
in the picture) evolves in a superposition of two semi-classical trajectories
from the perspective of particle $1$ ($C_{1}$ in the picture). In the most
general case, the state of particle $2$ from the point of view of particle $1$
is arbitrary. We describe a measurement performed at time $\tau_{2}^{*}$ in
the frame of clock $2$ from the point of view of $C_{1}$. a) The measurement,
which is localised in time in the frame of $C_{2}$, appears delocalised in
time in the frame $C_{1}$. In particular, it is performed in a superposition
of proper times $\tau_{1}^{\prime}$ and $\tau_{1}^{\prime\prime}$ in the
proper time of $C_{1}$. The two times $\tau_{1}^{\prime}$ and
$\tau_{1}^{\prime\prime}$ are related to $\tau_{2}^{*}$ by the expressions
$\tau_{1}^{\prime}=\Delta_{12}^{-1}(W_{1})\tau_{2}^{*}$ and
$\tau_{1}^{\prime\prime}=\Delta_{12}^{-1}(W_{2})\tau_{2}^{*}$, where
$\Delta_{12}^{-1}(W_{i})$, $i=1,2$, encodes the special-relativistic time
dilation or the gravitational redshift evaluated on the worldline $W_{i}$. b)
In general, $C_{1}$ “sees” the clock $C_{2}$ as ticking in a superposition of
times, depending on the state of the external degrees of freedom.
Specifically, this effect can be understood as a superposition of special-
relativistic time dilations, when the particles move at relativistic
velocities in a Minkowski background, or as a superposition of gravitational
redshifts, when the particles move slowly in a weak gravitational field.
We now restrict our consideration to $N=2$ particles and a source mass $M$ and
introduce the possibility of performing a measurement in the QRF of particle
$2$. Analogously to the measurement procedure of Ref. [36], we modify the
constraint $\hat{f}^{0}$ as
$\begin{split}&\hat{C}_{I}=\sqrt{g^{00}(\hat{\mathbf{x}}_{I}-\hat{\mathbf{x}}_{M})}\hat{p}^{I}_{0}-\hat{\omega}_{p}^{I}\qquad\text{for}\qquad
I=1,2;\\\
&\hat{f}_{Q}^{0}=\hat{p}^{1}_{0}+\hat{p}^{2}_{0}+\hat{p}_{0}^{M}+\Delta(\hat{\mathbf{x}}_{1}-\hat{\mathbf{x}}_{M},\hat{\mathbf{p}}_{1})\frac{\hat{H}_{1}}{c}+\Delta(\hat{\mathbf{x}}_{2}-\hat{\mathbf{x}}_{M},\hat{\mathbf{p}}_{2})\left[\frac{\hat{H}_{2}}{c}+\delta(\hat{T}_{2}-\tau_{2}*)\frac{\hat{Q}_{2}}{c}\right];\\\
&\hat{f}^{1}=\hat{\mathbf{p}}_{1}+\hat{\mathbf{p}}_{2}+\hat{\mathbf{p}}_{M},\end{split}$
(17)
where $\hat{Q}_{2}$ is an observable commuting with every constraint111Notice
that this condition can be relaxed by adding an auxiliary quantum state as a
measurement device. For the sake of simplicity, we do not consider this case..
The same procedure adopted in Section 2 can be used to calculate the history
state, with the only _caveat_ that the operator $\hat{\mathcal{T}}_{1}$ needs
to be modified to
$\hat{\mathcal{T}}_{Q,1}=e^{\frac{i}{\hbar}\frac{\log\sqrt{g_{00}(\hat{\mathbf{x}}_{M})}}{2}\sum_{I=1}^{N}(\hat{x}^{0}_{I}\hat{p}_{0}^{I}+\hat{p}_{0}^{I}\hat{x}^{0}_{I})}e^{\frac{\mathrm{i}}{\hbar}\hat{\mathbf{x}}_{1}\left(\hat{f}^{1}-\hat{\mathbf{p}}^{1}\right)}e^{\frac{\mathrm{i}}{\hbar}\hat{x}_{1}^{0}\left(\hat{f}_{Q}^{0}-\hat{p}^{1}_{0}\right)}.$
(18)
With a similar procedure to the one adopted in Section 2, we can calculate the
physical state $\ket{\Psi}_{ph}$ and the history state
$\ket{\psi}^{(1)}={}_{1}\bra{q_{1}=0}\hat{\mathcal{T}}_{Q,1}\ket{\Psi}_{ph}$.
The calculations are detailed in Appendix G, and the resulting history state
is
$\ket{\psi}^{(1)}=\int
d\tau_{1}\overleftarrow{T}\left\\{e^{-\frac{\mathrm{i}}{\hbar}\int_{0}^{\tau_{1}}ds\left[\hat{H}^{(1)}+\Delta_{12}\delta(\hat{T}_{2}+\Delta_{12}s-\tau_{2}^{*})\hat{Q}_{2}\right]}\right\\}\ket{\psi^{(1)}_{0}}\ket{\tau_{1}},$
(19)
where $\overleftarrow{T}$ denotes the time-ordering operator, defined as
$\overleftarrow{T}[\hat{O}(t_{1})\hat{O}(t_{2})]=\hat{O}(t_{1})\hat{O}(t_{2})$
if $t_{1}>t_{2}$ and
$\overleftarrow{T}[\hat{O}(t_{1})\hat{O}(t_{2})]=\hat{O}(t_{2})\hat{O}(t_{2})$
if $t_{2}>t_{1}$ for $\hat{O}(t_{1})$, $\hat{O}(t_{2})$ being any two
arbitrary operators. We have also defined
$\Delta_{12}=\frac{\Delta(\hat{\mathbf{q}}_{2}-\hat{\mathbf{q}}_{M},\hat{\mathbf{k}}_{2})}{\sqrt{g_{00}(\hat{\mathbf{q}}_{M})}\hat{\gamma}_{\Sigma
k,1}^{-1}}=\sqrt{\frac{g_{00}(\hat{\mathbf{q}}_{2}-\hat{\mathbf{q}}_{M})}{g_{00}(\hat{\mathbf{q}}_{M})}}\frac{\hat{\gamma}_{2}^{-1}}{\hat{\gamma}_{\Sigma
k,1}^{-1}},$ (20)
as the “worldline operator” of particle $2$ in the perspective of clock $1$.
Notice that the metric field in Eq. (20) coincides with the metric
$g^{\prime}_{00}(\hat{\mathbf{q}}_{2},\hat{\mathbf{q}}_{M})$ in the QLIF of
particle $1$, which was derived in Section 2. The Hamiltonian $\hat{H}^{(1)}$
was defined in Eq. (13), and $\ket{\psi^{(1)}_{0}}$ is explicitly calculated
in Appendix G.
In order to understand how a measurement happening at time $\tau_{2}^{*}$ in
the frame of particle $2$ is seen in the QLIF of clock $1$, we restrict our
attention to the Galilean, special relativistic, and Newtonian cases, detailed
in Appendix F. We want to study the case in which the clock $2$ is an ideal
clock, which corresponds to the requirement that the initial state
$\ket{\psi_{0}^{(1)}}$ is sharp in the clock $2$ time variable.
Mathematically, this condition can be expressed as
$\psi_{0}^{(1)}(\mathbf{p}_{2},p_{M},t_{2})\propto\delta(t_{2}-t_{2}^{*})\psi_{0}^{(1)}(\mathbf{p}_{2},p_{M})$.
In the following, we choose $t_{2}^{*}=0$, equivalent to the initial
synchronisation of clocks $1$ and $2$. Our results are summarised in Fig. 2.
In the Galilean case of Appendix F.1, there is no mass $M$ sourcing the
gravitational field, and the special relativistic effects are not present.
Hence, $\Delta_{12}$ reduces to the identity operator. We can rewrite the
state of Eq. (19) as
$\ket{\psi}^{(1)}=\int_{-\infty}^{\tau_{2}^{*}}d\tau_{1}e^{-\frac{\mathrm{i}}{\hbar}\hat{H}^{(1)}\tau_{1}}\ket{\psi^{(1)}_{0}}\ket{\tau_{1}}+\int^{\infty}_{\tau_{2}^{*}}d\tau_{1}e^{-\frac{\mathrm{i}}{\hbar}\hat{H}^{(1)}(\tau_{1}-\tau_{2}^{*})}e^{-\frac{\mathrm{i}}{\hbar}\hat{Q}_{2}}e^{-\frac{\mathrm{i}}{\hbar}\hat{H}^{(1)}\tau_{2}^{*}}\ket{\psi^{(1)}_{0}}\ket{\tau_{1}}.$
(21)
In this case we see that, in the QRF of clock $1$, the measurement happens at
the same time $\tau_{2}^{*}$ as it would in the QRF of clock $2$.
Let us consider now the special-relativistic case described in Appendix F.2.
In this case, we take into account the special relativistic effects, but we do
not have a mass $M$ sourcing the gravitational field. The operator
$\Delta_{12}$ reduces to $\Delta_{12}=\hat{\gamma}_{\Sigma
k,1}/\hat{\gamma}_{2}\approx\sqrt{1+\frac{\hat{\mathbf{k}}_{2}^{2}}{m_{1}^{2}c^{2}}-\frac{\hat{\mathbf{k}}_{2}^{2}}{m_{2}^{2}c^{2}}}$.
The history state is
$\begin{split}\ket{\psi}^{(1)}=\int
d\mathbf{k}_{2}dt_{2}&\left\\{\int_{-\infty}^{\tau_{k}}d\tau_{1}e^{-\frac{\mathrm{i}}{\hbar}\hat{H}^{(1)}\tau_{1}}\psi_{0}^{(1)}(\mathbf{k}_{2},t_{2})\ket{\mathbf{k}_{2},t_{2}}+\right.\\\
+&\left.\int^{\infty}_{\tau_{k}}d\tau_{1}e^{-\frac{\mathrm{i}}{\hbar}\hat{H}^{(1)}(\tau_{1}-\tau_{k})}e^{-\frac{\mathrm{i}}{\hbar}\hat{Q}_{2}}\psi_{\tau_{k}}^{(1)}(\mathbf{k}_{2},t_{2})\ket{\mathbf{k}_{2},t_{2}}\right\\}\ket{\tau_{1}},\end{split}$
(22)
where $\ket{\psi_{\tau_{k}}^{(1)}}=\int
dt_{2}e^{-\frac{\mathrm{i}}{\hbar}\hat{H}^{(1)}\tau_{k}}\psi_{0}^{(1)}(\mathbf{k}_{2},t_{2})\ket{\mathbf{k}_{2},t_{2}}$
and $\tau_{k}=\Delta_{12}^{-1}(\mathbf{k}_{2})\tau_{2}^{*}$.
From the previous equation we see that the measurement happening at time
$\tau_{2}^{*}$ in the QRF of clock $2$ is seen by clock $1$ in a quantum
superposition of special-relativistic time dilations. The relation to the
usual time-dilation can easily be understood by comparison with the classical
case by considering two particles, $1$ and $2$, which measure the same
interval in coordinate time with their own proper time. In this case, the
relation $dx^{0}=d\tau_{1}\gamma_{1}=d\tau_{2}\gamma_{2}$ is precisely the
classical equivalent of what we found when we consider two clocks $1$ and $2$
in their own QRF.
Here, for each value $\mathbf{k}_{2}$ of the momentum of particle $2$, there
is a well-defined value of the time dilation, which coincides with the
classical one. However, since the relational degrees of freedom between the
QRF and particle $2$ are in a quantum superposition of momentum eigenstates,
we obtain a coherent superposition of special-relativistic time dilations,
which implies that clock $1$ sees the measurement as a delocalised event in
time.
These results are in agreement with those presented in Refs. [40, 43] in the
case of a quantum relativistic particle with a sharp momentum, or in a
superposition of two momenta222Formally, each sharp momentum state
$\ket{p^{*}}$ is described in Refs. [40, 43] by a gaussian wavepacket centred
in the momentum $p^{*}$, and following a semi-classical trajectory., and
generalise them to arbitrary quantum states of the particles.
Finally, we consider the Newtonian case described in Appendix F.3. In this
case, we take into account the gravitational effects due to the presence of
the mass $M$, but we do not consider the special-relativistic effects. Hence,
$\Delta_{12}=\sqrt{g^{\prime}_{00}(\hat{\mathbf{q}}_{2},\hat{\mathbf{q}}_{M})}=\sqrt{\frac{g_{00}(\hat{\mathbf{q}}_{2}-\hat{\mathbf{q}}_{M})}{g_{00}(\hat{\mathbf{q}}_{M})}}$
and the history state can be written as
$\begin{split}\ket{\psi}^{(1)}=\int
d\mathbf{q}_{2}d^{2}q_{M}dt_{2}&\left\\{\int_{-\infty}^{\tau_{q}}d\tau_{1}e^{-\frac{\mathrm{i}}{\hbar}\hat{H}^{(1)}\tau_{1}}\psi_{0}^{(1)}(\mathbf{q}_{2},q_{M},t_{2})\ket{\mathbf{q}_{2},q_{M},t_{2}}+\right.\\\
+&\left.\int^{\infty}_{\tau_{q}}d\tau_{1}e^{-\frac{\mathrm{i}}{\hbar}\hat{H}^{(1)}(\tau_{1}-\tau_{q})}e^{-\frac{\mathrm{i}}{\hbar}\hat{Q}_{2}}\psi_{\tau_{q}}^{(1)}(\mathbf{q}_{2},q_{M},t_{2})\ket{\mathbf{q}_{2},q_{M},t_{2}}\right\\}\ket{\tau_{1}},\end{split}$
(23)
where $\ket{\psi_{\tau_{q}}^{(1)}}=\int
dt_{2}e^{-\frac{\mathrm{i}}{\hbar}\hat{H}^{(1)}\tau_{q}}\psi_{0}^{(1)}(\mathbf{q}_{2},q_{M}t_{2})\ket{\mathbf{q}_{2},q_{M},t_{2}}$
and $\tau_{q}=\Delta_{12}^{-1}(\mathbf{q}_{2},\mathbf{q}_{M})\tau_{2}^{*}$.
Analogously to the special-relativistic case, we have that the measurement
happening at time $\tau_{2}^{*}$ in the QRF of clock $2$ is delocalised in the
proper time of clock $1$. However, the effect in this case is a quantum
superposition of gravitational redshifts, as can be easily seen by recalling
the standard expression of the gravitational redshift
$\tau_{obs}=\sqrt{\frac{g_{00}(x_{obs})}{g_{00}(x_{em})}}\tau_{em}$, where
$\tau_{obs}$ and $x_{obs}$ are respectively the proper time and spacetime
coordinates of the observer, and $\tau_{em}$ and $x_{em}$ are respectively the
proper time and spacetime coordinates of the emitter. If we identify the
observer with clock $1$ and the emitter with clock $2$, we recover the
relation
$\tau_{q}=\Delta_{12}^{-1}(\mathbf{q}_{2},\mathbf{q}_{M})\tau_{2}^{*}$.
Interestingly, this result can be equivalently obtained in the initial set of
coordinates, where
$\Delta_{12}=\sqrt{\frac{g_{00}(\hat{\mathbf{x}}_{2}-\hat{\mathbf{x}}_{M})}{g_{00}(\hat{\mathbf{x}}_{1}-\hat{\mathbf{x}}_{M})}}$,
or knowing that it is always possible to transform to a QLIF where the metric
is locally inertial at the location of clock $1$. In this case, it is enough
to calculate $\Delta_{12}$ using the metric in the QLIF, i.e.,
$\Delta_{12}=\sqrt{\frac{g_{00}(x_{em})}{g_{00}(x_{obs})}}=\sqrt{g^{\prime}_{00}(\hat{\mathbf{q}}_{2},\hat{\mathbf{q}}_{M})}$,
because $g_{00}(x_{obs})=1$ in the QLIF.
The two effects discussed here, namely the _superposition of special-
relativistic time dilations_ and the _superposition of gravitational
redshifts_ are, in principle observable. A promising experimental system to
carry out tests on the generalisation of these concepts could be atom
interferometry [61].
An interesting consequence of the delocalisation of the measurement that we
have described is that, if we fix some time $\tau_{1}=\tau_{1}^{*}$, clock $1$
in general “sees” a superposition of the measurement having taken place or
not. Hence, it seems that whether a measurement has occurred or not is a QRF-
dependent feature. It is, however, an open question whether this would allow
for a change of the causal relations. Previous work on QRFs [36] and on causal
reference frames [62] suggests that the causal relations are preserved under
change of time reference frame, but other approaches [44] might allow for a
change of the causal relations.
The time delocalisation of events has been studied in Ref. [36]. However, in
that case the mechanism causing the delocalisation was the interaction between
the internal degrees of freedom of the clocks, or the choice of a clock with
an unsharp initial state. Here, the delocalisation happens due to the fact
that the external degrees of freedom of the particle are in a quantum state,
which is unsharp in the relevant basis (momentum in the special-relativistic
case, and position in the Newtonian case). Other authors have also studied,
with different techniques to the one that is developed here, the consequences
of proper time running in a superposition due to the external degrees of
freedom being in a quantum superposition of positions or momenta [53, 54, 56,
43, 63]. However, these results rely on an external and classical reference
frame. Here, we generalise these effects to QRFs in a quantum relationship
with each other, and show that they are purely due to the relation between the
QRF and a physical system that is described.
Furthermore, we show here that regardless of the state of the particle taken
as a QRF in the global model, it is always possible to find a QLIF in which,
locally, the metric field is flat and the particle is at rest. This result
corroborates the generalisation of the Einstein Equivalence Principle of Ref.
[37] and provides a concrete model which describes the QRF change including
multiple particles. In particular, this model is suitable to describe an
experimental test, such as an interferometric experiment [64], to verify the
superposition of gravitational redshifts and the extension of the Einstein
Equivalence Principle. It would be interesting to generalise this result to
more general metric fields and to a superposition of metric fields as
considered in Ref. [37], where only the dynamics of a single system was
considered.
## 5 Discussion
In this work, we have introduced the notion of a spacetime quantum reference
frame, i.e., a reference frame in spacetime associated to a quantum system
which can be in a superposition, or entangled from the point of view of
another quantum system. We have developed a formulation to describe a set of
relativistic quantum particles in a weak gravitational field from the
perspective of such spacetime quantum reference frame. In order to achieve
this, we developed a covariant formulation describing both the quantum state
associated to the position/momentum of the particles and its internal state.
The former is used to fix the spacetime quantum reference frame, and the
latter is used as the internal clock of the quantum reference frame ticking
according to its own proper time. This formulation allows us to compare the
dynamical evolution in the proper time of different quantum clocks, when the
state of their external variables (position or momentum) is in a quantum
superposition from the perspective of one of them. We describe such clocks as
being attached to a quantum particle evolving with a Hamiltonian operator, and
not having a (semi)classical worldline. We find a transformation to the
Quantum Locally Inertial Frame of such clocks, introduced in Ref. [37], and we
describe a _superposition of special-relativistic time dilations_ and a
_superposition of gravitational redshifts_ from the perspective of such
spacetime quantum reference frames.
We adopt a timeless formulation, according to which the global state of $N$
particles appears “frozen”, but the dynamical evolution is recovered in terms
of relational quantities of the particles and the quantum reference frame.
Formally, we achieve this by imposing a set of constraints on the model, some
of which encode the free evolution of the particles and treat space and time
on the same footing, and others impose total energy and momentum conservation.
The model that we derive also includes the Page-Wootters mechanism for non-
interacting clocks in Refs. [34, 35, 36] as a particular case, when the
external degrees of freedom are neglected.
This model is suitable to be extended to more general situations. Here, we
have only considered a description in $1+1$ dimensions, and a more general
model in $3+1$ dimensions will probably have to face technical difficulties
such as those solved in Ref. [27]. For instance, it would be interesting to
explore the possibility of imposing as a constraint the total angular momentum
tensor in spacetime, i.e., $M_{\mu\nu}$, with $\mu,\nu=0,\cdots,3$, rather
than just its spatial components, as done in Ref. [27].
Another possible generalisation involves the description of (gravitationally)
interacting systems. This will likely require a different way of handling the
constraints, which would no longer commute (and would thus be second-class
constraints). This generalisation, if successful, would allow us to fully
incorporate the interacting clocks model, i.e., the description of time
reference frames [36] in the spacetime description of quantum reference
frames.
In addition, it would be interesting to generalise the model to arbitrary
spacetimes and superpositions thereof and build the transformation to a
Quantum Locally Inertial Frame as in Ref. [37], in the general case.
Finally, we notice that the internal energies enter the model presented here
in the total energy balance. This means that they are considered as part of
the zero component of the total momentum. In other works, for instance in
Refs. [53, 56, 65, 43], the internal energies contribute instead to the total
mass, via the relativistic mass-energy equivalence. These two ways of
considering the role of the internal energies, as the zero-component of a
vector and as a relativistically invariant quantity are, in principle,
different. However, the results presented here are in agreement, to our order
of approximation, to those one obtains by considering the internal energy as
part of the total mass via the mass-energy equivalence. It would be
interesting to investigate the connection between these two methods further,
in order to gain a deeper insight on the role of the internal degrees of
freedom when both quantum and gravitational effects are relevant.
At the interface of quantum theory and gravity, the notion of a classical
spacetime is no longer adequate to describe physical phenomena. It is then
crucial to generalise the formulation of the laws of physics to situations in
which spacetime is non-classical. One of the main goals of the research on
quantum reference frames is to generalise the laws of physics to when
reference frames are associated to quantum systems. Here, we have taken a step
further and we have associated a reference frame in spacetime to a quantum
system. Generalisations of this approach to more general scenarios could pave
the way for formulating physics on a quantum spacetime.
I would like to thank Lin-Qing Chen, Thomas D. Galley, and Lee Smolin for
useful discussions at the early stages of this work. I am also grateful to
Achim Kempf, Lorenzo Maccone, and Alexander R. H. Smith for helpful comments
on the draft. Research at Perimeter Institute is supported in part by the
Government of Canada through the Department of Innovation, Science and
Economic Development and by the Province of Ontario through the Ministry of
Colleges and Universities.
## Appendix A Mass dispersion relation in a weak gravitational field
In the Newtonian limit, the metric field is described as in Eq. (2), i.e.,
$\begin{split}&g_{00}=1+\frac{2\Phi(\mathbf{x}-\mathbf{x}_{M})}{c^{2}};\\\
&g_{01}=g_{10}=0;\\\ &g_{11}=-1,\end{split}$ (24)
where $\Phi(\mathbf{x})$ is the Newtonian potential due to the mass $m_{M}$ of
the system $M$ and $|\Phi(\mathbf{x})|/c^{2}\ll 1$ in the spacetime region
considered. Hence, the general-relativistic dispersion relation
$\tilde{C}_{I}=g^{\mu\nu}p_{\mu}p_{\nu}-m_{I}^{2}c^{2}$ of a classical
particle $I$ with mass $m_{I}$ takes the form
$\tilde{C}_{I}=g^{00}(\mathbf{x}_{I}-\mathbf{x}_{M})p_{0}^{I2}-\mathbf{p}_{I}^{2}-m_{I}^{2}c^{2}.$
(25)
We impose the positive energy condition by enforcing that $p^{I}_{0}\geq
0$333Notice that this condition can only be enforced in the weak-field limit
of the gravitational field, i.e., when $|\Phi(\mathbf{x})|/c^{2}\ll 1$.. By
noting that $\tilde{C}_{I}$ can be decomposed as
$\tilde{C}_{I}=\left[\sqrt{g^{00}(\mathbf{x}_{I}-\mathbf{x}_{M})}p^{I}_{0}-\sqrt{\mathbf{p}_{I}^{2}+m_{I}^{2}c^{2}}\right]\left[\sqrt{g^{00}(\mathbf{x}_{I}-\mathbf{x}_{M})}p^{I}_{0}+\sqrt{\mathbf{p}_{I}^{2}+m_{I}^{2}c^{2}}\right],$
(26)
we see that, if the energies are to be positive, it is enough to consider the
new dispersion relation
$C_{I}=\sqrt{g^{00}(\mathbf{x}_{I}-\mathbf{x}_{M})}p^{I}_{0}-\sqrt{\mathbf{p}_{I}^{2}+m_{I}^{2}c^{2}}.$
(27)
Clearly, this relation only holds in the low-energy regime of the particle,
hence it holds perturbatively in $\frac{\mathbf{p}_{I}^{2}}{m_{I}^{2}c^{2}}$.
It is now straightforward to quantise Eq. (27) and obtain the constraint
$\hat{C}_{I}=\sqrt{g^{00}(\hat{\mathbf{x}}_{I}-\hat{\mathbf{x}}_{M})}\hat{p}^{I}_{0}-\sqrt{\hat{\mathbf{p}}_{I}^{2}+m_{I}^{2}c^{2}}.$
(28)
If we now define
$\hat{\omega}^{I}_{p}=\sqrt{\hat{\mathbf{p}}_{I}^{2}+m_{I}^{2}c^{2}}$, we find
the constraint $\hat{C}_{I}$ of Eq. (3). Notice that the weak-field
approximation $|\Phi(\mathbf{x})|/c^{2}\ll 1$ played a crucial role here to
avoid ordering ambiguities in the quantisation procedure, as a more general
form of the metric field would also have position-dependent spatial
components, which, when quantised, do not commute with their conjugate
momentum.
It is easy to show that the expression
$\bra{x_{I}^{0}}\hat{C}_{I}\ket{\psi_{I}}_{ph}=0$, where $\ket{\psi_{I}}_{ph}$
is the quantum state of particle $I$ solving the constraint $\hat{C}_{I}$, is
equivalent to
$\mathrm{i}\hbar\frac{d}{dx_{I}^{0}}\ket{\psi_{I}(x^{0}_{I})}=\sqrt{g_{00}(\hat{\mathbf{x}}_{I}-\hat{\mathbf{x}}_{M})}\hat{\omega}_{p}^{I}\ket{\psi_{I}(x^{0}_{I})},$
(29)
where $\ket{\psi_{I}(x^{0}_{I})}=\braket{x_{I}^{0}}{\psi_{I}}_{ph}$. By
defining $x^{0}_{I}=ct_{I}$ and expanding perturbatively, we find
$\mathrm{i}\hbar\frac{d}{dt_{I}}\ket{\psi_{I}(t_{I})}=\left[m_{I}c^{2}+\frac{\hat{\mathbf{p}}_{I}^{2}}{2m_{I}}-\frac{\hat{\mathbf{p}}_{I}^{4}}{8m_{I}^{3}c^{2}}+m_{I}\Phi(\hat{\mathbf{x}}_{I}-\hat{\mathbf{x}}_{M})\right]\ket{\psi_{I}(t_{I})},$
(30)
where we do not consider higher-order terms. The equation above corresponds to
Schrödinger equation of a particle in a Newtonian field, with first-order
special-relativistic corrections.
## Appendix B Comparison between relational approaches
In the literature, there are several, and in principle different, ways of
eliminating an external, absolute structure. For instance, the Page-Wootters
mechanism [34, 35] considers a set of clocks which, at the global level,
satisfy a Hamiltonian constraint as in Eq. (1). Upon conditioning on the state
of one of the clocks, one recovers the dynamics of the other clocks. This is
usually interpreted as the internal perspective of the clock on which one
conditions. The Page-Wootters mechanism was applied to interacting clocks in
Refs. [40, 36]. In particular, in the case of gravitationally interacting
clocks studied in Ref. [36] it was shown that the interaction term between the
internal degrees of freedom of the clocks leads to a relative temporal
localisation of events, due to the fact that different internal energy states
contribute differently to the total mass of the system. In addition, the
limits to the measurability of time studied in Ref. [65] could be recovered.
In a different approach to relational dynamics, it was shown in Refs. [26, 27]
that the formalism for QRFs introduced in Ref. [25] can be derived by imposing
a symmetry principle on a set of $N$ particles, and specifically by imposing
invariance under global translations and global rotations. This symmetry
principle is reminiscent of the construction of Shape Dynamics [66], which
also inspired the construction of a classical model in which a gravitational
arrow of time emerges [67]. The above-mentioned techniques have been derived
in different frameworks and are, in principle, different. However, a
connection between different approaches to relational dynamics can be found,
as shown in Refs. [41, 42] for the case of the Page-Wootters mechanism,
relational Dirac observables, and quantum deparametrisation.
## Appendix C Worldline operator of a quantum relativistic particle in a weak
gravitational field
Let us consider a classical relativistic particle of mass $m$ in a weak
gravitational field. The classical equivalent of the constraint considered in
the main text (up to factors of order $O(c^{-2})$, is
$C_{I}=\sqrt{g^{00}(\mathbf{x}_{I})}p_{0}^{I}-m_{I}c\sqrt{1+\frac{\mathbf{p}_{I}^{2}}{m^{2}_{I}c^{2}}},$
(31)
where $\mathbf{p}_{I}^{2}$ is the spatial norm of the momentum. Via the
Hamilton’s equations of motion we have
$\dot{x}_{I}^{\mu}=\frac{dx_{I}^{\mu}}{d\tau}=\frac{\partial C_{I}}{\partial
p^{I}_{\mu}}$, which yields
$\begin{split}&\dot{x}_{I}^{0}=\sqrt{g^{00}(\mathbf{x}_{I})},\\\
&\dot{\mathbf{x}}_{I}=\frac{\mathbf{p}_{I}}{m_{I}c}\left(1+\frac{\mathbf{p}_{I}^{2}}{m_{I}^{2}c^{2}}\right)^{-1/2}.\end{split}$
(32)
Hence, the line element is, to order $O(c^{-2})$,
$\begin{split}ds&=\sqrt{g_{00}(\mathbf{x}_{I})-\frac{1}{c^{2}}\left(\frac{d\mathbf{x}_{I}}{d\tau}\right)^{2}\left(\frac{d\tau}{dx^{0}_{I}}\right)^{2}}dx^{0}_{I}=\\\
&=\sqrt{g_{00}(\mathbf{x}_{I})}\sqrt{1+\frac{\mathbf{p_{I}}^{2}}{m^{2}_{I}c^{2}}}dx_{I}^{0}.\end{split}$
(33)
The quantisation of this expression is straightforward, given that in the
regime we consider we have
$\left[g_{00}(\hat{\mathbf{x}}),\sqrt{1+\frac{\hat{\mathbf{p}}_{I}^{2}}{m_{I}^{2}c^{2}}}\right]=0$.
We then write the “worldline operator” as
$\Delta(\hat{\mathbf{x}}_{I},\hat{\mathbf{p}}_{I})=\sqrt{g_{00}(\mathbf{\hat{x}}_{I})}\sqrt{1+\frac{\hat{\mathbf{p}}_{I}^{2}}{m_{I}^{2}c^{2}}}.$
(34)
## Appendix D Action of the $\mathcal{\hat{T}}_{1}$ operator on the position
and momentum operators
In this Appendix, we calculate the action of the operator
$\mathcal{\hat{T}}_{1}$ on the phase space operators of the quantum particle
in spacetime. We have
$\begin{split}&\mathcal{\hat{T}}_{1}\hat{x}_{1}^{0}\mathcal{\hat{T}}_{1}^{\dagger}=\sqrt{g^{00}(\hat{\mathbf{q}}_{M})}\hat{q}_{1}^{0},\hskip
10000.0pt\\\
&\mathcal{\hat{T}}_{1}\hat{\mathbf{x}}_{1}\mathcal{\hat{T}}_{1}^{\dagger}=\hat{\mathbf{q}}_{1},\\\
&\mathcal{\hat{T}}_{1}\hat{x}_{i}^{0}\mathcal{\hat{T}}_{1}^{\dagger}=\sqrt{g^{00}(\hat{\mathbf{q}}_{M})}\left(\hat{q}_{i}^{0}+\hat{q}_{1}^{0}\right),\\\
&\mathcal{\hat{T}}_{1}\hat{\mathbf{x}}_{i}\mathcal{\hat{T}}_{1}^{\dagger}=\hat{\mathbf{q}}_{i}+\hat{\mathbf{q}}_{1},\\\
&\mathcal{\hat{T}}_{1}\hat{x}_{M}^{0}\mathcal{\hat{T}}_{1}^{\dagger}=\hat{q}_{M}^{0}+\sqrt{g^{00}(\hat{\mathbf{q}}_{M})}\hat{q}_{1}^{0},\\\
&\mathcal{\hat{T}}_{1}\hat{\mathbf{x}}_{M}\mathcal{\hat{T}}_{1}^{\dagger}=\hat{\mathbf{q}}_{M}+\hat{\mathbf{q}}_{1},\end{split}$
$\begin{split}&\mathcal{\hat{T}}_{1}\hat{p}^{1}_{0}\mathcal{\hat{T}}_{1}^{\dagger}=\sqrt{g_{00}(\hat{\mathbf{q}}_{M})}\Big{(}\hat{k}^{1}_{0}-\sum_{i\neq
1}\hat{k}_{0}^{i}\Big{)}-\sum_{i\neq
1}\Delta(\hat{\mathbf{q}}_{i}-\hat{\mathbf{q}}_{M},\hat{\mathbf{k}}_{i})\frac{\hat{H}_{i}}{c}-\hat{k}_{0}^{M}-\Delta\Big{(}\hat{\mathbf{q}}_{M},\hat{\mathbf{k}}_{1}-\sum_{i\neq
1}\hat{\mathbf{k}}_{i}-\hat{\mathbf{k}}_{M}\Big{)}\frac{\hat{H}_{1}}{c},\\\
&\mathcal{\hat{T}}_{1}\hat{\mathbf{p}}_{1}\mathcal{\hat{T}}_{1}^{\dagger}=\hat{\mathbf{k}}_{1}-\sum_{i\neq
1}\hat{\mathbf{k}}_{i}-\hat{\mathbf{k}}_{M},\\\
&\mathcal{\hat{T}}_{1}\hat{p}^{i}_{0}\mathcal{\hat{T}}_{1}^{\dagger}=\sqrt{g_{00}(\hat{\mathbf{q}}_{M})}\hat{k}^{i}_{0},\\\
&\mathcal{\hat{T}}_{1}\hat{\mathbf{p}}_{i}\mathcal{\hat{T}}_{1}^{\dagger}=\hat{\mathbf{k}}_{i},\\\
&\mathcal{\hat{T}}_{1}\hat{p}^{M}_{0}\mathcal{\hat{T}}_{1}^{\dagger}=\hat{k}^{M}_{0},\\\
&\mathcal{\hat{T}}_{1}\hat{\mathbf{p}}_{M}\mathcal{\hat{T}}_{1}^{\dagger}=\hat{\mathbf{k}}_{M},\end{split}$
(35)
where $i=2,\cdots,N$.
We can also define a new transformation $\mathcal{\hat{T}}_{2}$ to the
relational variables of particle $2$ by swapping the labels $1$ and $2$ in the
transformation $\mathcal{\hat{T}}_{1}$. We define the relational phase-space
operators to particle $2$ as $\hat{r}_{j}$ (position operator) and
$\hat{u}_{j}$ (momentum operator), for $j=1,3,\cdots,N,M$, and we obtain
$\begin{split}&\mathcal{\hat{T}}_{2}\hat{x}_{2}^{0}\mathcal{\hat{T}}_{2}^{\dagger}=\sqrt{g^{00}(\hat{\mathbf{r}}_{M})}\hat{r}_{2}^{0},\hskip
10000.0pt\\\
&\mathcal{\hat{T}}_{2}\hat{\mathbf{x}}_{2}\mathcal{\hat{T}}_{2}^{\dagger}=\hat{\mathbf{r}}_{2},\\\
&\mathcal{\hat{T}}_{2}\hat{x}_{j}^{0}\mathcal{\hat{T}}_{2}^{\dagger}=\sqrt{g^{00}(\hat{\mathbf{r}}_{M})}\left(\hat{r}_{j}^{0}+\hat{r}_{2}^{0}\right),\\\
&\mathcal{\hat{T}}_{2}\hat{\mathbf{x}}_{j}\mathcal{\hat{T}}_{2}^{\dagger}=\hat{\mathbf{r}}_{j}+\hat{\mathbf{r}}_{2},\\\
&\mathcal{\hat{T}}_{2}\hat{x}_{M}^{0}\mathcal{\hat{T}}_{2}^{\dagger}=\hat{r}_{M}^{0}+\sqrt{g^{00}(\hat{\mathbf{r}}_{M})}\hat{r}_{2}^{0},\\\
&\mathcal{\hat{T}}_{2}\hat{\mathbf{x}}_{M}\mathcal{\hat{T}}_{2}^{\dagger}=\hat{\mathbf{r}}_{M}+\hat{\mathbf{r}}_{2},\\\
\end{split}$
$\begin{split}&\mathcal{\hat{T}}_{2}\hat{p}^{2}_{0}\mathcal{\hat{T}}_{2}^{\dagger}=\sqrt{g_{00}(\hat{\mathbf{r}}_{M})}\Big{(}\hat{u}^{2}_{0}-\sum_{j\neq
2}\hat{u}^{j}_{0}\Big{)}-\sum_{j\neq
2}\Delta(\hat{\mathbf{r}}_{j}-\hat{\mathbf{r}}_{M},\hat{\mathbf{u}}_{j})\frac{\hat{H}_{j}}{c}-\hat{u}_{0}^{M}-\Delta\Big{(}\hat{\mathbf{r}}_{M},\hat{\mathbf{u}}_{2}-\sum_{j\neq
2}\hat{\mathbf{u}}_{j}-\hat{\mathbf{u}}_{M}\Big{)}\frac{\hat{H}_{1}}{c}\\\
&\mathcal{\hat{T}}_{2}\hat{\mathbf{p}}_{2}\mathcal{\hat{T}}_{2}^{\dagger}=\hat{\mathbf{u}}_{2}-\sum_{j\neq
2}\hat{\mathbf{u}}_{j}-\hat{\mathbf{u}}_{M}\\\
&\mathcal{\hat{T}}_{2}\hat{p}^{j}_{0}\mathcal{\hat{T}}_{2}^{\dagger}=\sqrt{g_{00}(\hat{\mathbf{r}}_{M})}\hat{u}^{j}_{0}\\\
&\mathcal{\hat{T}}_{2}\hat{\mathbf{p}}_{j}\mathcal{\hat{T}}_{2}^{\dagger}=\hat{\mathbf{u}}_{j}\\\
&\mathcal{\hat{T}}_{2}\hat{p}^{M}_{0}\mathcal{\hat{T}}_{2}^{\dagger}=\hat{u}^{M}_{0}\\\
&\mathcal{\hat{T}}_{2}\hat{\mathbf{p}}_{M}\mathcal{\hat{T}}_{2}^{\dagger}=\hat{\mathbf{u}}_{M},\end{split}$
(36)
where $j=1,3,\cdots,N$.
We can also transform from one set of relational phase-space operators to
another, e.g., from particle $1$ to particle $2$. In order to do this, we
define
$\mathcal{\hat{T}}_{12}=\mathcal{\hat{T}}_{2}\mathcal{\hat{T}}_{1}^{\dagger}$
and find
$\begin{split}&\mathcal{\hat{T}}_{12}\hat{q}_{1}^{0}\mathcal{\hat{T}}_{12}^{\dagger}=\sqrt{\frac{g_{00}(\hat{\mathbf{r}}_{1}-\hat{\mathbf{r}}_{M})}{g_{00}(\hat{\mathbf{r}}_{M})}}\left(\hat{r}_{1}^{0}+\hat{r}_{2}^{0}\right),\hskip
10000.0pt\\\
&\mathcal{\hat{T}}_{12}\hat{\mathbf{q}}_{1}\mathcal{\hat{T}}_{12}^{\dagger}=\hat{\mathbf{r}}_{1}+\hat{\mathbf{r}}_{2},\\\
&\mathcal{\hat{T}}_{12}\hat{q}_{2}^{0}\mathcal{\hat{T}}_{12}^{\dagger}=-\sqrt{\frac{g_{00}(\hat{\mathbf{r}}_{1}-\hat{\mathbf{r}}_{M})}{g_{00}(\hat{\mathbf{r}}_{M})}}\hat{r}_{1}^{0},\\\
&\mathcal{\hat{T}}_{12}\hat{\mathbf{q}}_{2}\mathcal{\hat{T}}_{12}^{\dagger}=-\hat{\mathbf{r}}_{1},\\\
&\mathcal{\hat{T}}_{12}\hat{q}_{\ell}^{0}\mathcal{\hat{T}}_{12}^{\dagger}=\sqrt{\frac{g_{00}(\hat{\mathbf{r}}_{1}-\hat{\mathbf{r}}_{M})}{g_{00}(\hat{\mathbf{r}}_{M})}}(\hat{r}_{\ell}^{0}-\hat{r}_{1}^{0}),\\\
&\mathcal{\hat{T}}_{12}\hat{\mathbf{q}}_{\ell}\mathcal{\hat{T}}_{12}^{\dagger}=\hat{\mathbf{r}}_{\ell}-\hat{\mathbf{r}}_{1},\\\
&\mathcal{\hat{T}}_{12}\hat{q}_{M}^{0}\mathcal{\hat{T}}_{12}^{\dagger}=\hat{r}_{M}^{0}-\sqrt{g^{00}(\hat{\mathbf{r}}_{M})}\hat{r}_{1}^{0},\\\
&\mathcal{\hat{T}}_{12}\hat{\mathbf{q}}_{M}\mathcal{\hat{T}}_{12}^{\dagger}=\hat{\mathbf{r}}_{M}-\hat{\mathbf{r}}_{1},\end{split}$
$\begin{split}&\mathcal{\hat{T}}_{12}\hat{k}^{1}_{0}\mathcal{\hat{T}}_{12}^{\dagger}=\sqrt{g^{00}(\hat{\mathbf{r}}_{1}-\hat{\mathbf{r}}_{M})}\sqrt{g_{00}(\hat{\mathbf{r}}_{M})}\hat{u}^{2}_{0},\hskip
10000.0pt\\\
&\mathcal{\hat{T}}_{12}\hat{\mathbf{k}}_{1}\mathcal{\hat{T}}_{12}^{\dagger}=\hat{\mathbf{u}}_{2},\\\
&\mathcal{\hat{T}}_{12}\hat{k}^{2}_{0}\mathcal{\hat{T}}_{12}^{\dagger}=\sqrt{\frac{g^{00}(\hat{\mathbf{r}}_{1}-\hat{\mathbf{r}}_{M})}{g^{00}(\hat{\mathbf{r}}_{M})}}\Big{(}\hat{u}^{2}_{0}-\sum_{j\neq
2}\hat{u}_{0}^{j}\Big{)}-\sqrt{g^{00}(\hat{\mathbf{r}}_{1}-\hat{\mathbf{r}}_{M})}\left[\sum_{j\neq
2}\Delta_{j}\frac{\hat{H}_{j}}{c}+\hat{u}_{0}^{M}+\Delta_{\Sigma
u,2}\frac{\hat{H}_{2}}{c}\right],\\\
&\mathcal{\hat{T}}_{12}\hat{\mathbf{k}}_{2}\mathcal{\hat{T}}_{12}^{\dagger}=\hat{\mathbf{u}}_{2}-\sum_{j\neq
2}\hat{\mathbf{u}}_{j}-\hat{\mathbf{u}}_{M},\\\
&\mathcal{\hat{T}}_{12}\hat{k}^{\ell}_{0}\mathcal{\hat{T}}_{12}^{\dagger}=\sqrt{\frac{g^{00}(\hat{\mathbf{r}}_{1}-\hat{\mathbf{r}}_{M})}{g^{00}(\hat{\mathbf{r}}_{M})}}\hat{u}^{\ell}_{0},\\\
&\mathcal{\hat{T}}_{12}\hat{\mathbf{k}}_{\ell}\mathcal{\hat{T}}_{12}^{\dagger}=\hat{\mathbf{u}}_{\ell},\\\
&\mathcal{\hat{T}}_{12}\hat{k}^{M}_{0}\mathcal{\hat{T}}_{12}^{\dagger}=\hat{u}^{M}_{0},\\\
&\mathcal{\hat{T}}_{12}\hat{\mathbf{k}}_{M}\mathcal{\hat{T}}_{12}^{\dagger}=\hat{\mathbf{u}}_{M},\end{split}$
(37)
where $\ell=3,\cdots,N$ and
$\Delta_{j}=\Delta(\hat{\mathbf{r}}_{j}-\hat{\mathbf{r}}_{M},\hat{\mathbf{u}}_{j})$
and $\Delta_{\Sigma
u,2}=\Delta\Big{(}\hat{\mathbf{r}}_{M},\hat{\mathbf{u}}_{2}-\sum_{j\neq
2}\hat{\mathbf{u}}_{j}-\hat{\mathbf{u}}_{M}\Big{)}$. Notice that the relevant
relational operators are those of the particles $2,\cdots,N,M$ from the
perspective of particle $1$, which are mapped into quantities which only
depend on the operators of particle $2$ via $\hat{u}_{2}$, which is the
transformed constraint, since
$\hat{\mathcal{T}}_{2}\hat{f}^{0}\hat{\mathcal{T}}_{2}^{\dagger}=\sqrt{g_{00}(\hat{\mathbf{r}}_{M})}\hat{u}^{0}_{2}$
and
$\hat{\mathcal{T}}_{2}\hat{f}^{1}\hat{\mathcal{T}}_{2}^{\dagger}=\hat{\mathbf{u}}_{2}$.
## Appendix E Explicit calculation of the “history state” in the frame of
particle 1
We recover in this Appendix the dynamical evolution of the relational degrees
of freedom from the state of Eq. (6)
$\ket{\Psi}_{ph}\propto\int
d^{N}\mathcal{N}d^{2}ze^{\frac{\mathrm{i}}{\hbar}\mathcal{N}_{i}\hat{C}_{i}}e^{\frac{\mathrm{i}}{\hbar}z_{\mu}\hat{f}^{\mu}}\ket{\phi},$
(38)
where $\ket{\phi}$ is represented as
$\ket{\phi}=\int\Pi_{I}\left[d\mu(x_{I})dE_{I}\right]d^{2}x_{M}\phi(x_{1},\cdots,x_{N},x_{M},E_{1},\cdots,E_{N})\ket{x_{1},\cdots,x_{N},x_{M}}\ket{E_{1},\cdots,E_{N}},$
(39)
and $d\mu(x_{i})=\sqrt{g_{00}(\mathbf{x}_{I}-\mathbf{x}_{M})}d^{2}x_{I}$ is
the covariant integration measure. The set of constraints that we enforce in
this model is
$\begin{split}&\hat{C}_{I}=\sqrt{g^{00}(\hat{\mathbf{x}}_{I}-\hat{\mathbf{x}}_{M})}\hat{p}^{I}_{0}-\hat{\omega}_{p}^{I},\qquad\qquad
I=1,\cdots,N\\\
&\hat{f}^{0}=\sum_{I=1}^{N}\left[\hat{p}^{I}_{0}+\Delta(\hat{\mathbf{x}}_{I}-\hat{\mathbf{x}}_{M},\hat{\mathbf{p}}_{I})\frac{\hat{H}_{I}}{c}\right]+\hat{p}_{0}^{M};\\\
&\hat{f}^{1}=\sum_{I=1}^{N}\hat{\mathbf{p}}_{I}+\hat{\mathbf{p}}_{M},\end{split}$
(40)
where
$\Delta(\hat{\mathbf{x}}_{I}-\hat{\mathbf{x}}_{M},\hat{\mathbf{p}}_{I})=\sqrt{g_{00}(\hat{\mathbf{x}}_{I}-\hat{\mathbf{x}}_{M})}\left(1+\frac{\hat{\mathbf{p}}^{2}_{I}}{m_{I}^{2}c^{2}}\right)^{-1/2}$.
In order to find the dynamical evolution, we define an operator
$\hat{\mathcal{T}}_{1}$ which maps the observables to the relational
observables from the perspective of particle $1$. We define the operator
$\hat{\mathcal{T}}_{1}$ as
$\hat{\mathcal{T}}_{1}=e^{\frac{i}{\hbar}\frac{\log\sqrt{g_{00}(\hat{\mathbf{x}}_{M})}}{2}\sum_{I=1}^{N}(\hat{x}^{0}_{I}\hat{p}_{0}^{I}+\hat{p}_{0}^{I}\hat{x}^{0}_{I})}e^{\frac{\mathrm{i}}{\hbar}\hat{\mathbf{x}}_{1}\left(\hat{f}^{1}-\hat{\mathbf{p}}^{1}\right)}e^{\frac{\mathrm{i}}{\hbar}\hat{x}_{1}^{0}\left(\hat{f}^{0}-\hat{p}^{1}_{0}\right)},$
(41)
where
$\hat{f}^{1}-\hat{\mathbf{p}}^{1}=\sum_{i}\hat{\mathbf{p}}_{i}+\hat{\mathbf{p}}_{M}$,
$\hat{f}^{0}-\hat{p}^{1}_{0}=\sum_{i}\left[\hat{p}_{0}^{i}+\Delta(\hat{\mathbf{x}}_{i}-\hat{\mathbf{x}}_{M},\hat{\mathbf{p}}_{i})\frac{\hat{H}_{i}}{c}\right]+\Delta(\hat{\mathbf{x}}_{1}-\hat{\mathbf{x}}_{M},\hat{\mathbf{p}}_{1})\frac{\hat{H}_{1}}{c}+\hat{p}_{0}^{M}$
and the lowercase latin letters label all particles except for the particle
serving as the QRF, i.e., $i=2,\cdots,N$. The action of the operator
$\hat{\mathcal{T}}_{1}$ on the constraints is
$\begin{split}\hat{\mathcal{T}}_{1}\hat{f}^{0}\hat{\mathcal{T}}_{1}^{\dagger}=&\sqrt{g_{00}(\hat{\mathbf{q}}_{M})}\hat{k}^{1}_{0};\qquad\hat{\mathcal{T}}_{1}\hat{f}^{1}\hat{\mathcal{T}}_{1}^{\dagger}=\hat{\mathbf{k}}_{1};\qquad\hat{\mathcal{T}}_{1}\hat{C}_{i}\hat{\mathcal{T}}_{1}^{\dagger}=\hat{C}^{\prime}_{i}\,\,\,i\neq
1;\\\
\hat{\mathcal{T}}_{1}\hat{C}_{1}\hat{\mathcal{T}}_{1}^{\dagger}=&\hat{k}_{0}^{1}-\left\\{\sum_{i}\left[\hat{k}_{0}^{i}+\Delta^{\prime}(\hat{\mathbf{q}}_{i},\hat{\mathbf{q}}_{M},\hat{\mathbf{k}}_{i})\frac{\hat{H}_{i}}{c}\right]+\sqrt{g^{00}(\hat{\mathbf{q}}_{M})}\hat{k}_{0}^{M}\right\\}+\\\
&-m_{1}c\sqrt{1+\frac{\left(\hat{\mathbf{k}}_{1}-\sum_{i}\hat{\mathbf{k}}_{i}-\hat{\mathbf{k}}_{M}\right)^{2}}{m_{1}^{2}c^{2}}}-\left(1+\frac{\left(\hat{\mathbf{k}}_{1}-\sum_{i}\hat{\mathbf{k}}_{i}-\hat{\mathbf{k}}_{M}\right)^{2}}{m_{1}^{2}c^{2}}\right)^{-1/2}\frac{\hat{H}_{1}}{c},\end{split}$
(42)
where $\hat{C}^{\prime}_{i}$ has been defined in the main text as
$\hat{C}^{\prime}_{i}=\sqrt{g^{\prime
00}(\hat{\mathbf{q}}_{i},\hat{\mathbf{q}}_{M})}\hat{k}_{0}^{i}-\hat{\omega}_{k}^{i}$,
$\Delta^{\prime}(\hat{\mathbf{q}}_{i},\hat{\mathbf{q}}_{M},\hat{\mathbf{k}}_{i})=\sqrt{g^{\prime}_{00}(\hat{\mathbf{q}}_{i},\hat{\mathbf{q}}_{M})}\left(1+\frac{\hat{\mathbf{k}}_{i}^{2}}{m_{i}^{2}c^{2}}\right)^{-1/2}$,
and the transformed metric field is
$g^{\prime}_{00}(\hat{\mathbf{q}}_{i},\hat{\mathbf{q}}_{M})=\frac{g_{00}(\hat{\mathbf{q}}_{i}-\hat{\mathbf{q}}_{M})}{g_{00}(\hat{\mathbf{q}}_{M})}.$
(43)
The relational state from the perspective of system $1$ is then obtained as
$\ket{\psi}^{(1)}=\bra{q_{1}=0}\hat{\mathcal{T}}_{1}\ket{\Psi}_{ph}$
By defining the internal time operator $\hat{T}_{I}$ such that
$\left[\hat{T}_{I},\hat{H}_{J}\right]=i\hbar\delta_{IJ}$ and the state of the
internal clock of each particle $\ket{\tau_{I}}$ as the one satisfying the
relation $\hat{T}_{I}\ket{\tau_{I}}=\tau_{I}\ket{\tau_{I}}$, we find
$\ket{\psi}^{(1)}=2\pi\hbar\left(1-\frac{\Phi(\mathbf{\hat{q}}_{M})}{c^{2}}\right)\int
d^{N}\mathcal{N}e^{\frac{\mathrm{i}}{\hbar}\mathcal{N}_{i}\hat{C}^{\prime}_{i}}e^{-\frac{\mathrm{i}}{\hbar}\frac{\hat{\gamma}_{\Sigma
k,1}^{-1}}{c}\left(\hat{K}^{(1)}+\hat{H}_{1}\right)\mathcal{N}_{1}}\bra{q_{1}=0}\hat{\Pi}_{0}\mathcal{\hat{T}}_{1}\ket{\phi},$
(44)
where $\hat{K}^{(1)}$ is the operator encoding the relational dynamics of
particles $2,\cdots,N$ from the point of view of particle $1$, i.e.,
$\hat{K}^{(1)}=\hat{\gamma}_{\Sigma
k,1}\left[\sum_{i}c\hat{k}_{0}^{i}+c\sqrt{g^{00}(\hat{\mathbf{q}}_{M})}\hat{k}_{0}^{M}\right]+\sum_{i}\sqrt{g^{\prime}_{00}(\hat{\mathbf{q}}_{i},\hat{\mathbf{q}}_{M})}\frac{\hat{\gamma}_{i}^{-1}}{\hat{\gamma}_{\Sigma
k,1}^{-1}}\hat{H}_{i}+m_{1}c^{2}\hat{\gamma}_{\Sigma k,1}^{2},$ (45)
$\hat{\gamma}_{i}=\sqrt{1+\frac{\hat{\mathbf{k}}_{i}^{2}}{m_{i}^{2}c^{2}}}$,
$\hat{\gamma}_{\Sigma
k,1}=\sqrt{1+\frac{\left(\sum_{i}\hat{\mathbf{k}}_{i}+\hat{\mathbf{k}}_{M}\right)^{2}}{m_{1}^{2}c^{2}}}$,
and $\hat{\Pi}_{0}=\ket{k_{1}=0}\bra{k_{1}=0}$. We now define
$\ket{\phi_{/1}}=2\pi\hbar\left(1-\frac{\Phi(\mathbf{\hat{q}}_{M})}{c^{2}}\right)\bra{q_{1}=0}\hat{\Pi}_{0}\mathcal{\hat{T}}_{1}\ket{\phi}$
(46)
and act with the operator $\hat{H}_{1}$ on the internal state of clock $1$. We
then rewrite
$\ket{\psi}^{(1)}=\int
d\tau_{1}d^{N-1}\mathcal{N}e^{\frac{\mathrm{i}}{\hbar}\mathcal{N}_{i}\hat{C}^{\prime}_{i}}e^{-\frac{\mathrm{i}}{\hbar}\hat{K}^{(1)}\tau_{1}}\ket{\phi^{(1)}_{0}}\ket{\tau_{1}},$
(47)
where
$\begin{split}\ket{\phi^{(1)}_{0}}=&\int
dt_{1}\Pi_{i}[d^{2}k_{i}dE_{i}]d^{2}k_{M}e^{\frac{\mathrm{i}}{\hbar}\hat{K}^{(1)}t_{1}}\phi_{/1}(k_{2},\cdots,k_{N},k_{M},t_{1},E_{2},\cdots,E_{N})\times\\\
&\times\ket{k_{2},\cdots,k_{N},k_{M}}\ket{E_{2},\cdots,E_{N}}.\end{split}$
(48)
Following a similar procedure to the one outlined above, it is possible solve
all the constraints and to cast the previous expression as a “history state”
in the sense of Refs. [36, 37] evolving unitarily as
$\ket{\psi}^{(1)}=\int
d\tau_{1}e^{-\frac{\mathrm{i}}{\hbar}\hat{H}^{(1)}\tau_{1}}\ket{\psi^{(1)}_{0}}\ket{\tau_{1}},$
(49)
where the Hamiltonian $\hat{H}^{(1)}$ is
$\hat{H}^{(1)}=\hat{\gamma}_{\Sigma
k,1}\sum_{i}\sqrt{g^{\prime}_{00}(\hat{\mathbf{q}}_{i},\hat{\mathbf{q}}_{M})}\left\\{c\hat{\omega}_{k}^{i}+\hat{\gamma}_{i}^{-1}\hat{H}_{i}\right\\}+c\hat{\gamma}_{\Sigma
k,1}\sqrt{g^{00}(\hat{\mathbf{q}}_{M})}\hat{k}_{0}^{M}+m_{1}c^{2}\hat{\gamma}_{\Sigma
k,1}^{2},$ (50)
and
$\begin{split}\ket{\psi^{(1)}_{0}}=&\int\Pi_{i}\left(d\mu(q_{i})dq^{\prime
0}_{i}dE_{i}\sqrt{g^{\prime}_{00}(\mathbf{q}_{i},\mathbf{q}_{M})}\right)d^{2}q_{M}e^{\frac{\mathrm{i}}{\hbar}\sum_{i}(q^{\prime
0}_{i}-q^{0}_{i})\sqrt{g^{\prime}_{00}(\mathbf{q}_{i},\mathbf{q}_{M})}\hat{\omega}_{k}^{i}}\times\\\
&\times\phi^{(1)}_{0}\left(q_{2},\cdots,q_{N},q_{M},E_{2},\cdots,E_{N}\right)\ket{q^{\prime
0}_{2},\mathbf{q}_{2},\cdots,q^{\prime
0}_{N},\mathbf{q}_{N}}\ket{q_{M}}\ket{E_{2}\cdots E_{N}},\end{split}$ (51)
which reproduces the result of the main text.
The generalisation of this technique to general gravitational fields requires
dealing with second-class constraints, and is thus beyond the scope of this
work.
## Appendix F Limiting cases of the general $N$-particle model
In the following, we are going to study some relevant limits of the model
introduced in the main text. In particular, we will focus on a set of quantum
free particles moving slowly compared to the speed of light (Galilean case),
on a set of quantum special-relativistic particles, and on a set of quantum
Galilean particles in an external weak gravitational field (Newtonian case).
We remind the reader that operators and vectors with no indices are, unless
differently specified, operators and two-vectors in spacetime, while the
spatial component of operators and vectors is boldface. Greek letters label
spacetime indices, capital Latin letters label all the particles, e.g.,
$I=1,\cdots,N$, and lowercase Latin letters label all the particles except the
one serving as the QRF, e.g., $i=2,\cdots,N$.
### F.1 Galilean case
When the particles move slowly compared to the speed of light and are free,
the contraints introduced in the main text reduce to
$\displaystyle\hat{C}_{I}=\hat{p}_{I}^{0}-\frac{\hat{\mathbf{p}}^{2}_{I}}{2m_{I}c},$
(52)
$\displaystyle\hat{f}^{0}=\sum_{I=1}^{N}\left(\hat{p}_{I}^{0}+\frac{\hat{H}_{I}}{c}\right),$
(53) $\displaystyle\hat{f}^{1}=\sum_{I=1}^{N}\hat{\mathbf{p}}_{I},$ (54)
where here we have neglected particle $M$ because it plays no role. In
addition, we eliminated the constant term $m_{I}c$ in the expansion of
$\hat{C}_{I}$, because it simply amounts to a rescaling of the zero component
of the momentum.
We write the state in the Physical Hilbert space as
$\ket{\Psi}_{ph}\propto\int
d^{N}\mathcal{N}d^{2}ze^{\frac{\mathrm{i}}{\hbar}\mathcal{N}_{i}\hat{C}_{i}}e^{\frac{\mathrm{i}}{\hbar}z_{\mu}\hat{f}^{\mu}}\ket{\phi},$
(55)
where $\ket{\phi}$ is expanded in momentum basis as
$\ket{\phi}=\int\Pi_{I}\left[d^{2}p_{I}dE_{I}\right]\phi(p_{1},\cdots,p_{N},E_{1},\cdots,E_{N})\ket{p_{1},\cdots,p_{N}}\ket{E_{1},\cdots,E_{N}}.$
(56)
In this case, the operator $\hat{\mathcal{T}}_{1}$ simplifies to
$\hat{\mathcal{T}}_{1}=e^{\frac{\mathrm{i}}{\hbar}\hat{\mathbf{x}}_{1}\sum_{i}\hat{\mathbf{p}}_{i}}e^{\frac{\mathrm{i}}{\hbar}\hat{x}_{1}^{0}\left[\sum_{i}\left(\hat{p}_{0}^{i}+\frac{\hat{H}_{i}}{c}\right)+\frac{\hat{H}_{1}}{c}\right]},$
(57)
and the constraints are mapped to
$\begin{split}&\hat{\mathcal{T}}_{1}\hat{f}^{0}\hat{\mathcal{T}}_{1}^{\dagger}=\hat{k}^{1}_{0};\qquad\hat{\mathcal{T}}_{1}\hat{f}^{1}\hat{\mathcal{T}}_{1}^{\dagger}=\hat{\mathbf{k}}_{1};\qquad\hat{\mathcal{T}}_{1}\hat{C}_{i}\hat{\mathcal{T}}_{1}^{\dagger}=\hat{C}_{i}\,\,\,i\neq
1;\\\
&\hat{\mathcal{T}}_{1}\hat{C}_{1}\hat{\mathcal{T}}_{1}^{\dagger}=\hat{k}_{0}^{1}-\sum_{i}\left(\hat{k}_{0}^{i}+\frac{\hat{H}_{i}}{c}\right)-\frac{\hat{H}_{1}}{c}-\frac{\left(\hat{\mathbf{k}}_{1}-\sum_{i}\hat{\mathbf{k}}_{i}\right)^{2}}{2m_{1}}.\end{split}$
(58)
Defining the relational state from the perspective of particle $1$ to be
$\ket{\psi^{(1)}}={}_{1}\bra{q_{1}=0}\mathcal{T}_{1}\ket{\Psi}_{ph}$, we find
$\ket{\psi}^{(1)}=\int
d\tau_{1}d^{N-1}\mathcal{N}\ket{\tau_{1}}e^{\frac{\mathrm{i}}{\hbar}\mathcal{N}_{i}\hat{C}_{i}}e^{-\frac{\mathrm{i}}{\hbar}\hat{K}^{(1)}\tau_{1}}\ket{\phi^{(1)}_{0}},$
(59)
where
$\hat{K}^{(1)}=\sum_{i}\left(\hat{k}_{0}^{i}+\frac{\hat{H}_{i}}{c}\right)+\frac{\left(\sum_{i}\hat{\mathbf{k}}_{i}\right)^{2}}{2m_{1}},$
(60)
and
$\begin{split}&\ket{\phi^{(1)}_{0}}=\frac{1}{\sqrt{2\pi\hbar}}\int
dt_{1}dE_{1}\Pi_{i}(d^{2}k_{i}dE_{i})e^{\frac{\mathrm{i}}{\hbar}(\hat{K}^{(1)}+E_{1})t_{1}}\phi\left(\bar{k}_{1},k_{2},\cdots,k_{N},E_{1},\cdots,E_{N}\right)\times\\\
&\hskip 56.9055pt\times\ket{k_{2}\cdots k_{N}}\ket{E_{2}\cdots
E_{N}}.\end{split}$ (61)
By solving the constraint we have rewritten the two-vector
$\bar{k}_{1}=\left(-\sum_{i}\left(k_{0}^{i}+\frac{E_{i}}{c}\right)-\frac{E_{1}}{c},-\sum_{i}\mathbf{k}_{i}\right)$.
Notice that we can easily recover the framework of standard quantum mechanics
by integrating over the $N_{i}$ variables. After this operation, the state
$\ket{\psi^{(1)}}$ can be rewritten as the “history state”
$\ket{\psi}^{(1)}=\int
d\tau_{1}e^{-\frac{\mathrm{i}}{\hbar}\hat{H}^{(1)}\tau_{1}}\ket{\psi^{(1)}_{0}}\ket{\tau_{1}},$
(62)
where
$\displaystyle\hat{H}^{(1)}=\sum_{i}\frac{\hat{\mathbf{k}}_{i}^{2}}{2m_{i}}+\frac{(\sum_{{i}}\hat{\mathbf{k}}_{i})^{2}}{2m_{1}}+\sum_{i}\hat{H}_{i},$
$\displaystyle\begin{split}&\ket{\psi^{(1)}_{0}}=\int\Pi_{i}(d\mathbf{k}_{i}dE_{i})\phi^{(1)}_{0}\left(\frac{\mathbf{k}_{2}^{2}}{2m_{2}},\mathbf{k}_{2},\cdots,\frac{\mathbf{k}_{N}^{2}}{2m_{N}},\mathbf{k}_{N},E_{1},\cdots,E_{N}\right)\times\\\
&\hskip
56.9055pt\times\ket{\frac{\mathbf{k}_{2}^{2}}{2m_{2}},\mathbf{k}_{2},\cdots,\frac{\mathbf{k}_{N}^{2}}{2m_{N}},\mathbf{k}_{N}}\ket{E_{2}\cdots
E_{N}}.\end{split}$
### F.2 Special-relativistic case
In this section, we consider a set of N special-relativistic particles moving
freely. The set of constraints is
$\displaystyle\hat{C}_{I}=\hat{p}_{I}^{0}-\hat{\omega}_{p}^{I},$ (63)
$\displaystyle\hat{f}^{0}=\sum_{I=1}^{N}\left(\hat{p}_{I}^{0}+\hat{\gamma}_{I}^{-1}\frac{\hat{H}_{I}}{c}\right),$
(64) $\displaystyle\hat{f}^{1}=\sum_{I=1}^{N}\hat{\mathbf{p}}_{I},$ (65)
where we have defined
$\hat{\omega}_{p}^{I}=m_{I}c^{2}\sqrt{1+\frac{\hat{\mathbf{p}}_{I}^{2}}{m_{I}^{2}c^{2}}}$
and
$\hat{\gamma}_{I}=\sqrt{1+\frac{\hat{\mathbf{p}}_{I}^{2}}{m_{I}^{2}c^{2}}}$.
We write the state in the Physical Hilbert space as
$\ket{\Psi}_{ph}\propto\int
d^{N}\mathcal{N}d^{2}ze^{\frac{\mathrm{i}}{\hbar}\mathcal{N}_{i}\hat{C}_{i}}e^{\frac{\mathrm{i}}{\hbar}z_{\mu}\hat{f}^{\mu}}\ket{\phi},$
(66)
where $\ket{\phi}$ is expanded in momentum basis as
$\ket{\phi}=\int\Pi_{I}\left[d^{2}p_{I}dE_{I}\right]\phi(p_{1},\cdots,p_{N},E_{1},\cdots,E_{N})\ket{p_{1},\cdots,p_{N}}\ket{E_{1},\cdots,E_{N}},$
(67)
In this case, the trivialisation operator is
$\hat{\mathcal{T}}_{1}=e^{\frac{\mathrm{i}}{\hbar}\hat{\mathbf{x}}_{1}\sum_{i}\hat{\mathbf{p}}_{i}}e^{\frac{\mathrm{i}}{\hbar}\hat{x}_{1}^{0}\left[\sum_{i}\left(\hat{p}_{0}^{i}+\hat{\gamma}_{i}^{-1}\frac{\hat{H}_{i}}{c}\right)+\,\hat{\gamma}_{1}^{-1}\frac{\hat{H}_{1}}{c}\right]},$
(68)
and the constraints are mapped to
$\begin{split}&\hat{\mathcal{T}}_{1}\hat{f}^{0}\hat{\mathcal{T}}_{1}^{\dagger}=\hat{k}^{1}_{0};\qquad\hat{\mathcal{T}}_{1}\hat{f}^{1}\hat{\mathcal{T}}_{1}^{\dagger}=\hat{\mathbf{k}}_{1};\qquad\hat{\mathcal{T}}_{1}\hat{C}_{i}\hat{\mathcal{T}}_{1}^{\dagger}=\hat{C}_{i}\,\,\,i\neq
1;\\\
&\hat{\mathcal{T}}_{1}\hat{C}_{1}\hat{\mathcal{T}}_{1}^{\dagger}=\hat{k}_{0}^{1}-\sum_{i}\left(\hat{k}_{0}^{i}+\hat{\gamma}_{i}^{-1}\frac{\hat{H}_{i}}{c}\right)-\left(1+\frac{(\hat{\mathbf{k}}_{1}-\sum_{i}\hat{\mathbf{k}}_{i})^{2}}{m_{1}^{2}c^{2}}\right)^{-1/2}\frac{\hat{H}_{1}}{c}-m_{1}c\sqrt{1+\frac{(\hat{\mathbf{k}}_{1}-\sum_{i}\hat{\mathbf{k}}_{i})^{2}}{m_{1}^{2}c^{2}}}.\end{split}$
(69)
Following analogous steps to the Galilean case, we define
$\ket{\psi}^{(1)}=\bra{q_{1}=0}\hat{\mathcal{T}}_{1}\ket{\Psi}_{ph}$, where
$\ket{\psi}^{(1)}=\int
d\tau_{1}d^{N-1}\mathcal{N}\ket{\tau_{1}}e^{\frac{\mathrm{i}}{\hbar}\mathcal{N}_{i}\hat{C}_{i}}e^{-\frac{\mathrm{i}}{\hbar}\hat{K}^{(1)}\tau_{1}}\ket{\phi^{(1)}_{0}},$
(70)
where
$\hat{K}^{(1)}=\hat{\gamma}_{\Sigma
k,1}\left[\sum_{i}\left(c\hat{k}_{0}^{i}+\hat{\gamma}_{i}^{-1}\hat{H}_{i}\right)+c\hat{\omega}_{\Sigma
k}^{1}\right],$ (71)
and
$\begin{split}&\ket{\phi^{(1)}_{0}}=\frac{1}{\sqrt{2\pi\hbar}}\int\Pi_{i}(d^{2}k_{i}dE_{i})dt_{1}dE_{1}\gamma_{\Sigma
k,1}e^{\frac{\mathrm{i}}{\hbar}(\hat{K}^{(1)}+E_{1})t_{1}}\phi\left(\bar{k}_{1},k_{2},\cdots,k_{N},E_{1},\cdots,E_{N}\right)\times\\\
&\hskip 56.9055pt\times\ket{k_{2}\cdots k_{N}}\ket{E_{2}\cdots
E_{N}}.\end{split}$ (72)
Here, we have defined $\hat{\gamma}_{\Sigma
k,1}=\sqrt{1+\frac{(\sum_{i}\hat{\mathbf{k}}_{i})^{2}}{m_{1}^{2}c^{2}}}$ and
$\hat{\omega}_{\Sigma
k}^{1}=m_{1}c\sqrt{1+\frac{(\sum_{i}\hat{\mathbf{k}}_{i})^{2}}{m_{1}^{2}c^{2}}}$
and the two-vector
$\bar{k}_{1}=\left(-\sum_{i}\left(k_{0}^{i}+\gamma_{i}^{-1}\frac{E_{i}}{c}\right)-\gamma_{\Sigma
k,1}^{-1}\frac{E_{1}}{c},-\sum_{i}\mathbf{k}_{i}\right)$.
It is again possible, with a similar procedure to the Galilean case, to
recover the dynamical evolution of a quantum relativistic particle. We find
$\ket{\psi}^{(1)}=\int
d\tau_{1}e^{-\frac{\mathrm{i}}{\hbar}\hat{H}^{(1)}\tau_{1}}\ket{\psi^{(1)}_{0}}\ket{\tau_{1}},$
(73)
where
$\displaystyle\hat{H}^{(1)}=\hat{\gamma}_{\Sigma
k,1}\left[\sum_{i}\left(c\hat{\omega}_{k}^{i}+\hat{\gamma}_{i}^{-1}\hat{H}_{i}\right)+c\hat{\omega}_{\Sigma
k}^{1}\right],$
$\displaystyle\begin{split}&\ket{\psi^{(1)}_{0}}=\int\Pi_{i}(d\mathbf{k}_{i}dE_{i})\phi^{(1)}_{0}\left(\omega_{k}^{2},\mathbf{k}_{2},\cdots,\omega_{k}^{N},\mathbf{k}_{N},E_{2},\cdots,E_{N}\right)\times\\\
&\hskip
56.9055pt\times\ket{\omega_{k}^{2},\mathbf{k}_{2},\cdots,\omega_{k}^{N},\mathbf{k}_{N}}\ket{E_{2}\cdots
E_{N}}.\end{split}$
Note that the factor $\sqrt{1+\frac{\left(\sum_{i\neq
1}\hat{k}_{i}\right)^{2}}{m_{1}^{2}c^{2}}}$ multiplying the hamiltonian of the
$N-1$ particles from the point of view of particle $1$ is the special-
relativistic time dilation due to the fact that the QRF is moving in a
superposition of special relativistic velocities.
### F.3 Newtonian case
Finally, we consider the case in which a set of $N$ particles move in a
Newtonian gravitational field produced by a mass $M$. The gravitational field
in the Newtonian limit is the same as in the main text
$\begin{split}&g_{00}=1+\frac{2\Phi(\mathbf{x}-\mathbf{x}_{M})}{c^{2}};\\\
&g_{01}=g_{10}=0;\\\ &g_{11}=-1,\end{split}$ (74)
where $\Phi(\mathbf{x})$ is the Newtonian potential due to the mass $m_{M}$ of
the system $M$ and $|\Phi(\mathbf{x})|/c^{2}\ll 1$ in the spacetime region
considered.
The set of constraints for this situation are
$\begin{split}&\hat{C}_{I}=\sqrt{g^{00}(\hat{\mathbf{x}}_{I}-\hat{\mathbf{x}}_{M})}\hat{p}^{I}_{0}-m_{I}c-\frac{\hat{\mathbf{p}}^{2}_{I}}{2m_{I}c}\qquad\text{for}\qquad
I=1,\cdots,N;\\\
&\hat{f}^{0}=\sum_{I=1}^{N}\left[\hat{p}^{I}_{0}+\sqrt{g_{00}(\hat{\mathbf{x}}_{I}-\hat{\mathbf{x}}_{M})}\frac{\hat{H}_{I}}{c}\right]+\hat{p}_{0}^{M};\\\
&\hat{f}^{1}=\sum_{I=1}^{N}\hat{\mathbf{p}}_{I}+\hat{\mathbf{p}}_{M}.\end{split}$
(75)
Notice that, since this is a special case of the case considered in the main
text, the constraints are still first-class to our order of approximation.
The state in the Physical Hilbert space is
$\ket{\Psi}_{ph}\propto\int
d^{N}\mathcal{N}d^{2}ze^{\frac{\mathrm{i}}{\hbar}\mathcal{N}_{i}\hat{C}_{i}}e^{\frac{\mathrm{i}}{\hbar}z_{\mu}\hat{f}^{\mu}}\ket{\phi},$
(76)
where, in this case, it is convenient to expand $\ket{\phi}$ in position basis
as
$\ket{\phi}=\int\Pi_{I}\left[d\mu(x_{I})dE_{I}\right]d^{2}x_{M}\phi(x_{1},\cdots,x_{N},x_{M},E_{1},\cdots,E_{N})\ket{x_{1},\cdots,x_{N},x_{M}}\ket{E_{1},\cdots,E_{N}},$
(77)
where $d\mu(x_{I})=\sqrt{g_{00}(\mathbf{x}_{I}-\mathbf{x}_{M})}d^{2}x_{I}$ is
the covariant integration measure.
We define the operator $\hat{\mathcal{T}}_{1}$ as
$\hat{\mathcal{T}}_{1}=e^{\frac{i}{\hbar}\frac{\log\sqrt{g_{00}(\hat{\mathbf{x}}_{M})}}{2}\sum_{I=1}^{N}(\hat{x}^{0}_{I}\hat{p}_{0}^{I}+\hat{p}_{0}^{I}\hat{x}^{0}_{I})}e^{\frac{\mathrm{i}}{\hbar}\hat{\mathbf{x}}_{1}\left(\hat{f}^{1}-\hat{\mathbf{p}}^{1}\right)}e^{\frac{\mathrm{i}}{\hbar}\hat{x}_{1}^{0}\left(\hat{f}^{0}-\hat{p}^{1}_{0}\right)},$
(78)
where in this case
$\hat{f}^{1}-\hat{\mathbf{p}}^{1}=\sum_{i}\hat{\mathbf{p}}_{i}+\hat{\mathbf{p}}_{M}$
and
$\hat{f}^{0}-\hat{p}^{1}_{0}=\sum_{i}\left[\hat{p}_{0}^{i}+\sqrt{g_{00}(\hat{\mathbf{x}}_{i}-\hat{\mathbf{x}}_{M})}\frac{\hat{H}_{i}}{c}\right]+\sqrt{g_{00}(\hat{\mathbf{x}}_{1}-\hat{\mathbf{x}}_{M})}\frac{\hat{H}_{1}}{c}+\hat{p}_{0}^{M}$.
The action of the operator $\hat{\mathcal{T}}_{1}$ on the constraints is
$\begin{split}&\hat{\mathcal{T}}_{1}\hat{f}^{0}\hat{\mathcal{T}}_{1}^{\dagger}=\sqrt{g_{00}(\hat{\mathbf{q}}_{M})}\hat{k}^{1}_{0};\qquad\hat{\mathcal{T}}_{1}\hat{f}^{1}\hat{\mathcal{T}}_{1}^{\dagger}=\hat{\mathbf{k}}^{1};\qquad\hat{\mathcal{T}}_{1}\hat{C}_{i}\hat{\mathcal{T}}_{1}^{\dagger}=\hat{C}^{\prime}_{i}\,\,\,i\neq
1;\\\
&\hat{\mathcal{T}}_{1}\hat{C}_{1}\hat{\mathcal{T}}_{1}^{\dagger}=\hat{k}_{0}^{1}-\left[\sum_{i}\left(\hat{k}_{0}^{i}+\sqrt{g^{\prime}_{00}(\hat{\mathbf{q}}_{i},\hat{\mathbf{q}}_{M})}\frac{\hat{H}_{i}}{c}\right)+\sqrt{g^{00}(\hat{\mathbf{q}}_{M})}\hat{k}_{0}^{M}\right]+\\\
&\hskip 56.9055pt-
m_{1}c-\frac{(\hat{\mathbf{k}}_{1}-\sum_{i}\hat{\mathbf{k}}_{i}-\hat{\mathbf{k}}_{M})^{2}}{2m_{1}c}-\frac{\hat{H}_{1}}{c},\end{split}$
(79)
where
$\hat{C}^{\prime}_{i}=\sqrt{g^{\prime}_{00}(\hat{\mathbf{q}}_{i},\hat{\mathbf{q}}_{M})}\hat{k}^{i}_{0}-m_{i}c-\frac{\hat{\mathbf{k}}^{2}_{i}}{2m_{i}c}$
and $g^{\prime
00}(\hat{\mathbf{q}}_{i},\hat{\mathbf{q}}_{M})=\frac{g_{00}(\hat{\mathbf{q}}_{i}-\hat{\mathbf{q}}_{M})}{g_{00}(\hat{\mathbf{q}}_{M})}$
as in the main text.
Similarly to the other cases, the relational state from the perspective of
system $1$ is
$\ket{\psi}^{(1)}=\bra{q_{1}=0}\hat{\mathcal{T}}_{1}\ket{\Psi}_{ph}$. In order
to find its explicit expression, we here define
$\begin{split}\ket{\tilde{\phi}}=\hat{\mathcal{T}}_{1}\ket{\phi}=&\int
d^{2}k_{1}\Pi_{i}\left[d\mu(q_{i})dE_{i}\right]dE_{1}d^{2}q_{M}\tilde{\phi}(k_{1},q_{2},\cdots,q_{N},q_{M},t_{1},E_{2},\cdots,E_{N})\times\\\
&\times\ket{k_{1},q_{2},\cdots,q_{N},q_{M}}\ket{t_{1}}\ket{E_{2},\cdots,E_{N}}.\end{split}$
(80)
We then find
$\ket{\psi}^{(1)}=\int
d\tau_{1}d^{N-1}\mathcal{N}e^{\frac{\mathrm{i}}{\hbar}\mathcal{N}_{i}\hat{C}^{\prime}_{i}}e^{-\frac{\mathrm{i}}{\hbar}\hat{K}^{(1)}\tau_{1}}\ket{\phi^{(1)}_{0}}\ket{\tau_{1}},$
(81)
where
$\begin{split}\ket{\phi^{(1)}_{0}}&=c\int
dt_{1}\Pi_{i}[d\mu(q_{i})dE_{i}]d^{2}q_{M}\sqrt{g^{00}(\mathbf{q}_{M})}e^{\frac{\mathrm{i}}{\hbar}\hat{K}^{(1)}t_{1}}\times\\\
&\times\tilde{\phi}(k_{1}=0,q_{2},\cdots,q_{N},q_{M},t_{1},E_{2}\cdots,E_{N})\ket{q_{2},\cdots,q_{N},q_{M}}\ket{E_{2},\cdots,E_{N}},\end{split}$
(82)
and $\hat{K}^{(1)}$ is the operator encoding the relational dynamics of
particles $2,\cdots,N$ from the point of view of particle $1$, i.e.,
$\hat{K}^{(1)}=\sum_{i}c\hat{k}_{0}^{i}+c\sqrt{g^{00}(\hat{\mathbf{q}}_{M})}\hat{k}_{0}^{M}+\sum_{i}\sqrt{g^{\prime}_{00}(\hat{\mathbf{q}}_{i},\hat{\mathbf{q}}_{M})}\hat{H}_{i}+m_{1}c^{2}+\frac{(\sum_{i}\hat{\mathbf{k}}_{i}+\hat{\mathbf{k}}_{M})^{2}}{2m_{1}}.$
(83)
With a similar procedure to the previous cases, we find that we can recover
the usual dynamical evolution as
$\ket{\psi}^{(1)}=\int
d\tau_{1}\ket{\tau_{1}}e^{-\frac{\mathrm{i}}{\hbar}\hat{H}^{(1)}\tau_{1}}\ket{\psi^{(1)}_{0}},$
(84)
where the relational Hamiltonian in the perspective of particle $1$ is
$\hat{H}^{(1)}=\sqrt{g^{00}(\hat{\mathbf{q}}_{M})}c\hat{k}_{0}^{M}+\sum_{i}\sqrt{g^{\prime}_{00}(\hat{\mathbf{q}}_{i},\hat{\mathbf{q}}_{M})}\left[m_{i}c^{2}+\frac{\hat{\mathbf{k}}_{i}^{2}}{2m_{i}}+\hat{H}_{i}\right]+m_{1}c^{2}+\frac{(\sum_{i}\hat{\mathbf{k}}_{i}+\hat{\mathbf{k}}_{M})^{2}}{2m_{1}},$
(85)
and
$\begin{split}&\ket{\psi^{(1)}_{0}}=\int\Pi_{i}\left[d\mu(q_{i})dq^{\prime
0}_{i}\sqrt{g^{\prime}_{00}(\mathbf{q}_{i}-\mathbf{q}_{M})}dE_{i}\right]d^{2}q_{M}e^{\frac{\mathrm{i}}{\hbar}\sum_{i}(q^{\prime
0}_{i}-q^{0}_{i})\sqrt{g^{\prime}_{00}(\mathbf{q}_{i}-\mathbf{q}_{M})}\left[m_{i}c+\frac{\hat{\mathbf{k}}_{i}^{2}}{2m_{i}c}\right]}\times\\\
&\times\phi^{(1)}_{0}\left(q_{2},\cdots,q_{N},q_{M},E_{2},\cdots,E_{N}\right)\ket{q^{\prime
0}_{2},\mathbf{q}_{2},\cdots,q^{\prime
0}_{N},\mathbf{q}_{N}}\ket{q_{M}}\ket{E_{2}\cdots E_{N}}.\end{split}$ (86)
Notice that, if we expand the Hamiltonian to lowest order in the general
relativistic corrections for the external degrees of freedom, while still
allowing the internal degrees of freedom to have relativistic corrections, we
obtain
$\begin{split}\hat{H}^{(1)}=&\sum_{i}\left\\{m_{i}c^{2}+\frac{\hat{\mathbf{k}}_{i}^{2}}{2m_{i}}+m_{i}[\Phi(\hat{\mathbf{q}}_{i}-\hat{\mathbf{q}}_{M})-\Phi(\hat{\mathbf{q}}_{M})]+\left[1+\frac{[\Phi(\hat{\mathbf{q}}_{i}-\hat{\mathbf{q}}_{M})-\Phi(\hat{\mathbf{q}}_{M})]}{c^{2}}\right]\hat{H}_{i}\right\\}+\\\
&+m_{1}c^{2}+\frac{(\sum_{i}\hat{\mathbf{k}}_{i}+\hat{\mathbf{k}}_{M})^{2}}{2m_{1}}+\sqrt{g^{00}(\hat{\mathbf{q}}_{M})}c\hat{k}_{0}^{M}.\end{split}$
(87)
As a result, we see that the potential in the QRF of particle $1$ is the
difference between the gravitational potential between particle $i$ and the
mass $M$ and particle $1$ and the mass $M$. This is the straightforward
generalisation of the standard reference frame description.
## Appendix G Explicit calculation of the “history state” with a quantum
measurement
The calculation of the history state in the case with a measurement parallels
the one outlined in Appendix E. With a fully analogous procedure, fixing
$N=2$, and choosing the constraints of Eq. (17)
$\begin{split}&\hat{C}_{I}=\sqrt{g^{00}(\hat{\mathbf{x}}_{I}-\hat{\mathbf{x}}_{M})}\hat{p}^{I}_{0}-\hat{\omega}_{p}^{I}\qquad\text{for}\qquad
I=1,2;\\\
&\hat{f}_{Q}^{0}=\hat{p}^{1}_{0}+\hat{p}^{2}_{0}+\hat{p}_{0}^{M}+\Delta(\hat{\mathbf{x}}_{1}-\hat{\mathbf{x}}_{M},\hat{\mathbf{p}}_{1})\frac{\hat{H}_{1}}{c}+\Delta(\hat{\mathbf{x}}_{2}-\hat{\mathbf{x}}_{M},\hat{\mathbf{p}}_{2})\left[\frac{\hat{H}_{2}}{c}+\delta(\hat{T}_{2}-\tau_{2}*)\frac{\hat{Q}_{2}}{c}\right];\\\
&\hat{f}^{1}=\hat{\mathbf{p}}_{1}+\hat{\mathbf{p}}_{2}+\hat{\mathbf{p}}_{M},\end{split}$
(88)
where $\hat{Q}_{2}$ is an observable commuting with every constraint, and the
operator
$\hat{\mathcal{T}}_{Q,1}=e^{\frac{i}{\hbar}\frac{\log\sqrt{g_{00}(\hat{\mathbf{x}}_{M})}}{2}\sum_{I=1}^{N}(\hat{x}^{0}_{I}\hat{p}_{0}^{I}+\hat{p}_{0}^{I}\hat{x}^{0}_{I})}e^{\frac{\mathrm{i}}{\hbar}\hat{\mathbf{x}}_{1}\left(\hat{f}^{1}-\hat{\mathbf{p}}^{1}\right)}e^{\frac{\mathrm{i}}{\hbar}\hat{x}_{1}^{0}\left(\hat{f}_{Q}^{0}-\hat{p}^{1}_{0}\right)},$
(89)
we find that the history state can be cast in the form
$\ket{\psi}^{(1)}=\int
d\tau_{1}d^{N-1}\mathcal{N}e^{\frac{\mathrm{i}}{\hbar}\mathcal{N}_{i}\hat{C}^{\prime}_{i}}e^{-\frac{\mathrm{i}}{\hbar}\hat{K}_{Q}^{(1)}\tau_{1}}\ket{\phi^{(1)}_{0}}\ket{\tau_{1}},$
(90)
where $\hat{C}^{\prime}_{i}$ is the same as in the main text and in Appendix
E. In this case, we have that
$\hat{K}_{Q}^{(1)}=\hat{K}^{(1)}+\sqrt{g^{\prime}_{00}(\hat{\mathbf{q}}_{2},\hat{\mathbf{q}}_{M})}\frac{\hat{\gamma}_{2}^{-1}}{\hat{\gamma}_{\Sigma
k,1}^{-1}}\delta(\hat{T}_{2}-\tau_{2}^{*})\hat{Q}_{2},$ (91)
where $\hat{K}^{(1)}$ was defined in Eq. (45) for a general number of
particles $N$ (which here we have fixed to $N=2$). We also recall that we have
defined
$\hat{\gamma}_{2}=\sqrt{1+\frac{\hat{\mathbf{k}}_{2}^{2}}{m_{2}^{2}c^{2}}}$
and $\hat{\gamma}_{\Sigma
k,1}=\sqrt{1+\frac{\left(\hat{\mathbf{k}}_{2}+\hat{\mathbf{k}}_{M}\right)^{2}}{m_{1}^{2}c^{2}}}$.
Finally, we also write
$\ket{\phi^{(1)}_{0}}=c\int dt_{1}d^{2}k_{2}dt_{2}d^{2}k_{M}\gamma_{\Sigma
k,1}e^{\frac{\mathrm{i}}{\hbar}\hat{K}_{Q}^{(1)}t_{1}}\phi_{/1}(k_{2},k_{M},t_{1},t_{2})\ket{k_{2},k_{M},t_{2}}$
(92)
where the state $\ket{\phi_{/1}}$ has the same expression as in Eq. (46) with
the operator $\hat{\mathcal{T}}_{Q,1}$ instead of $\hat{\mathcal{T}}_{1}$. We
now consider the following equality, proved in Ref. [36],
$e^{-\frac{\mathrm{i}}{\hbar}\alpha[\hat{H}_{I}+\hat{f}(\hat{T}_{I})]}=e^{-\frac{\mathrm{i}}{\hbar}\alpha\hat{H}_{I}}\overleftarrow{T}\left\\{e^{-\frac{\mathrm{i}}{\hbar}\int_{0}^{1}ds\alpha\hat{f}(\hat{T}_{I}+\alpha
s)}\right\\},$ (93)
where $[\hat{T}_{I},\hat{H}_{I}]=\mathrm{i}\hbar$ and $\hat{f}(\hat{T}_{I})$
is an operator-valued function, and we slightly generalise it to
$e^{-\frac{\mathrm{i}}{\hbar}\alpha[\hat{K}_{R}+\hat{H}_{I}+\hat{f}(\hat{T}_{I})]}=e^{-\frac{\mathrm{i}}{\hbar}\alpha\hat{H}_{I}}\overleftarrow{T}\left\\{e^{-\frac{\mathrm{i}}{\hbar}\int_{0}^{1}ds\alpha\left[\hat{K}_{R}+\hat{f}(\hat{T}_{I}+\alpha
s)\right]}\right\\},$ (94)
where $\hat{K}_{R}$ is an arbitrary hermitian operator which does not
necessarily commute with $\hat{f}(\hat{T}_{I})$. Then, the expression of Eq:
(90) can be cast as
$\ket{\psi}^{(1)}=\int
d\tau_{1}\overleftarrow{T}\left\\{e^{-\frac{\mathrm{i}}{\hbar}\int_{0}^{\tau_{1}}ds\left[\hat{H}^{(1)}+\Delta_{12}\delta(\hat{T}_{2}+\Delta_{12}s-\tau_{2}^{*})\hat{Q}_{2}\right]}\right\\}\ket{\psi^{(1)}_{0}}\ket{\tau_{1}},$
(95)
where
$\Delta_{12}=\sqrt{g^{\prime}_{00}(\hat{\mathbf{q}}_{2},\hat{\mathbf{q}}_{M})}\frac{\hat{\gamma}_{2}^{-1}}{\hat{\gamma}_{\Sigma
k,1}^{-1}}$. The Hamiltonian $\hat{H}^{(1)}$ is the same as in the main text
$\hat{H}^{(1)}=\hat{\gamma}_{\Sigma
k,1}\sum_{i}\sqrt{g^{\prime}_{00}(\hat{\mathbf{q}}_{i},\hat{\mathbf{q}}_{M})}\left\\{c\hat{\omega}_{k}^{i}+\hat{\gamma}_{i}^{-1}\hat{H}_{i}\right\\}+c\hat{\gamma}_{\Sigma
k,1}\sqrt{g^{00}(\hat{\mathbf{q}}_{M})}\hat{k}_{0}^{M}+m_{1}c^{2}\hat{\gamma}_{\Sigma
k,1}^{2},$ (96)
and
$\begin{split}\ket{\psi^{(1)}_{0}}=&\int d\mu(q_{2})dq^{\prime
0}_{2}dt_{2}d^{2}q_{M}\sqrt{g^{\prime}_{00}(\mathbf{q}_{2}-\mathbf{q}_{M})}e^{\frac{\mathrm{i}}{\hbar}(q^{\prime
0}_{2}-q^{0}_{2})\sqrt{g^{\prime}_{00}(\mathbf{q}_{2}-\mathbf{q}_{M})}\hat{\omega}_{k}^{2}}\times\\\
&\times\phi^{(1)}_{0}\left(q_{2},q_{M},t_{2}\right)\ket{q^{\prime
0}_{2},\mathbf{q}_{2}}\ket{q_{M}}\ket{t_{2}},\end{split}$ (97)
## References
* Aharonov and Susskind [1967a] Yakir Aharonov and Leonard Susskind. Charge Superselection Rule. _Phys. Rev._ , 155:1428–1431, 1967a. doi: 10.1103/PhysRev.155.1428. URL https://link.aps.org/doi/10.1103/PhysRev.155.1428.
* Aharonov and Susskind [1967b] Yakir Aharonov and Leonard Susskind. Observability of the sign change of spinors under $2\pi$ rotations. _Phys. Rev._ , 158:1237–1238, 1967b. doi: 10.1103/PhysRev.158.1237. URL https://link.aps.org/doi/10.1103/PhysRev.158.1237.
* Aharonov and Kaufherr [1984] Y. Aharonov and T. Kaufherr. Quantum frames of reference. _Phys. Rev. D_ , 30:368–385, 1984. doi: 10.1103/PhysRevD.30.368. URL https://link.aps.org/doi/10.1103/PhysRevD.30.368.
* Bartlett et al. [2007] Stephen D. Bartlett, Terry Rudolph, and Robert W. Spekkens. Reference frames, superselection rules, and quantum information. _Rev. Mod. Phys._ , 79:555–609, 2007. doi: 10.1103/RevModPhys.79.555. URL https://link.aps.org/doi/10.1103/RevModPhys.79.555.
* Bartlett et al. [2009] Stephen D. Bartlett, Terry Rudolph, Robert W. Spekkens, and Peter S. Turner. Quantum communication using a bounded-size quantum reference frame. _New J. Phys._ , 11(6):063013, 2009. doi: 10.1088/1367-2630/11/6/063013. URL http://stacks.iop.org/1367-2630/11/i=6/a=063013.
* Gour and Spekkens [2008] Gilad Gour and Robert W Spekkens. The resource theory of quantum reference frames: manipulations and monotones. _New J. Phys._ , 10(3):033023, 2008. doi: 10.1088/1367-2630/10/3/033023. URL http://stacks.iop.org/1367-2630/10/i=3/a=033023.
* Kitaev et al. [2004] Alexei Kitaev, Dominic Mayers, and John Preskill. Superselection rules and quantum protocols. _Phys. Rev. A_ , 69(5):052326, 2004. doi: 10.1103/PhysRevA.69.052326.
* Palmer et al. [2014] Matthew C. Palmer, Florian Girelli, and Stephen D. Bartlett. Changing quantum reference frames. _Phys. Rev. A_ , 89:052121, 2014. doi: 10.1103/PhysRevA.89.052121. URL https://link.aps.org/doi/10.1103/PhysRevA.89.052121.
* Bartlett et al. [2006] Stephen D Bartlett, Terry Rudolph, Robert W Spekkens, and Peter S Turner. Degradation of a quantum reference frame. _New J. Phys._ , 8(4):58, 2006. doi: 10.1088/1367-2630/8/4/058. URL http://stacks.iop.org/1367-2630/8/i=4/a=058.
* Smith et al. [2016] Alexander R. H. Smith, Marco Piani, and Robert B. Mann. Quantum reference frames associated with noncompact groups: The case of translations and boosts and the role of mass. _Phys. Rev. A_ , 94:012333, 2016. doi: 10.1103/PhysRevA.94.012333. URL https://link.aps.org/doi/10.1103/PhysRevA.94.012333.
* Poulin and Yard [2007] David Poulin and Jon Yard. Dynamics of a quantum reference frame. _New J. Phys._ , 9(5):156, 2007. doi: 10.1088/1367-2630/9/5/156. URL http://stacks.iop.org/1367-2630/9/i=5/a=156.
* Girelli and Poulin [2008] Florian Girelli and David Poulin. Quantum reference frames and deformed symmetries. _Phys. Rev. D_ , 77:104012, 2008. doi: 10.1103/PhysRevD.77.104012. URL https://link.aps.org/doi/10.1103/PhysRevD.77.104012.
* Skotiniotis et al. [2013] Michael Skotiniotis, Borzu Toloui, Ian T. Durham, and Barry C. Sanders. Quantum Frameness for $CPT$ Symmetry. _Phys. Rev. Lett._ , 111:020504, 2013. doi: 10.1103/PhysRevLett.111.020504. URL https://link.aps.org/doi/10.1103/PhysRevLett.111.020504.
* Poulin [2006] David Poulin. Toy model for a relational formulation of quantum theory. _Int. J. Theor. Phys._ , 45(7):1189–1215, 2006\. doi: 10.1007/s10773-006-9052-0.
* Miyadera et al. [2016] Takayuki Miyadera, Leon Loveridge, and Paul Busch. Approximating relational observables by absolute quantities: a quantum accuracy-size trade-off. _J. Phys. A_ , 49(18):185301, 2016. doi: 10.1088/1751-8113/49/18/185301. URL http://stacks.iop.org/1751-8121/49/i=18/a=185301.
* Loveridge et al. [2017] L. Loveridge, P. Busch, and T. Miyadera. Relativity of quantum states and observables. _EPL (Europhysics Letters)_ , 117(4):40004, 2017\. doi: 10.1209/0295-5075/117/40004. URL http://stacks.iop.org/0295-5075/117/i=4/a=40004.
* Loveridge et al. [2018] Leon Loveridge, Takayuki Miyadera, and Paul Busch. Symmetry, reference frames, and relational quantities in quantum mechanics. _Found. Phys._ , 48(2):135–198, 2018. doi: 10.1007/s10701-018-0138-3.
* [18] Jacques Pienaar. A relational approach to quantum reference frames for spins. arXiv:1601.07320, 2016.
* Angelo et al. [2011] Renato M Angelo, Nicolas Brunner, Sandu Popescu, Anthony J Short, and Paul Skrzypczyk. Physics within a quantum reference frame. _J. Phys. A_ , 44(14):145304, 2011. doi: 10.1088/1751-8113/44/14/145304. URL https://doi.org/10.1088/1751-8113/44/14/145304.
* Angelo and Ribeiro [2012] R M Angelo and A D Ribeiro. Kinematics and dynamics in noninertial quantum frames of reference. _J. Phys. A_ , 45(46):465306, 2012. doi: 10.1088/1751-8113/45/46/465306. URL https://doi.org/10.1088/1751-8113/45/46/465306.
* Pereira and Angelo [2015] S. T. Pereira and R. M. Angelo. Galilei covariance and Einstein’s equivalence principle in quantum reference frames. _Phys. Rev. A_ , 91:022107, 2015. doi: 10.1103/PhysRevA.91.022107. URL https://link.aps.org/doi/10.1103/PhysRevA.91.022107.
* DeWitt [1967] Bryce S DeWitt. Quantum theory of gravity. I. The canonical theory. _Phys. Rev._ , 160(5):1113, 1967. doi: 10.1103/PhysRev.160.1113.
* Rovelli [1991] C Rovelli. Quantum reference systems. _Class. Quant. Grav._ , 8(2):317, 1991. doi: 10.1088/0264-9381/8/2/012. URL http://stacks.iop.org/0264-9381/8/i=2/a=012.
* Rovelli [1996] Carlo Rovelli. Relational quantum mechanics. _Int. J. Theor. Phys._ , 35(8):1637–1678, 1996\. doi: 10.1007/BF02302261.
* Giacomini et al. [2019a] Flaminia Giacomini, Esteban Castro-Ruiz, and Časlav Brukner. Quantum mechanics and the covariance of physical laws in quantum reference frames. _Nat. Commun._ , 10(1):494, 2019a. doi: 10.1038/s41467-018-08155-0. URL https://doi.org/10.1038/s41467-018-08155-0.
* Vanrietvelde et al. [2020] Augustin Vanrietvelde, Philipp A Höhn, Flaminia Giacomini, and Esteban Castro-Ruiz. A change of perspective: switching quantum reference frames via a perspective-neutral framework. _Quantum_ , 4:225, 2020. doi: 10.22331/q-2020-01-27-225.
* [27] Augustin Vanrietvelde, Philipp A Höhn, and Flaminia Giacomini. Switching quantum reference frames in the N-body problem and the absence of global relational perspectives. arXiv:1809.05093, 2018.
* Yang [2020] Jianhao M. Yang. Switching Quantum Reference Frames for Quantum Measurement. _Quantum_ , 4:283, June 2020. doi: 10.22331/q-2020-06-18-283. URL https://doi.org/10.22331/q-2020-06-18-283.
* Giacomini et al. [2019b] Flaminia Giacomini, Esteban Castro-Ruiz, and Časlav Brukner. Relativistic quantum reference frames: the operational meaning of spin. _Phys. Rev. Lett._ , 123(9):090404, 2019b. doi: 10.1103/PhysRevLett.123.090404.
* Streiter et al. [2021] Lucas F. Streiter, Flaminia Giacomini, and Časlav Brukner. Relativistic bell test within quantum reference frames. _Phys. Rev. Lett._ , 126:230403, Jun 2021. doi: 10.1103/PhysRevLett.126.230403. URL https://link.aps.org/doi/10.1103/PhysRevLett.126.230403.
* de la Hamette and Galley [2020] Anne-Catherine de la Hamette and Thomas D. Galley. Quantum reference frames for general symmetry groups. _Quantum_ , 4:367, November 2020. doi: 10.22331/q-2020-11-30-367. URL https://doi.org/10.22331/q-2020-11-30-367.
* [32] Marius Krumm, Philipp A. Hoehn, and Markus P. Mueller. Quantum reference frame transformations as symmetries and the paradox of the third particle. arXiv:2011.01951, 2020.
* Ballesteros et al. [2021] Angel Ballesteros, Flaminia Giacomini, and Giulia Gubitosi. The group structure of dynamical transformations between quantum reference frames. _Quantum_ , 5:470, 2021. doi: 10.22331/q-2021-06-08-470.
* Page and Wootters [1983] Don N. Page and William K. Wootters. Evolution without evolution: Dynamics described by stationary observables. _Phys. Rev. D_ , 27:2885, 1983. doi: 10.1103/PhysRevD.27.2885.
* Giovannetti et al. [2015] Vittorio Giovannetti, Seth Lloyd, and Lorenzo Maccone. Quantum time. _Phys. Rev. D_ , 92(4):045033, 2015. doi: 10.1103/PhysRevD.92.045033.
* Castro-Ruiz et al. [2020] Esteban Castro-Ruiz, Flaminia Giacomini, Alessio Belenchia, and Časlav Brukner. Quantum clocks and the temporal localisability of events in the presence of gravitating quantum systems. _Nat. Commun._ , 11(1):1–12, 2020. doi: 10.1038/s41467-020-16013-1.
* [37] Flaminia Giacomini and Časlav Brukner. Einstein’s Equivalence principle for superpositions of gravitational fields and quantum reference frames. arXiv:2012.13754, 2020.
* Höhn and Vanrietvelde [2020] Philipp A. Höhn and Augustin Vanrietvelde. How to switch between relational quantum clocks. _New J. Phys._ , 22(12):123048, 2020. doi: 10.1088/1367-2630/abd1ac.
* Höhn [2019] Philipp A Höhn. Switching internal times and a new perspective on the ‘wave function of the universe’. _Universe_ , 5(5):116, 2019. doi: 10.3390/universe5050116.
* Smith and Ahmadi [2019] Alexander R. H. Smith and Mehdi Ahmadi. Quantizing time: Interacting clocks and systems. _Quantum_ , 3:160, 2019. doi: 10.22331/q-2019-07-08-160.
* [41] Philipp A. Höhn, Alexander R. H. Smith, and Maximilian P. E. Lock. The Trinity of Relational Quantum Dynamics. arXiv:1912.00033, 2019.
* Hoehn et al. [2021] Philipp A. Hoehn, Alexander R. H. Smith, and Maximilian P. E. Lock. Equivalence of Approaches to Relational Quantum Dynamics in Relativistic Settings. _Front. in Phys._ , 9:181, 2021. doi: 10.3389/fphy.2021.587083.
* Smith and Ahmadi [2020] Alexander R. H. Smith and Mehdi Ahmadi. Quantum clocks observe classical and quantum time dilation. _Nat. Commun._ , 11(1):1–9, 2020. doi: 10.1038/s41467-020-18264-4.
* [44] Lucien Hardy. The Construction Interpretation: Conceptual Roads to Quantum Gravity. arXiv:1807.10980, 2020.
* [45] Magdalena Zych, Fabio Costa, and Timothy C Ralph. Relativity of quantum superpositions. arXiv:1809.04999, 2018.
* Isham [1993] C. J. Isham. Canonical quantum gravity and the problem of time. volume 409, pages 157–287. 1993.
* Rovelli [2004] Carlo Rovelli. _Quantum Gravity_. Cambridge University Press, 2004. doi: 10.1017/CBO9780511755804.
* Kuchař [2011] Karel V Kuchař. Time and interpretations of quantum gravity. _Int. J. Mod. Phys. D_ , 20(supp01):3–86, 2011\. doi: 10.1142/S0218271811019347.
* Rovelli [1990] Carlo Rovelli. Quantum mechanics without time: a model. _Phys. Rev. D_ , 42(8):2638, 1990. doi: 10.1103/PhysRevD.42.2638.
* Reisenberger and Rovelli [2002] Michael Reisenberger and Carlo Rovelli. Spacetime states and covariant quantum theory. _Phys. Rev. D_ , 65(12):125016, 2002. doi: 10.1103/PhysRevD.65.125016.
* Hellmann et al. [2007] Frank Hellmann, Mauricio Mondragon, Alejandro Perez, and Carlo Rovelli. Multiple-event probability in general-relativistic quantum mechanics. _Phys. Rev. D_ , 75(8):084033, 2007. doi: 10.1103/PhysRevD.75.084033.
* Zych et al. [2019] Magdalena Zych, Łukasz Rudnicki, and Igor Pikovski. Gravitational mass of composite systems. _Phys. Rev. D_ , 99(10):104029, 2019. doi: 10.1103/PhysRevD.99.104029.
* Zych et al. [2011] Magdalena Zych, Fabio Costa, Igor Pikovski, and Časlav Brukner. Quantum interferometric visibility as a witness of general relativistic proper time. _Nat. Commun._ , 2:505, 2011. doi: 10.1038/ncomms1498.
* Pikovski et al. [2017] Igor Pikovski, Magdalena Zych, Fabio Costa, and Časlav Brukner. Time dilation in quantum systems and decoherence. _New J. Phys._ , 19(2):025011, 2017. doi: 10.1088/1367-2630/aa5d92. URL http://stacks.iop.org/1367-2630/19/i=2/a=025011.
* Zych et al. [2016] Magdalena Zych, Igor Pikovski, Fabio Costa, and Časlav Brukner. General relativistic effects in quantum interference of “clocks”. In _Journal of Physics: Conference Series_ , volume 723, page 012044\. IOP Publishing, 2016. doi: 10.1088/1742-6596/723/1/012044.
* Pikovski et al. [2015] Igor Pikovski, Magdalena Zych, Fabio Costa, and Časlav Brukner. Universal decoherence due to gravitational time dilation. _Nat. Phys._ , 11(8):668, 2015. doi: 10.1038/nphys3366.
* Ashtekar [1991] A. Ashtekar. _Lectures on nonperturbative canonical gravity_ , volume 6. 1991\. doi: 10.1142/1321.
* Marolf [1995] Donald Marolf. Refined algebraic quantization: Systems with a single constraint. _Banach Center Publications_ , 39, 09 1995. doi: 10.4064/-39-1-331-344.
* Hartle and Marolf [1997] James B. Hartle and Donald Marolf. Comparing formulations of generalized quantum mechanics for reparametrization - invariant systems. _Phys.Rev._ , D56:6247–6257, 1997. doi: 10.1103/PhysRevD.56.6247.
* Kempf and Klauder [2001] Achim Kempf and John R Klauder. On the implementation of constraints through projection operators. _J Phys A_ , 34(5):1019, 2001. doi: 10.1088/0305-4470/34/5/307.
* Tino [2020] Guglielmo M Tino. Testing gravity with cold atom interferometry: Results and prospects. _Quantum Science and Technology_ , 2020. doi: 10.1088/2058-9565/abd83e.
* Guérin and Brukner [2018] Philippe Allard Guérin and Časlav Brukner. Observer-dependent locality of quantum events. _New J. Phys._ , 20(10):103031, 2018. doi: 10.1088/1367-2630/aae742. URL http://stacks.iop.org/1367-2630/20/i=10/a=103031.
* Grochowski et al. [2021] Piotr T. Grochowski, Alexander R. H. Smith, Andrzej Dragan, and Kacper Dębski. Quantum time dilation in atomic spectra. _Phys. Rev. Research_ , 3:023053, Apr 2021. doi: 10.1103/PhysRevResearch.3.023053. URL https://link.aps.org/doi/10.1103/PhysRevResearch.3.023053.
* Tino et al. [2020] GM Tino, L Cacciapuoti, S Capozziello, G Lambiase, and F Sorrentino. Precision gravity tests and the Einstein Equivalence Principle. _Prog. Part. Nucl. Phys._ , page 103772, 2020. doi: 10.1016/j.ppnp.2020.103772.
* Castro Ruiz et al. [2017] Esteban Castro Ruiz, Flaminia Giacomini, and Časlav Brukner. Entanglement of quantum clocks through gravity. _PNAS_ , 114(12):E2303–E2309, 2017. ISSN 0027-8424. doi: 10.1073/pnas.1616427114. URL https://www.pnas.org/content/114/12/E2303.
* Mercati [2018] Flavio Mercati. _Shape dynamics: Relativity and relationalism_. Oxford University Press, 2018. doi: 10.1093/oso/9780198789475.001.0001.
* Barbour et al. [2014] Julian Barbour, Tim Koslowski, and Flavio Mercati. Identification of a gravitational arrow of time. _Phys. Rev. Lett._ , 113(18):181101, 2014. doi: 10.1103/PhysRevLett.113.181101.
|
figuret
# Using an atom interferometer to infer gravitational entanglement generation
Daniel Carney Joint Center for Quantum Information and Computer Science/Joint
Quantum Institute, University of Maryland/NIST, College Park, MD 20742, USA
Fermi National Accelerator Laboratory, Batavia, IL 60510, USA Holger Müller
Department of Physics, University of California, Berkeley, CA 94720, USA
Jacob M. Taylor Joint Center for Quantum Information and Computer
Science/Joint Quantum Institute, University of Maryland/NIST, College Park, MD
20742, USA
###### Abstract
If gravitational perturbations are quantized into gravitons in analogy with
the electromagnetic field and photons, the resulting graviton interactions
should lead to an entangling interaction between massive objects. We suggest a
test of this prediction. To do this, we introduce the concept of interactive
quantum information sensing. This novel sensing protocol is tailored to
provable verification of weak dynamical entanglement generation between a pair
of systems. We show that this protocol is highly robust to typical thermal
noise sources. The sensitivity can moreover be increased both using an initial
thermal state and/or an initial phase of entangling via a non-gravitational
interaction. We outline a concrete implementation testing the ability of the
gravitational field to generate entanglement between an atomic interferometer
and mechanical oscillator. Preliminary numerical estimates suggest that near-
term devices could feasibly be used to perform the experiment.
## I Introduction
If a particle is in a superposition of two locations, will its gravitational
field also be in a superposition, and can this field generate entanglement
with another system? This foundational question feynman1971lectures ;
page1981indirect has received considerable attention diosi1984gravitation ;
penrose1996gravity ; kafri2015bounds ; bose2017spin ; Marletto:2017kzi ;
Haine:2018bwu ; Chevalier:2020uvv ; Howl:2020isj ; Anastopoulos:2020cdp ;
Matsumura:2020law ; carney2019tabletop . Proposed experimental tests to detect
entanglement due to gravity based on Bell tests (or more generally,
entanglement witnesses HORODECKI19961 ; terhal2000bell ) require performing
measurements on both subsystems and are challenging in practice. As a result,
there is still no direct experimental evidence as to whether gravitational
interactions generate entanglement. Here, we propose a test that only requires
observing a single subsystem feynman1963theory ; caldeira1983path ;
joos1985emergence ; zurek2003decoherence . We show that, if an interaction
(such as gravity) between two systems can cause both decoherence (collapse)
and recoherence (revival) of a subsystem, then for restricted classes of
systems the interaction is _necessarily_ capable of generating entanglement.
We propose a concrete implementation based on atom interferometry
kasevich1992measurement ; santarelli1999quantum ; gross2010nonlinear ;
xu2019probing , in which an atom in a superpositon of being in one of two
interferometer arms interacts with a low-frequency mechanical resonator
lee2020new ; catano2020high ; the signal for entanglement-generation is a
collapse and revival of the atomic interference fringes due to the periodic
motion of the resonator. The experiment does not require preparing a non-
classical state of the oscillator and can in fact be enhanced by placing the
oscillator in a thermal state, which appears to make this experiment feasible
with near-term devices.
The relation of such an experiment to the quantization gravity is a subject of
intense current study belenchia2018quantum ; Christodoulou:2018cmk ;
Marshman:2019sne ; Galley:2020qsf . These experiments operate in a regime
where the energy density (or equivalently, spacetime curvature), is far below
the Planck scale $\rho\ll m_{\rm Pl}/\ell_{\rm Pl}^{3}\sim 10^{123}~{}{\rm
eV}/{\rm cm^{3}}$. Thus the non-linearity of the gravitational interaction is
very weak, and one can treat the metric $g_{\mu\nu}$ as a linear perturbation
around flat spacetime. In this limit, one can quantize the gravitational
perturbations (“gravitons”) in exact analogy with quantum electrodynamics;
graviton exchange generates a two-body Newton potential operator
$V_{N}=-\frac{G_{N}m_{1}m_{2}}{|\mbox{\boldmath$x$}_{1}-\mbox{\boldmath$x$}_{2}|}$
(1)
between a pair of masses, just as photons generate the Coulomb potential
Feynman:1963ax ; t1974one ; deser1974one ; Veltman:1975vx ;
donoghue1994general ; burgess2004quantum . We review some standard
demonstrations of this in appendix A. In equation (1),
$\mbox{\boldmath$x$}_{1,2}$ are the position _operators_ on a pair of masses,
and thus this interaction can generate entanglement. However, there are
dissenting opinions penrose1996gravity ; howl2019exploring ; Tilloy:2019hxe ;
Bruschi:2020xbm about whether gravity should be quantized in this way, and
indeed one can produce models where classical gravitational interactions can
arise but without generating entanglement Kibble:1979jn ; kafri2014classical ;
Oppenheim:2018igd ; Kent:2020gov , providing substantial motivation to perform
tests of (1).
Figure 1: Implementation of the basic protocol using an atom interferometer
and a suspended pendulum (see Section VI). A trapped atom (labeled A) is
prepared some distance $L$ away from a mechanical resonator (B, here pictured
as a pendulum). The atom is then put into a superposition of two different
locations separated by $\ell$, effecting a Hadamard gate $H$. This generates a
state-dependent force between the atoms and resonator, leading to motion in
opposite directions for some time $\Delta t$. Finally, the atom state is
recombined using the inverse Hadamard gate and measured to check for
decoherence caused by the atom-mechanical interaction. When the resonator
undergoes a complete period of motion, its state no longer depends upon the
atoms and coherence is recovered for the interferometer.
The ability to test such a weak entanglement signal relies entirely on our
central technical result, a novel sensing protocol which we refer to as
_interactive quantum information sensing_. This is a detection scheme tailored
specifically to the verification of weak dynamical entanglement generation.
The traditional methods to detect entanglement in bipartite systems
$H_{A}\otimes H_{B}$ use non-local measurements HORODECKI19961 ;
terhal2000bell , and can be very difficult in practice with noisy systems and
weak entanglement. However, in the past two decades, more sophisticated
methods have been developed to address these types of problems
guhne2009entanglement ; pezze2018quantum . We suggest here a new protocol
which relies on time-dependent measurements on a single subsystem. Within
standard quantum mechanics, system $A$ will decohere—evolve from a pure to
mixed state—if it becomes entangled with another system $B$ which is not
measured feynman1963theory ; caldeira1983path ; joos1985emergence ;
zurek2003decoherence . This loss of coherence can be observed via an
interference measurement on $A$ alone. Simple decoherence could be explained
by entanglement but also by, for example, random classical noise
stern1990phase . However, if the same interaction can cause both decoherence
and recoherence of $A$, in a manner controlled by $B$, then for certain
classes of systems we prove that the interaction is necessarily capable of
generating entanglement between subsystems $A$ and $B$. This protocol provides
an indirect test of the quantum communication capabilities of the two systems,
and is a limited probe of the family of quantum channels associated with the
interaction between the two systems. The interplay between the information
theoretic channel properties and the physical interaction provides our
suggested nomenclature.
We outline the interactive sensing protocol in sections II and III. We find
the remarkable result that using an initial state at high temperature can
_increase_ the sensitivity of the protocol, because it can increase the rate
of entanglement generation and lead to a thermally-enhanced collapse and
revival signal. In section IV we demonstrate that this conclusion is robust to
typical sources of noise, essentially because the test does not involve
producing large superpositions of the non-observed subsystem. In section V we
show how to further enhance the protocol using pre-entangled initial
conditions. Finally, we outline an experimental realization with gravitational
entanglement generation between an atom interferometer and a mechanical
oscillator in section VI, before concluding with a discussion of implications
and loopholes in section VII.
## II Collapse and revival dynamics
To begin, we illustrate the basic idea of the collapse-and-revival dynamics
with an example. The setup is similar to electron spin echo envelope
modulation rowan1965electron ; dikanov1992electron and the cavity QED
experiments of Haroche _et al._ raimond2001manipulating . Consider an harmonic
oscillator $B$ coupled to a two-state system $A$ through the Hamiltonian
$H=\omega a^{\dagger}a+g(a+a^{\dagger})\sigma_{z}.$ (2)
In section VI we give an implementation of this Hamiltonian where the
oscillator $B$ is a mechanical resonator, the two-state system $A$ corresponds
to an atom located in one of two spatial locations, and $g\ll\omega$ is set by
the atom-oscillator gravitational interaction (1), so $g$ is proportional to
Newton’s constant $G_{N}$. The essential idea is to do an interferometry
measurement on the two-state system $A$ (the “control”) in the presence of
system $B$ (the “target”). The key is the dynamical response of the target
system $B$ to a superposition of $A$.
(a) (b) (c)
Figure 2: Equivalent circuit (left) and phase space (center) descriptions of
the experiment, and schematic interferometric data output (right). In the
circuit, the top line represents the atom and bottom line the resonator. The
large box represents joint evolution of the trapped atom and the resonator,
which can be decomposed into conditional displacement of the resonator,
followed by free evolution and an inverse displacement operator. This sequence
can be visualized in the phase space of the oscillator, where the solid and
dashed lines represent the two oscillator evolutions conditioned on the two
possible atomic locations. Interferometric measurement of the atom population
will show rapid fringes with frequency $\omega_{\rm DC}$ due to any stray DC
accelerations (e.g., due to electric fields, Earth’s gravity, or off-center
location of the resonator or atom), modulated by an overall reduction and then
increase due to the atom-resonator entanglement. Resonator motion over a full
period leads to nominal full recovery of the fringes.
To understand the entanglement dynamics generated by (2), it is useful to note
that the time evolution operator can be re-written
$U(t)=e^{-iHt}=D^{\dagger}\left(\sigma_{z}\lambda\right)e^{-i\omega
a^{\dagger}at}D\left(\sigma_{z}\lambda\right)$ (3)
up to an overall phase, where
$D(\alpha)\equiv\exp\left\\{\alpha a^{\dagger}-\alpha^{*}a\right\\}$ (4)
is the usual displacement operator.111To see this, note that
$D^{\dagger}(\alpha)a^{\dagger}aD(\alpha)=|a+\alpha|^{2}$ and expand the free
evolution operator $e^{-i\omega a^{\dagger}at}$ in the middle of (3). Here and
throughout, we will use the dimensionless quantity
$\lambda\equiv\frac{g}{\omega}.$ (5)
This is the length, measured in units of the zero-point length $x_{0}$, that
the oscillator equilibrium is displaced under the force from the atom. This
ratio will set the scale of all observables considered in this paper.
Observing the collapse-and-revival can be done with a typical interferometric
measurement. Consider starting the full system in its decoupled ground state
$\ket{0}_{A}\otimes\ket{0}_{B}$. The interferometry experiment then proceeds
by performing a Hadamard gate (or any other beamsplitter operation) on the
two-state system $A$, $\ket{0}\to(\ket{0}+\ket{1})/\sqrt{2}$, evolving the
joint system for some time $t$, performing the inverse Hadamard gate to
recombine the two-level system, and then measuring its population.
Mathematically, this proceeds as follows:
$\displaystyle\begin{split}\ket{\psi}&=\ket{0}_{A}\otimes\ket{0}_{B}\\\
&\xrightarrow{H}\frac{\ket{0}_{A}+\ket{1}_{A}}{\sqrt{2}}\otimes\ket{0}_{B}\\\
&\xrightarrow{U_{\rm
int}}\frac{\ket{0}_{A}\ket{\delta}_{B}+\ket{1}_{A}\ket{-\delta}_{B}}{\sqrt{2}}\\\
&\xrightarrow{H^{\dagger}}\ket{0}_{A}\frac{\ket{\delta}_{B}+\ket{-\delta}_{B}}{2}+\ket{1}_{A}\frac{\ket{\delta}_{B}-\ket{-\delta}_{B}}{2}.\end{split}$
(6)
Here, the conditionally-evolved states of the oscillator are simply coherent
states
$\ket{\pm\delta}_{B}=D\left(\pm\lambda(e^{-i\omega t}-1)\right)\ket{0}.$ (7)
If we now measure the two-state system $A$, we find for example that the
probability of being in the $\ket{0}$ state is
$P_{A}(0)=\frac{1}{2}+\frac{1}{2}{\rm
Re}\braket{\delta}{-\delta}_{B}=\frac{1}{2}\left(1+e^{-8\lambda^{2}\sin^{2}(\omega
t/2)}\right).$ (8)
We see that the interference term is reduced, with a period set by the
oscillator frequency $\omega$. In particular, at half-period we have a maximum
reduction of the phase contrast, and after a full period the contrast is
completely restored, as in Fig. 2.
Before moving on, we mention for later use an alternative calculation of the
same effect. Consider the Pauli lowering operator
$\sigma_{-}=(\sigma_{x}-i\sigma_{y})/2$ on the two-level system. The
expectation value $\braket{\sigma_{-}(t)}$ tracks the loss of phase contrast;
we will refer to the absolute value as the interferometric visibility
$V=|\braket{\sigma_{-}}|$. Using the time-evolution operator (3), we have
$\displaystyle\begin{split}\sigma_{-}(t)&=U^{\dagger}(t)\sigma_{-}U(t)\\\
&=D^{\dagger}(-\lambda)e^{i\omega
a^{\dagger}at}D(-\lambda)\sigma_{-}D^{\dagger}(\lambda)e^{-i\omega
a^{\dagger}at}D(\lambda)\\\ &=\sigma_{-}D(2\lambda(1-e^{i\omega
t})).\end{split}$ (9)
This is easy to show by working with explicit components in the $\sigma_{z}$
basis, where $\sigma_{-}=\ket{1}\bra{0}$. With an oscillator initially in the
ground state, this gives
$\braket{\sigma_{-}(\pi/\omega)}=\braket{\sigma_{-}(0)}e^{-8\lambda^{2}},\ \ \
\braket{\sigma_{-}(2\pi/\omega)}=\braket{\sigma_{-}(0)}.$ (10)
Here we see again the loss of phase contrast at half period followed by the
revival at a full period.
Up to this point, we have assumed that the oscillator was initialized in its
ground state $\ket{0}$. In a realistic implementation–particularly one where
the oscillator is a massive mechanical object–the oscillator will instead
start in a mixed state, such as a thermal state, due to its coupling to an
environment. Although one may be concerned that this would destroy the revival
of coherence in the atom, it turns out that not only does the revival persist,
but in fact the relative contrast between decoherence and revival is
_enhanced_ so long as the thermalization time scale remains very long. That
the revival persists is a consequence of the harmonic potential: after a full-
period, the state of the oscillator must return to its initial condition.
To see this, consider first the oscillator initialized to an arbitrary
coherent state $\ket{\alpha}$. Using (9), we have
$\displaystyle\begin{split}\braket{\alpha}{\sigma_{-}(t)}{\alpha}&=e^{-2\lambda[\alpha^{*}(1-e^{i\omega
t})-\alpha(1-e^{-i\omega t})]}\\\ &\times e^{-8\lambda^{2}\sin^{2}(\omega
t/2)}\braket{\sigma_{-}(0)}.\end{split}$ (11)
We see the complete revival after a full period, while at half period we now
pick up a phase involving the initial oscillator momentum
$p_{\alpha}=\alpha+\alpha^{*}$. To obtain the thermal-state result, one can
now average over the coherent states (i.e. use the oscillator density matrix
$\rho_{\rm th}=\int d^{2}\alpha
e^{-|\alpha|^{2}/\bar{n}}/(\pi\bar{n})\ket{\alpha}\bra{\alpha}$, with
$\bar{n}$ the thermal phonon occupancy). The result for the qubit visibility
is
$V_{\rm th}(t)=\exp\left[-8\lambda^{2}(2\bar{n}+1)\sin^{2}(\omega
t/2)\right].$ (12)
In particular, we have $V_{\rm th}(2\pi/\omega)=1$, showing a full revival of
the qubit coherence after a full oscillator period. On the other hand, at
half-period, we have $V_{\rm
th}(\pi/\omega)=\exp\left[-8\lambda^{2}\left(2\bar{n}+1\right)\right]$, an
enhancement to the loss of visibility by a factor of $\bar{n}$. Thus, starting
with a thermal state increases the contrast between the ‘dip’ of coherence
halfway through oscillation and the recovery at full oscillation. The
experiment is _easier_ with a hot oscillator.
## III Revival verifies entanglement generation
As this example clearly shows, entanglement generation between two systems $A$
and $B$ can cause periodic collapse and revival of $A$’s wavefunction. The
crucial question is then: does observation of this collapse and revival
_necessarily_ require entanglement generation between $A$ and $B$? Our central
result says that the answer is yes, under some particular assumptions. We
characterize this with a theorem:
###### Theorem 1
Let $L$ be a channel on $H_{A}\otimes H_{B}$, where $H_{A}$ is a two-state
system and $H_{B}$ is arbitrary. Assume that
1. (a)
The channel $L$ generates time evolution, in a manner consistent with time
translation invariance, thus obeying a semigroup composition law $L_{t\to
t^{\prime\prime}}=L_{t\to t^{\prime}}L_{t^{\prime}\to t^{\prime\prime}}$ for
all $t\leq t^{\prime}\leq t^{\prime\prime}$,
2. (b)
The two-level subsystem $H_{A}$ has its populations preserved under the time
evolution, $\sigma_{z}(t)=\sigma_{z}(0)$, and
3. (c)
$L$ is a separable channel rains1999rigorous : all of its Krauss operators are
simple products. In particular, this means that any initial separable (non-
entangled) state evolves to a separable state: $\rho(t)=L_{t}[\rho(0)]$ is
separable for all separable initial states $\rho(0)$.
Then the visibility $V(t)=|\braket{\sigma_{-}(t)}|$ is a monotonic function of
time.
Here, we have modeled the time evolution of the $A$-$B$ system as a quantum
channel $L$, a map on density matrices $\rho(t)=L_{t}[\rho(0)]$. For example,
within standard quantum mechanics, the unitary evolution of the universe
($A,B$ and their environment $C$, including the experimentalist) generates
such a channel for the reduced $A-B$ evolution. Suppose that we can
experimentally convince ourselves that time-translation invariance in the form
(a) and population condition (b) hold. Then the theorem says that if $L$
cannot generate entanglement (c), then the only possible evolution for the
qubit $A$ is to have its interferometric visibility decay monotonically. Thus
if we observe non-monotonic visibility like the oscillatory signal described
above, we can conclude that the channel must be capable of generating
entanglement.
We note that non-entangling channels still allow for non-trivial interactions.
For example, semiclassical gravity $G_{\mu\nu}=8\pi\braket{T_{\mu\nu}}$
(appropriately completed by a modified version of the Schrödinger equation) is
of this form kafri2015bounds . On the other hand, the graviton model will
produce an entangling channel.
We now give a proof of this theorem. By assumption (a), there exists a
generator $\mathcal{L}$ of $L_{t}$ of Lindblad form lindblad1976generators ;
gorini1976completely :
$\dot{\rho}=\mathcal{L}\rho=-i[H,\rho]-\sum_{j}\gamma_{j}\left[E_{j}^{\dagger}E_{j}\rho+\rho
E_{j}^{\dagger}E_{j}-2E_{j}\rho E_{j}^{\dagger}\right].$ (13)
These Lindblad operators $E_{j}$ are highly constrained by the separability
assumption, because they cannot be used to generate $A-B$ entanglement. To
make this precise, we write the channel in its Krauss representation
$L[\rho]=\sum_{j\geq 0}L_{j}\rho L_{j}^{\dagger}$. Expanding for small times
and comparing to (13), one finds that the Krauss operators $L_{j}$ take the
form, to lowest order in $dt$,
$L_{0}=1-iHdt+Kdt,\ \ L_{j}=E_{j}\sqrt{dt},\ \
K=-\frac{1}{2}\sum_{j>0}E_{j}^{\dagger}E_{j}.$ (14)
See, for example, chapter 3 of preskill1998lecture . Now we invoke the
separability criterion (c), which says that the Krauss operators for $j>0$
take the form of simple product operators, i.e. $E_{j}=A_{j}\otimes B_{j}$
rains1999rigorous . Furthermore, the separability of $L_{0}$ to order $dt$
means that $L_{0}=(1+A_{0}dt)\otimes(1+B_{0}dt)$ for some $A_{0},B_{0}$, and
this can only be satisfied if both $H$ and $E_{j}^{\dagger}E_{j}$ can be
written as sums of operators acting either on $\mathcal{H}_{A}$ _or_
$\mathcal{H}_{B}$. This in turn requires that for each $j>0$, either
$A_{j}^{\dagger}A_{j}=1_{A}$ or $B_{j}^{\dagger}B_{j}=1_{B}$. Finally, we
impose the requirement (b) that the atom populations are invariant. This means
that $\dot{\sigma}_{z}=0$. The only possible non-trivial interaction term
which satisfies these requirements is $E_{z}=\sigma_{z}\otimes B$, with $B$
any operator on $H_{B}$.
We are then left with the the very simple form of the Lindblad generator:
$\mathcal{L}\rho=-\gamma\left[B^{\dagger}B\rho+\rho
B^{\dagger}B-2B\sigma_{z}\rho\sigma_{z}B^{\dagger}\right]+\mathcal{L}_{A}+\mathcal{L}_{B}.$
(15)
Here $\mathcal{L}_{A(B)}$ are Lindblad operators (including Hamiltonians)
acting only on $\mathcal{H}_{A(B)}$, and $\mathcal{L}_{A}(\sigma_{z})=0$. With
this result for the channel’s structure, we can compute the time derivative of
the interferometric visibility $V(t)=|\braket{\sigma_{-}(t)}|$. Since
$[H,\sigma_{z}]=0$, the most general qubit Hamiltonian is a sum of
$\sigma_{z}$ and the identity. We thus have, in the Heisenberg picture,
$\displaystyle\begin{split}\Braket{\frac{d\sigma_{-}}{dt}}&=-i\braket{[H,\sigma_{-}]}+\gamma\left[\braket{E_{z}^{\dagger}\sigma_{-}E_{z}}-\frac{1}{2}\braket{\left\\{E_{z}^{\dagger}E_{z},\sigma_{-}\right\\}}\right]\\\
&=2(-i\omega_{0}-\gamma)\braket{\sigma_{-}},\end{split}$ (16)
where the oscillatory term is generated by the qubit Hamiltonian. Taking the
absolute value to compute the visibility $V=|\braket{\sigma_{-}}|$ removes the
oscillating phase and we have
$\frac{dV}{dt}=-2\gamma V,$ (17)
so it is monotonically decreasing, as we set out to prove.
## IV Effects of noise during evolution
The sensing protocol is subject to errors caused by random noise during the
time evolution. In a typical realization, the dominant sources of this
continuous noise will consist of thermal load on the oscillator and dephasing
in the atomic system (from, e.g., background fields and gas weakly measuring
the atomic position joos1985emergence ; gallis1990environmental ). These
sources of noise can be modeled by a Lindblad evolution of the form
$\dot{\rho}=-i[H,\rho]-\sum_{i}\frac{1}{2}\\{L_{i}^{\dagger}L_{i},\rho\\}-L_{i}\rho
L_{i}^{\dagger},$ (18)
where the error operators are
$L_{i}\in\\{\sqrt{\bar{n}\gamma_{m}}a^{\dagger},\sqrt{(\bar{n}+1)\gamma_{m}}a,\sqrt{\gamma_{a}}\sigma_{z}\\}$.
The decay rates of the oscillator and atom are $\gamma_{m},\gamma_{a}$,
respectively, and $\bar{n}$ is the thermal phonon occupancy. This description
should be accurate for times similar to or shorter than the damping time
$1/\gamma_{m}$, and assuming only small changes over time in the mechanical
frequency.
It is possible to analytically solve for the atomic visibility
$\eqref{visibility}$ in the presence of this noise, using an explicit Ohmic
heating model where the bath is taken to be an infinite set of bosonic modes
linearly coupled to the mechanical system. The same displacement-operator
picture used in (3) generalizes to this linear bath (see appendix B). One
finds that the visibility at half and full-period evolution is given by
$\displaystyle\begin{split}V(\pi/\omega)&=\exp[-\pi\gamma_{a}/\omega]\exp[-8\lambda^{2}(2\bar{n}+1)]\\\
V(2\pi/\omega)&=\exp[-2\pi\gamma_{a}/\omega]\exp[-8\lambda^{2}(2\bar{n}+1)/Q].\end{split}$
(19)
Here we have assumed the mechanical damping factor $Q=\omega/\gamma_{m}\gg 1$.
This recovers the previous result for the visibility (12), up to an overall
exponential damping from the atomic dephasing and small correction from
mechanical heating. Neglecting atomic dephasing, the visibility at half period
is exactly the same as (12), while at full period, for $Q\gg
1\gg\bar{n}\lambda^{2}$, we have $V(2\pi/\omega)\approx 1$, i.e. we have full
recovery up to a correction at order $1/Q$. Thus, with a sufficiently high-$Q$
oscillator, and with atomic coherence times longer than the mechanical period
$\gamma_{a}\lesssim\omega$, damping does not pose a substantial barrier to the
experiment.
Before moving on, we consider the effects of decoherence from another
inevitable source: blackbody radiation of the oscillator. Here we are
discussing position superpositions of the oscillator at distances of about
$\lambda x_{0}$. With the sorts of experimental parameters we suggest later,
this will be a length many orders of magnitude smaller than a typical
blackbody photon wavelength (or ambient gas molecule’s de Broglie wavelength).
Thus these interactions will be incapable of efficiently decohering the
oscillator, because they are too long-wavelength to efficiently measure the
oscillator’s position gallis1990environmental .
## V Protocol linear in the weak coupling
Figure 3: Left: Experimental realization of the atomic system as a lattice
interferometer. The lines marked “x” denote populations that do not interfere.
Right: Some example implementations with one or more mechanical masses
connected rigidly. Small black dots represent the atom. In each case, the
mechanical system is restricted to oscillate along the $z$-axis. More masses
enable a stronger gravitational coupling. A natural limiting case would be to
use a toroidal mass. In the example with a single sphere, we have
$R=\sqrt{L^{2}+(\ell/2)^{2}}$ and $\kappa=1$.
Our basic observable (8) is quadratic in the ratio $\lambda=g/\omega$, which
for a weak coupling is a small dimensionless number. Here we suggest a
“boosted” method in which linear sensitivity can be achieved by first
preparing an _entangled_ state of the atom-oscillator system (as demonstrated,
for example, in karg2020light ; thomas2021entanglement ).
Let $\lambda^{\prime}=g^{\prime}/\omega$, where $g$ is the coupling of
interest (e.g., gravity) and $g^{\prime}$ is some other coupling. Consider
performing a $\pi$ gate with the coupling $V_{\rm
int}=(g+g^{\prime})\sigma_{z}x$. This will produce an initial entanglement set
by displacement operators $D(\pm(\lambda+\lambda^{\prime}))$, as in equation
(9). Turning off the non-gravitational $g^{\prime}$ coupling then leads to
only a partial revival of the atomic signal at later times $t>\pi/\omega$.
This leads to the visibility, for $t>\pi/\omega$,
$V_{\rm b}(t)=\exp\left[-8(2\bar{n}+1)\left(\lambda^{\prime
2}+2\lambda\lambda^{\prime}\sin^{2}\frac{\omega
t}{2}+\lambda^{2}\sin^{2}\frac{\omega t}{2}\right)\right].$ (20)
A detailed calculation is given in appendix (C). For times $0<t<\pi/\omega$,
the visibility is given by the previous result (12) but with
$\lambda\to\lambda+\lambda^{\prime}$.
The observable we are interested in is the difference in visibility at half-
period and full-period:
$\displaystyle\begin{split}\Delta V_{\rm
b}&=V_{b}(2\pi/\omega)-V_{b}(\pi/\omega)\\\
&\approx\exp[-8(2\bar{n}+1)\lambda^{\prime
2}](1-16(2\bar{n}+1)\lambda^{\prime}\lambda+O(\lambda^{2})),\end{split}$ (21)
assuming $\lambda\ll\lambda^{\prime}$. We see again that using an initially
“hot” resonator increases the relative visibility. However, here the
observable is _linear_ in the weak gravitational coupling $\lambda$. We note
that if $\bar{n}$ or $\lambda^{\prime}$ are too large, the signal will be
destroyed by the overall prefactor $e^{-8(2\bar{n}+1)\lambda^{\prime 2}}$. The
optimal solution is to tune the non-gravitational coupling to satisfy
$\lambda^{\prime}_{\rm opt}=1/\sqrt{8(2\bar{n}+1)}$, in which case the
prefactor is order one, and the relative visibility is given roughly by
$\Delta V_{b}\approx\sqrt{8(2\bar{n}+1)}\lambda$. Use of this boosted protocol
substantially improves the viability of an experiment with a weak coupling
$g$. We note that this protocol does not violate our assumptions about time-
translation invariance in Theorem 1: once the extra $g^{\prime}$ coupling is
turned off, the entire system proceeds in a time-independent fashion.
## VI Experimental implementation with atom interferometry
Figure 4: Logarithm $\log_{10}$ of the visibility change $\Delta V$ as
function of the logarithms of the hold time $\tau$ in seconds and the
temperature $T$ in Kelvin for the unboosted scheme (left) and the boosted
scheme (right). The plots assume $\ell=1\,$mm, $\rho=20$g/cm3, $m=m_{\rm Cs}$.
We now show how to apply our sensing protocol to a test of quantum gravity.
The idea is to realize the qubit in the Hamiltonian (2) as an optical-lattice
atom interferometer xu2019probing with a hold-time $\tau$ and splitting
$\ell$ between the matter wave packets. The majority of the interferometer
time sees the atoms trapped in one of two different potential wells created by
the lattice. The atom position thus becomes a two-state system with
$\sigma_{z}$-eigenvalues corresponding to the two locations. The mechanical
oscillator has a mass $M$ and fundamental frequency $\omega$. Expanding the
Newtonian atom-oscillator potential (1), we then have the total Hamiltonian
$H=\omega a^{\dagger}a-g\sigma_{z}(a+a^{\dagger}).$ (22)
Here, $a,a^{\dagger}$ are oscillator operators, so the second term represents
the position-position coupling. The coupling strength is
$g=\kappa\frac{G_{N}mM\ell x_{0}}{\hbar R^{3}}$ (23)
where $x_{0}=\sqrt{\hbar/2M\omega}$ is the ground state oscillator
uncertainty, $R$ parametrizes the distance between the oscillator and atom,
and $\kappa$ is a dimensionless number of order one which depends on the
specific oscillator mass geometry (see Fig. 3).
Technical challenge | Examples | Possible strategies
---|---|---
Non-gravitational interactions | Van der Waals, stray fields, scattered laser light | Superconducting shielding, place atoms in waveguide Xin2018fiber
Mean field shift | Parasitic atom-atom interactions leading to inhomogeneous dephasing Jannin2015Meanfield | Spin-echo techniques Laudat (see also appendix E), fermionic atoms (e.g., Yb-171 or 173) McAlpine ; Niederriter
Exponential decay of signal | Atomic dephasing | Interleaved differential measurement, e.g. by toggling the mass between near and far positions Jaffe2017subgrav
Deviations from harmonicity | Time-dependent oscillator frequency, anharmonic perturbations | Keep effective temperature below nonlinear thresholds; change materials, mounting, or frequency
Table 1: Some systematic effects and other perturbations expected in a realistic implementation. Loophole or pathology | Typical sources | Problematic behavior allowed | Possible solutions
---|---|---|---
Non-gravitational interactions between atom and oscillator | Casimir/van der Waals interactions | Can generate entanglement (reproduce the full desired signal), can generate extra noise | Vary parameters (masses and distance) to check proper scaling with $V=G_{N}m_{1}m_{2}/r$ law
Stationarity assumption on bath (and/or experimentalist) violated | Explicit time-dependence introduced by experimentalist (e.g. spin-echo protocol); low-frequency noise (e.g., gravity gradients, seismic noise) | Violates assumption of theorem in Sec. III. In principle, could mimic collapse and revival | Adjust theorem to allow for bath relaxation timescale; experimentally verify Markovian nature of oscillator noise
Non-locality | Time of interaction for experiment is much longer than light-crossing time $T_{\rm int}\gg T_{com}$ | Allows for non-local, hidden variable model explaining the entanglement (same as Bell test) | Long baseline version?
Table 2: Some loopholes and pathologies in our proposed test.
The information sensing protocol requires generation of an initial state
$\ket{0}+\ket{1}$. This can be generated, e.g., by a pair of Raman pulses
separated by a free evolution time xu2019probing , by spin-dependent kicks
Jaffe2018SDK , by optical lattice techniques Pagel2020lattice , or by rapidly
splitting a single-well potential to the double-well. Measuring in the
$\sigma_{z}$ basis at the end of the protocol corresponds to closing the atom
interferometer and counting the atoms in the two output ports. To implement
the “boosted” protocol of section V, we can use a number of non-gravitational
interactions to generate the initial entanglement. For example, a hyperfine or
Rydberg atomic state could be magnetically or optically coupled to the
oscillator. Entanglement of this type has been recently demonstrated
experimentally karg2020light ; thomas2021entanglement .
Let us consider how we can obtain a visibility change that is large enough to
be measured. In order to observe at least one full cycle of decay and revival,
we choose $\omega=2\pi/\tau$ where $\tau$ is the atom hold-time. In this case,
the visibility change is given by
$\displaystyle\begin{split}\Delta V&=\frac{\pi
G_{N}^{2}m^{2}\rho}{3\sqrt{2}\ell\omega^{3}\hbar}\left(8+\bar{n}\right)\xrightarrow[k_{\rm
B}T/\omega\to\infty]{}\frac{\pi}{3\sqrt{2}}K^{2},\\\ \Delta V_{\rm
b}&=2^{1/4}G_{N}m\sqrt{\frac{\pi\rho}{3\ell\omega^{3}\hbar}(8+\bar{n})}\xrightarrow[k_{\rm
B}T/\omega\to\infty]{}\frac{2^{1/4}}{\sqrt{3}}K\end{split}$ (24)
in the unboosted and boosted scheme respectively, where
$\displaystyle\begin{split}K^{2}&=\frac{G_{N}^{2}m^{2}\rho
k_{B}T}{\ell\omega^{4}\hbar^{2}}\\\ &\approx 1.04\times
10^{-14}\left(\frac{T}{300\,{\rm K}}\right)\left(\frac{\ell}{1\,{\rm
mm}}\right)^{-1}\left(\frac{\tau}{10\,{\rm s}}\right)^{4}.\end{split}$ (25)
Here we took a solid density $\rho=20~{}{\rm g/cm}^{3},$ cesium atoms
$m=m_{\rm Cs}=133~{}{\rm amu}$, used the four-sphere configuration (Fig. 3)
for definiteness, and maximized the coupling $g$ for a given splitting $\ell$
by choosing a sphere radius of $R_{s}=\ell/(\sqrt{8})$. Longer atomic
interrogation times $\tau$ are preferable. This would require a
correspondingly low-frequency oscillator, e.g. a mHz-scale torsional pendulum.
While 20 s have been experimentally realized xu2019probing , 100 s may be a
reasonable expectation for the future. Using a small matter-wave splitting
$\ell$ is desirable, but subject to mechanical constraints. Choosing, e.g.,
$\ell=1\,$mm, $L=1/\sqrt{2}\,$ mm and $R_{s}=0.35\,$mm would leave about
$0.15\,$ mm free space between the spheres. For $\tau=100\,$s and $T=300$ K we
obtain $\Delta V\sim 10^{-10}$; but for the boosted scheme, it will be as
large as $\Delta V=7\times 10^{-6}$ (see Fig. 4). At the standard quantum
limit, this can be detected with $5-\sigma$ significance by running the
experiment with $\sim 5\times 10^{11}\,$ atoms (see appendix D for details on
noise scaling with many atoms). Assuming that the experiment has $10^{7}$
atoms per run, and each run takes 2 minutes, this will be possible in two
months total run time.
Remarkably, this suggests that the experiment may be feasible in the near
future. A number of systematic effects and technical issues will need to be
understood. We postpone detailed discussion to future work, but flag some
likely issues and ways to handle them in Table 1.
## VII Implications, loopholes, and conclusions
Our interactive information sensing protocol is a novel strategy for
verification of dynamical entanglement generation. While a standard Bell-type
test requires measurements on both parts of a bipartite system, our protocol
can verify entanglement generation with only single-body measurements.
Crucially, the test verifies the ability of an interaction channel to generate
entanglement, without needing to directly verify the entanglement of the final
state. However, it is important to note that this test is subject to
loopholes. Some are analogous to those in standard Bell tests and others are
particular to our proposal. We suggest a few of these in Table 2.
In our view, the most important loophole stems from our time-translation
invariance assumption, which we used to write the atom-oscillator dynamics in
Lindblad form (13). Non-Markovian time dependence introduced by an
experimentalist or Maxwell’s demon could, in principle, reproduce the observed
collapse and revival dynamics. One way to improve the situation would be to
reformulate the theorem to include some level of non-Markovianity, for example
a bath relaxation time scale. A more robust option would be to prove
experimentally that it is simply the Markovian thermalizing channel acting on
the mechanical system. Methods for this include precision quantum thermometry
purdy2017quantum , which can support the hypothesis of detailed balance. In
any case, extending the results here beyond the strictly Markovian assumption
will be a crucial next step.
The central technical advances suggested here are the interactive sensing
protocol and the use of atoms as a sensor. The key advantage of the periodic
collapse-and-revival protocol is that it enables a huge enhancement with a
thermal state of the mechanical system; understanding if this can be extended
beyond the specific context here would be very interesting. While using
trapped atoms is perhaps counter-intuitive since it decreases the strength of
the signal (the Newton potential), we emphasize that the extremely long
coherence lifetime and ability to generate spatially well-separated
superpositions of the atoms lead to similar parametric scaling of the overall
signal strength.
We have shown how the interactive sensing protocol can be used to test the
ability of the gravitational field to communicate quantum information. If the
answer is yes, this would constitute the first direct evidence that the
gravitational field itself is a quantum mechanical degree of freedom
carney2019tabletop ; belenchia2018quantum ; Christodoulou:2018cmk ;
Marshman:2019sne ; Galley:2020qsf . On the contrary, if the answer is
negative, the existence of the graviton is ruled out carney2019tabletop . The
simple estimates of section VI suggest that this experiment is feasible with
realistic devices, even in the presence of noise. We will present a more
detailed proposal and analysis of systematic effects in a future paper.
## Acknowledgements
We thank Thomas Guff, Jack Harris, John Kitching, and Jess Riedel for
discussions. H.M. has been supported in part by Jet Propulsion Laboratory
under grant number 1652036, the Office of Naval Research under grant numbers
GRANT12980618 and N00014-20-1-2656, as well as the National Science Foundation
under grant number 1708160.
## Appendix A Newtonian entanglement from graviton exchange
For completeness, we review here some standard arguments about the
perturbative quantization of gravity and its relation to entanglement
generation via the Newton potential (1). Our goal is to explain the standard
logic by which one treats small fluctuations of the metric as a quantum field
and uses this to make predictions in non-relativistic systems. We do not mean
to say that this derivation somehow proves that this is the correct model of
low-energy quantum gravity—on the contrary, determining if this is the correct
set of predictions is a central objective of the experiment proposed in this
paper.
By far the most common and efficient method to compare a field theoretical
description to the non-relativistic setting relevant to these experiments is
to do a “matching” calculation. For example, one can compute scattering
amplitudes in the field theory, compare these to the same amplitude computed
in a potential scattering model, and thus obtain the effective non-
relativistic potential. Since the scattering states form a complete basis for
the Hilbert space (other than bound states), if these two calculations agree
for all scattering states, we can conclude that the two descriptions are
equivalent quantum-mechanically in the regime in which the calculations match.
Figure 5: Feynman diagrams for single photon and graviton exchange,
respectively.
For example, consider electrodynamics: photons $A_{\mu}$ coupled to a Dirac
fermion $\psi$ (e.g., the electron, with charge $e$). See section 4.8 of
peskin2018introduction for a textbook treatment of what follows. The
Lagrangian is
$\mathcal{L}=\frac{1}{4}F_{\mu\nu}F^{\mu\nu}+\bar{\psi}\gamma^{\mu}(\partial_{\mu}-ieA_{\mu})\psi$
(26)
with $F_{\mu\nu}=\partial_{\mu}A_{\nu}-\partial_{\nu}A_{\mu}$ the
electromagnetic field tensor and $\gamma^{\mu}$ the usual Dirac matrices. With
this Lagrangian we can perturbatively compute the elastic $e^{-}e^{-}\to
e^{-}e^{-}$ scattering amplitude to lowest order in the charge using standard
methods. See the Feynman diagram of figure 5. One finds222There is also a
$u$-channel diagram coming from the indistinguishability of the two electrons,
which will also appear in the potential scattering computation, but this is
not important for our argument here so we drop it for simplicity.
$\braket{\mbox{\boldmath$p$}_{1}^{\prime}\mbox{\boldmath$p$}_{2}^{\prime}}{T}{\mbox{\boldmath$p$}_{1}\mbox{\boldmath$p$}_{2}}=4\pi
e^{2}\frac{N_{\rm EM}}{-t-i\epsilon}$ (27)
where $t=-(p_{1}^{\prime}-p_{1})^{2}$ is the Lorentz-invariant four-momentum
transfer, and the numerator is
$N_{\rm
EM}=\bar{u}(\mbox{\boldmath$p$}_{1}^{\prime})\gamma^{\mu}u(\mbox{\boldmath$p$}_{1})\bar{u}(\mbox{\boldmath$p$}_{2}^{\prime})\gamma_{\mu}u(\mbox{\boldmath$p$}_{2}).$
(28)
Here $u,\bar{u}$ are Dirac spinors. In the non-relativistic limit where the
rest masses dominate over the spatial momenta, the numerator reduces to
$N_{\rm EM}\to 1$ and
$t\to(\mbox{\boldmath$p$}_{1}^{\prime}-\mbox{\boldmath$p$}_{1})^{2}$. In the
center-of-mass frame, this means that the amplitude for the relative momentum
$\mbox{\boldmath$p$}=(\mbox{\boldmath$p$}_{1}-\mbox{\boldmath$p$}_{2})/2$ to
transition to $\mbox{\boldmath$p$}^{\prime}$ is given by
$\braket{\mbox{\boldmath$p$}^{\prime}}{T}{\mbox{\boldmath$p$}}=\frac{4\pi
e^{2}}{(\mbox{\boldmath$p$}^{\prime}-\mbox{\boldmath$p$})^{2}},$ (29)
where we used conservation of the total momentum to write
$\mbox{\boldmath$p$}_{1}^{\prime}-\mbox{\boldmath$p$}_{1}=\mbox{\boldmath$p$}^{\prime}-\mbox{\boldmath$p$}$.
We can then compare this to the amplitude computed non-relativistically with a
Hamiltonian description
$H=H_{1}+H_{2}+V(r)$ (30)
including a central potential
$V(r)=V(|\mbox{\boldmath$x$}_{1}-\mbox{\boldmath$x$}_{2}|)$. In the center of
mass frame, the first Born approximation gives
$\braket{\mbox{\boldmath$p$}^{\prime}}{T}{\mbox{\boldmath$p$}}=\tilde{V}(\mbox{\boldmath$p$}^{\prime}-\mbox{\boldmath$p$}),$
(31)
where $\tilde{V}$ is the Fourier transform of the potential. Comparing these
expressions, we see that the effective potential is
$\tilde{V}(\mbox{\boldmath$q$})=\frac{4\pi e^{2}}{\mbox{\boldmath$q$}^{2}},$
(32)
or in position space,
$V(\mbox{\boldmath$x$})=\int\frac{d^{3}\mbox{\boldmath$q$}}{(2\pi)^{3}}e^{i\mbox{\boldmath$q$}\cdot\mbox{\boldmath$x$}}\tilde{V}(\mbox{\boldmath$q$})=\frac{e^{2}}{|\mbox{\boldmath$x$}|},$
(33)
recovering the usual Coulomb force.
We emphasize that this line of argument is at the level of transition
_amplitudes_ $\braket{\psi_{f}}{U(t_{f},t_{i})}{\psi_{i}}$, where $U$ means
time evolution and we take $t_{i}\to-\infty,t_{f}\to+\infty$. Thus, the
potential (33) is the potential operator in the full meaning of the term—it is
a two-body operator and can entangle particles, and so forth. Indeed,
entanglement generated by this Coulomb potential is exactly what underlies,
for example, quantum information processing with chains of trapped ions in
Coulomb crystals.
The same type of matching calculation can be performed in gravity. The
additional complication is that we need to work perturbatively around a
background spacetime, and only quantize the perturbations. For an experiment
well-localized in spacetime like ours, we can expand around a locally flat
spacetime metric $\eta_{\mu\nu}$, by the equivalence principle.333A more
accurate statement would be that we should expand around a Schwarzschild
solution, with Schwarzschild mass given by the mass of Earth. This term would
generate a constant background external potential and is inessential to the
core quantum mechanics of the following argument. We write
$g_{\mu\nu}=\eta_{\mu\nu}+\frac{h_{\mu\nu}}{m_{\rm Pl}}$ (34)
where we scale out a factor of the Planck mass to give $h_{\mu\nu}$ dimensions
of mass (i.e. the same dimensions as a canonical bosonic field in four
dimensions). In this expansion, the Einstein-Hilbert Lagrangian becomes
$\mathcal{L}=\frac{1}{2}\partial_{\alpha}h_{\mu\nu}\partial^{\alpha}h^{\mu\nu}+\cdots+\frac{1}{m_{\rm
Pl}}h_{\mu\nu}T^{\mu\nu},$ (35)
where the dots represent other similar second-derivative kinetic terms for the
perturbations $h_{\mu\nu}$ and $T_{\mu\nu}$ is the matter stress tensor. These
metric perturbations $h_{\mu\nu}$ can be quantized through the exact same
procedure as the electromechanical potential $A_{\mu}$ Feynman:1963ax ;
t1974one ; deser1974one ; Veltman:1975vx . There is a key difference—namely
that the interaction is non-renormalizable—which means that we have an
effective quantum field theory kadanoff1966scaling ; wilson1971renormalization
; weinberg1979phenomenological ; donoghue1994general ; burgess2004quantum
which can only make reliable predictions at energy densities well below the
Planck scale.444To be more precise: the non-renormalizable nature of the
interaction means that we have to include all possible generally-covariant
terms, in particular curvature-curvature couplings, in the action $S=m_{\rm
Pl}^{2}\int d^{4}x\sqrt{-g}[R+m_{\rm
Pl}^{-2}(c_{1}R^{2}+c_{2}R_{\mu\nu}R^{\mu\nu})+m_{\rm
Pl}^{-4}c_{3}R^{3}+\cdots]$. Since this is an infinite series of terms each
with an unknown constant coefficient $c_{i}$, the model becomes non-predictive
once these $R^{n\geq 2}$ terms become important. In the regime of these
experiments $R\ll m_{\rm Pl}^{2}$, so all of these terms are extraordinarily
small and the dynamics is determined entirely by the first term $R$, i.e., the
usual Einstein-Hilbert action. However, we are well within this limit in the
kind of experiment envisaged here, as discussed in the introduction.
Proceeding accordingly, the scattering of a pair of masses (here modelled as
single-particle excitations of a massive spinless field) via gravitons is
given by donoghue1994general ; burgess2004quantum
$\braket{\mbox{\boldmath$p$}_{1}^{\prime}\mbox{\boldmath$p$}_{2}^{\prime}}{T}{\mbox{\boldmath$p$}_{1}\mbox{\boldmath$p$}_{2}}=\frac{4\pi}{m_{\rm
Pl}^{2}}\frac{N_{\rm grav}}{-t-i\epsilon}.$ (36)
The numerator is more complicated due to the tensorial nature of the
interaction,
$\displaystyle\begin{split}N_{\rm grav}&=2(p_{1}\cdot
p_{1}^{\prime})(p_{2}\cdot p_{2}^{\prime})+2(p_{1}\cdot
p_{2}^{\prime})(p_{1}^{\prime}\cdot p_{2})\\\ &\ \ \ +8(p_{1}\cdot
p_{1}^{\prime}+m^{2})(p_{2}\cdot p_{2}^{\prime}+m^{2})\end{split}$ (37)
but reduces in the non-relativistic limit to the simple value $N_{\rm grav}\to
m^{2}$. Recognizing that $m_{\rm Pl}^{2}=1/G_{N}$ in terms of the Newton
constant, we can compare this again to the Born approximation (31), and
determine the effective potential
$\tilde{V}(\mbox{\boldmath$q$})=\frac{4\pi
G_{N}m^{2}}{\mbox{\boldmath$q$}^{2}}$ (38)
which again is just $V(r)=G_{N}m^{2}/r$ in real space. In this way, we see
that “graviton exchange” leads to the Newton potential operator (1) in the
non-relativistic limit.
Finally, we note that can one directly obtain a Hamiltonian operator for the
field theory and read off the non-relativistic potential directly, without
resorting to scattering or other matching calculations. In contrast to the
gauge-invariant scattering amplitude approach, this is complicated by the
gauge symmetries of the model (in both the electrodynamics and gravity cases).
To see how this works, consider the electrodynamical Lagrangian (26). To
perform the transformation from the Lagrangian to Hamiltonian we have to fix a
gauge, say Coulomb gauge $\partial_{i}A^{i}=0$. This gauge leads to a second-
class Dirac constraint $\partial_{i}F^{i0}=-J^{0}$, so that
$\nabla^{2}A^{0}=-J^{0}$, i.e., the $A^{0}$ part of the potential is non-
dynamical and simply fixed by the current
$A^{0}(\mbox{\boldmath$x$},t)=-\int
d^{3}\mbox{\boldmath$y$}\frac{J^{0}(\mbox{\boldmath$y$},t)}{|\mbox{\boldmath$x$}-\mbox{\boldmath$y$}|}.$
(39)
Performing the Legendre transformation to obtain the Hamiltonian, one then
finds a coupling
$\displaystyle\begin{split}H_{\rm Coul}&=-\int
d^{3}\mbox{\boldmath$x$}A^{0}(\mbox{\boldmath$x$})J^{0}(\mbox{\boldmath$x$})\\\
&=\int
d^{3}\mbox{\boldmath$x$}d^{3}\mbox{\boldmath$y$}\frac{J^{0}(\mbox{\boldmath$x$})J^{0}(\mbox{\boldmath$y$})}{|\mbox{\boldmath$x$}-\mbox{\boldmath$y$}|},\end{split}$
(40)
which is just the usual non-relativistic Coulomb interaction, since $J^{0}$ is
the charge density. Again, everything here is at the level of operators. In
the case of gravity, the exactly analogous calculation can be performed and
one finds an instantaneous Newton interaction $H=m_{\rm Pl}^{-2}\int
d^{3}\mbox{\boldmath$x$}d^{3}\mbox{\boldmath$y$}T^{00}(\mbox{\boldmath$x$})T^{00}(\mbox{\boldmath$y$})/{|\mbox{\boldmath$x$}-\mbox{\boldmath$y$}|}$.
See chapter 8 of weinberg1995quantum for a thorough treatment of this kind of
Hamiltonian approach in the case of electrodynamics, and e.g.
anastopoulos2013master for the calculation in perturbative gravity.
The fact that the field component responsible for the Coulomb/Newton
interaction is non-dynamical (e.g., $A^{0}$ in the above example) has led some
authors to argue that observing Newtonian entanglement would tell us nothing
about the quantization of the “physical” (i.e., dynamical) degrees of freedom
of the gravitational field; for a prototypical expression of this view, see
Anastopoulos:2018drh . There are, however, strong arguments against this
belenchia2018quantum ; Christodoulou:2018cmk ; Marshman:2019sne ;
Galley:2020qsf , which essentially say that there is no consistent way to have
both an entangling Newton interaction and non-quantized metric fluctuations.
We anticipate substantial further debate on this topic, and will present a
detailed discussion in a separate paper carneyprep .
## Appendix B Detailed calculation of oscillator noise
To get a quantitative estimate of the effects of thermal loading on the
oscillator, let us assume that we can safely neglect atomic dephasing
$\gamma_{a}t\ll 1$ for the experimental timescale of interest. To develop an
exact result, we will use an input-operator method, in which we include an
explicit heat bath for the oscillator. The Lindblad system (18) can be derived
through this method by tracing out the oscillator bath. See
clerk2010introduction for a lucid review of this technique.
Let $H_{0}$ denote the Hamiltonian for the oscillator _and its bath_. The
total time evolution operator is then
$U(t)=e^{-i(H_{0}+g\sigma_{z}x)t}=e^{-iH_{0}t}\hat{T}e^{-ig\sigma_{z}\int_{0}^{t}x_{I}(t^{\prime})dt^{\prime}}.$
(41)
Here $x_{I}(t)$ is the oscillator position operator in the interaction picture
and $\hat{T}$ is the time-ordering operator. For the case of a linear bath,
such as that assumed in quantum optics or in quantum Brownian motion, we can
explicitly find $x_{I}$. Writing $x=(a+a^{\dagger})/\sqrt{2}$, we have
$\displaystyle\begin{split}a_{I}(t)&=\exp[-i(\omega+\gamma_{m}/2)t]a(0)\\\
&+\sqrt{\gamma_{m}}\int_{0}^{t}\exp[-i(\omega+\gamma_{m}/2)(t-t^{\prime})]a_{\rm
in}(t^{\prime})dt^{\prime}\end{split}$ (42)
where $a_{\rm in}(t)$ is the vacuum noise fluctuation operator, satisfying
$[a_{\rm in}(t),a_{\rm in}^{\dagger}(t^{\prime})]=\delta(t-t^{\prime})$. Using
the linearity of this expression and the Baker-Campbell-Hausdorf relation, we
then have that
$\displaystyle\begin{split}&\hat{T}\exp\left(-ig\sigma_{z}\int_{0}^{\tau}x_{I}(t)dt\right)\\\
&=\exp\left(-ig\sigma_{z}\int_{0}^{\tau}x_{I}(t)dt\right)\exp(-ig^{2}C(t))\end{split}$
(43)
where $C(t)$ is a real, time-dependent number, arising from the non-commuting
elements of $x_{I}(t)$.
Having dispensed with the time-ordering, we can now explicitly perform the
time integral (including a change of integration order in the $a_{\rm in}$
term). Dropping the $e^{-ig^{2}C(t)}$ phase, which will cancel out of our
observable, we find the time evolution reduces to a simple product of
displacement operators, one for the oscillator and one for each mode $a_{\rm
in}(t^{\prime})$ for $0\leq t^{\prime}\leq t$, that is
$U(t)=e^{-iH_{0}t}D_{a}[\sigma_{z}\alpha(t)]\prod_{0\leq t^{\prime}\leq
t}D_{a_{in}(t^{\prime})}[\sigma_{z}\alpha_{\rm in}(t^{\prime})]$ (44)
where
$\displaystyle\begin{split}\alpha(t)&=\frac{ig}{i\omega-\gamma_{m}/2}(1-e^{(i\omega-\gamma_{m}/2)t}),\\\
\alpha_{\rm
in}(t^{\prime})&=\frac{ig}{i\omega-\gamma_{m}/2}(1-e^{(i\omega-\gamma_{m}/2)(t-t^{\prime})}).\end{split}$
(45)
Finally, we can evaluate our visibility
$\sigma_{-}(t)=U^{\dagger}(t)\sigma_{-}U(t)$, assuming an initial thermal
state for the oscillator and each bath mode and the $\ket{+}$ state for the
atom. Using the same results for coherent states as above, one finds
$\displaystyle\begin{split}\braket{\sigma_{-}}&=\braket{D_{a}[2\alpha(t)]}\prod_{t^{\prime}}\braket{D_{a_{\rm
in}(t^{\prime})}[\alpha_{\rm in}(t^{\prime})]}\\\
&=\exp\left[-8\lambda^{2}(2\bar{n}+1)f(t)\right]\end{split}$ (46)
with
$\displaystyle\begin{split}f(t)&=\frac{\omega^{2}/4}{\omega^{2}+\gamma_{m}^{2}/4}\Big{(}2-2\cos(\omega
t)e^{-\gamma_{m}t/2}+\gamma_{m}t\\\ &-\frac{8\gamma_{m}}{\omega}\sin(\omega
t)e^{-\gamma_{m}t/2}+O(1/Q^{2})\Big{)}\end{split}$ (47)
where $Q=\omega/\gamma_{m}$ is assumed much larger than one. In particular, at
full and half-period this gives
$\displaystyle\begin{split}V(\pi/\omega)&=\exp[-8\lambda^{2}(2\bar{n}+1)],\\\
V(2\pi/\omega)&=\exp[-8\lambda^{2}(2\bar{n}+1)/Q].\end{split}$ (48)
Here we have assumed the mechanical damping factor $Q=\omega/\gamma_{m}\gg 1$.
Re-inserting the exponential damping factor for atomic dephasing then
reproduces the results in (19).
In Fig. 6, we compare this analytical model with a numerical simulation of the
Lindblad equation (13), showing excellent agreement.
Figure 6: Examples of the signal of interest, the phase contrast
$V=|\braket{\sigma_{-}(t)}|$, compared with its initial value $V(0)=1/2$.
Left: Direct simulation of the Lindblad evolution (18) in blue, and we see
good agreement with our analytic solution including noise (46) in orange,
dashed. Normalizing all units to the oscillator frequency $\omega=1$, here we
use values $g=10^{-2}$ for the gravitational coupling, $\gamma_{m}=5\times
10^{-3}$ for the mechanical damping, and $T=2$ for the temperature. (Numerical
simulation with a much higher $T\gg\omega$ as discussed in the paper is
infeasible due to restrictions on the oscillator Hilbert space dimension).
Same parameters as left figure, but with an initial $\pi$-pulse using a non-
gravitational coupling $g^{\prime}=10^{-1}$. The difference between the first
collapse and revival is now much larger than in the un-boosted protocol, as
predicted in (21).
## Appendix C Detailed calculation of the boosted protocol
Here we give the full computation of the visibility in our entanglement-
enhanced, “boosted” protocol of Section V. The total evolution is a product of
two unitaries, one for the first half-period under the coupling $g+g^{\prime}$
and the second under only $g$. We write these as
$\displaystyle\begin{split}U_{g+g^{\prime}}&=D^{\dagger}((\lambda+\lambda^{\prime})\sigma_{z})e^{-i\omega
n}D((\lambda+\lambda^{\prime})\sigma_{z}),\\\
U_{g}(t)&=D^{\dagger}(\lambda\sigma_{z})e^{-i\omega
n(t-\pi/\omega)}D(\lambda\sigma_{z}).\end{split}$ (49)
With this notation, the visibility of the atom, given some initial coherent
state $\ket{\alpha}$ for the oscillator, is given by (defining
$\tilde{\lambda}=\lambda+\lambda^{\prime}$ for brevity)
$\displaystyle\begin{split}V_{{\rm
b},\alpha}(t)&=\braket{\alpha}{U^{\dagger}_{g+g^{\prime}}U_{g}(t)\sigma_{-}U_{g}(t)U_{g+g^{\prime}}(t)}{\alpha}\\\
&=\bra{\alpha}D^{\dagger}(-\tilde{\lambda})e^{i\omega
n/\pi}D(-\tilde{\lambda})U^{\dagger}_{g}(t)\\\ &\ \ \
\times\sigma_{-}U_{g}(t)D^{\dagger}(\tilde{\lambda})e^{-i\omega
n/\pi}D(\tilde{\lambda})\ket{\alpha}\\\
&=\braket{\alpha}{D(2\tilde{\lambda})e^{i\omega
n/\pi}U^{\dagger}_{g}(t)\sigma_{-}U_{g}(t)e^{-i\omega
n/\pi}D(2\tilde{\lambda})}{\alpha}\\\
&=\braket{\alpha}{D\left[2\tilde{\lambda}-\lambda(1+e^{i\omega
t})\right]D\left[2\tilde{\lambda}-\lambda(1+e^{i\omega t})\right]}{\alpha}\\\
&=\bra{0}D(-\alpha)D\left[2\tilde{\lambda}-\lambda(1+e^{i\omega t})\right]\\\
&\ \ \ \times D\left[2\tilde{\lambda}-\lambda(1+e^{i\omega
t})\right]D(\alpha)\ket{0}\\\
&=e^{\phi}\braket{\alpha-2\tilde{\lambda}+\lambda(1+e^{i\omega
t})}{\alpha+2\tilde{\lambda}-\lambda(1+e^{i\omega t})}\\\
&=e^{2\phi}e^{-|4\tilde{\lambda}^{2}-2\lambda(1+e^{i\omega
t})|^{2}/2}.\end{split}$ (50)
To go from the second to third line, we inserted a pair of identity operators
$1=e^{-i\omega n/\pi}e^{i\omega n/\pi}$ and used $e^{i\omega
n/\pi}D(\tilde{\lambda})e^{-i\omega n/\pi}=D(-\tilde{\lambda})$. From the
third to fourth we used the same trick and the more general time evolution
$e^{i\omega nt}D(\tilde{\lambda})e^{-i\omega nt}=D(\tilde{\lambda}e^{i\omega
t})$. In the last few lines the “phase” is
$\phi=\alpha^{*}(2\tilde{\lambda}-\lambda(1+e^{i\omega t}))/2+{\rm c.c.}.$
(51)
Note that we got two factors of this: one in the fifth line, from the braiding
relation $D(\alpha)D(\beta)=e^{(\alpha\beta^{*}-\alpha^{*}\beta)/2}$, and then
another in the subsequent line from the inner product
$\braket{\beta}{\alpha}=e^{-|\beta-\alpha|^{2}/2}e^{(\alpha\beta^{*}-\alpha^{*}\beta)/2}$.
Notice also that the second exponential does not depend on the coherent state
parameter $\alpha$. Thus we only need to average this phase term over the
Glauber representation, which gives
$\displaystyle\begin{split}&\int\frac{d^{2}\alpha}{\pi\bar{n}}e^{-|\alpha|^{2}/\bar{n}}e^{2\phi}\\\
&\ \ \ =\exp\left(8\lambda
n(\lambda+2\lambda^{\prime})\cos(t\omega)-8n\left(\lambda^{2}+2\lambda\lambda^{\prime}+2\lambda^{\prime
2}\right)\right),\end{split}$ (52)
where we used the explicit coefficient
$\tilde{\lambda}=\lambda+\lambda^{\prime}$. Doing the same with the second
term in (50), and simplifying the terms, we finally obtain
$\displaystyle\begin{split}&V_{\rm
b}(t)=\int\frac{d^{2}\alpha}{\pi\bar{n}}V_{{\rm b},\alpha}(t)\\\
&=\exp\left[-8(2\bar{n}+1)\left(\lambda^{\prime
2}+2\lambda\lambda^{\prime}\sin^{2}\frac{\omega
t}{2}+\lambda^{2}\sin^{2}\frac{\omega t}{2}\right)\right],\end{split}$ (53)
as quoted in (20). Note the limit $\lambda^{\prime}\to 0$ reproduces the
basic, un-boosted protocol. We show the form of this visibility evolution in
Fig. 6.
## Appendix D Using many atoms
Figure 7: Spin-echo variant of the basic protocol. After a half-period of
evolution, the two pathways through oscillator phase space are maximally
distant. The atomic positions are then flipped ($X$ gate), followed by another
half-period of evolution. This procedure can be repeated arbitrarily, leading
to a net amplification of the basic protocol. The bottom-right figure shows
the resulting conditioned paths of the oscillator through phase space, through
one iteration.
The sensitivity of the protocol can be substantially improved by moving from
using a single atom to use a collection of atoms, as is typical in an atom
interferometer kasevich1992measurement ; santarelli1999quantum ;
gross2010nonlinear . For simplicity, we will take $g$ the same for all the
atoms, though that is not necessary in practice. In that limit, we can define
$J_{z}=\sum_{j}\sigma_{z}^{j}$ and $J_{-}=\sum_{j}\sigma_{-}^{j}$ as the
collective variables that will enter.
Consider the extension of (9) to the case of $N$ atoms prepared in the initial
state $\ket{+++\cdots}$. The “observable” of interest is $\braket{J_{-}(t)}$.
This is easiest to calculate term by term for each atom. The total time
evolution operator, following the same logic as in (3), is
$U(t)=\exp\left(-i\beta J_{z}^{2}\right)D^{\dagger}(\lambda J_{z})e^{i\omega
a^{\dagger}at}D(\lambda J_{z}),$ (54)
with
$\beta=\frac{g^{2}t}{\omega}.$ (55)
This total $J_{z}^{2}$ term is peculiar to the case of $N>1$ atoms; for $N=1$
it is just an overall phase which we dropped in (3). Here, however, it is a
non-trivial operator, physically representing the ponderomotive squeezing of
the spins due to the gravitational coupling with the oscillator. At time $t>0$
we have, for each $i=1,\ldots,N$, the operator evolution
$\sigma_{-}^{i}(t)=U^{\dagger}(t)\sigma_{-}^{i}U(t).$ (56)
If we define $\tilde{J}_{z}^{i}=J_{z}-\sigma_{z}^{i}$, we can see that we can
write the $J_{z}^{2}$ contribution to (56) as
$\exp\left(i\beta J_{z}^{2}\right)\sigma_{-}^{i}\exp\left(-i\beta
J_{z}^{2}\right)=\exp\left(-2i\beta\tilde{J}_{z}\right)\sigma_{-}^{i}.$ (57)
This prefactor then commutes with the rest of the operators in (56). Using
this and the same basic logic as in (9), we find that all the $i\neq j$ spins
will just give a phase proportional to $\tilde{J}_{z}$:
$\sigma_{-}^{i}(t)=\sigma_{-}^{i}e^{-2i\beta\tilde{J}_{z}}D(-\lambda)D(2\lambda
e^{i\omega t})D(-\lambda).$ (58)
Acting on the initial state $\ket{0,+++\cdots}$ with the oscillator prepared
in $\ket{0}$ and each atom in the $\ket{+}$ state, that is to say with $N$
unentangled atoms, we obtain
$\braket{\sigma_{-}^{i}(t)}=\cos^{N-1}(2\beta)\braket{\sigma_{-}(t)}_{1}$ (59)
where the term $\braket{\sigma_{-}(t)}_{1}$ means the answer with a single
spin, as in equations (9) and (10). For $N\gg 1$ and $\beta\ll 1$ (recall
$\beta=g^{2}t/\omega$, so this condition is certainly satisfied for us), we
can Taylor expand the cosine and match it to an exponential for convenience,
i.e., $\cos^{N-1}(2\beta)\approx e^{-2N\beta^{2}}$. Thus since
$\braket{J_{-}}$ has $N$ terms of the form (59), we finally get
$\braket{J_{-}(t)}=Ne^{-2Ng^{4}t^{2}/\omega^{2}}\braket{\sigma_{-}(t)}_{1}.$
(60)
For example, at a half period and full period we then have
$\displaystyle\begin{split}\braket{J_{-}(\pi/\omega)}&=Ne^{-2\pi^{2}N\lambda^{4}}e^{-8\lambda^{2}},\\\
\braket{J_{-}(2\pi/\omega)}&=Ne^{-4\pi^{2}N\lambda^{4}}.\end{split}$ (61)
Note that this noisy phase is independent of the oscillator’s initial state,
so for example we get the same answer if the oscillator begins in a thermal
state. We see the basic $N$ enhancement to the signal here as the prefactor.
The phase noise scales like $N\lambda^{4}$. For our particular implementation
with parameters like those quoted in (25), we have $\lambda\sim 10^{-13}$, so
for $N\sim 10^{10}$ atoms these phase noise exponentials are completely
negligible. The overall $N$ factor here represents the usual $\sqrt{N}$
statistical enhancement in the signal-to-noise, assuming uncorrelated atom
errors.
The same calculation extends directly to the entanglement-enhanced, $g$-linear
protocol (20). This is clear by the algebraic structure of the argument given
above. Explicitly, we now have two time evolution operators of the form (54),
one with a coupling $g+g^{\prime}$ from $t=-\pi/\omega$ to $t=0$, followed by
another with only the gravitational $g$ coupling from $t=0$ onward. In an
obvious notation we can write
$\sigma^{i}_{-}(t)=U^{\dagger}_{g}(t)U^{\dagger}_{g+g^{\prime}}\sigma_{-}^{i}U_{g+g^{\prime}}U_{g}(t).$
(62)
In these $U$ operators, we have the same phase-noise terms, namely $e^{i\beta
J_{z}^{2}}$ in the $U_{g}$ and another factor $e^{i\beta^{\prime}J_{z}^{2}}$,
with $\beta^{\prime}=-(g+g^{\prime})^{2}\pi/\omega$, from the
$U_{g+g^{\prime}}$ factor. These depend only on the $J_{z}$ operator and thus
commute with the other terms (displacement operators and free oscillator
evolution) in $U_{g}$ and $U_{g+g^{\prime}}$. Thus we get an expression
$\displaystyle\begin{split}\sigma^{i}_{-}(t)&\sim\exp(i(\beta+\beta^{\prime})J_{z}^{2})\sigma_{-}^{i}\exp(-i(\beta+\beta^{\prime})J_{z}^{2})\\\
&=\exp(-2i(\beta+\beta^{\prime})\tilde{J}_{z})\sigma_{-}^{i},\end{split}$ (63)
just as in (57) except now with $\beta$ replaced by
$\beta+\beta^{\prime}=\frac{g^{2}t}{\omega}-\frac{(g+g^{\prime})^{2}\pi}{\omega}.$
(64)
In particular, all the terms other than these phase noise exponentials
contribute as given by (20). The overall signal is still increased linearly in
$N$ as in (61), times a negligible contribution from this ponderomotive
squeezing noise.
## Appendix E Spin-echo version for faster physical oscillators
The physical oscillator frequency is crucially important to the size of the
observable effect. The interferometric contrast scales as a power of
$\lambda=g/\omega$, so a low-frequency oscillator is ideal. However, in
practice, using a very low-frequency (sub-Hz) oscillator would present
substantial technical problems. This can be alleviated by using a high-
frequency oscillator and a spin-echo like sequence to mimic the effect of a
low-frequency oscillator. Specifically, after every $\pi/\omega$ half-period,
we swap the two atom locations, i.e. perform a $\sigma_{x}$ operation (see
Fig. 7). This produces the evolution
$\displaystyle\begin{split}U&=\sigma_{x}e^{-iH\tau/2}\sigma_{x}e^{-iH\tau/2}\\\
&=e^{-i(H_{p}-V)\tau/2}e^{-i(H_{p}+V)\tau/2}\\\
&=D(\lambda\sigma_{z})e^{-iH_{p}\tau/2}D(-\lambda\sigma_{z})D(-\lambda\sigma_{z})e^{-iH_{p}\tau/2}D(\lambda\sigma_{z})\\\
&=D(\lambda\sigma_{z})D(2\lambda\sigma_{z})D(\lambda\sigma_{z})=D(4\lambda\sigma_{z})\end{split}$
(65)
where we used $\tau=2\pi/\omega$ for the final line. With $N_{\pi}$ iterations
of this, we produce the total evolution
$U(N_{\pi})=D(4N_{\pi}\lambda\sigma_{z})$. Performing $N_{\pi}$ iterations,
followed by a $\sigma_{x}$ operation, followed by $N_{\pi}$ further
iterations, we recover the revival:
$\displaystyle\begin{split}U_{\rm spin-
echo}&=\sigma_{x}D(4N_{\pi}\lambda\sigma_{z})\sigma_{x}D(4N_{\pi}\lambda\sigma_{z})\\\
&=D(-4N_{\pi}\lambda\sigma_{z})D(4N_{\pi}\lambda\sigma_{z})\\\
&=\mathbb{I}.\end{split}$ (66)
In this spin-echo style variant, the wavefunction overlap after a total time
$t=N_{\pi}\tau$, i.e. to the halfway point, is given by
$O=\exp\left[-32\frac{N_{\pi}^{2}g^{2}}{\omega^{2}}\right].$ (67)
Thus we have an effect scaling like $g^{2}/\omega_{\rm eff}^{2}$, with the
effective frequency $\omega_{\rm eff}=2\pi/t=\omega/N_{\pi}$. This means we
can have an effectively slow oscillator (which is beneficial for the signal
strength) while using a faster physical oscillator (beneficial for noise
reasons), at the cost of having to perform some $\sigma_{x}$ operations on the
two-state system.
We note that application of this spin-echo protocol would violate the Markov
assumption used to prove the theorem in section III. Adjusting that proof to
accommodate this specific type of non-Markovian control would be necessary to
draw the same conclusion, namely, that the gravitational entanglement in this
protocol is necessarily due to the mass-atom entangling interaction.
## References
* (1) R. P. Feynman, F. B. Morinigo, and W. G. Wagner, Lectures on Gravitation, 1962-63. California Institute of Technology, 1971.
* (2) D. N. Page and C. Geilker, “Indirect evidence for quantum gravity,” Physical Review Letters 47 no. 14, (1981) 979.
* (3) L. Diósi, “Gravitation and quantum-mechanical localization of macro-objects,” Physics Letters, Section A: General, Atomic and Solid State Physics 105 no. 4-5, (1984) 199–202.
* (4) R. Penrose, “On gravity’s role in quantum state reduction,” General relativity and gravitation 28 no. 5, (1996) 581–600.
* (5) D. Kafri, G. Milburn, and J. Taylor, “Bounds on quantum communication via Newtonian gravity,” New Journal of Physics 17 no. 1, (2015) 015006.
* (6) S. Bose, A. Mazumdar, G. W. Morley, H. Ulbricht, M. Toroš, M. Paternostro, A. A. Geraci, P. F. Barker, M. Kim, and G. Milburn, “Spin entanglement witness for quantum gravity,” Physical Review Letters 119 no. 24, (2017) 240401.
* (7) C. Marletto and V. Vedral, “Gravitationally-induced entanglement between two massive particles is sufficient evidence of quantum effects in gravity,” Phys. Rev. Lett. 119 no. 24, (2017) 240402, arXiv:1707.06036 [quant-ph].
* (8) S. A. Haine, “Searching for Signatures of Quantum Gravity in Quantum Gases,” arXiv:1810.10202 [quant-ph].
* (9) H. Chevalier, A. J. Paige, and M. S. Kim, “Witnessing the nonclassical nature of gravity in the presence of unknown interactions,” Phys. Rev. A 102 no. 2, (2020) 022428, arXiv:2005.13922 [quant-ph].
* (10) R. Howl, V. Vedral, M. Christodoulou, C. Rovelli, D. Naik, and A. Iyer, “Testing Quantum Gravity with a Single Quantum System,” arXiv:2004.01189 [quant-ph].
* (11) C. Anastopoulos and B.-L. Hu, “Quantum Superposition of Two Gravitational Cat States,” Class. Quant. Grav. 37 no. 23, (2020) 235012, arXiv:2007.06446 [quant-ph].
* (12) A. Matsumura and K. Yamamoto, “Gravity-induced entanglement in optomechanical systems,” Phys. Rev. D 102 no. 10, (2020) 106021, arXiv:2010.05161 [gr-qc].
* (13) D. Carney, P. C. E. Stamp, and J. M. Taylor, “Tabletop experiments for quantum gravity: a user’s manual,” Classical and Quantum Gravity 36 no. 3, (2019) 034001.
* (14) M. Horodecki, P. Horodecki, and R. Horodecki, “Separability of mixed states: necessary and sufficient conditions,” Physics Letters A 223 no. 1, (1996) 1–8.
* (15) B. M. Terhal, “Bell inequalities and the separability criterion,” Physics Letters A 271 no. 5-6, (2000) 319–326.
* (16) R. P. Feynman and F. L. Vernon, “The theory of a general quantum system interacting with a linear dissipative system,” Annals of Physics 24 no. 11, (1963) 173.
* (17) A. O. Caldeira and A. J. Leggett, “Path integral approach to quantum brownian motion,” Physica A: Statistical mechanics and its Applications 121 no. 3, (1983) 587–616.
* (18) E. Joos and H. D. Zeh, “The emergence of classical properties through interaction with the environment,” Zeitschrift für Physik B Condensed Matter 59 no. 2, (1985) 223–243.
* (19) W. H. Zurek, “Decoherence, einselection, and the quantum origins of the classical,” Reviews of Modern Physics 75 no. 3, (2003) 715\.
* (20) M. Kasevich and S. Chu, “Measurement of the gravitational acceleration of an atom with a light-pulse atom interferometer,” Applied Physics B 54 no. 5, (1992) 321–332.
* (21) G. Santarelli, P. Laurent, P. Lemonde, A. Clairon, A. G. Mann, S. Chang, A. N. Luiten, and C. Salomon, “Quantum projection noise in an atomic fountain: A high stability cesium frequency standard,” Physical Review Letters 82 no. 23, (1999) 4619.
* (22) C. Gross, T. Zibold, E. Nicklas, J. Esteve, and M. K. Oberthaler, “Nonlinear atom interferometer surpasses classical precision limit,” Nature 464 no. 7292, (2010) 1165–1169.
* (23) V. Xu, M. Jaffe, C. D. Panda, S. L. Kristensen, L. W. Clark, and H. Müller, “Probing gravity by holding atoms for 20 seconds,” Science 366 no. 6466, (2019) 745–749.
* (24) J. Lee, E. Adelberger, T. Cook, S. Fleischer, and B. Heckel, “New test of the gravitational $1/r^{2}$ law at separations down to 52 $\mu$m,” Physical Review Letters 124 no. 10, (2020) 101101.
* (25) S. B. Cataño-Lopez, J. G. Santiago-Condori, K. Edamatsu, and N. Matsumoto, “High-Q milligram-scale monolithic pendulum for quantum-limited gravity measurements,” Physical Review Letters 124 no. 22, (2020) 221102\.
* (26) A. Belenchia, R. M. Wald, F. Giacomini, E. Castro-Ruiz, Č. Brukner, and M. Aspelmeyer, “Quantum superposition of massive objects and the quantization of gravity,” Physical Review D 98 no. 12, (2018) 126009.
* (27) M. Christodoulou and C. Rovelli, “On the possibility of laboratory evidence for quantum superposition of geometries,” Phys. Lett. B 792 (2019) 64–68, arXiv:1808.05842 [gr-qc].
* (28) R. J. Marshman, A. Mazumdar, and S. Bose, “Locality and entanglement in table-top testing of the quantum nature of linearized gravity,” Phys. Rev. A 101 no. 5, (2020) 052110, arXiv:1907.01568 [quant-ph].
* (29) T. D. Galley, F. Giacomini, and J. H. Selby, “A no-go theorem on the nature of the gravitational field beyond quantum theory,” arXiv:2012.01441 [quant-ph].
* (30) R. P. Feynman, “Quantum theory of gravitation,” Acta Phys. Polon. 24 (1963) 697–722.
* (31) G. t Hooft and M. Veltman, “One-loop divergencies in the theory of gravitation,” in Annales de l’IHP Physique Théorique, vol. 20, pp. 69–94. 1974\.
* (32) S. Deser and P. van Nieuwenhuizen, “One-loop divergences of quantized Einstein-Maxwell fields,” Physical Review D 10 no. 2, (1974) 401.
* (33) M. J. G. Veltman, “Quantum Theory of Gravitation,” Conf. Proc. C 7507281 (1975) 265–327.
* (34) J. F. Donoghue, “General relativity as an effective field theory: The leading quantum corrections,” Physical Review D 50 no. 6, (1994) 3874\.
* (35) C. P. Burgess, “Quantum gravity in everyday life: General relativity as an effective field theory,” Living Reviews in Relativity 7 no. 1, (2004) 1–56.
* (36) R. Howl, R. Penrose, and I. Fuentes, “Exploring the unification of quantum theory and general relativity with a bose–einstein condensate,” New Journal of Physics 21 no. 4, (2019) 043047.
* (37) A. Tilloy, “Does gravity have to be quantized? Lessons from non-relativistic toy models,” J. Phys. Conf. Ser. 1275 no. 1, (2019) 012006, arXiv:1903.01823 [quant-ph].
* (38) D. E. Bruschi and F. K. Wilhelm, “Self gravity affects quantum states,” arXiv:2006.11768 [quant-ph].
* (39) T. W. B. Kibble and S. Randjbar-Daemi, “Nonlinear Coupling of Quantum Theory and Classical Gravity,” J. Phys. A13 (1980) 141.
* (40) D. Kafri, J. Taylor, and G. Milburn, “A classical channel model for gravitational decoherence,” New Journal of Physics 16 no. 6, (2014) 065020.
* (41) J. Oppenheim, “A post-quantum theory of classical gravity?,” arXiv:1811.03116 [hep-th].
* (42) A. Kent, “Tests of Quantum Gravity near Measurement Events,” arXiv:2010.11811 [gr-qc].
* (43) O. Gühne and G. Tóth, “Entanglement detection,” Physics Reports 474 no. 1-6, (2009) 1–75.
* (44) L. Pezze, A. Smerzi, M. K. Oberthaler, R. Schmied, and P. Treutlein, “Quantum metrology with nonclassical states of atomic ensembles,” Reviews of Modern Physics 90 no. 3, (2018) 035005.
* (45) A. Stern, Y. Aharonov, and Y. Imry, “Phase uncertainty and loss of interference: A general picture,” Physical Review A 41 no. 7, (1990) 3436.
* (46) L. Rowan, E. Hahn, and W. Mims, “Electron-spin-echo envelope modulation,” Physical Review 137 no. 1A, (1965) A61.
* (47) S. A. Dikanov and Y. Tsvetkov, Electron spin echo envelope modulation (ESEEM) spectroscopy. CRC press, 1992.
* (48) J.-M. Raimond, M. Brune, and S. Haroche, “Manipulating quantum entanglement with atoms and photons in a cavity,” Reviews of Modern Physics 73 no. 3, (2001) 565.
* (49) E. M. Rains, “Rigorous treatment of distillable entanglement,” Physical Review A 60 no. 1, (1999) 173.
* (50) G. Lindblad, “On the generators of quantum dynamical semigroups,” Communications in Mathematical Physics 48 no. 2, (1976) 119–130.
* (51) V. Gorini, A. Kossakowski, and E. C. G. Sudarshan, “Completely positive dynamical semigroups of n-level systems,” Journal of Mathematical Physics 17 no. 5, (1976) 821–825.
* (52) J. Preskill, “Lecture notes on quantum information and computation,” 1998.
* (53) M. R. Gallis and G. N. Fleming, “Environmental and spontaneous localization,” Physical Review A 42 no. 1, (1990) 38.
* (54) T. M. Karg, B. Gouraud, C. T. Ngai, G.-L. Schmid, K. Hammerer, and P. Treutlein, “Light-mediated strong coupling between a mechanical oscillator and atomic spins 1 meter apart,” Science 369 no. 6500, (2020) 174–179.
* (55) R. A. Thomas, M. Parniak, C. Østfeldt, C. B. Møller, C. Bærentsen, Y. Tsaturyan, A. Schliesser, J. Appel, E. Zeuthen, and E. S. Polzik, “Entanglement between distant macroscopic mechanical and spin systems,” Nature Physics 17 no. 2, (2021) 228–233.
* (56) M. X. Xin, W. S. Leong, C. Zilong, and S.-Y. Lan, “An atom interferometer inside a hollow-core photonic crystal fiber,” Science Advances 4 (2018) e1701723.
* (57) R. Jannin, P. Cladé, and S. Guellati-Khélifa, “Phase shift due to atom-atom interactions in a light-pulse atom interferometer,” Physical Review A 92 (2015) 013616.
* (58) T. Laudat, V. Dugrain, T. Mazzoni, M.-Z. Huang, C. L. G. Alzar, A. Sinatra, P. Rosenbusch, and J. Reichel, “Spontaneous spin squeezing in a rubidium bec,” New Journal of Physics 20 (2018) 073018.
* (59) K. E. McAlpine, D. Gochnauer, and S. Gupta, “Excited-band Bloch oscillations for precision atom interferometry,” Physical Review A 101 (2020) 023614.
* (60) R. D. Niederriter, C. Schlupf, and P. Hamilton, “Cavity probe for real-time detection of atom dynamics in an optical lattice,” Physical Review A 102 (2020) 051301(R).
* (61) M. Jaffe, P. Haslinger, V. X. Xu, P. Hamilton, A. Upadhye, B. Elder, J. Khoury, and H. Müller, “Testing sub-gravitational forces on atoms from a miniature, in-vacuum source mass,” Nature Physics 13 (2017) 938–942.
* (62) M. Jaffe, V. Xu, P. Haslinger, and P. Hamilton, “Efficient adiabatic spin-dependent kicks in an atom interferometer,” Physical Review Letters 121 (2018) 040402.
* (63) Z. Pagel, W. Zhong, R. H. Parker, C. T. Olund, N. Y. Yao, and H. Müller, “Symmetric bloch oscillations of matter waves,” Physical Review A 102 (2020) 053312.
* (64) T. Purdy, K. Grutter, K. Srinivasan, and J. Taylor, “Quantum correlations from a room-temperature optomechanical cavity,” Science 356 no. 6344, (2017) 1265–1268.
* (65) M. Peskin and D. Schroeder, An introduction to quantum field theory. CRC press, 2018.
* (66) L. P. Kadanoff, “Scaling laws for Ising models near $T_{c}$,” Physics Physique Fizika 2 no. 6, (1966) 263.
* (67) K. G. Wilson, “Renormalization group and critical phenomena. I. Renormalization group and the Kadanoff scaling picture,” Physical review B 4 no. 9, (1971) 3174.
* (68) S. Weinberg, “Phenomenological Lagrangians,” Physica A 96 no. 1-2, (1979) 327–340.
* (69) S. Weinberg, The quantum theory of fields, vol. 2. Cambridge university press, 1995.
* (70) C. Anastopoulos and B. Hu, “A master equation for gravitational decoherence: probing the textures of spacetime,” Classical and Quantum Gravity 30 no. 16, (2013) 165007.
* (71) C. Anastopoulos and B.-L. Hu, “Comment on ”A Spin Entanglement Witness for Quantum Gravity” and on ”Gravitationally Induced Entanglement between Two Massive Particles is Sufficient Evidence of Quantum Effects in Gravity”,” arXiv:1804.11315 [quant-ph].
* (72) D. Carney, “Newton, entanglement, and the graviton,” In preparation .
* (73) A. A. Clerk, M. H. Devoret, S. M. Girvin, F. Marquardt, and R. J. Schoelkopf, “Introduction to quantum noise, measurement, and amplification,” Reviews of Modern Physics 82 no. 2, (2010) 1155.
|
# Simple and maximally robust processes with no classical common-cause or
direct-cause explanation
Marcello Nery<EMAIL_ADDRESS>Departamento de Física, Universidade Federal
de Minas Gerais, Av. Pres. Antonio Carlos 6627 - Belo Horizonte, MG, Brazil -
31270-901. Marco Túlio Quintino Institute for Quantum Optics and Quantum
Information (IQOQI), Austrian Academy of Sciences, Boltzmanngasse 3, A-1090
Vienna, Austria Faculty of Physics, University of Vienna, Boltzmanngasse 5,
1090 Vienna, Austria Philippe Allard Guérin Faculty of Physics, University
of Vienna, Boltzmanngasse 5, 1090 Vienna, Austria Institute for Quantum
Optics and Quantum Information (IQOQI), Austrian Academy of Sciences,
Boltzmanngasse 3, A-1090 Vienna, Austria Perimeter Institute for Theoretical
Physics, 31 Caroline St. N, Waterloo, Ontario, N2L 2Y5, Canada Thiago O.
Maciel Departamento de Física, Universidade Federal de Santa Catarina,
Florianópolis, SC, 88040-900, Brazil Universidade Federal do Rio de Janeiro,
Caixa Postal 68528, Rio de Janeiro, RJ 21941-972, Brazil Reinaldo O. Vianna
Departamento de Física, Universidade Federal de Minas Gerais, Av. Pres.
Antonio Carlos 6627 - Belo Horizonte, MG, Brazil - 31270-901.
(September 02, 2021)
###### Abstract
Guided by the intuition of coherent superposition of causal relations, recent
works presented quantum processes without classical common-cause and direct-
cause explanation, that is, processes which cannot be written as probabilistic
mixtures of quantum common-cause and quantum direct-cause relations (CCDC). In
this work, we analyze the minimum requirements for a quantum process to fail
to admit a CCDC explanation and present “simple” processes, which we prove to
be the most robust ones against general noise. These simple processes can be
realized by preparing a maximally entangled state and applying the identity
quantum channel, thus not requiring an explicit coherent mixture of common-
cause and direct-cause, exploiting the possibility of a process to have both
relations simultaneously. We then prove that, although all bipartite direct-
cause processes are bipartite separable operators, there exist bipartite
separable processes which are not direct-cause. This shows that the problem of
deciding whether a process is direct-cause _is not_ equivalent to entanglement
certification and points out the limitations of entanglement methods to detect
non-classical CCDC processes. We also present a semi-definite programming
hierarchy that can detect and quantify the non-classical CCDC robustnesses of
every non-classical CCDC process. Among other results, our numerical methods
allow us to show that the simple processes presented here are likely to be
also the maximally robust against white noise. Finally, we explore the
equivalence between bipartite direct-cause processes and bipartite processes
without quantum memory, to present a separable process which cannot be
realized as a process without quantum memory.
## Introduction
Common-cause and direct-cause relations are the building blocks of classical
causal models, as a causal model of multiple variables consists of
combinations of common-cause and direct-cause relations between them.
Understanding the relationship between cause and effects is one of the
fundamental goals of several physical theories and of statistical analysis
[1]. Also, determining the causal relation behind a correlation of two objects
is a fundamental problem in causal inference theory and plays a main role in
topics such as hypothesis testing, social sciences, medicine, and machine
learning [2, 3]. In order to analyze causality in quantum phenomena, recent
works proposed new frameworks of theory of Bayesian inference [4] and causal
modeling [5], where understanding common-cause and direct-cause relations in
quantum mechanics plays a fundamental role.
Quantum causal processes consist of a sequence of quantum operations and may
be analyzed from different equivalent perspectives, such as sequential quantum
operations via non-Markovian processes [6, 7, 8, 9], fragments of quantum
circuits via quantum combs [10, 11], quantum channels with memory [12], and
quantum strategies for playing an $n$-turn game [13]. When referring to causal
modeling, a process which can be written as a probabilistic mixture of common-
cause (CC) and direct-cause (DC) processes is said to admit a classical
common-cause or direct-cause (CCDC) explanation, since the causal relations in
such process could be simulated classically, by sampling from a probability
distribution and then implementing either a common-cause or direct-cause
process.
In Ref. [14] the authors consider the different possibilities of combining CC
and DC processes and, inspired by coherent mixture of quantum channels, the
authors of Ref. [15] experimentally certify the existence of a quantum process
with no classical CCDC explanation. Also, inspired by the _Quantum Switch_
[16, 17, 18], Ref. [19] presents a coherent superposition of common-cause and
direct-cause processes which cannot be explained by a classical CCDC.
In this work, we analyze quantum processes which cannot be decomposed as
probabilistic mixtures of common-cause and direct-cause ones, hence admitting
no classical CCDC explanation. We focus on the bipartite case, the simplest
scenario where such processes can exist. In this simplest scenario, exploring
the possibility of a process to have both CC and DC relations simultaneously,
we present a process which can be simply realized by preparing a maximally
entangled state and an identity channel, not needing an interpretation of it
as a coherent mixture of causal relations. When attempting to work with the
minimum non-trivial dimensions, we propose another process, which requires a
controlled-NOT operation. We prove that these processes are the most robust
ones against their worst possible noise, known as generalized noise. We also
develop a semi-definite programming (SDP) numerical approach to quantify the
non-classical CCDC property of a process, based on its robustness against
white and general noise. Our numerical methods allow us to show that these
simple processes presented here are also maximally robust against white noise
when considering qubits, and to quantify the non-classical CCDC property in
any bipartite ordered process.
We also investigate the differences and similarities between quantum
entanglement and processes without a direct-cause explanation, showing that
there exist separable processes which are not direct-cause. This example
points out the limitations of purely entanglement-based methods to detect non
direct-cause processes and answers a conjecture first raised in Ref. [20].
Finally, we connect our results with a different related field by proving that
for the bipartite case, processes without quantum memory [20] are equivalent
to processes having a direct-cause decomposition. With that, we present a
bipartite separable process which cannot be realized as a process without
quantum memory and contribute to the understanding of the relation between
entanglement and of quantum memory [21, 20, 22, 23, 24].
## 1 Mathematical Preliminaries
### 1.1 The Choi-Jamiołkowski isomorphism
As is standard in several branches of quantum information theory, we will make
use of the Choi-Jamiołkowski isomorphism to represent linear operators and
linear maps [25, 26, 27]. We first present the “pure” version of the
isomorphism, which allows us to represent linear operators as bipartite
vectors. Here $A_{I}$ and $A_{O}$ are finite-dimensional complex vector
spaces, which will late be identified with Alice’s input and Alice’s output,
respectively.
###### Definition 1.1 (“Pure” Choi-Jamiołkowski isomorphism).
Let $\textrm{U}:(A_{I})\rightarrow(A_{O})$ be a linear transformation (which
is not necessarily unitary). The Choi vector $|{\textrm{U}}\rangle\rangle\in
A_{I}\otimes A_{O}$ of the operator U is defined as
$|{\textrm{U}}\rangle\rangle:=\sum_{i}|{i}\rangle\otimes(\textrm{U}\,|{i}\rangle),$
(1)
where $\\{|{i}\rangle\\}_{i=0}^{d_{A_{I}}-1}$ is the computational basis for
$A_{I}$ and $d_{A_{I}}$ is the dimension of $A_{I}$
In quantum information theory, linear transformations between operators are
sometimes referred to as linear maps, or simply, as maps. In this paper, we
identify linear maps with a single tilde. For instance, let
$\mathcal{L}(A_{I})$ be the set of linear operators mapping $A_{I}$ to itself,
$\widetilde{\Lambda}:\mathcal{L}(A_{I})\to\mathcal{L}(A_{O})$ is a map
transforming operators from $\mathcal{L}(A_{I})$ to $\mathcal{L}(A_{O})$. We
now present the standard version of the Choi-Jamiołkowski isomorphism, which
allows us to represent linear maps as bipartite operators.
###### Definition 1.2 (Choi-Jamiołkowski isomorphism).
Let $\widetilde{\Lambda}:\mathcal{L}(A_{I})\rightarrow\mathcal{L}(A_{O})$ be a
linear map and $\\{|{i}\rangle\\}_{i=0}^{d_{A_{I}}-1}$ be the computational
basis for $A_{I}$. The Choi operator $\Lambda\in\mathcal{L}(A_{I}\otimes
A_{O})$ of the map $\widetilde{\Lambda}$ is defined as
$\Lambda:=\sum_{ij}|{i}\rangle\langle{j}|\otimes\widetilde{\Lambda}(|{i}\rangle\langle{j}|).$
(2)
A common equivalent way to define the Choi operator of a map
$\widetilde{\Lambda}$ is given by the equation
$\Lambda:=\left[\left(\widetilde{\mathds{1}}\otimes\widetilde{\Lambda}\right)(|{\mathds{1}}\rangle\rangle\langle\langle{\mathds{1}}|)\right],$
(3)
where $\widetilde{\mathds{1}}:\mathcal{L}(A_{I})\rightarrow\mathcal{L}(A_{I})$
is an identity channel, and
$|{\mathds{1}}\rangle\rangle=\sum_{i=0}^{d_{A_{I}}-1}|{ii}\rangle\in
A_{I}\otimes A_{I}$.
The action of the map
$\widetilde{\Lambda}:\mathcal{L}(A_{I})\rightarrow\mathcal{L}(A_{O})$ on an
operator $\rho\in\mathcal{L}(A_{I})$ can be obtained from the Choi operator
$\Lambda$ by means of the relation
$\widetilde{\Lambda}(\rho)=\mathrm{tr}_{A_{I}}\left[({\rho^{A_{I}}}^{T}\otimes\mathds{1}^{A_{O}})\cdot\Lambda\right],$
(4)
where $\mathds{1}$ is the identity operator and $(\cdot)^{T}$ stands for the
transposition in the computational basis.
A linear map is a quantum channel iff it is completely positive (CP) and
trace-preserving (TP). In the Choi-Jamiołkowski representation, a map
$\widetilde{\Lambda}:\mathcal{L}(A_{I})\rightarrow\mathcal{L}(A_{O})$ is CP
iff $\Lambda\succeq 0$, _i.e._ , $\Lambda$ is positive semi-definite, and TP
iff $\mathrm{tr}_{A_{O}}(\Lambda)=\mathds{1}^{A_{I}}$.
For the sake of clarity, some equations in this manuscript will explicitly
identify the input and output Hilbert spaces of a map with a superscript in
the symbol representing the Choi operator of the map. In the above situation,
for example, the Choi operator is equivalently represented by
$\Lambda^{A_{I}/A_{O}}$, indicating that the map $\widetilde{\Lambda}$ takes
an operator from $\mathcal{L}(A_{I})$ to one in $\mathcal{L}(A_{O})$.
### 1.2 The link product operation
First defined in Ref. [11], the link product is a useful mathematical tool to
deal with composition of elements in a quantum circuit presented in the Choi-
Jamiołkowski representation.
###### Definition 1.3 (Link product).
Let $\Lambda_{1}\in\mathcal{L}(A\otimes A^{\prime})$ and
$\Lambda_{2}\in\mathcal{L}(A^{\prime}\otimes A^{\prime\prime})$ be linear
operators. The link product between $\Lambda_{1}$ and $\Lambda_{2}$ is defined
as
$\displaystyle\begin{split}\Lambda_{2}*\Lambda_{1}:=\mathrm{tr}_{A^{\prime}}\Bigg{[}\bigg{(}\Lambda_{1}^{T_{A^{\prime}}}\
\otimes\mathds{1}^{A^{\prime\prime}}\bigg{)}\cdot\bigg{(}\mathds{1}^{A}\otimes\Lambda_{2}\bigg{)}\Bigg{]},\end{split}$
(5)
where $\Lambda_{1}^{T_{A^{\prime}}}$ is the partial transpose of $\Lambda_{1}$
on the space $A^{\prime}$.
As stated previously, the link product is useful for composing linear maps and
quantum objects. If
$\widetilde{\Lambda}_{1}:\mathcal{L}(A)\rightarrow\mathcal{L}(A^{\prime})$ and
$\widetilde{\Lambda}_{2}:\mathcal{L}(A^{\prime})\rightarrow\mathcal{L}(A^{\prime\prime})$
are linear maps with Choi operators $\Lambda_{1}\in\mathcal{L}(A\otimes
A^{\prime})$ and $\Lambda_{2}\in\mathcal{L}(A^{\prime}\otimes
A^{\prime\prime})$, respectively, it can be shown that the Choi operator of
the composition
$\widetilde{\Phi}:=\widetilde{\Lambda}_{2}\circ\widetilde{\Lambda}_{1}:\mathcal{L}(A)\rightarrow\mathcal{L}(A^{\prime\prime})$
is $\Phi=\Lambda_{2}*\Lambda_{1}.$
In particular, when $\rho\in\mathcal{L}(A)$, is a linear operator with no
components on $A^{\prime}$, and $\Lambda\in\mathcal{L}(A\otimes A^{\prime})$,
we have
$\displaystyle\begin{split}\Lambda*\rho:=&\mathrm{tr}_{A}\left[\left(\rho^{T_{A}}\
\otimes\mathds{1}^{A^{\prime}}\right)\cdot\left(\Lambda\right)\right]\\\
=&\widetilde{\Lambda}(\rho).\end{split}$ (6)
Hence, we can use the link product to represent quantum operations being
performed on quantum states. Also, if $\rho_{1}\in\mathcal{L}(A_{1})$ and
$\rho_{2}\in\mathcal{L}(A_{2})$, acting on different spaces, we have
$\rho_{1}*\rho_{2}:=\rho_{1}\otimes\rho_{2}$. Therefore, the link product of
independent systems is simply the tensor product.
Additionally, when $\rho,M\in\mathcal{L}(A)$ act in the same linear space, the
link product is given by $\rho*M:=\mathrm{tr}(M^{\textrm{T}}\cdot\rho)$, which
is simply the trace of their product with an extra transposition. This form
can also be used to write Born’s rule.
## 2 The CCDC scenario
The main objects analyzed in this work are the bipartite ordered processes,
which may be understood as a physical dynamics which flows from Alice to Bob.
For all definitions in this section, we consider a scenario where Alice is a
quantum physicist in a laboratory who can perform any quantum operation that
transforms states from system $A_{I}$ (Alice input) to $A_{O}$ (Alice output),
and Bob is a quantum physicist in a laboratory who can perform any quantum
measurement on states defined on system $B_{I}$ (Bob input).
### 2.1 Markovian processes
We start by presenting the definition of bipartite Markovian processes, which
admits a direct-cause interpretation [6, 28, 29] and are also known in the
literature as quantum processes with no memory [20], and cause-effect causal
maps [14]. In a Markovian scenario, Alice receives a known quantum state
$\rho\in\mathcal{L}(A_{I})$ which acts on her input system space. Alice can
then perform an arbitrary quantum operation111Which may be a quantum channel
(deterministic quantum operation) or a quantum instrument (probabilistic
quantum
operation).$\widetilde{\Lambda}:\mathcal{L}(A_{I})\to\mathcal{L}(A_{O})$ to
obtain a quantum state on system $A_{O}$, which will then be subjected to a
known deterministic dynamics described by a channel
$\widetilde{D}:\mathcal{L}(A_{O})\to\mathcal{L}(B_{I})$. Finally, the state
$\widetilde{D}\Big{(}\widetilde{\Lambda}(\rho)\Big{)}\in\mathcal{L}(B_{I})$
arrives to Bob, who can perform an arbitrary measurement on it. See Fig. 1 for
a circuit based pictorial illustration.
A bipartite Markovian process can then be described by the known quantum state
$\rho\in\mathcal{L}(A_{I})$ and the known dynamics represented by the channel
$\widetilde{D}:\mathcal{L}(A_{O})\to\mathcal{L}(B_{I})$. In the Choi-
Jamiołkowski representation, this can be conveniently described by
$W_{\textrm{Markov}}:=\rho^{A_{I}}\otimes D^{A_{O}/B_{I}}$. Under this
definition, for any quantum operation
$\widetilde{\Lambda}:\mathcal{L}(A_{I})\to\mathcal{L}(A_{O})$ performed by
Alice, the quantum state arriving in Bob’s input space can be obtained via the
link product as
$\displaystyle\begin{split}W_{\textrm{Markov}}*\Lambda^{A_{I}/A_{O}}&=\rho^{A_{I}}\otimes
D^{A_{O}/B_{I}}*\Lambda^{A_{I}/A_{O}}\\\
&=\rho^{A_{I}}*D^{A_{O}/B_{I}}*\Lambda^{A_{I}/A_{O}}\\\
&=D^{A_{O}/B_{I}}*\Lambda^{A_{I}/A_{O}}*\rho^{A_{I}}\\\
&=D^{A_{O}/B_{I}}*\widetilde{\Lambda}(\rho^{A_{I}})\\\
&=\widetilde{D}\Big{(}\widetilde{\Lambda}(\rho)\Big{)}.\end{split}$ (7)
Figure 1: Circuit representation of Markovian processes: Alice receives a
fixed state $\rho\in\mathcal{L}(H_{A_{I}})$ on which she can perform an
arbitrary quantum operation. After Alice’s operation, the system is subjected
to a fixed dynamics described by a quantum channel
$\widetilde{D}:\mathcal{L}(A_{O})\to\mathcal{L}(B_{I})$, then arriving at
Bob’s input space laboratory. No auxiliary system is used in this circuit,
which implies that correlations between the parties are only due to the direct
communication from Alice to Bob.
###### Definition 2.1 (Markovian process).
A linear operator $W_{\text{Markov}}\in\mathcal{L}(A_{I}\otimes A_{O}\otimes
B_{I})$ is bipartite Markovian process if it can be written as
$\displaystyle\begin{split}W_{\textrm{Markov}}:=&\,\rho^{A_{I}}*D^{A_{O}/B_{I}},\\\
=&\,\rho^{A_{I}}\otimes D^{A_{O}/B_{I}},\end{split}$ (8)
where $\rho^{A_{I}}$ is a quantum state i.e., $\rho\succeq 0$,
$\mathrm{tr}(\rho)=1$ and $D^{A_{O}/B_{I}}$ is the Choi operator of a quantum
channel from the output of Alice to Bob’s input i.e., $D\succeq 0$,
$\mathrm{tr}_{B_{I}}(D)=\mathds{1}_{A_{O}}$. We denote the set of Markovian
processes by $\mathcal{L}_{\textrm{Markov}}$.
### 2.2 Direct-cause processes
In a Markovian process, all correlations between Alice and Bob admit a direct-
cause interpretation, since no auxiliary system or environment is required at
any point. In a scenario where Alice and Bob may share prior _classical_
correlations, it is natural to define direct-cause processes as any process
which can be written as a probabilistic mixture of Markovian processes. Hence,
the correlations arising from these processes can always be explained by
classical mixture of direct-cause ones.
###### Definition 2.2 (Direct-cause process).
A linear operator $W_{\text{DC}}\in\mathcal{L}(A_{I}\otimes A_{O}\otimes
B_{I})$ is direct-cause if it is a classical mixture of Markovian processes,
that is, $W_{\text{DC}}$ can be written as
$\displaystyle W_{\textrm{DC}}:=$
$\displaystyle\sum_{i}p_{i}\rho_{i}^{A_{I}}*D_{i}^{A_{O}/B_{I}}$ (9a)
$\displaystyle=$ $\displaystyle\sum_{i}p_{i}\rho_{i}^{A_{I}}\otimes
D_{i}^{A_{O}/B_{I}},\ p_{i}\in[0,1],$ (9b)
where $\rho^{A_{I}}_{i}$ are quantum states, i.e., $\rho_{i}\succeq 0$ and
$\mathrm{tr}(\rho_{i})=1\ \forall i$, whereas $D^{A_{O}/B_{I}}_{i}$ are Choi
operators of quantum channels from the output of Alice to Bob’s input, i.e.,
$D_{i}\succeq 0$ and $\mathrm{tr}_{B_{I}}(D_{i})=\mathds{1}_{A_{O}}\ \forall
i$. We denote the set of direct-cause processes by
$\mathcal{L}_{\textrm{DC}}$.
In the bipartite scenario, an ordered process is direct-cause if and only if
it is a process without quantum memory, as defined and shown in Ref. [20]. See
Appendix A for details. Although Ref. [19] analyzes tripartite ordered
processes, when bipartite processes are considered, their definition is
equivalent to the one presented in Refs. [14, 15], which is also equivalent to
the one used here. We discuss this in more details in Section 6.2.
#### 2.2.1 Quantum separability as an outer approximation for the set of DC
processes
A positive semi-definite linear operator $W\in\mathcal{L}(A_{I}\otimes
A_{O}\otimes B_{I})$ is separable in the bipartition $A_{I}|A_{O}B_{I}$ if
there exist a probability distribution $p_{i}$ and positive semi-definite
operators $\rho_{i}^{A_{I}}\succeq 0$ and $D_{i}^{A_{O}/B_{I}}\succeq 0$ such
that $W=\sum_{i}p_{i}\rho_{i}^{A_{I}}\otimes D_{i}^{A_{O}/B_{I}}$. Which
coincides with the definition of a DC process (Def. 2.2) without the trace
normalization constraints. It follows directly from Eq. (9b) that every
direct-cause process is separable in the bipartition $A_{I}|A_{O}B_{I}$.
Conversely, if a process is entangled in the bipartition $A_{I}|A_{O}B_{I}$,
such process cannot have a DC explanation. This allows us to employ techniques
from entanglement theory to certify whether a process is not direct-cause. A
simple outer approximation for the set of separable states is the set of
states with positive partial transpose (PPT) [30, 31]. Also, it is known that
the set of states with a PPT k-symmetric extension [32, 33] forms a hierarchy
of sets which converges to the set of separable states when $k\to\infty$. For
every fixed $k$, checking if a state has a PPT symmetric k-extension can be
done via SDP. The precise definition of a PPT k-symmetric extension is
provided in Appendix E.
###### Definition 2.3 (Separable outer approximation for DC processes).
A bipartite ordered process $W$ belongs to
$\mathcal{L}_{\text{DC}}^{\text{out},\text{PPT}_{k}}$ if $W$ has a PPT
k-symmetric extension on the bipartition $A_{I}|A_{O}B_{I}$.
One could be tempted to assume that all process which are separable in
$A_{I}|A_{O}B_{I}$ bipartition have a direct-cause explanation. However, in
Section 7, we prove that this is not the case by constructing an explicit
example. This ensures that the problem of certifying if a given process is not
direct-cause is not equivalent to certifying entanglement.
### 2.3 Common-cause processes
Common-cause processes are those where the correlations are not due to
communication from Alice to Bob, but only due to a bipartite quantum state
$\rho\in\mathcal{L}(A_{I}\otimes B_{I})$ initially shared by them. In this
way, a common-cause process does not involve any particular quantum dynamic
and can be fully characterized by a fixed bipartite state. In order to
establish a notation which dialogues to general scenarios which may have non-
trivial dynamics between Alice and Bob, we describe common-cause processes by
$W_{\textrm{CC}}=\rho^{A_{I}B_{I}}\otimes\mathds{1}^{A_{O}}$.
###### Definition 2.4 (Common-cause process).
A linear operator $W_{\text{CC}}\in\mathcal{L}(A_{I}\otimes A_{O}\otimes
B_{I})$ is a common-cause process if it can be written as
$W_{\textrm{CC}}=\rho^{A_{I}B_{I}}\otimes\mathds{1}^{A_{O}},$ (10)
where $\rho^{A_{I}B_{I}}$ is a quantum state shared between Alice and Bob,
i.e., $\rho^{A_{I}B_{I}}\succeq 0$, $\mathrm{tr}(\rho^{A_{I}B_{I}})=1$, and
$\mathds{1}^{A_{O}}$ is the identity operator acting on $A_{O}$. We denote the
set of common-cause processes by $\mathcal{L}_{\textrm{CC}}$.
### 2.4 Classical CCDC processes
Classical CCDC processes are processes which can be decomposed as a convex
combination of a common-cause and a direct-cause ones. Such processes can
always be understood as simple classical statistical mixtures between quantum
common-cause and quantum direct-cause processes [14, 19].
###### Definition 2.5 (Classical CCDC processes).
A linear operator $W_{\text{CCDC}}\in\mathcal{L}(A_{I}\otimes A_{O}\otimes
B_{I})$ is a classical CCDC process if it can be written as
$W_{\textrm{CCDC}}:=pW_{\textrm{CC}}+(1-p)W_{\textrm{DC}},$ (11)
where $W_{\textrm{CC}}\in\mathcal{L}_{\textrm{CC}}$,
$W_{\textrm{DC}}\in\mathcal{L}_{\textrm{DC}}$, and $p\in[0,1]$. We denote the
set of classical CCDC processes by $\mathcal{L}_{\textrm{CCDC}}$, which is the
convex hull of $\mathcal{L}_{\textrm{CC}}\cup\mathcal{L}_{\textrm{DC}}$.
### 2.5 Bipartite ordered processes
Previously, we have only considered processes which do not require any
explicit use of auxiliary systems or environments. In a more general process,
the initial state $\rho$ received by Alice may be defined in a larger space,
and Alice’s lab can only operate on part of it. In order to tackle this
situation, we now consider that the initial state is defined not only on
Alice’s input system, but also in some auxiliary system with unconstrained
finite dimension, that is, $\rho\in\mathcal{L}(A_{I}\otimes\textrm{aux})$.
Now, Alice can only operate on the part of $\rho$ that enters her laboratory,
that is, she can perform any operation
$\widetilde{\Lambda}:\mathcal{L}(A_{I})\to\mathcal{L}(A_{O})$, which will
transform the initial state $\rho$ into222Here
$\widetilde{\mathds{1}}^{\text{aux}}$ stands for the identity map on the
auxiliary space, that is, for any operator $A\in\mathcal{L}(\textrm{aux})$ we
have $\widetilde{\mathds{1}}^{\text{aux}}(A)=A$.$\widetilde{\Lambda}^{A_{I}\to
A_{O}}\otimes\widetilde{\mathds{1}}^{\text{aux}}(\rho)$, which will be
subjected to a global dynamics described by a quantum channel
$\widetilde{D}:\mathcal{L}(A_{O}\otimes\textrm{aux})\to\mathcal{L}(B_{I})$.
See Fig. 2 for a circuit based pictorial illustration.
Figure 2: A circuit illustrating a bipartite ordered process $W=\rho*D$ where
$\rho\in\mathcal{L}(A_{I}\otimes\textrm{aux})$ is a quantum state and $D$ is
the Choi operator of a channel
$\widetilde{D}:\mathcal{L}(A_{O}\otimes\textrm{aux})\rightarrow\mathcal{L}(B_{I})$.
When a quantum channel
$\widetilde{\Lambda}:\mathcal{L}(A_{I})\rightarrow\mathcal{L}(A_{O})$ with
Choi operator $\Lambda$ is “plugged” into the process $W$, the state
$\rho^{\prime}=W*\Lambda\in\mathcal{L}(B_{I})$ is obtained.
A bipartite ordered process can then be described by the known quantum state
$\rho\in\mathcal{L}(A_{I}\otimes\textrm{aux})$ and the known dynamics
represented by the channel
$\widetilde{D}:\mathcal{L}(A_{O}\otimes\textrm{aux})\to\mathcal{L}(B_{I})$. In
the Choi-Jamiołkowski representation, this can be conveniently described by
$W:=\rho^{A_{I}\textrm{aux}}*D^{A_{O}\textrm{aux}/B_{I}}$. Under this
definition, for any quantum operation
$\widetilde{\Lambda}:\mathcal{L}(A_{I})\to\mathcal{L}(A_{O})$ performed by
Alice, the quantum state arriving in Bob’s input space can be obtained via the
link product as
$\displaystyle\begin{split}W*\Lambda^{A_{I}/A_{O}}&=\rho^{A_{I}\textrm{aux}}*D^{A_{O}\textrm{aux}/B_{I}}*\Lambda^{A_{I}/A_{O}}\\\
&=D^{A_{O}/B_{I}}*\Lambda^{A_{I}/A_{O}}*\rho^{A_{I}\textrm{aux}}\\\
&=D^{A_{O}/B_{I}}*\widetilde{\Lambda}^{A_{I}\to
A_{O}}\otimes\widetilde{\mathds{1}}^{\text{aux}}\big{(}\rho^{A_{I}\textrm{aux}}\big{)}\\\
&=\widetilde{D}\Big{(}\widetilde{\Lambda}^{A_{I}\to
A_{O}}\otimes\widetilde{\mathds{1}}^{\text{aux}}\big{(}\rho^{A_{I}\textrm{aux}}\big{)}\Big{)}.\end{split}$
(12)
###### Definition 2.6 (Bipartite ordered process).
A linear operator $W\in\mathcal{L}(A_{I}\otimes A_{O}\otimes B_{I})$ is a
bipartite ordered process if there exists a quantum state
$\rho^{A_{I}\textrm{aux}}\in\mathcal{L}(A_{I}\otimes\textrm{aux})$,i.e.,
$\rho^{A_{I}\textrm{aux}}\succeq 0$, $\mathrm{tr}(\rho^{A_{I}\textrm{aux}})=1$
and a quantum channel
$\widetilde{D}:\mathcal{L}(A_{I}\otimes\textrm{aux})\rightarrow\mathcal{L}(B_{I})$
with Choi operator $D^{A_{O}\textrm{aux}/B_{I}}$, i.e.,
$D^{A_{O}\textrm{aux}/B_{I}}\succeq 0$,
$\mathrm{tr}_{B_{O}}(D^{A_{O}\textrm{aux}/B_{I}})=\mathds{1}_{A_{0},\textrm{aux}}$,
such that
$\displaystyle\begin{split}W:=&\rho^{A_{I}\textrm{aux}}*D^{A_{O}\textrm{aux}/B_{I}}\\\
=&\mathrm{tr}_{\textrm{aux}}\Bigg{(}\left(\rho^{A_{I}\textrm{aux}}\otimes\mathds{1}^{A_{O}B_{I}}\right)^{T_{\textrm{aux}}}\\\
&\hskip 19.91684pt\cdot\left(\mathds{1}^{A_{I}}\otimes
D^{A_{O}\textrm{aux}/B_{I}}\right)\Bigg{)}.\end{split}$ (13)
We denote the set of bipartite ordered processes by $\mathcal{L}_{A\rightarrow
B}$.
Bipartite ordered process may be seen as quantum objects transforming a
quantum operation into a quantum state333In Ref. [34, 11] the authors prove
that under a set of assumptions, it can be proven that bipartite ordered
process are indeed the most general method to transform arbitrary quantum
operations into quantum states.and are also known in the literature as quantum
non-Markovian processes [7], causal maps [14], quantum co-strategies [13], and
ordered process matrices [35]. Also, ordered processes may also be seen as a
particular instance of quantum channels with memory [12] and quantum combs
[11].
In this general definition of bipartite ordered processes, Markovian processes
form the particular case where there is no auxiliary space aux. Common-cause
processes can also be represented as bipartite ordered processes. Indeed, if
$W_{\textrm{CC}}=\rho^{A_{I}B_{I}}\otimes\mathds{1}^{A_{O}}$, we can set the
auxiliary space $\mathcal{H}_{\text{aux}}$ as isomorphic to
$\mathcal{H}_{B_{I}}$ and the initial state $\rho^{A_{I}\text{aux}}$ as
isomorphic to $\rho^{A_{I}B_{I}}$. Then, one can set the channel
$\widetilde{D}$ as an operation that sends the auxiliary system to $B_{I}$ and
discard the system in $A_{O}$. This can be done formally via
$\displaystyle W_{\textrm{CC}}$
$\displaystyle=\mathrm{tr}_{\textrm{aux}^{\prime}}\left[\rho^{A_{I}\textrm{aux}}*|{\textrm{U}_{\textrm{SWAP}}}\rangle\rangle\langle\langle{\textrm{U}_{\textrm{SWAP}}}|^{A_{O}\textrm{aux}/B_{I}\textrm{aux}^{\prime}}\right]$
$\displaystyle=\rho^{A_{I}B_{I}}\otimes\mathds{1}^{A_{O}},$ (14)
where $\textrm{U}_{\textrm{SWAP}}:=\sum_{ij}|{ij}\rangle\langle{ji}|$ is the
swap operator. A circuit illustration of common-cause process as bipartite
ordered ones is presented in Fig. 3.
Figure 3: A general bipartite common-cause process can be represented by a
circuit of this form. The shared quantum state between Alice and Bob comes
from an initial correlated state between Alice and an auxiliary system. The
state of the auxiliary system is exchanged with the system coming from the
output of Alice and sent to Bob, and the auxiliary system is then discarded,
which makes the output state of Alice to have no impact on the input of Bob.
This makes only the input states of Alice and Bob to be relevant for the
process, which is why the correlations between them stem from the common-cause
correlations on $\rho$.
With all the important processes defined, Fig. 4 shows the relationship
between all the sets defined in this section.
Figure 4: Illustrative representation of the sets in the CCDC scenario. The
sets of common-cause processes ($\mathcal{L}_{\textrm{CC}}$) and of direct-
cause processes ($\mathcal{L}_{\textrm{DC}}$) are proper subsets of the set of
bipartite ordered process $\mathcal{L}_{A\rightarrow B}$, with a non-empty
intersection between $\mathcal{L}_{\textrm{CC}}$ and
$\mathcal{L}_{\textrm{DC}}$. Classical CCDC processes are all processes in the
convex hull of $\mathcal{L}_{\textrm{CC}}\cup\mathcal{L}_{\textrm{DC}}$,
represented by $\mathcal{L}_{\textrm{CCDC}}$.
In addition to its definition, bipartite ordered processes admit a
characterization in terms of linear and positive semi-definite constraints,
which will be useful later in this work. If follows from direct inspection
that all bipartite ordered process $W\in\mathcal{L}(A_{I}\otimes A_{O}\otimes
B_{I})$ respect
$\displaystyle W$ $\displaystyle\succeq 0,$ (15a)
$\displaystyle\mathrm{tr}_{B_{I}}(W)$
$\displaystyle=\sigma^{A_{I}}\otimes\mathds{1}^{A_{O}},$ (15b)
where $\sigma^{A_{I}}$ is a quantum state. Interestingly, this condition is
also sufficient [11]. For any linear operator $W$ respecting the conditions in
Eq. (15), one can find a quantum state
$\rho\in\mathcal{L}(A_{I}\otimes\textrm{aux})$ and a quantum channel
$\widetilde{D}:\mathcal{L}(A_{O}\otimes\textrm{aux})\to\mathcal{L}(B_{I})$
such that $\rho^{A_{I}\text{aux}}*D^{\text{aux}B_{I}/B_{O}}=W$. One possible
construction is done by setting the auxiliary space aux to be isomorphic to
$A_{I}$ and $\rho\in\mathcal{L}(A_{I}\otimes\textrm{aux})$ be be isomorphic to
a purification of $\sigma^{A_{I}}$, for instance,
$\rho:=\left(\mathds{1}^{A_{I}}\otimes\sqrt{\sigma^{\textrm{aux}}}^{T}\right)\;|{\mathds{1}}\rangle\langle{\mathds{1}}|^{A_{I}\text{aux}}\;\left(\mathds{1}^{A_{I}}\otimes\sqrt{\sigma^{\textrm{aux}}}^{T}\right).$
(16)
Now, one can define a quantum channel via
$\displaystyle\begin{split}D:=&\left(\sqrt{\sigma^{\textrm{aux}}}^{-1}\otimes\mathds{1}^{A_{O}B_{I}}\right)\\\
&W^{\text{aux}A_{O}B_{I}}\;\left(\sqrt{\sigma^{\textrm{aux}}}^{-1}\otimes\mathds{1}^{A_{O}B_{I}}\right),\end{split}$
(17)
where $\sqrt{\sigma}$ is the unique positive semi-definite square root of
$\sigma$ and $\sigma^{-1}$ is the Moore–Penrose inverse of $\sigma$, that is,
the inverse of $\sigma$ on its range. In this way, direct inspection shows
that $D$ is the Choi operator of a quantum channel and that $W=\rho*D$.
Inspired by Ref. [36], we now present another characterization for bipartite
ordered processes which will be useful for proving our results. A linear
operator $W\in\mathcal{L}(A_{I}\otimes A_{O}\otimes B_{I})$ is a bipartite
ordered process if and only if it respects
$\displaystyle W$ $\displaystyle\succeq 0,$ (18a) $\displaystyle W$
$\displaystyle=L_{A\rightarrow B}(W),$ (18b) $\displaystyle\mathrm{tr}(W)$
$\displaystyle=d_{A_{O}},$ (18c)
where
$L_{A\rightarrow B}(W):=W+\prescript{}{A_{O}B_{I}}{W}-\prescript{}{B_{I}}{W},$
(19)
with
$\prescript{}{X}{(\cdot)}=\mathrm{tr}_{X}(\cdot)\otimes\frac{\mathds{1}^{X}}{d_{X}}$
and $L_{A\rightarrow B}(W)$ is the map projecting an operator $W$ into the
linear space spanned by bipartite ordered processes. Eqs. (15) and (18) are
equivalent conditions for an operator $W\in\mathcal{L}(A_{I}\otimes
A_{O}\otimes B_{I})$ to be a valid bipartite ordered process. In particular,
conditions given by Eqs. (18) are rather useful for implementing our numerical
methods, due to the use of the projector $L_{A\rightarrow B}(W)$.
## 3 Detecting and quantifying non-classical CCDC
Inspired by the robustness of entanglement [37, 38], one can quantify the
violation of a classical CCDC decomposition in a process
$W\in\mathcal{L}(A_{I}\otimes A_{O}\otimes B_{I})$ in terms of its generalized
robustness of non-classical CCDC444Note that the definition of generalized
robustness $R_{G}$ here has a one-to-one relation to the “non-classicality of
causality” $\mathcal{C}$ presented in Ref.[19] via
$R_{G}(W)=\frac{\mathcal{C}(W)}{1+\mathcal{C}(W)}$. [19] by
$\displaystyle R_{G}(W):=\textrm{min}\hskip 5.69046pt$ $\displaystyle r$
subject to $\displaystyle(1-r)W+r\Omega\in\mathcal{L}_{\textrm{CCDC}},$ (20a)
$\displaystyle 0\leq r\leq 1,$ (20b)
$\displaystyle\Omega\in\mathcal{L}_{A\rightarrow B},$ (20c)
where $\mathcal{L}_{A\rightarrow B}$ is the set of bipartite ordered processes
defined in Eqs. (18). The generalized robustness $R_{G}(W)$ corresponds to how
resistant is the non-classical CCDC property of the process $W$ against its
worst possible noise.
In Appendix B, we show that the generalized robustness of all bipartite
processes is upper-bounded by $R_{G}(W)\leq 1-\frac{1}{d_{A_{I}}}$, which is
attainable with the examples of non-classical CCDC processes we present in
Section 5.
Another quantification of non-classical CCDC is in terms of its _white-noise
robustness_ , given by
$\displaystyle R_{WN}(W):=\textrm{min}\hskip 5.69046pt$ $\displaystyle r$
subject to
$\displaystyle(1-r)W+r\frac{\mathds{1}}{d_{A_{I}}d_{B_{I}}}\in\mathcal{L}_{\textrm{CCDC}},$
(21a) $\displaystyle 0\leq r\leq 1,$ (21b)
where $\mathds{1}$ is the identity operator in $\mathcal{L}(A_{I}\otimes
A_{O}\otimes B_{I})$. The value $R_{WN}(W)$ corresponds to how resistant is
the non-classical CCDC property of the process $W$ against white noise. Figure
5 gives a representation of both types of robustness.
In Appendix C, we show that the white noise robustness of all bipartite
processes is upper-bounded by $R_{WN}(W)\leq
1-\frac{1}{d_{A_{I}}d_{A_{O}}d_{B_{I}}+1}$. Such bound is not tight, but is
useful to demonstrate that the convex optimization problem of Eq.(21) respects
strong duality (see Appendix D for more details).
Figure 5: Representation of the problem of determining if a given process $W$
is non-classical CCDC. The above figure illustrates an ordered process
operator $\Omega\in\mathcal{L}_{A\rightarrow B}$ such that the minimum convex
combination of it with the process $W$ results in a classical CCDC process.
### 3.1 Witnessing non-classical CCDC processes
A non-classical CCDC witness is a Hermitian operator
$S\in\mathcal{L}(A_{I}\otimes A_{O}\otimes B_{I})$ which respects
$\mathrm{tr}(SW_{\text{CCDC}})\geq 0$ for every classical CCDC process
$W_{\text{CCDC}}$ [19]. Non-classical CCDC witnesses that are not trivial are
the ones having $\mathrm{tr}(SW)<0$ for some non-classical CCDC process $W$,
as the main purpose of a witness is to certify that a given process is non-
classical CCDC. Since the set of classical CCDC processes is closed and
convex, similarly to entanglement [39], every non-classical CCDC process may
be certified by a non-classical CCDC witness. Also, as proven in Appendices D
and E, the convex optimization problems used to define generalized and white
noise robustness of non-classical CCDC respect a strong duality condition and
the dual formulation of our outer approximation can be used to explicitly
construct non-classical CCDC witness.
Similarly to entanglement witnesses, verifying if an arbitrary operator
$S\in\mathcal{L}(A_{I}\otimes A_{O}\otimes B_{I})$ is a non-classical CCDC
witness is likely to be computationally hard [40], as it would require, in
principle, to verify that $\mathrm{tr}(SW_{\text{CCDC}})\geq 0$ for all
classical CCDC processes $W_{\text{CCDC}}$. Despite the hardness of the
general problem, we now provide simple sufficient (but not necessary)
analytical conditions to ensure that $S$ is a non-classical CCDC witness.
Any operator $S\in\mathcal{L}(A_{I}\otimes A_{O}\otimes B_{I})$ respecting
$\mathrm{tr}_{A_{O}}(S)\succeq 0$ and $S^{T_{A_{I}}}\succeq 0$ is a non-
classical CCDC witness. This claim follows from the fact that every CC process
respects $W_{\text{CC}}=\prescript{}{A_{O}}{W_{\text{CC}}}\succeq 0$ and every
DC process respects $W_{\text{DC}}^{T{{}_{A_{I}}}}\succeq 0$. Hence,
$\displaystyle\begin{split}\mathrm{tr}(SW_{\text{CCDC}})&=p\leavevmode\nobreak\
\mathrm{tr}(SW_{\text{CC}})+(1-p)\mathrm{tr}(SW_{\text{DC}})\\\
&=p\leavevmode\nobreak\
\mathrm{tr}(S\prescript{}{A_{O}}{W_{\text{CC}}})+(1-p)\mathrm{tr}(S^{T_{A_{I}}}W^{T_{A_{I}}}_{\text{DC}})\\\
&=p\leavevmode\nobreak\
\mathrm{tr}(\prescript{}{A_{O}}{S}W_{\text{CC}})+(1-p)\mathrm{tr}(S^{T_{A_{I}}}W^{T_{A_{I}}}_{\text{DC}})\\\
&\geq 0.\end{split}$ (22)
## 4 Beyond the entanglement approximation: SDP hierarchies for tight non-
classical CCDC certification
As stated in Sec. 2.2.1, the set of separable processes in the bipartition
$A_{I}|A_{O}B_{I}$ provides an outer approximation for the set of DC
processes. As pointed in Ref. [19, 20], this provides a SDP approach that
gives lower bounds for the non-DC and non-classical CCDC robustness of a
bipartite ordered process. Indeed, since the relation
$\mathcal{L}_{\text{DC}}\subseteq\mathcal{L}_{\text{DC}}^{\text{PPT}_{k}}$
holds for every natural number $k$, we can define the convex hull
$\mathcal{L}_{\text{CCDC}}^{\text{out},\text{PPT}_{k}}:=\text{conv}(\mathcal{L}_{\text{CC}},\mathcal{L}_{\text{DC}}^{\text{out},\text{PPT}_{k}})$,
and the quantity
$\displaystyle R_{G}^{\text{low},\text{PPT}_{k}}(W):=\textrm{min}\hskip
5.69046pt$ $\displaystyle r$ subject to
$\displaystyle(1-r)W+r\Omega\in\mathcal{L}_{\text{CCDC}}^{\text{out},\text{PPT}_{k}}$
(23a) $\displaystyle 0\leq r\leq 1,$ (23b)
$\displaystyle\Omega\in\mathcal{L}_{A\rightarrow B},$ (23c)
which gives a lower-bound for the actual generalized robustness, since the
relation $R_{G}^{\text{low,PPT}_{k}}(W)\leq R_{G}(W)$ holds for every $W$.
Analogously, we can also obtain a lower-bound for the white noise robustness
by defining $R_{WN}^{\text{low,PPT}_{k}}(W)$.
As we show in Sec. 7, approximating the set of DC processes by the set of
separable process is rather limited. In particular, there exist processes
which are separable in the bipartition $A_{I}|A_{O}B_{I}$ but are non-
classical CCDC. Hence, such processes would never be detected by a method
solely based in quantum entanglement techniques. In this section, we overcome
this problem by presenting two novel SDP hierarchies of sets which converge to
the set of CCDC processes. Also, one hierarchy is based in an inner
approximation, whereas the other constitutes an outer approximation. Hence, we
can obtain a sequence of converging upper and lower bounds on the non-
classical CCDC robustnesses. All the sets discussed in this section are
illustrated in Fig. 6.
Figure 6: Hierarchical relation between the set of processes separable in the
bipartition $A_{I}|A_{O}B_{I}$ ($\mathcal{L}_{\text{SEP}}^{A_{I}|A_{O}B_{I}}$)
and the set of direct-cause processes $\mathcal{L}_{\text{DC}}$. In the limit
of large $k$, $\mathcal{L}_{\text{SEP}}^{\text{PPT}_{k}}$ converges to
$\mathcal{L}_{\text{SEP}}^{A_{I}|A_{O}B_{I}}$. By increasing the set of $N$
states, the inner approximation
($\mathcal{L}_{\text{DC}}^{\text{in},\mathcal{E}_{N}}$) converges to
$\mathcal{L}_{\text{DC}}$. Similarly, by increasing the set of $N$ trace-one
operators $\\{\hat{\rho}_{i=1}^{N}\\}_{i}$, the outer approximation
$\mathcal{L}_{\text{DC}}^{\text{out},\hat{\mathcal{E}}_{N}}$ also converges to
$\mathcal{L}_{\text{DC}}$. The point $W_{\text{SEP}}$ stands for the example
of process which is separable in the bipartition $A_{I}|A_{O}B_{I}$, but is
non-classical CCDC, defined and discussed in Sec. 7.
### 4.1 Inner approximation for direct-cause processes
The following inner approximation leads to a hierarchy that converges to the
set of direct-cause processes and is inspired by the SDP approach for
entanglement detection introduced in [41].
###### Definition 4.1 (Inner approximation for the direct-cause set).
Let $\mathcal{E}_{N}:=\\{|{\psi_{i}}\rangle^{A_{I}}\\}_{i=1}^{N}$ be a fixed
set of pure quantum states $|{\psi_{i}}\rangle^{A_{I}}$. The set
$\mathcal{L}_{\text{DC}}^{\text{in},\mathcal{E}_{N}}$ corresponds to the inner
approximation of the set $\mathcal{L}_{\text{DC}}$ and is composed by every
process $W$ that can be written as
$W=\sum_{i}p_{i}|{\psi_{i}}\rangle\langle{\psi_{i}}|^{A_{I}}\otimes
D_{i}^{A_{O}/B_{I}},$ (24a) where $\sum_{i}p_{i}=1\;\text{ and }\;p_{i}\geq
0,\;\forall i,$ (24b) and $D_{i}^{A_{O}/B_{I}}$ are the Choi operators of
quantum channels, i.e., $D_{i}^{A_{O}/B_{I}}\succeq 0\;\text{ and
}\;\mathrm{tr}_{B_{I}}\big{(}D_{i}^{A_{O}/B_{I}}\big{)}=\mathds{1}^{A_{O}},\;\forall
i.$ (24c)
Note that in Def. 4.1, we imposed the partial trace quantum channel
normalization condition
$\mathrm{tr}_{B_{I}}\big{(}D_{i}^{A_{O}/B_{I}}\big{)}=\mathds{1}^{A_{O}}$,
which does not appear in an entanglement-based approach where we only impose
positivity and the full trace constraint. As we will see in the further
sections, this corresponds to a fundamental difference, since the set of
separable quantum states is not equivalent to the set of direct-cause
processes.
By construction, for any set of quantum states $\mathcal{E}_{N}$, we have the
relation
$\mathcal{L}_{\text{DC}}^{\text{in},\mathcal{E}_{N}}\subseteq\mathcal{L}_{\text{DC}}$.
Also, if $\mathcal{E}_{\infty}$ is the set of all pure quantum states in
$\mathcal{L}(A_{I})$, we have
$\mathcal{L}_{\text{DC}}^{\text{in},\mathcal{E}_{\infty}}=\mathcal{L}_{\text{DC}}$.
We can now define the convex hull
$\mathcal{L}_{\text{CCDC}}^{\text{in},\mathcal{E}_{N}}:=\text{conv}(\mathcal{L}_{\text{CC}},\mathcal{L}_{\text{DC}}^{\text{in},\mathcal{E}_{N}})$,
and the quantity
$\displaystyle\begin{split}R_{G}^{\text{up,}\mathcal{E}_{N}}(W):=\textrm{min}\hskip
5.69046pt&r\\\ \textrm{subject to}\hskip
5.69046pt&(1-r)W+r\Omega\in\mathcal{L}_{\textrm{CCDC}}^{\text{in,}\mathcal{E}_{N}},\\\
&0\leq r\leq 1,\\\ &\Omega\in\mathcal{L}_{A\rightarrow B},\end{split}$ (25)
which provides an upper-bound for the generalized robustness $R_{G}(W)$, since
the relation $R_{G}(W)\leq R_{G}^{\text{up},\mathcal{E}_{N}}(W)$ holds for
every bipartite ordered process $W$ and $\mathcal{E}_{N}$.
As discussed in details in Appendix E, the quantity
$R_{G}^{\text{up},\mathcal{E}_{N}}(W)$ can be evaluated by an SDP. Moreover,
this SDP converges to the exact value of $R_{G}(W)$ when the set of states
$\mathcal{E}_{N}$ contains all pure states in $\mathcal{L}(A_{I})$.
Analogously, we can also obtain an upper-bound for the white noise robustness
by defining $R_{WN}^{\text{up},\mathcal{E}_{N}}(W)$.
### 4.2 Outer approximation for direct-cause processes
For the outer approximation for the set of direct-cause processes, we will
make use of an outer approximation for the set of quantum states. Since the
set of quantum states is convex, we can always construct an outer
approximation given by a polytope, similarly to the strategy used to construct
a polytopes providing outer approximations for the set of quantum measurements
[42, 43, 44, 45, 46]. Let $\mathcal{S}_{d}$ be the set of quantum states with
dimension $d$, there exists a set of linear operators
$\hat{\rho}_{i}\in\mathcal{L}(\mathbb{C}_{d})$ that satisfy
$\mathrm{tr}(\hat{\rho}_{i})=1$ and such that the convex hull of
$\\{\hat{\rho}_{i}\\}_{i}$ contains $\mathcal{S}_{d}$. Note that the operators
$\hat{\rho}_{i}$ are _not_ required to be positive semi-definite, hence they
do not correspond to quantum states.
For the moment, let us assume that we know a set of trace-one operators
$\\{\hat{\rho}_{i=1}^{N}\\}_{i}$ such that
$\text{conv}(\\{\hat{\rho}_{i}\\}_{i=1}^{N})\supseteq\mathcal{S}_{d}$. We can
then use this finite set to construct an outer approximation for the set of DC
processes, similarly to the inner approximation in Def. 4.1.
###### Definition 4.2 (Outer approximation for the direct-cause set).
Let $\hat{\mathcal{E}}_{N}:=\\{\hat{\rho}_{i}^{A_{I}}\\}_{i=1}^{N}$ be a fixed
set of linear operators with $\mathrm{tr}(\hat{\rho}_{i})=1\ \forall i$, such
that the convex hull of $\hat{\mathcal{E}}_{N}$ contains the set of all
quantum states acting on $A_{I}$. A bipartite ordered process $W$ is in the
set $\mathcal{L}_{\text{DC}}^{\text{out},\hat{\mathcal{E}}_{N}}$ if it can be
written as
$W=\sum_{i}p_{i}\hat{\rho}_{i}^{A_{I}}\otimes D_{i}^{A_{O}/B_{I}},$ (26a)
where $\sum_{i}p_{i}=1\;\text{ and }\;p_{i}\geq 0\;\forall i,$ (26b) and
$D_{i}^{A_{O}/B_{I}}$ are the Choi operators of quantum channels i.e.
$D_{i}^{A_{O}/B_{I}}\succeq 0\;\text{ and
}\;\mathrm{tr}_{B_{I}}\big{(}D_{i}^{A_{O}/B_{I}}\big{)}=\mathds{1}^{A_{O}}\;\forall
i.$ (26c)
By construction, we have that
$\mathcal{L}_{\text{DC}}\subseteq\mathcal{L}_{\text{DC}}^{\text{out},\hat{\mathcal{E}}_{N}}$
and the set
$\mathcal{L}_{\text{CCDC}}^{\text{out},\hat{\mathcal{E}}_{N}}:=\text{conv}(\mathcal{L}_{\text{CC}}\cup\mathcal{L}_{\text{DC}}^{\text{out},\hat{\mathcal{E}}_{N}})$
is an outer approximation for $\mathcal{L}_{\text{CCDC}}$. The generalized and
white noise robustnesses are defined in a way analogous to what is done in
Sec. 4.1, thus being denoted by $R_{G}^{\text{low},\hat{\mathcal{E}}_{N}}(W)$
and $R_{WN}^{\text{low},\hat{\mathcal{E}}_{N}}(W)$, and these are lower bounds
for the robustness $R_{G}(W)$ and $R_{WN}(W)$ respectively.
We now address the problem of finding sets of trace-one operators which
contain the set of quantum states. Let us first start with the simple qubit
case ($d=2$) where the set of quantum states can be faithfully represented in
terms of the Bloch sphere [47]. For this case, we can construct a set of
trace-one operators $\\{\hat{\rho}_{i}\\}_{i}$ such that
$\text{conv}(\\{\hat{\rho}_{i}\\}_{i})\supseteq\mathcal{S}_{2}$ simply by
finding a polyhedron that includes a sphere of unit radius. One method to find
such polyhedron goes as following:
1. 1.
Sort $N$ normalized vectors $\\{\vec{v_{i}}\\}_{i=1}^{N}$ in $\mathbb{R}^{3}$.
These will be the vertices of a polyhedron inscribed by the Bloch sphere;
2. 2.
Find the radius $r_{\text{in}}<1$ of the largest sphere inscribed by the
polyhedron;
3. 3.
Construct a polyhedron with vertices given by
$\frac{\vec{v_{i}}}{r_{\text{in}}}$. This polyhedron includes the Bloch
sphere.
The last step is ensured by the symmetry of the problem, a sphere of radius
$r_{\text{in}}$ is contained in the polytope with norm-one vertices
$\\{\vec{v_{i}}\\}_{i=1}^{N}$ if and only if a sphere of radius one contains
the polytope with vertices $\frac{\vec{v_{i}}}{r_{\text{in}}}$. Also, the
radius of the largest sphere inscribed by a polytope (required in step 2) can
be made by first finding the facet representation of this polytope with
vertices $\\{\vec{v_{i}}\\}_{i=1}^{N}$, step which can be done via Fourier-
Motzkin elimination and can be tackled with the aid of numerical packages such
as lrs [48] or the Matlab code vert2lcon [49]. With the facet representation,
the radius of the largest inscribed sphere can be found by evaluating the
distance of the origin to the hyperplane represented by each facet. A concrete
example for the qubit case can be found in the Appendix B: “A family of
polyhedra”, of Ref. [50]. In such work, the authors provide a family of
polyhedra parameterized by $n\in\mathbb{N}$ where the radius of the larger
inscribed sphere is greater than or equals to
$r_{n}:=\cos^{2}\big{(}\frac{\pi}{2n}\big{)}$ and if $n$ is an odd number, the
number of vertices is given by $N=2n^{2}$. This family of polyhedra can be
used for a systematic approach to the problem.
For the general $d>3$, we start by pointing out that every convex set can be
approximated by a polytope. Now, in order to obtain a concrete set of
operators $\\{\hat{\rho}_{i}\\}_{i}$, we refer to the methods described in
Appendix A: “Calculating shrinking factors”, of Ref. [43] (see also Red.
[51]). In a nutshell, the method starts by sampling random pure (which are
extremal) states $|{\psi_{i}}\rangle\langle{\psi_{i}}|$ in the set
$\mathcal{S}_{d}$ and by obtaining the facet representation for the polytope
associated to this set. Then, for any fixed “shrinking factor” $\eta$, we can
verify whether a noisy quantum state $\eta\rho+(1-\eta)\mathds{1}/d$ can be
written as a convex combination of
$\\{|{\psi_{i}}\rangle\langle{\psi_{i}}|\\}_{i}$ by means of an SDP. If all
$\eta$-noisy states can be written as convex combinations of
$\\{|{\psi_{i}}\rangle\langle{\psi_{i}}|\\}_{i}$, the convex hull of operators
is given by
$\hat{\rho}_{i}:=\frac{1}{\eta}|{\psi_{i}}\rangle\langle{\psi_{i}}|+1-\frac{1}{\eta}\mathds{1}/d$.
We observe that we can tighten the outer approximation presented in Def. 4.2
by imposing that the process $W$ should be separable in the bipartition
$A_{I}|A_{O}B_{I}$. Also, if we do this by means of the $PPT$ k-symmetric
extension (as in Def. 2.3), the problem can still be tackled by means of an
SDP.
Set | Robustness | Requirements for set membership
---|---|---
$\mathcal{L}_{\text{DC}}^{\text{out,PPT}_{k}}$ | $R^{\text{low,PPT}_{k}}$ | $W$ is a valid bipartite ordered process and has a PPT $k$-symmetric extension.
(This hierarchy converges to the set of separable processes, which is strictly
larger than $\mathcal{L}_{\text{DC}}$)
$\mathcal{L}_{\textrm{DC}}^{\text{out,}\hat{\mathcal{E}}_{N}}$ | $R^{\text{low},\hat{\mathcal{E}}_{N}}$ | Given a set of operators $\hat{\mathcal{E}}_{N}=\\{\hat{\rho}^{A_{I}}_{i}\\}_{i=1}^{k}$ forming an outer approximation for the set of quantum states, there exist quantum channels $D_{i}^{A_{O}/B_{I}}$ s.t. $W=\sum_{i}p_{i}\hat{\rho}^{A_{I}}_{i}\otimes D_{i}^{A_{O}/B_{I}}$.
$\mathcal{L}_{\textrm{DC}}^{\text{in,}\mathcal{E}_{N}}$ | $R^{\text{up},\mathcal{E}_{N}}$ | Given a set of states $\mathcal{E}_{N}=\\{\rho^{A_{I}}_{i}\\}_{i=1}^{k}$, there exist quantum channels $D_{i}^{A_{O}/B_{I}}$ s.t.
$W=\sum_{i}p_{i}{\rho}^{A_{I}}_{i}\otimes D_{i}^{A_{O}/B_{I}}$.
Table 1: Summary of the sets used as approximation to the set of direct-cause
processes, $\mathcal{L}_{\text{DC}}$, ordered from the outside to the inside
of the set. The operators $D_{i}^{A_{O}/B_{I}}$ stands for Choi operators of
quantum channels, $D_{i}^{A_{O}/B_{I}}\succeq 0$,
$\mathrm{tr}_{B_{I}}(D_{i}^{A_{O}/B_{I}})=\mathds{1}$ and $\\{p_{i}\\}_{i}$ is
a probability distribution. The operators $\hat{\rho}^{A_{I}}_{i}$ are not
positive semi-definite, hence they are not quantum states, but the convex hull
of the set $\\{\hat{\rho}^{A_{I}}_{i}\\}_{i=1}^{k}$ is a valid outer
approximation for the set of quantum states. The precise definitions of these
sets are given in Def. 2.3, Def. 4.2, and Def. 4.1. These approximations can
be used for generating approximations for the set of classical CCDC processes,
as
$\mathcal{L}_{\text{CCDC}}=\text{conv}(\mathcal{L}_{\text{CC}}\cup\mathcal{L}_{\text{DC}})$.
We present a summary of all the sets defined as approximations to the set
$\mathcal{L}_{\text{DC}}$ in Table 1. For obtaining the corresponding inner or
outer approximations for the set of classical CCDC processes
$\mathcal{L}_{\text{CCDC}}$, it is necessarily to take the convex hull of
$\mathcal{L}_{\text{CC}}$ with the chosen set for approximating
$\mathcal{L}_{\text{DC}}$. Before finishing this section, we state that,
throughout this work, whenever we calculate the non-classical CCDC
robustnesses of the inner and outer approximations for
$\mathcal{L}_{\text{CCDC}}$, the first lower bounds to be calculated are
always provided by the outer approximation
$\mathcal{L}_{\text{CCDC}}^{\text{out,PPT}_{k}}$. If the lower bounds obtained
with such approximation does not match the upper bounds, then we take the
outer approximation provided by
$\mathcal{L}_{\text{CCDC}}^{\text{out,}\hat{\mathcal{E}}_{N}}$.
## 5 Simplest non-classical CCDC processes
### 5.1 The scenario where $d_{A_{I}}=d_{A_{O}}=d$ and $d_{B_{I}}=d^{2}$
We now present a “simple” process which has the highest generalized robustness
in the scenario where $d_{A_{I}}=d_{A_{O}}=d$ and $d_{B_{I}}=d^{2}$. In this
process, Alice shares a $d$-dimensional maximally entangled state
$|{\phi^{+}}\rangle=\frac{1}{\sqrt{d}}\sum_{i=0}^{d-1}|{ii}\rangle$ with the
auxiliary system, and the channel of communication to Bob (decoder) is
composed of an identity channel,
$D=|{\mathds{1}}\rangle\rangle\langle\langle{\mathds{1}}|^{\textrm{aux}A_{O}/B_{I}^{1}B_{I}^{2}}$
(see Fig. 7). This process is mathematically represented by
$\displaystyle\begin{split}W_{ddd^{2}}&:=|{\phi^{+}}\rangle\langle{\phi^{+}}|^{A_{I}\textrm{aux}}*|{\mathds{1}}\rangle\rangle\langle\langle{\mathds{1}}|^{\textrm{aux}A_{O}/B_{I}^{1}B_{I}^{2}}\\\
&=|{\phi^{+}}\rangle\langle{\phi^{+}}|^{A_{I}B_{I}^{1}}\otimes|{\mathds{1}}\rangle\rangle\langle\langle{\mathds{1}}|^{A_{O}/B_{I}^{2}}.\end{split}$
(27)
Figure 7: Circuit representation of the process $W_{ddd^{2}}$. This process
consists in the preparation of a maximally entangled state between Alice and
an auxiliary system and an identity channel of communication to Bob.
We argue that this process has a conceptually simple realization, since it
only requires the preparation of a maximally entangled state and the identity
channel. The process $W_{ddd^{2}}$ has the interesting property of
transforming a quantum channel
$\widetilde{\Lambda}:\mathcal{L}(A_{I})\rightarrow\mathcal{L}(A_{O})$ into a
state which is proportional to its Choi operator
$\Lambda=\sum_{ij}|{i}\rangle\langle{j}|\otimes\widetilde{\Lambda}(|{i}\rangle\langle{j}|)$,
that is
$\displaystyle\begin{split}W_{ddd^{2}}*\Lambda&=\left(\widetilde{\mathds{1}}\otimes\widetilde{\Lambda}\right)|{\phi^{+}}\rangle\langle{\phi^{+}}|,\\\
&=\frac{\Lambda}{d}.\end{split}$ (28)
Despite the simplicity, we now show that $W_{ddd^{2}}$ is non-classical CCDC
and, moreover, it is the one with highest generalized robustness in the
scenario where $d_{A_{I}}=d_{A_{O}}=d$ and $d_{B_{I}}=d^{2}$.
###### Theorem 5.1.
The bipartite ordered process
$W_{ddd^{2}}=|{\phi^{+}}\rangle\langle{\phi^{+}}|^{A_{I}B_{I}^{1}}\otimes|{\mathds{1}}\rangle\rangle\langle\langle{\mathds{1}}|^{A_{O}/B_{I}^{2}}$
attains the maximum generalized robustness of all processes with the same
dimensions. That is,
$\displaystyle\begin{split}R_{G}(W_{ddd^{2}})=&\max_{W\in\mathcal{L_{A\rightarrow
B}}}\left[R_{G}(W)\right]\\\ =&1-\frac{1}{d_{A_{I}}}\end{split}$ (29)
###### Proof.
We start the proof by defining the operator
$\displaystyle\begin{split}S:=&\mathds{1}^{A_{I}A_{O}B_{I}}-W_{ddd^{2}}\\\
=&\mathds{1}^{A_{I}A_{O}B_{I}}-|{\phi^{+}}\rangle\langle{\phi^{+}}|^{A_{I}B_{I}^{1}}\otimes|{\mathds{1}}\rangle\rangle\langle\langle{\mathds{1}}|^{A_{O}/B_{I}^{2}}\end{split}$
(30)
and showing that $S$ is a valid non-classical CCDC witness, _i.e._ , every
CCDC process $W_{\text{CCDC}}$ satisfies $\mathrm{tr}(SW_{\text{CCDC}})\geq
0$. To ensure that $S$ is a witness, we start by pointing that
$\displaystyle\begin{split}\mathrm{tr}_{A_{O}}S&=d\mathds{1}^{A_{I}B_{I}}-|{\phi^{+}}\rangle\langle{\phi^{+}}|^{A_{I}B_{I}^{1}}\otimes\mathds{1}^{B_{I}^{2}}\\\
&\succeq 0,\end{split}$ (31)
where the last inequality holds because the smallest eigenvalue of
$|{\phi^{+}}\rangle\langle{\phi^{+}}|^{A_{I}B_{I}^{1}}\otimes\mathds{1}^{B_{I}^{2}}$
is $1$.
Now, note that
$\displaystyle\begin{split}S^{T_{A_{I}}}&=\mathds{1}-\sum_{ij}\frac{|{ij}\rangle\langle{ji}|}{d}^{A_{I}B_{I}^{1}}\otimes|{\mathds{1}}\rangle\rangle\langle\langle{\mathds{1}}|^{A_{O}/B_{I}^{2}}\\\
&=\mathds{1}-\frac{\textrm{U}_{\textrm{SWAP}}^{A_{I}B_{I}^{1}}}{d}\otimes|{\mathds{1}}\rangle\rangle\langle\langle{\mathds{1}}|^{A_{O}/B_{I}^{2}}\\\
&\succeq 0,\end{split}$ (32)
where the last inequality holds because the eigenvalues of
$\textrm{U}_{\textrm{SWAP}}$ are $+1$ or $-1$. Conditions (31) and (32)
together ensure that $S$ is a witness, as shown in Eqs. (22).
Direct calculation shows that $\mathrm{tr}(SW_{ddd^{2}})=d-d^{2}=d(1-d)$ and
that, for any bipartite ordered process $\Omega$, we have
$\displaystyle\mathrm{tr}(S\Omega)$
$\displaystyle=\mathrm{tr}(\Omega)-\mathrm{tr}(\Omega\,W_{ddd^{2}})$ (33a)
$\displaystyle\leq\mathrm{tr}(\Omega)=d.$ (33b)
Since $S$ is a non-classical CCDC witness, if $(1-r)W_{ddd^{2}}+r\Omega$ is a
CCDC process, it holds that
${\mathrm{tr}\left(S\left[(1-r)W_{ddd^{2}}+r\Omega\right]\right)\geq 0}$. By
combining $\mathrm{tr}(SW_{ddd^{2}})=d(1-d)$ with the inequality (33b), we
have
$\displaystyle(1-r)d(1-d)+rd\geq 0,$ (33c)
thus $r\geq 1-\frac{1}{d}$.
We finish the proof by invoking Theorem B.3, which states that $r\leq
1-\frac{1}{d}$. ∎
We now analyze the robustness of $W_{ddd^{2}}$ against the white noise
process. For that, we use the SDP formulation to numerically tackle the case
where $d=2$ and $d=3$. For the inner approximation, we have used the set
$\mathcal{L}_{\text{CCDC}}^{\text{in},\mathcal{E}_{N}}$ with
${\mathcal{E}_{N}}$ being a set with $N=10^{4}$ uniformly random pure states
for $d=2$ and $N=200$ uniformly random pure states for $d=3$. For the outer
approximation, it was enough to use the loose approximation
$\mathcal{L}_{\text{CCDC}}^{\text{out,PPT}}$. The results obtained were
$\displaystyle
R_{WN}^{\text{up},\mathcal{E}_{10^{4}}}(W_{222^{2}})=R_{WN}^{\textrm{low,PPT}}(W_{222^{2}})=0.8421$
(34a) $\displaystyle
R_{WN}^{\text{up},\mathcal{E}_{200}}(W_{333^{2}})=R_{WN}^{\textrm{low,PPT}}(W_{333^{2}})=0.9529,$
(34b)
where equations hold up to numerical precision. Since the upper and lower-
bound coincides, these values are the actual values of robustnesses.
For $d=2$, we believe that $W_{222^{2}}$ is the process with maximum white
noise robustness on its scenario. Our conjecture is based on a heuristic see-
saw technique inspired by [52, 53], which suggests that the highest value of
white noise robustness in the scenario where $d=2$ is $0.8421$.
In a nutshell, our see-saw technique goes as follows: for a fixed scenario
$d_{A_{I}}=d_{A_{O}}=d$ and $d_{B_{I}}=d^{2}$, we sample a random bipartite
ordered process $W$. Then, we obtain its optimal non-classical CCDC witness
$S$ from the dual formulation of the PPT white noise robustness (see Appendix
E). We then find the bipartite ordered process which maximally violates the
witness $S$, then finding its optimal witness after that. This process is re-
iterated until it converges to a robustness value, which we expect to be
considerably greater than the robustness of the initially sampled process $W$.
The see-saw method is described in details in Appendix F.
Moreover, in the scenario where $d=3$, our see-saw method could find a
bipartite ordered process $W_{\text{max}}$ which has
$R_{WN}^{\text{up},\mathcal{E}_{200}}(W_{\text{max}})=R_{WN}^{\text{low,PPT}}(W_{\text{max}})=0.9643>R_{WN}^{\text{up,}\mathcal{E}_{200}}(W_{333^{2}})$.
This result shows the power of the see-saw method in finding processes which
are robust against white noise. It also shows that, even though $W_{ddd^{2}}$
has maximum generalized robustness for every $d$, which we proved
analytically, in the case where $d=3$, $W_{333^{2}}$ is not the most robust
process against white noise. This leads us to conjecture that $W_{ddd^{2}}$
may be not the most robust process against white noise for $d>3$.
Before finishing this section we mention that, following the same steps from
the demonstration of Theorem 5.1, we can also obtain an analytical lower-bound
for white noise robustness of $W_{ddd^{2}}$, which is
$R_{WN}(W_{ddd^{2}})\geq\frac{d^{3}}{(d^{2}+1)(d+1)}.$ (35)
Differently from the lower-bound presented for generalized robustness, the
above inequality is not tight. For instance, when $d=2$, this lower-bound
provides $R_{WN}(W_{ddd^{2}})\geq\frac{8}{15}\approx 0.5333$, which is
considerably lower than $R_{WN}^{\text{low,PPT}}(W_{ddd^{2}})=0.8421$.
However, it is interesting to point out that the above expression shows that
$R_{WN}(W_{ddd^{2}})\rightarrow 1$ when $d\rightarrow\infty$.
### 5.2 The scenario with minimum dimensions:
$d_{A_{I}}=d_{A_{O}}=d_{B_{I}}=2$
Figure 8: Circuit representation of the process $W_{222}$, constructed with
the smallest possible space dimensions $d_{A_{I}}=d_{A_{O}}=d_{B_{I}}=2$. A
maximally entangled state is initially shared between Alice and the auxiliary
system. After Alice’s operation, a control-NOT is applied, then
$\textrm{aux}^{\prime}$ is discarded.
We now consider the scenario where $d_{A_{I}}=d_{A_{O}}=d_{B_{I}}=2$, that is,
the minimum non-trivial dimensions. For this scenario, we propose the process
composed by a two-qubit maximally entangled state between Alice and the
auxiliary system, and the decoder channel from Alice output to Bob’s input
being a control-NOT channel, where the auxiliary system is later discarded
(see Fig. 8). Mathematically, such process is described by
$\displaystyle\begin{split}W_{222}:=\
\mathrm{tr}_{\textrm{aux}^{\prime}}\Big{(}&|{\phi^{+}}\rangle\langle{\phi^{+}}|^{A_{I}\textrm{aux}}\\\
*&|{\textrm{U}_{\textrm{CNOT}}}\rangle\rangle\langle\langle{\textrm{U}_{\textrm{CNOT}}}|^{\textrm{aux}A_{O}/\textrm{aux}^{\prime}B_{I}}\Big{)}\end{split}$
(36)
where $|{\textrm{U}_{\textrm{CNOT}}}\rangle\rangle$ is the Choi vector of the
control-NOT unitary gate $\textrm{U}_{\textrm{CNOT}}$, given by
$\textrm{U}_{\textrm{CNOT}}=\left(\begin{matrix}1&0&0&0\\\ 0&1&0&0\\\
0&0&0&1\\\ 0&0&1&0\\\ \end{matrix}\right).$ (37)
Direct calculation shows that the process $W_{222}$ can also be written in
terms of a un-normalized $GHZ$ state
$|{GHZ}\rangle\rangle:=|{000}\rangle+|{111}\rangle\in\mathcal{L}(A_{I}\otimes
A_{O}\otimes B_{I})$ :
$W_{\textrm{222}}=\frac{1}{2}\left(|{GHZ}\rangle\rangle\langle\langle{GHZ}|+\sigma_{X}^{A_{O}}|{GHZ}\rangle\rangle\langle\langle{GHZ}|\sigma_{X}^{A_{O}}\right),$
(38)
where $\sigma_{X}^{A_{O}}$ is the Pauli matrix $\sigma_{X}$ applied on
subsystem $A_{O}$. The decomposition of Eq. (38) expresses $W_{222}$ as a
probabilistic mixture of pure “GHZ-like” non-normalized
states555Interestingly, one can verify that for every
$\epsilon\in\left(0,\frac{1}{2}\right]$, the operator
$W=\left(\frac{1}{2}+\epsilon\right)|{GHZ}\rangle\rangle\langle\langle{GHZ}|+\left(\frac{1}{2}-\epsilon\right)\sigma_{X}^{A_{O}}|{GHZ}\rangle\rangle\langle\langle{GHZ}|\sigma_{X}^{A_{O}}$
is outside $\mathcal{L}_{A\rightarrow B}$. This ensures that, as illustrated
in Fig. 9, $W_{222}$ is on the boundary of $\mathcal{L}_{A\rightarrow B}$. and
it will be useful to prove that $W_{222}$ attains the maximum generalized
robustness value for its scenario.
Another decomposition of the process $W_{222}$ is in terms of Pauli matrices:
$\displaystyle\begin{split}W_{222}=\frac{1}{4}&\bigg{(}\mathds{1}^{A_{I}A_{O}B_{I}}+\sigma_{Z}^{A_{I}}\mathds{1}^{A_{O}}\sigma_{Z}^{B_{I}}\\\
&+\sigma_{X}^{A_{I}}\sigma_{X}^{A_{O}}\sigma_{X}^{B_{I}}-\sigma_{Y}^{A_{I}}\sigma_{X}^{A_{O}}\sigma_{Y}^{B_{I}}\bigg{)},\end{split}$
(39)
with implicit tensor product between the operators.
Figure 9: The process $W_{222}$ seen as a convex combination of two GHZ-type
un-normalized states acting on $A_{I}$, $A_{O}$ and $B_{I}$. Due to the
ordered process causal constraints, GHZ states do not lead to valid bipartite
ordered processes, but such convex combination of these two GHZ-type results
in the valid bipartite ordered process $W_{222}$.
###### Theorem 5.2.
The bipartite ordered process $W_{222}$ attains the maximum generalized
robustness of all processes with the same dimensions. That is,
$\displaystyle R_{G}(W_{222})=$
$\displaystyle\max_{W\in\mathcal{L_{A\rightarrow B}}}\left[R_{G}(W)\right]$
(40a) $\displaystyle=$ $\displaystyle\frac{1}{2}.$ (40b)
###### Proof.
The proof of this theorem follows similar steps to the proof of theorem 5.1.
We start by defining the operator
$\displaystyle S:=$ $\displaystyle\mathds{1}-2W_{222}$ (41a) $\displaystyle=$
$\displaystyle\mathds{1}-\left(|{GHZ}\rangle\rangle\langle\langle{GHZ}|+\sigma_{x}^{A_{O}}|{GHZ}\rangle\rangle\langle\langle{GHZ}|\sigma_{x}^{A_{O}}\right),$
(41b)
and showing that $S$ is a non-classical CCDC witness. For this, we need to
show that $\mathrm{tr}_{A_{O}}(S)\succeq 0$ and $S^{T_{A_{I}}}\succeq 0$, as
shown in Eqs. (22).
$\displaystyle\begin{split}\frac{1}{2}\mathrm{tr}_{A_{O}}(S)&=\mathds{1}^{A_{I}B_{I}}-\left(|{00}\rangle\langle{00}|+|{11}\rangle\langle{11}|\right)\\\
&=|{01}\rangle\langle{01}|+|{10}\rangle\langle{10}|\\\ &\succeq 0.\end{split}$
(42)
Now, for the second condition, we have
$\displaystyle\begin{split}S^{T_{A_{I}}}&=\mathds{1}-\left(|{GHZ}\rangle\rangle\langle\langle{GHZ}|+\sigma_{x}^{A_{O}}|{GHZ}\rangle\rangle\langle\langle{GHZ}|\sigma_{x}^{A_{O}}\right)^{T_{A_{I}}}\\\
&=\mathds{1}-\sum_{ij}\left(|{jii}\rangle\langle{ijj}|+\sigma_{x}^{A_{O}}|{jii}\rangle\langle{ijj}|\sigma_{x}^{A_{O}}\right)\\\
&\succeq 0,\end{split}$ (43)
where the last inequality holds true because
$|{GHZ}\rangle\rangle\langle\langle{GHZ}|^{T_{A_{I}}}$ and
$\left(\sigma_{x}^{A_{O}}|{GHZ}\rangle\rangle\langle\langle{GHZ}|\sigma_{x}^{A_{O}}\right)^{T_{A_{I}}}$
have orthogonal support, i.e.,
$|{GHZ}\rangle\rangle\langle\langle{GHZ}|^{T_{A_{I}}}\left(\sigma_{x}^{A_{O}}|{GHZ}\rangle\rangle\langle\langle{GHZ}|\sigma_{x}^{A_{O}}\right)^{T_{A_{I}}}=0,$
(44)
and the eigenvalues of $\sum_{ij}|{jii}\rangle\langle{ijj}|$ and
$\sum_{ij}\sigma_{x}^{A_{O}}|{jii}\rangle\langle{ijj}|\sigma_{x}^{A_{O}}$ are
$+1$ and $-1$.
Direct calculation shows that $\mathrm{tr}(SW_{222})=-2$ and, for every
process $\Omega$ in this scenario, we have $\mathrm{tr}(S\Omega)\leq 2$, by an
argument analogous to Eq. (33b).
Since $S$ is a non-classical CCDC witness, if $(1-r)W_{222}+r\Omega$ is a CCDC
process, it holds that
${\mathrm{tr}\left(S\left[(1-r)W_{222}+r\Omega\right]\right)\geq 0}$. It is
also true that
$-2r+2(1-r)\geq 0,$ (45)
thus $r\geq\frac{1}{2}$.
We finish the proof by invoking Theorem B.3, which states that
$r\leq\frac{1}{2}$ in this scenario. ∎
We also evaluated the robustness of $W_{222}$ against white noise, obtaining
evidences that $W_{222}$ attains the maximal white noise robustness on its
scenario.
###### Theorem 5.3.
The white noise robustness of the bipartite ordered processes $W_{222}$ is
$R_{WN}(W_{222})=\frac{2}{3}.$ (46)
###### Proof.
We provide a lower-bound for $R_{WN}(W_{222})$ by using similar steps to the
proof of theorem 5.2, starting with the non-classical CCDC witness
$\displaystyle\begin{split}S:=&\mathds{1}-2W_{222}\\\
=&\mathds{1}-\left(|{GHZ}\rangle\rangle\langle\langle{GHZ}|+\sigma_{X}^{A_{O}}|{GHZ}\rangle\rangle\langle\langle{GHZ}|\sigma_{X}^{A_{O}}\right).\end{split}$
(47)
We have $\mathrm{tr}(SW_{222})=-2$ and $\mathrm{tr}(S)=2^{3}-4=4$. As $S$ is a
non-classical CCDC witness, if $(1-r)W_{222}+r\Omega$ is a CCDC process,
${\mathrm{tr}\left(S\left[(1-r)W_{222}+r\frac{\mathds{1}}{4}\right]\right)\geq
0}$. So, it is true that
$\displaystyle-2(1-r)+r\geq 0,$ (48)
thus $r\geq\frac{2}{3}$.
We now show, using techniques which are similar to the proof of Theorem B.3,
that the above lower-bound can be attained. First notice that the qubit
depolarizing channel
$\widetilde{D}_{\eta}(\rho)=(1-\eta)\rho+\eta\mathrm{tr}(\rho)\frac{\mathds{1}}{2}$
is entanglement breaking when $\eta=\frac{2}{3}$ [54, 55]. Also, notice that,
since
$\mathrm{tr}_{A_{I}}(W_{222})=\frac{\mathds{1}^{A_{O}B_{I}}}{d_{B_{I}}}$, if
we apply the depolarizing channel on the subspace $A_{I}$ of the process
$W_{222}$, we obtain
$\widetilde{D}_{\eta}^{A_{I}}\otimes\widetilde{\mathds{1}}^{A_{O}B_{I}}(W_{222})=(1-\eta)W_{222}+\eta\frac{\mathds{1}}{4}$.
Since $\widetilde{D}_{\eta}$ is entanglement breaking for $\eta=\frac{2}{3}$,
lemma B.1 ensures that
$(1-\frac{2}{3})W_{222}+\frac{2}{3}\frac{\mathds{1}}{4}$ is CCDC, thus
concluding the proof. ∎
It is worth to mention that every witness introduced in the proofs of Theorems
5.1 and 5.2 are the optimal witnesses for these particular processes.
Similarly to the scenario with $d_{A_{I}}=d_{A_{O}}=d$ and $d_{B_{I}}=d^{2}$,
we implemented the see-saw algorithm that indicates that, in the scenario with
$d_{A_{I}}=d_{A_{O}}=d_{B_{I}}=2$, the highest white noise robustness is
exactly $\frac{2}{3}$. This suggests that $W_{222}$ has also maximum white
noise robustness in its scenario.
## 6 Relation with previous research
### 6.1 Comparison with the non-classical CCDC process of Ref. [15]
We now compare our proposed process with the non-classical CCDC process
presented and experimentally implemented in Ref. [15]. The process consists of
the preparation of a maximally entangled state $|{\phi^{+}}\rangle$ shared
between Alice and the auxiliary system, a partial swap channel from
$\mathcal{L}(A_{O}\otimes\textrm{aux})$ to
$\mathcal{L}(B_{I}\otimes\textrm{aux}^{\prime})$, which is simply a unitary
composed by a coherent mixture of an identity and a SWAP gate, and a partial
trace on the output of the auxiliary system. This process can be explicitly
written as
$\displaystyle\begin{split}W_{\textrm{MRSR}}=\
\mathrm{tr}_{\textrm{aux}^{\prime}}&\Big{(}|{\phi^{+}}\rangle\langle{\phi^{+}}|^{A_{I}\textrm{aux}}\\\
*&|{\textrm{U}_{\textrm{PS}}}\rangle\rangle\langle\langle{\textrm{U}_{\textrm{PS}}}|^{A_{O}\textrm{aux}/B_{I}\textrm{aux}^{\prime}}\Big{)},\end{split}$
(49)
where
$|{\textrm{U}_{\textrm{PS}}}\rangle\rangle^{A_{O}\textrm{aux}/B_{I}\textrm{aux}^{\prime}}$
is the Choi vector of the unitary partial SWAP666 The authors from Ref. [15]
use an equivalent way to represent the unitary partial SWAP gate, which is
$\textrm{U}_{\textrm{PS}}^{A_{I}\textrm{aux}/B_{I}\textrm{aux}^{\prime}}=\frac{1}{\sqrt{2}}\left(\mathds{1}^{A_{O}\textrm{aux}/B_{I}\textrm{aux}^{\prime}}+i\mathds{1}^{A_{O}\textrm{aux}/\textrm{aux}^{\prime}B_{I}}\right).$
(50) In the second term, the identity channel exchanges the outputs $B_{I}$
and $\textrm{aux}^{\prime}$, in comparison to the identity channel in the
first term. This part of the channel does exactly the same as
$\textrm{U}_{\textrm{SWAP}}$ does, differing only by the fact that the output
labels are not explicitly exchanged. gate $\textrm{U}_{\textrm{PS}}$, given by
$\textrm{U}_{\textrm{PS}}=\frac{1}{\sqrt{2}}\left(\mathds{1}+i\,\textrm{U}_{\textrm{SWAP}}\right),$
(51)
with $\textrm{U}_{\textrm{SWAP}}$ being the SWAP gate for qubits, given by
$\textrm{U}_{\textrm{SWAP}}=\left(\begin{matrix}1&0&0&0\\\ 0&0&1&0\\\
0&1&0&0\\\ 0&0&0&1\end{matrix}\right).$ (52)
The label MRSR makes reference to the names MacLean, Ried, Spekkens and Resch,
authors of Ref. [15]. Fig. 10 ilustrates the process $W_{\text{MRSR}}$.
Figure 10: Circuit representation of the process $W_{\text{MRSR}}$ presented
in Ref. [15]. A maximally entangled state is initially shared between Alice
and the auxiliary system. After Alice’s operation, the partial-SWAP gate
$\text{U}_{\text{PS}}$ is applied, then $\textrm{aux}^{\prime}$ is discarded.
Using our numerical methods, we can evaluate the values of robustnesses for
$W_{\textrm{MRSR}}$, which are
$\displaystyle
R_{G}^{\textrm{low,PPT}}(W_{\textrm{MRSR}})=R_{G}^{\text{up},\mathcal{E}_{10^{4}}}(W_{\textrm{MRSR}})=0.3506,$
(53a) $\displaystyle
R_{WN}^{\textrm{low,PPT}}(W_{\textrm{MRSR}})=R_{WN}^{\text{up},\mathcal{E}_{10^{4}}}(W_{\textrm{MRSR}})=0.5000.$
(53b)
When comparing the robustness values of $W_{\textrm{MRSR}}$ with
$W_{\textrm{222}}$ (see Section 5.2 and Eqs. (53)), which is a process defined
in the same scenario, we verify that $W_{\textrm{222}}$ is strictly more
robust against both generalized and white noise than $W_{\text{MRSR}}$.
For the case of the process $W_{222^{2}}$, we argue that the construction of
$W_{222^{2}}$ is simpler than $W_{\textrm{MRSR}}$. Both processes require the
preparation of a maximally entangled qubit state, but $W_{\textrm{MRSR}}$
requires the implementation of the control swap operation, which is a coherent
superposition between the identity channel and the swap channel, while
$W_{222^{2}}$ only requires the identity channel.
### 6.2 Comparison with the non-classical CCDC process of Ref. [19]
In Ref. [19], the authors proposed to study non-classical CCDC by considering
quantum superpositions of both relations. Their example is a the tripartite
process ordered as $A\rightarrow B\rightarrow C$. A tripartite process ordered
as $A\rightarrow B\rightarrow C$ is an operator $W\in\mathcal{L}(A_{I}\otimes
A_{O}\otimes B_{I}\otimes B_{O}\otimes C_{I})$ which can be written as
$W=\rho^{A_{I}\text{aux}}*D_{A}^{\text{aux}A_{O}/B_{I}\text{aux'}}*D_{B}^{\text{aux'}B_{O}/C_{I}},$
(54)
where $D_{A}$ and $D_{B}$ are Choi operators of quantum channels.
In a tripartite scenario, common-cause and direct-cause relations can be more
complex than in the bipartite scenario. In the case of common-cause relations,
more general situations are discussed in Refs. [56, 57]. However, following
the definitions presented in Ref. [19], we restrict the attention to the same
common-cause and direct-cause relations of the bipartite case, recovering them
when taking $d_{B_{O}}{=}d_{C_{I}}{=}1$. The following definitions are taken
considering the ones presented in Ref. [19].
A tripartite ordered process $W_{\text{CC}}$ is common-cause if
$\mathrm{tr}_{B_{O}C_{I}}(W_{\text{CC}}):=d_{B_{O}}\rho^{A_{I}B_{I}}\otimes\mathds{1}^{A_{O}},$
(55)
where $\rho^{A_{I}B_{I}}$ is a quantum state in $\mathcal{L}(A_{I}\otimes
B_{I})$. A tripartite ordered process $W_{\text{DC}}$ is direct-cause if
$\mathrm{tr}_{B_{O}C_{I}}(W_{\text{DC}}):=d_{B_{O}}\sum_{i}p_{i}\rho_{i}^{A_{I}}\otimes
D_{i}^{A_{O}/B_{I}},$ (56)
where $\rho_{i}^{A_{I}}$ are quantum states in $\mathcal{L}(A_{I})$ and
$D_{i}^{A_{O}/B_{I}}$ are quantum channels from $A_{O}$ to $B_{I}$.
A tripartite ordered process $W_{\text{CCDC}}$ is classical CCDC if it can be
decomposed in a convex combination of a common-cause process and a direct-
cause process. Note that when bipartite processes are considered, _i.e._ , the
dimensions of $B_{O}$ and $C_{I}$ are equal to one, their definition are
equivalent to the ones presented in section 2.
The example of non-classical CCDC process presented by the authors is the
operator777The state $|{\phi^{+}}\rangle$ appearing in the expression of
$W_{\text{FB}}$ is originally written, in Ref. [19], as generic bipartite
state $|{\psi}\rangle$. However, we verified that the numerical robustness
results obtained in Ref. [19] are reproduced when $|{\psi}\rangle$ is a
maximally entangled state, which lead us to take
$|{\psi}\rangle=|{\phi^{+}}\rangle$, as this state is used in every other
process mentioned before in this work.
$W_{\text{FB}}=|{W_{\text{FB}}}\rangle\rangle\langle\langle{W_{\text{FB}}}|$,
where
$\displaystyle\begin{split}|{W_{\text{FB}}}\rangle\rangle=\frac{1}{\sqrt{2}}\Big{(}&|{\phi^{+}}\rangle^{A_{I}B_{I}}|{\mathds{1}}\rangle\rangle^{A_{O}/C_{I}^{1}}|{\mathds{1}}\rangle\rangle^{B_{O}/C_{I}^{2}}|{0}\rangle^{C_{I}^{3}}\\\
+&|{\phi^{+}}\rangle^{A_{I}C_{I}^{1}}|{\mathds{1}}\rangle\rangle^{A_{O}/B_{I}}|{\mathds{1}}\rangle\rangle^{B_{O}/C_{I}^{2}}|{1}\rangle^{C_{I}^{3}}\Big{)},\end{split}$
(57)
The label FB makes reference to the names Feix and Brukner, authors of Ref.
[19].
This process can be seen as a superposition of a common-cause and a direct-
cause processes, as the first term corresponds to a common-cause process,
whereas the second term corresponds to a direct-cause process. We can also
represent $W_{\text{FB}}$ in terms of ordered quantum circuits888Indeed, every
ordered quantum process can be represented in terms of ordered quantum
circuits by concatenating quantum states and quantum operations [11]. as
illustrated in Fig. 11.
Figure 11: Circuit representation of the process $W_{\text{FB}}$ presented in
Ref. [19]. A fixed qubit
$|{+}\rangle=\frac{1}{\sqrt{2}}(|{0}\rangle+|{1}\rangle)$ in $C_{I}^{3}$
controls the swap operation from the auxiliary system in $C_{I}^{2}$ and
$A_{O}$ to $B_{I}$ and from $C_{I}^{2}$ and $B_{O}$ to $C_{I}^{1}$. As the
control qubit is fixed, the swapping operation occurs in the same way in both
channels. This combination of operations generate the pure non-classical CCDC
process $W_{\text{FB}}$
In Ref. [19], the authors numerically obtained a lower-bound to the
generalized robustness of $W_{\text{FB}}$, using the PPT outer approximation
of DC processes, obtaining999Strictly speaking, the authors have evaluated
quantity before mentioned here, the non-classicality of causality
$\mathcal{C}$, which has a one-to-one relation with the generalized robustness
via $R_{G}^{\text{low,PPT}}(W)=\frac{\mathcal{C}(W)}{1+\mathcal{C}(W)}$.
$R_{G}^{\text{low,PPT}}(W_{\text{FB}})=0.1855.$ (58)
Using our inner approximation method with $N=200$, we could obtain, up to
numerical precision, the upper-bound
$R_{G}^{\text{up,}\mathcal{E}_{200}}(W_{\text{FB}})=0.1855,$ (59)
showing that $R_{G}(W_{\text{FB}})=0.1855$, up to numerical precision. For
completeness, we have also computed the white noise robustness, obtaining
$R_{WN}^{\text{low,PPT}}(W_{\text{FB}})=R_{WN}^{\text{up,}\mathcal{E}_{200}}(W_{\text{FB}})=0.3324.$
(60)
### 6.3 The role of coherent mixture of causal relations
As discussed in this section, previous research has shown that one way to
obtain processes with the non-classical CCDC property is by coherently
superposing causal relations. For instance, the tripartite processes presented
in Ref. [19] and discussed in Section 6.2 is constructed in a way to be a
coherent superposition of a purely common-cause and a purely direct-cause
processes. Also, the bipartite process presented in Ref. [15] is inspired by a
coherent mixture of causal relations which is mathematically formalized by the
application of the partial swap operation (see Eq. (50)).
Differently from previous works, we have shown in this work that the
connection between non-classical CCDC and coherent superpositions of causal
relations may be more subtle than it seems at first glance. In particular,
although the non-classical CCDC process $W_{ddd^{2}}$ presented in section 5
may be viewed as a superposition of a process which is initialized in
$|{00}\rangle$ with a process which is initialized in state $|{11}\rangle$, it
also admits a natural interpretation as a process with both common-cause and
direct-cause relations simultaneously, without explicitly considering any
superposition of causal relations. We then argue that the process
$W_{ddd^{2}}$ does not need to be interpreted as a coherent superposition of
causal relations.
## 7 Separable process without a direct-cause explanation
As mentioned in previous sections, the definition of direct-cause processes
(Def.2.2) reminds one of the definition of separable quantum states. More
precisely, if we do not impose that the operators $D_{i}^{A_{O}/B_{I}}$ have
to respect the quantum channel condition
$\mathrm{tr}_{B_{I}}D_{i}^{A_{O}/B_{I}}=\mathds{1}^{A_{O}}$, equation (9b) is
precisely the definition of a separable state on the bipartition
$A_{I}|A_{O}B_{I}$. We now show that, despite being related, these two
definitions are not equivalent.
Consider the process
$\displaystyle\begin{split}W_{\text{SEP}}:=\frac{1}{2}\Big{(}&|{0}\rangle\langle{0}|^{A_{I}}\otimes|{0}\rangle\langle{0}|^{A_{O}}\otimes|{0}\rangle\langle{0}|^{B_{I}}\\\
+&|{1}\rangle\langle{1}|^{A_{I}}\otimes|{0}\rangle\langle{0}|^{A_{O}}\otimes|{1}\rangle\langle{1}|^{B_{I}}\\\
+&|{+}\rangle\langle{+}|^{A_{I}}\otimes|{1}\rangle\langle{1}|^{A_{O}}\otimes|{+}\rangle\langle{+}|^{B_{I}}\\\
+&|{-}\rangle\langle{-}|^{A_{I}}\otimes|{1}\rangle\langle{1}|^{A_{O}}\otimes|{-}\rangle\langle{-}|^{B_{I}}\Big{)},\end{split}$
(61)
which is a separable operator by construction. First, notice that
$\displaystyle\mathrm{tr}_{B_{I}}W_{\text{SEP}}=\frac{1}{2}\mathds{1}^{A_{I}}\otimes\mathds{1}^{A_{O}},$
(62)
which shows that $W$ is a valid bipartite ordered process.
The process $W_{\text{SEP}}$ can be physically realized by preparing a
maximally entangled qubit state between $A_{I}$ and aux, and using a “decoder”
described by
$\displaystyle D^{A_{O}\textrm{aux}/B_{I}}$
$\displaystyle=|{0}\rangle\langle{0}|^{A_{O}}\otimes\left(|{00}\rangle\langle{00}|+|{11}\rangle\langle{11}|\right)^{\textrm{aux}B_{I}}$
$\displaystyle+|{1}\rangle\langle{1}|^{A_{O}}\otimes\left(|{++}\rangle\langle{++}|+|{--}\rangle\langle{--}|\right)^{\textrm{aux}B_{I}}.$
(63)
In this way we have
$\displaystyle
W_{\text{SEP}}=|{\phi^{+}}\rangle\langle{\phi^{+}}|^{A_{I}\textrm{aux}}*D^{A_{O}\textrm{aux}/B_{I}}.$
(64)
Note that the channel $D^{A_{O}\textrm{aux}/B_{I}}$ can be implemented as
follows: first, perform a computational basis measurement on $A_{O}$. If the
outcome is $0$, perform a computational basis measurement on aux and send the
output qubit to $B_{I}$. If the outcome is $1$, perform a measurement on aux
in the $X$-basis instead.
###### Theorem 7.1.
The bipartite ordered process $W_{\text{SEP}}$ is separable in the bipartition
$A_{I}|A_{O}B_{I}$, but is not direct-cause.
###### Proof.
Equation (61) represents $W_{\text{SEP}}$ as a convex combination of product
states, ensuring that $W_{\text{SEP}}$ is separable in the bipartition
$A_{I}|A_{O}B_{I}$. In order to show that $W_{\text{SEP}}$ is not a direct-
cause process, let us assume that $W_{\text{SEP}}$ can be written as a convex
combination $W_{\text{SEP}}=\sum_{i}p_{i}\rho_{i}^{A_{I}}\otimes
D_{i}^{A_{O}/B_{I}}$, where $\rho_{i}^{A_{I}}$ are normalized quantum states
and every $D_{i}^{A_{O}/B_{I}}$ satisfy
$\mathrm{tr}_{A_{I}}D_{i}^{A_{O}/B_{I}}=\mathds{1}^{A_{O}}$. Note that each
$\rho_{i}^{A_{I}}$ has non-trivial overlap with, at least, 3 out of the 4
states $|{0}\rangle\langle{0}|$, $|{1}\rangle\langle{1}|$,
$|{+}\rangle\langle{+}|$,$|{-}\rangle\langle{-}|$. Indeed, let
$\rho=\frac{1}{2}\left(\mathds{1}+\sum_{i}\alpha_{i}\sigma_{i}\right)$ be an
arbitrary state where $\\{\alpha_{i}\\}$ are real numbers that satisfy
$\sum_{i}\alpha_{i}^{2}\leq 1$ and $\sigma_{i}$ are Pauli matrices. Suppose
that $\rho$ has zero overlap with some pure state
$|{\psi}\rangle\langle{\psi}|=\frac{1}{2}\left(\mathds{1}+\sum_{i}\beta_{i}\sigma_{i}\right)$
with $\sum_{i}\beta_{i}^{2}=1$, then we have that
$(\vec{\alpha},\vec{\beta})=-1$, where $(\cdot,\cdot)$ is the Euclidean inner
product. By the Cauchy-Schwarz inequality, we have
$1=(\vec{\alpha},\vec{\beta})^{2}\leq(\vec{\alpha},\vec{\alpha})(\vec{\beta},\vec{\beta})=(\vec{\alpha},\vec{\alpha}),$
(65)
with equality if and only if $\vec{\alpha}$ is a multiple of $\vec{\beta}$.
This shows that $\vec{\alpha}=-\vec{\beta}$ and therefore $\rho$ cannot be
orthogonal to any other pure quantum state.
Let us choose some fixed index $j$ in the sum and suppose, without lack of
generality, that $\rho_{j}$ has non-zero overlap with
$|{1}\rangle,|{+}\rangle,|{-}\rangle$. Then, from the above definition of
$W_{\text{SEP}}$, we calculate
$\displaystyle\mathrm{tr}(|{1}\rangle\langle{1}|^{A_{I}}\otimes|{0}\rangle\langle{0}|^{A_{O}}\otimes|{0}\rangle\langle{0}|^{B_{I}}W_{\text{SEP}})=0,$
(66)
From the decomposition $W_{\text{SEP}}=\sum_{i}p_{i}\rho_{i}^{A_{I}}\otimes
D_{i}^{A_{O}/B_{I}}$, we get
$\sum_{i}p_{i}\mathrm{tr}(|{1}\rangle\langle{1}|\rho_{i}^{A_{I}})\mathrm{tr}(|{0}\rangle\langle{0}|^{A_{O}}\otimes|{0}\rangle\langle{0}|^{B_{I}}D_{i}^{A_{O}/B_{I}})=0.$
(67)
By positivity of $D^{A_{O}/B_{I}}$ and $\rho^{A_{I}}$, each term in the sum
has to be zero. Since, by assumption,
$\mathrm{tr}(|{1}\rangle\langle{1}|\rho_{j})\neq 0$, it must be the case that
$D_{j}$ obeys
$\mathrm{tr}(|{0}\rangle\langle{0}|^{A_{O}}\otimes|{0}\rangle\langle{0}|^{B_{I}}D_{j}^{A_{O}/B_{I}})=0.$
(68)
Similarly, by calculating other projectors we get
$\displaystyle\mathrm{tr}(|{1}\rangle\langle{1}|^{A_{O}}\otimes|{-}\rangle\langle{-}|^{B_{I}}D_{j}^{A_{O}/B_{I}})=0,$
(69a)
$\displaystyle\mathrm{tr}(|{1}\rangle\langle{1}|^{A_{O}}\otimes|{+}\rangle\langle{+}|^{B_{I}}D_{j}^{A_{O}/B_{I}})=0.$
(69b)
From this we get
$\mathrm{tr}(|{1}\rangle\langle{1}|^{A_{O}}\otimes\mathds{1}^{B_{I}}D_{j}^{A_{O}/B_{I}})=0$,
which means that
$\mathrm{tr}_{B_{I}}D_{j}^{A_{O}/B_{I}}\neq\mathds{1}^{A_{O}}$. ∎
By construction, $W_{\text{SEP}}$ is separable in the bipartition
$A_{I}|A_{O}B_{I}$, implying that $W_{\text{SEP}}$ has a PPT $k$-symmetric
extension for every $k\in\mathbb{N}$ and $R_{G}^{\text{low,
PPT}_{k}}(W)=R_{WN}^{\text{low, PPT}_{k}}(W)=0$. This means that the non-
classical CCDC property of $W_{\text{SEP}}$ cannot be certified by any purely
entanglement-based criterion, such as the ones explored in Refs.[15, 19, 20].
Using the inner and outer approximations presented in Secs. 4.1 and 4.2 with
the family of qubits presented in the Appendix B of Ref.[50] ($n=171$, which
corresponds to $N=2*171^{2}$ states), we obtain upper and lower bounds for the
generalized and white noise robustnesses of $W_{\text{SEP}}$
$\displaystyle
R_{G}^{\text{up},\mathcal{E}_{N}}(W_{\text{SEP}})=R_{G}^{\text{low},\hat{\mathcal{E}}_{N}}(W_{\text{SEP}})=0.1465,$
(70a) $\displaystyle
R_{WN}^{\text{up},\mathcal{E}_{N}}(W_{\text{SEP}})=R_{WN}^{\text{low},\hat{\mathcal{E}}_{N}}(W_{\text{SEP}})=0.2930,$
(70b)
with equality holding up to numerical precision.
In Ref. [20], it was conjectured that the set of separable processes and the
set of processes without quantum memory are not the same, the latter being a
strict subset of the first. Since, the definitions of bipartite processes
without quantum memory and bipartite direct-cause processes are equivalent
(see Appendix A), we have then proven the conjecture presented in Ref. [20] by
explicitly constructing the bipartite separable process $W_{\text{SEP}}$,
which cannot be realized by processes without quantum memory.
It is interesting to point that our numerical methods allowed us to obtain a
relatively high robustness of $R_{WN}(W_{\text{SEP}})=0.2930$, but techniques
exclusively based on entanglement would lead to the trivial lower bound
$R^{\text{sep}}_{WN}(W_{\text{SEP}})\geq 0$. Since
$R_{WN}(W_{\text{SEP}})=0.2930$ is considerably greater than zero, we see that
the approximating the set of direct-cause processes by separable processes may
lead to very unsatisfactory results.
## 8 Certifying non-classical CCDC on PPT processes
In this section, we present an example of a non-classical CCDC process which
has $R_{G}^{\text{low,PPT}}(W)=R_{WN}^{\text{low,PPT}}(W)=0$, _i.e._ , its
non-classical CCDC property cannot be certified by the PPT approximation used
in Refs. [15, 19]. Such a process can be obtained by exploiting a class of
entangled states with positive partial transpose, presented in Ref. [58]. The
class of states of our interest is
$\displaystyle\begin{split}&\rho_{a}^{2\times 4}:=\\\
&\frac{1}{7a+1}\begin{bmatrix}a&0&0&0&0&a&0&0\\\ 0&a&0&0&0&0&a&0\\\
0&0&a&0&0&0&0&a\\\ 0&0&0&a&0&0&0&0\\\
0&0&0&0&\tfrac{1}{2}(1+a)&0&0&\tfrac{1}{2}\sqrt{1-a^{2}}\\\ a&0&0&0&0&a&0&0\\\
0&a&0&0&0&0&a&0\\\
0&0&a&0&\tfrac{1}{2}\sqrt{1-a^{2}}&0&0&\tfrac{1}{2}(1+a)\end{bmatrix},\end{split}$
(71)
being entangled for $a\in(0,1)$ and separable for $a=0$ or $a=1$.
We now set $a=\frac{1}{2}$ and use $\rho_{\frac{1}{2}}^{2\times 4}$ to define:
$W_{\textrm{PPT}}:=d_{A_{O}}\cdot\rho_{\frac{1}{2}}^{2\times 4},$ (72)
with $d_{A_{O}}=2$. $W_{\textrm{PPT}}$ is a valid bipartite ordered process,
as it satisfies every condition from Eqs. (18) by direct inspection. Also, it
has dimensions $d_{A_{I}}{=}d_{A_{O}}{=}d_{B_{I}}=2$ and has bound
entanglement in the bipartition $A_{I}|A_{O}B_{I}$.
‘We will now ensure that $W_{\textrm{PPT}}$ is a non-classical CCDC process by
using a better approximation $\mathcal{L}_{\text{DC}}^{\text{out,PPT}_{k}}$.
In particular, we set $k=2$, then we obtain
$\displaystyle\begin{split}R_{G}^{\text{low,PPT}_{2}}(W_{\textrm{PPT}})&=0.0083,\\\
R_{WN}^{\text{low,PPT}_{2}}(W_{\textrm{PPT}})&=0.0230.\end{split}$ (73)
For $k=3$, the robustnesses do not change, which indicates that using greater
values of $k$ does not improve the values of generalized and white noise
robustnesses.
When using the inner and outer approximations presented in Secs. 4.1 and 4.2
with the family of qubits presented in the Appendix B of Ref.[50] ($n=171$,
which corresponds to $N=2*171^{2}$ states), up to numerical precision, we
obtain
$\begin{split}R_{G}^{\text{up},\mathcal{E}_{10^{4}}}(W_{\textrm{PPT}})&=R_{G}^{\text{low},\hat{\mathcal{E}}_{10^{4}}}(W_{\textrm{PPT}})=0.1085,\\\
R_{WN}^{\text{up},\mathcal{E}_{10^{4}}}(W_{\textrm{PPT}})&=R_{WN}^{\text{low},\hat{\mathcal{E}}_{10^{4}}}(W_{\textrm{PPT}})=0.2782.\end{split}$
(74)
We verify that the lower bounds for the robustnesses obtained with the
entanglement criterium in Eqs. (73) are rather loose in comparison to the
actual robustnesses values from Eqs. (74). This example also illustrates the
limitations of certifying non-classical CCDC solely based on entanglement
criteria.
## 9 Summary of generalized and white noise robustnesses
Process (Eq.) | $R_{G}$ | $R_{G}^{\text{low, PPT}_{k=2}}$ | $R_{WN}$ | $R_{WN}^{\text{low, PPT}_{k=2}}$
---|---|---|---|---
$W_{333^{2}}$ (27) | $\frac{2}{3}$ | $\frac{2}{3}$ | $0.9529$ | $0.9529$
$W_{222^{2}}$ (27) | $\frac{1}{2}$ | $\frac{1}{2}$ | $0.8421$ | $0.8421$
$W_{222}$ (36) | $\frac{1}{2}$ | $\frac{1}{2}$ | $\frac{2}{3}$ | $\frac{2}{3}$
$W_{\text{MRSR}}$ (49) | $0.3506$ | $0.3506$ | $0.5000$ | $0.5000$
$W_{\text{FB}}$ (57) | $0.1855$ | $0.1855$ | $0.3324$ | $0.3324$
$W_{\text{SEP}}$ (61) | $0.1465$ | $0$ | $0.2930$ | $0$
$W_{\text{PPT}}$ (72) | $0.1085$ | $0.0083$ | $0.2782$ | $0.0230$
Table 2: Table presenting generalized and white noise robustnesses for every
process analyzed in this work. Values represented in fractions were obtained
by mathematical theorems and coincide with SDP optimization. Values with
decimal digits were obtained only via SDP optimization, where our upper and
lower bounds are identical up to $4$ decimals. The lower bounds obtained by
the approximations of the CCDC set based on entanglement are the values
provided in columns $R_{G}^{\text{low, PPT}_{k}}$ and $R_{WN}^{\text{low,
PPT}_{k}}$. Lower-bounds that match the actual robustnesses are highlighted in
green, while lower bounds that are rather different from the actual
robustnesses are highlighted in red. We can observe that entanglement based
criteria could never detect $W_{\text{SEP}}$ as a non-classical CCDC process,
while the PPT $k$-symmetric extension bound for $W_{\text{PPT}}$ provides the
loose lower bounds $R_{G}^{\text{low, PPT}_{k}}(W_{\text{PPT}})\geq 0.0083$
$R_{WN}^{\text{low, PPT}_{k}}(W_{\text{PPT}})\geq 0.0230$, which are values
obtained both with $k=2$ and $k=3$.
In previous sections, we presented several examples of non-classical CCDC
processes with different values of generalized and white noise robustness.
Table 2 summarizes the non-classical CCDC robustnesses of several processes
presented in this work and compare the actual value of robustness with the
values obtained with methods based on entanglement criteria.
## 10 Conclusions
In this work, we have introduced a class of bipartite ordered processes, and a
process with dimension of three qubits, which are maximally robust against
general noise and very likely to be the most robust against white noise for
the qubit case. This class of processes can be implemented by preparing a pair
of maximally entangled states and an identity channel, admitting a natural
interpretation of a process with both common-cause and direct-cause relations
simultaneously. Hence, in contrast to previously known non-classical CCDC
processes [14, 19], the class presented here does not require either the
construction or the interpretation directly based on coherent superposition of
causal relations.
Several analytical results proved in this work employed general convex
analysis arguments based on witness hyper-planes, combined with entanglement
theory concepts, such as entanglement breaking channels. We believe that the
techniques developed here may find applications in related problems.
We have also presented a systematic semi-definite approach to characterize the
set of non-classical CCDC processes. In particular, we provided a hierarchy of
inner and outer approximations that converge to the set of classical CCDC
processes, and an entanglement-based hierarchy which, despite not converging
to the set of CCDC processes, provides us several tight and non-trivial
bounds.
In order to tackle situations where we could not prove the value of the
highest robustness of a given scenario analytically, we constructed a
heuristic see-saw method, which provided numerical evidence of the highest
robustnesses attainable on such scenario, also providing a valid lower-bound
for the value of the highest robustness.
Finally, we have shown that, although all bipartite processes that are
entangled in the bipartition $A_{I}|A_{O}B_{I}$ do not have a direct-cause
decomposition, the converse does not hold. Our proof consists in explicitly
constructing a process which is separable on the bipartition
$A_{I}|A_{O}B_{I}$, but does not have a direct-cause decomposition. Since
bipartite processes without quantum memory are equivalent to bipartite direct-
cause ones, our results prove a conjecture first raised in Ref. [20] and
contributes towards a better understanding of the particularities of quantum
memory, spacial entanglement and temporal entanglement [21, 20, 22, 23, 24].
All SDP optimization problems presented in this manuscript were implemented
using MATLAB, the convex optimization package Yalmip [59] and CVX [60], the
solvers MOSEK, SeDuMi and SDPT3 [61, 62, 63], and the toolbox for quantum
information QETLAB [64]. All our codes are available in the public repository
[65] and can be freely used under the GNU Lesser General Public License v3.0.
We are grateful to Mateus Araújo, Jessica Bavaresco, Rafael Chaves, Simon Milz
and Philip Taranto for interesting discussions. MN, RV and TM acknowledge
financial support by the Brazilian agencies INCT-IQ (National Institute of
Science and Technology for Quantum Information), FAPEMIG, CNPq, CAPES and
Instituto Serrapilheira. MTQ has received funding from the European Union’s
Horizon 2020 research and innovation programme under the Marie Skłodowska-
Curie grant agreement No 801110 and the Austrian Federal Ministry of
Education, Science and Research (BMBWF). It reflects only the authors’ view,
the EU Agency is not responsible for any use that may be made of the
information it contains. MTQ and PG acknowledge the support of the Austrian
Science Fund (FWF) through the SFB project "BeyondC", a grant from the
Foundational Questions Institute (FQXi) Fund and a grant from the John
Templeton Foundation (Project No. 61466) as part of the Quantum Information
Structure of Spacetime (QISS) Project (qiss.fr). The opinions expressed in
this publication are those of the authors and do not necessarily reflect the
views of the John Templeton Foundation. Research at Perimeter Institute is
supported in part by the Government of Canada through the Department of
Innovation, Science and Economic Development Canada and by the Province of
Ontario through the Ministry of Colleges and Universities.
## References
* Reichenbach [1991] H. Reichenbach, _The direction of time_, Vol. 65 (Univ of California Press, 1991).
* Pearl [2009] J. Pearl, _Causality_, 2nd ed. (Cambridge University Press, 2009).
* Illari _et al._ [2011] P. M. Illari, F. Russo, and J. Williamson, _Causality in the Sciences_ (Oxford University Press, 2011).
* Leifer and Spekkens [2013] M. S. Leifer and R. W. Spekkens, Towards a formulation of quantum theory as a causally neutral theory of bayesian inference, Phys. Rev. A 88, 052130 (2013), arXiv:1107.5849 [quant-ph].
* Costa and Shrapnel [2015] F. Costa and S. Shrapnel, Quantum causal modelling, New Journal of Physics 18, 1–17 (2015), arXiv:1512.07106 [quant-ph].
* Modi [2012] K. Modi, Operational approach to open dynamics and quantifying initial correlations, Scientific Reports 2, 10.1038/srep00581 (2012), arXiv:1011.6138 [quant-ph].
* Pollock _et al._ [2018a] F. A. Pollock, C. Rodríguez-Rosario, T. Frauenheim, M. Paternostro, and K. Modi, Non-Markovian quantum processes: Complete framework and efficient characterization, Phys. Rev. A 97 (2018a), arXiv:1512.00589 [quant-ph].
* Li _et al._ [2018] L. Li, M. J. Hall, and H. M. Wiseman, Concepts of quantum non-markovianity: A hierarchy, Physics Reports 759, 1–51 (2018), arXiv:1712.08879 [quant-ph].
* Milz and Modi [2021] S. Milz and K. Modi, Quantum Stochastic Processes and Quantum non-Markovian Phenomena, PRX Quantum 2, 030201 (2021), arXiv:2012.01894 [quant-ph].
* Chiribella _et al._ [2008] G. Chiribella, G. M. D’Ariano, and P. Perinotti, Quantum Circuit Architecture, Phys. Rev. Lett. 101, 060401 (2008), arXiv:0712.1325 [quant-ph].
* Chiribella _et al._ [2009] G. Chiribella, G. M. D’Ariano, and P. Perinotti, Theoretical framework for quantum networks, Phys. Rev. A 80 (2009), arXiv:0904.4483 [quant-ph].
* Kretschmann and Werner [2005] D. Kretschmann and R. F. Werner, Quantum channels with memory, Phys. Rev. A 72, 062323 (2005), arXiv:quant-ph/0502106 [quant-ph].
* Gutoski and Watrous [2007] G. Gutoski and J. Watrous, Toward a general theory of quantum games, in _Proceedings of the Annual ACM Symposium on Theory of Computing_ (2007) pp. 565–574, arXiv:quant-ph/0611234 [quant-ph].
* Ried _et al._ [2015] K. Ried, M. Agnew, L. Vermeyden, D. Janzing, R. W. Spekkens, and K. J. Resch, A quantum advantage for inferring causal structure, Nature Physics 11, 414–420 (2015), arXiv:1406.5036 [quant-ph].
* MacLean _et al._ [2016] J.-P. W. MacLean, K. Ried, R. W. Spekkens, and K. J. Resch, Quantum-coherent mixtures of causal relations, Nature Communications 8, 15149 (2016), arXiv:1606.04523 [quant-ph].
* Chiribella _et al._ [2013] G. Chiribella, G. M. D’Ariano, P. Perinotti, and B. Valiron, Quantum computations without definite causal structure, Phys. Rev. A 88 (2013), arXiv:0912.0195v4 [quant-ph].
* Procopio _et al._ [2015] L. M. Procopio, A. Moqanaki, M. Araújo, F. Costa, I. Alonso Calafell, E. G. Dowd, D. R. Hamel, L. A. Rozema, Č. Brukner, and P. Walther, Experimental superposition of orders of quantum gates, Nature Communications 6, 7913 (2015), arXiv:1412.4006 [quant-ph].
* Rubino _et al._ [2017] G. Rubino, L. A. Rozema, A. Feix, M. Araújo, J. M. Zeuner, L. M. Procopio, Č. Brukner, and P. Walther, Experimental verification of an indefinite causal order, Science Advances 3, 11 (2017), arXiv:1608.01683 [quant-ph].
* Feix and Brukner [2017] A. Feix and Č. Brukner, Quantum superpositions of ‘common-cause’ and ‘direct-cause’ causal structures, New Journal of Physics 19, 123028 (2017), arXiv:1606.09241 [quant-ph].
* Giarmatzi and Costa [2021] C. Giarmatzi and F. Costa, Witnessing quantum memory in non-markovian processes, Quantum 5, 440 (2021), arXiv:1811.03722 [quant-ph].
* Milz _et al._ [2018] S. Milz, F. A. Pollock, T. P. Le, G. Chiribella, and K. Modi, Entanglement, non-Markovianity, and causal non-separability, New Journal of Physics 20, 033033 (2018), arXiv:1711.04065 [quant-ph].
* Taranto _et al._ [2019] P. Taranto, F. A. Pollock, and K. Modi, Memory Strength and Recoverability of Non-Markovian Quantum Stochastic Processes, arXiv e-prints (2019), arXiv:1907.12583 [quant-ph].
* Taranto [2020] P. Taranto, Memory effects in quantum processes, International Journal of Quantum Information 18, 1941002-574 (2020), arXiv:1909.05245 [quant-ph].
* Milz _et al._ [2021] S. Milz, C. Spee, Z.-P. Xu, F. Pollock, K. Modi, and O. Gühne, Genuine multipartite entanglement in time, SciPost Physics 10, 141 (2021), arXiv:2011.09340 [quant-ph].
* de Pillis [1967] J. de Pillis, Linear transformations which preserve hermitian and positive semidefinite operators, Pacific Journal of Mathematics 23, 129–137 (1967).
* Jamiołkowski [1972] A. Jamiołkowski, Linear transformations which preserve trace and positive semidefiniteness of operators, Reports on Mathematical Physics 3, 275–278 (1972).
* Choi [1975] M.-D. Choi, Completely positive linear maps on complex matrices, Linear Algebra and its Applications 10, 285 – 290 (1975).
* Milz _et al._ [2017] S. Milz, F. A. Pollock, and K. Modi, An introduction to operational quantum dynamics, Open Systems & Information Dynamics 24, 1740016 (2017), arXiv:1708.00769 [quant-ph].
* Pollock _et al._ [2018b] F. A. Pollock, C. Rodríguez-Rosario, T. Frauenheim, M. Paternostro, and K. Modi, Operational markov condition for quantum processes, Phys. Rev. Lett. 120 (2018b), arXiv:1801.09811 [quant-ph].
* Peres [1996] A. Peres, Separability criterion for density matrices, Phys. Rev. Lett. 77, 1413–1415 (1996), arXiv:quant-ph/9604005 [quant-ph].
* and Horodecki, Michał and Horodecki and Horodecki [1996] P. and Horodecki, Michał and Horodecki and R. Horodecki, Separability of mixed states: Necessary and sufficient conditions, Physics Letters, Section A: General, Atomic and Solid State Physics 223, 1–8 (1996), arXiv:quant-ph/9605038 [quant-ph].
* Doherty _et al._ [2002] A. C. Doherty, P. A. Parrilo, P. A. Parrilo, and F. M. Spedalieri, Distinguishing separable and entangled states, Phys. Rev. Lett. 88, 1879041–1879044 (2002), arXiv:quant-ph/0112007 [quant-ph].
* Doherty _et al._ [2004] A. C. Doherty, P. A. Parrilo, and F. M. Spedalieri, Complete family of separability criteria, Phys. Rev. A 69, 20 (2004), arXiv:quant-ph/0308032 [quant-ph].
* Chiribella _et al._ [2008] G. Chiribella, G. M. D’Ariano, and P. Perinotti, Transforming quantum operations: Quantum supermaps, EPL (Europhysics Letters) 83, 30004 (2008), arXiv:0804.0180 [quant-ph].
* Oreshkov _et al._ [2012] O. Oreshkov, F. Costa, and Č. Brukner, Quantum correlations with no causal order, Nature Communications 3, 1092 (2012), arXiv:1105.4464 [quant-ph].
* Araújo _et al._ [2015] M. Araújo, C. Branciard, F. Costa, A. Feix, C. Giarmatzi, and Č. Brukner, Witnessing causal nonseparability, New Journal of Physics 17, 1–28 (2015), arXiv:1506.03776 [quant-ph].
* Vidal and Tarrach [1999] G. Vidal and R. Tarrach, Robustness of entanglement, Phys. Rev. A 59, 141–155 (1999), arXiv:quant-ph/9806094 [quant-ph].
* Steiner [2003] M. Steiner, Generalized robustness of entanglement, Phys. Rev. A 67, 4 (2003), arXiv:quant-ph/0304009 [quant-ph].
* Gühne and Tóth [2009] O. Gühne and G. Tóth, Entanglement detection, Physics Reports 474, 1–75 (2009), arXiv:0811.2803 [quant-ph].
* Gurvits [2004] L. Gurvits, Classical complexity and quantum entanglement, Journal of Computer and System Sciences 69, 448–484 (2004), arXiv:quant-ph/0303055 [quant-ph].
* Brandao and Vianna [2004] F. G. S. L. Brandao and R. O. Vianna, Separable multipartite mixed states - operational asymptotically necessary and sufficient conditions, Phys. Rev. Lett. (2004), arXiv:quant-ph/0405063 [quant-ph].
* Quintino _et al._ [2016] M. T. Quintino, J. Bowles, F. Hirsch, and N. Brunner, Incompatible quantum measurements admitting a local-hidden-variable model, Physical Review A 93, 10.1103/physreva.93.052115 (2016), arXiv:1510.06722 [quant-ph].
* Hirsch _et al._ [2016] F. Hirsch, M. T. Quintino, T. Vértesi, M. F. Pusey, and N. Brunner, Algorithmic Construction of Local Hidden Variable Models for Entangled Quantum States, Phys. Rev. Lett. 117, 190402 (2016), arXiv:1512.00262 [quant-ph].
* Oszmaniec _et al._ [2017] M. Oszmaniec, L. Guerini, P. Wittek, and A. Acín, Simulating Positive-Operator-Valued Measures with Projective Measurements, Phys. Rev. Lett. 119, 190501 (2017), arXiv:1609.06139 [quant-ph].
* Hirsch _et al._ [2018] F. Hirsch, M. T. Quintino, and N. Brunner, Quantum measurement incompatibility does not imply Bell nonlocality, Phys. Rev. A 97, 012129 (2018), arXiv:1707.06960 [quant-ph].
* Bene and Vértesi [2018] E. Bene and T. Vértesi, Measurement incompatibility does not give rise to Bell violation in general, New Journal of Physics 20, 013021 (2018), arXiv:1705.10069 [quant-ph].
* Nielsen and Chuang [2000] M. Nielsen and I. Chuang, _Quantum Computation and Quantum Information_, Cambridge Series on Information and the Natural Sciences (Cambridge University Press, 2000).
* [48] D. Avis, lrs, http://cgm.cs.mcgill.ca/~avis/C/lrs.html.
* [49] Stanford ASL, vert2lcon, https://github.com/StanfordASL/KinoFMT/blob/master/vert2lcon.m.
* Hirsch _et al._ [2017] F. Hirsch, M. T. Quintino, T. Vértesi, M. Navascués, and N. Brunner, Better local hidden variable models for two-qubit Werner states and an upper bound on the Grothendieck constant KG(3), Quantum 1, 3 (2017), 1609.06114 [quant-ph].
* Cavalcanti _et al._ [2016] D. Cavalcanti, L. Guerini, R. Rabelo, and P. Skrzypczyk, General Method for Constructing Local Hidden Variable Models for Entangled Quantum States, Phys. Rev. Lett. 117, 190401 (2016), arXiv:1512.00277 [quant-ph].
* Cavalcanti and Skrzypczyk [2017] D. Cavalcanti and P. Skrzypczyk, Quantum steering: A review with focus on semidefinite programming, Reports on Progress in Physics 80 (2017), arXiv:1604.00501 [quant-ph].
* Bavaresco _et al._ [2017] J. Bavaresco, M. T. Quintino, L. Guerini, T. O. Maciel, D. Cavalcanti, and M. T. Cunha, Most incompatible measurements for robust steering tests, Phys. Rev. A 96 (2017), arXiv:1704.02994 [quant-ph].
* Horodecki and Horodecki [1999] M. Horodecki and P. Horodecki, Reduction criterion of separability and limits for a class of distillation protocols, Phys. Rev. A 59, 4206–4216 (1999), arXiv:quant-ph/9708015 [quant-ph].
* Horodecki _et al._ [2003] M. Horodecki, P. W. Shor, and M. B. Ruskai, Entanglement breaking channels, Reviews in Mathematical Physics 15, 629–641 (2003), arXiv:quant-ph/0302031 [quant-ph].
* Guo _et al._ [2021] Y. Guo, P. Taranto, B.-H. Liu, X.-M. Hu, Y.-F. Huang, C.-F. Li, and G.-C. Guo, Experimental Demonstration of Instrument-Specific Quantum Memory Effects and Non-Markovian Process Recovery for Common-Cause Processes, Phys. Rev. Lett. 126, 230401 (2021), arXiv:2003.14045 [quant-ph].
* Ringbauer _et al._ [2018] M. Ringbauer, F. Costa, M. E. Goggin, A. G. White, and A. Fedrizzi, Multi-time quantum correlations with no spatial analog, npj Quantum Information 4, 1–6 (2018).
* Horodecki [1997] P. Horodecki, Separability criterion and inseparable mixed states with positive partial transposition, Physics Letters, Section A: General, Atomic and Solid State Physics 232, 333–339 (1997), arXiv:quant-ph/9703004 [quant-ph].
* Lofberg [2004] J. Lofberg, Yalmip: A toolbox for modeling and optimization in matlab, https://yalmip.github.io/ (2004).
* Grant _et al._ [2009] M. Grant, S. Boyd, and Y. Ye, Cvx: Matlab software for disciplined convex programming, http://cvxr.com/cvx (2009).
* ApS" [2019] M. ApS", "the mosek optimization toolbox for matlab manual. version 9.0.", http://docs.mosek.com/9.0/toolbox/index.html (2019).
* Sturm [1999] J. F. Sturm, Using sedumi 1.02, a matlab toolbox for optimization over symmetric cones, Optimization Methods and Software 11, 625–653 (1999), https://doi.org/10.1080/10556789908805766.
* Toh _et al._ [1999] K. C. Toh, M. J. Todd, and R. H. Tütüncü, Sdpt3 — a matlab software package for semidefinite programming, version 1.3, Optimization Methods and Software 11, 545–581 (1999), https://doi.org/10.1080/10556789908805762.
* Johnston [2016] N. Johnston, QETLAB: A MATLAB toolbox for quantum entanglement, version 0.9, http://qetlab.com (2016).
* ccd [2021] https://github.com/marcellongvb/non_ccdc_processes (2021).
* Boyd and Vandenberghe [2004] S. Boyd and L. Vandenberghe, _Convex Optimization_ (Cambridge University Press, 2004).
* Navascués _et al._ [2009] M. Navascués, M. Owari, and M. B. Plenio, Power of symmetric extensions for entanglement detection, Phys. Rev. A 80 (2009), arXiv:0906.2731 [quant-ph].
* Feix _et al._ [2016] A. Feix, M. Araújo, and Č. Brukner, Causally nonseparable processes admitting a causal model, New Journal of Physics 18, 083040 (2016), arXiv:1604.03391 [quant-ph].
* Bavaresco _et al._ [2020] J. Bavaresco, M. Murao, and M. T. Quintino, Strict hierarchy between parallel, sequential, and indefinite-causal-order strategies for channel discrimination, (2020), arXiv:2011.08300 [quant-ph].
## Appendix A Ordered processes without quantum memory are equivalent to
direct-cause processes on the bipartite case
We now present the definition of ordered (non-Markovian) processes without
quantum memory, first introduced in Ref.[20].
###### Definition A.1 (Process without quantum memory[20]).
A linear operator $W\in\mathcal{L}(A_{I}\otimes A_{O}\otimes B_{I})$ is a
bipartite ordered process without quantum memory if it can be written as
$W=\rho_{\text{SEP}}^{A_{I}\textrm{aux}}*D^{A_{O}\textrm{aux}/B_{I}},$ (75)
where $\rho_{\text{SEP}}^{A_{I}\textrm{aux}}$ is a separable state and
$D^{A_{O}\textrm{aux}/B_{I}}$ is the Choi operator of a quantum channel.
We now show that all bipartite processes without quantum memory are bipartite
direct-cause, vice-versa. This proof was first presented in Appendix A.3.2 of
Ref. [20], but we reproduce it here for the sake of completeness.
###### Theorem A.1 (Appendix A.3.2 of Ref.[20]).
A bipartite ordered process $W$ is a process without quantum memory if and
only if $W$ is direct-cause.
###### Proof.
We first show that every process without quantum memory is direct-cause. Since
$\rho_{\text{SEP}}^{A_{I}\textrm{aux}}$ is separable, there exists
$\rho_{i}^{A_{I}}$ and $\sigma_{i}^{\textrm{aux}}$ and some probabilities
$p_{i}$ such that
$\rho_{\text{SEP}}^{A_{I}\textrm{aux}}=\sum_{i}p_{i}\rho_{i}^{A_{I}}\otimes\sigma_{i}^{\textrm{aux}}$.
Thus, we can write
$\displaystyle\left(\sum_{i}p_{i}\rho_{i}^{A_{I}}\otimes\sigma_{i}^{\textrm{aux}}\right)*D^{\textrm{aux}A_{O}/B_{I}}=\sum_{i}p_{i}\rho_{i}^{A_{I}}\otimes
D_{i}^{A_{O}/B_{I}},$ (76)
where
$D_{i}^{A_{O}/B_{I}}:=\sigma_{i}^{\textrm{aux}}*D^{\textrm{aux}A_{O}/B_{I}}$
are valid quantum channels, since $D_{i}^{A_{O}/B_{I}}\succeq 0$ and
$\mathrm{tr}_{B_{I}}(D_{i}^{A_{O}/B_{I}})=\mathds{1}^{A_{O}}$.
Now, we need to show that every direct-cause process is a process without
quantum memory in the bipartite case. Let us assume that $W$ is direct-cause.
Then $W$ can be written as
$W=\sum_{i}p_{i}\rho_{i}^{A_{I}}\otimes D_{i}^{A_{O}/B_{I}}$ (77)
for some states $\rho_{i}^{A_{I}}$ and channels $D_{i}^{A_{O}/B_{I}}$. Now,
let us define the separable state
$\sigma^{A_{I}\textrm{aux}}:=\sum_{i}p_{i}\rho_{i}^{A_{I}}\otimes|{i}\rangle\langle{i}|^{\textrm{aux}}$,
and the quantum channel
$D^{\textrm{aux}A_{O}/B_{I}}:=\sum_{i}|{i}\rangle\langle{i}|^{\textrm{aux}}\otimes
D_{i}^{A_{O}/B_{I}}$. Direct calculation shows that
$\sigma^{A_{I},\textrm{aux}}*D^{\textrm{aux}A_{O}/B_{I}}=W$, ensuring that $W$
is a process without quantum memory. ∎
## Appendix B A tight upper bound for the generalized robustness
In this section, we prove that the generalized robustness of any process is
upper-bounded by $R_{G}(W)\leq 1-\frac{1}{d_{A_{I}}}$. This bound is saturated
by the processes $W_{ddd^{2}}$ (Eq. (27)) for every dimension $d$ and by
$W_{222}$ (Eq. (36)).
###### Lemma B.1.
Let $W$ be a bipartite ordered process. If
$\widetilde{\Lambda}:\mathcal{L}(A_{I})\to\mathcal{L}(A_{I})$ is an
entanglement breaking channel, the process
$\widetilde{\Lambda}^{A_{I}}\otimes\widetilde{\mathds{1}}^{A_{O}B_{I}}(W)$ is
direct-cause.
###### Proof.
By definition, any bipartite ordered process can be written as
$W=\rho^{A_{I}\textrm{aux}}*D^{\textrm{aux}A_{O}/B_{I}}$. Since
$\widetilde{\Lambda}$ is entanglement breaking, it holds that
$\widetilde{\Lambda}^{A_{I}}\otimes\widetilde{\mathds{1}}^{\textrm{aux}}(\rho^{A_{I}\textrm{aux}})$
is a separable state. Therefore, we can use the same argument presented in the
proof of theorem A.1 to ensure that
$\widetilde{\Lambda}^{A_{I}}\otimes\widetilde{\mathds{1}}^{A_{O}B_{I}}(W)$ is
a direct-cause process. ∎
###### Lemma B.2.
Let $\widetilde{\Lambda}:\mathbb{C}_{d}\to\mathbb{C}_{d}$ be a quantum
channel, $\omega=e^{\frac{2\pi\sqrt{-1}}{d}}$, and
$Z:=\sum_{i=0}^{d-1}\omega^{i}|{i}\rangle\langle{i}|$ be the $d$-dimensional
clock operator. The channel
$\widetilde{\Lambda}(\rho)=\frac{1}{d}\sum_{k=0}^{d-1}Z^{k}\rho Z^{-k}$ (78)
is an entanglement breaking channel.
###### Proof.
A necessary and sufficient condition [55] for $\widetilde{\Lambda}$ to be
entanglement-breaking is that its Choi operator $\Lambda$ is separable between
input and output spaces. We now show that $\Lambda$ is separable by
$\displaystyle\Lambda:=$
$\displaystyle\sum_{ab}|{a}\rangle\langle{b}|\otimes\widetilde{\Lambda}(|{a}\rangle\langle{b}|)$
(79a) $\displaystyle=$
$\displaystyle\frac{1}{d}\sum_{abk}|{a}\rangle\langle{b}|\otimes
Z^{k}(|{a}\rangle\langle{b}|)Z^{-k}$ (79b) $\displaystyle=$
$\displaystyle\frac{1}{d}\sum_{abkij}|{a}\rangle\langle{b}|\otimes|{i}\rangle\langle{i}|\omega^{ik}(|{a}\rangle\langle{b}|)|{j}\rangle\langle{j}|\omega^{-jk}$
(79c) $\displaystyle=$
$\displaystyle\frac{1}{d}\sum_{kij}|{i}\rangle\langle{j}|\otimes|{i}\rangle\langle{j}|\omega^{ik}\omega^{-jk}$
(79d) $\displaystyle=$
$\displaystyle\frac{1}{d}\sum_{ij}|{i}\rangle\langle{j}|\otimes|{i}\rangle\langle{j}|\sum_{k}\omega^{ik}\omega^{-jk}$
(79e) $\displaystyle=$
$\displaystyle\frac{1}{d}\sum_{ij}|{i}\rangle\langle{j}|\otimes|{i}\rangle\langle{j}|\sum_{k}\omega^{k(i-j)}$
(79f) $\displaystyle=$
$\displaystyle\frac{1}{d}\sum_{ij}|{i}\rangle\langle{j}|\otimes|{i}\rangle\langle{j}|d\delta_{ij}$
(79g) $\displaystyle=$
$\displaystyle\sum_{i}|{i}\rangle\langle{i}|\otimes|{i}\rangle\langle{i}|.$
(79h)
∎
###### Theorem B.3.
The generalized robustness of a bipartite ordered process
$W\in\mathcal{L}(A_{I}\otimes A_{O}\otimes B_{I})$ is upper-bounded by
$R_{G}(W)\leq 1-\frac{1}{d_{A_{I}}}$.
###### Proof.
Let $\widetilde{\Lambda}:\mathcal{L}(A)\to\mathcal{L}(A)$ be the entanglement-
breaking channel defined in Lemma B.2. Note that since $Z^{0}=\mathds{1}$, the
action of $\widetilde{\Lambda}$ can be written as
$\displaystyle\widetilde{\Lambda}(\rho)$
$\displaystyle=\frac{1}{d}\sum_{k=0}^{d-1}Z^{k}\rho Z^{-k}$ (80a)
$\displaystyle=\frac{1}{d}\rho+\frac{1}{d}\sum_{k=1}^{d-1}Z^{k}\rho Z^{-k}$
(80b)
$\displaystyle=\frac{1}{d}\rho+\left(1-\frac{1}{d}\right)\widetilde{\Lambda}_{\setminus}(\rho),$
(80c)
where
$\widetilde{\Lambda}_{\setminus}(\rho):=\frac{1}{d-1}\sum_{k=1}^{d-1}Z^{k}\rho
Z^{-k}$ is a valid quantum channel.
Lemma B.1 states that, for any bipartite ordered process $W$, the process
$\widetilde{\Lambda}^{A_{I}}\otimes\widetilde{\mathds{1}}^{A_{O}B_{I}}(W)$ is
direct-cause, thus being CCDC. Hence, by making use of Eq. (80a), we see that
the resulting process
$\frac{1}{d_{A_{I}}}W+\left(1-\frac{1}{d_{A_{I}}}\right)\widetilde{\Lambda}_{\setminus}^{A_{I}}\otimes\widetilde{\mathds{1}}^{A_{O}B_{I}}(W)$
(81)
is CCDC, with
$\widetilde{\Lambda}_{\setminus}^{A_{I}}\otimes\widetilde{\mathds{1}}^{A_{O}B_{I}}(W)$
being a valid bipartite ordered process.
By analyzing the definition of generalized robustness presented in Eq. (20),
we see that setting
$\Omega=\widetilde{\Lambda}_{\setminus}^{A_{I}}\otimes\widetilde{\mathds{1}}^{A_{O}B_{I}}(W)$
ensures that the relation
$R_{G}(W)\leq 1-\frac{1}{d_{A_{I}}}$ (82)
holds for any bipartite ordered process $W$. ∎
## Appendix C An upper bound for the white noise robustness
In this section, we present an upper bound for the white noise robustness.
Differently from the generalized robustness case, this bound is not tight, but
is useful for proving the strong duality relation for the problem of
evaluating the white noise robustness for non-classical CCDC processes in
Section D.
###### Theorem C.1.
The white noise robustness of a bipartite ordered process
$W\in\mathcal{L}(A_{I}\otimes A_{O}\otimes B_{I})$ is upper-bounded by
$R_{WN}(W)\leq 1-\frac{1}{d_{A_{I}}d_{A_{O}}d_{B_{I}}+1}$.
###### Proof.
The depolarizing channel
$\widetilde{D}_{\eta}(\rho):=(1-\eta)\rho+\eta\frac{\mathds{1}}{d}$ is known
to be entanglement breaking when $\eta\geq\frac{d}{d+1}$ [54, 55]. Hence,
Lemma B.1 ensures that, for any bipartite ordered process
$W\in\mathcal{L}(A_{I}\otimes A_{O}\otimes B_{I})$, the process
$\widetilde{\Lambda}^{A_{I}}\otimes\widetilde{\mathds{1}}^{A_{O}B_{I}}(W)=\frac{1}{d_{A_{I}}+1}W+\frac{d_{A_{I}}}{d_{A_{I}}+1}\left(\frac{\mathds{1}^{A_{I}}}{d_{A_{I}}}\otimes\mathrm{tr}_{A_{I}}(W)\right)$
(83)
is direct-cause. We now define the operator
$W^{\prime}:=\frac{\mathds{1}^{A_{I}}}{d_{A_{I}}}\otimes\frac{\left(d_{A_{O}}\mathds{1}^{A_{O}B_{I}}-\mathrm{tr}_{A_{I}}(W)^{A_{O}B_{I}}\right)}{d_{A_{O}}d_{B_{I}}-1},$
(84)
which is a valid direct-cause process by direct inspection. Taking a convex
combination of
$\widetilde{\Lambda}^{A_{I}}\otimes\widetilde{\mathds{1}}^{A_{O}B_{I}}(W)$ and
$W^{\prime}$, we obtain
$\displaystyle\begin{split}&q\widetilde{\Lambda}^{A_{I}}\otimes\widetilde{\mathds{1}}^{A_{O}B_{I}}(W)+(1-q)W^{\prime}\\\
=&q\left(\frac{1}{d_{A_{I}}+1}W+\frac{d_{A_{I}}}{d_{A_{I}}+1}\left(\frac{\mathds{1}^{A_{I}}}{d_{A_{I}}}\otimes\mathrm{tr}_{A_{I}}(W)\right)\right)+(1-q)\frac{\mathds{1}^{A_{I}}}{d_{A_{I}}}\otimes\frac{\left(d_{A_{O}}\mathds{1}^{A_{O}B_{I}}-\mathrm{tr}_{A_{I}}(W)^{A_{O}B_{I}}\right)}{d_{A_{O}}d_{B_{I}}-1}\\\
=&\frac{q}{d_{A_{I}}+1}W+\left(\frac{(1-q)d_{A_{O}}d_{B_{I}}}{d_{A_{O}}d_{B_{I}}-1}\right)\frac{\mathds{1}^{A_{I}}}{d_{A_{I}}}\otimes\frac{\mathds{1}^{A_{O}B_{I}}}{d_{B_{I}}}+\left(\frac{qd_{A_{I}}}{d_{A_{I}}+1}-\frac{(1-q)}{(d_{A_{O}}d_{B_{I}}-1)}\right)\frac{\mathds{1}^{A_{I}}}{d_{A_{I}}}\otimes\mathrm{tr}_{A_{I}}(W),\end{split}$
(85)
which is direct-cause by construction. Now, by setting
$q=\frac{d_{A_{I}}+1}{d_{A_{I}}d_{A_{O}}d_{B_{I}}+1}$, we obtain
$\displaystyle\frac{qd_{A_{I}}}{d_{A_{I}}+1}-\frac{(1-q)}{(d_{A_{O}}d_{B_{I}}-1)}=0,$
(86)
and the process
$\frac{1}{d_{A_{I}}d_{A_{O}}d_{B_{I}}+1}W+\left(1-\frac{1}{d_{A_{I}}d_{A_{O}}d_{B_{I}}+1}\right)\frac{\mathds{1}}{d_{A_{I}}d_{B_{I}}}$
(87)
is guaranteed to be direct-cause. ∎
## Appendix D Strong duality for CCDC robustness problems
###### Theorem D.1.
The convex optimization problems for non-classical CCDC generalized robustness
(Eq. (20)) and non-classical CCDC white noise robustness (Eq. (21)) satisfy
strong duality.
###### Proof.
We recall that every convex optimization problem admitting a strictly feasible
solution i.e., all equality constraints are satisfied, and all inequality
constraints are strictly satisfied, necessarily satisfies strong duality
(Slater condition [66]).
From Theorem C.1 we see that for any value
$1-\frac{1}{d_{A_{I}}d_{A_{O}}d_{B_{I}}+1}<r<1$, for any process $W$, the
process $\Omega:=(1-r)W+r\frac{\mathds{1}}{d_{A_{I}}d_{B_{I}}}$ is a strictly
feasible solution, ensuring that both robustness problems respect strong
duality. ∎
## Appendix E SDPs for CCDC separability
In this section, we present the explicit forms of the SDPs mentioned in
Section 3 to obtain the generalized and white noise robustnesses of a
bipartite ordered process $W$.
Consider the problem of obtaining the PPTk generalized robustness of a process
$W$, _i.e._ , finding the robustnesses of a process against its worst noise
$\Omega$, so the resulting combination process lies in
$\mathcal{L}_{A\rightarrow B}^{\text{out,PPT}_{k}}$. This is represented by
the following optimization program:
$\displaystyle R_{G}^{\text{low,PPT}_{k}}(W):=\textrm{min}\hskip 5.69046ptr$
(88a) $\displaystyle\textrm{s.t.}\hskip
5.69046pt(1-r)W+r\Omega=qW_{\textrm{CC}}+(1-q)W_{\textrm{DC}}^{\text{PPT}_{k}},$
(88b) $\displaystyle\hskip 22.76228pt0\leq r\leq 1,$ (88c)
$\displaystyle\hskip 22.76228pt0\leq q\leq 1,$ (88d) $\displaystyle\hskip
22.76228ptW_{\textrm{CC}}\in\mathcal{L}_{\text{CC}},\
W_{\textrm{DC}}^{\text{PPT}_{k}}\in\mathcal{L}_{\text{DC}}^{\text{out,PPT}_{k}},\
\Omega\in\mathcal{L}_{A\rightarrow B},$ (88e)
with the process $W_{\text{DC}}^{\text{PPT}_{k}}$ having a $k$-symmetric PPT
extension $W_{\text{DC}}^{A_{I}^{\otimes
k}|A_{O}B_{I}}\in\mathcal{L}(A_{I}^{\otimes k}\otimes A_{O}\otimes B_{I})$
[32, 33]. This means that there exists a positive semi-definite operator
$W_{\text{DC}}^{A_{I}^{\otimes k}|A_{O}B_{I}}$ such that
$\displaystyle\mathrm{tr}_{A_{I}^{\otimes k-1}}(W_{\text{DC}}^{A_{I}^{\otimes
k}|A_{O}B_{I}})$ $\displaystyle=W_{\textrm{DC}}^{\text{PPT}_{k}}$ (89)
$\displaystyle W_{\text{DC}}^{A_{I}^{\otimes k}|A_{O}B_{I}}$
$\displaystyle=(P_{\text{sym}}^{A_{I}^{\otimes
k}}\otimes\mathds{1}^{A_{O}B_{I}})W_{\text{DC}}^{A_{I}^{\otimes
k}|A_{O}B_{I}}(P_{\text{sym}}^{A_{I}^{\otimes
k}}\otimes\mathds{1}^{A_{O}B_{I}})$ (90)
with positive partial transposition on $A_{I}^{\otimes k}$, that is
$\left(W_{\text{DC}}^{{A_{I}^{\otimes
k}|A_{O}B_{I}}}\right)^{T_{A_{I}}}\succeq 0,$ (91)
and where $P_{\text{sym}}^{A_{I}^{\otimes k}}$ is the projector onto the
symmetric subspace of $A_{I}^{\otimes k}$. We point that, in practice, instead
of Eq. (90), it is more advantageous to impose the Bose k-symmetric extension
condition, that is, to impose
$W_{\text{DC}}^{A_{I}^{\otimes k}|A_{O}B_{I}}=(P_{\text{sym}}^{A_{I}^{\otimes
k}}\otimes\mathds{1}^{A_{O}B_{I}})W_{\text{DC}}^{A_{I}^{\otimes
k}|A_{O}B_{I}}.$ (92)
This follows from the fact that, computationally, imposing the Bose
k-symmetric extension condition instead of the k-symmetric extension condition
does not add any complexity to the problem, but the Bose k-symmetric extension
condition detects more entangled states than the standard k-symmetric one
[67].
Since Eqs. (88) contains products of optimization variables, such as
$qW_{\textrm{CC}}$, the optimization program above is not linear, hence not an
SDP. This issue can be circumvented by rewriting the problem in the following
equivalent form:
$\displaystyle R_{G}^{\text{low, PPT}_{k}}(W):=\textrm{min}\hskip
5.69046pt\mathrm{tr}\left(\frac{\overline{\Omega}}{d_{A_{O}}}\right)$ (93a)
$\displaystyle\textrm{s.t.}\hskip
5.69046pt\left(1-\mathrm{tr}\left(\frac{\overline{\Omega}}{d_{A_{O}}}\right)\right)W+\overline{\Omega}=\overline{W_{\textrm{CC}}}+\overline{W_{\textrm{DC}}},$
(93b) $\displaystyle\hskip
22.76228pt\overline{W_{\textrm{CC}}}=\rho^{A_{I}B_{I}}\otimes\mathds{1}^{A_{O}}$
(93c) $\displaystyle\hskip
19.91684pt\left(\overline{W_{\textrm{DC}}}^{{A_{I}^{\otimes
k}|A_{O}B_{I}}}\right)^{T_{A_{I}}}\succeq 0,$ (93d) $\displaystyle\hskip
22.76228pt\overline{W_{\textrm{DC}}}=L_{A\rightarrow
B}(\overline{W_{\textrm{DC}}}),\ \overline{\Omega}=L_{A\rightarrow
B}(\overline{\Omega}),$ (93e) $\displaystyle\hskip
22.76228pt\overline{\Omega},\ \overline{W_{\textrm{DC}}},\
\rho^{A_{I}B_{I}}\succeq 0.$ (93f)
In the above SDP we have used the fact that valid processes should respect
$\mathrm{tr}(W)=d_{A_{O}}$ to embed the scalar variables into the operators.
This means that the new variables relate to the variables from Eqs. (88) with
$\overline{\Omega}=r\Omega,\ \overline{W_{\textrm{CC}}}=qW_{\textrm{CC}},\
\overline{W_{\textrm{DC}}}=(1-q)W_{\textrm{DC}}^{\text{PPT}_{k}}.$ (94)
Analogously, the same steps are applied to the PPT white noise robustness,
thus being
$\displaystyle R_{WN}^{\text{low,PPT}_{k}}(W):=\textrm{min}\hskip 5.69046ptr$
(95a) $\displaystyle\textrm{s.t.}\hskip
5.69046pt(1-r)W+r\frac{\mathds{1}^{A_{I}A_{O}B_{I}}}{d_{A_{I}}d_{B_{I}}}=\overline{W_{\textrm{CC}}}+\overline{W_{\textrm{DC}}},$
(95b) $\displaystyle\hskip
22.76228pt\overline{W_{\textrm{CC}}}=\rho^{A_{I}B_{I}}\otimes\mathds{1}^{A_{O}}$
(95c) $\displaystyle\hskip
19.91684pt\left(\overline{W_{\textrm{DC}}}^{{A_{I}^{\otimes
k}|A_{O}B_{I}}}\right)^{T_{A_{I}}}\succeq 0,$ (95d) $\displaystyle\hskip
22.76228pt\overline{W_{\textrm{DC}}},\ \rho^{A_{I}B_{I}}\succeq 0.$ (95e)
From the Lagrangian of the SDPs, we can obtain their dual problems, which
generate the optimal non-classical CCDC witnesses $S$ for the given process
$W$ [66]. For instance, considering $k=1$, the dual form101010Strictly
speaking, the dual objective function is to “maximize $-\mathrm{tr}(SW)$”,
which results in $R(W)=\mathrm{tr}(SW)$. Here, we adopted a convention of
changing the sign of $S$ and replacing this objective function by “minimize
$\mathrm{tr}(SW)$”. With this convention, the operator $S$ is a non-classical
CCDC witnesses as defined in Section 3.1, that is, it satisfies
$\mathrm{tr}(SW_{\text{CCDC}})\geq 0$ for all classical CCDC processes
$W_{\text{CCDC}}$. of the PPT generalized robustness is given by
$\displaystyle\textrm{min}\hskip 5.69046pt\mathrm{tr}\left(SW\right)$ (96a)
$\displaystyle\textrm{s.t.}\hskip
14.22636pt\mathds{1}^{A_{I}A_{O}B_{I}}\left(1+\mathrm{tr}(SW)\right)-d_{A_{O}}S\succeq
0$ (96b) $\displaystyle\hskip 28.45274ptS-
S_{\text{DC}}+S_{\text{DC}}^{\perp}\succeq 0$ (96c) $\displaystyle\hskip
28.45274ptL_{A\rightarrow B}(S_{\text{DC}}^{\perp})=0,$ (96d)
$\displaystyle\hskip 28.45274ptS_{\text{DC}}^{T_{A_{I}}}\succeq 0,$ (96e)
$\displaystyle\hskip 28.45274pt\mathrm{tr}_{A_{O}}(S)\succeq 0,$ (96f)
which is the one we use in our heuristic see-saw presented in Section F. The
variable $S_{\text{DC}}^{\perp}$ in Eq. (96c) is associated with an orthogonal
projection onto $\mathcal{L}_{A\rightarrow B}$. Also, the same methods
presented in Section D to prove that the robustness optimization problem
satisfies strong duality can be used to show that this upper bound problem
also respects strong duality. We, then, have $R_{G}^{\text{low,
PPT}}=-\mathrm{tr}(SW)$, when $S$ is the optimal witness for $W$.
We can see that Eq. (96c) together with Eqs. (96e) and (96f) ensure that $S$
is a non-classical CCDC witness. Also, Eq. (96b) corresponds to the
normalization condition, which determines that it is a witness for the
generalized robustness measure.
Analogously, the dual form of the PPT white noise robustness is given by
$\displaystyle\textrm{min}\hskip 5.69046pt\mathrm{tr}\left(SW\right)$ (97a)
$\displaystyle\textrm{s.t.}\hskip
14.22636pt\frac{\mathrm{tr}(S)}{d_{A_{I}}d_{B_{I}}}-\mathrm{tr}(SW)\leq 1,$
(97b) $\displaystyle\hskip 28.45274ptS-
S_{\text{DC}}^{T_{A_{I}}}+S_{\text{DC}}^{\perp}\succeq 0$ (97c)
$\displaystyle\hskip 28.45274ptL_{A\rightarrow B}(S_{\text{DC}}^{\perp})=0,$
(97d) $\displaystyle\hskip 28.45274ptS_{\text{DC}}\succeq 0,$ (97e)
$\displaystyle\hskip 28.45274pt\mathrm{tr}_{A_{O}}(S)\succeq 0,$ (97f)
where Eq. (97b) is the normalization condition for the witness $S$, which
corresponds to the white noise robustness measure.
A similar method can be used to characterize our inner approximation
$\mathcal{L}_{\text{CCDC}}^{\text{in},\mathcal{E}_{N}}$. In this case we just
need to absorb the probabilities $p_{i}$ of Eq. (93d) into the quantum channel
$D_{i}$ to obtain $\overline{D_{i}}:=p_{i}D_{i}$. This allows us to replace
the constraint of Eq. (93d) by
$\displaystyle
W_{\textrm{DC}}=\sum_{i=1}^{N}|{\psi_{i}}\rangle\langle{\psi_{i}}|^{A_{I}}\otimes\overline{D_{i}}^{A_{O}/B_{I}},$
(98a) $\displaystyle\overline{D_{i}}^{A_{O}/B_{I}}\succeq 0,$ (98b)
$\displaystyle\prescript{}{B_{I}}{\overline{D_{i}}^{A_{O}/B_{I}}}=\prescript{}{A_{O}B_{I}}{\overline{D_{i}}^{A_{O}/B_{I}}},$
(98c)
where $\\{|{\psi_{i}}\rangle^{A_{I}}\\}$ is a set of random states in
$\mathcal{L}(A_{I})$ and $\\{D_{i}^{A_{O}/B_{I}}\\}$ is a set of optimization
variables in $\mathcal{L}(A_{O}\otimes B_{I})$.
We can also obtain the dual form of the inner approximation problems of
robustnesses from the Lagrangian, which results in a similar SDP to Eqs. (96)
and Eqs. (97), but replacing Eqs. (96c) and (96e) by
$\displaystyle\mathrm{tr}_{A_{I}}\left[S(|{\psi_{i}}\rangle\langle{\psi_{i}}|^{A_{I}}\otimes\mathds{1}^{A_{O}B_{I}})+\prescript{}{B_{I}}{S_{i}}-\prescript{}{A_{O}B_{I}}{S_{i}}\right]\succeq
0\ \forall i\in\mathbb{N},$ (99)
where $\\{S_{i}\\}$ is a set of variables in $\mathcal{L}(A_{I}\otimes
A_{O}\otimes B_{I})$.
## Appendix F Heuristic algorithm that seeks for the maximum robustness on a
given scenario
Given the dimensions $d_{A_{I}},d_{A_{O}},d_{B_{I}}$, what is the maximum
robustness values $R_{G}(W)$ or $R_{WN}(W)$ that a process
$W\in\mathcal{L}_{A\rightarrow B}$ can obtain? Inspired by the see-saw
techniques of Refs. [52, 53], we now present a heuristic iterative method to
seek for the highest robustness values for a given scenario. This method works
either for the generalized or white noise robustness, which is why in the
following we do not make explicit which robustness quantifier to work with.
Our heuristic algorithm works as follows. First, we sample a bipartite ordered
process $W_{1}$ using one of the methods we describe in Appendix G. Then, we
perform the dual robustness problem for $W_{1}$ and obtain its optimal non-
classical CCDC witness $S_{1}$, which gives
$R(W_{1})=-\mathrm{tr}(S_{1}W_{1})$. Next, we find a process $W_{2}$ which
maximally violates this first witness $S_{1}$, a problem that can be solved by
the following SDP:
$\displaystyle\textrm{min}\hskip 39.83368pt\mathrm{tr}(S_{1}W_{2})$ (100a)
$\displaystyle\textrm{s.t.}\hskip 44.10185ptW_{2}\succeq 0,$ (100b)
$\displaystyle\hskip 56.9055ptW_{2}=L_{A\rightarrow B}(W_{2}),$ (100c)
$\displaystyle\hskip 56.9055pt\mathrm{tr}(W_{2})=d_{A_{O}}.$ (100d)
Now we repeat the previous steps, that is, we evaluate the dual robustness
program for $W_{2}$, and its optimal non-classical CCDC witness $S_{2}$, then
find the process $W_{3}$ which maximally violates $S_{2}$. These steps are
taken iteratively until some stopping criterion is satisfied. In our code, the
stopping criterion used is $R(W_{i+1})-R(W_{i})\leq\epsilon=0.0001$. In the
end of this procedure, we obtain a process $W$ which attains a non-classical
CCDC robustness, providing a lower bound on the maximal value for the given
scenario.
In order to increase the confidence in this heuristic method, we perform this
algorithm several times with for various initial processes $W_{1}$ randomly
sampled from different manners. We have implemented this heuristic methods for
three different scenarios: $d_{A_{I}}{=}d_{A_{O}}{=}d_{B_{I}}{=}2$,
$d_{A_{I}}{=}d_{A_{O}}{=}2,d_{B_{I}}{=}4$, and
$d_{A_{I}}{=}d_{A_{O}}{=}3,d_{B_{I}}{=}9$. For these cases, our see-saw
algorithm led to the same value of robustness for several different random
initial processes, suggesting that the heuristic method may have attained the
global maximum.
## Appendix G Sampling random ordered process
In this section, we describe the methods for sampling random processes used in
this work. We remark that Ref. [68] presents a method for generating uniformly
distributed random process matrices which are different from the ones
considered here.
### Method 1
This method is similar to the technique for generating random quantum channels
used in Ref. [69].
1. 1.
Sort a random density operator $\rho\in\mathcal{L}(A_{I}\otimes A_{O}\otimes
B_{I})$ with the Hilbert-Schmidt measure, _i.e_ , sort a random pure quantum
state with the Haar measure
$|{\psi}\rangle\langle{\psi}|\in\mathcal{L}(A_{I}\otimes A_{O}\otimes
B_{I}\otimes\textrm{aux})$ with
$d_{\textrm{aux}}=d_{A_{I}}d_{A_{O}}d_{B_{I}}$, then trace out the auxiliary
space;
2. 2.
Project $\rho$ into the subspace of bipartite ordered process to obtain
$\overline{W}=L_{A\rightarrow B}(\rho)$;
3. 3.
Evaluate the minimum eigenvalue $\lambda_{\text{min}}$ of $\overline{W}$ and
output the bipartite ordered process:
$W=d_{A_{O}}\cdot\frac{\overline{W}-\lambda_{\text{min}}\mathds{1}^{d_{A_{I}}d_{A_{O}}d_{B_{I}}}}{\mathrm{tr}\left(\overline{W}-\lambda_{\text{min}}\mathds{1}^{d_{A_{I}}d_{A_{O}}d_{B_{I}}}\right)},$
(101)
which is positive semi-definite by construction.
### Method 2
This method is good for generating random processes with high values of
generalized and white noise robustnesses. However, it only works when
$\frac{d_{A_{I}}d_{A_{O}}}{d_{B_{I}}}$ is an integer;
1. 1.
Set $d_{\textrm{aux}}=d_{A_{I}}$ and sort a random pure state
$|{\psi}\rangle\langle{\psi}|^{A_{I}\textrm{aux}}\in\mathcal{L}(A_{I}\otimes\textrm{aux})$
according to the Haar measure.
2. 2.
Set $d_{\textrm{aux}^{\prime}}=\frac{d_{A_{O}}d_{\textrm{aux}}}{d_{B_{I}}}$
and sort a random unitary operator $\text{U}:A_{O}\otimes\textrm{aux}\to
B_{I}\otimes\textrm{aux}^{\prime}$ according to the Haar measure. Then, obtain
its Choi operator
$|{\text{U}}\rangle\rangle\langle\langle{\text{U}}|^{A_{O}\textrm{aux}/B_{I}\textrm{aux}^{\prime}}$
and define the channel
$D^{A_{O}\textrm{aux}/B_{I}}:=\mathrm{tr}_{\textrm{aux}^{\prime}}\left(|{\text{U}}\rangle\rangle\langle\langle{\text{U}}|^{A_{O}\textrm{aux}/B_{I}\textrm{aux}^{\prime}}\right)$;
3. 3.
Output the operator:
$\displaystyle W$
$\displaystyle=\rho^{A_{I}\textrm{aux}}*D^{A_{O}\textrm{aux}/B_{I}},$ (102)
which is a valid bipartite ordered process.
### Method 3
1. 1.
Set $d_{\textrm{aux}}=d_{A_{I}}$ and sort a random pure state
$|{\psi}\rangle\langle{\psi}|^{A_{I}\textrm{aux}}$ according to the Haar
measure;
2. 2.
Sort a random density operator
$\rho\in\mathcal{L}(A_{O}\otimes\textrm{aux}\otimes B_{I})$ according to the
Hilbert-Schmidt measure;
3. 3.
Define
$D^{A_{O}\textrm{aux}/B_{I}}=\left(\sigma^{-\frac{1}{2}}\otimes\mathds{1}^{B_{I}}\right)\rho\left(\sigma^{-\frac{1}{2}}\otimes\mathds{1}^{B_{I}}\right)$,
with $\sigma=\mathrm{tr}_{B_{I}}(\rho)$. The resulting operator
$D^{A_{O}\textrm{aux}/B_{I}}$ is the Choi operator of a channel
$\widetilde{D}:\mathcal{L}(A_{I}\otimes\textrm{aux})\rightarrow\mathcal{L}(B_{I})$,
as $D^{A_{O}\textrm{aux}/B_{I}}\succeq 0$ and
$\mathrm{tr}_{B_{I}}(D^{A_{O}\textrm{aux}/B_{I}})=\mathds{1}^{A_{O}\textrm{aux}}$
by direct inspection;
4. 4.
Output the operator
$W=|{\psi}\rangle\langle{\psi}|^{A_{I}\textrm{aux}}*D^{A_{O}\textrm{aux}/B_{I}},$
(103)
which is a valid bipartite ordered process.
|
# Impact of correlations and heavy-tails on quantum error correction
B.D. Clader<EMAIL_ADDRESS>Colin J. Trout Jeff P. Barnes Kevin
Schultz Gregory Quiroz Paraj Titum Johns Hopkins University Applied Physics
Laboratory
11100 Johns Hopkins Road, Laurel, MD 20723, USA
###### Abstract
We show that space- and time-correlated single-qubit rotation errors can lead
to high-weight errors in a quantum circuit when the rotation angles are drawn
from heavy-tailed distributions. This leads to a breakdown of quantum error
correction, yielding reduced or in some cases no protection of the encoded
logical qubits. While heavy-tailed phenomena are prevalent in the natural
world, there is very little research as to whether noise with these statistics
exist in current quantum processing devices. Furthermore, it is an open
problem to develop tomographic or noise spectroscopy protocols that could test
for the existence of noise with such statistics. These results suggest the
need for quantum characterization methods that can reliably detect or reject
the presence of such errors together with continued first-principles studies
of the origins of space- and time-correlated noise in quantum processors. If
such noise does exist, physical or control-based mitigation protocols must be
developed to mitigate this noise as it would severely hinder the performance
of fault-tolerant quantum computers.
## I Introduction
The theory of fault tolerance and the associated threshold theorem demonstrate
that if the physical error rate per gate can be lowered below some threshold,
then one can perform quantum computation with arbitrary accuracy with
polynomial overhead (see e.g. Shor (1995); Steane (1996); Aliferis _et al._
(2006); Aharonov and Ben-Or (2008); Terhal (2015)). The prevailing noise model
for analyzing quantum error correcting codes is noise that manifests itself as
bit flips and phase flips that are local in both time and space, meaning there
are no spatial or temporal correlations Nielsen and Chuang (2010). However, it
is well known that at least some spatial correlations are inevitable in a
quantum system due to such physical effects as common baths shared amongst
qubits or control-line crosstalk Harper _et al._ (2020). In addition, time-
correlated noise is generally always present (e.g. $1/f^{\alpha}$ type noise
in superconducting qubit system Bialczak _et al._ (2007); Bylander _et al._
(2011); Burnett _et al._ (2014); Paladino _et al._ (2014); Kumar _et al._
(2016); Burnett _et al._ (2019)).
For these reasons, there have been a number of studies that have examined the
impact of spatial and temporal noise correlations on quantum error correction
(see e.g. Clemens _et al._ (2004); Klesse and Frank (2005); Terhal and
Burkard (2005); Aharonov _et al._ (2006); Aliferis _et al._ (2006); Alicki
_et al._ (2006); Novais and Baranger (2006); Novais _et al._ (2008); Shabani
(2008)). Despite the proliferation of studies examining noise correlations,
the complete impact on error correction is mixed. Some studies suggest that
noise correlations are fairly detrimental Clemens _et al._ (2004); Klesse and
Frank (2005); Alicki _et al._ (2006); Novais _et al._ (2008); Cafaro and
Mancini (2010) while others suggest that most realistic models of correlated
noise can still be handled via quantum error correction with a manageable
overhead Clemens _et al._ (2004); Terhal and Burkard (2005); Aharonov _et
al._ (2006); Aliferis _et al._ (2006); Novais and Baranger (2006); Shabani
(2008).
These disagreements arise due to the difficulty in analyzing and simulating a
quantum error correction (QEC) code operating on realistic multi-level quantum
systems interacting with a general open quantum system bath. Approximations
are always required to create manageable calculations such as considering only
the two-level subspace of the multi-level system and severe approximations as
to the impact of the bath on the qubits. The most restrictive of these bath
approximations is the assumption that it can be modeled using the Pauli error
model where random bit flips and phase flips are inserted stochastically into
the circuit. Despite the simple nature of this error model, it has proven
highly useful in the general theory of fault-tolerant QEC and the development
of new QEC codes.
The standard theory of fault-tolerant QEC assumes that if a weight-one Pauli
error occurs with probability $p$ then a weight-two Pauli error occurs with
probability of order $p^{2}$ and so on. Under these assumptions, a QEC code
will yield a logical error rate proportional to $p^{(d+1)/2}$ where $d$ is the
code distance. Correlated Pauli errors are clearly detrimental and reduce the
effectiveness of QEC. If a weight-two error occurs with an error rate
proportional to $p$ then the effective distance of the code is reduced by 1
and so on. Generating these types of errors can happen if the interaction
Hamiltonian contains entangling terms (sometimes called weight-two
generators). These types of errors might arise from a common bath shared by
all qubits that generates entanglement or residual coupling between qubits
that cannot be effectively turned off to a sufficient level. It is generally
assumed that qubits at best might share some portion of the bath, but that the
correlations would be short range at worst. Unfortunately, models of open
quantum systems with spatially correlated baths are not well studied, although
some progress has been made Jeske and Cole (2013). Despite this, it is
generally assumed that these higher weight errors can be effectively
controlled and mitigated in future quantum systems.
In this report, we analyze a related model of decoherence, but one that avoids
the difficulties required when performing general open quantum system
calculations. The errors are modeled as classical random variables within the
Hamiltonian. This model is often called the semi-classical system bath model.
It is an alternative model to a general open quantum system coupled to a
quantum bath, and is considered a reasonable approximation in certain
instances. In particular, the bath must be in thermal equilibrium, there is no
back action on the environment from the qubits (i.e. the bath dephases
instantly), and the bath is at infinite temperature yielding equal populations
of the qubit states after long term decay Kubo (1963); Haken and Strobl
(1973); Čápek (1993); Cheng and Silbey (2004); Gardiner (2004); Cheng and
Silbey (2005); Van Kampen (1997). These assumptions apply directly to
classical noise from the classical control system. The advantage to such a
model from our perspective is that spatial and temporal correlations are
fairly straightforward to model in this manner. There are known situations
where this approximation breaks down and the quantum nature of the bath leads
to observable signatures such as nontrivial phase evolution in addition to
pure decoherence Paz-Silva _et al._ (2017). This can occur for example for
baths at low temperature or systems that are strongly coupled. We do not
consider these more general quantum bath models here, and leave that for
future work.
Using the semiclassical noise model allows us to demonstrate a rather
surprising result. We show that single-qubit rotation errors arising from
weight-one error generators can lead to high-weight errors in a quantum system
when the noise is either spatially or temporally correlated and drawn from
certain types of heavy-tailed distributions Bryson (1974). This results in a
reduction in the effective distance of the QEC code that depends on the tails
of the distribution. Various definitions of heavy-tailed (sometimes referred
to as fat-tailed) distributions exist. Generally speaking they are
distributions whose tails decay more slowly than exponential (e.g. as a power
law) and they have undefined (or infinite) variance. They are well-studied in
the quantitative finance literature Haas and Pigorsch (2009) as they are
routinely used to model things such as exogenous shocks to financial markets
(see e.g. Refs. Fernandez-Villaverde and Rubio-Ramirez (2007); Fagiolo _et
al._ (2008); Mishkin (2011); Ascari _et al._ (2012)) or even the value of
returns for asset prices recognized early on with the seminal work of
Mandelbrot Mandelbrot (1963).
Relatively little research has been conducted as to whether semi-classical
noise with heavy-tailed distributions exists in quantum computing devices, but
this not true for quantum systems in general. Events with heavy-tailed
distributions have been discussed extensively in the context of physical
models that generate $1/f^{\alpha}$ noise spectra. It has been shown that
signals from systems with dynamics that adhere to families of point process
models Niemann _et al._ (2013); Eliazar and Klafter (2010); Davidsen and
Schuster (2002); Kaulakys _et al._ (2005); Lowen and Teich (1993); Lukovic
and Grigolini (2008) and linear/nonlinear stochastic differential equations
(SDEs) Kaulakys _et al._ (2013); Ruseckas and Kaulakys (2010) generate
$1/f^{\alpha}$ spectra when the signal probability density functions (PDFs)
obey heavy-tailed statistics. Various physical systems have exhibited signals
with power-law statistics. Similar point process models have been utilized to
describe the power-law behavior of fluorescent blinking in quantum dots Kuno
_et al._ (2000); Shimizu _et al._ (2001); Pelton _et al._ (2004); Margolin
_et al._ (2006); Pelton _et al._ (2007); Mahler _et al._ (2008); Frantsuzov
_et al._ (2009, 2013) and single-molecule fluorescence of organic molecules
Haase _et al._ (2004); Schuster _et al._ (2005); Hoogenboom _et al._
(2005); Yeow _et al._ (2006). Power-law behavior has been observed in
trapping times for charge transport in amorphous semiconductors Lowen and
Teich (1993); Dhariwal and Deoraj (1991); Tiedje and Rose (1981); Lowen and
Teich (1992); Adler _et al._ (1985) and nanoscale electrodes Krapf (2013).
The family of SDEs that exhibit power law behavior generate dynamics that
violate the fluctuation dissipation and equipartition of energy theorems which
has been observed in finite-dimensional spin-glasses Lobaskin and Kehrein
(2006). Furthermore, spin glasses exhibit random couplings and relaxation
rates that obey power-law behavior Klein (1968); Klein _et al._ (1976);
Cizeau and Bouchaud (1993); Berkov (1996); Pickup _et al._ (2009); Neri _et
al._ (2010).
For quantum processors, non-Gaussian noise spectroscopy techniques have been
developed Norris _et al._ (2016) and demonstrated experimentally Sung _et
al._ (2019), but this approach only applies to noise with distributions
tighter than Gaussian. The difficulty in developing characterization
techniques for heavy-tailed distributions is that most quantum
characterization techniques that seek to characterize the statistics of noise
correlations rely on expanding the statistics of the noise into moments (or
cumulants) Kubo (1962); Kampen (1974a, b). Since higher-order moments are
undefined or infinite for heavy-tailed distributions these existing techniques
do not apply. Our results show that a new characterization technique is needed
to test for this type of harmful noise in quantum systems. If this type of
noise is present, it is imperative that physical or control-based mitigation
schemes are developed to reduce its effect as QEC will not suffice.
## II Impact of correlated noise on QEC: Analytic Model
We begin our analysis by considering an analytic model of QEC where we
consider noise arising from stochastic unitary rotation errors on the data
qubits only in an $n$-qubit perfect code. This model is referred to as the
code-capacity error model. Physically, these correlations would correspond to
spatial correlations between qubits. The term perfect here refers to the fact
that our analytical analysis assumes that a distance $d$ code can correct
exactly $(d-1)/2$ errors on the encoded qubits and no more. Certain codes,
such as surface codes, can correct certain types of high-weight errors that
will break this assumption. Numerical simulations shown later demonstrate that
our code-capacity model accurately predicts the behavior of fully fault-
tolerant implementations of QEC with noise affecting data and ancilla qubits
at all locations in the circuit.
Our model starts by assuming that we have perfectly encoded an $n$-qubit
logical state into a QEC code denoted as $\ket{\psi_{L}}$. We then apply a
single-qubit rotation about an arbitrary axis to all of the data qubits in the
encoded state yielding
$\ket{\psi_{L}}\to\prod_{j=1}^{n}\left[\cos\left(\frac{\theta_{j}}{2}\right)\hat{I}-i\sin\left(\frac{\theta_{j}}{2}\right)\vec{v}\cdot\hat{\vec{\sigma}}^{(j)}\right]\ket{\psi_{L}},$
(1)
where $\hat{I}$ is the identity matrix, $\vec{v}$ is an arbitrary unit three
vector of real numbers specifying the direction of rotation,
$\hat{\vec{\sigma}}^{(j)}$ is a three vector containing the Pauli $x,y$, and
$z$ operators, and $\theta_{j}$ is the angle of rotation for qubit $j$. We
consider two cases. In the uncorrelated case, each angle of rotation is drawn
from a probability distribution and taken to be independent for a total of $n$
independent random variables. In the correlated case, we assume that a single
angle is drawn from the same probability distribution and applied to each
qubit. The average probability of a logical error is then given by
$P_{unc}=1-\sum_{k=0}^{w}\binom{n}{k}\left\langle\cos^{2}\left(\frac{\theta}{2}\right)\right\rangle^{n-k}\left\langle\sin^{2}\left(\frac{\theta}{2}\right)\right\rangle^{k}$
(2a)
$P_{cor}=1-\sum_{k=0}^{w}\binom{n}{k}\left\langle\cos^{2(n-k)}\left(\frac{\theta}{2}\right)\sin^{2k}\left(\frac{\theta}{2}\right)\right\rangle,$
(2b)
where we have assumed that our error correcting code can correct all Pauli
errors of weight $w=(d-1)/2$, $\theta$ is a random variable, and the angular
brackets $\langle\cdot\rangle$ denote an ensemble average. After some
algebraic manipulations, we can transform Eqs. (2) into
$P_{unc}=1-\frac{1}{2^{n}}\sum_{k=0}^{w}\sum_{l=0}^{n-k}\sum_{m=0}^{k}\binom{n}{k}\binom{n-k}{l}\binom{k}{m}(-1)^{m-k}f(t=1)^{n-l-m}$
(3a)
$P_{cor}=1-\frac{1}{2^{2n}}\sum_{k=0}^{w}\sum_{l=0}^{2(n-k)}\sum_{m=0}^{2k}\binom{n}{k}\binom{2(n-k)}{l}\binom{2k}{m}(-1)^{m-k}f(t=n-l-m),$
(3b)
where $f(t)$ is the characteristic function of the probability distribution.
In deriving Eqns. (3) we have assumed distributions that are symmetric about 0
such that the characteristic function is even $f(t)=f(-t)$. For a single
physical qubit, we define the failure probability to be the probability that
upon measurement we get either a bit-flip or phase-flip error. For a qubit
rotated by an angle $\theta$, this is given by
$P=\sin^{2}\left(\frac{\theta}{2}\right)$ (4)
with the corresponding expectation value in terms of the characteristic
function given by
$\langle P\rangle\equiv P_{ph}=\frac{1}{2}\left[1-f(t=1)\right],$ (5)
where $P_{ph}$ stands for the probability of a physical error occurring.
Evaluation and comparison of Eqs. (3) and (5) allow us to examine the impact
of various noise distributions by inserting the known characteristic function
and computing the formula for given code sizes and comparing the logical error
rate with the physical error rate. For the purpose of our analysis, the number
of qubits in the code will be equal to the the Knill–Laflamme bound $n\geq
4w+1$ unless otherwise specified.
Table 1: Comparison of the leading order term in a $\sigma\ll 1$ expansion of the failure probability for a physical qubit and various distance perfect codes with uncorrelated and correlate noise. The rows are for various $\nu=2r-1$ parameters in the Student’s $t$-distribution as given in Eq. (8). As $r$ increases the tails of the distribution are reduced, resulting in a tighter distribution. As this occurs, the reduction in effective code distance is pushed out to higher distances. As an example, for a $d=3$ code at $r=3$ the effective distances are equivalent for uncorrelated and correlated noise. We find that when $d\leq 2r-3$ the code distance is equal for correlated and uncorrelated noise, but when $d>2r-3$ the effective distance is reduced for correlated noise. For all code distances we set the number of qubits to $n=4w+1$, which is the minimum Knill-Laflamme bound. $r$ | Physical | $d=3$ | $d=5$ | $d=7$ | $d=9$
---|---|---|---|---|---
| | Unc | Cor | Unc | Cor | Unc | Cor | Unc | Cor
1 | $\frac{1}{2}\sigma^{\phantom{1}}$ | $\frac{5}{2}\sigma^{2}$ | $\frac{35}{64}\sigma^{\phantom{1}}$ | $\frac{21}{2}\sigma^{3}$ | $\frac{9009}{16384}\sigma^{\phantom{1}}$ | $\frac{715}{16}\sigma^{4}$ | $\frac{1154725}{2097152}\sigma^{\phantom{1}}$ | $\frac{1547}{8}\sigma^{5\phantom{1}}$ | $\frac{591534125}{1073741824}\sigma^{\phantom{10}}$
2 | $\frac{3}{4}\sigma^{2}$ | $\frac{45}{8}\sigma^{4}$ | $\frac{175\sqrt{3}}{64}\sigma^{3}$ | $\frac{567}{16}\sigma^{6}$ | $\frac{27027\sqrt{3}}{16384}\sigma^{3}$ | $\frac{57915}{256}\sigma^{8}$ | $\frac{3002285\sqrt{3}}{2097152}\sigma^{3}$ | $\frac{375921}{256}\sigma^{10}$ | $\frac{1436582875\sqrt{3}}{1073741824}\sigma^{3\phantom{1}}$
3 | $\frac{5}{12}\sigma^{2}$ | $\frac{125}{72}\sigma^{4}$ | $\frac{125}{8}\sigma^{4}$ | $\frac{875}{144}\sigma^{6}$ | $\frac{705705\sqrt{5}}{16384}\sigma^{5}$ | $\frac{446875}{20736}\sigma^{8}$ | $\frac{345262775\sqrt{5}}{18874368}\sigma^{5}$ | $\frac{4834375}{62208}\sigma^{10}$ | $\frac{130729041625\sqrt{5}}{9663676416}\sigma^{5\phantom{1}}$
4 | $\frac{7}{20}\sigma^{2}$ | $\frac{49}{40}\sigma^{4}$ | $\frac{49}{8}\sigma^{4}$ | $\frac{7203}{2000}\sigma^{6}$ | $\frac{7203}{16}\sigma^{6}$ | $\frac{343343}{32000}\sigma^{8}$ | $\frac{113011411513\sqrt{7}}{94371840}\sigma^{7}$ | $\frac{26000429}{800000}\sigma^{10}$ | $\frac{3787119775075\sqrt{7}}{9663676416}\sigma^{7\phantom{1}}$
5 | $\frac{9}{28}\sigma^{2}$ | $\frac{405}{392}\sigma^{4}$ | $\frac{243}{56}\sigma^{4}$ | $\frac{2187}{784}\sigma^{6}$ | $\frac{2187}{16}\sigma^{6}$ | $\frac{4691115}{614656}\sigma^{8}$ | $\frac{4691115}{256}\sigma^{8}$ | $\frac{13049829}{614656}\sigma^{10}$ | $\frac{1086610719708657}{7516192768}\sigma^{9\phantom{1}}$
6 | $\frac{11}{36}\sigma^{2}$ | $\frac{605}{648}\sigma^{4}$ | $\frac{605}{168}\sigma^{4}$ | $\frac{9317}{3888}\sigma^{6}$ | $\frac{1331}{16}\sigma^{6}$ | $\frac{10468315}{1679616}\sigma^{8}$ | $\frac{10468315}{2304}\sigma^{8}$ | $\frac{249145897}{15116544}\sigma^{10}$ | $\frac{249145897}{256}\sigma^{10}$
Figure 1: Ratio of $P_{cor}/P_{unc}$ of the leading order term in powers of
$\sigma$ in the error expansion of Eqs. (3) for correlated and uncorrelated
Gaussian noise, assuming $\sigma\ll 1$. The labels on the data points
correspond to the numerical value of the ratio. As the code distance
increases, the ratio increases as $d!!$.
### II.1 Gaussian
First, we examine the case where the noise is drawn from a Gaussian
distribution with zero mean and standard deviation $\sigma$. The
characteristic function of a Gaussian random variable with zero mean is
$f(t;\sigma)=e^{-\frac{1}{2}\sigma^{2}t^{2}}.$ (6)
To understand the impact on QEC, we study the low noise scaling of the
physical and logical error rates with respect to the width parameter of the
probability distribution. For the Gaussian distribution this implies taking a
series expansion about $\sigma\to 0$ for Eqs. (3). Evaluation of the leading
order terms for both correlated and uncorrelated noise reveals logical error
rates that scale as $\sigma^{2(w+1)}$ where $w$ is the number of correctable
errors for a code of distance $d=2w+1$. Thus, Gaussian correlated noise does
not yield high-weight errors that might impact the code distance, however it
does affect the series coefficient. To see this, we plot the ratio of the
first non-zero term for correlated noise to uncorrelated noise in Fig. 1. As
the distance of the code increases, the ratio of the series coefficient grows
at a rate proportional to $d!!$ implying that the logical error rate is
increased by a related factor for correlated noise. This implies that if
correlated single-qubit rotation noise drawn from a Gaussian distribution is
present, its impact may be minor for small codes, but grows as code distance
increases. We note as an aside that the $\sigma\to 0$ approximation of Eqs.
(2) must be done with care as the number of terms in the binomial expansion
grows for larger codes. Thus the approximation must technically be in the
regime where $\sigma\ll(1/d!!)^{2/(d+1)}$. These results agree with previously
published work and imply that a threshold does not exist Clemens _et al._
(2004). Despite this, the code can still suppress noise and a pseudo-threshold
does exist for each distance.
For a code of distance 3, the leading order terms assuming $\sigma\ll 1$ are
$\displaystyle P_{ph}$ $\displaystyle\approx\frac{\sigma^{2}}{4}$ (7)
$\displaystyle P_{unc}$ $\displaystyle\approx\frac{5\sigma^{4}}{8}$
$\displaystyle P_{cor}$ $\displaystyle\approx\frac{15\sigma^{4}}{8}.$
The logical error rate for both correlated and uncorrelated noise is reduced
by a squared factor of $\sigma$, while the pre-factor of the correlated noise
logical error rate increases by a factor of three relative to the uncorrelated
case. The quadratic reduction of the logical error rate suggests the code is
behaving as expected by a distance $3$ code in both cases.
### II.2 Student’s t
Next, we examine the Student’s $t$-distribution to see what impact varying the
probability distribution of the random variable has on the physical and
logical error rates. The Student’s $t$-distribution is an example of a heavy-
tailed distribution with a discrete parameter $\nu$ that governs the heaviness
of the tails of the distribution. This can be seen by considering the
probability density function of the Student’s $t$-distribution
$PDF_{t}(\theta;\nu,\sigma)=\frac{(\nu\sigma^{2})^{\frac{\nu}{2}}\Gamma\left(\frac{\nu+1}{2}\right)}{\sqrt{\pi}\Gamma\left(\frac{\nu}{2}\right)}\frac{1}{\left(\nu\sigma^{2}+\theta^{2}\right)^{\frac{\nu+1}{2}}}.$
(8)
Here $\nu\geq 1$ is an integer corresponding to the number of degrees of
freedom of the distribution and $\Gamma(x)$ is the gamma function. We will
restrict our attention to cases where $\nu$ is odd. Taking $\nu=1$ gives
exactly the Cauchy distribution, while larger values of $\nu$ tightens the
tails of the probability distribution via this discrete parameter. In the
limit of $\nu\to\infty$ the Gaussian distribution is recovered. The
characteristic function of a Student’s $t$-distribution is
$f(t;\sigma,\nu)=\frac{\sigma^{\frac{\nu}{2}}\nu^{\frac{\nu}{4}}|t|^{\frac{\nu}{2}}}{2^{\frac{\nu}{2}-1}\Gamma\left(\frac{\nu}{2}\right)}K_{\nu/2}\left(\sigma\sqrt{\nu}|t|\right),$
(9)
where $K_{n}(x)$ is the modified Bessel function of the second kind.
Again, we just report the approximate expressions in the $\sigma\ll 1$ limit.
For additional ease of analysis, we take $\nu$ to be odd, but we lift this
restriction in our numerical simulations. We display the results in Table 1
for the physical and logical errors rates with both correlated and
uncorrelated noise, where we have replaced $\nu$ with $\nu=2r-1$ and $r\geq 1$
is a positive integer. For correlated Cauchy random noise ($r=1$) error
correction provides no error suppression as the effective code distance is
reduced to 1 for all code distances considered. We conjecture that this
extends to arbitrary code distance. This implies that single-qubit rotations
with correlated rotation angles drawn from a Cauchy distribution result in at
least weight $(d+1)/2$ errors in the quantum circuit. As $r$ increases the
tails of the distribution are tightened, and the reduction in the code
distance is pushed out to higher distances. We find that when the distance is
$d\leq 2r-3$, for positive $d$, the effective distance of the QEC is
equivalent for the correlated and uncorrelated cases. However, once $d>2r-3$
the effective distance begins to be reduced for correlated noise with the
maximum exponent on $\sigma$ appearing to be $2r-1$ for correlated noise.
These results imply that correlated single-qubit noise drawn from the
Student’s $t$-distribution yields higher-weight Pauli errors with the weight
of the error related to the tail index of the distribution.
Figure 2: The various plots correspond to the value of the error term
proportional to $\sigma^{\alpha}$ for all valid values of $\alpha$ in the
correlated noise error expansion, $P_{cor}$. The dashed line is the first
order term of the physical error rate $P_{ph}=0.5$ of Eqs. (3). Code distance
does not have a large impact as seen by the nearly indistinguishable curves.
For all cases plotted, the logical error rate containing correlated noise from
an alpha stable distribution will not provide protection of the encoded
logical qubits as this distribution results in errors with weight of $(d+1)/2$
or greater expect at the exact point where $\alpha=2$, which corresponds to
the Gaussian distribution.
### II.3 Lévy alpha-stable
Finally, we consider the Lévy alpha-stable distributions. These are a family
of probability distributions that contain the Gaussian and Cauchy
distributions and allow for continuous interpolation between them. It is also
useful as it is a stable distribution meaning that sums of independent random
variates have the same distribution up to location and scale parameters. This
allows us to efficiently model time correlations in our numerical simulations
that we show in the next section. The probability density function for the
stable distribution is not analytically expressible. However the
characteristic function is expressible. We write a simplified form of the
characteristic function where we take the skewness and location parameters to
be zero yielding
$f(t;\sigma,\alpha)=e^{-|\sigma t|^{\alpha}}.$ (10)
The parameter $\alpha$ is the stability parameter and it lies in the range
$(0,2]$. The stable distribution corresponds to a Gaussian when $\alpha=2$ and
a Cauchy when $\alpha=1$, and allows for continuous interpolation between
Gaussian and heavy-tailed distributions. The formulas are complicated for the
general case, so we examine the simplest case for a distance 3 code. The
leading order terms, assuming $\sigma\ll 1$, for the physical and logical
error rates are
$\displaystyle P_{ph}$ $\displaystyle\approx\frac{\sigma^{\alpha}}{2}$ (11)
$\displaystyle P_{unc}$ $\displaystyle\approx\frac{5\sigma^{2\alpha}}{2}$
$\displaystyle P_{cor}$
$\displaystyle\approx\frac{1}{128}\left[5\left(2^{\alpha+2}\right)-5\left(3^{\alpha}\right)-5\left(4^{\alpha}\right)-5^{\alpha}+70\right]\sigma^{\alpha}$
$\displaystyle+\frac{1}{256}\left[5\left(-4^{\alpha+1}+9^{\alpha}+16^{\alpha}-14\right)+25^{\alpha}\right]\sigma^{2\alpha}.$
Only for exactly $\alpha=2$, which corresponds to a Gaussian distribution,
does the lowest order term (the term of order $\sigma^{\alpha}$) of the
correlated error rate disappear. Any $\alpha<2$ yields a term that is
proportional to the physical error rate, causing the error correcting code to
become ineffective against this type of noise. To examine larger code
distances, we plot the value of the term proportional to $\sigma^{\alpha}$ for
the correlated noise logical error rate in Fig. 2. The figure shows that for
all code distances considered the logical error rate scaling is proportional
to the physical error rate scaling. In other words, the QEC code offers no
protection for this type of noise no matter how large of a code is used. Only
exactly when $\alpha=2$ does the term dissapear and we recover protection of
the encoded logical state as shown previously in the discussion on Gaussian
noise.
## III Numerical Simulations
Our analytic model has shown the detrimental effect of spatially correlated
noise on the data qubits within a single error correction block, with
correlated single-qubit (weight one) rotation errors leading to uncorrectable
multi-qubit (high-weight) errors. However, determining fault tolerance also
requires the accounting of errors within the time-length of a decoding block
Aliferis _et al._ (2006). Therefore, we anticipate that time correlations of
weight-one error generators would have an equally harmful impact on quantum
error correction as they would lead to a breakdown of fault-tolerance due to
high-weight errors occurring in time across the decoding boundaries. To study
this, we use numerical simulations of low-distance surface codes to examine
the case of time-correlated errors from heavy-tailed distributions. In
addition to perfect correlations that we studied with our analytical model, we
also generalize to the situation where the noise correlations are defined by a
correlation function. Our simulations show that time-correlations do result in
a similar breakdown as our analytical results predict for spatially-correlated
noise on the data qubits only, but with non-trivial dependence on the
correlation function of the noise.
Because our errors are stochastic unitary errors, we must simulate the entire
state-vector. This limits us to low-distance codes and for this report, we
limit ourselves to just simulations of a distance-3 rotated surface code. We
make use of the same simulation framework and circuits that we used in our
study of random coherent errors Barnes _et al._ (2017). We model the errors
as random unitary gates
$U_{k}^{(\ell)}(\theta_{k}^{(\ell)})=\exp(-i\theta_{k}^{(\ell)}\sigma_{y}^{(\ell)})$
that create a rotation by a random angle $\theta_{k}^{(\ell)}$ (here drawn
from Gaussian, Cauchy, Student’s t, and Lévy alpha-stable distributions) about
the $Y$ axis at a circuit time location $k$ for qubit $\ell$. We insert these
errors across all the qubits in the code (both data and ancilla) after every
single gate in the circuit. To compute the performance of the code we start
with a random initial state
$\ket{\psi_{0}}=\cos\alpha\ket{0}+e^{i\beta}\sin\alpha\ket{1},$ (12)
where $0\leq\alpha<2\pi$ and $0\leq\beta<2\pi$ are both uniform random
variables. This random initialization covers the Bloch sphere, but it is not
uniform. We define the fidelity to be
$\mathcal{F}^{2}=\frac{1}{(2\pi)^{2}}\iint_{0}^{2\pi}d\alpha
d\beta\int_{-\infty}^{\infty}d\theta
p(\theta)|\braket{\psi_{0}}{e^{-i\theta\sigma_{y}}}{\psi_{0}}|^{2},$ (13)
where $p(\theta)$ is the probability distribution for the error terms. This
expression simply computes the overlap of the initial state and the final
state and averages over the probability distributions. We note, as before,
that this is not the standard definition of fidelity, since the distribution
of $\alpha$ and $\beta$ is uniform this is not a Haar average over the Bloch
sphere. This yields slightly different error rates for different errors on
different axes, but this does not have any meaningful impact on our overall
results. Using Eqs. (12) and (13), it is straightforward to analytically
calculate the physical fidelity for the various probability distributions
considered in this manuscript. It is
$\mathcal{F}^{2}=\frac{5}{8}+\frac{3}{8}f(t=2)$ (14)
where $f(t=2)$ is the characteristic function of the probability distribution
and is given in Eqs. (6), (9), and (10) for the Gaussian, Student’s t, and
Stable distributions respectively.
To calculate the logical fidelity, we perfectly encode the random state
defined in Eq. (12), simulate three rounds of faulty syndrome extraction with
errors inserted after every location where a gate exists (a circuit-level
noise model), and follow that by decoding and perfect correction. We conclude
the simulation with a round of perfect error correction to correct any
trailing errors. We estimate the logical fidelity by numerically estimating
the integral given in Eq. (13) by computing the overlap of the final decoded
logical state with the initial state and Monte Carlo sampling over the initial
random states and error terms. We use bootstrap resampling to report the
$95\%$ confidence intervals with $10^{3}$ samples with replacement used. Each
data point is the result of $10^{7}$ independent trials.
For arbitrary time correlations we leverage SchWARMA Schultz _et al._ (2020)
to simulate time-correlated noise in quantum circuits. The SchWARMA method
leverages a classical time-series modeling approach called autoregressive-
moving-average (ARMA) models where the angle of rotation at circuit time
location $k$ is an ARMA model
$\theta_{k}^{(\ell)}=\underbrace{\sum_{i=1}^{p}a_{i}\theta_{k-i}^{(\ell)}}_{AR}+\underbrace{\sum_{j=0}^{q}b_{j}x_{k-j}^{(\ell)}}_{MA}\,.$
(15)
The set $\\{a_{i}\\}$ defines the autoregressive portion of the model, and
$\\{b_{j}\\}$ the moving-average portion with $p$ and $q+1$ elements of each
set, respectively, and the $x_{k-j}^{(\ell)}$ are random variables drawn from
the user-defined probability distribution. Because ARMA models require one to
add random variates we must restrict ourselves to stable distributions when
using this method to ensure that the probability distribution of the output
model remains the same as the random variables. The Student’s $t$-distribution
is not stable. Therefore, in that case we consider white noise (uncorrelated
in time) and direct-current (DC) noise in which we draw a single random
variable at the beginning of the quantum circuit for each qubit and we use
that same angle at all subsequent times in the circuit. These two limits
correspond to an ARMA model with $p=0,q=0$ and $p=0,q\to\infty$ respectively.
For the other cases we can consider more general time correlations. For the
purposes of this study, we interpolate between white noise and DC noise by
considering exponential moving-averages (EMAs) where the terms
$b_{j}=N\exp(-\ln(2)j/T_{h})$ where $N$ is chosen such that $\sum_{j}b_{j}=1$
and $T_{h}$ is the “half-life” of the moving average. We do not consider the
AR portion in this paper, so $p=0$. We set each gate in the circuit to take a
single unit of time, so the parameter $j$ corresponds exactly to the circuit
depth to that point and we set the number of terms in the moving average to
$q=10\lceil T_{h}\rceil$. The various syndrome extraction circuits for the
surface code take anywhere from 2 to 6 time ticks in those units.
### III.1 Gaussian
The results for Gaussian noise are plotted in Fig. 3 where we plot the logical
infidelity versus the physical infidelity. There is no discernible difference
between the DC and white noise case. Below the pseudo-threshold the logical
error rate is reduced relative to the physical error rate and the reduction is
quadratic as expected for a distance three code. This is all in agreement with
our analytical results given in Eq. (7).
Figure 3: Pseudo-threshold plots obtained from simulating the rotated
distance-3 surface code with independent and time-correlated single-qubit
rotation errors drawn from a Gaussian distribution. The SchWARMA simulations
denoted with titles EMA $T_{h}$ interpolate between the white and DC noise
cases by using an exponential moving average filter. The various curves are
denoted by their color, but are indistinguishable on this scale. This
demonstrates that time-correlated noise drawn from a Gaussian distribution has
minimal impact to low distance QEC codes. The top and bottom dashed lines show
slopes where the logical infidelity is proportional to the physical infidelity
and the square of the physical infidelity respectively. Error bars are the 95%
confidence intervals obtained from bootstrap resampling. They are cutoff at
the lowest error rates for display purposes.
Figure 4: Pseudo-threshold plots obtained from simulating the rotated
distance-3 surface code with independent (top) and DC correlated (bottom)
single-qubit rotation errors drawn from a Student’s $t$-distribution for
multiple values of the tail index $\nu$ denoted in the legend. For the DC
noise case, the curves move from top to bottom in the same order as the legend
is displayed with $\nu=1$ having the highest logical error rate and $\nu=5$
the lowest. For white noise, the curves are indistinguishable on this scale.
Odd values of $\nu$ correspond to the analytical calculations shown in Table
1. The numerical results agree with our analytical results presented in Tab. 1
with the slope of the correlated DC noise logical error rate varying as the
tail index increases, while the slope of the uncorrelated white noise is not
impacted. Once $\nu\geq 4$ the correlated logical infidelity scales as the
square of the physical infidelity in agreement with our analytical
predictions. The top and bottom dashed lines show slopes where the logical
infidelity is proportional to the physical infidelity and the square of the
physical infidelity respectively. Error bars are the 95% confidence intervals
obtained from bootstrap resampling. They are cutoff at the lowest error rates
for display purposes.
### III.2 Student’s t
Next, we consider the Student’s $t$-distribution. We plot the simulation
results in Fig. 4. We consider the white-noise and DC noise cases and vary the
tail index $\nu$ for each simulation. The results are consistent with our
analytical predictions. For white (uncorrelated) noise, the code suppresses
the logical error rate and is relatively immune to different tail indices.
Meanwhile, for DC noise (infinite time correlation), the slope of the logical
to physical infidelity depends on the tail index. When $\nu=1$ the
distributions have the fattest tails and the simulations show that the logical
infidelity is proportional to the physical infidelity. This implies that the
QEC code provides no protection for the encoded qubits. For $\nu=\\{2,3\\}$
the code offers some protection, but not full protection. Finally, when
$\nu\geq 4$ the code offers full protection with the logical infidelity
scaling quadratically with the physical infidelity. We note that our
analytical results only considered odd values of $\nu$, so the appearance of
full code protection was only predicted for $\nu\geq 5$ from our analytical
results.
### III.3 Lévy alpha-stable
Finally, we show simulations of correlated noise drawn from the Lévy alpha-
stable distribution. Since the distribution is stable, we can use the SchWARMA
formalism to simulate time-correlated noise that interpolates between the
white noise and DC noise cases. We plot these results in Fig. 5. The results
once again agree with the analytical results for the white-noise and DC-noise
cases. For intermediate correlations times, enabled by our SchWARMA
simulations, we observe that the slope of the logical error rate is related to
the half-life of the exponential moving average. The time units of the half
life correspond exactly to the gate ticks of the circuit. For the surface
code, the syndrome extraction cycles take between 2 and 6 ticks (2-4 CNOT
gates for the weight-two and weight-four operators and two rotation gates for
the X syndrome), so the slope reduction occurs as the half-life approaches the
syndrome cycle as one would expect. The noise begins to generate weight-two
errors across syndrome boundaries. We have not examined any mitigation schemes
in this paper, but our results suggest that if the noise correlations can be
mitigated, through decoupling techniques as one example, that these effects
can be reduced and fault-tolerant quantum computing could be realized even if
heavy-tailed noise exists.
Figure 5: Pseudo-threshold plots obtained from simulating the rotated
distance-3 surface code with DC, time-correlated, and independent single-qubit
rotation errors drawn from an alpha-stable distribution with $\alpha=1.5$. The
curves move from top to bottom in the same order as the legend is displayed
with DC noise having the highest logical error rate and White noise the
lowest. The numerical results agree with our analytical results with the
uncorrelated white-noise case yielding a pseudo-threshold. For DC noise, the
code offers no protection with the logical error rate scaling proportionally
to the physical error rate. The SchWARMA simulations denoted with titles EMA
$T_{h}$ interpolate between the white and DC noise cases by using an
exponential moving average filter. These simulations show that time-correlated
noise drawn from a non-Gaussian heavy-tailed distribution can have a strongly
detrimental impact to QEC. The top and bottom dashed lines show slopes where
the logical infidelity is proportional to the physical infidelity and the
square of the physical infidelity respectively. Error bars are the 95%
confidence intervals obtained from bootstrap resampling.
## IV Conclusions
We have presented analytical evidence that space- or time-correlated single-
qubit noise drawn from heavy-tailed distributions can lead to high-weight
errors in a quantum circuit. This can lead to a breakdown of quantum error
correction via a reduced code distance (in some instances yielding no
protection at all). The exact predictions depend upon the type of distribution
used for the noise.
For Gaussian noise, correlations cause a reduction in leading order
coefficient of the pseudo-threshold, but the code can still suppress the
noise. This leads to logical error rates scaling as expected with a slope
proportional to $\sigma^{(d+1)/2}$ for a distance $d$ code. Meanwhile for
noise with heavy-tailed behavior such as Cauchy, Student’s $t$, or Lévy alpha-
stable distributions we find more interesting behavior. The quantum error
correcting codes that we considered could not correct correlated noise drawn
from Cauchy distribution. This leads to logical failure rates that scale
proportionally to the physical error rate. There is no suppression relative to
$\sigma$ for any code distance. The Student’s $t$-distribution allows us to
interpolate between the Gaussian and Cauchy case via the tail-index parameter
$r$. As $r$ increases, the error correcting code is able to suppress the
logical error rate relative to the physical error rate with increasing power.
Once the code distance $d\leq 2r-3$ the code achieves its expected error
suppression. Finally, the Lévy alpha-stable distribution also allows us to
interpolate between the Gaussian and Cauchy distributions, but since it is a
stable distribution we can also consider arbitrary time correlations. Here we
find that the correlation time of the noise has a direct impact on the slope
of the logical error rate with longer correlation times leading to more
reduction in error suppression ability.
These results all reinforce the notion that the ability of quantum error
correction to suppress noise is highly dependent on the physical noise model
Cafaro and van Loock (2014); Barnes _et al._ (2017); Iyer and Poulin (2018).
It is known that many complex physical systems have non-Gaussian noise and we
have pointed to many examples in Sec. I of quantum systems where noise with
heavy-tailed statistics is present Niemann _et al._ (2013); Eliazar and
Klafter (2010); Davidsen and Schuster (2002); Kaulakys _et al._ (2005); Lowen
and Teich (1993); Lukovic and Grigolini (2008); Kaulakys _et al._ (2013);
Ruseckas and Kaulakys (2010); Kuno _et al._ (2000); Shimizu _et al._ (2001);
Pelton _et al._ (2004); Margolin _et al._ (2006); Pelton _et al._ (2007);
Mahler _et al._ (2008); Frantsuzov _et al._ (2009, 2013); Haase _et al._
(2004); Schuster _et al._ (2005); Hoogenboom _et al._ (2005); Yeow _et al._
(2006); Lowen and Teich (1993); Dhariwal and Deoraj (1991); Tiedje and Rose
(1981); Lowen and Teich (1992); Adler _et al._ (1985); Krapf (2013); Lobaskin
and Kehrein (2006); Klein (1968); Klein _et al._ (1976); Cizeau and Bouchaud
(1993); Berkov (1996); Pickup _et al._ (2009); Neri _et al._ (2010); Sung
_et al._ (2019). Our results suggest that it is important to develop quantum
characterization techniques that can reliably determine the distribution of
the underlying noise to ensure the performance of future quantum error
corrected systems. Finally, we close with a positive note. Even if heavy-
tailed noise with correlations are present in quantum computing systems, our
results do not in any way preclude the possibility that quantum control or
other physical-level noise mitigation schemes may be able to reduce or
eliminate their detrimental impact. We leave that as an another avenue for
further research.
###### Acknowledgements.
Computing resources for the numerical simulations were provided by JHU/APL
through the Abyss cluster. We thank Leigh Norris for helpful discussions
during the preparation of this manuscript.
## References
* Shor (1995) P. W. Shor, Phys. Rev. A 52, R2493 (1995).
* Steane (1996) A. M. Steane, Phys. Rev. Lett. 77, 793 (1996).
* Aliferis _et al._ (2006) P. Aliferis, D. Gottesman, and J. Preskill, Quantum Info. Comput. 6, 97 (2006).
* Aharonov and Ben-Or (2008) D. Aharonov and M. Ben-Or, SIAM J. on Comp. 38, 1207 (2008).
* Terhal (2015) B. M. Terhal, Rev. Mod. Phys. 87, 307 (2015).
* Nielsen and Chuang (2010) M. A. Nielsen and I. L. Chuang, _Quantum Computation and Quantum Information_ , 10th ed. (Cambridge University Press, 2010).
* Harper _et al._ (2020) R. Harper, S. T. Flammia, and J. J. Wallman, Nature Physics (2020), 10.1038/s41567-020-0992-8.
* Bialczak _et al._ (2007) R. C. Bialczak, R. McDermott, M. Ansmann, M. Hofheinz, N. Katz, E. Lucero, M. Neeley, A. D. O’Connell, H. Wang, A. N. Cleland, and J. M. Martinis, Phys. Rev. Lett. 99, 187006 (2007).
* Bylander _et al._ (2011) J. Bylander, S. Gustavsson, F. Yan, F. Yoshihara, K. Harrabi, G. Fitch, D. G. Cory, Y. Nakamura, J.-S. Tsai, and W. D. Oliver, Nature Physics 7, 565 (2011).
* Burnett _et al._ (2014) J. Burnett, L. Faoro, I. Wisby, V. L. Gurtovoi, A. V. Chernykh, G. M. Mikhailov, V. A. Tulin, R. Shaikhaidarov, V. Antonov, P. J. Meeson, A. Y. Tzalenchuk, and T. Lindström, Nature Communications 5, 4119 (2014).
* Paladino _et al._ (2014) E. Paladino, Y. M. Galperin, G. Falci, and B. L. Altshuler, Rev. Mod. Phys. 86, 361 (2014).
* Kumar _et al._ (2016) P. Kumar, S. Sendelbach, M. A. Beck, J. W. Freeland, Z. Wang, H. Wang, C. C. Yu, R. Q. Wu, D. P. Pappas, and R. McDermott, Phys. Rev. Applied 6, 041001 (2016).
* Burnett _et al._ (2019) J. J. Burnett, A. Bengtsson, M. Scigliuzzo, D. Niepce, M. Kudra, P. Delsing, and J. Bylander, npj Quantum Information 5, 54 (2019).
* Clemens _et al._ (2004) J. P. Clemens, S. Siddiqui, and J. Gea-Banacloche, Phys. Rev. A 69, 062313 (2004).
* Klesse and Frank (2005) R. Klesse and S. Frank, Phys. Rev. Lett. 95, 230503 (2005).
* Terhal and Burkard (2005) B. M. Terhal and G. Burkard, Phys. Rev. A 71, 012336 (2005).
* Aharonov _et al._ (2006) D. Aharonov, A. Kitaev, and J. Preskill, Phys. Rev. Lett. 96, 050504 (2006).
* Alicki _et al._ (2006) R. Alicki, D. A. Lidar, and P. Zanardi, Phys. Rev. A 73, 052311 (2006).
* Novais and Baranger (2006) E. Novais and H. U. Baranger, Phys. Rev. Lett. 97, 040501 (2006).
* Novais _et al._ (2008) E. Novais, E. R. Mucciolo, and H. U. Baranger, Phys. Rev. A 78, 012314 (2008).
* Shabani (2008) A. Shabani, Phys. Rev. A 77, 022323 (2008).
* Cafaro and Mancini (2010) C. Cafaro and S. Mancini, Phys. Rev. A 82, 012306 (2010).
* Jeske and Cole (2013) J. Jeske and J. H. Cole, Phys. Rev. A 87, 052138 (2013).
* Kubo (1963) R. Kubo, J. of Math. Phys. 4, 174 (1963).
* Haken and Strobl (1973) H. Haken and G. Strobl, Zeitschrift für Physik A Hadrons and nuclei 262, 135 (1973).
* Čápek (1993) Čápek, Chemical Physics 171, 79 (1993).
* Cheng and Silbey (2004) Y. C. Cheng and R. J. Silbey, Phys. Rev. A 69, 052325 (2004).
* Gardiner (2004) C. Gardiner, _Handbook of stochastic methods for physics, chemistry, and the natural sciences_ , 3rd ed., Springer Series in Synergetics, Vol. 13 (Springer-Verlag, Berlin, 2004).
* Cheng and Silbey (2005) Y. C. Cheng and R. J. Silbey, Phys. Rev. A 72, 012320 (2005).
* Van Kampen (1997) N. Van Kampen, _Stochastic Processes in physics and chemistry_ , 2nd ed. (Elsevier Science, Amsterdam, 1997).
* Paz-Silva _et al._ (2017) G. A. Paz-Silva, L. M. Norris, and L. Viola, Phys. Rev. A 95, 022121 (2017).
* Bryson (1974) M. C. Bryson, Technometrics 16, 61 (1974).
* Haas and Pigorsch (2009) M. Haas and C. Pigorsch, “Financial economics, fat-tailed distributions,” in _Encyclopedia of Complexity and Systems Science_, edited by R. A. Meyers (Springer New York, New York, NY, 2009) pp. 3404–3435.
* Fernandez-Villaverde and Rubio-Ramirez (2007) J. Fernandez-Villaverde and J. F. Rubio-Ramirez, The Review of Economic Studies 74, 1059 (2007).
* Fagiolo _et al._ (2008) G. Fagiolo, M. Napoletano, and A. Roventini, Journal of Applied Econometrics 23, 639 (2008).
* Mishkin (2011) F. S. Mishkin, _Monetary Policy Strategy: Lessons from the Crisis_, Working Paper 16755 (National Bureau of Economic Research, 2011).
* Ascari _et al._ (2012) G. Ascari, G. Fagiolo, and A. Roventini, Macroeconomic Dynamics 19 (2012), 10.2139/ssrn.1987486.
* Mandelbrot (1963) B. Mandelbrot, The Journal of Business 36, 394 (1963).
* Niemann _et al._ (2013) M. Niemann, H. Kantz, and E. Barkai, Phys. Rev. Lett. 110, 140603 (2013).
* Eliazar and Klafter (2010) I. Eliazar and J. Klafter, Phys. Rev. E 82, 021109 (2010).
* Davidsen and Schuster (2002) J. Davidsen and H. G. Schuster, Phys. Rev. E 65, 026120 (2002).
* Kaulakys _et al._ (2005) B. Kaulakys, V. Gontis, and M. Alaburda, Phys. Rev. E 71, 051105 (2005).
* Lowen and Teich (1993) S. B. Lowen and M. C. Teich, Phys. Rev. E 47, 992 (1993).
* Lukovic and Grigolini (2008) M. Lukovic and P. Grigolini, The Journal of Chemical Physics 129, 184102 (2008), https://doi.org/10.1063/1.3006051 .
* Kaulakys _et al._ (2013) B. Kaulakys, R. Kazakevičius, and J. Ruseckas, in _2013 22nd International Conference on Noise and Fluctuations (ICNF)_ (2013) pp. 1–4.
* Ruseckas and Kaulakys (2010) J. Ruseckas and B. Kaulakys, Phys. Rev. E 81, 031105 (2010).
* Kuno _et al._ (2000) M. Kuno, D. P. Fromm, H. F. Hamann, A. Gallagher, and D. J. Nesbitt, The Journal of Chemical Physics 112, 3117 (2000), https://doi.org/10.1063/1.480896 .
* Shimizu _et al._ (2001) K. T. Shimizu, R. G. Neuhauser, C. A. Leatherdale, S. A. Empedocles, W. K. Woo, and M. G. Bawendi, Phys. Rev. B 63, 205316 (2001).
* Pelton _et al._ (2004) M. Pelton, D. G. Grier, and P. Guyot-Sionnest, Appl. Phys. Lett. 85, 819 (2004), https://doi.org/10.1063/1.1779356 .
* Margolin _et al._ (2006) G. Margolin, V. Protasenko, M. Kuno, and E. Barkai, The Journal of Physical Chemistry B 110, 19053 (2006), pMID: 16986903, https://doi.org/10.1021/jp061487m .
* Pelton _et al._ (2007) M. Pelton, G. Smith, N. F. Scherer, and R. A. Marcus, Proceedings of the National Academy of Sciences 104, 14249 (2007), https://www.pnas.org/content/104/36/14249.full.pdf .
* Mahler _et al._ (2008) B. Mahler, P. Spinicelli, S. Buil, X. Quelin, J.-P. Hermier, and B. Dubertret, Nature Materials 7, 659 (2008).
* Frantsuzov _et al._ (2009) P. A. Frantsuzov, S. Volkán-Kacsó, and B. Jankó, Phys. Rev. Lett. 103, 207402 (2009).
* Frantsuzov _et al._ (2013) P. A. Frantsuzov, S. Volkán-Kacsó, and B. Jankó, Nano Lett. 13, 402 (2013), pMID: 23272638, https://doi.org/10.1021/nl3035674 .
* Haase _et al._ (2004) M. Haase, C. G. Hübner, E. Reuther, A. Herrmann, K. Müllen, and T. Basché, The Journal of Physical Chemistry B 108, 10445 (2004).
* Schuster _et al._ (2005) J. Schuster, F. Cichos, and C. von Borczyskowski, Applied Physics Letters 87, 051915 (2005), https://doi.org/10.1063/1.2006217 .
* Hoogenboom _et al._ (2005) J. P. Hoogenboom, E. M. H. P. van Dijk, J. Hernando, N. F. van Hulst, and M. F. García-Parajó, Phys. Rev. Lett. 95, 097401 (2005).
* Yeow _et al._ (2006) E. K. L. Yeow, S. M. Melnikov, T. D. M. Bell, F. C. De Schryver, and J. Hofkens, The Journal of Physical Chemistry A 110, 1726 (2006).
* Dhariwal and Deoraj (1991) S. Dhariwal and B. Deoraj, Solid State Communications 79, 521 (1991).
* Tiedje and Rose (1981) T. Tiedje and A. Rose, Solid State Communications 37, 49 (1981).
* Lowen and Teich (1992) S. B. Lowen and M. C. Teich, Phys. Rev. B 46, 1816 (1992).
* Adler _et al._ (1985) D. Adler, B. B. Schwartz, and M. C. Steele, _Physical Properties of Amorphous Materials_ (Springer, Boston, MA, 1985).
* Krapf (2013) D. Krapf, Phys. Chem. Chem. Phys. 15, 459 (2013).
* Lobaskin and Kehrein (2006) D. Lobaskin and S. Kehrein, J. Stat. Phys. 123, 301 (2006).
* Klein (1968) M. W. Klein, Phys. Rev. 173, 552 (1968).
* Klein _et al._ (1976) M. W. Klein, C. Held, and E. Zuroff, Phys. Rev. B 13, 3576 (1976).
* Cizeau and Bouchaud (1993) P. Cizeau and J. P. Bouchaud, Journal of Physics A: Mathematical and General 26, L187 (1993).
* Berkov (1996) D. V. Berkov, Phys. Rev. B 53, 731 (1996).
* Pickup _et al._ (2009) R. M. Pickup, R. Cywinski, C. Pappas, B. Farago, and P. Fouquet, Phys. Rev. Lett. 102, 097202 (2009).
* Neri _et al._ (2010) I. Neri, F. L. Metz, and D. Bollé, Journal of Statistical Mechanics: Theory and Experiment 2010, P01010 (2010).
* Norris _et al._ (2016) L. M. Norris, G. A. Paz-Silva, and L. Viola, Phys. Rev. Lett. 116, 150503 (2016).
* Sung _et al._ (2019) Y. Sung, F. Beaudoin, L. M. Norris, F. Yan, D. K. Kim, J. Y. Qiu, U. von Lüpke, J. L. Yoder, T. P. Orlando, S. Gustavsson, L. Viola, and W. D. Oliver, Nature Communications 10, 3715 (2019).
* Kubo (1962) R. Kubo, Journal of the Physical Society of Japan 17, 1100 (1962), https://doi.org/10.1143/JPSJ.17.1100 .
* Kampen (1974a) N. V. Kampen, Physica 74, 215 (1974a).
* Kampen (1974b) N. V. Kampen, Physica 74, 239 (1974b).
* Barnes _et al._ (2017) J. P. Barnes, C. J. Trout, D. Lucarelli, and B. D. Clader, Phys. Rev. A 95, 062338 (2017).
* Schultz _et al._ (2020) K. Schultz, G. Quiroz, P. Titum, and B. Clader, arXiv preprint arXiv:2010.04580 (2020).
* Cafaro and van Loock (2014) C. Cafaro and P. van Loock, Phys. Rev. A 89, 022316 (2014).
* Iyer and Poulin (2018) P. Iyer and D. Poulin, Quantum Science and Technology 3, 030504 (2018).
*[QEC]: quantum error correction
*[ARMA]: autoregressive-moving-average
|
# Statistical physics through the lens of real-space mutual information
Doruk Efe Gökmen Institute for Theoretical Physics, ETH Zurich, 8093 Zurich,
Switzerland Zohar Ringel Racah Institute of Physics, The Hebrew University
of Jerusalem, Jerusalem 9190401, Israel Sebastian D. Huber Institute for
Theoretical Physics, ETH Zurich, 8093 Zurich, Switzerland Maciej Koch-Janusz
Institute for Theoretical Physics, ETH Zurich, 8093 Zurich, Switzerland
Department of Physics, University of Zurich, 8057 Zurich, Switzerland James
Franck Institute, The University of Chicago, Chicago, Illinois 60637, USA
###### Abstract
Identifying the relevant degrees of freedom in a complex physical system is a
key stage in developing effective theories in and out of equilibrium. The
celebrated renormalization group provides a framework for this, but its
practical execution in unfamiliar systems is fraught with ad hoc choices,
whereas machine learning approaches, though promising, lack formal
interpretability. Here we present an algorithm employing state-of-art results
in machine-learning-based estimation of information-theoretic quantities,
overcoming these challenges, and use this advance to develop a new paradigm in
identifying the most relevant operators describing properties of the system.
We demonstrate this on an interacting model, where the emergent degrees of
freedom are qualitatively different from the microscopic constituents. Our
results push the boundary of formally interpretable applications of machine
learning, conceptually paving the way towards automated theory building.
Fundamental physical theories, in a reductionist spirit, are often formulated
at the smallest scales, describing the interactions of elementary
constituents. Nonetheless, the experimentally accessible features typically
arise from their collective behaviour. Indeed, there exist profound examples
of effective theories, e.g. classical hydrodynamics and thermodynamics,
consistently describing complex phenomena in terms of a few macroscopic
variables, without making any reference to individual particles.
Bridging this scale gap to _derive_ the emergent macroscopic properties from
microscopic models is a perpetual challenge. The renormalization group
(RG),PhysicsPhysiqueFizika.2.263 ; Wilson1974 ; Wilson1975 ; Fisher1998
provides a powerful framework for this, associating physical theories at
different length scales by iteratively coarse-graining configurations of local
degrees of freedom (DOFs). The induced RG transformation acts as a telescope
in the space of models, generating the RG flow, whose structure around the
fixed point eventually reveals the relevant DOFs. They are the scaling
operators, which determine the correlations, and thus the physical properties,
at large scales.
In practice, executing this program in real-space RG is very difficult. The
accuracy of the procedure is improved by optimizing the coarse-graining to
retain the highest real-space mutual information (RSMI),Koch-Janusz2018 ;
optimalRSMI quantifying correlations to distant parts of the system. However,
this still misses a crucial insight: any long-range information is due to the
scaling operators and thus its optimal compression not only can serve as a
better RG transformation, but should allow to extract all the operators
themselves, without ever explicitly executing the RG flow. This was recently
proven in part: in critical systems, the formal solutions to the RSMI
compression problem are determined by the most relevant operators,Gordon2020
_in theory_ allowing to access them directly. Unfortunately, solving this
mathematical problem is notoriously hard in a general
setting.poole2019variational
Figure 1: Extracting the relevant operators with RSMI NE. (a) The most
relevant operators are _learnt_ as the optimal compressions of long-range
information $I(\mathcal{H}:\mathcal{E})$ at each point in the phase diagram.
The learnt maps can be associated to the physical operators by computing the
correlators and extracting the scaling dimensions. (b) The architecture of
RSMI-NE: the relevant operators are extracted via the transformations
$\Lambda$ and discretizing step $\tau$. The long-range information which
$\Lambda$ maximize is estimated by $f_{\Theta}$, all of which are parametrized
by neural networks and co-trained together.
Here, using state-of-art machine learning results in estimating mutual
information,poole2019variational we overcome this challenge to develop a
highly efficient algorithm extracting relevant operators of the theory from
real-space configurations. In contrast to standard approaches no RG maps are
iterated: scaling operators are _not_ constructed from the RG flow, but
instead using their definition as dominant contributions to RSMI, _in a single
step_. The RSMI neural estimator (RSMI-NE) returns them parametrized as neural
networks, which can be assigned scaling dimensions and used in computations
(see Fig. 1.a). Moreover, we empirically demonstrate the power of the method
across the whole phase diagram, also away from criticality.
In particular, the algorithm can, unsupervised, construct order parameters,
locate phase transitions, and identify spatial correlations and symmetries for
complex and large dimensional real-space data. Our findings, elevating the
coarse-graining transformations to formal operators, give a new paradigm in
investigating statistical systems, and a numerical toolbox to do so.
An often raised criticism of the use of machine learning in physics is the
lack of interpretability of the results.doi:10.1098/rsta.2016.0153
Particularly, the extent to which architecture- and training-dependent
conclusions from machine learning relate to formal concepts in physical
theories is unclear. RSMI-NE overcomes this challenge: its outputs are
explicitly identified with the scaling operatorsGordon2020 on the lattice.
Thus, in contrast to generic data-driven approaches, RSMI-NE executes a
physical principle using machine learning tools to produce theoretically
interpretable results.
Below we give an overview of the general RSMI setup, introducing the
probabilistic language of the coarse-graining optimisation. We then present
the RSMI-NE algorithm, and the theoretical and numerical results in machine
learning underlying its efficiency. We validate its capabilities on an
interacting model, whose non-trivial RG flow was a subject of a detailed
theoretical analysis.PhysRevLett.94.235702 ; PhysRevE.74.041124 We
investigate the physical data contained in the ensemble of RSMI filters. We
conclude with a discussion of further applications, most notably to non-
equilibrium problems.
Consider a system of classical DOFs in any dimension denoted by a collective
random variable $\mathcal{X}$, whose physics is specified by a probability
measure $p(x)$, either Gibbsian dictated by the energy of the realization $x$
of $\mathcal{X}$, or a generic non-equilibrium distribution. A coarse-graining
(CG) rule $\mathcal{X}\to\mathcal{X}^{\prime}$ is defined as a conditional
distribution $p_{\Lambda}(x^{\prime}|x)$, determined by a set of parameters
$\Lambda$ to be optimised. The coarse-graining is typically carried out on
disjoint spatial blocks $\mathcal{V}_{i}\subset\mathcal{X}$, and it
factorises: $p(x^{\prime}|x)=\prod_{i}p_{\Lambda_{i}}(h_{i}|v_{i})$, such that
$\mathcal{X}=\bigcup_{i}\mathcal{V}_{i}$ and
$\mathcal{X}^{\prime}=\bigcup_{i}\mathcal{H}_{i}$, with
$p_{\Lambda_{i}}(h_{i}|v_{i})$ the CG rule applied to block $i$. In
translation invariant systems a fixed $\Lambda_{i}\equiv\Lambda$ suffices;
with disorder each block can be individually optimised.
The RSMI principle identifies CG rules extracting the most relevant long-range
features as the ones retaining the most information shared by a block
$\mathcal{V}\subset\mathcal{X}$ to be coarse-grained, and its distant
environment $\mathcal{E}$,Koch-Janusz2018 ; optimalRSMI _i.e._ those that
optimally _compress_ this information. The environment is separated from
$\mathcal{V}$ by a shell of non-zero thickness constituting the buffer
$\mathcal{B}$, and forms the remainder of the system (see Fig. 1.a). The
“shared information” between the random variables $\mathcal{H}$ and
$\mathcal{E}$ is given by the mutual information:
$I_{\Lambda}(\mathcal{H}:\mathcal{E})=\sum_{h,e}p_{\Lambda}(e,h)\log\left(\frac{p_{\Lambda}(e,h)}{p_{\Lambda}(h)p(e)}\right),$
(1)
where $p_{\Lambda}(e,h)$ and $p(h)$ are the marginal probability distributions
of $p_{\Lambda}(h,x)=p_{\Lambda}(h|v)p(x)$ obtained by summing over the DOFs
in $\\{\mathcal{V}$, $\mathcal{B}\\}$ and
$\\{\mathcal{V},\mathcal{B},\mathcal{E}\\}$, respectively. Finding such
optimal coarse-graining requires thus maximizing $I_{\Lambda}$ as a function
of parameters $\Lambda$.
The conceptual importance of the buffer $\mathcal{B}$ cannot be overstated: it
sets the length-scale filtering out contributions of short-range correlations
between $\mathcal{V}$ and $\mathcal{E}$. Increasing its thickness
$L_{\mathcal{B}}$ corresponds to growing the RG scale, preserving only the
long-range physics. With an arbitrary fixed CG rule this can only be achieved
in RG by iterating the coarse-graining, with all the ensuing difficulties,
particularly amplifying the errors in the formulation of the rule. In our
approach _the CG rules themselves contain long-range information_ , and are
obtained in a single shot, by solving the $I_{\Lambda}$ optimization problem
directly at large $L_{\mathcal{B}}$, at different points in the phase diagram.
The optimization problem of Eq. (1) is, however, difficult, as estimating or
maximizing mutual information is notoriously hard.poole2019variational This
was a major weakness of the RSMI proposal,Koch-Janusz2018 hindering numerical
and theoretical progress. We can now overcome this challenge. At the heart of
our approach, encapsulated in the RSMI-NE algorithm, is a series of recent
results combining mathematically rigorous variational bounds on mutual
information doi:10.1002/cpa.3160360204 ; NIPS2003_2410 ; 5605355 with deep
learning.belghazi2018mine ; poole2019variational A family of _differentiable_
lower bounds to $I_{\Lambda}$ is introduced, parametrised by neural networks
$f_{\Theta}$ (see Fig. 1.b), which in the course of gradient descent training
on the joint samples of $\mathcal{H}$ and $\mathcal{E}$ become accurate, and
in the limit exact, estimators of $I_{\Lambda}$, see the Supplemental Material
(SM).111See Supplemental Material for details about the dimer model, more
technical details of the RSMI-NE algorithm and its dependence on length
scales, where also Refs. oord2018representation, ;
899a65b4919f47c8a06d115df85dad11, ; Goodfellow-et-al-2016, ; NIPS2014_5449, ;
gumbel1954, ; kingma2014adam, are included. The transformation
$p_{\Lambda}(h|v)$ feeding the coarse-grained variables into the estimator is
also expressed by a neural network _ansatz_. We use the following composite
architecture (see Fig. 1.b):
$h=\tau\circ(\Lambda\cdot v).$ (2)
Here $\Lambda$ become parameters of a convolutional neural network (CNN)
applied to the configurations, and $\tau$ differentiably maps $\Lambda\cdot v$
into states of variable $h$ of pre-determined type (_e.g._ pseudo-binary
spins). This last embedding step is both crucial,PhysRevX.10.031056 and
algorithmically non-trivial.jang2016categorical We emphasize that while the
CNN choice is motivated by convenience, any sufficiently expressive _ansatz_
can be used.
Figure 2: RSMI analysis of the interacting dimer model. (a) RG flow of the
model (see Eq.3) and representative configurations (top panel). (b) Total RSMI
extracted with the optimal filters as a function of $T$ and its scaling with
the buffer size. (c) Information extracted by the pristine staggered and
plaquette filters at different T. (d) Samples of optimal filters obtained with
RSMI-NE for different $T$ [columnar (C), plaquette (P1, P2) and staggered (S1,
S2)]. (e) The average overlap of the optimal filters at $T$ with the pristine
components. (f) The dimer symmetry breaking and plaquette order parameters
extracted using the low-$T$ pristine filters.
We have thus cast _both_ the CG rule and the lower-bound to the cost function
it optimizes as differentiable neural networks. Next, we can chain them
together (see Fig. 1.b), and _simultaneously_ optimize via stochastic gradient
descent, improving the RSMI estimator and the CG rule in each pass. Note, that
it is this numerical breakthrough which enables the exploration of new
theoretical ideas and renders the RSMI algorithm a promising new approach to
tackle open challenges in complex domains.
We demonstrate this on the example of an interacting dimer model. This is an
optimal test-bed for the illustration of our algorithm. First, a large class
of classical statistical physics problems can be mapped to interacting dimer
models.doi:10.1063/1.1703953 ; Blote_1982 ; Henley_1997 ; 10.2307/2692028 ;
Kenyon_2002 ; Cimasoni_2007 ; PhysRevLett.101.155702 ; PhysRevB.80.045112 ;
PhysRevB.80.134413 ; PhysRevLett.122.080601 Moreover, aspects of the quantum
dimer modelPhysRevLett.61.2376 ; PhysRevB.69.224415 leave their footprint on
the phase diagram.PhysRevLett.94.235702 ; PhysRevE.74.041124 Second, in the
dimer model, the relevant low-energy degrees of freedom are profoundly
different form the microscopic building blocks of the theory and change
qualitatively throughout the phase diagram. Hence, the algorithm is presented
a non-trivial task.
The model is defined by the partition function
$Z(T)=\sum_{\\{C\\}}\exp{(-E_{C}/T)}$ at a given temperature $T$ and the
configurations $C$ involve binary-valued microscopic degrees of freedom,
dimers, that sit on the edges of the square lattice. They obey the constraint
of exactly one dimer being connected to every vertex. The energy
$E_{C}=N_{C}(||)+N_{C}(=)$ counts plaquettes covered by parallel dimers
favoured by the interaction, see Fig. 2.a.
The essence of this system is in the interplay of aligning interaction energy
and entropic effects due to the non-local cooperation of local dimer covering
constraints. At low $T$ the former facilitates a long-range order (LRO)
crystallizing the system into one of the four translation symmetry breaking
columnar states, see Fig. 2.a. With increasing $T$ the system undergoes a
Berezinskii-Kosterlitz-Thouless (BKT) transition at $T_{\scriptscriptstyle\rm
BKT}=0.65(1)$,PhysRevE.74.041124 entering a critical phase characterised by
algebraic decay of correlations (also at $T\to\infty$) with exponents
continuously changing with $T$. This is reflected in the effective continuum
field theory, which, via the mapping of dimer configurations to height-field
$\varphi(\mathbf{r})$fradkin_2013 (see also SM for the definition) is given
by a sine-Gordon (SG) action:PhysRevE.74.041124
$S[\varphi(\mathbf{r})]=\int\mathrm{d}^{2}\mathbf{r}\left[\frac{g(T)}{2}|\nabla\varphi(\mathbf{r})|^{2}+V\cos\left(4\varphi(\mathbf{r})\right)\right].$
(3)
The potential $V$ locks $\varphi(\mathbf{r})$ into four values corresponding
to the columnar states. The stiffness $g(T)$ controls fluctuations of
$\varphi(\mathbf{r})$: large $g(T)$ favours “flat” fields of high-entropy, low
$g(T)$ allows large gradients corresponding to the staggered configurations,
which are not suppressed in the algebraic phase. The RG flow is shown in Fig.
2.a: the $T<T_{\scriptscriptstyle\rm BKT}$ fixed point with finite $g(T)$ and
$V\to\infty$ leaves energy minimisation as the sole relevant constraint; the
_line_ of fixed points at $V=0$ at $T>T_{\scriptscriptstyle\rm BKT}$ indicates
that the energetic interactions are irrelevant and exponents vary with $T$.
The flow reveals the physical nature of the algebraic correlations:
$\nabla\varphi(\mathbf{r})$ obeys Gauss’ law and so the fixed point theory is
that of electrical fields.
To showcase RSMI-NE, we input Monte Carlo samples of the model across the
whole temperature range to the algorithm. For concreteness, we restrict the
coarse-grained variables $\mathcal{H}$ to a two-component binary vector
$\\{\pm 1,\pm 1\\}$ (the optimal dimensionality can be found
systematicallyrsmine ). Hence, we are looking for a two component vector of
filters $\Lambda_{1}$, $\Lambda_{2}$ determining how the visible region
$\mathcal{V}$ is mapped onto $\mathcal{H}$. Optimizing the filters
$\Lambda_{1}$, $\Lambda_{2}$ for all $T$ separately gives a comprehensive
picture of the long-wavelength physics, culminating in the construction of the
relevant operators on the lattice, as we now show.
First, we find that already the curve $I_{\Lambda}(T)$, _i.e._ the amount of
long-range information attained _with the optimal_ $\Lambda$, reveals the
structure of the phase diagram (see Fig. 2.b). To wit, for
$T<T_{\scriptscriptstyle\rm BKT}$ its value is constant and equal to $\log 4$.
The information shared between distant parts of the system in the ordered
phase is precisely which of the four columnar states they are in.
Phase transitions are reflected by non-analyticities in $I_{\Lambda}(T)$
(_cf._ Wilms_2011, ; PhysRevE.87.022128, ). Moreover, the algebraic decay of
$I_{\Lambda}(T)$ with the buffer size for $T>T_{\rm BKT}$ is indicative of a
critical phase, see Fig. 2.b and Fig 6.c in Ref. rsmine, . This behaviour
should also be contrasted with the exponential decay for the paramagnetic
phase of 2D Ising model in Ref. rsmine, .
Going beyond the mutual information and examining the filters $\Lambda(T)$
yields further insight about spatial correlations. As conjectured, the optimal
CG rules depend on the tuning parameters of the system. In the high- and low-
temperature limits, three classes of filters emerge: independent optimizations
return exclusively sets of $\Lambda_{1,2}$ that correspond to columnar and
plaquette at low temperatures, and staggered ones at high temperatures, see
Fig. 2.d. We call these filters “pristine” as they reflect simple limiting
cases. They are orthogonal to each other and represent independent degrees of
freedom. The filters for intermediate temperatures and their overlap with the
pristine ones is shown in Figs. 2.d and e, respectively. We first discuss in
detail the individual filters $\Lambda_{1,2}$ in the different temperature
regimes $T\to 0$, $T\sim T_{\scriptscriptstyle\rm BKT}$ and $T\gg
T_{\scriptscriptstyle\rm BKT}$, and then explicitly match them with the RG-
relevant operators of the continuum sine-Gordon theory.
The pristine plaquette and columnar filters at $T\to 0$ break translation or
rotation symmetry, respectively. Any pair of $\Lambda_{1,2}$ drawn out of
these filters is a _bijection_ between the four ordered columnar states and
the four states $(\pm 1,\pm 1)$ taken by the compressed degrees of freedom
$\mathcal{H}$. This degeneracy of plaquette and columnar filters is lifted
when the rotation symmetry is restored: the pristine columnar filter is not
found above $T_{\scriptscriptstyle\rm BKT}$. Strikingly, its modulus acquires
an expectation value for $T<T_{\scriptscriptstyle\rm BKT}$ (see Fig. 2.f).
This filter is thus an order parameter _discovered_ by RSMI-NE, and is in fact
equal to the dimer symmetry breaking (DSB) order parameter identified in Ref.
PhysRevLett.94.235702, .
The optimal CG rules around $T_{\scriptscriptstyle\rm BKT}$ hold yet further
insights. Particularly, the plaquette filters give rise to a _putative_
plaquette order parameter (see Fig. 2.g). The corresponding regime where it
attains a non-vanishing value does not survive in the thermodynamic limit.
However, the non-zero expectation value at finite system sizes (see Fig. 2.f)
reveals the importance of such plaquette correlations, which are stabilised in
the quantum dimer model (QDM).PhysRevLett.61.2376 RSMI-NE indicates this
without any prior insights about QDM, which inspired previous
studies.PhysRevLett.94.235702 ; PhysRevE.74.041124
Finally, the critical phase $T>T_{\scriptscriptstyle\rm BKT}$ interpolates
between pristine plaquette and staggered filters, due to the competition
between the electric field operator and plaquette correlations in the finite
system, as per Eq. (3). The value of RSMI attained with _fixed_ rules reflects
this competition: the plaquette filter retains more information until well
above $T_{\scriptscriptstyle\rm BKT}$, where the staggered one takes over as
plaquette correlations dwindle (see Fig. 2.c). The staggered filters _are_ the
electric fields _viz._ they define coarse-grained variables
$E_{1,2}(\mathbf{r}):=\tau\circ\Lambda_{1,2}\left(\mathcal{V}(\mathbf{r})\right)$,
which precisely target the operator $\nabla\varphi(\mathbf{r})$ (see SM).
Table 1: Pairs of filters drawn from the columnar and plaquette coarse-
graining rules unambiguously label each of the four columnar ground states.
The mapping is given by the sign of the scalar product of the filter (blue=-1,
red=+1) with a dimer configuration (1 for occupied, 0 for unoccupied link). As
the columnar configurations correspond to uniform height field, the electric
charge operators $\mathcal{O}_{n=1,2}$, acting on the height-field $\varphi$
also serve as order parameters for the columnar phase and they directly
correspond to the RSMI-optimal filters at low $T$.
The RSMI-NE finding the order parameters or the electric fields is no
accident: the pristine $\Lambda$ filters define the relevant operators on the
lattice. The considerable technical machinery behind this is the subject of
Ref. Gordon2020, , here we show it using the field theory of the dimer model,
also away from criticality. To wit, the columnar and the DSB order parameters
in Fig. 2.f correspond to the relevant electric charge operators
$\mathcal{O}_{n}(\varphi)=(\cos(n\varphi),\sin(n\varphi))$,PhysRevB.76.134514
for $n=\pm 1$ and $n=\pm 2$, respectively. This is seen explicitly, using the
height-field map in Table 1 as a dictionary:
$\displaystyle\left(\Lambda_{\rm P1},\Lambda_{\rm P2}\right)\circ\varphi$
$\displaystyle=\left(\cos(\varphi+3\pi/4),\sin(\varphi+3\pi/4)\right),$ (4)
$\displaystyle\Lambda_{\rm C}\circ\varphi$ $\displaystyle=\cos(2\varphi),$ (5)
where on the left dimer configurations (on which the $\Lambda$ act) mapped to
height field value $\varphi$ are denoted by $\varphi$ itself.
Though competing correlations, especially in finite-size systems, may result
in mixing of the pristine components, they can be identified by applying
standard machine learning tools _to the ensemble of filters_. Note that RSMI-
NE is a stochastic algorithm, and through independent runs produces a
distribution of optimal CG rules. Thus Fig. 2.d shows a sample of filters at
each $T$, and in Fig. 2.e the overlap is averaged over the filter ensemble at
each temperature. The distribution contains crucial information, _e.g._ the
disappearance of the columnar filter above $T_{\scriptscriptstyle\rm BKT}$
signals the lifting of the columnar/plaquette degeneracy (consistent with the
scaling dimensions of $\mathcal{O}_{n}$, which go as $n^{2}$), and restoration
of the rotation symmetry. More concretely, representations of the broken
symmetries can be identified in the distribution, whereas at high-$T$ it can
be used to retrieve even the emergent $U(1)$ symmetry of the electrical field!
See Ref. rsmine, for a more detailed discussion.
We thus managed to automatically sequence the operators of the theory,
returning their lattice representations which are modular, reusable and may be
formally labelled by their scaling dimensions. Indeed, evaluating a correlator
of two neural networks parametrized by the plaquette filters, we fit a scaling
dimension of $1.00037$ at $T\rightarrow\infty$ (see SM), in excellent
agreement with $1.0$ predicted for $\mathcal{O}_{1}$.PhysRevB.76.134514 This
raises the remarkable prospect of building a complete effective theory from
raw data using machine learning.
Though the discussion centred around an equilibrium example in two dimensions,
our procedure works in any dimension, can be adapted to
disorder,PhysRevX.10.031056 and does not require the existence of a
Hamiltonian, as it only uses probability distributions. While a formal
understanding of this approach for non-equilibrium distributions, extending
the results of Ref. Gordon2020, , is missing, in the companion paperrsmine we
validated the concept on the example of lattice model with aggregation and
chipping,PhysRevE.63.036114 for which RSMI-NE locates precisely the non-
equilibrium phase transition. We believe complex systems, such as realized in
_e.g._ active matterPhysRevE.58.4828 ; Bialek4786 or atmospheric
phenomena,Peters2006 to be a natural arena where information-theoretic
methods can be applied,Dewar_2003 and our conceptual and numerical
advancements may provide new theoretical insights (see also Ref. Nir30234, ).
The understanding of challenging higher dimensional interacting and
quasiperiodic statistical systems PhysRevB.82.085114 ; PhysRevB.86.214414 ;
quasiperiodic_spin_chains ; PhysRevX.10.011005 may also benefit from this new
method.
Code availability Source code for the RSMI-NE is available online at
https://github.com/RSMI-NE/RSMI-NE.
Acknowledgements M.K.-J. is grateful to F. Alet for his comments on the
physics of the interacting dimer model. D.E.G., S.D.H., and M.K.-J. gratefully
acknowledge financial support from the Swiss National Science Foundation and
the NCCR QSIT, and the European Research Council under the Grant Agreement No.
771503 (TopMechMat), as well as from European Union’s Horizon 2020 programme
under Marie Sklodowska-Curie Grant Agreement No. 896004 (COMPLEX ML). Z.R.
acknowledges support from ISF grant 2250/19. Some of the computations were
performed using the Leonhard cluster at ETH Zurich. This work was supported by
a grant from the Swiss National Supercomputing Centre (CSCS) under project ID
eth5b.
Supplemental Material: Statistical physics through the lens of real-space
mutual information
## Appendix A RSMI optimization
Here we provide a description of the algorithmic components of the RSMI-NE,
referencing the necessary definitions and results from machine learning and
statistics. The estimation of of real-space mutual information (RSMI) is three
to four orders of magnitude faster than in Ref. Koch-Janusz2018, ; allowing to
explore large systems (we tested up to the size of 256x256), and the limit of
large buffer, while showing superior convergence and stability. The typical
run-times are _ca._ 30 seconds. This dramatic improvement in performance is
due to very recent advances in estimating information theoretic
quantities,belghazi2018mine ; poole2019variational and it, in turn, enables
development of an entirely new approach to real-space renormalization based on
RSMI, in particular enabling the extraction of relevant operators (see main
text, and section C).
### A.1 Some properties of mutual information
The Shannon mutual information (MI) of two random variables $X$ and $Y$
quantifies the amount of knowledge we gain about one of them, when observing
the other. Formally it is defined as a difference of entropies:
$I(X:Y):=H(X)-H(X|Y)=H(X)+H(Y)-H(X,Y).$ (6)
Typically RSMI can take values at most on the order of a few units of
information when coarse-graining small blocks $\mathcal{V}$ that contain
$N_{\mathcal{V}}$ individual degrees of freedom with a discrete alphabet of
$n$ symbols. Indeed:
$I_{\Lambda}(\mathcal{H}:\mathcal{E})\leq
I(\mathcal{V}:\mathcal{E})=H(\mathcal{V})-H(\mathcal{V}|\mathcal{E})\leq
H(\mathcal{V})\leq N_{\mathcal{V}}\log n,$ (7)
where, since $\Lambda$ compresses $\mathcal{V}$ into $\mathcal{H}$, the first
step is the data-processing inequality, and the second inequality follows from
the positive semi-definiteness of Shannon entropies. This is important, as MI
estimation by variational lower-bounds can suffer from a bias-variance trade-
off in the opposite regime, _i.e._ when the MI is large.
An alternative expression for $I(X:Y)$ is in terms of the Kullback-Leibler
(KL) divergence $D_{\rm KL}$:222The Kullback-Leibler divergence is a measure
of distance (but formally not a metric) between the two probability
distributions in its argument.
$I(X:Y)=D_{\rm KL}\left[p(x,y)||p(x)p(y)\right],$ (8)
The Gibbs’ inequality $D_{KL}(p||q)\geq 0$ predicates on a useful
interpretation of MI: given $(X,Y)$ jointly distributed according to $p(x,y)$,
$I(X:Y)$ measures the information lost when encoding $(X,Y)$ as a pair of
independent random variables while they may not be so. This is $0$ if and only
if $X$ and $Y$ are actually independent, _i.e._ $p(x,y)=p(x)p(y)$.
### A.2 Noise-contrastive lower-bound of mutual information
The results in the text were obtained with the noise-contrastive lower-bound
of MI (which we refer to as InfoNCE). Here we introduce it and briefly
motivate its form. We begin by casting the MI as follows:
$I(X:Y)=\mathbb{E}_{p(x,y)}\left[\log\frac{p(x|y)}{p(x)}\right]=\mathbb{E}_{p(x,y)}\left[\log\frac{p(y|x)}{p(y)}\right].$
To proceed, let the conditional probability distribution $q(x|y)$ be a
variational ansatz for the conditional probability distribution $p(x|y)$
appearing in the definition of MI above. Consider the KL divergence between
$p(x|y)$ and $q(x|y)$:
$\displaystyle D_{\rm{KL}}(p(x|y)||q(x|y))$
$\displaystyle=\mathbb{E}_{p(x|y)}\left[\log\frac{p(x|y)}{q(x|y)}\right]$
$\displaystyle=\sum_{x}\frac{p(x,y)}{p(y)}\left(\log p(x|y)-\log
q(x|y)\right),$ (9)
where the joint and the marginal distributions are related by
$p(x,y)=p(x|y)p(y)=p(y|x)p(x)$. The variational distribution $q(x|y)$ serves
to approximate $p(x|y)$: the KL divergence vanishes for $q^{*}(x|y)=p(x|y)$.
Positive semi-definiteness of the KL divergence yields a lower-bound for
$I(X:Y)$, known as the Barber-Agakov (BA) bound:NIPS2003_2410
$\displaystyle I(X:Y)\geq$
$\displaystyle\mathbb{E}_{p(x,y)}\left[\log\frac{q(x|y)}{p(x)}\right]$
$\displaystyle=\mathbb{E}_{p(x,y)}\left[\log q(x|y)\right]+H(X)=:I_{\rm
BA}(X:Y),$ (10)
where $H(X)$ is the entropy of the random variable $X$. $I_{\rm BA}$ is a
functional of the marginal distribution $q(x|y)$
$I_{\rm BA}(X:Y)=I_{\rm BA}(X:Y)[q(x|y)].$
Since $D_{\rm KL}=0$ if and only if $q(x|y)=p(x|y)$, the BA bound is tight
only for the optimal ansatz $I_{\rm BA}(X:Y)[q(x|y)=p(x|y)]=I(X:Y)$.
MI being the KL divergence between the joint and the product of marginal
distributions motivates the idea of circumventing the exhaustive modelling of
$p(x|y)$ by $q(x|y)$, focusing instead on the correlations between the
variables $X$ and $Y$. Consider then a variational ansatz $q(x|y)$ constrained
into an energy-based family of functions:
$q(x|y):=\frac{p(x)}{Z(y)}e^{f(x,y)},$ (11)
with the partition function $Z(y):=\mathbb{E}_{p(x)}\left[e^{f(x,y)}\right]$.
By forcing $q(x|y)$ to this form, the complex correlations within the possibly
high-dimensional data $X$ are stripped-out into the marginal distribution
$p(x)$. The resulting lower-bounds are sensitive mainly to the variables’
interdependency. In other words, maximising the lower-bound of MI is rephrased
as a search for a function $f(x,y)$ modelling the relationships (or shared
information) between $X$ and $Y$ very well, ignoring the surplus relationships
within $X$ or $Y$ (whose contribution to the MI are insignificant). As noted
in Ref. poole2019variational, , this conceptual step underlies the leap in
terms of practical efficiency gained in the corresponding energy-based MI
lower-bounds _e.g._ by Nguyen, Wainwright and Jordan (NWJ)5605355 and
tractable unnormalised BA (TUBA) bound.poole2019variational .
Even so, such so-called single-sample bounds are known to suffer from the
large variance of the resulting MI estimators.poole2019variational ;
oord2018representation An improved approach is to divide a single batch of MC
samples for the pair of random variables $(X,Y)$ into minibatches of K-fold
replicated random variables $(X_{i},Y_{i})_{i=1}^{K}$, and to derive the
corresponding “multi-sample” lower-bounds (note the confusing, but standard,
dual usage of the term “sample”). These are obtained by taking the average of
the single-sample bounds introduced above, and address the issue of large
variance by means of noise-contrastive estimation
(NCE)899a65b4919f47c8a06d115df85dad11 , as first proposed in the context of MI
estimation in Ref. oord2018representation, .
More concretely, a multi-sample (or replica) bound estimates $I(X_{1},Y)$,
where $(X_{1},Y)\sim p(x_{1},y)$, given $K-1$ additional independent replicas
for one of the random variables, say $X$ (drawn from the marginal
distribution), which are gathered into a multidimensional variable:
$X_{2:K}\sim\prod_{j=2}^{K}p(x_{j})$
Considering the average of $I_{\rm NWJ}(X_{i}:Y_{i})$ over the $K$ replica
random variables such that $(X_{i},Y_{i})\sim p(x_{i},y_{i})$, _i.e._ with
each $Y_{j}$ playing the role of $Y$ in turn, leads to the InfoNCE lower-bound
of MI:poole2019variational
$I(X:Y)\geq I_{\rm NCE}(X:Y):=\langle I_{\rm
NWJ}(X:Y)\rangle=\frac{1}{K}\mathbb{E}_{\prod_{k=1}^{K}p(x_{k},y_{k})}\left[\sum_{j=1}^{K}\log\frac{e^{f(x_{j},y_{j})}}{\frac{1}{K}\sum_{i=1}^{K}e^{f(x_{i},y_{j})}}\right].$
(12)
Maximisation of the InfoNCE lower-bound for RSMI is the method used in the
main text, where maximisation is over the discriminator functions $f$. A key
algorithmic idea here, originally introduced in Ref. belghazi2018mine, , is to
parametrize them by neural networks $f_{\Theta}$ and instead optimize over the
network parameters $\Theta$ using stochastic gradient descent.
### A.3 Deep neural network architecture for RSMI optimisation
Our goal in using the lower-bounds derived above is to optimize the RSMI, i.e.
MI between the random variable $\mathcal{H}$ representing the coarse-grained
degrees of freedom in the block $\mathcal{V}$, and $\mathcal{E}$ representing
the degrees of freedom in the environment.
There exist several admissible multi-layer perceptron (MLP) architectures for
the corresponding energy-based ansatz $f\equiv f_{\Theta}(h,e)$ (see above),
with a set of variational parameters $\Theta$. Here, we opt for a separable
form, such that:
$f_{\Theta}(h,e)=v^{\rm T}(h)u(e),$ (13)
where $v$ and $u$ are array-valued functions (here, neural networks, whose
weights constitute $\Theta$) that depend only on hidden variables and the
environment, respectively. The networks $v$ and $u$ independently map
$\mathcal{H}$ and $\mathcal{E}$ to a so-called embedding space. The advantage
of this choice is the ability to construct the elements of the scores matrix,
storing the values of the ansatz $f_{\Theta}$ for all pairs of jointly and
independently drawn samples, in $N$ passes of the MLP ($N$ passes for both $v$
and $u$ networks) for a sample dataset of size $N$. This is in contrast to the
requirement of $N^{2}$ passes for all $N(N-1)$ independent samples and $N$
joint samples in a concatenated architecture
$f_{\Theta}(h,e)=f_{\Theta}([h,e])$, where the data for the pair of variables
are concatenated before they are fed to the MLP.
Table 2: Details of architecture for the RSMI-estimation module of RSMI-NE. The multi-layer perceptron architecture expresses the variational ansatz $f_{\Theta}(h,e)=v^{\rm T}(h)u(e)$ for the RSMI lower-bound. For each of $v$ and $u$, we use a single fully-connected hidden layer, which is a tensor of shape (hidden dimension=32, embed dimension=8). $f_{\Theta}(h,e)$ is obtained by taking the inner product along the embedding axis. The neurons are activated by the ReLU function. We use the InfoNCE lower-bound, hence the baseline function is the constant $e$. lower-bound | ansatz type | $\\#$ layers | hidden dim. | embed dim. | activation function
---|---|---|---|---|---
InfoNCE | separable | 2 | 32 | 8 | ReLU
Tabulated in Tab.2 are the details of the network architecture for
$f_{\Theta}(h,e)$ that we have used in this work. Nevertheless, we note that
the results of the RSMI-NE are not very sensitive to a specific choice of
these parameters. Specifically, we opted for two hidden layers each with 32
neurons fully-connected to the layer containing the
$(\mathcal{H},\mathcal{E})$ data. The embedding dimension is 8. The neurons
are activated by the rectified linear unit (ReLU) function (see, e.g. Ref.
Goodfellow-et-al-2016, ).
#### A.3.1 Gumbel-softmax reparametrisation trick for discretization of
coarse-grained variables
In the RSMI-NE architecture the coarse-grained variables $h$ are fed into the
MI estimator. Since the MI value depends on what kind of distribution $h$
belongs to, we need to ensure this estimation step is not falsified by _e.g._
neglecting to force the output of the coarse-grainer into a discrete form,
rather than a real number, if we decided $h$ to be Ising spins. The apparent
problem is that discreteness of $h$ seems to spoil the differentiability of
the whole setup. This is somewhat similar to the problem encountered in
Variational Autoencoders (VAEs), which is solved there using the so-called
_reparametrization trick_ , effectively allowing to only differentiate w.r.t.
to the parameters of the latent space probability distribution. This is the
intuition behind the solution to the issue in RSMI-NE, which goes under the
name of _Gumbel-softmax reparametrisation trick_.NIPS2014_5449
Let $h$ be a categorical random variable which can be in one of the states
$\\{i\\}_{i=1}^{N}$ with the set of probabilities $\\{\pi_{i}\\}_{i=1}^{N}$.
Given $\\{g_{i}\\}_{i=1}^{N}$, random variables drawn from the Gumbel
distribution gumbel1954 ; NIPS2014_5449 , we define a vector-valued random
variable utilizing the softmax function, whose j-th component takes the form:
${\rm
softmax}_{j,\epsilon}\left(\\{g_{i}+\log\pi_{i}\\}_{i=1}^{N}\right)=\frac{\exp\left[(\log\pi_{j}+g_{j})/\epsilon\right]}{\sum_{i=1}^{N}\exp\left[(\log\pi_{i}+g_{i})/\epsilon\right]},$
(14)
where $\epsilon$ is the smearing parameter. For $\epsilon\rightarrow 0$ the
softmax becomes the argmax function, mapping the argument vector
$y=\\{g_{i}+\log\pi_{i}\\}_{i=1}^{N}$ into a $N$-component one-hot vector
(one-hot encoding maps each of $N$ possible states $i$ of a discrete variable
into a $N$-dimensional vector, with $1$ on $i$-th position, and zeros
elsewhere) with some $k^{*}$-th entry taking the value $1$, thereby marking
$y_{k^{*}}=\max y$. The resulting random variable is called a Gumbel-softmax
random variable; it is only approximately (or pseudo-) discrete, for small
enough $\epsilon$ (do not confuse with a discrete random variable defined by
taking the maximum component of the softmax function). In practice though, the
error coming from using a finite $\epsilon$ can be made comparable to machine
precision. The samples from the Gumbel-softmax approximation of a certain
categorical distribution $\\{\pi_{k}\\}_{k=1}^{N}$ are approximately one-hot
vectors for small $\epsilon$.
We anneal $\epsilon$ during the training, from $\epsilon_{\rm max}$ to
$\epsilon_{\rm min}$, exponentially, with a decay exponent $r$.
#### A.3.2 Network architecture for coarse-graining
In Tab. 3 we tabulate the details of the architecture for the coarse-graining
network we used for the 2D dimer model (and other examples in the companion
paper Ref. rsmine, ). We stack a single layer convolutional neural network
(CNN) (generally with multiple kernels, corresponding to different components
of $\mathcal{H}$) and the Gumbel-softmax reparametrisation layer to embed the
components of $\mathcal{H}$ into (pseudo-) binary variables. While we
determine the relaxation parameter $r$ by experimentation and fix it for all
models, we tune the initial value of the Gumbel-softmax temperature
$\epsilon_{\rm max}$ according to the total number of iterations during
training.
Table 3: Architecture details of the coarse-graining module of RSMI-NE for the 2-d interacting dimer model on a square lattice. $L_{\mathcal{V}}$ | 8
---|---
$L_{\mathcal{B}}$ | $\\{2,4,6,8\\}$
$L_{\mathcal{E}}$ | 4
number of components of $\mathcal{H}$ | 2
embedding of $\mathcal{H}$ | binary
$(\epsilon_{\rm max}$, $\epsilon_{\rm min})$ for Gumbel-softmax (GS) | $(0.75,0.1)$
GS annealing parameter $r$ | $5\times 10^{-3}$
Representing the coarse-graining map using a CNN enables interpreting the
weights of the network to be directly understood in terms of a generalized
Kadanoff block-spin construct for real-space RG.Kadanoff:1976vx Nevertheless,
the RSMI principle does not restrict the specific type of the variational
ansatz for coarse-graining, and the CNN form is not imperative.
### A.4 Unsupervised learning scheme
We now describe the (unsupervised) training, maximising our neural network
ansatz $I_{\Lambda}(\mathcal{H}:\mathcal{E})[f_{\Theta}]$. The inputs of the
RSMI-NE can be _e.g._ the Monte Carlo (MC) samples from the desired model (as
in the main text), but equally well experimentally measured data. As usual in
numerical investigations, ensuring good quality sampling is important. Since
we use the InfoNCE bound, the sampling is divided into mini-batches, each
containing $K$ samples. We separate in each sample the visible patch
$\mathcal{V}$ and its environment $\mathcal{E}$, dismissing a finite buffer
separating them. Then a single mini-batch is denoted by the multi-dimensional
random variable $(v_{1:K},e_{1:K})=(v_{1},\cdots,v_{K},e_{1},\cdots,e_{K})$.
Let $\Lambda^{s}$ and $\Theta^{s}$ be the network parameters for the coarse-
graining, and the InfoNCE ansatz $f$, respectively. At the outset we
initialize these as tensors containing random numbers. At each step $s$ of
training, the samples in the mini-batch are used to coarse-grain the samples
$v_{i}$ into $h_{i}[\Lambda^{s}]$ and compute the scores matrix
$F_{ij}(\Theta^{s},\Lambda^{s})=f(h_{i}[\Lambda^{s}],e_{j};\Theta^{s})$ for
the InfoNCE ansatz and the current values of the network parameters. Here
$\Lambda^{s}$ and $\Theta^{s}$ denote value of the network parameters in
$s$’th training step. In the scores matrix, the entries with $i=j$ denote the
jointly drawn samples and the rest denote independently drawn samples for the
coarse-grained degree of freedom and the environment.
The ansatz for RSMI can be formulated in a way which fixes the allowed
alphabet from which the coarse-grained degrees of freedom are drawn. In our
implementation, this is done by choosing a layer $\tau$ to generate
$\mathcal{H}$. For discrete-valued degrees of freedom we use the Gumbel-
softmax layer with an annealing schedule as described above. More generally,
one can specify the number of convolutional channels according to the
symmetries of the system.
The InfoNCE prediction (that of $p(h,e)$ being equal to $p(h)p(e)$ or not) for
the mini-batch is computed via the scores matrix as (compare Eq. 12):
$Q(h_{1:K},e_{1:K};\Theta^{s},\Lambda^{s})=\sum_{j=1}^{K}\frac{F_{jj}(\Theta^{s},\Lambda^{s})}{\sum_{ij=1}^{K}\exp{F_{ij}(\Theta^{s},\Lambda^{s})}}.$
(15)
Then
$\log Q(h_{1:K},e_{1:K};\Theta^{s},\Lambda^{s})+\log K$
gives our single mini-batch estimate of RSMI.
The gradients of the mini-batch estimate of RSMI with respect to $\Lambda$ and
$\Theta$ are then used for proposing the updated set of network parameters.
More concretely, we use the adam optimizerkingma2014adam to perform
stochastic gradient-ascent. We have found that using the same learning rate
for both parameter sets $\Lambda$ and $\Theta$ leads to efficient training.
We repeat this procedure over all mini-batches until all samples are fed to
the network once. This constitutes one epoch of training. We repeat it for
multiple epochs until convergence (see Ref. rsmine, for a more detailed
discussion of convergence criteria), when we are left with an optimized
coarse-graining filter represented by the final convolutional network
parameters $\Lambda$, and an estimate of the RSMI given by a moving average of
the time-series of mini-batch estimates.
## Appendix B Dependence on the buffer, block and system size
Here we provide a more detailed discussion of the length scales set by the
linear sizes of the buffer $\mathcal{B}$ and the block $\mathcal{V}$,
respectively denoted by $L_{\mathcal{B}}$ and $L_{\mathcal{V}}$, and their
physical interpretation as well as practical importance.
The nature of the scales $L_{B}$ and $L_{\mathcal{V}}$ is different. $L_{B}$
sets the distinction between long- and short-range contributions to RSMI.
Naively, one may assume $L_{B}\rightarrow\infty$ is necessary, but this is not
entirely true. To determine the functional form of the operator in terms of
local degrees of freedom, the scale $L_{B}$ needs only to be sufficiently
large, so that correlations due to the _second-most_ relevant operator decayed
to the point of being indistinguishable from statistical noise always present
in the sampling, while those of the most relevant one are still discernable
(assuming the leading operator is not degenerate). For a given total system
size, further increasing $L_{B}$ cannot change the answer, as the most
relevant operator has already “won”, and is the sole contributor to RSMI. In
fact, with finite sampling, it would make the problem computationally harder,
or impossible, as the leading correlations would be drowned by the noise.
Moreover, even when $L_{B}$ is not large enough to perfectly distinguish
leading from subleading contributions, _both of them_ can be extracted using
the _filter ensemble_ analysis descrived in the companion manuscript Ref.
rsmine, . Thus, in practice, already small buffers of a dozen sites are
sufficient to find the correct operator content (but not necessarily their
correct dimensions yet, see below). In use, one should systematically scan
(increase) the size of $L_{B}$.
The block $V$, on the other hand, is the support (in the mathematical sense)
of the relevant operator/order parameter. Assuming this operator is _local_ ,
in the first instance $V$ needs only to be sufficiently large so that the
appropriate function of the original degrees of freedom can be constructed.
For example, the “plaquette” operators in the dimer model require a block
twice smaller than what we used, and the plaquette pattern is repeated in the
block. This effectively corresponds to computing _block-averages_ of the
operator values, which produces quasi-continuous outputs even in completely
discrete systems. The degree of smearing is controlled by $L_{\mathcal{V}}$.
This is discussed in detail in Sec.III.D.1 and Sec.III.D.2 of the companion
Ref.rsmine, , where we show it allows to recover continuous emergent
symmetries. Note also that for _e.g._ experimental measurements of a quantity
performed on a grid of an arbitrary spatial resolution, the relevant
operator/order parameter may require a large support in such arbitrary units.
As for $L_{B}$, one should then scan the block size, or set it based on prior
knowledge. In typical cases, the size $V$ is very small though.
Though $L_{B}$ and $L_{\mathcal{V}}$ need not be large to extract the
operators, the above _does not imply_ that larger system sizes are
superfluous. In smaller systems, certain operators may appear more relevant
than others in parameter regimes (_e.g._ above a phase transition) where they
would be already subleading, had the system been closer to the thermodynamic
limit (see plaquette filters in Fig. 2c in the main text). The correctness and
accuracy of determination of their scaling dimension from the correlator will
be improved with the system size (and the number of samples). We emphasize,
however, that RSMI-NE discovers the parametrized form, in terms of the
original local degrees of freedom, of both the leading and subleading
operators. This allows to “export” them to analyse their properties with other
methods (even analytically, perhaps). We also expect that standard MC scaling
analysis to simplify, as we already have _the correct operators_ and only need
to compute their correlators. To wit, we fit for the charge operator
$\mathcal{O}_{1}$ at $T\rightarrow\infty$ a dimension of 1.00037, in excellent
agreement with the predicted value of $1.0$, already at system size of
$64\times 64$. Thus, the method has very favorable properties, extracting
parametrized leading and subleading operators on the lattice even from modest
system sizes, but large systems and more samples allow to more accurately
determine their properties, particularly the scaling dimensions.
## Appendix C Details of operator extraction in the interacting dimer model
### C.1 Dimer model: Height-field mapping of dimers on 2-d square lattice
Figure 3: Height field mapping of the dimer model on a square lattice. a The
height fields are defined on the plaquettes, and the electric fields related
to their gradients are shown by the arrows on the bonds (see explanation in
the text). The color-map shows the average height $\langle
H(\mathbf{r}_{i})\rangle$ over the four plaquettes surrounding an even-
sublattice site $i$ (denoted by green dots). Left: for a columnar ground-state
configuration the height profile $\langle H(\mathbf{r})\rangle$ is _uniform_.
Correspondingly, the average electric field vanishes over the whole
configuration. Middle: upon flipping a pair of dimers around a plaquette,
$\langle H(\mathbf{r})\rangle$ changes by one. Right: staggered configurations
have a net electric field over the whole configuration (observe the
orientation of the arrows). Correspondingly the height field has a _tilted_
profile. b $\varphi=\frac{\pi}{2}\langle H\rangle$ is interpreted as an angle
and is directly related to the orientation of the dimer at the given site.
Here we describe the classic mapping of the dimer models to height fields, as
per _e.g._ Ref. fradkin_2013, . Since the 2-d square lattice is bipartite, we
can map each dimer configuration $C(\mathbf{r}_{i})$ onto a unique height
profile $H(\mathbf{r}_{i})$ living on the plaquettes of the lattice. Without
loss of generality, we start by assigning a reference height
$H(\mathbf{r}_{0})=1/2$ to the plaquette at $i=0$. The heights of the
neighboring plaquettes are then determined by winding clockwise (counter-
clockwise) around a vertex of the even (odd) sublattice, and changing the
height by $-3$ if a dimer is crossed on the link connecting the two
plaquettes, or changing by $+1$ otherwise, see Fig. 3. The fact that each site
in the fully packed dimer model can have a single dimer with only one of the
four orientations attached to it is reflected in the height profile as an
invariance under a uniform shift by 4 of the entire profile:
$H(\mathbf{r})\equiv H(\mathbf{r})+4$. This makes it natural to define the
$2\pi$ periodic rescaled field
$\varphi(\mathbf{r}_{i})=\frac{\pi}{2}H(\mathbf{r}_{i})$, which can be
interpreted as the orientation angle of the dimer connected to a given site.
As shown in Fig. 3, we can assign to each site in the odd sublattice the
average height of the four plaquettes that surround it. This assignment takes
four distinct values $\left\\{0,\frac{\pi}{2},\pi,\frac{3\pi}{2}\right\\}$,
respectively corresponding to $\\{\text{right, up, left, down}\\}$ oriented
dimers connected to the site.
The mapping can also be realized in terms of the electric fields on the links,
see Fig. 3. Imagine giving each dimer a fixed orientation from even to odd
sites. Then the above mapping corresponds to assigning 3 units of electric
field pointing along the orientation of the dimer (_i.e._ flowing from the
even to the odd sublattice site), and 1 unit of electric field, flowing out
along the three remaining unoccupied links connected to each odd site. It
follows that for fully packed dimers these electric fields locally conserve
the flux, _i.e._ have a vanishing lattice divergence
$\nabla\cdot\mathbf{E}=0$, which is the Gauss’ law. Observe that the height
can be treated as a scalar potential for this field:
$\nabla\varphi=\mathbf{E}$ if we identify $\mathbf{E}=(E_{|},E_{-})$, that is:
the gradient of the height field in x-direction is the y-component of the
electric field, and _vice versa_ (see Fig. 3). This height mapping serves as a
basis for deriving the continuum effective field theory, ultimately giving the
sine-Gordon action in the main text.PhysRevB.76.134514 ; fradkin_2013
### C.2 Order parameters from RSMI-optimal coarse-graining filters
As mentioned in the main text, any two-component vector of filters with the
components drawn from among the columnar (C) and the two plaquette filters
(P1, P2) defines a coarse-graining which deterministically encodes the four
columnar ground-states without any ambiguity, as shown in Tab.1 in the main
text. As all these filters recover two bits of information, they occur
degenerately in the long-range ordered phase (LRO).
In fact, the filters can be interpreted as components of an order parameter,
which distinguish the four distinct sectors of the phase space arising due to
the spontaneous breaking of the $C_{4}$ symmetry. More concretely, one can
define a global two component _plaquette_ order parameter $(D_{1},D_{2})$:
$D_{i}:=\mathbb{E}\left[\frac{1}{N_{\mathcal{V}}}\sum_{k}\tau\circ\left(\Lambda_{i}\cdot\mathcal{V}_{k}\right)\right],$
(16)
using _e.g._ $\Lambda_{1}=\Lambda_{\rm C}$ or $\Lambda_{1}=\Lambda_{\rm P1}$,
and $\Lambda_{2}=\Lambda_{\rm P2}$, where the sum inside the expectation is
taken over all disjoint blocks $\mathcal{V}_{k}$ on the lattice, and
$N_{\mathcal{V}}$ is the number of such blocks. The norm of this order
parameter is plotted as order parameter $P$ in Fig. 2.f in the main text. The
columnar filter $\Lambda_{\rm C}$ alone defines the dimer orientational
symmetry breaking (DSB) order parameter, exactly equivalent to the one
introduced by Alet _et al._PhysRevE.74.041124 (also plotted in Fig. 2.f in
the main text):
${\rm DSB}:=\mathbb{E}\left[\sum_{k}\tau\circ\left(\Lambda_{\rm
C}\cdot\mathcal{V}_{k}\right)\right].$ (17)
Although both columnar and plaquette filters signal the columnar order with a
non-vanishing expectation value, their behavior differs above the BKT
transition temperature. As shown in Fig.2.f in the main text, while DSB decays
quickly to $0$ for $T>T_{\scriptscriptstyle\rm BKT}$ even for small lattices
the plaquette ordered parameter decays much more slowly. This distinction is
readily understood. The lowest-lying excitations above the columnar ground
states are plaquette flips, which cost two units of energy. However, the
mappings defined by the $\Lambda_{\rm P1/P2}$ filters are invariant under
subsets of plaquette flips in the configuration, and therefore the expectation
stays non-zero for finite-size systems at $T>T_{\scriptscriptstyle\rm BKT}$
(in an infinite system the plaquette order parameter also decays to zero).
### C.3 Discovering the relevant operators
##### Plaquette/columnar filters, and the electric charge operators.
Table 4: Scaling dimensions of the electric charge operators at the BKT transition, and at the free dimer point. electric charge operator | $\mathcal{O}_{1}$ | $\mathcal{O}_{2}$
---|---|---
$T\to\infty$ | $d_{1}=1$ | $d_{2}=4$
$T=T_{\scriptscriptstyle\rm BKT}$ | $d_{1}=\frac{1}{8}$ | $d_{2}=\frac{1}{2}$
For $T\geq T_{\scriptscriptstyle\rm BKT}$, the temperature-dependent RG
scaling dimensions of the $\mathcal{O}_{n=1,2}$ operators, given in Ref.
PhysRevB.76.134514, , are proportional to $n^{2}$. We list the values they
take at the BKT transition, and in the limit of infinite $T$ in Tab. 4. This
field-theoretic result underlies the evolution of the RSMI-optimal filters
with $T$. The orientational symmetry breaking and plaquette operators are
degenerate in the long-range ordered phase; in the critical phase, however,
the columnar order parameter corresponding to $n=2$ has the higher scaling
dimension, therefore its correlations decay faster, and this is why we obtain
the plaquette filters but not the columnar ones beyond the BKT point. We thus
verify for the case of interacting fully-packed dimers that the RSMI-NE indeed
finds the operators with the lowest-scaling dimension, _i.e._ the RG relevant
operators.
We emphasize that the extracted RSMI filters can be used exactly as one
usually uses operators. For instance, their correlation functions can be
evaluated numerically from the samples, to recover (fit) their scaling
dimensions. Though above we gave an explicit mapping between the filters and
the scaling operators of the dimer model, so it is superfluous in this case,
in more complex scenarios, with systems not fully understood, this would be
necessary. As an illustration, in Fig. 4 we evaluate and plot the correlation
function (on a $128\times 128$ dimer system, utilizing size $4\times 4$
filters) of the plaquette operators in the limit $T\rightarrow\infty$, and
verify accurately it does indeed decay as $r^{-2}$, _i.e._ the we extract the
scaling dimension $d_{1}=1$ for the $\mathcal{O}_{1}$ operator, a
predicted.PhysRevB.76.134514 The decay of correlation function of the
columnar filters is much faster, consistent with the prediction of $r^{-8}$
(we do not fit the exponent, as the fast decay requires better sample
statistics). Thus, RSMI-NE allows to gain a formal understanding of the long-
distance properties of the system, unsupervised and without inputting any
prior knowledge.
Figure 4: Using RSMI filters to compute correlation functions and scaling
dimensions: The plaquette and the columnar coarse-graining filters correspond
to the discretized $\mathcal{O}_{1}$ and $\mathcal{O}_{2}$ operators, with
scaling dimensions $d_{1}=1,d_{2}=4$ at $T\to\infty$. (a) The correlators
obtained using these filters confirm this picture. We fit a power-law decay to
the plaquette correlator (on a $128\times 128$ dimer system, using size
$4\times 4$ filters), obtaining a best-fit of $2.00074$, in excellent
agreement with the field theoretic result $2d_{1}=2$, thus extracting the
scaling dimension. The columnar correlator is consistent with the theoretical
prediction of $r^{-8}$, but the fast decay would necessitate larger sample
size for an accurate fitting of the exponent. (b) More precisely, the
correlation function for $\mathcal{O}_{1,2}$ on the height field $\varphi$ is
directly computed from the MC samples as the correlator of the trained neural
networks parametrizing the coarse-graining filters $\Lambda_{\rm P,C}$,
evaluated at spatial separations $r$ on the dimer configurations $C$.
##### Staggered filters, electric fields.
The local action of pristine staggered filters which are RSMI-optimal at
$T\to\infty$ has the form
$\displaystyle\Lambda_{\rm S+}\cdot\mathcal{V}(\mathbf{r})$
$\displaystyle=(-1)^{x+y}N_{|}(\mathbf{r}),$ (18) $\displaystyle\Lambda_{\rm
S-}\cdot\mathcal{V}(\mathbf{r})$
$\displaystyle=(-1)^{x+y+1}N_{-}(\mathbf{r}).$ (19)
Note that here $\left(\Lambda_{\rm S+}:=\frac{\Lambda_{\rm S1}+\Lambda_{\rm
S2}}{\sqrt{2}},\,\Lambda_{\rm S-}:=\frac{\Lambda_{\rm S1}-\Lambda_{\rm
S2}}{\sqrt{2}}\right)$ corresponds to the orthogonal basis $(E_{x},E_{y})$ for
the electrical field.333S1 and S2 refers to the pristine staggered filters
introduced in the main text. On the other hand, any unitary rotation
$R(\vartheta)(\Lambda_{\rm S+},\Lambda_{\rm
S-})\sim\left(\cos(\vartheta)E_{x}-\sin(\vartheta)E_{y},\sin(\vartheta)E_{x}+\cos(\vartheta)E_{y}\right)$
recovers the same amount of RSMI, indicating the emergent $U(1)$ symmetry of
the free dimer model. For the interested reader, we point Ref. rsmine, for
more details.
Ref. PhysRevB.76.134514, gives an expansion of the dimer densities on the
lattice in terms of the observables of the height-field continuum model:
$\displaystyle
N_{|}(\mathbf{r})=\frac{1}{4}+\frac{(-1)^{x+y+1}}{2\pi}\partial_{x}\varphi(\mathbf{r})+(-1)^{y}\sin\varphi(\mathbf{r})$
, (20) $\displaystyle
N_{-}(\mathbf{r})=\frac{1}{4}+\frac{(-1)^{x+y}}{2\pi}\partial_{y}\varphi(\mathbf{r})+(-1)^{x}\cos\varphi(\mathbf{r})$
. (21)
Here, $1/4$ is the average dimer density (due to the $C_{4}$ symmetry). The
terms $(-1)^{y}\sin\varphi$ and $(-1)^{x}\cos\varphi$, respectively, select
the vertical and horizontal dimers connected to point $\mathbf{r}$ (_cf._ Fig.
3.b).
Substituting these in Eq. 18 and averaging over the degrees of freedom in the
block $\mathcal{V}$ to which the filter is applied, we obtain the coarse-
grained degrees of freedom:
$\displaystyle\mathcal{H}_{1}\sim\sum_{\mathbf{r}\in\mathcal{V}}\Lambda_{\rm
S+}\cdot\mathcal{V}(\mathbf{r})$
$\displaystyle=\sum_{\mathbf{r}\in\mathcal{V}}\left[\frac{(-1)^{x+y}}{4}+\frac{\partial_{x}\varphi(\mathbf{r})}{2\pi}+(-1)^{x}\sin\varphi(\mathbf{r})\right]=\sum_{\mathbf{r}\in\mathcal{V}}\left[\frac{\partial_{x}\varphi(\mathbf{r})}{2\pi}+(-1)^{x}\sin\varphi(\mathbf{r})\right],$
(22)
$\displaystyle\mathcal{H}_{2}\sim\sum_{\mathbf{r}\in\mathcal{V}}\Lambda_{\rm
S-}\cdot\mathcal{V}(\mathbf{r})$
$\displaystyle=\sum_{\mathbf{r}\in\mathcal{V}}\left[\frac{(-1)^{x+y}}{4}+\frac{\partial_{y}\varphi(\mathbf{r})}{2\pi}+(-1)^{y}\cos\varphi(\mathbf{r})\right]=\sum_{\mathbf{r}\in\mathcal{V}}\left[\frac{\partial_{y}\varphi(\mathbf{r})}{2\pi}+(-1)^{y}\cos\varphi(\mathbf{r})\right],$
(23)
where the $\sim$ symbol is used to denote that the coarse-grained variable
$\mathcal{H}$ is the result of applying the binary mapping $\tau$ to the right
hand side. We see that the first term in the summand vanishes since there are
an even number of links in the region $\mathcal{V}$. The last term probes the
tilt of the height configuration either in the $x$ or in the $y$ direction.
This is because the term $(-1)^{x}\sin\varphi(\mathbf{r})$ has a non-zero
average only if the region $\mathcal{V}$ contains a vertical dimer with a
missing parallel neighbor, in which case the height profile has a slope in the
$x$ direction, while it vanishes for a uniform height field (similarly for the
other component in the $y$ direction). Equivalently, only not-constant-in-x
function $\varphi(\mathbf{r})$ survives averaging with $-1^{x}$ over the
block, and so under the average we may replace $\sin\varphi(\mathbf{r})$ with
$\partial_{x}\varphi(\mathbf{r})$. Hence, in total:
$\mathcal{H}\propto\tau\circ\nabla\left\langle\varphi(\mathbf{r})\right\rangle_{\mathbf{r}\in\mathcal{V}},$
(24)
where the average is over the sites in the block and $\tau$ maps the gradient
into binary values. In other words, the staggered filters extract the coarse-
grained gradients of the height field, _i.e._ they correspond to the electric
fields described in Fig. 3.a.
Acknowledgements This work was supported by a grant from the Swiss National
Supercomputing Centre (CSCS) under project ID eth5b.
## References
* (1) Kadanoff, L. P. Scaling laws for Ising models near ${T}_{c}$. _Phys. Phys. Fiz._ 2, 263–272 (1966).
* (2) Wilson, K. G. & Kogut, J. The renormalization group and the $\epsilon$ expansion. _Phys. Rep._ 12, 75 – 199 (1974).
* (3) Wilson, K. G. The renormalization group: Critical phenomena and the Kondo problem. _Rev. Mod. Phys._ 47, 773–840 (1975).
* (4) Fisher, M. E. Renormalization group theory: Its basis and formulation in statistical physics. _Rev. Mod. Phys._ 70, 653–681 (1998).
* (5) Koch-Janusz, M. & Ringel, Z. Mutual information, neural networks and the renormalization group. _Nat. Phys._ 14, 578–582 (2018).
* (6) Lenggenhager, P. M., Gökmen, D. E., Ringel, Z., Huber, S. D. & Koch-Janusz, M. Optimal renormalization group transformation from information theory. _Phys. Rev. X_ 10, 011037 (2020).
* (7) Gordon, A., Banerjee, A., Koch-Janusz, M. & Ringel, Z. Relevance in the renormalization group and in information theory. _Phys. Rev. Lett._ 126, 240601 (2021).
* (8) Poole, B., Ozair, S., van den Oord, A., Alemi, A. A. & Tucker, G. On variational bounds of mutual information (2019). eprint arXiv:1905.06922.
* (9) Coveney, P. V., Dougherty, E. R. & Highfield, R. R. Big data need big theory too. _Philos. Trans. R. Soc. A_ 374 (2016).
* (10) Alet, F. _et al._ Interacting classical dimers on the square lattice. _Phys. Rev. Lett._ 94, 235702 (2005).
* (11) Alet, F., Ikhlef, Y., Jacobsen, J. L., Misguich, G. & Pasquier, V. Classical dimers with aligning interactions on the square lattice. _Phys. Rev. E_ 74, 041124 (2006).
* (12) Donsker, M. D. & Varadhan, S. R. S. Asymptotic evaluation of certain Markov process expectations for large time. IV. _Commun. Pure Appl. Math._ 36, 183–212 (1983).
* (13) Barber, D. & Agakov, F. V. Information maximization in noisy channels: A variational approach. In Thrun, S., Saul, L. K. & Schölkopf, B. (eds.) _Advances in Neural Information Processing Systems 16_ , 201–208 (MIT Press, 2004).
* (14) Nguyen, X., Wainwright, M. J. & Jordan, M. I. Estimating divergence functionals and the likelihood ratio by penalized convex risk minimization. In Platt, J. C., Koller, D., Singer, Y. & Roweis, S. T. (eds.) _Advances in Neural Information Processing Systems 20_ , 1089–1096 (Neural Information Processing Systems Foundation, 2009).
* (15) Belghazi, M. I. _et al._ MINE: Mutual Information Neural Estimation (2018). eprint arXiv:1801.04062.
* (16) Lu, P. Y., Kim, S. & Soljačić, M. Extracting interpretable physical parameters from spatiotemporal systems using unsupervised learning. _Phys. Rev. X_ 10, 031056 (2020).
* (17) Jang, E., Gu, S. & Poole, B. Categorical Reparameterization with Gumbel-Softmax (2016). eprint arXiv:1611.01144.
* (18) Kasteleyn, P. W. Dimer Statistics and Phase Transitions. _J. Math. Phys._ 4, 287–293 (1963).
* (19) Blote, H. W. J. & Hilborst, H. J. Roughening transitions and the zero-temperature triangular Ising antiferromagnet. _J. Phys. A_ 15, L631–L637 (1982).
* (20) Henley, C. L. Relaxation time for a dimer covering with height representation. _J. Stat. Phys._ 89, 483–507 (1997).
* (21) Kenyon, R. Dominos and the Gaussian Free Field. _Ann. Probab._ 29, 1128–1137 (2001).
* (22) Kenyon, R. The Laplacian and Dirac operators on critical planar graphs. _Invent. Math._ 150, 409–439 (2002).
* (23) Cimasoni, D. & Reshetikhin, N. Dimers on Surface Graphs and Spin Structures. I. _Commun. Math. Phys._ 275, 187–208 (2007).
* (24) Powell, S. & Chalker, J. T. SU(2)-Invariant Continuum Theory for an Unconventional Phase Transition in a Three-Dimensional Classical Dimer Model. _Phys. Rev. Lett._ 101, 155702 (2008).
* (25) Chen, G., Gukelberger, J., Trebst, S., Alet, F. & Balents, L. Coulomb gas transitions in three-dimensional classical dimer models. _Phys. Rev. B_ 80, 045112 (2009).
* (26) Powell, S. & Chalker, J. T. Classical to quantum mapping for an unconventional phase transition in a three-dimensional classical dimer model. _Phys. Rev. B_ 80, 134413 (2009).
* (27) Sreejith, G. J., Powell, S. & Nahum, A. Emergent SO(5) Symmetry at the Columnar Ordering Transition in the Classical Cubic Dimer Model. _Phys. Rev. Lett._ 122, 080601 (2019).
* (28) Rokhsar, D. S. & Kivelson, S. A. Superconductivity and the Quantum Hard-Core Dimer Gas. _Phys. Rev. Lett._ 61, 2376–2379 (1988).
* (29) Fradkin, E., Huse, D. A., Moessner, R., Oganesyan, V. & Sondhi, S. L. Bipartite Rokhsar–Kivelson points and Cantor deconfinement. _Phys. Rev. B_ 69, 224415 (2004).
* (30) Fradkin, E. _Field Theories of Condensed Matter Physics_ (Cambridge University Press, 2013), 2 edn.
* (31) Gökmen, D. E., Ringel, Z., Huber, S. D. & Koch-Janusz, M. Symmetries and phase diagrams with real-space mutual information neural estimation (2021). eprint arXiv:2103.16887.
* (32) Wilms, J., Troyer, M. & Verstraete, F. Mutual information in classical spin models. _J. Stat. Mech.: Theory Exp._ 2011, P10011 (2011).
* (33) Lau, H. W. & Grassberger, P. Information theoretic aspects of the two-dimensional Ising model. _Phys. Rev. E_ 87, 022128 (2013).
* (34) Papanikolaou, S., Luijten, E. & Fradkin, E. Quantum criticality, lines of fixed points, and phase separation in doped two-dimensional quantum dimer models. _Phys. Rev. B_ 76, 134514 (2007).
* (35) Rajesh, R. & Majumdar, S. N. Exact phase diagram of a model with aggregation and chipping. _Phys. Rev. E_ 63, 036114 (2001).
* (36) Toner, J. & Tu, Y. Flocks, herds, and schools: A quantitative theory of flocking. _Phys. Rev. E_ 58, 4828–4858 (1998).
* (37) Bialek, W. _et al._ Statistical mechanics for natural flocks of birds. _Proc. Natl. Acad. Sci. U.S.A._ 109, 4786–4791 (2012).
* (38) Peters, O. & Neelin, J. D. Critical phenomena in atmospheric precipitation. _Nat. Phys._ 2, 393–396. 5 p (2006).
* (39) Dewar, R. Information theory explanation of the fluctuation theorem, maximum entropy production and self-organized criticality in non-equilibrium stationary states. _J. Phys. A_ 36, 631–641 (2003).
* (40) Nir, A., Sela, E., Beck, R. & Bar-Sinai, Y. Machine-learning iterative calculation of entropy for physical systems. _Proc. Natl. Acad. Sci. U.S.A._ 117, 30234–30240 (2020).
* (41) Tupitsyn, I. S., Kitaev, A., Prokof’ev, N. V. & Stamp, P. C. E. Topological multicritical point in the phase diagram of the toric code model and three-dimensional lattice gauge higgs model. _Phys. Rev. B_ 82, 085114 (2010).
* (42) Moon, E.-G. & Xu, C. Exotic continuous quantum phase transition between ${\mathbb{z}}_{2}$ topological spin liquid and néel order. _Phys. Rev. B_ 86, 214414 (2012).
* (43) Agrawal, U., Gopalakrishnan, S. & Vasseur, R. Universality and quantum criticality in quasiperiodic spin chains. _Nat. Commun._ 11, 2225 (2020).
* (44) Flicker, F., Simon, S. H. & Parameswaran, S. A. Classical dimers on penrose tilings. _Phys. Rev. X_ 10, 011005 (2020).
* (45) van den Oord, A., Li, Y. & Vinyals, O. Representation learning with contrastive predictive coding (2018). eprint arXiv:1807.03748.
* (46) Gutmann, M. & Hyvärinen, A. Noise-contrastive estimation: A new estimation principle for unnormalized statistical models. In Teh, Y. W. & Titterington, M. (eds.) _Proceedings of the Thirteenth International Conference on Artificial Intelligence and Statistics_ , vol. 9 of _Proceedings of Machine Learning Research_ , 297–304 (JMLR Workshop and Conference Proceedings, 2010).
* (47) Goodfellow, I., Bengio, Y. & Courville, A. _Deep Learning_ (MIT Press, 2016). URL http://www.deeplearningbook.org.
* (48) Maddison, C. J., Tarlow, D. & Minka, T. $\text{A}^{\ast}$ sampling. In Ghahramani, Z., Welling, M., Cortes, C., Lawrence, N. D. & Weinberger, K. Q. (eds.) _Advances in Neural Information Processing Systems 27_ , 3086–3094 (Curran Associates, Inc., 2014).
* (49) Gumbel, E. J. The maxima of the mean largest value and of the range. _Ann. Math. Stat._ 25, 76–84 (1954).
* (50) Kingma, D. P. & Ba, J. Adam: A method for stochastic optimization (2014). eprint arXiv:1412.6980.
## References
* (1) Kadanoff, L. P. Scaling laws for Ising models near ${T}_{c}$. _Phys. Phys. Fiz._ 2, 263–272 (1966).
* (2) Wilson, K. G. & Kogut, J. The renormalization group and the $\epsilon$ expansion. _Phys. Rep._ 12, 75 – 199 (1974).
* (3) Wilson, K. G. The renormalization group: Critical phenomena and the Kondo problem. _Rev. Mod. Phys._ 47, 773–840 (1975).
* (4) Fisher, M. E. Renormalization group theory: Its basis and formulation in statistical physics. _Rev. Mod. Phys._ 70, 653–681 (1998).
* (5) Koch-Janusz, M. & Ringel, Z. Mutual information, neural networks and the renormalization group. _Nat. Phys._ 14, 578–582 (2018).
* (6) Lenggenhager, P. M., Gökmen, D. E., Ringel, Z., Huber, S. D. & Koch-Janusz, M. Optimal renormalization group transformation from information theory. _Phys. Rev. X_ 10, 011037 (2020).
* (7) Gordon, A., Banerjee, A., Koch-Janusz, M. & Ringel, Z. Relevance in the renormalization group and in information theory. _Phys. Rev. Lett._ 126, 240601 (2021).
* (8) Poole, B., Ozair, S., van den Oord, A., Alemi, A. A. & Tucker, G. On variational bounds of mutual information (2019). eprint arXiv:1905.06922.
* (9) Coveney, P. V., Dougherty, E. R. & Highfield, R. R. Big data need big theory too. _Philos. Trans. R. Soc. A_ 374 (2016).
* (10) Alet, F. _et al._ Interacting classical dimers on the square lattice. _Phys. Rev. Lett._ 94, 235702 (2005).
* (11) Alet, F., Ikhlef, Y., Jacobsen, J. L., Misguich, G. & Pasquier, V. Classical dimers with aligning interactions on the square lattice. _Phys. Rev. E_ 74, 041124 (2006).
* (12) Donsker, M. D. & Varadhan, S. R. S. Asymptotic evaluation of certain Markov process expectations for large time. IV. _Commun. Pure Appl. Math._ 36, 183–212 (1983).
* (13) Barber, D. & Agakov, F. V. Information maximization in noisy channels: A variational approach. In Thrun, S., Saul, L. K. & Schölkopf, B. (eds.) _Advances in Neural Information Processing Systems 16_ , 201–208 (MIT Press, 2004).
* (14) Nguyen, X., Wainwright, M. J. & Jordan, M. I. Estimating divergence functionals and the likelihood ratio by penalized convex risk minimization. In Platt, J. C., Koller, D., Singer, Y. & Roweis, S. T. (eds.) _Advances in Neural Information Processing Systems 20_ , 1089–1096 (Neural Information Processing Systems Foundation, 2009).
* (15) Belghazi, M. I. _et al._ MINE: Mutual Information Neural Estimation (2018). eprint arXiv:1801.04062.
* (16) Lu, P. Y., Kim, S. & Soljačić, M. Extracting interpretable physical parameters from spatiotemporal systems using unsupervised learning. _Phys. Rev. X_ 10, 031056 (2020).
* (17) Jang, E., Gu, S. & Poole, B. Categorical Reparameterization with Gumbel-Softmax (2016). eprint arXiv:1611.01144.
* (18) Kasteleyn, P. W. Dimer Statistics and Phase Transitions. _J. Math. Phys._ 4, 287–293 (1963).
* (19) Blote, H. W. J. & Hilborst, H. J. Roughening transitions and the zero-temperature triangular Ising antiferromagnet. _J. Phys. A_ 15, L631–L637 (1982).
* (20) Henley, C. L. Relaxation time for a dimer covering with height representation. _J. Stat. Phys._ 89, 483–507 (1997).
* (21) Kenyon, R. Dominos and the Gaussian Free Field. _Ann. Probab._ 29, 1128–1137 (2001).
* (22) Kenyon, R. The Laplacian and Dirac operators on critical planar graphs. _Invent. Math._ 150, 409–439 (2002).
* (23) Cimasoni, D. & Reshetikhin, N. Dimers on Surface Graphs and Spin Structures. I. _Commun. Math. Phys._ 275, 187–208 (2007).
* (24) Powell, S. & Chalker, J. T. SU(2)-Invariant Continuum Theory for an Unconventional Phase Transition in a Three-Dimensional Classical Dimer Model. _Phys. Rev. Lett._ 101, 155702 (2008).
* (25) Chen, G., Gukelberger, J., Trebst, S., Alet, F. & Balents, L. Coulomb gas transitions in three-dimensional classical dimer models. _Phys. Rev. B_ 80, 045112 (2009).
* (26) Powell, S. & Chalker, J. T. Classical to quantum mapping for an unconventional phase transition in a three-dimensional classical dimer model. _Phys. Rev. B_ 80, 134413 (2009).
* (27) Sreejith, G. J., Powell, S. & Nahum, A. Emergent SO(5) Symmetry at the Columnar Ordering Transition in the Classical Cubic Dimer Model. _Phys. Rev. Lett._ 122, 080601 (2019).
* (28) Rokhsar, D. S. & Kivelson, S. A. Superconductivity and the Quantum Hard-Core Dimer Gas. _Phys. Rev. Lett._ 61, 2376–2379 (1988).
* (29) Fradkin, E., Huse, D. A., Moessner, R., Oganesyan, V. & Sondhi, S. L. Bipartite Rokhsar–Kivelson points and Cantor deconfinement. _Phys. Rev. B_ 69, 224415 (2004).
* (30) Fradkin, E. _Field Theories of Condensed Matter Physics_ (Cambridge University Press, 2013), 2 edn.
* (31) Gökmen, D. E., Ringel, Z., Huber, S. D. & Koch-Janusz, M. Symmetries and phase diagrams with real-space mutual information neural estimation (2021). eprint arXiv:2103.16887.
* (32) Wilms, J., Troyer, M. & Verstraete, F. Mutual information in classical spin models. _J. Stat. Mech.: Theory Exp._ 2011, P10011 (2011).
* (33) Lau, H. W. & Grassberger, P. Information theoretic aspects of the two-dimensional Ising model. _Phys. Rev. E_ 87, 022128 (2013).
* (34) Papanikolaou, S., Luijten, E. & Fradkin, E. Quantum criticality, lines of fixed points, and phase separation in doped two-dimensional quantum dimer models. _Phys. Rev. B_ 76, 134514 (2007).
* (35) Rajesh, R. & Majumdar, S. N. Exact phase diagram of a model with aggregation and chipping. _Phys. Rev. E_ 63, 036114 (2001).
* (36) Toner, J. & Tu, Y. Flocks, herds, and schools: A quantitative theory of flocking. _Phys. Rev. E_ 58, 4828–4858 (1998).
* (37) Bialek, W. _et al._ Statistical mechanics for natural flocks of birds. _Proc. Natl. Acad. Sci. U.S.A._ 109, 4786–4791 (2012).
* (38) Peters, O. & Neelin, J. D. Critical phenomena in atmospheric precipitation. _Nat. Phys._ 2, 393–396. 5 p (2006).
* (39) Dewar, R. Information theory explanation of the fluctuation theorem, maximum entropy production and self-organized criticality in non-equilibrium stationary states. _J. Phys. A_ 36, 631–641 (2003).
* (40) Nir, A., Sela, E., Beck, R. & Bar-Sinai, Y. Machine-learning iterative calculation of entropy for physical systems. _Proc. Natl. Acad. Sci. U.S.A._ 117, 30234–30240 (2020).
* (41) Tupitsyn, I. S., Kitaev, A., Prokof’ev, N. V. & Stamp, P. C. E. Topological multicritical point in the phase diagram of the toric code model and three-dimensional lattice gauge higgs model. _Phys. Rev. B_ 82, 085114 (2010).
* (42) Moon, E.-G. & Xu, C. Exotic continuous quantum phase transition between ${\mathbb{z}}_{2}$ topological spin liquid and néel order. _Phys. Rev. B_ 86, 214414 (2012).
* (43) Agrawal, U., Gopalakrishnan, S. & Vasseur, R. Universality and quantum criticality in quasiperiodic spin chains. _Nat. Commun._ 11, 2225 (2020).
* (44) Flicker, F., Simon, S. H. & Parameswaran, S. A. Classical dimers on penrose tilings. _Phys. Rev. X_ 10, 011005 (2020).
* (45) van den Oord, A., Li, Y. & Vinyals, O. Representation learning with contrastive predictive coding (2018). eprint arXiv:1807.03748.
* (46) Gutmann, M. & Hyvärinen, A. Noise-contrastive estimation: A new estimation principle for unnormalized statistical models. In Teh, Y. W. & Titterington, M. (eds.) _Proceedings of the Thirteenth International Conference on Artificial Intelligence and Statistics_ , vol. 9 of _Proceedings of Machine Learning Research_ , 297–304 (JMLR Workshop and Conference Proceedings, 2010).
* (47) Goodfellow, I., Bengio, Y. & Courville, A. _Deep Learning_ (MIT Press, 2016). URL http://www.deeplearningbook.org.
* (48) Maddison, C. J., Tarlow, D. & Minka, T. $\text{A}^{\ast}$ sampling. In Ghahramani, Z., Welling, M., Cortes, C., Lawrence, N. D. & Weinberger, K. Q. (eds.) _Advances in Neural Information Processing Systems 27_ , 3086–3094 (Curran Associates, Inc., 2014).
* (49) Gumbel, E. J. The maxima of the mean largest value and of the range. _Ann. Math. Stat._ 25, 76–84 (1954).
* (50) Kingma, D. P. & Ba, J. Adam: A method for stochastic optimization (2014). eprint arXiv:1412.6980.
|
Present address: ]Centre for Engineered Quantum Systems, School of Mathematics
and Physics, University of Queensland, QLD 4072, Australia
# A Bright Source of Telecom Single Photons Based on Quantum Frequency
Conversion
Christopher L. Morrison<EMAIL_ADDRESS>Markus Rambach [ Zhe Xian Koong
Francesco Graffitti Fiona Thorburn Ajoy K. Kar Institute of Photonics and
Quantum Sciences, School of Engineering and Physical Sciences, Heriot-Watt
University, Edinburgh EH14 4AS, UK Yong Ma College of Optoelectronic
Engineering, Chongqing University of Posts and Telecommunications, Chongqing
400065, China Suk-In Park Jin Dong Song Center for Opto-Electronic
Materials and Devices Research, Korea Institute of Science and Technology,
Seoul 02792, Republic of Korea Nick G. Stoltz Materials Department,
University of California, Santa Barbara, California 93106, USA Dirk
Bouwmeester Huygens-Kamerlingh Onnes Laboratory, Leiden University, P.O. Box
9504, 2300 RA Leiden, Netherlands Department of Physics, University of
California, Santa Barbara, California 93106, USA Alessandro Fedrizzi Brian
D. Gerardot Institute of Photonics and Quantum Sciences, School of
Engineering and Physical Sciences, Heriot-Watt University, Edinburgh EH14 4AS,
UK
(August 27, 2024)
###### Abstract
On-demand indistinguishable single photon sources are essential for quantum
networking and communication. Semiconductor quantum dots are among the most
promising candidates, but their typical emission wavelength renders them
unsuitable for use in fibre networks. Here, we present quantum frequency
conversion of near-infrared photons from a bright quantum dot to the
telecommunication C-band, allowing integration with existing fibre
architectures. We use a custom-built, tunable 2400 nm seed laser to convert
single photons from 942 nm to 1550 nm in a difference frequency generation
process. We achieve an end-to-end conversion efficiency of $\sim$35%,
demonstrate count rates approaching 1 MHz at 1550 nm with
$g^{\left(2\right)}\left(0\right)=0.04$, and achieve Hong-Ou-Mandel
visibilities of 60%. We expect this scheme to be preferable to quantum dot
sources directly emitting at telecom wavelengths for fibre based quantum
networking.
Semiconductor quantum dots (QDs) are a leading technology for bright,
indistinguishable single-photon sources. QDs emitting in the 920 nm - 980 nm
window have been used as a source of single photons with count rates upwards
of 10 MHz, indistinguishability greater than 95%, and low multi-photon
contributions on the order of 1% [1, 2, 3]. Yet, in order to be compatible
with existing telecommunication technology, an ideal single-photon source
should operate in the telecommunication C-band, around 1550 nm, where fibre
loss is minimal. While there has been recent progress in producing QD sources
that directly emit single photons in the C-band [4, 5, 6], achieving high-
quality single-photon emission at high rates remains an open challenge, with
the best-performing single photon sources to date being confined to near-
infrared (NIR) [7] wavelengths.
One route to bridge the gap to the C-band is quantum frequency conversion
(QFC), converting single photons from a NIR QD to telecommunication
wavelengths. Quantum frequency conversion is a nonlinear process where a
single-photon input is mixed with a strong seed beam producing a single-photon
output at either the sum or difference frequency. QFC can in principle be
noise free and therefore preserve the quantum statistics of near-infrared
emitters. Frequency conversion of QD sources has been demonstrated from 700 nm
to the telecommunication O-band [8], and from 900 nm to the C-band [9, 10],
culminating in the remote two-photon interference between independent
downconverted QD sources [11]. Spin-photon entanglement between a 910 nm QD
and a 1560-nm photon has been demonstrated using frequency conversion seeded
by ultrafast pulses [12, 13]. QFC of QDs has also been demonstrated in nano-
photonic circuits using four-wave mixing in silicon nitride [14]. However, a
source that is simultaneously bright, pure and coherent has not been
demonstrated in QDs emitting directly at telecom wavelengths. Here we
demonstrate a frequency-converted InGaAs quantum dot source approaching 1 MHz
count rates at 1550 nm, with $g^{\left(2\right)}\left(0\right)$ around 4% and
HOM visibilities of 60%.
Figure 1: Difference-frequency generation schematic. Blue lines indicate the
optical path of the QD photons. Pink lines represent the path of the 2401 nm
seed light. The polarisation of both beams is aligned to the extraordinary
axis of the ppLN crystal with a quarter-wave plate (QWP) and half-wave plate
(HWP). A 100 mm focal length lens is used in the QD beam path to mode match
the seed beam at the waveguide facet. The green lines represent the converted
1550 nm light after the frequency conversion. The converted light is sent
through two short pass filters at 2050 nm (SP 2050), a 1400 nm long pass
filter (LP 1400) and a bandpass filter at 1550 nm (BP 1550) before being
collected in a single mode fibre for detection. The inset shows the
experimental layout used to produce the 2401 nm laser light for QFC. The seed
laser is pumped by a commercial Thulium fibre laser (yellow lines). The pump
beam polarisation is prepared with a HWP to reduce loss in the laser cavity
due to Fresnel reflections. The light enters the cavity through a partially
reflective curved mirror which acts as the input coupler at 1900 nm and a
focusing mirror at 2401 nm. A longpass filter with a cutoff wavelength of 2000
nm (LP 2000) is placed after the output coupler to remove unabsorbed 1900 nm
pump light.
For difference frequency generation, energy conservation demands
$\left(1/\lambda_{\text{in}}-1/\lambda_{\text{seed}}\right)^{-1}=\lambda_{\text{out}}$.
Our 942 nm InGaAs QD source requires a seed wavelength of 2401 nm to generate
output photons at 1550 nm. The seed beam is produced in a chromium doped zinc
selenide (Cr:ZnSe) laser while the difference frequency generation occurs in a
periodically poled lithium niobate (ppLN) waveguide.
Near-infrared photons are generated by a single self-assembled InGaAs/GaAs QD
coupled to a high quality ($Q\approx 4.4\times 10^{4}$) oxide-apertured
micropillar cavity, as shown in Fig. 1. The QDs are embedded in a p-i-n diode
structure [15, 16] which enables charge control and tuning of the QD emission
to the cavity mode via the quantum-confined Stark effect. The sample is kept
at a temperature of $4\,\mathrm{K}$ in a closed-cycle helium flow cryostat. A
dark-field confocal microscope is used to excite and collect the scattered
photons from the QD before filtering in a cross-polarisation scheme with a
$\sim 10^{7}$ extinction ratio to suppress the excitation laser.
Pulsed excitation of the QD is performed using a mode-locked titanium:sapphire
laser with a repetition rate of 80.3 MHz and pulse duration of 10 ps. The QD
output is then detected by superconducting nanowire single photon detectors
(SNSPDs), with a nominal quantum efficiency of $\sim 90\,\%$ at 950 nm. We
exploit a neutral exciton QD transition ($X^{0}$), resonantly coupled to the
cavity with a Purcell factor of $\sim 4$ and an emission wavelength of 942.33
nm. The $T_{2}$ coherence time of the emission, measured using standard
Fourier spectroscopy under $\pi$-pulse resonant excitation, shows
$T_{2}=0.348\,(2)\,\mathrm{ns}$, corresponding to a linewidth of
$915\,(5)\,\mathrm{MHz}$. This value is $\approx 1.5$ times larger than the
transform-limited linewidth ($h/T_{1}=607\,\mathrm{MHz}$), with an
independently measured lifetime of $T_{1}=0.2622\,(1)\,\mathrm{ns}$.
The seed laser for the QFC stage consists of a z-cavity resonator with a
Cr:ZnSe crystal, a gain medium with an emission spectrum spanning 1900-3300 nm
[17, 18, 19], see inset in Fig. 1. This laser design allows for both
continuous-wave [20] and mode-locked operation generating pulses as short as
43 fs [21]. Here, we operate with a narrowband CW seed to drive the QFC. Our
2401 nm laser system is pumped by a thulium-doped fibre laser (IPG Photonics
TLR-20-LP) which has a maximum CW output power of 20 W at 1900 nm. The pump
light is focused into the Cr:ZnSe crystal using a 100 mm $\text{CaF}_{2}$
lens. The cavity consists of a dichroic input coupler (50 mm radius of
curvature (ROC), transmissive at 1900 nm reflective at 2400 nm), a gold mirror
(50 mm ROC), two plane silver mirrors and the Cr:ZnSe crystal. The crystal is
placed at Brewster angle to minimise losses due to reflection. A diffraction
grating (450 lines/mm) is inserted into the cavity to control the emission
wavelength. The output coupler (Layertec) has a transmission of 60% at 2401
nm, allowing a good trade-off between the cavity enhancement and available
output power.
Fig. 1 shows our difference-frequency generation (DFG) setup. The DFG, one
special case of QFC, takes place in a 48 mm periodically-poled lithium niobate
crystal (ppLN, NTT Electronics). The chip contains multiple ridge waveguides
with poling periods ranging from 26.00 $\mu$m to 26.25 $\mu$m. These poling
periods are designed for type-0 DFG from 942 nm to 1550 nm. Quarter- and half-
wave plates are used to align the polarisation of the incoming single photons
at 942 nm and the generated pump light at 2401 nm to the extraordinary axis of
the crystal. Seed light from the laser is overlapped with single photons from
the QD using a dichroic mirror (Omega Optical) and coupled into the waveguide
using a NIR coated aspheric lens with a focal length of 11 mm. A NIR coated
lens is used to match the beam size of the single photons to the seed beam and
to compensate for the chromatic aberration of the aspheric lenses. The
converted 1550 nm light is collimated with an 11 mm NIR-coated aspheric lens
and sent towards a filtering stage.
The filtering stage consists of two shortpass filters at 2050 nm ($>$OD 4),
which are used to remove seed light impinging on the collection fibre; a
longpass filter at 1400 nm ($>$OD 5) to remove weakly phase-matched second-
harmonic generation from the seed beam and unconverted quantum dot light; and
finally, a 2.8 nm full-width-at-half-maximum (FWHM) bandpass filter ($>$OD 4,
BP 1550) to isolate the converted single photons. The converted 1550 nm light
is collected into a single mode fibre with 86% coupling efficiency and sent to
SNSPDs with a nominal quantum efficiency greater than 80%.
The DFG conversion efficiency is characterised by sending CW coherent light
from a 942 nm laser (Toptica DL Pro) into the QFC setup. For CW-seeded QFC,
the conversion efficiency is almost independent of the temporal mode of the
input light [22]. This allows characterisation with a CW beam despite the
single photons’ decaying exponential wavepacket. In the case that we can treat
the seed beam as unamplified, QFC is expressed as a beam-splitter Hamiltonian
between two different frequency modes [23]. This means the measured conversion
efficiency is independent of the input intensity. These two factors ensure the
conversion efficiency measured with a low power $\left(500\text{
$\mu$W}\right)$ CW coherent field is equivalent to the single-photon
conversion efficiency.
Figure 2: Conversion efficiency of the difference frequency generation process
as a function of seed power coupled into the waveguide. The power in the
waveguide is determined by measuring the pump power after the waveguide and
factoring out the loss through the c-coated aspheric collimation lens. The
transmission through this lens is measured to be 64 % at the pump wavelength.
The data is fitted with Eq. 1. Inset shows the signal to noise ratio for off-
resonant excitation with a measured noise count rate of $12\,(1)$ Hz/mW.
Figure 3: Photo-luminescence emission profile of the QD under 820 nm
excitation shows cavity modes up to the $5^{\mathrm{th}}$ order. The QD is
resonantly coupled to the 1st cavity mode. Inset shows the detected count rate
as a function of the excitation power when the excitation laser is resonant to
the 1st (resonant, blue) or 3rd (non-resonant, green) cavity mode. Figure 4:
Characterisation of the single photon properties before (upper row) and after
(lower row) QFC. (a) Time-resolved emission spectra under pulsed resonant
excitation reveals an exponential decay which gives the emitter’s lifetime
$T_{1}$ and a fast oscillation indicating the quantum beating between the
fine-structure peaks of the neutral exciton emission, $\Delta_{\rm
fss}=4.807\,(3)\,\mathrm{GHz}$. (b, c) Second-order intensity correlation
histogram $g^{(2)}$, of the emitted photons under off-resonant (b) and
resonant (c) excitation. The lack of coincidences in the central peak
indicates the low probability of multi-photon emission. (d) Two-photon
interference of consecutively scattered photons delayed by
$12.5\,\mathrm{ns}$, prepared in cross ($g^{(2}_{\perp}$) and parallel
($g^{(2}_{\parallel}$) polarisations, under resonant $\pi$-pulse excitation.
The extracted photon indistinguishability, given by ratio of zero-delay-
coincidences from both configurations, along with the extracted values
($T_{1}$ and $g^{(2)}(0)$) from the fits (solid lines) are summarised in
Table. 1.
Figure 2 shows the internal conversion efficiency for 942 nm light, measured
by comparing output 1550 nm light to the 942 nm coupled through the waveguide
with the seed laser blocked. This factors out coupling losses into the
waveguide which are measured to be $17\,\%$. The data is fitted with [24]
$\eta=\eta_{max}\sin^{2}\left(\sqrt{\eta_{nor}P}\leavevmode\nobreak\
L\right),$ (1)
where $\eta_{max}$ is the maximum possible conversion efficiency, $\eta_{nor}$
is the normalised conversion efficiency of the process, $P$ is the input
power, and $L$ is the waveguide length. The fit gives a normalised conversion
efficiency (to waveguide length in the limit of small pump powers [24]) is
$\eta_{nor}=44\left(1\right)\,\%\text{ /}\left(\text{W cm}^{2}\right)$. The
maximal external conversion efficiency, the ratio of photons collected into
single-mode fibre after the conversion stage to the number of NIR photons
impinging on the waveguide, $\eta_{max}=38\,\%$, leading to a maximum internal
conversion efficiency of $56.7\left(4\right)\,\%$ when taking losses into
account. This $\eta_{max}$ is higher than previously reported values for NIR
QD frequency conversion to 1550 nm with similar waveguides [11]. We would like
to highlight that we achieve SNRs $>250$ for all seed powers (inset Fig. 2),
meaning that the noise contribution of the DFG process towards the converted
single photons is minimal.
We now compare the characteristics of the converted telecom photons to the QD
NIR photons. We tune the QD into resonance with the first cavity mode, and
excite either resonantly or non-resonantly into the third cavity mode; see
Fig. 3 for spectral properties of the cavity under 820 nm excitation. The
inset in Fig. 3 shows the detected count rates for these two excitation
scenarios as a function of power: Rabi oscillations are observed for resonant
excitation while a clear maximum is observed for non-resonant excitation. For
non-resonant characterisation, the QD photons are spectrally filtered with a
grating filter with a 30 GHz FWHM to suppress the excitation laser. We detect
a count rate of $1.85(5)\leavevmode\nobreak\ \mathrm{MHz}$ at an excitation
power of $6.8\,\mu\mathrm{W}$. The grating filter was removed when
characterising the converted photons as low-loss bandpass filters were used at
1550 nm. For resonant driving, we optimize the excitation power to the
$\pi$-pulse and detect a count rate of $1.46(4)\leavevmode\nobreak\
\mathrm{MHz}$. This value is slightly lower than for off-resonant excitation
due to the presence of spectral fluctuations [25]. After QFC, the detected
count rate at 1550 nm, for the off-resonant and on-resonant case is
$856(18)\leavevmode\nobreak\ \mathrm{kHz}$ and $456(14)\leavevmode\nobreak\
\mathrm{kHz}$, respectively. Comparing the NIR and telecom counts under
resonant excitation gives an end-to-end conversion efficiency of $\approx
35\,\%$, after accounting for the difference in the detection efficiency of
both NIR ($\sim 90\%$) and telecom C-band ($\sim 80\%$) detectors. This agrees
well with the measured loss budget through the optical components including
the conversion efficiency. The difference in efficiency for off-resonant
excitation is accounted for by the loss of the grating filter.
Figure 4 shows the comparison between the performance of the QD signal before
and after QFC. The lifetime measured under resonant excitation in Fig. 4(a)
remains unchanged within experimental error after conversion. The oscillation
in the time-resolved emission, indicative of the quantum beating of the
$X^{0}$ fine-structure splitting, shows a frequency of
$4.807\,(3)\,\mathrm{GHz}$. The equivalent oscillation after the QFC process
is unchanged ($4.803\,(1)\,\mathrm{GHz}$), indicating that the CW-seeded
frequency conversion preserves the temporal mode of the input photons.
Next, we measure the second-order intensity correlation $g^{(2)}$ using a
Hanbury-Brown and Twiss (HBT) interferometer. For a perfect single-photon
source $g^{(2)}(0)=0$, indicating the absence of multi-photon emissions. Under
off-resonant driving, Fig. 4(b), we observe a slight increase from
$g^{(2)}(0)=0.045\,(0)$ to $g^{(2)}(0)=0.051\,(1)$ before and after the QFC
process, respectively. We observe similar values under resonant driving, Fig.
4(c), demonstrating near-ideal single-photon emission with
$g^{(2)}(0)=0.040\,(0)$ and $g^{(2)}(0)=0.043\,(1)$ before and after the QFC
process, respectively. The slight increase in the normalized coincidences in
the uncorrelated side peaks in the HBT histogram is due to blinking of the
emitters, a common effect resulting from QD coupling to the solid-state charge
environment [26]. The imperfection in $g^{(2)}(0)$ can be due to imperfect
suppression of the cavity emission due to cavity feeding [27, 28, 29], slight
imperfection in the wave-plate retarders used in our confocal microscope, and
presence of multi-photon capture processes [30, 31]. Nevertheless, with a
modest increase in $g^{(2)}(0)$ after the QFC process, we have demonstrated
near background-free single photon frequency conversion from NIR to telecom
C-band, with the photon number purity predominately limited by the quantum
dot.
To demonstrate that our QFC setup preserves photon coherence, we perform Hong-
Ou-Mandel (HOM) interference between photons emitted from two consecutive
excitation pulses. We use an unbalanced Mach-Zehnder interferometer with a
delay of 12.5 ns to match photons temporally on a 50/50 beam splitter. We
measure the coincidence counts for parallel and perpendicular polarised
photons and evaluate the visibility as
$\mathrm{V_{HOM}}=1-g^{(2)}_{\parallel}/g^{(2)}_{\perp}$. For a pair of
indistinguishable photons, $\mathrm{V_{HOM}}=1$. For resonant excitation we
achieve an interference visibility of $\mathrm{V_{HOM}}=0.88\,(1)$ before QFC.
We calculate the single photon indistinguishability $M_{s}$ as
$M_{s}=(\mathrm{V_{HOM}}+g^{(2)}(0))/(1-g^{(2)}(0))$ [32]. This gives an upper
bound to the HOM visibility taking the finite
$g^{\left(2\right)}\left(0\right)$ into account. Before conversion, $\rm
M_{s}$ is equal to $0.95\,(1)$. After conversion we find the raw visibility
and corrected indistinguishability to be $0.60\,(1)$ and $0.67\,(2)$
respectively. The results of lifetime, HBT and HOM measurements are summarised
in Table 1. The reduced interference visibility originates in spectral
instability introduced by fluctuating power in multiple longitudinal modes of
the seed laser. The line width of the seed laser is around 4 GHz with a free
spectral range estimated to be 177 MHz, corresponding to $\sim 22$ modes.
Despite this, we show that our QFC setup indeed preserves the coherence of
single photons as measured from non-classical two-photon interference which
can be improved by increased control over the cavity dispersion and active
stabilization of the cavity length.
| 942 nm | 1550 nm
---|---|---
Lifetime $T_{1}\,(\mathrm{ns}$) | 0.2622 (2) | 0.2621 (2)
Resonant count rate (kHz) | 1,460 (40) | 456 (14)
Off-resonant count rate (kHz) | 1,850 (50) | 856 (18)
Off-resonant $g^{(2)}(0)$ | 0.045 (0) | 0.051 (1)
Resonant $g^{(2)}(0)$ | 0.040 (0) | 0.043 (1)
Resonant $\rm V_{HOM}$ | 0.88 (1) | 0.60 (1)
Resonant $M_{s}$ | 0.95 (1) | 0.67 (2)
Table 1: Summary of the lifetime, count rate, $g^{(2)}(0)$ and
indistinguishability $\rm V_{HOM}$ for converted and unconverted photons.
Values are obtained from measurement results, illustrated in Fig. 4. The
corrected photon indistinguishability $M_{S}$ is estimated based on the
measured $g^{(2)}(0)$ and uncorrected $\rm V_{HOM}$ [32]. The error, given by
the standard deviation from the fit, is included in brackets.
Modest improvements to the current QFC system will allow us to improve the
converted two-photon interference visibility to equal the unconverted
visibility. The external conversion efficiency could be further improved with
lower-loss filtering and improved mode matching between the single-mode fibre
and the waveguide mode. We believe that this source will find applications in
fibre-based quantum communication where a source of bright and highly pure
single photons in the C-band is required. This can lead to demonstrations of
various quantum communication protocols including measurement-device-
independent quantum key distribution, teleportation and entanglement swapping
between distant quantum nodes.
###### Acknowledgements.
We acknowledge Pierre M. Petroff for his contribution to the sample design and
fabrication. This work was supported by the EPSRC (Grants No. EP/L015110/1,
No. EP/M013472/1, No. EP/P029892/1, EP/N002962/1 and EP/T001011/1), the ERC
(Grant No. 725920), and the EU Horizon 2020 research and innovation program
under Grant Agreement No. 820423. B. D. G. thanks the Royal Society for a
Wolfson Merit Award and the Royal Academy of Engineering for a Chair in
Emerging Technology. Y. M. acknowledges the support from Chongqing Research
Program of Basic Research and Frontier Technology (No.cstc2016jcyjA0301). The
authors in K.I.S.T. acknowledge the support from the KIST institutional
program, the program of quantum sensor core technology through IITP and the
IITP grant funded by the Korea government(MSIT) (No. 20190004340011001).
## References
## References
* Tomm _et al._ [2020] N. Tomm, A. Javadi, N. O. Antoniadis, D. Najer, M. C. Löbl, A. R. Korsch, R. Schott, S. R. Valentin, A. D. Wieck, A. Ludwig, and R. J. Warburton, “A bright and fast source of coherent single photons,” arXiv:2007.12654 (2020).
* Wang _et al._ [2019a] H. Wang, Y.-M. He, T.-H. Chung, H. Hu, Y. Yu, S. Chen, X. Ding, M.-C. Chen, J. Qin, X. Yang, R.-Z. Liu, Z.-C. Duan, J.-P. Li, S. Gerhardt, K. Winkler, J. Jurkat, L.-J. Wang, N. Gregersen, Y.-H. Huo, Q. Dai, S. Yu, S. Höfling, C.-Y. Lu, and J.-W. Pan, “Towards optimal single-photon sources from polarized microcavities,” Nat. Photonics 13, 770–775 (2019a).
* Somaschi _et al._ [2016] N. Somaschi, V. Giesz, L. De Santis, J. C. Loredo, M. P. Almeida, G. Hornecker, S. L. Portalupi, T. Grange, C. Antón, J. Demory, C. Gómez, I. Sagnes, N. D. Lanzillotti-Kimura, A. Lemaítre, A. Auffeves, A. G. White, L. Lanco, and P. Senellart, “Near-optimal single-photon sources in the solid state,” Nat. Photonics 10, 340–345 (2016).
* Olbrich _et al._ [2017] F. Olbrich, J. Höschele, M. Müller, J. Kettler, S. Luca Portalupi, M. Paul, M. Jetter, and P. Michler, “Polarization-entangled photons from an InGaAs-based quantum dot emitting in the telecom C-band,” Applied Physics Letters 111, 133106 (2017).
* Anderson _et al._ [2020] M. Anderson, T. Müller, J. Huwer, J. Skiba-Szymanska, A. B. Krysa, R. M. Stevenson, J. Heffernan, D. A. Ritchie, and A. J. Shields, “Quantum teleportation using highly coherent emission from telecom C-band quantum dots,” npj Quantum Information 6, 14 (2020).
* Zeuner _et al._ [2019] K. D. Zeuner, K. D. Jöns, L. Schweickert, C. Reuterskiöld Hedlund, C. Nuñez Lobato, T. Lettner, K. Wang, S. Gyger, E. Schöll, S. Steinhauer, M. Hammar, and V. Zwiller, “On-demand generation of entangled photon pairs in the telecom C-band for fiber-based quantum networks,” arXiv:1912.04782 (2019).
* Wang _et al._ [2019b] H. Wang, J. Qin, X. Ding, M.-C. Chen, S. Chen, X. You, Y.-M. He, X. Jiang, L. You, Z. Wang, C. Schneider, J. J. Renema, S. Höfling, C.-Y. Lu, and J.-W. Pan, “Boson Sampling with 20 Input Photons and a 60-Mode Interferometer in a 1 0 14 -Dimensional Hilbert Space,” Phys. Rev. Lett. 123, 250503 (2019b).
* Zaske _et al._ [2012] S. Zaske, A. Lenhard, C. A. Keßler, J. Kettler, C. Hepp, C. Arend, R. Albrecht, W.-M. Schulz, M. Jetter, P. Michler, and C. Becher, “Visible-to-Telecom Quantum Frequency Conversion of Light from a Single Quantum Emitter,” Phys. Rev. Lett. 109, 147404 (2012).
* Kambs _et al._ [2016] B. Kambs, J. Kettler, M. Bock, J. N. Becker, C. Arend, A. Lenhard, S. L. Portalupi, M. Jetter, P. Michler, and C. Becher, “Low-noise quantum frequency down-conversion of indistinguishable photons,” Opt. Express 24, 22250–22260 (2016).
* Weber _et al._ [2018] J. H. Weber, J. Kettler, H. Vural, M. Müller, J. Maisch, M. Jetter, S. L. Portalupi, and P. Michler, “Overcoming correlation fluctuations in two-photon interference experiments with differently bright and independently blinking remote quantum emitters,” Phys. Rev. B 97, 195414 (2018).
* Weber _et al._ [2019] J. H. Weber, B. Kambs, J. Kettler, S. Kern, J. Maisch, H. Vural, M. Jetter, S. L. Portalupi, C. Becher, and P. Michler, “Two-photon interference in the telecom C-band after frequency conversion of photons from remote quantum emitters,” Nature Nanotechnology 14, 23–26 (2019).
* De Greve _et al._ [2012] K. De Greve, L. Yu, P. L. McMahon, J. S. Pelc, C. M. Natarajan, N. Y. Kim, E. Abe, S. Maier, C. Schneider, M. Kamp, S. Höfling, R. H. Hadfield, A. Forchel, M. M. Fejer, and Y. Yamamoto, “Quantum-dot spin–photon entanglement via frequency downconversion to telecom wavelength,” Nature 491, 421–425 (2012).
* Pelc _et al._ [2012] J. S. Pelc, L. Yu, K. De Greve, P. L. McMahon, C. M. Natarajan, V. Esfandyarpour, S. Maier, C. Schneider, M. Kamp, S. Höfling, R. H. Hadfield, A. Forchel, Y. Yamamoto, and M. M. Fejer, “Downconversion quantum interface for a single quantum dot spin and 1550-nm single-photon channel,” Optics Express 20, 27510 (2012).
* Singh _et al._ [2019] A. Singh, Q. Li, S. Liu, Y. Yu, X. Lu, C. Schneider, S. Höfling, J. Lawall, V. Verma, R. Mirin, S. W. Nam, J. Liu, and K. Srinivasan, “Quantum frequency conversion of a quantum dot single-photon source on a nanophotonic chip,” Optica 6, 563 (2019).
* Rakher [2008] M. T. Rakher, _Quantum optics with quantum dots in microcavities_ , Ph.D. thesis, University of California, Santa Barbara (2008).
* Rakher _et al._ [2009] M. T. Rakher, N. G. Stoltz, L. A. Coldren, P. M. Petroff, and D. Bouwmeester, “Externally mode-matched cavity quantum electrodynamics with charge-tunable quantum dots,” Phys. Rev. Lett. 102, 097403 (2009).
* Macdonald _et al._ [2015] J. R. Macdonald, S. J. Beecher, A. Lancaster, P. A. Berry, K. L. Schepler, and A. K. Kar, “Ultrabroad Mid-Infrared Tunable Cr:ZnSe Channel Waveguide Laser,” IEEE Journal of Selected Topics in Quantum Electronics 21, 375–379 (2015).
* Wagner _et al._ [1999] G. J. Wagner, T. J. Carrig, R. H. Page, K. I. Schaffers, J.-O. Ndap, X. Ma, and A. Burger, “Continuous-wave broadly tunable Cr${}^{\textrm{2+}}$:ZnSe laser,” Optics Letters 24, 19–21 (1999).
* Sorokina _et al._ [2002] I. T. Sorokina, E. Sorokin, S. Mirov, V. Fedorov, V. Badikov, V. Panyutin, and K. I. Schaffers, “Broadly tunable compact continuous-wave Cr${}^{\textrm{2+}}$:ZnS laser,” Optics Letters 27, 1040–1042 (2002).
* Wagner _et al._ [2004] G. J. Wagner, B. G. Tiemann, W. J. Alford, and T. J. Carrig, “Single-Frequency Cr:ZnSe Laser,” in _Advanced Solid-State Photonics (TOPS) (2004), paper 371_ (Optical Society of America, 2004) p. 371.
* Vasilyev, Mirov, and Gapontsev [2014] S. Vasilyev, M. Mirov, and V. Gapontsev, “High Power Kerr-Lens Mode-Locked Femtosecond mid-IR Laser with Efficient Second Harmonic Generation in Polycrystalline Cr2+:ZnS and Cr2+:ZnSe,” in _Advanced Solid State Lasers (2014), paper AM3A.3_ (Optical Society of America, 2014) p. AM3A.3.
* Brecht _et al._ [2011] B. Brecht, A. Eckstein, A. Christ, H. Suche, and C. Silberhorn, “From quantum pulse gate to quantum pulse shaper—engineered frequency conversion in nonlinear optical waveguides,” New Journal of Physics 13, 065029 (2011).
* Kumar [1990] P. Kumar, “Quantum frequency conversion,” Optics Letters 15, 1476–1478 (1990).
* Roussev _et al._ [2004] R. V. Roussev, C. Langrock, J. R. Kurz, and M. M. Fejer, “Periodically poled lithium niobate waveguide sum-frequency generator for efficient single-photon detection at communication wavelengths,” Optics Letters 29, 1518–1520 (2004).
* Santana _et al._ [2017] T. S. Santana, Y. Ma, R. N. E. Malein, F. Bastiman, E. Clarke, and B. D. Gerardot, “Generating indistinguishable photons from a quantum dot in a noisy environment,” Phys. Rev. B 95, 201410 (2017).
* Davanço _et al._ [2014] M. Davanço, C. S. Hellberg, S. Ates, A. Badolato, and K. Srinivasan, “Multiple time scale blinking in inas quantum dot single-photon sources,” Phys. Rev. B 89, 161303 (2014).
* Kaniber _et al._ [2008] M. Kaniber, A. Laucht, A. Neumann, J. M. Villas-Bôas, M. Bichler, M.-C. Amann, and J. J. Finley, “Investigation of the nonresonant dot-cavity coupling in two-dimensional photonic crystal nanocavities,” Phys. Rev. B 77, 161303 (2008).
* Winger _et al._ [2009] M. Winger, T. Volz, G. Tarel, S. Portolan, A. Badolato, K. J. Hennessy, E. L. Hu, A. Beveratos, J. Finley, V. Savona, and A. İmamoğlu, “Explanation of photon correlations in the far-off-resonance optical emission from a quantum-dot–cavity system,” Phys. Rev. Lett. 103, 207403 (2009).
* Suffczyński _et al._ [2009] J. Suffczyński, A. Dousse, K. Gauthron, A. Lemaître, I. Sagnes, L. Lanco, J. Bloch, P. Voisin, and P. Senellart, “Origin of the optical emission within the cavity mode of coupled quantum dot-cavity systems,” Phys. Rev. Lett. 103, 027401 (2009).
* Peter _et al._ [2007] E. Peter, S. Laurent, J. Bloch, J. Hours, S. Varoutsis, I. Robert-Philip, A. Beveratos, A. Lemaître, A. Cavanna, G. Patriarche, P. Senellart, and D. Martrou, “Fast radiative quantum dots: From single to multiple photon emission,” Applied Physics Letters 90, 223118 (2007).
* Lanco and Senellart [2015] L. Lanco and P. Senellart, “A highly efficient single photon-single quantum dot interface,” in _Engineering the Atom-Photon Interaction: Controlling Fundamental Processes with Photons, Atoms and Solids_, edited by A. Predojević and M. W. Mitchell (Springer International Publishing, New York, NY, 2015) Chap. 2, pp. 39–71.
* Ollivier _et al._ [2020] H. Ollivier, S. E. Thomas, S. C. Wein, I. M. de Buy Wenniger, N. Coste, J. C. Loredo, N. Somaschi, A. Harouri, A. Lemaitre, I. Sagnes, L. Lanco, C. Simon, C. Anton, O. Krebs, and P. Senellart, “Hong-ou-mandel interference with imperfect single photon sources,” arXiv:2005.01743 (2020).
|
††institutetext: Institut für Theoretische Physik und Astrophysik
and Würzburg-Dresden Cluster of Excellence ct.qmat,
Julius-Maximilians-Universität Würzburg, Am Hubland, 97074 Würzburg, Germany
# Monodromy methods for torus conformal blocks and entanglement entropy at
large central charge
Marius Gerbershagen<EMAIL_ADDRESS>
###### Abstract
We compute the entanglement entropy in a two dimensional conformal field
theory at finite size and finite temperature in the large central charge limit
via the replica trick. We first generalize the known monodromy method for the
calculation of conformal blocks on the plane to the torus. Then, we derive a
monodromy method for the zero-point conformal blocks of the replica partition
function. We explain the differences between the two monodromy methods before
applying them to the calculation of the entanglement entropy. We find that the
contribution of the vacuum exchange dominates the entanglement entropy for a
large class of CFTs, leading to universal results in agreement with
holographic predictions from the RT formula. Moreover, we determine in which
regime the replica partition function agrees with a correlation function of
local twist operators on the torus.
## 1 Introduction
Entanglement entropy is a measure for the amount of entanglement between two
parts of a quantum system. It is defined as the von Neumann entropy of the
reduced density matrix $\rho_{A}$ for a subsystem $A$. In general, the
entanglement entropy depends on details of the theory and state in question
such as the spectrum and operator content. However, certain universal features
are common to all quantum field theories. For example, the leading order
divergence in the UV cutoff usually scales with the area of the boundary of
the subregion $A$ Bombelli:1986rw ; Srednicki:1993im . Conformal field
theories in two dimensions admit more general universal features. In
particular, the entanglement entropy of a single interval $A$ at zero
temperature is given by Calabrese:2004eu
$S_{A}=\frac{c}{3}\log(l/\epsilon_{\text{UV}}),$ (1)
depending only on the central charge, irrespective of any other details such
as the OPE coefficients or the spectrum of the theory. For subsystems $A$
consisting of multiple intervals, the entanglement entropy is no longer
universal for any CFT. However, as shown in Hartman:2013mia , in the
semiclassical large central charge limit and at zero temperature, the
entanglement entropy again becomes universal for a large class of conformal
field theories. These CFTs are characterized by a sparse spectrum of light
operators and at most exponentially growing OPE coefficients. By using
conformal transformations, the universal results of Calabrese:2004eu ;
Hartman:2013mia translate to the case of either finite temperature or finite
size. This publication is dedicated to the study of universal features of the
entanglement entropy in a system with both finite size and finite temperature.
The computational approach most commonly used to determine entanglement
entropies is the replica trick. It is based on the calculation of the Rényi
entropies
$S_{A}^{(n)}=\frac{1}{1-n}\log{\mathrm{Tr}}\rho_{A}^{n},$ (2)
via a partition function $Z_{n}$ on a higher genus Riemann surface
${\mathcal{R}}_{n}$ obtained by gluing $n$ copies of the complex plane
cyclically together along the entangling interval $A$. This partition function
is then mapped to a correlation function of twist operators, i.e. local
operators with scaling dimension $h=\bar{h}=c/24(n-1/n)$ inserted at the
endpoints of the entangling interval. For a subsystem $A$ consisting of $N$
disjoint intervals, the correlation function contains $2N$ twist operator
insertions. Finally, the entanglement entropy is obtained by analytically
continuing $n$ to the real numbers and taking the limit $n\to 1$. The
universality of the entanglement entropy for a single interval follows
immediately from the universality of the two-point function in any conformal
field theory Calabrese:2004eu . For multiple intervals, the Rényi entropy is
mapped to a higher-point correlation function of twist operators which
decomposes into a sum over conformal blocks. The universality in the
semiclassical limit observed in Hartman:2013mia is explained by the fact that
only a single conformal block (the vacuum block) contributes to the
entanglement entropy. More precisely, in the semiclassical limit of large
central charge, the contribution of other conformal blocks is exponentially
suppressed in the central charge, assuming the aforementioned restrictions on
the theory, i.e. a sparse spectrum of light operators and at most
exponentially growing OPE coefficients.
In the case of a system with both finite size and finite temperature, the
replica trick instructs us to calculate the partition function on a higher
genus surface constructed by gluing $n$ copies of the torus along the
entangling interval $A$. In the small interval limit, this partition function
can be obtained analogously to the zero temperature case as a correlation
function on the torus of two (local) twist operators inserted at the endpoints
of the entangling interval (see e.g. Azeyanagi:2007bj ; Calabrese:2009qy ;
Cardy:2014jwa ; Chen:2016lbu ). For large intervals, on the other hand, the
replica partition function does not agree with a correlation function of local
twist operators, as can be seen by the following argument. It is well known
that for a pure state $\rho$, the entanglement entropy for $A$ is equal to the
entanglement entropy of the complement $A^{c}$, $S_{A}=S_{A^{c}}$. However,
for mixed states such as the thermal states described by the CFT on the torus,
this property no longer holds. Since the correlation function of local twist
operators contains no information about whether we compute the entanglement
entropy for $A$ or for $A^{c}$ (the location of the branch cuts between the
twist operators is not fixed by the twist correlator), it cannot give the
correct answer on the torus111In the case of a free CFT, a discrepancy between
the correlator of local twist operators and the higher genus partition
function has been observed in Lokhande:2015zma ; Mukhi:2017rex based on
previous work Headrick:2012fk ; Datta:2013hba ; Chen:2015cna . In particular,
it was found in Lokhande:2015zma that the twist operator result is not
modular covariant and violates Bose-Fermi equivalence.. This issue can be
resolved for instance by defining non-local twist operators on the torus222We
would like to thank the referee for pointing this out to us. as in
Chen:2014hta , however we will not do this and instead phrase the calculation
directly in terms of the replica partition function. Therefore, in the
following the term “twist operator” will always refer to a local operator of
scaling dimension $h=\bar{h}=c/24(n-1/n)$.
A convenient way to calculate the entanglement entropy on the plane works by
expanding the twist correlator in conformal blocks and calculating these using
the well-known monodromy method first described in Zamolodchikov1987 . This
monodromy method is derived as follows. By inserting a degenerate field into
the correlation function, one obtains a differential equation for an auxiliary
function $\Psi(z)$. This differential equations contains derivatives of the
sought after conformal block as accessory parameters. The accessory parameters
are fixed by demanding a certain monodromy of the solution $\Psi(z)$ around
cycles which encircle a number of operator insertion points. Which insertion
points are encircled depends on the channel in which the correlation function
is expanded. We review this method in detail in sec. 2.1 before generalizing
it to the case of finite temperature and finite size.
A different perspective on this method was offered in Faulkner:2013yia , where
it was related to a uniformization problem on the replica surface
${\mathcal{R}}_{n}$. In general, a compact Riemann surface $\Sigma$ can be
obtained as the quotient of the complex plane by a subgroup of
$PSL(2,{\mathbb{C}})$ Zograf_1988 . Thus, there exists a single valued map
$w\to z$ from the complex plane to ${\mathcal{R}}_{n}$. This uniformization
map is given as the quotient $w=\Psi_{1}(z)/\Psi_{2}(z)$ of two independent
solutions of a differential equation. It turns out that this differential
equation is equal to the one from the monodromy method for the conformal
blocks of the twist correlator on the plane Faulkner:2013yia . Using the
equivalence between the CFT partition function and the gravitational action in
the dual AdS space, this yields a proof of the RT formula at zero temperature
Faulkner:2013yia . We use similar arguments from the uniformization problem on
the replica surface of the torus to determine a monodromy method for the zero-
point conformal blocks of the replica partition function at finite
temperature. This new monodromy method is closely connected to the monodromy
method for the conformal blocks of the correlator of local twist operators on
the torus, with the crucial difference being that the new method allows for
choosing a larger set of cycles around which to impose the monodromy. These
new cycles are necessary to reproduce the $S_{A}\neq S_{A^{c}}$ property for
thermal states.
In the context of the AdS/CFT correspondence, universal features of the
entanglement entropy for holographic CFTs at large central charge are
predicted by the Ryu-Takayanagi formula (RT formula for short) Ryu:2006bv .
The RT formula states that the entanglement entropy of a subregion $A$ in the
boundary field theory corresponds to the area of a minimal surface
$\gamma_{A}$ in the bulk, anchored at $\partial A$ on the boundary of the AdS
space. Apart from reproducing the universal results obtained in
Calabrese:2004eu ; Hartman:2013mia , the RT formula also predicts interesting
universal features of the entanglement entropy in the case of both finite
temperature and finite size. In particular, there are two phase transitions
Azeyanagi:2007bj . First, there is a Hawking-Page transition in the bulk from
thermal AdS3 to the BTZ black hole phase as the temperature increases. This
induces a corresponding phase transition in the entanglement entropy. Second,
the entanglement entropy in the BTZ phase also shows a phase transition as the
size of the entangling interval increases. We explain how these features
appear from the CFT side.
Related work on the entanglement entropy in conformal field theories at finite
size and finite temperature includes Chen:2014hta ; Chen:2014ehg ;
Chen:2014unl ; Barrella:2013wja ; Chen:2015kua . Chen:2014hta ; Chen:2014ehg ;
Chen:2014unl is concerned with the entanglement entropy in various limits of
high and low temperature or size of the entangling interval $A$, in which case
universal results for arbitrary values of the central charge can be obtained.
In Barrella:2013wja ; Chen:2015kua , the holographic entanglement entropy for
a single entangling interval on the boundary of a thermal AdS3 space and the
BTZ black hole was calculated using a monodromy method on the gravity side.
The monodromy method derived from the CFT side in this publication will turn
out to be equivalent to the monodromy method on the gravity side used in
Barrella:2013wja ; Chen:2015kua . Related work on torus conformal blocks
includes KashaniPoor:2012wb which derived the monodromy method for the
special case of one-point Virasoro conformal blocks on the torus and
Alkalaev:2016fok ; Alkalaev:2017bzx which performed explicit calculations of
one- and two-point Virasoro conformal blocks in various limits including the
semiclassical one which we study in this publication.
Our paper is organized as follows. In sec. 2, we derive the monodromy methods
used in this publication. After a review of the standard monodromy method for
conformal blocks on the plane in sec. 2.1, we generalize to torus conformal
blocks in sec. 2.2. Sec. 2.3 explains how to obtain a monodromy method for
zero-point conformal blocks of the partition function on the replica surface
and the differences between it and the monodromy method for conformal blocks
on the torus. Following this, we apply the newly derived monodromy methods to
the calculation of the entanglement entropy in sec. 3. Assuming that the
vacuum exchange dominates the partition function on the higher genus Riemann
surface, we find universal results in agreement with the RT formula. For the
partition function on the replica surface, we find in particular agreement
with the phase transition in the entanglement entropy at large interval size
and high temperature. This feature cannot be reproduced from the correlator of
local twist operators. We check the assumption on the dominance of the vacuum
exchange numerically in sec. 3.4. Finally, we conclude with a brief discussion
and outlook in sec. 4.
## 2 Monodromy methods
This section contains an overview over the monodromy methods used in this
publication. We start with a review of the standard monodromy method for
conformal blocks on the plane, then turn to the case of conformal blocks on
the torus and finally explain how to derive a monodromy method for zero-point
blocks of the partition function on the replica surface relevant to the
computation of entanglement entropy on the torus.
### 2.1 Conformal blocks on the plane
In this section, we review the monodromy method for the calculation of four-
point semiclassical conformal blocks on the plane first derived in
Zamolodchikov1986 (see also Harlow:2011ny for a more detailed explanation).
The starting point of the derivation is the correlation function of four
primary fields ${\mathcal{O}}_{i}$
$\langle{\mathcal{O}}_{1}(z_{1},\bar{z}_{1}){\mathcal{O}}_{2}(z_{2},\bar{z}_{2}){\mathcal{O}}_{3}(z_{3},\bar{z}_{3}){\mathcal{O}}_{4}(z_{4},\bar{z}_{4})\rangle.$
(3)
Let us parametrize the central charge as $c=1+6(b+1/b)^{2}$ and take the
semiclassical limit $c\to\infty,b\to 0$ in which the conformal weights $h_{i}$
of the operators ${\mathcal{O}}_{i}$ as well as the internal conformal weight
$h_{p}$ scale proportional to the central charge. In the correlation function
(3), we insert a degenerate operator $\Psi(z,\bar{z})$ with conformal weight
$h_{\Psi}=-1/2-3b^{2}/4\sim{\mathcal{O}}(c^{0})$ obeying
$\left(\hat{L}_{-2}+\frac{1}{b^{2}}\hat{L}_{-1}^{2}\right)\Psi(z,\bar{z})=0$
(4)
From the conformal Ward identities, this yields the following differential
equation known as the decoupling equation,
$\left[\frac{1}{b^{2}}\partial_{z}^{2}+\sum_{i}\left(\frac{h_{i}}{(z-z_{i})^{2}}+\frac{\partial_{z_{i}}}{z-z_{i}}\right)\right]\langle{\mathcal{O}}_{1}{\mathcal{O}}_{2}\Psi{\mathcal{O}}_{3}{\mathcal{O}}_{4}\rangle=0.$
(5)
To get to the $s$-channel conformal block, we then insert the operator product
expansion
${\mathcal{O}}_{1}(z_{1},\bar{z}_{1}){\mathcal{O}}_{2}(z_{2},\bar{z}_{2})=\sum_{p}C^{p}_{21}\sum_{k,\bar{k}}(z_{2}-z_{1})^{h_{p}-h_{1}-h_{2}+|k|}(\bar{z}_{2}-\bar{z}_{1})^{\bar{h}_{p}-\bar{h}_{1}-\bar{h}_{2}+|\bar{k}|}\beta^{pk}_{21}\beta^{p\bar{k}}_{21}{\mathcal{O}}_{p}^{\\{k,\bar{k}\\}}(z_{1},\bar{z}_{1})$
(6)
into the correlation function which yields terms containing
$\langle{\mathcal{O}}_{p}^{\\{k,\bar{k}\\}}\Psi{\mathcal{O}}_{3}{\mathcal{O}}_{4}\rangle$.
At large central charge, these terms can be approximated by
$\langle{\mathcal{O}}_{p}^{\\{k,\bar{k}\\}}\Psi{\mathcal{O}}_{3}{\mathcal{O}}_{4}\rangle\approx\Psi_{p}\langle{\mathcal{O}}_{p}^{\\{k,\bar{k}\\}}{\mathcal{O}}_{3}{\mathcal{O}}_{4}\rangle,$
(7)
where $\Psi_{p}$ is defined by
$\Psi_{p}=\frac{\langle{\mathcal{O}}_{p}\Psi{\mathcal{O}}_{3}{\mathcal{O}}_{4}\rangle}{\langle{\mathcal{O}}_{p}{\mathcal{O}}_{3}{\mathcal{O}}_{4}\rangle}.$
(8)
This can be shown by employing the form of
$\langle{\mathcal{O}}_{p}^{\\{k,\bar{k}\\}}\Psi{\mathcal{O}}_{3}{\mathcal{O}}_{4}\rangle$
in terms of a string of differential operators
${\mathcal{L}}_{k_{i}},\bar{\mathcal{L}}_{k_{i}}$ acting on
$\langle{\mathcal{O}}_{p}\Psi{\mathcal{O}}_{3}{\mathcal{O}}_{4}\rangle$, where
${\mathcal{L}}_{-k_{i}}^{\Psi}=-\sum_{j=3,4,\Psi}\left(\frac{(1-k_{i})h_{j}}{(z_{j}-z_{1})^{k_{i}}}+\frac{1}{(z_{j}-z_{1})^{k_{i}-1}}\partial_{z_{j}}\right).$
(9)
Now $\langle{\mathcal{O}}_{p}\Psi{\mathcal{O}}_{3}{\mathcal{O}}_{4}\rangle$
scales as $e^{-c/6S_{\text{cl.}}}$ in the semiclassical limit where
$S_{\text{cl.}}\sim{\mathcal{O}}(c^{0})$ while
$\Psi_{p}\sim{\mathcal{O}}(c^{0})$ and $h_{\Psi}\sim{\mathcal{O}}(c^{0})$.
Hence, we can neglect the derivatives acting on $\Psi_{p}$ and on the
$h_{\Psi}$ term to obtain (7) in the leading order in $c$. Now, use a
conformal transformation to send $z_{1}\to 0$, $z_{3}\to 1$, $z_{4}\to\infty$
and $z_{2}$ to the cross ratio $x$. This implies
$\langle{\mathcal{O}}_{1}{\mathcal{O}}_{2}\Psi{\mathcal{O}}_{3}{\mathcal{O}}_{4}\rangle\approx\sum_{p}\Psi_{p}(z,x,\bar{x})C^{p}_{21}C^{p}_{43}{\mathcal{F}}^{p}_{12,34}(x)\bar{\mathcal{F}}^{p}_{12,34}(\bar{x}),$
(10)
where ${\mathcal{F}}^{p}_{12,34}(x)$ is the desired conformal block which in
the semiclassical limit scales as ${\mathcal{F}}^{p}_{12,34}(x)\sim
e^{-\frac{c}{6}f_{\text{cl.}}(x)}$ as was conjectured in Zamolodchikov1986
and recently shown in Besken:2019jyw . The semiclassical conformal block
$f_{\text{cl.}}$ depends only on the cross ratio $x$ and on
$b^{2}h_{i},b^{2}h_{p}$. The decoupling equation (5) then yields at leading
order in $c$
$\left[\partial_{z}^{2}+\sum_{i}\left(\frac{b^{2}h_{i}}{(z-z_{i})^{2}}-\frac{\partial_{z_{i}}f_{\text{cl.}}(x)}{z-z_{i}}\right)\right]\Psi_{p}=0.$
(11)
There is one separate decoupling equation for each term in the sum over $p$
since generically, each term has a different monodromy and thus must vanish
separately. All terms involving derivatives of $\Psi_{p}$ vanish to leading
order due to $\Psi_{p}\sim{\mathcal{O}}(c^{0})$. From the expression for the
cross ratio $x=\frac{(z_{1}-z_{2})(z_{4}-z_{3})}{(z_{4}-z_{2})(z_{1}-z_{3})}$,
we obtain linear relations among the $\partial_{z_{i}}f_{\text{cl.}}$
$\sum_{i}\partial_{z_{i}}f_{\text{cl.}}=\sum_{i}(z_{i}\partial_{z_{i}}f_{\text{cl.}}-b^{2}h_{i})=\sum_{i}(z_{i}^{2}\partial_{z_{i}}f_{\text{cl.}}-2z_{i}b^{2}h_{i})=0.$
(12)
These follow from $\partial_{z_{i}}f_{\text{cl.}}=(\frac{\partial x}{\partial
z_{i}})\partial_{x}f_{\text{cl.}}$ and $\sum_{i}\frac{\partial x}{\partial
z_{i}}=\sum_{i}\frac{\partial x}{\partial z_{i}}z_{i}=\sum_{i}\frac{\partial
x}{\partial z_{i}}z_{i}^{2}=0$ as can easily be shown from the definition of
$x$ and the conformal transformation properties of correlation functions of
primary operators. This yields the final form of the decoupling equation,
$\left[\partial_{z}^{2}+\frac{b^{2}h_{1}}{z^{2}}+\frac{b^{2}h_{2}}{(z-x)^{2}}+\frac{b^{2}h_{3}}{(z-1)^{2}}-\frac{b^{2}(h_{1}+h_{2}+h_{3}-h_{4})}{z(z-1)}+\frac{x(1-x)\partial_{x}f_{\text{cl.}}}{z(z-x)(z-1)}\right]\Psi_{p}=0.$
(13)
To obtain $f_{\text{cl.}}$ from this equation, we use the fact that the
solutions $\Psi_{p}$ must have a certain monodromy when $z$ is taken in a loop
around $0,x$. This monodromy can be derived from the decoupling equation of
$\langle{\mathcal{O}}_{p}\Psi{\mathcal{O}}_{3}{\mathcal{O}}_{4}\rangle$,
$\biggl{[}\frac{1}{b^{2}}\partial_{z}^{2}+\left(\frac{h_{p}}{(z-z_{1})^{2}}+\frac{1}{z-z_{1}}\partial_{z_{1}}\right)+\sum_{i=3,4}\left(\frac{h_{i}}{(z-z_{i})^{2}}+\frac{1}{z-z_{i}}\partial_{z_{i}}\right)\biggr{]}\langle{\mathcal{O}}_{p}\Psi{\mathcal{O}}_{3}{\mathcal{O}}_{4}\rangle=0.$
(14)
As $z\to z_{1}$, the leading coefficient of the OPE between $\Psi$ and
${\mathcal{O}}_{p}$ is given by
$(z-z_{1})^{\kappa}{\mathcal{O}}_{p}^{\prime}(z_{1})$ where $\kappa$ can be
determined by inserting this coefficient into (14),
$\displaystyle\biggl{[}$
$\displaystyle\frac{1}{b^{2}}\kappa(\kappa-1)(z-z_{1})^{\kappa-2}+\sum_{i=3,4}\left(\frac{h_{i}}{(z-z_{i})^{2}}+\frac{1}{z-z_{i}}\partial_{z_{i}}\right)(z-z_{1})^{\kappa}$
(15)
$\displaystyle+h_{p}(z-z_{1})^{\kappa-2}-\kappa(z-z_{1})^{\kappa-2}\biggr{]}\langle{\mathcal{O}}_{p}^{\prime}{\mathcal{O}}_{3}{\mathcal{O}}_{4}\rangle=0.$
The leading contribution in $z\to z_{1}$ is given by
$\biggl{[}\frac{1}{b^{2}}\kappa(\kappa-1)+h_{p}-\kappa\biggl{]}(z-z_{1})^{\kappa-2}=0.$
(16)
Thus as $b^{2}\to 0$ and $z\to z_{1}$,
$\kappa=\frac{1}{2}\left(1\pm\sqrt{1-4h_{p}b^{2}}\right)\text{\leavevmode\nobreak\
\leavevmode\nobreak\ and\leavevmode\nobreak\ \leavevmode\nobreak\
}\Psi_{p}\sim(z-z_{1})^{\frac{1}{2}(1\pm\sqrt{1-4h_{p}b^{2}})}.$ (17)
Therefore, the monodromy matrix around $0,x$ is given by
$M_{0,x}=\left(\begin{array}[]{cc}e^{i\pi(1+\sqrt{1-4h_{p}b^{2}})}&0\\\
0&e^{i\pi(1-\sqrt{1-4h_{p}b^{2}})}\\\ \end{array}\right).$ (18)
The trace of the monodromy matrix, which is independent of the basis in which
the two solutions of (13) are decomposed, is given by
${\mathrm{Tr}}M_{0,x}=-2\cos\left(\pi\sqrt{1-4h_{p}b^{2}}\right).$ (19)
Thus, the torus conformal block can be extracted from (13) by choosing
$\partial_{x}f_{\text{cl.}}$ such that the monodromy of the solution
$\Psi_{p}$ around a loop enclosing $z_{1}$ and $z_{2}$ is given by (19) 333The
loop needs to enclose both $z_{1}$ and $z_{2}$ in order for the OPE between
${\mathcal{O}}_{1}(z_{1})$ and ${\mathcal{O}}_{2}(z_{2})$ to converge..
Finally, the conformal block is obtained by integrating the chosen
$\partial_{x}f_{\text{cl.}}$.
The four point conformal block in other channels is obtained from the same
decoupling equation by imposing different monodromy conditions. For example,
for the $t$-channel block we impose the monodromy condition around the
insertion points of ${\mathcal{O}}_{2}$ and ${\mathcal{O}}_{3}$,
${\mathrm{Tr}}M_{1,x}=-2\cos\left(\pi\sqrt{1-4h_{p}b^{2}}\right)$. Higher
point conformal blocks on the plane are computed from similar monodromy
methods derived analogously to the four-point case. For $n$ point blocks, the
decoupling contains $n-3$ independent derivatives fixed by $n-3$ monodromy
conditions around the operator insertion points which are contracted in the
OPE.
### 2.2 Conformal blocks on the torus
We now continue with the derivation of a monodromy method for conformal blocks
on the torus. The derivation of this monodromy method proceeds in a very
similar way to the one on the plane. We illustrate the derivation using the
two-point function on the torus
$\langle{\mathcal{O}}_{1}(z_{1}){\mathcal{O}}_{2}(z_{2})\rangle_{\tau}={\mathrm{Tr}}[e^{2\pi
i\tau(L_{0}-c/24)}e^{-2\pi
i\bar{\tau}(\bar{L}_{0}-c/24)}{\mathcal{O}}_{1}(z_{1}){\mathcal{O}}_{2}(z_{2})],$
(20)
however conformal blocks for other correlation functions on the torus are
obtained in a similar fashion, as we briefly discuss as the end of this
section. The modular parameter of the torus is denoted by $\tau$, related to
the inverse temperature by $\beta=-2\pi i\tau$. We also introduce the
parameter
$Q=e^{-\beta}=e^{2\pi i\tau},$ (21)
written with an uppercase $Q$ instead of the standard lowercase $q$ to
distinguish it from the conformal dimension $h_{q}$ of the internal index of
the conformal block which we are about to derive.
As on the plane, we insert the degenerate operator $\Psi(z,\bar{z})$ into
(20). To derive the corresponding decoupling equation, we use the conformal
Ward identity on the torus Eguchi:1986sb 444Note that Eguchi:1986sb uses a
convention where correlation functions on the torus are normalized by the
inverse of the partition function $Z(\tau)$ and thus the expression for the
conformal Ward identity there contains an additional term $(2\pi
i\partial_{\tau}Z(\tau))\langle\prod_{i}{\mathcal{O}}_{i}(z_{i})\rangle_{\tau}$.,
$\displaystyle\langle
T(z)\prod_{i}{\mathcal{O}}_{i}(z_{i})\rangle_{\tau}=\left[\sum_{i}\bigl{(}h_{i}\left(\wp(z-z_{i})+2\eta_{1}\right)+\left(\zeta(z-z_{i})+2\eta_{1}z_{i}\right)\partial_{z_{i}}\bigr{)}+2\pi
i\partial_{\tau}\right]\langle\prod_{i}{\mathcal{O}}_{i}(z_{i})\rangle_{\tau}.$
(22)
Here, $z\sim z+1\sim z+\tau$ are the coordinates on the torus with modular
parameter $\tau$ and $\wp(z),\zeta(z)$ denote Weierstraß elliptic functions
with associated $\eta_{1}$ parameter (see App. A for more details on the
Weierstraß functions). Using the definition of the Virasoro generators,
$\hat{L}_{-n}\Psi(z)=\int\frac{dw}{2\pi i}\frac{1}{(w-z)^{n-1}}T(w)\Psi(z).$
(23)
and the conformal Ward identity (22), we see that
$\displaystyle\langle\prod_{i}{\mathcal{O}}_{i}(z_{i})\hat{L}_{-2}\Psi(z)\rangle_{\tau}=\biggl{[}$
$\displaystyle\sum_{i}\left(h_{i}(\wp(z-z_{i})+2\eta_{1})+(\zeta(z-z_{i})+2\eta_{1}z_{i})\partial_{z_{i}}\right)$
(24) $\displaystyle+2\eta_{1}z\partial_{z}+2h_{\Psi}\eta_{1}+2\pi
i\partial_{\tau}\biggr{]}\langle\prod_{i}{\mathcal{O}}_{i}(z_{i})\Psi(z)\rangle_{\tau}$
and
$\langle\prod_{i}{\mathcal{O}}_{i}(z_{i})\hat{L}_{-1}\Psi(z)\rangle_{\tau}=\partial_{z}\langle\prod_{i}{\mathcal{O}}_{i}(z_{i})\Psi(z)\rangle_{\tau}.$
(25)
Thus $\langle\prod_{i}{\mathcal{O}}_{i}(z_{i})\Psi(z)\rangle_{\tau}$ obeys the
decoupling equation
$\displaystyle\biggl{[}$
$\displaystyle\frac{1}{b^{2}}\partial_{z}^{2}+\sum_{i}\left(h_{i}(\wp(z-z_{i})+2\eta_{1})+(\zeta(z-z_{i})+2\eta_{1}z_{i})\partial_{z_{i}}\right)$
(26) $\displaystyle+2\eta_{1}z\partial_{z}+2h_{\Psi}\eta_{1}+2\pi
i\partial_{\tau}\biggr{]}\langle\prod_{i}{\mathcal{O}}_{i}(z_{i})\Psi(z)\rangle_{\tau}=0.$
To relate
$\langle{\mathcal{O}}_{1}(z_{1}){\mathcal{O}}_{2}(z_{2})\Psi(z)\rangle_{\tau}$
to a conformal block, we decompose the trace over states into contributions
from a primary ${\mathcal{O}}_{q}$ and its descendants and insert the
appropriate OPE contractions.
12$p$$q$12$p$$q$ Figure 1: Conformal blocks for the two point function on the
torus. Left: OPE channel, right: projection channel.
For the two-point function, there are two possible channels (see fig. 1). The
projection block is obtained by OPE contracting ${\mathcal{O}}_{2}$ and
${\mathcal{O}}_{q}$. On the other hand, for the OPE block we contract
${\mathcal{O}}_{2}$ and ${\mathcal{O}}_{1}$,
$\displaystyle\langle{\mathcal{O}}_{1}(z_{1}){\mathcal{O}}_{2}(z_{2})\Psi(z)\rangle_{\tau}=\sum_{q}\sum_{l}Q^{h_{q}-c/24+|l|}\langle{\mathcal{O}}_{q}^{\\{l\\}}(z_{0}){\mathcal{O}}_{1}(z_{1}){\mathcal{O}}_{2}(z_{2})\Psi(z){\mathcal{O}}^{\\{l\\}}_{q}(z_{\infty})\rangle\text{(c.c.)}$
(27)
$\displaystyle=\sum_{p,q}C^{p}_{21}\sum_{k,l}(z_{2}-z_{1})^{h_{p}-h_{1}-h_{2}+|k|}Q^{h_{q}-c/24+|l|}\beta^{pk}_{21}\langle{\mathcal{O}}^{\\{l\\}}_{q}(z_{0}){\mathcal{O}}^{\\{k\\}}_{p}(z_{1})\Psi(z){\mathcal{O}}^{\\{l\\}}_{q}(z_{\infty})\rangle\text{(c.c.)},$
where (c.c.) denotes schematically the antiholomorphic parts of the expression
and $z_{0}\to-i\infty$ while $z_{\infty}\to+i\infty$. We define
$\Psi_{pq}=\frac{\langle{\mathcal{O}}_{q}(z_{0})\Psi(z){\mathcal{O}}_{p}(z_{1}){\mathcal{O}}_{q}(z_{\infty})\rangle}{\langle{\mathcal{O}}_{q}(z_{0}){\mathcal{O}}_{p}(z_{1}){\mathcal{O}}_{q}(z_{\infty})\rangle}.$
(28)
As on the plane, in the large $c$ limit
$\langle{\mathcal{O}}^{\\{l\\}}_{q}(z_{0}){\mathcal{O}}^{\\{k\\}}_{p}(z_{1})\Psi(z){\mathcal{O}}^{\\{l\\}}_{q}(z_{\infty})\rangle\text{(c.c.)}\approx\Psi_{pq}\langle{\mathcal{O}}^{\\{l\\}}_{q}(z_{0}){\mathcal{O}}^{\\{k\\}}_{p}(z_{1}){\mathcal{O}}^{\\{l\\}}_{q}(z_{\infty})\rangle\text{(c.c.)}$
(29)
This yields
$\langle{\mathcal{O}}_{1}{\mathcal{O}}_{2}\Psi\rangle_{\tau}\approx\sum_{p,q}C^{p}_{21}C^{q}_{pq}\Psi_{pq}{\mathcal{F}}_{21,pq}\bar{\mathcal{F}}_{21,pq},$
(30)
where ${\mathcal{F}}_{21,pq}$ is the conformal block which we want to compute.
Assuming that exponentiation of the conformal blocks in the semiclassical
limit holds, ${\mathcal{F}}_{21,pq}\sim e^{-c/6\,f_{\text{cl.}}}$, and using
that $\partial_{z_{1}}f_{\text{cl.}}=-\partial_{z_{2}}f_{\text{cl.}}$, we
obtain
$\biggl{[}\partial_{z}^{2}+\sum_{i=1,2}\left(b^{2}h_{i}(\wp(z-z_{i})+2\eta_{1})+\partial_{z_{2}}f_{\text{cl.}}(-1)^{i+1}(\zeta(z-z_{i})+2\eta_{1}z_{i})\right)-2\pi
i\partial_{\tau}f_{\text{cl.}}\biggr{]}\Psi_{pq}=0.$ (31)
From the definition of $\Psi_{pq}$ we derive the monodromy conditions in the
same way as on the plane. For the OPE block, these are
${\mathrm{Tr}}M_{z_{1},z_{2}}=-2\cos(\pi\sqrt{1-4h_{p}b^{2}}),\leavevmode\nobreak\
{\mathrm{Tr}}M_{z_{0}}=-2\cos(\pi\sqrt{1-4h_{q}b^{2}}).$ (32)
The subscripts of the monodromy matrices show around which cycles the
monodromy is taken. The derivation for the projection block works analogously.
Here we contract ${\mathcal{O}}_{q}(z_{0}){\mathcal{O}}_{2}(z_{2})$:
$\displaystyle\langle{\mathcal{O}}_{1}(z_{1}){\mathcal{O}}_{2}(z_{2})\Psi(z)\rangle_{\tau}$
(33)
$\displaystyle=\sum_{p,q}C^{p}_{2q}\sum_{k,l}(z_{2}-z_{0})^{h_{p}-h_{q}-h_{2}+|k|}Q^{h_{q}-c/24+|l|}\beta^{pk}_{2q}\langle{\mathcal{O}}^{\\{k\\}}_{p}(z_{0}){\mathcal{O}}_{1}(z_{1})\Psi(z){\mathcal{O}}^{\\{l\\}}_{q}(z_{\infty})\rangle\text{(c.c.)}.$
Using $\Psi_{pq}$ defined by
$\Psi_{pq}=\frac{\langle{\mathcal{O}}_{p}(z_{0})\Psi(z){\mathcal{O}}_{1}(z_{1}){\mathcal{O}}_{q}(z_{\infty})\rangle}{\langle{\mathcal{O}}_{p}(z_{0}){\mathcal{O}}_{1}(z_{1}){\mathcal{O}}_{q}(z_{\infty})\rangle}$
(34)
and related to the two point correlator by
$\langle{\mathcal{O}}_{1}{\mathcal{O}}_{2}\Psi\rangle_{\tau}\approx\sum_{p,q}C^{p}_{2q}C^{q}_{1p}\Psi_{pq}{\mathcal{F}}_{2q,1p}\bar{\mathcal{F}}_{2q,1p}$
(35)
we find the same decoupling equation (31). However, the monodromy conditions
differ. They are given by
${\mathrm{Tr}}M_{z_{0},z_{2}}=-2\cos(\pi\sqrt{1-4h_{p}b^{2}}),\leavevmode\nobreak\
{\mathrm{Tr}}M_{z_{\infty}}=-2\cos(\pi\sqrt{1-4h_{q}b^{2}}).$ (36)
To solve the decoupling equation, it is useful to perform a coordinate
transformation $u=e^{-2\pi iz}$. Using the transformation of primary operators
under conformal transformations as well as the series representations of the
Weierstrass elliptic functions from app. A, the decoupling equation becomes
$\displaystyle\biggl{[}$
$\displaystyle\partial_{u}^{2}+y(h_{2}-(1+y)\partial_{y}f_{\text{cl.}})\sum_{m=-\infty}^{\infty}\frac{Q^{m}}{u(u-Q^{m})(u-Q^{m}(1+y))}+\frac{1/4-Q\partial_{Q}f_{\text{cl.}}}{u^{2}}$
(37)
$\displaystyle+h_{1}\sum_{m=-\infty}^{\infty}\frac{Q^{m}}{u(u-Q^{m})^{2}}+h_{2}\sum_{m=-\infty}^{\infty}\frac{Q^{m}(1+y)}{u(u-Q^{m}(1+y))^{2}}\biggr{]}\Psi_{pq}=0,$
where we have chosen w.l.o.g. $z_{1}=0$ and $e^{-2\pi iz_{2}}=1+y$. In these
coordinates, the monodromy conditions become
$\displaystyle{\mathrm{Tr}}M_{1,1+y}$
$\displaystyle=-2\cos(\pi\sqrt{1-4h_{p}b^{2}})$
$\displaystyle,\leavevmode\nobreak\ {\mathrm{Tr}}M_{0}=$
$\displaystyle-2\cos(\pi\sqrt{1-4h_{q}b^{2}})$ (OPE block) (38)
$\displaystyle{\mathrm{Tr}}M_{0,1+y}$
$\displaystyle=-2\cos(\pi\sqrt{1-4h_{p}b^{2}})$
$\displaystyle,\leavevmode\nobreak\ {\mathrm{Tr}}M_{\infty}=$
$\displaystyle-2\cos(\pi\sqrt{1-4h_{q}b^{2}})$ (projection block)
This representation of the decoupling equation is immediately applicable for
the calculation of the OPE block, which is defined through a series expansion
in $y$ and $Q$. Using this series expansion as well as a WKB approximation for
large $h_{p},h_{q}$, (37) can be solved order by order. For example, to first
order in $y$ and $Q$ we get
$f_{\text{cl.}}^{\text{OPE}}=-b^{2}(h_{p}-h_{1}-h_{2})\log
y-(b^{2}h_{q}-1/4)\log
Q+\frac{1}{2}b^{2}(h_{p}+h_{2}-h_{1})y-b^{2}\frac{h_{p}^{2}}{2h_{q}}Q+...$
(39)
The projection block, on the other hand, can be expanded in a series in
$q_{1}=Q/(1+y)$ and $q_{2}=1+y$. The decoupling equation can then be solved in
the same way as for the OPE block order by order in $q_{1}$ and $q_{2}$. For
example, to first order in $q_{1}$ and $q_{2}$ we obtain
$\displaystyle f_{\text{cl.}}^{\text{projection}}=$
$\displaystyle-(b^{2}(h_{p}-h_{2})-1/4)\log q_{2}-(b^{2}h_{q}-1/4)\log q_{1}$
(40)
$\displaystyle-b^{2}\frac{(h_{1}-h_{p}+h_{q})(h_{2}-h_{p}+h_{q})}{2h_{q}}q_{1}-b^{2}\frac{(h_{1}+h_{p}-h_{q})(h_{2}+h_{p}-h_{q})}{2h_{p}}q_{2}+...$
We have checked that the results for both the OPE and the projection block are
in agreement with the recursion formulas derived in Cho:2017oxl (see app. B
for detailed expressions) as well as explicit calculations up to third order.
It is clear that the above derivation can be easily generalized to other
conformal blocks on the torus. The simplest case is the zero-point block on
the torus, i.e. the Virasoro character. Performing a similar derivation as
above or equivalently taking the limit $h_{1,2,p}\to 0$ in (37), we arrive at
the following decoupling equation
$\left[\partial_{u}^{2}+\frac{1/4-Q\partial_{Q}f_{\text{cl.}}}{u^{2}}\right]\Psi_{q}=0,$
(41)
together with the monodromy condition
${\mathrm{Tr}}M_{0}=-2\cos(\pi\sqrt{1-4h_{q}b^{2}})$. In this case, we can
give the full solution. The decoupling equation is solved by
$\Psi_{q}=u^{1/2\pm\sqrt{Q\partial_{Q}f_{\text{cl.}}}}$, from which we obtain
$f_{\text{cl.}}=(1/4-b^{2}h_{q})\log Q$ which correctly reproduces the leading
order contribution in $c$ of the Virasoro character
$\chi_{q}=\frac{1}{\eta(\tau)}Q^{h_{q}-(c-1)/24}\sim e^{-c/6f_{\text{cl.}}}$.
For a general $n$-point conformal block, the decoupling equation is given by
$\biggl{[}\partial_{z}^{2}+\sum_{i=1}^{n}\left(b^{2}h_{i}(\wp(z-z_{i})+2\eta_{1})+\partial_{z_{i}}f_{\text{cl.}}(\zeta(z-z_{i})+2\eta_{1}z_{i})\right)-2\pi
i\partial_{\tau}f_{\text{cl.}}\biggr{]}\Psi=0,$ (42)
and there are $n$ monodromy conditions around non-trivial cycles determined by
the OPE contractions. By conformal transformations, the insertion point of one
of the operators can be fixed, for example to $z_{1}\to 0$. Then, there are
$n$ independent accessory parameters $\partial_{\tau}f_{\text{cl.}}$ and
$\partial_{z_{i}}f_{\text{cl.}}$ for $i\geq 2$ fixed by these monodromy
conditions.
### 2.3 Partition function on the replica surface
We now turn to the computation of the partition function $Z_{n}$ on the
replica surface ${\mathcal{R}}_{n}$. In general, the partition function on any
higher genus Riemann surface can be expanded in zero-point conformal blocks,
which can again be calculated via a monodromy method. This monodromy method
can be derived by inserting the degenerate operator directly on the higher
genus Riemann surface – in contrast to the last section, where we inserted it
in a correlation function on the torus – and inserting projection operators in
the appropriate places. The difficulty of this approach is of course that
deriving the decoupling equation for an arbitrary Riemann surface is quite
hard. However, we will see that things simplify for the special higher genus
surface that we are interested in, that is the replica surface
${\mathcal{R}}_{n}$. Assuming that the dominant contribution to the partition
function depends only on the temperature and size of the entangling interval
and not on any other moduli of ${\mathcal{R}}_{n}$, we find the same
decoupling equation as for the twist operator correlator on the torus. The
difference to the last section lies in the monodromy conditions. The zero-
point block on ${\mathcal{R}}_{n}$ admits more general monodromy conditions
(corresponding to different channels) than the conformal block on the torus.
One of these more general monodromy conditions will give the dominant
contribution to the entanglement entropy for large intervals.
Before deriving the decoupling equation on ${\mathcal{R}}_{n}$, we collect
some facts about the topology and moduli of ${\mathcal{R}}_{n}$. For
simplicity, we specialize again to the single interval case. The replica
surface is given by $n$ copies of a torus with modular parameter $\tau$,
joined at a branch cut along the entangling interval $A$. We use coordinates
$z,\bar{z}$ to parametrize ${\mathcal{R}}_{n}$ with identifications $z\sim
z+1$ and $z\sim z+\tau$. In these coordinates, ${\mathcal{R}}_{n}$ is
described by a branched cover of the torus with branch points located at
$z=z_{1,2}+k+l\tau$ for $k,l\in{\mathbb{Z}}$. Near the branch points, the
covering map is given by $y^{n}\propto(z-z_{1}-k-l\tau)$ and $y^{n}\propto
1/(z-z_{2}-k-l\tau)$. The genus of ${\mathcal{R}}_{n}$ is then obtained by the
Riemann-Hurwitz theorem. The ramification index at each branch point is equal
to $n$, yielding $g=n$. Since the Euler characteristic is $\chi<0$, there are
no conformal Killing vectors. This implies by the Riemann-Roch theorem that
there exist $3(n-1)$ holomorphic quadratic differentials $\omega_{zz}^{(i)}$
parametrizing deformations of the complex structure of the Riemann surface.
The $\omega_{zz}^{(i)}$ are meromorphic doubly periodic functions that are
regular everywhere on the covering surface, i.e.
$\omega_{yy}^{(i)}dy^{2}=\omega_{zz}^{(i)}\bigl{(}\frac{dz}{dy}\bigr{)}^{2}dz^{2}$
is non-singular for all $y$. Simple examples include
$\omega_{zz}^{(1)}=\text{const.}$ which is trivially regular and doubly
periodic as well as
$\omega_{zz}^{(2)}=\zeta(z-z_{1})-\zeta(z-z_{2})+2\eta_{1}(z_{1}-z_{2})$.
$\omega_{zz}^{(2)}$ is regular since near $z=z_{1}+k+l\tau$ we have
$\omega_{yy}^{(2)}\propto y^{n-2}$ which is regular at $y=0$ for $n\geq 2$.
Near $z=z_{2}+k+l\tau$, regularity can be shown in an analogous way. In fact,
these two examples are the only ones relevant for the following arguments
since they are the only ones that respect the ${\mathbb{Z}}_{n}$ replica
symmetry permuting the different copies of the torus with each other555This
can be seen as follows. The replica symmetry acts as $y\to ye^{2\pi i/n}$.
Therefore, only the $\omega_{zz}\sim(z-z_{i})^{\alpha_{i}}$ with
$\alpha_{i}\in{\mathbb{Z}}$, $i=1,2$ are invariant under this symmetry. The
case $\alpha_{i}<-1$ is singular at $z=z_{i}$. $\alpha_{i}>0$ is singular at
some other point since any non-constant elliptic function has at least two
poles inside the fundamental parallelogram, which lead to singularities in
$\omega_{yy}$. This leaves only $\alpha_{i}=0,-1$ which are the two examples
described above..
The derivation of the decoupling equation on the replica surface then proceeds
in a similar fashion as in the previous section. Assuming exponentiation of
the zero-point block in the semiclassical limit, the conformal Ward identities
for a general Riemann surface Eguchi:1986sb imply a decoupling equation of
the form
$\left[\partial_{z}^{2}+\langle
T_{zz}\rangle+\sum_{i=1}^{n}\omega_{zz}^{(i)}\partial_{w_{i}}f_{\text{cl.}}\right]\Psi(z)=0,$
(43)
where $w_{i}$ are the modular parameters associated to $\omega_{zz}^{(i)}$.
$\langle T_{zz}\rangle$ is the expectation value of the energy momentum
tensor. It can be derived along the lines of Faulkner:2013yia : $\langle
T_{zz}\rangle$ transforms with a Schwarzian derivative,
$\langle T_{yy}\rangle=\left(\frac{\partial z}{\partial y}\right)^{2}\langle
T_{zz}\rangle+\frac{nc}{12}\\{z,y\\},$ (44)
and $\langle T_{yy}\rangle$ must be regular. The Schwarzian derivative term
comes with a $nc/12$ prefactor since the stress-energy tensor on the replica
surface is given as the sum of the stress-energy tensors of the $n$ tori.
Therefore, the Schwarzian for the transformation of the stress-energy tensor
on the replica surface is given by the sum of $n$ identical Schwarzian terms
with prefactor $c/12$. Together with the requirement that $\langle
T_{zz}\rangle$ be doubly periodic, regularity of $\langle T_{yy}\rangle$
implies
$\langle
T_{zz}\rangle=\frac{c}{24}\left(n-\frac{1}{n}\right)\sum_{i}(\wp(z-z_{i})+2\eta_{1}).$
(45)
The $1/(z-z_{i})^{2}$ poles in $\wp(z-z_{i})$ give a $1/y^{2}$ contribution to
$\langle T_{yy}\rangle$ that cancels with the Schwarzian derivative term666For
ease of comparison with the previous section we have also added a constant
term $\frac{c}{24}\left(n-\frac{1}{n}\right)4\eta_{1}$ to $\langle
T_{zz}\rangle$ which is not strictly necessary for regularity and could be
absorbed into the prefactor of $\omega_{zz}^{(1)}$.. Letting the sum over $i$
in (43) run only over $i=1,2$, we recover the decoupling equation (31) for the
twist correlator. Restricting the sum to this range means that we assume
$\partial_{w_{i}}f_{\text{cl.}}=0$ for $i>2$, i.e. we assume that the result
for the partition function on the replica surface does not depend on other
moduli of the replica surface than the size of the torus $\tau$ and the length
of the entangling interval $z_{2}-z_{1}$.
$\langle(P_{p}{\mathcal{O}}_{1}{\mathcal{O}}_{2}P_{p}){\mathcal{O}}_{3}{\mathcal{O}}_{4}\rangle\sim$$4$$3$$p$$1$$2$$\sim$$P_{p}$${\mathcal{O}}_{1}$${\mathcal{O}}_{2}$${\mathcal{O}}_{3}$${\mathcal{O}}_{4}$$\langle{\mathcal{O}}_{1}(P_{p}{\mathcal{O}}_{2}{\mathcal{O}}_{3}P_{p}){\mathcal{O}}_{4}\rangle\sim$$4$$1$$p$$3$$2$$\sim$${\mathcal{O}}_{1}$$P_{p}$${\mathcal{O}}_{2}$${\mathcal{O}}_{3}$${\mathcal{O}}_{4}$
Figure 2: Inserting projection operators $P_{p}$ onto the Verma module of a
primary ${\mathcal{O}}_{p}$ into a correlator yields the conformal block with
internal weight $h_{p}$.
However, as mentioned in the beginning of this section, the admitted monodromy
conditions for the decoupling equation (43) are more general than those of
(31) for the twist correlator. To see this, recall that conformal blocks of
any correlation function can be obtained in two equivalent ways. Either we can
perform OPE contractions between two or more operators and then keep only
terms of particular primaries and their descendants in the OPE or equivalently
we can insert projection operators onto the Verma modules of these primaries
in the correlation function at appropriate places. The projectors of the
latter approach can be thought of as non-local operators acting in a closed
line around the operators whose OPE contractions are performed in the former
approach (see fig. 2). For the zero-point block on an arbitrary higher genus
Riemann surface, there are in general $3(n-1)$ projectors to be inserted
corresponding to $3(n-1)$ monodromy conditions. However, as mentioned above we
assume that the partition function on the higher genus Riemann surface depends
only on two of the moduli and thus we consider only two of the $3(n-1)$
monodromy conditions.
Which monodromy conditions are appropriate for the calculation of the
entanglement entropy? For the conformal block, the monodromy conditions must
be taken around the spatial circle and around $z_{1},z_{2}$ 777It is also
possible to calculate the modular transformed block for the twist correlator,
which we expect to be the dominant contribution at large temperature and small
intervals. In this case, the monodromy conditions are taken around
$z_{1},z_{2}$ and around the time circle.. On the other hand, for the zero-
point block on the replica surface the prescription described in this section
still leaves open the question of where to put the monodromy conditions – i.e.
which channel to choose – in order to obtain the dominant contribution to the
partition function from the vacuum block. Taking the limits of high and low
temperature, it is clear that for small intervals one of the monodromy
conditions must be taken around the spatial circle for low temperatures and
the time circle for high temperatures, while the other monodromy condition
must be imposed around the entangling interval $A$ between $z_{1}$ and
$z_{2}$. For large intervals, the correct monodromy condition is obtained by
reformulating the problem along the lines of Cardy:2014jwa ; Chen:2015kua . We
separate the branch cut on the torus along $A$ yielding the replica surface
into a branch cut along the full spatial circle and a branch cut in the
opposite direction along $A^{c}$ (see fig. 3). We then impose trivial
monodromy around $A^{c}$ to fix the dependence on the size of the entangling
interval. For small temperatures, the monodromy condition around the spatial
circle remains unchanged. However, for high temperatures the monodromy
condition around the time circle is now transformed into a monodromy condition
around a time circle of size $n\tau$, since the branch cut along the full
spatial circle connects all $n$ replica copies together to effectively create
a torus with modular parameter $n\tau$.
$A$$A^{c}$ Figure 3: Branch cut structure for large intervals. We can
decompose the branch cut along $A$ (denoted in red on the left) into a branch
cut along the full spatial circle (denoted in red on the right) and a branch
cut in the opposite direction along $A^{c}$ (denoted in blue on the right).
Note again that it is perfectly valid to use any of the above monodromy
conditions for all values of the temperature and entangling interval size.
However, outside of the regimes of validity of the monodromy conditions
described above, we don’t expect the vacuum block to give the dominant
contribution in the semiclassical limit and thus the partition function in
this case would be obtained by summing up all of the conformal blocks for
different values of the dimensions of the exchanged operators. The cross-over
point between the regimes must be determined by an analysis of these
contributions from the exchange of non-identity operators.
Let us also note that explicit calculations for the free fermion case support
the arguments presented in this section with regards to the differences
between twist correlators and partition functions on ${\mathcal{R}}_{n}$ and
with regards to the monodromy condition for large intervals. Namely, in
Lokhande:2015zma it was observed that the twist correlator on the torus does
not give the correct answer for the entanglement entropy and in particular
violates Bose-Fermi equivalence and modular covariance. This was traced back
in Mukhi:2017rex to the way in which different spin structures of the replica
surface ${\mathcal{R}}_{n}$ combine to give the total answer for the partition
function $Z_{n}$. The replica surface, being composed of $n$ copies of a
torus, has $2n$ nontrivial cycles around which the fermions have either
periodic or antiperiodic boundary conditions. To calculate the total partition
function $Z_{n}$ on the replica surface, it is necessary to sum over all
possible spin structures each of which corresponds to a particular choice of
boundary conditions around the nontrivial cycles of the replica surface. For
small intervals, the replica surface essentially factorizes into $n$
unconnected torus copies. In this limit, it was shown in Mukhi:2017rex that
$Z_{n}$ is given by an “uncorrelated” sum where the summation over spin
structures is performed for each of the $n$ tori separately. For large
intervals, $Z_{n}$ is given by a “correlated” sum where only one sum over spin
structures is performed, i.e. we take equal boundary conditions around the
time resp. space circles of each of the $n$ replica copies Mukhi:2017rex . In
this case, the partition function $Z_{n}$ of the replica surface is
essentially given by the partition function on a torus with modular parameter
$n\tau$.
## 3 Entanglement entropy at large central charge
In this section, we present the calculation of the entanglement entropy on the
torus at large central charge. As explained in the previous section, the
entanglement entropy can be obtained from the partition function $Z_{n}$ on
the replica surface ${\mathcal{R}}_{n}$ which decomposes into zero-point
conformal blocks. The claim we want to investigate is that at large central
charge $c\to\infty$ and for $n\to 1$ the dominant contribution to $Z_{n}$
comes from the vacuum block with $h_{p}=h_{q}=0$. The derivation of this
statement proceeds as follows. First, it is necessary to show that the
semiclassical limit is well-defined not only for $h_{p,q}={\mathcal{O}}(c)$
but also for $h_{p,q}={\mathcal{O}}(c^{0})$. This means that for
$h_{p,q}=\gamma c$ the limits $\lim_{c\to\infty}$ and $\lim_{\gamma\to 0}$ of
the conformal block commute. A discussion of this point starting from the
recursion relation for torus conformal blocks is relegated to app. B. In the
next step we solve the decoupling equation (31) perturbatively in
$\varepsilon=n-1$ up to first order. Imposing the monodromy conditions derived
in sec. 2.2 then yields the conformal block from which we finally extract the
entanglement entropy.
We first consider the limits of high and low temperature as well as small and
large entangling interval size in sec. 3.1. Each combination of these limits
comes with different monodromy conditions as explained in the previous
section. We find agreement with the known universal results in the limits
where the torus degenerates into a cylinder. We then examine the conditions on
the CFT spectrum and OPE coefficients under which these results extend to
intermediate temperature and interval size regimes in sec. 3.2, leading to the
conclusion that for holographic CFTs the results from sec. 3.1 are valid for
all temperatures and interval sizes. Moreover, we consider the case of
multiple intervals on the torus. Finally, we numerically check the assumption
on the dominance of the vacuum block in sec. 3.4.
### 3.1 Limiting cases
#### 3.1.1 Low temperature and small intervals
In the low temperature limit $\beta\to\infty$ the torus degenerates into a
cylinder with periodic space direction. For the cylinder, the entanglement
entropy of a single interval can be obtained directly by mapping this cylinder
to the plane and using the known formula for the entanglement entropy on the
plane Calabrese:2004eu ,
$S_{A}=\frac{c}{3}\log\left(\sin(\pi(z_{2}-z_{1}))\right)+\text{const.}$ (46)
To obtain the same result from the monodromy method, we series expand
$\Psi_{pq}$ and $f_{\text{cl.}}$ in $n-1$:
$\Psi_{pq}=\sum_{k}\Psi_{pq}^{(n)}(n-1)^{k}$ and
$f_{\text{cl.}}=\sum_{k}f_{n}(n-1)^{k}$. The decoupling equation (31) at
zeroth order in $n-1$ becomes
$[\partial_{z}^{2}-2\pi i\partial_{\tau}f_{0}]\Psi_{pq}^{(0)}(z)=0.$ (47)
This is solved by
$\Psi^{(0)}_{pq}(z)=\exp(\pm\sqrt{2\pi i\partial_{\tau}f_{0}}z).$ (48)
By a coordinate transformation $u=\exp(-2\pi iz)$, we obtain
$\tilde{\Psi}_{pq}^{(0)}(u)=u^{h_{\Psi}}\Psi_{pq}^{(0)}(z=i\log(u)/(2\pi))$.
Imposing trivial monodromy of $\tilde{\Psi}_{pq}^{(0)}(u)$ around $u=0$ is
equivalent to antiperiodic monodromy conditions for $\Psi_{pq}^{(0)}(z)$
around the spatial circle of the torus,
$\Psi_{pq}^{(0)}(z+1)=-\Psi_{pq}^{(0)}(z)$. As expected, this implies that
$f_{0}$ is equal to the leading order in $c$ of the vacuum character on the
torus,
$f_{0}=\pi i\tau/2=-\beta/4\hskip 28.45274pt\Leftrightarrow
e^{-c/6f_{0}}=e^{c/24\beta}=\left.\chi_{h=0}(\beta)\right|_{c\to\infty}.$ (49)
At first order in $n-1$, the decoupling equation is given by
$[\partial_{z}^{2}-2\pi
i\partial_{\tau}f_{0}]\Psi_{pq}^{(1)}(z)+m(z)\Psi_{pq}^{(0)}(z)=0,$ (50)
yielding
$\Psi_{pq}^{(1)}(z)=\frac{e^{-i\pi z}}{2\pi i}\int^{z}dx\,m(x)e^{i\pi
x}\Psi_{pq}^{(0)}(x)-\frac{e^{i\pi z}}{2\pi i}\int^{z}dx\,m(x)e^{-i\pi
x}\Psi_{pq}^{(0)}(x),$ (51)
where $m(z)$ is given by
$m(z)=\sum_{i}\left(\frac{1}{2}(\wp(z-z_{i})+2\eta_{1})+(-1)^{i+1}(\zeta(z-z_{i})+2\eta_{1}z_{i})\partial_{z_{2}}f_{1}\right)-2\pi
i\partial_{\tau}f_{1}.$ (52)
To compute the conformal block we impose trivial monodromy around
$z_{1},z_{2}$ which is equivalent to the vanishing of
$\oint_{z_{1},z_{2}}dx\,m(x)e^{\pm i\pi x}\Psi_{pq}^{(0)}(x).$ (53)
This gives the $z_{1},z_{2}$ dependence of $f_{1}$,
$f_{1}=\log(\sin(\pi(z_{2}-z_{1})))+C_{1}(\tau).$ (54)
From trivial monodromy around $u=0$ we find the $\tau$ dependence to be
$\partial_{\tau}f_{1}=0$, which implies that $C_{1}(\tau)=\text{const.}$ is
independent of $\tau$. The antiholomorphic conformal block $\bar{f}_{1}$ gives
the same result as the holomorphic one. Then the entanglement entropy is given
by $S_{A}=\frac{c}{6}(f_{1}+\bar{f}_{1})$, in agreement with (46). In this
limit, the OPE vacuum block of the twist correlator gives the same results,
since it is given by the same monodromy method as the zero-point vacuum block
of the replica partition function computed in this section.
#### 3.1.2 Low temperature and large intervals
In this limit, we demand trivial monodromy around the spatial circle (i.e.
around $u=e^{-2\pi iz}=0$) as well as trivial monodromy around $A^{c}$ (i.e.
around $z_{1},z_{2}-1$). This gives the same vacuum block and thus the same
entanglement entropy (46) as in the small interval case at low temperature.
Also in this case, the twist correlator gives the correct result.
#### 3.1.3 High temperature and small intervals
In the high temperature limit $\beta\to 0$ the torus again degenerates into a
cylinder, now with periodic time direction. As in the low temperature case,
the entanglement entropy of a single interval can be obtained by mapping this
cylinder to the plane Calabrese:2004eu ,
$S_{A}=\frac{c}{3}\log\left(\frac{\tau}{i\pi}\sinh\left(\frac{i\pi}{\tau}(z_{2}-z_{1})\right)\right)+\text{const.}$
(55)
In the monodromy method, we impose trivial monodromy around the time circle
and around $z_{1},z_{2}$. As in the low temperature limit, we solve the
decoupling equation (31) in a series expansion around $n-1$. An analogous
calculation as above yields
$f_{\text{cl.}}=-\frac{\pi
i}{2\tau}+\epsilon\log\left(\tau\sinh\left(\frac{\pi
i}{\tau}(z_{2}-z_{1})\right)\right)+\text{const.}=f_{0}+\epsilon f_{1}$ (56)
At zeroth order in $\epsilon$ we recover the leading order in $c$ of the
vacuum character
$\chi_{h=0}(\beta)=e^{\frac{c}{24}\frac{4\pi^{2}}{\beta}}=e^{-\frac{c}{6}(-\frac{\pi
i}{2\tau})}$. The entanglement entropy given by the first order contribution,
$S_{A}=\frac{c}{6}(f_{1}+\bar{f}_{1})$, agrees with the result from the
cylinder (55).
From the twist correlator point of view, we obtain the high temperature limit
by a modular transformation $\tau\to-1/\tau$ from the low temperature result.
The two point function on the torus transforms covariantly,
$\langle{\mathcal{O}}_{1}(-z_{1}/\tau,-\bar{z}_{1}/\bar{\tau}){\mathcal{O}}_{2}(-z_{2}/\tau,-\bar{z}_{2}/\bar{\tau})\rangle_{-1/\tau}=(-\tau)^{h_{1}+h_{2}}(-\bar{\tau})^{\bar{h}_{1}+\bar{h}_{2}}\langle{\mathcal{O}}_{1}(z_{1},\bar{z}_{1}){\mathcal{O}}_{2}(z_{2},\bar{z}_{2})\rangle_{\tau}.$
(57)
If the two point function of twist operators is dominated by a single
conformal block, the conformal block for the modular transformed $\tau$ is
obtained from the modular transformation properties of the correlation
function as
$f_{\text{cl.}}^{-1/\tau}(-z_{1}/\tau,-z_{2}/\tau)=-b^{2}(h_{1}+h_{2})\log(-\tau)+f_{\text{cl.}}^{\tau}(z_{1},z_{2}).$
(58)
Therefore, we can immediately read off the high temperature behavior of the
twist correlator from
$\left.f_{\text{cl.}}^{\tau}(z_{1},z_{2})\right|_{\tau\to\infty}=\left[f_{\text{cl.}}^{-1/\tau}(-z_{1}/\tau,-z_{2}/\tau)+b^{2}(h_{1}+h_{2})\log(-\tau)\right]_{\tau\to\infty}$
(59)
For $h_{1}=h_{2}=1/2\epsilon$,
$\left.f_{\text{cl.}}^{\tau}(z_{1},z_{2})\right|_{\tau\to 0}=\frac{\pi
i\tau}{2}+\epsilon\log(\sin(\pi(z_{2}-z_{1})))+\text{const.}$ we again obtain
(56). Thus the twist correlator agrees with the replica partition in this
limit.
Let us note that is also possible to determine the high temperature expansion
of twist correlator by applying the modular transformation to the monodromy
conditions. At low temperatures, we demand trivial monodromy around the
spatial circle of the torus at time. Since the modular transformation
$\tau\to-1/\tau$ exchanges the time and space directions of the torus, we can
obtain the high temperature behavior of the twist correlator by demanding
trivial monodromy around the time circle of the torus, showing directly the
equivalence between the twist correlator result and the replica partition
function in the limit of high temperature and small entangling intervals.
#### 3.1.4 High temperature and large intervals
As explained in sec. 2.3, the correct monodromy condition of the zero-point
block on the replica surface for large intervals imposes trivial monodromy
around $A^{c}$ and $z\to z+n\tau$. To zeroth order in $n-1$ the decoupling
equation (31) is solved by $\Psi_{pq}^{(0)}(z)=\exp(\pm\sqrt{2\pi
i\partial_{\tau}f_{0}})$. Imposing
$\Psi_{pq}^{(0)}(z+n\tau)=-\Psi_{pq}^{(0)}(z)$ yields
$f_{0}=-\frac{\pi i}{2n^{2}\tau}.$ (60)
To first order, the solution reads
$\Psi_{pq}^{(1)}(z)=n\tau\frac{e^{-\frac{i\pi z}{n\tau}}}{2\pi
i}\int^{z}dx\,m(x)e^{\frac{i\pi
x}{n\tau}}\Psi_{pq}^{(0)}(x)-n\tau\frac{e^{\frac{i\pi z}{n\tau}}}{2\pi
i}\int^{z}dx\,m(x)e^{-\frac{i\pi x}{n\tau}}\Psi_{pq}^{(0)}(x).$ (61)
Imposing trivial monodromy around $A^{c}$ and expanding in $n-1$ we obtain
$f_{\text{cl.}}=-\frac{\pi
i}{2\tau}+(n-1)\left(\frac{i\pi}{\tau}+\log\left(\tau\sinh\left(\frac{i\pi}{\tau}(1-(z_{2}-z_{1}))\right)\right)\right)+\text{const.}$
(62)
giving
$S_{A}=\frac{c}{3}\left(\frac{i\pi}{\tau}+\log\left(\frac{\tau}{i\pi}\sinh\left(\frac{i\pi}{\tau}(1-(z_{2}-z_{1}))\right)\right)\right)+\text{const.}$
(63)
As expected from general arguments Azeyanagi:2007bj ; Cardy:2014jwa , the
difference between the entanglement entropy for $A$ and for the complement
$A^{c}$ in the limit of large interval size is given by the thermal entropy
$S(\beta)=\beta^{2}(-\beta^{-1}\partial_{\beta}\log
Z(\beta))=\frac{c}{3}\frac{2\pi^{2}}{\beta}=\frac{c}{3}\frac{i\pi}{\tau}$,
using the partition function $Z(\beta\to
0)=\exp(\frac{c}{12}\frac{4\pi^{2}}{\beta})$ from the Cardy formula. The twist
correlator, on the other hand, cannot reproduce this feature as is easy to see
by applying the modular transformation argument from the last section which
gives (56) in disagreement with (62).
### 3.2 Holographic CFTs
We now argue that the results in the limits considered in the previous section
are valid for all temperatures and interval sizes in holographic CFTs. This
statement holds if the vacuum block computed in the previous section gives the
dominant contribution to the partition function on the replica manifold
${\mathcal{R}}_{n}$ in the large central charge limit.
Let us first consider the case $n=1$, i.e. the zeroth order in the $n-1$
expansion. In this order, the monodromy method calculates the vacuum character
to leading order in $c$. If this vacuum character dominates the partition
function, $Z(\beta)$ takes on the universal form
$Z(\beta)=\left\\{\begin{array}[]{ll}\exp\bigl{(}\frac{c}{12}\beta\bigr{)}&,\leavevmode\nobreak\
\beta>2\pi\\\
\exp\bigl{(}\frac{c}{12}\frac{4\pi^{2}}{\beta}\bigr{)}&,\leavevmode\nobreak\
\beta<2\pi\end{array}\right.$ (64)
for all temperatures. For consistency of the computation method, dominance of
the conformal block to first order in $n-1$ also requires dominance of the
zeroth order in $n-1$. It has been argued in Hartman:2014oaa that a partition
function of the form (64) is characteristic of a holographic CFT. Thus it is a
necessary condition that the CFT in question be holographic in order for the
results of sec. 3.1 to hold at arbitrary temperatures.
More explicit conditions on the CFT in question can be given following an
argument for dominance of the vacuum block in the zero temperature case which
goes as follows Hartman:2013mia . The arguments below hold for CFTs with OPE
coefficients $C^{p}_{21}$ that grow at most exponentially with $c$ 888This is
equivalent to the requirement that correlation functions obey the cluster
decomposition principle and are smooth in a neighborhood of the point where
multiple operator insertion points coincide Hartman:2013mia . and a density of
states for light operators that does not grow with $c$. Here light operators
mean operators with conformal weight $h,\bar{h}<h_{\text{gap}}$ where
$h_{\text{gap}}$ is of the order of the central charge. In the large central
charge limit, the conformal block expansion in the $s$-channel of the four-
point twist correlator on the plane takes on the form
$\langle\sigma_{n}(0)\sigma_{n}(x)\sigma_{n}(1)\sigma_{n}(\infty)\rangle=\sum_{p}C^{p}_{21}\exp\left(-\frac{c}{6}\bigl{(}f_{\text{cl.}}(h_{\sigma_{n}}/c,h_{p}/c,x)+\bar{f}_{\text{cl.}}(\bar{h}_{\sigma_{n}}/c,\bar{h}_{p}/c,\bar{x})\bigr{)}\right).$
(65)
On account of the cluster decomposition principle, the vacuum exchange is the
leading contribution in the $s$-channel around $x=0$ and in the $t$-channel
around $x=1$. This implies that the contribution of the heavy operators with
$h,\bar{h}>h_{\text{gap}}$ in the $s$-channel in some finite region around
$x=0$ is suppressed exponentially and similarly for the $t$-channel around
$x=1$. The sparse spectrum of the light operators allows ignoring the
multiplicity factors for $h,\bar{h}<h_{\text{gap}}$. Consider first the
scenario that the OPE coefficients grow subexponentially with $c$. In this
case, we can also ignore the coefficient $C^{p}_{21}$ and the sum over the
light operators in (65) is dominated exponentially in $e^{-c}$ by its largest
term. If the semiclassical conformal block $f_{\text{cl.}}$ is a monotonically
increasing function with $h_{p}/c$, this implies that the vacuum block with
the lowest possible $h_{p}/c=0$ dominates. In the case that the OPE
coefficients grow exponentially with $c$, the sum over the light operators
gives
$\int_{0}^{h_{\text{gap}}}dh_{p}d\bar{h}_{p}\exp\left(\frac{c}{6}\bigl{(}g(h_{p}/c,\bar{h}_{p}/c)-f_{\text{cl.}}(h_{\sigma_{n}}/c,h_{p}/c,x)-\bar{f}_{\text{cl.}}(\bar{h}_{\sigma_{n}}/c,\bar{h}_{p}/c,\bar{x})\bigr{)}\right),$
(66)
where $g(h_{p}/c,\bar{h}_{p}/c)$ contains the contribution of the OPE
coefficients and multiplicities. This integral is dominated either by the
endpoints of the integration or by a saddle point. Near coincident points, the
leading universal term in any correlation function is given by the vacuum
exchange. Therefore, in a finite region around $x=0$ and $x=1$, the integral
in (66) is dominated by the endpoint at $h_{p}=0$. However, now – unlike for
subexponentially growing OPE coefficients – it is not possible to exclude
saddle points dominating the integral (66) in some finite subset of
$x\in[0,1]$. Examples of such saddle points have been found in Belin:2017nze
for genus two partition functions computing the third Rényi entropy on the
plane.
In summary, the vacuum block dominates correlation functions in a finite
region around coincident operator insertion points if the conformal block
$f_{\text{cl.}}$ increases monotonically with the internal conformal weights
$h_{p,q}$. In sec. 3.4, we present numerical evidence that this is indeed the
case. Therefore, the results of sec. 3.1 are valid for all temperatures and
interval sizes assuming subexponentially growing OPE coefficients and a sparse
spectrum of light operators. For exponentially growing OPE coefficients, the
results are still valid in a finite region around the respective limiting
points. The points at which the different limits exchange dominance are
obtained as follows. From the requirement of consistency of our results at
$n=1$ with the partition function (64) we see that the low and high
temperature regimes exchange dominance at the Hawking-Page phase transition
point $\beta=2\pi$. The transition between small and large interval behavior
in the high temperature regime is estimated by equating the results for the
conformal block in the small and large interval regime. The dominant
contribution comes from the smaller conformal block.
Assuming the aforementioned restrictions on the CFT spectrum and OPE
coefficients, the results match perfectly with the predictions from the RT
formula Ryu:2006bv ; Azeyanagi:2007bj . There is also a direct way to
implement the monodromy computation for the calculation of the entanglement
entropy in the dual gravity theory, as was shown in Faulkner:2013yia at zero
temperature. This was generalized to the finite temperature case in
Barrella:2013wja ; Chen:2015kua . We obtain the same monodromy method as
Barrella:2013wja ; Chen:2015kua from the CFT side. This clearly shows that
our results are valid for holographic CFTs.
Furthermore, the high temperature and large interval size result (62) agrees
with a CFT calculation of the vacuum sector contribution to the entanglement
entropy done in Chen:2015kua using complementary techniques to obtain the
vacuum block. Specifically, the conformal block is obtained in Chen:2015kua
by an explicit construction of the Virasoro generators and descendant states
in the twisted sector of a ${\mathbb{Z}}_{n}$ orbifold. This allows for a
series expansion of the Rényi entropy. Moreover, in the limit $n\to 1$, the
authors of Chen:2015kua find that the leading order of the semiclassical
vacuum block is given by an expression in terms of the four-point function of
twist-operators on the plane, which gives the same conformal block (62) that
we obtain using the monodromy method.
Let us also note that the universal form of the partition function (64) in
holographic theories can be derived from the same vacuum block dominance
argument as the universal form of the entanglement entropy. The partition
function can be expanded in zero-point blocks on the torus either in a low
temperature expansion (zero-point blocks are Virasoro characters) or in a high
temperature expansion (zero-point blocks are modular transformed Virasoro
characters). From the known form of the Virasoro characters, the leading order
contribution in the central charge of the characters is given by $\chi\sim
Q^{h_{q}-c/24}$ where $Q=e^{2\pi i\tau}$ resp. $Q=e^{-2\pi i/\tau}$ in the low
and high temperature expansions. Since $(-6/c)\log\chi$ is an increasing
function of $h_{q}$, the same arguments as for the entanglement entropy given
above apply to the partition function which is dominated by the vacuum
character with $h_{q}=0$.
### 3.3 Multiple intervals
The generalization to an entangling interval consisting of the union of
multiple intervals, $A=[z_{1},z_{2}]\cup[z_{3},z_{4}]...[z_{2N-1},z_{2N}]$, is
straightforward. The decoupling equation is given by
$\left[\partial_{z}^{2}+\sum_{i=1}^{2N}\left(\frac{1}{4}(n-1/n)(\wp(z-z_{i})+2\eta_{1})-\partial_{z_{i}}f_{\text{cl.}}(\zeta(z-z_{i})+2\eta_{1}z_{i})\right)-2\pi
i\partial_{\tau}f_{\text{cl.}}\right]\Psi(z)=0.$ (67)
We impose trivial monodromy around $N$ pairs $(i,j)$ of interval endpoints
$z_{i},z_{j}$ to fix the $\partial_{z_{i}}f_{\text{cl.}}$. The temperature
dependence is fixed by demanding trivial monodromy around either the spatial
circle, a time circle of size $\tau$ or a time circle of size $n\tau$
depending on the temperature and total entangling interval size
$|A|=\sum_{i}|z_{2i}-z_{2i-1}|$. This yields
* •
low temperature: trivial monodromy for $z\to z+1$
$S_{A}=\frac{c}{3}\sum_{(i,j)}\log\left(\sin(\pi(z_{i}-z_{j}))\right)+\text{const.}$
(68)
* •
high temperature and small total interval size: trivial monodromy for $z\to
z+\tau$
$S_{A}=\frac{c}{3}\sum_{(i,j)}\log\left(\frac{\tau}{i\pi}\sinh\left(\frac{i\pi}{\tau}(z_{i}-z_{j})\right)\right)+\text{const.}$
(69)
* •
high temperature and large total interval size: trivial monodromy for $z\to
z+n\tau$
$S_{A}=\frac{c}{3}\left(\frac{i\pi}{\tau}+\sum_{(i,j)}\log\left(\frac{\tau}{i\pi}\sinh\left(\frac{i\pi}{\tau}(z_{i}-z_{j})\right)\right)\right)+\text{const.}$
(70)
Which monodromy condition and which combination of pairs $(i,j)$ to take, i.e.
in which channel to expand the conformal block, depends on the interval size.
We expect the dominant contribution to come from the channel with the smallest
$f_{\text{cl.}}$, in which case we find agreement with the RT formula.
However, we caution that this argument depends on the vacuum block dominating
the partition function, which we have checked only for a single interval.
One particular interesting special case of the above calculation is the time
dependence of the entanglement entropy between two intervals on opposite
boundaries of a two-sided BTZ black hole. The two-sided BTZ black hole is dual
to the thermofield double state. On the CFT side, the entanglement entropy is
obtained by positioning the two intervals at a distance $\tau/2$ in the time
coordinate on the torus. To obtain the time evolution of the entanglement
entropy, it is necessary to perform an analytic continuation of the endpoints
Hartman:2013qma ,
$\displaystyle z_{1}=\bar{z}_{1}=0$ $\displaystyle z_{2}=\bar{z}_{2}=L$ (71)
$\displaystyle z_{3}=2t+L+\tau/2,\bar{z}_{3}=-2t+L+\bar{\tau}/2$
$\displaystyle z_{4}=2t+\tau/2,\bar{z}_{4}=-2t+\bar{\tau}/2,$
where $L$ is the size of the entangling interval taken to be equal on both
boundaries and $t$ is the time coordinate at which both parts of the
entangling interval are placed on the asymptotic boundaries of the
wormhole999Note that this time coordinate has nothing to do with the euclidean
time coordinate on the torus on which we calculate the finite temperature
correlator in the euclidean CFT. (see Hartman:2013qma for more details on the
setup). For small interval size $L$, there are two conformal blocks to
consider. At early times, i.e. for small $t$, the dominant contribution comes
from imposing trivial monodromy around $z_{1},z_{4}$ and $z_{2},z_{3}$, while
at late times the vacuum block with trivial monodromy around $z_{1},z_{2}$ and
$z_{3},z_{4}$ dominates. Taking into account that due to the analytic
continuation $f_{\text{cl.}}\neq\bar{f}_{\text{cl.}}$, the corresponding
entanglement entropy is given by
$S_{A}=\frac{2c}{3}\log\left(\frac{\tau}{i\pi}\cosh\left(\frac{2\pi
i}{\tau}t\right)\right)+\text{const.}$ (72)
at early times and by
$S_{A}=\frac{2c}{3}\log\left(\frac{\tau}{i\pi}\sinh\left(\frac{i\pi}{\tau}L\right)\right)+\text{const.}$
(73)
at late times. This reproduces the phase transition in the dual RT surfaces
from geodesics that connect the two boundaries through the interior of the
two-sided black hole at early times to disconnected geodesics on opposite
boundaries that do not enter the black hole interior at late times
Hartman:2013qma .
### 3.4 Vacuum block dominance
In this section, we provide numerical evidence that the vacuum block
exponentially dominates the twist correlator in the large $c$ limit. For
simplicity, we restrict to the single interval case. Assuming the same
conditions on the spectrum and OPE coefficients of the CFT detailed in the
last section, we need to show that the conformal block monotonically increases
with the weight of the internal operators $h_{p,q}$.
For the zero temperature case, this was done numerically in Hartman:2013mia
for arbitrary $n$, giving evidence that the Rényi entropies are given by the
vacuum conformal block contribution only. However, the calculation is much
simpler if we restrict to $n$ close to one which implies $h_{p}/c\gg
h_{i}/c\to 0$ for $i=1,2,3,4$. In this limit, the conformal block can be
obtained in closed form from the monodromy calculation by a WKB expansion in
$1/(h_{p}/c)$ Zamolodchikov1987 ,
$\frac{c}{6}f_{\text{cl.}}(0,h_{p}/c,x)=h_{p}\left(\pi\frac{K(1-x)}{K(x)}-\log
16\right),$ (74)
where $K(x)$ is the complete elliptic integral of the first kind. Thus,
$f_{\text{cl.}}$ is an increasing function of $h_{p}/c$ if $\pi
K(1-x)/K(x)-\log 16>0$ which can easily be checked to be fulfilled for
$x<1/2$. In fact, $f_{\text{cl.}}(0,h_{p}/c,x)-f_{\text{cl.}}(0,h_{p}/c,1-x)$
reaches its crossover point exactly at $x=1/2$, confirming that at this point
dominance is exchanged from the $s$ to the $t$-channel block.
Applying the same arguments as on the plane to the case of the torus, it is
clear that the vacuum block dominates if the semiclassical block is an
increasing function of $h_{p}$ and $h_{q}$. Without loss of generality, we
take $z_{1}=0$ in the following. Restricting again to $n\approx 1$, we need to
find the semiclassical block in the limit $h_{p,q}/c\gg h_{1,2}/c\to 0$.
Unlike on the plane, however, $f_{\text{cl.}}$ can not easily be obtained in a
closed form expression from the monodromy calculation in this limit101010The
reason for this is that the solution of the decoupling equation on the torus
takes on the schematic form of an integral over
$\sqrt{A\partial_{z_{2}}f_{\text{cl.}}+B\partial_{\tau}f_{\text{cl.}}}$ for
some functions $A$ and $B$, from which it is not easily possible to extract
$\partial_{z_{2}}f_{\text{cl.}}$ and $\partial_{\tau}f_{\text{cl.}}$. The
solution of the decoupling equation on the plane, on the other hand, is given
by an integral over $\sqrt{A\partial_{x}f_{\text{cl.}}}$ from which
$\partial_{x}f_{\text{cl.}}$ can be factored out immediately.. Thus, we apply
a series expansion in $y=e^{-2\pi iz_{2}}-1$ and $Q=e^{2\pi i\tau}$ on top of
the WKB approximation in $1/(h_{p,q}/c)$. While this yields a very precise
numerical approximation to the true value of the conformal block if enough
terms are included in the series, the expansion has a restricted domain of
validity in the $y,Q$ plane. In particular, the series expansion in the cross-
ratio $x$ of the four-point block on the plane converges for $|x|<1$
Zamolodchikov1987 , therefore we expect the series expansion of the two-point
block on the torus to have a convergence radius of $|y|=1$ (the torus block
reduces to the block on the plane in the limit $Q\to 0$). The numerics confirm
this expectation. For $|y|>1$, we observe large fluctuations in the value of
the conformal block as we include more terms in the series expansion. The
numerics for the convergence radius in $Q$ is less clear, but also in this
variable we observe large fluctuations close to $Q=e^{-2\pi}$. Thus, we can
check the vacuum block dominance only in a restricted region around the origin
in $y$, corresponding to small intervals. The restricted convergence radius in
$Q$ is not limiting since above the Hawking-Page transition temperature given
by $Q=e^{-2\pi}$, we expect the block in the high temperature expansion to
dominate. The high temperature expansion of the conformal block is given by a
series expansion in $e^{2\pi iz_{2}/\tau}-1$ and $e^{-2\pi i/\tau}$. In the
$h_{p,q}/c\gg h_{1,2}/c$ limit, the series coefficients are equal to those of
the low temperature expansion. Moreover, in the same limit at high
temperatures and for $n\to 1$ the conformal block in the large interval limit
(where the monodromy condition is taken around a time circle of size $n\tau$)
is equivalent to the conformal block in the small interval limit with the
replacement $z_{2}\to 1-z_{2}$.
We show some plots of $\frac{\partial f_{\text{cl.}}}{\partial h_{p}}$ and
$\frac{\partial f_{\text{cl.}}}{\partial h_{q}}$ in fig. 4 for small
temperatures and values of $y$ and $Q$ inside the convergence radius.
Figure 4: Derivative of the semiclassical conformal block $f_{\text{cl.}}$
w.r.t. $h_{p}$ (left) and $h_{q}$ (right) for $z_{2}=0.1$, $\beta=4\pi$ and
different values of $h_{p}$ and $h_{q}$ in the range $[0,c/24]$. Note that the
plotted value is greater than zero in all cases, showing that the
semiclassical conformal block increases with $h_{p}$ and $h_{q}$. The series
expansion used in this figure was truncated at order 10 in both $y$ and $Q$.
The plots for the conformal block in the high temperature limit show no
significant differences from the ones in the low temperature limit. We find in
all cases that inside the convergence radius of the series expansion
$\frac{\partial f_{\text{cl.}}}{\partial h_{p,q}}>0$, i.e. the assumption of
vacuum block dominance is fulfilled
While it is not possible to find an analytic continuation for the conformal
block from a truncated series expansion, we can use a Padé approximant to get
a heuristic approximation of the series outside its convergence radius. The
Padé approximation works by replacing the truncated series expansion by a
rational function whose Taylor expansion agrees with the series expansion up
to the order in which the truncation was performed BakerGraves-Morris . In
many cases, this approximation has a better radius of convergence than the
original series expansion due to poles in the function limiting the radius of
convergence of its Taylor expansion being taken into account in the Padé
approximation. We plot the Padé approximant of $f_{\text{cl.}}$ in the special
case $h_{p}=h_{q}$ for different orders of the denominator polynomial in fig.
5 depending on $z_{2}$ (the size of the entangling interval).
Figure 5: Derivative of the Padé approximation of the semiclassical conformal
block w.r.t. the internal conformal weight plotted against $z_{2}$. The upper
plot shows the conformal block at $\beta=4\pi$ in the low temperature
expansion and the lower plot the block at $\beta=\pi$ in the high temperature
expansion. For plotting convenience we set $h_{p}=h_{q}$ which gives a
derivative of $f_{\text{cl.}}$ that is constant for all $h_{p}$. As we
increase the order of the denominator polynomial in the Padé approximation,
the fluctuations in $f_{\text{cl.}}$ outside of the convergence radius of the
series expansion decrease. Moreover, multiple Padé approximants converge to
the same value outside of the convergence radius. The convergence radius
$|y|=1$ is reached at $z_{2}=1/6$ in the upper plot and
$z_{2}=\frac{\log(2)}{4\pi}=0.0552$ in the lower plot. The series expansion
approximated by the Padé approximants in this figure was truncated at order 10
in both $y$ and $Q$.
We find that different Padé approximants for $f_{\text{cl.}}$ converge to the
same value and yield $\frac{\partial f_{\text{cl.}}}{\partial h_{p,q}}>0$ in a
finite region outside of the convergence radius $|y|=1$ of the series
expansion of $f_{\text{cl.}}$. While this is not a formal proof, it does
indicate that vacuum block dominance holds also outside of the convergence
radius.
## 4 Discussion
Let us briefly summarize the main points investigated in this publication. In
the first part in sec. 2, we took a detailed look at monodromy methods for the
computation of arbitrary conformal blocks on the torus as well as zero-point
conformal blocks on the special higher genus Riemann surface relevant to the
calculation of the entanglement entropy via the replica trick. We then applied
the derived monodromy methods to the calculation of the entanglement entropy
in sec. 3. We found that for holographic CFTs, the vacuum conformal block
dominates the partition function on the replica surface and therefore the
entanglement entropy takes on a universal form in agreement with the RT
formula.
The main advancement compared to previous work on entanglement entropy for
finite size and finite temperature systems at large central charge is as
follows.
First of all, we clarified the role of the correlation function of twist
operators on the torus compared to the partition function $Z_{n}$ on the
replica surface ${\mathcal{R}}_{n}$. The twist correlator only gives the
correct results for small intervals or low temperature, while the replica
partition function $Z_{n}$ must give the entanglement entropy in all cases by
construction. The reason why the twist correlator agrees with $Z_{n}$ for
small intervals or low temperature is evident from the monodromy method. The
conformal blocks of the twist correlator obey the same monodromy method as the
zero-point conformal blocks for $Z_{n}$ apart from the set of allowed
monodromy conditions, which for the twist correlator blocks is a subset of
those of the replica partition function.
Secondly, our derivation of the monodromy method is based solely on CFT
techniques and thus provides a non-trivial check of both the RT formula as
well as previous calculations of Barrella:2013wja ; Chen:2015kua using the
monodromy method on the gravity side. Moreover, the monodromy method presented
here is derived from first principles, further strengthening the heuristic
arguments that were used to justify the monodromy method in Barrella:2013wja ;
Chen:2015kua . The derivation from the CFT side furthermore allows for a
determination of the restrictions on the operator content and OPE coefficients
that must be obeyed by the CFT in question in order for the vacuum block to
give the dominant contribution and the RT results for the entanglement entropy
to be valid. Perhaps unsurprisingly, these restrictions are equivalent to
those imposed in Hartman:2013mia for the entanglement entropy at zero
temperature. That is, the entanglement entropy for any CFT with large central
charge, sparse spectrum of light operators and at most exponentially growing
OPE coefficients agrees with the RT formula not only at zero temperature as
was already known from the results of Hartman:2013mia , but also at finite
temperature and finite size. In addition, the same arguments that imply the
universal form of the entanglement entropy in these CFTs also imply a
universal form of the partition function at all temperatures, another feature
of holographic CFT investigated also in Hartman:2014oaa .
We close with a short outlook on possible future directions. One interesting
direction is an investigation in full generality of the issue under which
conditions the twist correlator yields the same results as the replica
partition function $Z_{n}$. Previous work in the free fermion case
Lokhande:2015zma ; Mukhi:2017rex indicates that for short intervals the twist
correlator is equivalent to the replica partition function, a result which we
can corroborate for large $c$ CFTs. Does this hold in general? Other
directions can be found in further applications of the monodromy method. Since
conformal blocks are the basic building blocks of any correlation function,
the derivation of the monodromy methods from first principles lays the
groundwork for a number of applications. In particular, the monodromy methods
found in this paper may prove useful in deriving recursion relations in the
dimensions of the exchanged operators for torus conformal blocks with better
convergence properties than the known recursion relations in the central
charge found in Cho:2017oxl . Moreover, the monodromy methods also pave the
way to the study of other information theoretic quantities closely connected
to the entanglement entropy such as entanglement negativity Calabrese:2012nk ;
Calabrese:2012ew ; Calabrese:2014yza , symmetry resolved entanglement
Goldstein:2017bua ; Bonsignori:2019naz ; Murciano:2020lqq ; Capizzi:2020 ;
Zhao:2020qmn or entwinement Balasubramanian:2014sra ; Lin:2016fqk ;
Balasubramanian:2016xho ; Balasubramanian:2018ajb ; Erdmenger:2019lzr for
finite temperature states, providing opportunities for new insights into the
physics of the dual AdS black holes. We hope to be able to report on further
results in this direction in the future.
##### Acknowledgements
I would like to thank Johanna Erdmenger and Christian Northe for useful
discussions. I acknowledge financial support by the Deutsche
Forschungsgemeinschaft (DFG, German Research Foundation) under Germany’s
Excellence Strategy through Würzburg-Dresden Cluster of Excellence on
Complexity and Topology in Quantum Matter - ct.qmat (EXC 2147, project-id
390858490).
## Appendix A Conventions for elliptic functions
This appendix contains an overview over conventions and useful identities for
the Weierstraß elliptic functions used in the rest of the publication (see
e.g. NIST:DLMF for more details about these functions). All elliptic
functions are defined for a lattice $\Lambda$ generated by the identifications
$z\sim z+1$ and $z\sim z+\tau$. The Weierstraß elliptic functions
$\wp(z),\zeta(z)$ and $\sigma(z)$ are defined by
$\displaystyle\wp(z)$
$\displaystyle=-\zeta^{\prime}(z)=\frac{1}{z^{2}}+\sum_{(m,n)\neq(0,0)}\left(\frac{1}{(z+n+m\tau)^{2}}-\frac{1}{(n+m\tau)^{2}}\right),$
(75) $\displaystyle\zeta(z)$
$\displaystyle=\frac{\sigma^{\prime}(z)}{\sigma(z)}=\frac{1}{z}+\sum_{(m,n)\neq(0,0)}\left(\frac{1}{(z+n+m\tau)}-\frac{1}{(n+m\tau)}+\frac{z}{(n+m\tau)^{2}}\right),$
$\displaystyle\sigma(z)$
$\displaystyle=z\prod_{(m,n)\neq(0,0)}\left(1-\frac{z}{n+m\tau}\right)\exp\left(\frac{z}{n+m\tau}+\frac{1}{2}\frac{z^{2}}{(n+m\tau)^{2}}\right).$
$\wp(z)$ is a true elliptic (i.e. doubly periodic) function while $\zeta(z)$
and $\sigma(z)$ are quasiperiodic:
$\displaystyle\wp(z+1)=\wp(z+\tau)=\wp(z)$ (76)
$\displaystyle\zeta(z+1)=\zeta(z)+2\eta_{1},$
$\displaystyle\zeta(z+\tau)=\zeta(z)+2\eta_{3}$
$\displaystyle\sigma(z+1)=-\exp(2\eta_{1}(z+1/2))\sigma(z),$
$\displaystyle\sigma(z+\tau)=-\exp(2\eta_{3}(z+\tau/2))\sigma(z),$
where $\eta_{1}=\zeta(1/2)$ and $\eta_{3}=\zeta(\tau/2)=\tau\eta_{1}-\pi i$.
Another useful definition of $\wp(z)$ and $\zeta(z)$ is given by
$\displaystyle\wp(z)+2\eta_{1}=(2\pi
i)^{2}\sum_{m=-\infty}^{\infty}\frac{Q^{m}u}{(u-Q^{m})^{2}}=\left(\frac{2\pi
i}{\tau}\right)^{2}\sum_{m=-\infty}^{\infty}\frac{\tilde{Q}^{m}\tilde{u}}{(\tilde{u}-\tilde{Q}^{m})^{2}}$
(77)
$\displaystyle\zeta(z)-2\eta_{1}z=i\pi\sum_{m=-\infty}^{\infty}\frac{Q^{m}+u}{Q^{m}-u}=-\frac{i\pi}{\tau}\sum_{m=-\infty}^{\infty}\frac{\tilde{Q}^{m}+\tilde{u}}{\tilde{Q}^{m}-\tilde{u}},$
where $u=e^{-2\pi iz},Q=e^{2\pi i\tau}$ and $\tilde{u}=e^{2\pi iz/\tau}$,
$\tilde{Q}=e^{-2\pi i/\tau}$.
## Appendix B Recursion relations for torus conformal blocks
For completeness, this appendix shows the recursion formulas for the two-point
conformal blocks on the torus following as a special case from the general
method derived in Cho:2017oxl . For details of the derivation, see Cho:2017oxl
and also Hadasz:2009db for the one-point torus block. For the OPE block, the
conformal block is given by the following recursion scheme
$\displaystyle{\mathcal{F}}^{\text{OPE}}_{21,pq}(h_{p},h_{q},c)$
$\displaystyle=U^{\text{OPE}}(h_{p},h_{q},c)$ (78) $\displaystyle-\sum_{r\geq
2,s\geq 1}$ $\displaystyle\frac{\partial c_{rs}(h_{q})}{\partial
h_{q}}Q^{rs}\frac{A^{c_{rs}(h_{q})}_{rs}{P^{rs}_{c_{rs}(h_{q})}\biggl{[}\begin{array}[]{c}h_{p}\\\
h_{q}+rs\end{array}\biggr{]}}{P^{rs}_{c_{rs}(h_{q})}\biggl{[}\begin{array}[]{c}h_{p}\\\
h_{q}\end{array}\biggr{]}}}{c-c_{rs}(h_{q})}{\mathcal{F}}^{\text{OPE}}_{21,pq}(h_{q},h_{q}+rs,c_{rs}(h_{q}))$
$\displaystyle-\sum_{r\geq 2,s\geq 1}$ $\displaystyle\frac{\partial
c_{rs}(h_{p})}{\partial
h_{p}}y^{rs}\frac{A^{c_{rs}(h_{p})}_{rs}{P^{rs}_{c_{rs}(h_{p})}\biggl{[}\begin{array}[]{c}h_{q}\\\
h_{q}\end{array}\biggr{]}}{P^{rs}_{c_{rs}(h_{p})}\biggl{[}\begin{array}[]{c}h_{2}\\\
h_{1}\end{array}\biggr{]}}}{c-c_{rs}(h_{p})}{\mathcal{F}}^{\text{OPE}}_{21,pq}(h_{p}+rs,h_{q},c_{rs}(h_{p}))$
where the fusion polynomials are given by
${P^{rs}_{c}\biggl{[}\begin{array}[]{c}h_{1}\\\
h_{2}\end{array}\biggr{]}}=\prod_{m=1-r,m\in
1-r+2{\mathbb{N}}}^{r-1}\prod_{n=1-s,n\in
1-s+2{\mathbb{N}}}^{s-1}\frac{\lambda_{1}+\lambda_{2}+mb+nb^{-1}}{2}\frac{\lambda_{1}-\lambda_{2}+mb+nb^{-1}}{2}$
(79)
with $\lambda_{i}=\sqrt{(b+b^{-1})^{2}-4h_{i}}$ while the prefactor is
$A^{c}_{rs}=\frac{1}{2}\prod_{\stackrel{{\scriptstyle(m,n)=(1-r,1-s)}}{{(m,n)\neq(0,0),(r,s)}}}^{(r,s)}(mb+nb^{-1})^{-1}.$
(80)
$c_{rs}$ denotes the value of the central charge where degenerate
representations of the Virasoro algebra appear,
$\displaystyle c_{rs}(h)=1+6(b_{rs}(h)+b_{rs}^{-1}(h))^{2},$ (81)
$\displaystyle
b_{rs}(h)^{2}=\frac{rs-1+2h+\sqrt{(r-s)^{2}+4(rs-1)h+4h^{2}}}{1-r^{2}}.$
The $c$-regular part $U$ is given by
$U^{\text{OPE}}(h_{p},h_{q},c)=\left[\prod_{n=2}^{\infty}\frac{1}{1-Q^{n}}\right]\sum_{i,j\geq
0}Q^{i}y^{j}\frac{s_{ij}(h_{q},h_{p},h_{q})(1-h_{p}-j)_{j}(h_{p}+h_{1}-h_{2})_{j}}{i!(2h_{q})_{i}j!(2h_{p})_{j}},$
(82)
where we define the rising and falling Pochhammer symbols by
$(a)_{n}=\prod_{k=0}^{n-1}(a+k)\hskip
56.9055pt(a)^{(n)}=\prod_{k=0}^{n-1}(a-k)$ (83)
and
$\displaystyle s_{ij}(h_{1},h_{2},h_{3})=\langle
h_{1}|(L_{-1}^{i})^{\dagger}{\mathcal{O}}_{h_{2}}(1)L_{-1}^{j}|h_{3}\rangle$
(84) $\displaystyle=\left\\{\begin{aligned}
&\sum_{p=0}^{i}\biggl{(}\begin{array}[]{c}i\\\
p\end{array}\biggr{)}(j)^{(p)}(2h_{3}+j-1)^{(p)}(h_{1}+h_{2}-h_{3}-(j-p))_{i-p}(h_{3}+h_{2}-h_{1})_{j-p}\leavevmode\nobreak\
,\leavevmode\nobreak\ j\geq i\\\
&\sum_{p=0}^{j}\biggl{(}\begin{array}[]{c}j\\\
p\end{array}\biggr{)}(i)^{(p)}(2h_{1}+i-1)^{(p)}(h_{3}+h_{2}-h_{1}-(i-p))_{j-p}(h_{1}+h_{2}-h_{3})_{i-p}\leavevmode\nobreak\
,\leavevmode\nobreak\ i\geq j\end{aligned}\right.$
The recursion formula for the projection block is given by 111111The authors
of Cho:2017oxl also claim to have found a different recursion formula in the
conformal dimensions of the exchanged primaries $h_{p,q}$ for a class of
conformal blocks termed the “necklace blocks” which include the projection
block considered here as a special case. We cannot confirm this claim since
the formula presented in Cho:2017oxl agrees neither with the recursion
relation in the central charge derived in Cho:2017oxl nor with an explicit
calculation in the first few orders of the series expansion in $q_{1},q_{2}$
for the projection block.
$\displaystyle{\mathcal{F}}^{\text{projection}}_{2p,1q}(h_{p},h_{q},c)=$
$\displaystyle U^{\text{projection}}(h_{p},h_{q},c)$ (85)
$\displaystyle-\sum_{r\geq 2,s\geq 1}$ $\displaystyle\frac{\partial
c_{rs}(h_{q})}{\partial
h_{q}}q_{1}^{rs}\frac{A^{c_{rs}(h_{q})}_{rs}{P^{rs}_{c_{rs}(h_{q})}\biggl{[}\begin{array}[]{c}h_{p}\\\
h_{1}\end{array}\biggr{]}}{P^{rs}_{c_{rs}(h_{q})}\biggl{[}\begin{array}[]{c}h_{p}\\\
h_{2}\end{array}\biggr{]}}}{c-c_{rs}(h_{q})}{\mathcal{F}}^{\text{projection}}_{2p,1q}(h_{p},h_{q}+rs,c_{rs}(h_{q}))$
$\displaystyle-\sum_{r\geq 2,s\geq 1}$ $\displaystyle\frac{\partial
c_{rs}(h_{p})}{\partial
h_{p}}q_{2}^{rs}\frac{A^{c_{rs}(h_{p})}_{rs}{P^{rs}_{c_{rs}(h_{p})}\biggl{[}\begin{array}[]{c}h_{q}\\\
h_{1}\end{array}\biggr{]}}{P^{rs}_{c_{rs}(h_{p})}\biggl{[}\begin{array}[]{c}h_{q}\\\
h_{2}\end{array}\biggr{]}}}{c-c_{rs}(h_{p})}{\mathcal{F}}^{\text{projection}}_{2p,1q}(h_{p}+rs,h_{q},c_{rs}(h_{p})),$
where the $c$-regular part is given by
$U^{\text{projection}}(h_{p},h_{q},c)=\left[\prod_{n=2}^{\infty}\frac{1}{1-Q^{n}}\right]\sum_{i,j\geq
0}q_{1}^{i}q_{2}^{j}\frac{s_{ij}(h_{q},h_{2},h_{q})s_{ji}(h_{p},h_{1},h_{q})}{i!(2h_{q})_{i}j!(2h_{p})_{j}}.$
(86)
By explicit calculation, it is easy to check in the first few orders of the
series expansion that for $h_{1}=h_{2}$ and $h_{p,q}=\gamma c$, the limits
$\lim_{\gamma\to 0}$ and $\lim_{c\to\infty}$ of the OPE block commute. We have
checked this up to fourth order in $y,Q$. A more convenient proof is possible
with a recursion relation in the conformal weights of the exchanged operators
$h_{p,q}$, as derived for the conformal block on the plane in
Zamolodchikov1987 . In fact, the singular parts proportional to $\sim
1/(h_{p,q}-h_{rs})$ of such a recursion relation are proportional to the
singular parts $\sim 1/(c-c_{rs}(h_{p,q}))$ of the above recursion relations
in $c$ Cho:2017oxl . Using these known singular parts, one can show that the
limits $\gamma\to 0$ and $c\to\infty$ commute for the singular parts of this
recursion relation to all orders, assuming the above conditions on
$h_{1,2,p,q}$. Together, these calculations provide some evidence that the
semiclassical limit is well-defined for the vacuum block on the torus derived
in sec. 3. Since the vacuum block on the torus in most limits is equivalent to
a zero-point block on the replica surface ${\mathcal{R}}_{n}$, this also
provides evidence that the semiclassical limit for the zero-point vacuum block
on ${\mathcal{R}}_{n}$ is well-defined.
## References
* (1) L. Bombelli, R.K. Koul, J. Lee and R.D. Sorkin, _A Quantum Source of Entropy for Black Holes_ , _Phys. Rev. D_ 34 (1986) 373.
* (2) M. Srednicki, _Entropy and area_ , _Phys. Rev. Lett._ 71 (1993) 666 [hep-th/9303048].
* (3) P. Calabrese and J.L. Cardy, _Entanglement entropy and quantum field theory_ , _J. Stat. Mech._ 0406 (2004) P06002 [hep-th/0405152].
* (4) T. Hartman, _Entanglement Entropy at Large Central Charge_ , 1303.6955.
* (5) T. Azeyanagi, T. Nishioka and T. Takayanagi, _Near Extremal Black Hole Entropy as Entanglement Entropy via AdS(2)/CFT(1)_ , _Phys. Rev. D_ 77 (2008) 064005 [0710.2956].
* (6) P. Calabrese and J. Cardy, _Entanglement entropy and conformal field theory_ , _J. Phys. A_ 42 (2009) 504005 [0905.4013].
* (7) J. Cardy and C.P. Herzog, _Universal Thermal Corrections to Single Interval Entanglement Entropy for Two Dimensional Conformal Field Theories_ , _Phys. Rev. Lett._ 112 (2014) 171603 [1403.0578].
* (8) B. Chen, J.-B. Wu and J.-j. Zhang, _Short interval expansion of Rényi entropy on torus_ , _JHEP_ 08 (2016) 130 [1606.05444].
* (9) S.F. Lokhande and S. Mukhi, _Modular invariance and entanglement entropy_ , _JHEP_ 06 (2015) 106 [1504.01921].
* (10) S. Mukhi, S. Murthy and J.-Q. Wu, _Entanglement, Replicas, and Thetas_ , _JHEP_ 01 (2018) 005 [1706.09426].
* (11) M. Headrick, A. Lawrence and M. Roberts, _Bose-Fermi duality and entanglement entropies_ , _J. Stat. Mech._ 1302 (2013) P02022 [1209.2428].
* (12) S. Datta and J.R. David, _Rényi entropies of free bosons on the torus and holography_ , _JHEP_ 04 (2014) 081 [1311.1218].
* (13) B. Chen and J.-q. Wu, _Rényi entropy of a free compact boson on a torus_ , _Phys. Rev. D_ 91 (2015) 105013 [1501.00373].
* (14) B. Chen and J.-q. Wu, _Large interval limit of Rényi entropy at high temperature_ , _Phys. Rev. D_ 92 (2015) 126002 [1412.0763].
* (15) A.B. Zamolodchikov, _Conformal symmetry in two-dimensional space: recursion representation of conformal block_ , _Teoreticheskaya i Matematicheskaya Fizika_ 73 (1987) 103.
* (16) T. Faulkner, _The Entanglement Renyi Entropies of Disjoint Intervals in AdS/CFT_ , 1303.7221.
* (17) P.G. Zograf and L.A. Takhtadzhyan, _ON UNIFORMIZATION OF RIEMANN SURFACES AND THE WEIL-PETERSSON METRIC ON TEICHMÜLLER AND SCHOTTKY SPACES_ , _Mathematics of the USSR-Sbornik_ 60 (1988) 297.
* (18) S. Ryu and T. Takayanagi, _Holographic derivation of entanglement entropy from AdS/CFT_ , _Phys. Rev. Lett._ 96 (2006) 181602 [hep-th/0603001].
* (19) B. Chen and J.-q. Wu, _Universal relation between thermal entropy and entanglement entropy in conformal field theories_ , _Phys. Rev. D_ 91 (2015) 086012 [1412.0761].
* (20) B. Chen and J.-q. Wu, _Single interval Renyi entropy at low temperature_ , _JHEP_ 08 (2014) 032 [1405.6254].
* (21) T. Barrella, X. Dong, S.A. Hartnoll and V.L. Martin, _Holographic entanglement beyond classical gravity_ , _JHEP_ 09 (2013) 109 [1306.4682].
* (22) B. Chen and J.-q. Wu, _Holographic calculation for large interval Rényi entropy at high temperature_ , _Phys. Rev. D_ 92 (2015) 106001 [1506.03206].
* (23) A.-K. Kashani-Poor and J. Troost, _The toroidal block and the genus expansion_ , _JHEP_ 03 (2013) 133 [1212.0722].
* (24) K.B. Alkalaev, R.V. Geiko and V.A. Rappoport, _Various semiclassical limits of torus conformal blocks_ , _JHEP_ 04 (2017) 070 [1612.05891].
* (25) K.B. Alkalaev and V.A. Belavin, _Holographic duals of large-c torus conformal blocks_ , _JHEP_ 10 (2017) 140 [1707.09311].
* (26) A. Zamolodchikov, _Two-dimensional Conformal Symmetry and Critical Four-spin Correlation Functions in the Ashkin-Teller Model_ , _Zh. Eksp. Teor. Fiz._ 90 (1986) 1808.
* (27) D. Harlow, J. Maltz and E. Witten, _Analytic Continuation of Liouville Theory_ , _JHEP_ 12 (2011) 071 [1108.4417].
* (28) M. Beşken, S. Datta and P. Kraus, _Semi-classical Virasoro blocks: proof of exponentiation_ , _JHEP_ 01 (2020) 109 [1910.04169].
* (29) T. Eguchi and H. Ooguri, _Conformal and Current Algebras on General Riemann Surface_ , _Nucl. Phys. B_ 282 (1987) 308.
* (30) M. Cho, S. Collier and X. Yin, _Recursive Representations of Arbitrary Virasoro Conformal Blocks_ , _JHEP_ 04 (2019) 018 [1703.09805].
* (31) T. Hartman, C.A. Keller and B. Stoica, _Universal Spectrum of 2d Conformal Field Theory in the Large c Limit_ , _JHEP_ 09 (2014) 118 [1405.5137].
* (32) A. Belin, C.A. Keller and I.G. Zadeh, _Genus two partition functions and Rényi entropies of large c conformal field theories_ , _J. Phys. A_ 50 (2017) 435401 [1704.08250].
* (33) T. Hartman and J. Maldacena, _Time Evolution of Entanglement Entropy from Black Hole Interiors_ , _JHEP_ 05 (2013) 014 [1303.1080].
* (34) G.A. Baker and P.R. Graves-Morris, _Padé approximants_ , Encyclopedia of mathematics and its applications, Addison-Wesley, Reading, Mass. [u.a.] (1981).
* (35) P. Calabrese, J. Cardy and E. Tonni, _Entanglement negativity in extended systems: A field theoretical approach_ , _J. Stat. Mech._ 1302 (2013) P02008 [1210.5359].
* (36) P. Calabrese, J. Cardy and E. Tonni, _Entanglement negativity in quantum field theory_ , _Phys. Rev. Lett._ 109 (2012) 130502 [1206.3092].
* (37) P. Calabrese, J. Cardy and E. Tonni, _Finite temperature entanglement negativity in conformal field theory_ , _J. Phys. A_ 48 (2015) 015006 [1408.3043].
* (38) M. Goldstein and E. Sela, _Symmetry-resolved entanglement in many-body systems_ , _Phys. Rev. Lett._ 120 (2018) 200602 [1711.09418].
* (39) R. Bonsignori, P. Ruggiero and P. Calabrese, _Symmetry resolved entanglement in free fermionic systems_ , _J. Phys. A_ 52 (2019) 475302 [1907.02084].
* (40) S. Murciano, P. Ruggiero and P. Calabrese, _Symmetry resolved entanglement in two-dimensional systems via dimensional reduction_ , _J. Stat. Mech._ 2008 (2020) 083102 [2003.11453].
* (41) L. Capizzi, P. Ruggiero and P. Calabrese, _Symmetry resolved entanglement entropy of excited states in a CFT_ , _Journal of Statistical Mechanics: Theory and Experiment_ 2020 (2020) 073101 [2003.04670].
* (42) S. Zhao, C. Northe and R. Meyer, _Symmetry-Resolved Entanglement in AdS 3/CFT2 coupled to $U(1)$ Chern-Simons Theory_, 2012.11274.
* (43) V. Balasubramanian, B.D. Chowdhury, B. Czech and J. de Boer, _Entwinement and the emergence of spacetime_ , _JHEP_ 01 (2015) 048 [1406.5859].
* (44) J. Lin, _A Toy Model of Entwinement_ , 1608.02040.
* (45) V. Balasubramanian, A. Bernamonti, B. Craps, T. De Jonckheere and F. Galli, _Entwinement in discretely gauged theories_ , _JHEP_ 12 (2016) 094 [1609.03991].
* (46) V. Balasubramanian, B. Craps, T. De Jonckheere and G. Sárosi, _Entanglement versus entwinement in symmetric product orbifolds_ , _JHEP_ 01 (2019) 190 [1806.02871].
* (47) J. Erdmenger and M. Gerbershagen, _Entwinement as a possible alternative to complexity_ , _JHEP_ 03 (2020) 082 [1910.05352].
* (48) “NIST Digital Library of Mathematical Functions.” http://dlmf.nist.gov/.
* (49) L. Hadasz, Z. Jaskolski and P. Suchanek, _Recursive representation of the torus 1-point conformal block_ , _JHEP_ 01 (2010) 063 [0911.2353].
|
# Green function estimates on complements of low-dimensional uniformly
rectifiable sets
Guy David Guy David. Université Paris-Saclay, CNRS, Laboratoire de
mathématiques d’Orsay, 91405, Orsay, France guy.david@universite-paris-
saclay.fr , Joseph Feneuil Joseph Feneuil. Université Paris-Saclay, CNRS,
Laboratoire de mathématiques d’Orsay, 91405, Orsay, France
<EMAIL_ADDRESS>and Svitlana Mayboroda Svitlana Mayboroda. School
of Mathematics, University of Minnesota, Minneapolis, MN 55455, USA
<EMAIL_ADDRESS>
###### Abstract.
It has been recently established in [DM2] that on uniformly rectifiable sets
the Green function is almost affine in the weak sense, and moreover, in some
scenarios such Green function estimates are equivalent to the uniform
rectifiability of a set. The present paper tackles a strong analogue of these
results, starting with the “flagship” degenerate operators on sets with lower
dimensional boundaries.
We consider the elliptic operators
$L_{\beta,\gamma}=-\operatorname{div}D^{d+1+\gamma-n}\nabla$ associated to a
domain $\Omega\subset\mathbb{R}^{n}$ with a uniformly rectifiable boundary
$\Gamma$ of dimension $d<n-1$, the now usual distance to the boundary
$D=D_{\beta}$ given by
$D_{\beta}(X)^{-\beta}=\int_{\Gamma}|X-y|^{-d-\beta}d\sigma(y)$ for
$X\in\Omega$, where $\beta>0$ and $\gamma\in(-1,1)$. In this paper we show
that the Green function $G$ for $L_{\beta,\gamma}$, with pole at infinity, is
well approximated by multiples of $D^{1-\gamma}$, in the sense that the
function
$\big{|}D\nabla\big{(}\ln\big{(}\frac{G}{D^{1-\gamma}}\big{)}\big{)}\big{|}^{2}$
satisfies a Carleson measure estimate on $\Omega$. We underline that the
strong and the weak results are different in nature and, of course, at the
level of the proofs: the latter extensively used compactness arguments, while
the present paper relies on some intricate integration by parts and the
properties of the “magical” distance function from [DEM].
Résumé. Dans [DM2] il est démontré que pour les domaines à bord uniformément
rectifiable, la fonction de Green vérifie des estimations faibles de bonne
approximation par des fonctions affines, avec une réciproque vraie dans
certains cas encourageants. Ici on part de la rectifiabilité uniforme et on
démontre les estimations fortes naturelles d’approximation de la fonction de
Green, et aussi des solutions, par des applications affines (ou, de manière
équivalente, des multiples de la distance au bord adoucie). L’étude inclut les
analogues naturels du Laplacien dans les domaine dont la frontière est de
grande co-dimension.
On considère les opérateurs elliptiques
$L_{\beta,\gamma}=\operatorname{div}D^{d+1+\gamma-n}\nabla$ associés à un
domaine $\Omega\subset\mathbb{R}^{n}$ dont le bord $\Gamma$ est Ahlfors
régulier et uniformément rectifiable de dimension $d<n-1$ et à la distance au
bord maintenant usuelle $D=D_{\beta}$ définie par
$D_{\beta}(X)^{-\beta}=\int_{\Gamma}|X-y|^{-d-\beta}d\sigma(y)$ pour
$X\in\Omega$, où $\beta>0$ et $\gamma\in(-1,1)$ sont des paramètres et
$\sigma$ une mesure Ahlfors régulière sur $\Gamma$. Les auteurs ont montré
précédemment que la mesure elliptique associée à $L_{\beta,\gamma}$ est bien
définie et est mutuellement absolument continue par rapport à $\sigma$, avec
un poids de $A_{\infty}$. Ici on démontre que la fonction de Green $G$ avec
pôle à l’infini associée à $L_{\beta,\gamma}$ est bien approchée par les
multiples de $D$, au sens où la fonction
$\big{|}D\nabla\big{(}\ln\big{(}\frac{G}{D^{1-\gamma}}\big{)}\big{)}\big{|}^{2}$
vérifie une condition de Carleson sur $\Omega$. Ces nouvelles estimations sont
différentes en nature. Les estimations de [DM2] reposaient sur des arguments
de compacité; ici on a besoin d’estimations plus précises, obtenues par
intégration par parties et en utilisant les propriétés algébriques de la
fonction $D_{\alpha}$ dans le cas“magique” de [DEM].
David was partially supported by the European Community H2020 grant GHAIA
777822, and the Simons Foundation grant 601941, GD. Mayboroda was supported in
part by the NSF grant DMS 1839077 and the Simons foundation grant 563916, SM
Key words/Mots clés. Uniformly rectifiable sets, degenerate elliptic
operators, Estimates on Green functions, harmonic measure in higher
codimension.
AMS classification: 42B37, 31B25, 35J25, 35J70.
###### Contents
1. 1 Introduction
2. 2 Definitions and anterior results
3. 3 Proof of Theorem 2.21
## 1\. Introduction
Rectifiable sets are an important notion in geometric measure theory and the
calculus of variation, in particular because the sets that minimize an energy
often enter this category. In the past decades, many mathematicians worked on
finding characterizations of rectifiability by properties apparently unrelated
to geometric measure theory. In the early 90’s, the quantifiable version -
uniform rectifiability - was introduced in [DS1, DS2] along with many
characterizations in terms of geometry (such as big pieces of Lipschitz
images, or using Peter Jones’ $\beta$ numbers) and in terms of singular
integrals. Later, it was observed that uniformly rectifiable sets may be the
right extension of Lipschitz graphs for elliptic boundary value problems, that
is, if $\Omega\subset\mathbb{R}^{n}$ is an open set with uniformly rectifiable
boundary and $\Omega$ provides enough access its the boundary, then we can
control the oscillations of the harmonic functions in $\Omega$. It is even
more noteworthy that a criterion for rectifiability can be obtained using
harmonic functions. Indeed, Hofmann, Martell, and Uriarte-Tuero proved in
[HM1, HMU] that under some conditions regarding the access to the boundary,
$\partial\Omega$ is uniformly rectifiable if and only if the harmonic measure
on $\partial\Omega$ is absolutely continuous in a quantitative way - called
$A_{\infty}$ \- with respect to the surface measure (see also [AHMNT]). The
optimal topological conditions in this regard have been identified in [AHMMT].
These were accompanied but a rich array of beautiful and difficult alternative
characterizations, exploring Carleson estimates for the solutions, behavior of
the singular integral operators, extensions to more general elliptic
operators, to mention just a few. Here we do not aim to provide a survey of
the related literature; the reader can consult, e.g., [DFM4] for a more
detailed presentation of the literature.
A weakness of the above theory is the fact that the harmonic measure on
$\Gamma$ only makes sense when $\Gamma\subset\mathbb{R}^{n}$ is of dimension
$d>n-2$, because lower dimensional sets have probability zero to be hit by a
Brownian motion. As a consequence, rectifiability, a notion that exists for
all integer dimensions, can only be characterized by means of the harmonic
measure for sets $\Gamma\subset\mathbb{R}^{n}$ of dimension $d=n-1$. To
overcome this obstacle, the authors of the present article developed a theory
of degenerate elliptic operators. The idea was to define a ‘harmonic measure’
on a set $\Gamma$ of dimension $d<n-1$ by replacing the Laplacian by an
operator on $\Omega:=\mathbb{R}^{n}\setminus\Gamma$ in the form
$L=-\operatorname{div}w(x)\nabla$, where $w(x)$ goes to infinity at an
appropriate rate when $x$ is approaching $\Gamma$, so that the corresponding
‘Brownian motion’ is attracted by $\Gamma$ and hits the boundary with
probability one. The articles [DFM2, DFM5] set the elliptic theory for this.
In particular an appropriate elliptic measure is constructed for a large class
of sets of any dimension $d<n$ (or even mixed dimensions) and operators $L$ as
above, with the usual nice properties such as the non-degeneracy and doubling
properties for the harmonic measure, the Harnack inequality and the comparison
principle for solutions, and estimates for the change of poles.
Then we tested the relevancy of our new elliptic measure. Dahlberg proved in
[Da] that the classical harmonic measure is absolutely continuous with respect
to the surface measure, and even $A_{\infty}$, whenever the boundary is
Lipschitz. In [DFM3] we assumed that $\Gamma$ is the graph of a Lipschitz
function $\varphi:\mathbb{R}^{d}\to\mathbb{R}^{n-d}$ with small Lipschitz
constant, and proved the same $A_{\infty}$ property for correctly chosen
operators. Yet we had to be careful about the operator we picked because as
showed in [DFM4], based on counterexamples from [CFK, MM], not every operator
$L$ will work. We wanted an explicit operator, which could be constructed by a
single method for any set of dimension $d$. We set our choice on
(1.1) $L_{\beta,\gamma}:=-\operatorname{div}(D_{\beta})^{d+1+\gamma-n}\nabla,$
where $\beta>0$, $\gamma\in(-1,1)$, and $D_{\beta}$ is defined on $\Omega$ as
(1.2)
$D_{\beta}(X):=\left(\int_{\Gamma}|X-y|^{-d-\beta}d\sigma(y)\right)^{-1/\beta}.$
The quantity $D_{\beta}$ is equivalent to the distance to the boundary, i.e.,
there exists $C>0$ such that
(1.3) $C^{-1}\,\mathrm{dist}(X,\Gamma)\leq D_{\beta}\leq
C\,\mathrm{dist}(X,\Gamma)\qquad\text{ for }X\in\Omega,$
but the advantage of $D_{\beta}$ over $\,\mathrm{dist}(.,\Gamma)$ is to always
be smooth in a certain quantitative way.
When $\gamma=0$, the level of the degeneracy $D_{\beta}^{d+1-n}$ in the
coefficients of the operator
$L_{\beta,0}:=-\operatorname{div}(D_{\beta})^{d+1-n}\nabla$ makes it a perfect
analogue of the Laplacian for the sets with a $d$-dimensional boundary when
$d<n-1$. In particular, we proved that the harmonic measure associated to
$L_{\beta,\gamma}$ is $A_{\infty}$ with respect to the $d$-dimensional
Hausdorff measure. Moreover, recently, in two distinct papers [DM1, Fen1] we
extended this result to the more general case where
$\Gamma\subset\mathbb{R}^{n}$ is uniformly rectifiable. Studying these
operators and the dimension of the support of the corresponding elliptic
measure, we were naturally drawn to introducing a parameter $\gamma$ which
seemingly unbalances the situation. Besides, for $d=n-1$, the operator
$L_{\beta,\gamma}:=-\operatorname{div}(D_{\beta})^{\gamma}\nabla$ is the
celebrated Caffarelli-Silvester extension of the fractional Laplacian operator
(cf. [CS]). However, we were surprised to realize that the argument in [DFM3]
extends to all $\gamma\in(-1,1)$ rather simply, and the generalization to
uniformly rectifiable set is stated for any $\gamma\in(-1,1)$ in [Fen1]. The
proof in [DM1] relies on geometric arguments, such as corona decompositions
and the construction of sawtooth domains, and an extrapolation argument, which
allow one to reduce to the case of small Lipschitz graphs. The proof of [Fen1]
is substantially simpler and more direct, and is based on a trick unique to
the case where the dimension of $\Gamma$ is at most $n-2$.
###### Theorem 1.4 ([DM1, Fen1]).
Let $\Gamma\subset\mathbb{R}^{n}$ be a $d$-Ahlfors regular uniformly
rectifiable set with $d<n-1$, and let $\sigma$ be an Ahlfors regular measure
that satisfies (2.1). Take $\beta>0$, $\gamma\in(-1,1)$, define
$L_{\beta,\gamma}$ as in (1.1), and construct the associated harmonic measure
$\omega_{\beta,\gamma}^{X}$ as in Definition 2.20. Then
$\omega_{\beta,\gamma}^{X}$ is $A_{\infty}$-absolutely continuous with respect
to $\sigma$. This means that for every choice of $\epsilon\in(0,1)$, there
exists $\delta\in(0,1)$, that depends only on $C_{\sigma}$, $C_{0}$,
$\epsilon$, $n$, $d$, $\beta$, and $\gamma$, such that for each choice of
$x\in\Gamma$, $r>0$, a Borel set $E\subset B(x,r)\cap\Gamma$, and a corkscrew
point $X=A_{x,r}$ as in (2.3),
(1.5)
$\frac{\sigma(E)}{\sigma(B(x,r)\cap\Gamma)}<\delta\Rightarrow\frac{\omega^{X}_{\Omega,L}(E)}{\omega^{X}_{\Omega,L}(B(x,r)\cap\Gamma)}<\epsilon.$
It is known that in the present context where all our measures are doubling,
the $A_{\infty}$ condition also implies that under the assumptions of (1.5),
(1.6)
$\frac{\omega^{X}_{\Omega,L}(E)}{\omega^{X}_{\Omega,L}(B(x,r)\cap\Gamma)}<\delta\Rightarrow\frac{\sigma(E)}{\sigma(B(x,r)\cap\Gamma)}<\epsilon.$
Observe that in the previous theorem, contrary to the case of co-dimension 1,
i.e. when $d=n-1$, we don’t assume any topological condition. And that is
perfectly natural, since the domain $\Omega:=\mathbb{R}^{n}\setminus\Gamma$
has a lot of paths and ample access to the boundary (see Lemmas 2.1 and 11.6
in [DFM2]). The next big objective would be to prove the reverse implication,
meaning that if the harmonic measure on $\Gamma$ is $A_{\infty}$ with respect
to the Hausdorff measure, then $\Gamma$ is uniformly rectifiable.
Unfortunately (and surprisingly) this fails brutally when $d+\beta=n+2$ (see
[DEM]), and although we expect this to be the only exception, the methods used
to prove the converse in co-dimension 1 do not appear to be adaptable to the
higher codimension case.
The purpose of the present paper is to provide different estimates on the
harmonic functions, which we hope will ultimately furnish one side of the
desired criterion. Indeed, as established in [DM2], some weak bounds on the
Green function are equivalent to the uniform rectifiability even in lower
dimensional settings. The general idea is that instead of trying to
characterize rectifiable sets using the harmonic measure, we would do so using
the property that the Green function behaves like a distance to the boundary.
This is not a surprising idea, as the Green functions and harmonic measure are
deeply connected, and are both prominent in the analysis of the free boundary
problems, and the proof of the properties of the harmonic measure in [DFM2,
DFM5] heavily relies on a comparison between the harmonic measure and Green
functions. Yet, the results we are about to prove are not known in the
“classical” setting of domains with an $n-1$ dimensional boundary, except for
a few simple situations and certainly not in the generality of the uniformly
rectifiable sets which we attack in this paper.
We will use the Green function with a pole at infinity, which is constructed
in [DEM, Definition 6.2, Lemma 6.5] with the following properties.
###### Proposition 1.7 ([DEM]).
Let $\Gamma\subset\mathbb{R}^{n}$ be a $d$-Ahlfors regular uniformly
rectifiable set with $d<n-1$. Take $\beta>0$, $\gamma\in(-1,1)$, and construct
$L_{\beta,\gamma}$ on $\Omega:=\mathbb{R}^{n}\setminus\Gamma$ as in (1.1).
There exists a continuous function $G^{\infty}=G_{\beta,\gamma}^{\infty}$ on
$\mathbb{R}^{n}$ such that
1. (i)
$G^{\infty}>0$ on $\Omega$,
2. (ii)
$G^{\infty}=0$ on $\Gamma$,
3. (iii)
there exists a positive Borel measure
$\omega_{\infty}=\omega^{\infty}_{\beta,\gamma}$ on $\Gamma$ such that
$\int_{\Omega}\nabla
G^{\infty}\cdot\nabla\varphi\,D_{\beta}^{d+1-n+\gamma}dx=\int_{\Gamma}\varphi(y)d\omega^{\infty}(y)\qquad\text{
for }\varphi\in C^{\infty}_{0}(\mathbb{R}^{n}).$
We call $G^{\infty}$ the Green function with pole at infinity, and
$G^{\infty}$ is unique up to multiplication by a scalar constant.
The Green function with pole at infinity for the Laplacian in
$\mathbb{R}^{n}_{+}=\\{(x,t)\in\mathbb{R}^{n-1}\times(0,+\infty)\\}$ is
$G^{\infty}(x,t)=t$. The one for the operator $-\operatorname{div}|t|^{d+1-n}$
in
$\mathbb{R}^{n}\setminus\mathbb{R}^{d}=\\{(x,t)\in\mathbb{R}^{n}\times\mathbb{R}^{n-d},\,t\neq
0\\}$ is $G^{\infty}(x,t)=|t|$. There is a third case where $G^{\infty}$ can
be computed, which is when $\Gamma$ is a $d$-Ahlfors regular set with $d<n-2$,
and we choose the specific operator $L_{n-d-2,0}$; then one can easily compute
that $G^{\infty}=D_{n-d-2}$. In [DEM] this is called the magic case. The rough
idea of what follows is that, when $\gamma=0$, a good Green function at
infinity should behave a bit like the distance to the boundary $\Gamma$, or
like $D_{\beta}$ which is its smooth substitute. When $\gamma\neq 0$, similar
homogeneity considerations lead us to expect that in the good cases,
$G^{\infty}$ will behave like $D_{\beta}^{1-\gamma}$ instead. Here is our main
result.
###### Theorem 1.8.
Let $\Gamma\subset\mathbb{R}^{n}$ be a $d$-Ahlfors regular uniformly
rectifiable set with $d<n-1$, and let $\sigma$ be an Ahlfors regular measure
on $\Gamma$ that satisfies (2.1). For $\beta>0$ and $\gamma\in(-1,1)$, let
$L_{\beta,\gamma}$ be as in (1.1) and define $G^{\infty}_{\beta,\gamma}$ as in
Proposition 1.7. Then for any $\alpha>0$, there exists $C>0$ that depends only
on $C_{\sigma}$, $C_{0}$, $\alpha$, $\beta$, $\gamma$, $d$, $n$ such that for
any ball $B:=B(x,r)$ centered on $\Gamma$, one has
(1.9)
$\int_{B}\left|\nabla\ln\left(\frac{G^{\infty}_{\beta,\gamma}}{D_{\alpha}^{1-\gamma}}\right)\right|^{2}\,D_{\alpha}^{d+2-n}\,dX\leq
C\sigma(B).$
Remarks:
* •
Remember that $G^{\infty}_{\beta,\gamma}$ is only defined up to a constant.
Yet, the above theorem makes perfect sense, because the value of left-hand
side of (1.9) does not change when we replace $G^{\infty}_{\beta,\gamma}$ by
$KG^{\infty}_{\beta,\gamma}$, where $K$ is a positive constant.
* •
Theorem 1.8 is just the application of Theorem 2.21 to the Green function at
infinity: we shall generalize the estimate of Theorem 1.8 to a large class of
solutions that is interesting by itself. However, we decided to highlight the
above statement, which was our true purpose in our search for
characterizations of rectifiability.
* •
Here we decided to use the same measure $\sigma$ for the definitions of
$L_{\beta,\gamma}$ and $D_{\alpha}$, but a minor modification of the proof
would allow us to take two different Ahlfors regular measures $\sigma$ and
$\widetilde{\sigma}$ to define $L_{\beta,\gamma}$ and and $D_{\alpha}$.
* •
As we have pointed out above, these results are not known for $d=n-1$ in the
generality of the uniformly rectifiable sets. In the half-space one can see
somewhat similar estimates in [FKP] and the bounds on the second derivatives
of the Green function for a special class of operators were, in disguise,
obtained in [HMT]. However, it is not clear how to deduce from [HMT] a co-
dimension 1 analogue of (1.9), directly or using interpolation.
In the present paper we only show that good geometric properties (the uniform
rectifiability of $\Gamma$) imply precise approximation properties of
$G^{\infty}$ by $D_{\alpha}$, and the proof will rely heavily on the
$A_{\infty}$ property of the harmonic measure and some intermediate results in
[Fen1] concerning the uniform rectifiability of $\Gamma$. We do not address
the issue of the converse in this paper. Yet there are good reasons to believe
that it may be easier to prove than for the absolute continuity of the
harmonic measure. In a parallel paper [DM2], the authors study a less precise
(weaker) approximation property of the Green function, and show that in some
cases (but where $d>n-2$) it already implies the uniform rectifiability of
$\Gamma$, while in the case of the present paper it implies another strange
property of potentials defined on $\Gamma$. But we did not manage to prove
that this strange property is impossible to obtain when $d+\alpha\neq n+2$.
The reader should be aware that even though we think of the Green function
estimate (1.9) as a possible alternative to the $A_{\infty}$ absolute
continuity of the harmonic measure for the characterization of uniform
rectifiability, we already know that in the general context of elliptic
operator in the form $L=-\operatorname{div}A\nabla$, the $A_{\infty}$ absolute
continuity (1.5) - or (1.6) - is not equivalent to (1.9). A counterexample in
$\mathbb{R}^{n}_{+}$, that can be extended to any codimensions using the
construction (4.6) in [DFM4], is given in the next lines. Denote the running
point in $\mathbb{R}^{n}_{+}$ as $(x,r)\in\mathbb{R}^{n-1}\times(0,+\infty)$,
and observe that in this case $D_{\alpha}(x,r)=c_{\alpha}r$. We set
$b(r):=1/(2+\cos(r))$ and we construct $L_{b}:=-\operatorname{div}b(r)\nabla$.
By uniqueness of the Green function with pole at infinity (Proposition 1.7),
we have
$G^{\infty}(x,r)=\int_{0}^{r}\frac{ds}{b(s)}=2r+\sin(r).$
On one hand, $r\leq G^{\infty}(x,r)\leq 3r$ for all
$(x,r)\in\mathbb{R}^{n}_{+}$, and that is enough, by using a comparison
principle with the harmonic measure, to show that the harmonic measure
associated to $L$ is equivalent to the surface measure on $\mathbb{R}^{n-1}$
(see for instance Theorem 1.17 in [Fen2] for details), hence it is
$A_{\infty}$ absolutely continuous with respect to the surface measure. On the
other hand, we have
$\begin{split}\left|\nabla\ln\Big{(}\frac{G_{\infty}}{D_{\alpha}}\Big{)}\right|=\left|\frac{\nabla
G^{\infty}}{G^{\infty}}-\frac{\nabla
D_{\alpha}}{D_{\alpha}}\right|=\left|\frac{2+\cos(r)}{2r+\sin(r)}-\frac{1}{r}\right|\geq\frac{1}{3r}\left|\cos(r)-\frac{\sin(r)}{r}\right|.\end{split}$
That is, for any ball $B(x,R)\subset\mathbb{R}^{n}$ centered on the boundary
$\mathbb{R}^{n-1}$, we have
$\begin{split}\int_{B(x,R)}\left|\nabla\ln\Big{(}\frac{G_{\infty}}{D_{\alpha}}\Big{)}\right|^{2}D_{\alpha}\,dX\geq
c_{n}R^{n-1}\int_{0}^{R/2}\left|\cos(r)-\frac{\sin(r)}{r}\right|^{2}\frac{dr}{r}\geq
c^{\prime}_{n}R^{n-1}\ln(R),\end{split}$
when $R$ is large (e.g. $R\geq 100$), and where $c_{n},c^{\prime}_{n}$ are
constants that depends only on $n$. As a consequence, (1.9) fails for the
elliptic operator $L_{b}$. This highlights the difference between Green
function and harmonic measure estimates for general elliptic operators.
Coming back to the Green function, the reader might wonder about the
comparison with the results in [DM2]. It is not easy to describe the results
in [DM2] avoiding technicalities, but roughly speaking, we only proved there
the weak statement that the set where the Green function behaves like a
distance to the boundary is a Carleson-prevalent set. This carries the
structural information saying that there are a lot of points and a lot of
scales where the desired estimate is true, but it carries no norm control.
Respectively, the methods heavily rely on compactness arguments. In the
present paper the arguments and the results are completely different, aiming
at a strong norm control in the form of (1.9) and using a new idea that the
“magical” distance function identified in [DEM], being an explicit solution to
a certain PDE, can be effectively used in the integration-by-parts arguments
(cf. [Fen1]). A good comparison is a familiar to many experts integration by
parts with the weight $t$ which disappears when the Laplacian hits $t$,
although the details in our case are necessarily considerably more involved.
In the next section, we shall give the definitions that we skipped up to now
to lighten the introduction, such as the definition of Ahlfors regular and
uniformly rectifiable sets. We also introduce the results from [Fen1] which we
will rely upon. The remainder of the article will be devoted to the proof of
our main result.
We shall use the notation $A(x)\lesssim B(x)$ when $A(x)\leq CB(x)$ and $C$ is
a constant whose dependence into the various parameters will be either
recalled or obvious from context. We also write $A(x)\approx B(x)$ when
$A(x)\lesssim B(x)$ and $A(x)\gtrsim B(x)$.
## 2\. Definitions and anterior results
For the rest of the article, we take $\Gamma\subset\mathbb{R}^{n}$ and
$\Omega:=\mathbb{R}^{n}\setminus\Gamma$. We assume that $\Gamma$ is a
$d$-Ahlfors regular set with $d<n-1$, that is $\Gamma$ is closed and there
exists a measure $\sigma$ supported on $\Gamma$ and a constant $C_{\sigma}\geq
1$ such that
(2.1) $C_{\sigma}^{-1}r^{d}\leq\sigma(B(x,r))\leq C_{\sigma}r^{d}\qquad\text{
for }x\in\Gamma,\ r>0.$
It is known that if the above property (2.1) is true for some measure
$\sigma$, then it is also true when $\sigma$ is replaced by
$\mathcal{H}^{d}|_{\Gamma}$ \- the $d$-dimensional Hausdorff measure
restricted to $\Gamma$.
The Ahlfors regularity of $\Gamma$ and the low dimension $d<n-1$ are
sufficient conditions to obtain the aforementioned equivalence (1.3). Indeed,
Lemma 5.1 in [DFM3] gives us that
(2.2) $C^{-1}\,\mathrm{dist}(X,\Gamma)\leq D_{\beta}\leq
C\,\mathrm{dist}(X,\Gamma)\qquad\text{ for }X\in\Omega,$
where the constant $C>0$ above depends only on $\beta>0$, $C_{\sigma}$ and
$n-d>1$.
Moreover, Lemma 11.6 in [DFM2] entails the existence of a constant $C$ that
depends only on $C_{\sigma}$ and $n-d>1$ such that for any $x\in\Gamma$ and
$r>0$, we can find a point $A_{x,r}$ such that
(2.3) $C^{-1}r\leq\,\mathrm{dist}(A_{x,r},\Gamma)\leq|A_{x,r}-x|\leq Cr.$
In other words, when $\Gamma$ is $d$-Ahlfors regular with $d<n-1$, its
complement automatically satisfies the interior corkscrew condition.
We shall also assume that $\Gamma$ is uniformly rectifiable. Equivalent
definitions of uniform rectifiability were given in [DS1, DS2], and the reader
may use their preferred one, but since we will only use the uniform
rectifiability of $\Gamma$ via results from [Fen1] that rely on the
summability properties of Tolsa’s $\alpha$-numbers, we will use these
properties as a definition of uniform rectifiability. We need some notation
first.
We denote by $\Xi$ the set of affine $d$-dimensional planes in
$\mathbb{R}^{n}$. Each plane $P\in\Xi$ is associated with a measure $\mu_{P}$,
which is the restriction to $P$ of the $d$-dimensional Hausdorff measure (i.e.
$\mu_{P}$ is the Lebesgue measure on the plane). A flat measure is a measure
$\mu$ that can be written $\mu=c\mu_{P}$ where $c$ is a positive constant and
$P\in\Xi$. The set of flat measure is called $\mathcal{F}$.
We need the following variant of Wasserstein distances to quantify the
difference between two measures, and then measure how far $\sigma$ is from
flat measures.
###### Definition 2.4.
For $x\in\mathbb{R}^{n}$ and $r>0$, denote by $Lip(x,r)$ the set of
$1$-Lipschitz functions $f$ supported in $\overline{B(x,r)}$, that is the set
of functions $f:\mathbb{R}^{n}\to\mathbb{R}$ such that $f(y)=0$ for
$y\in\mathbb{R}^{n}\setminus B(x,r)$ and $|f(y)-f(z)|\leq|y-z|$ for
$y,z\in\mathbb{R}^{n}$. The normalized Wasserstein distance in B(x,r) between
two measures $\sigma$ and $\mu$ is
(2.5) $\,\mathrm{dist}_{x,r}(\mu,\sigma)=r^{-d-1}\sup_{f\in
Lip(x,r)}\Big{|}\int fd\sigma-\int fd\mu\Big{|}.$
The distance to flat measures is then defined by
(2.6)
$\alpha_{\sigma}(x,r)=\inf_{\mu\in{\mathcal{F}}}\,\mathrm{dist}_{x,r}(\mu,\sigma).$
One can easily check that when (2.1) holds, the quantity $\alpha_{\sigma}$ is
uniformly bounded, i.e. there exists a constant $C$ that depends only on $d$,
$n$, and $C_{\sigma}$ such that $\alpha_{\sigma}(x,r)\leq C$ for $x\in\Gamma$
and $r>0$.
Let $\Gamma$ be a $d$-Ahlfors regular set, and $\sigma$ a measure that
satisfies (2.1). Tolsa’s characterization of uniform rectifiability, Theorem
1.2 in [Tol], is as follows111Tolsa’s characterization of rectifiability in
[Tol] is given with dyadic cubes but one can easily check that our bound (2.7)
is equivalent to Tolsa’s one.: $\Gamma$ is uniformly rectifiable if and only
if there exists a constant $C_{0}>0$ such that
(2.7) $\int_{0}^{r}\int_{\Gamma\cap
B(x,r)}|\alpha_{\sigma}(y,s)|^{2}\,d\sigma(y)\,\frac{ds}{s}\leq
C_{0}\sigma(B(x,r))\qquad\text{ for }x\in\Gamma\text{ and }r>0.$
Here we will only use the fact that (2.7) holds when $\Gamma$ is uniformly
rectifiable. That is, we will only use (2.7) and do not need to know other
properties of uniformly rectifiable sets. The property (2.7) will allow us to
obtain additional estimates on the smooth distance $D_{\beta}$. The
presentation of those bounds will be easier after the following definition.
###### Definition 2.8.
Let the function $f$ be defined on $\Omega$. We say that $f$ satisfies the
Carleson measure condition when $f\in L^{\infty}(\Omega)$ and
$|f(X)|^{2}\,\mathrm{dist}(X,\Gamma)^{d-n}dX$ is a Carleson measure, that is,
(2.9) $\int_{B(x,r)}|f(X)|^{2}\,\mathrm{dist}(X,\Gamma)^{d-n}dX\leq
C\sigma(B(x,r))$
for $x\in\Gamma$ and $r>0$, with a constant $C$ that does not depend on $x$ or
$r$.
Thus this is actually a quadratic Carleson condition. For short, we shall
write $f\in CM$, or $f\in CM(C)$ when we want to refer to the constant in
(2.9).
Due to (2.2), we can replace $\,\mathrm{dist}(X,\Gamma)^{d-n}$ with
$D_{\beta}^{d-n}(X)$, and even choose $\beta$ to fit our purposes; we shall
often do this without additional explanations. We shall rely strongly on Lemma
1.24 in [Fen1], which says the following.
###### Lemma 2.10.
Let $\Gamma$ be uniformly rectifiable, so that (2.1) and (2.7) hold. Let
$\beta>0$. Then there exist a scalar function $b$ and a vector function
$\mathcal{V}$, both defined on $\Omega$, such that
(2.11) $\int_{\Gamma}|X-y|^{-n}(X-y)d\sigma(y)=(b\nabla
D_{\beta}+\mathcal{V})D_{\beta}^{d+1-n}\qquad\text{ for }X\in\Omega$
and a constant $C_{1}$ that depends only on $C_{\sigma}$, $C_{0}$, $\beta$,
$n$, and $d$, such that
(2.12) $C_{1}^{-1}\leq b\leq C_{1},$ (2.13) $D_{\beta}\nabla b\in CM(C_{1}),$
(2.14) $|\mathcal{V}|\leq C_{1},$
and
(2.15) $\mathcal{V}\in CM(C_{1}).$
Observe that the left-hand side of (2.11) is divergence free. So if we use
(2.11) and we write the divergence free condition in weak terms, we obtain
(2.16) $\int_{\Omega}(b\nabla
D_{\beta}+\mathcal{V})\cdot\nabla\varphi\,D_{\beta}^{d+1-n}\,dX=0\qquad\text{
for }\varphi\in C^{\infty}_{0}(\Omega).$
We also need the following, which is Lemma 1.26 in [Fen1].
###### Lemma 2.17.
Let $\Gamma$ be uniformly rectifiable, i.e., assume that (2.1) and (2.7) hold.
Let $\alpha,\beta>0$. Then $D_{\alpha}\nabla[D_{\beta}/D_{\alpha}]$ satisfies
the Carleson measure condition with a constant that depends only on
$C_{\sigma}$, $C_{0}$, $\alpha$, $\beta$, $n$, and $d$.
We are now finished with the geometric background and turn to the elliptic
theory. Pick $\beta>0$ and $\gamma\in(-1,1)$. We whall use the operator
$L_{\beta,\gamma}$ constructed in (1.1); hence $L_{\beta,\gamma}$ enters the
scope of the theory developed in [DFM5], and in particular there is an
elliptic measure $\omega^{X}$ which we shall describe now.
We first need a Hilbert space $W_{\gamma}$, which is the same as in [DFM5],
Definition 3.1, but is more easily defined as
(2.18) $W_{\gamma}=\big{\\{}u\in
L^{1}_{loc}(\mathbb{R}^{n}),\|u\|_{\gamma}:=\int_{\Omega}|\nabla
u(X)|^{2}\,\mathrm{dist}(X,\Gamma)^{d+1+\gamma-n}dX<+\infty\big{\\}};$
the equivalence between the two definition is proved as in [DFM2, Lemma 3.3
and Lemma 5.21].
Each $f\in W_{\gamma}$ has a trace on $\Gamma$, which lies in a corresponding
Sobolev space $H_{\gamma}$ (which is equal to $H^{1/2}(\Gamma)$ when
$\gamma=0$); then we denote by $W_{\gamma,0}$ the set of functions in
$W_{\gamma}$ with zero trace; $W_{\gamma,0}$ is also the completion of
$C^{\infty}_{0}(\Omega)$ under the norm $\|.\|_{\gamma}$. And for any open set
$E\subset\mathbb{R}^{n}$, we write that $u\in W_{\gamma}(E)$ [respectively,
$u\in W_{\gamma,0}(E)$] if $u\varphi\in W_{\gamma}$ [respectively,
$u\varphi\in W_{\gamma,0}$] for any $\varphi\in C^{\infty}_{0}(E)$.
Then there is a notion of weak solution for $L_{\beta,\gamma}=0$, which the
reader may also find in [DFM5], such that in particular
(2.19) $\int_{\Omega}(\nabla
u\cdot\nabla\varphi)\,D_{\beta}^{d+1+\gamma-n}=0\qquad\text{ for }\varphi\in
C^{\infty}_{0}(\Omega).$
when $u$ is a weak solution for $L_{\beta,\gamma}=0$ (in $\Omega$). Here and
below, we remove the variable $X$ and the integration symbol $dX$ from the
notation when they are not entirely needed; unless otherwise specified, all
our integrals on $\Omega$ will be against the Lebesgue measure $dX$.
With all this notation, the main properties of our elliptic measures
$\omega^{X}$ are as follows.
###### Definition 2.20.
For each $X\in\Omega$, we can define a unique probability measure
$\omega^{X}:=\omega^{X}_{\beta,\gamma}$ on $\Gamma$ with the following
properties. For any $g\in C_{0}(\Gamma)$ (i.e., continuous function on
$\Gamma$ and compactly supported), the function $u_{g}$ defined as
$u_{g}(X)=\int_{\Gamma}g(y)d\omega^{X}(y)$
is a weak solution to $L_{\beta,\gamma}$ and, if in addition $g$ lies in the
Sobolev space $H_{\gamma}$, then $u_{g}\in W_{\gamma}$ and the trace of
$u_{g}$ is equal to $g$.
The space $H_{\gamma}\cap C_{0}(\Gamma)$ is dense in $C_{0}(\Gamma)$ (with the
sup norm), so the last condition is our way of solving a Dirichlet problem. We
are now ready to state the general version of our main theorem.
###### Theorem 2.21.
Let $\Gamma\subset\mathbb{R}^{n}$ be a $d$-Ahlfors regular uniformly
rectifiable set with $d<n-1$, and let $\sigma$ be an Ahlfors regular measure
on $\Gamma$ that satisfies (2.1). For $\beta>0$ and $\gamma\in(-1,1)$, define
$L_{\beta,\gamma}$ as in (1.1). Then for any $\alpha>0$, there exists $C>0$
that depends only on $C_{\sigma}$, $C_{0}$, $\alpha$, $\beta$, $\gamma$, $d$,
$n$, such that for any ball $B:=B(x,r)$ centered on $\Gamma$ and any non-
negative non identically zero weak solution $u$ of $L_{\beta,\gamma}u=0$ in
$\Omega\cap 3B$ which lies in $W_{\gamma,0}(3B)$, one has
(2.22)
$\int_{B}\left|\nabla\ln\left(\frac{u}{D_{\alpha}^{1-\gamma}}\right)\right|^{2}\,D_{\alpha}^{d+2-n}\leq
C\sigma(B).$
The proof of the Theorem will use the uniform rectifiability of $\Gamma$ via
Theorem 1.4 and Lemma 2.10. The Lemma will be used to estimate the left-hand
side of (2.22) via simple integration techniques, except for one more
complicated term in the form
$\int\left|\nabla\ln\left(\frac{u}{D_{\alpha}^{1-\gamma}}\right)\right||\nabla\phi_{B,\epsilon}|\,D_{\alpha}^{d+1-n},$
where $\phi_{B,\epsilon}$ is a well chosen cut off function, which is related
to the integral of the logarithm of the Poisson kernel and that will be
estimated using Theorem 1.4, that is, the $A_{\infty}$ absolute continuity of
the harmonic measure.
## 3\. Proof of Theorem 2.21
Let us recall
###### Lemma 3.1.
Let $\Gamma\subset\mathbb{R}^{n}$ be a $d$-Ahlfors regular uniformly
rectifiable set with $d<n-2$, and let $\sigma$ be an Ahlfors regular measure
on $\Gamma$ that satisfies (2.1). Define $L_{\beta,\gamma}$ as in (1.1). Then
there exist $C>0$ and $\theta\in(0,1]$, that depend only on $C_{\sigma}$,
$C_{0}$, $\beta$, $\gamma$, $n$ and $d$, such that for each choice of
$x\in\Gamma$, $r>0$, any Borel set $E\subset B(x,r)\cap\Gamma$, and any
corkscrew point $X=A_{x,r}$ as in (2.3), one has
(3.2)
$\frac{\omega_{\beta,\gamma}^{X}(E)}{\omega_{\beta,\gamma}^{X}(B(x,r))}\leq
C\left(\frac{\sigma(E)}{\sigma(B(x,r))}\right)^{\theta}$
and
(3.3) $\frac{\sigma(E)}{\sigma(B(x,r))}\leq
C\left(\frac{\omega_{\beta,\gamma}^{X}(E)}{\omega_{\beta,\gamma}^{X}(B(x,r))}\right)^{\theta}.$
Here and below, we assume implicitly that the constant $C$ in (2.3) (the
definition of corkscrew points) is chosen to depend on $C_{\sigma}$, $n$ and
$d$ only. If we allow a larger $C$ in (2.3), the constants in (3.2) and (3.3)
depend on $C$ as well.
Proof. The conditions (3.2) and (3.3) are another characterizations of the
fact that $\omega^{X}\in A_{\infty}(\sigma)$, which is true by Theorem 1.4.
The fact that (3.2)–(3.3) is implied by (1.6) can be found in [Ken, Theorem
1.4.13] and its proof in [CF, Lemma 5]. $\Box$
For our next result, we want to establish that the logarithm of the Poisson
kernel, that is, $\ln(\frac{d\omega^{X}}{d\sigma})$, is integrable. We want a
quantitative version, and moreover, we shall state this in a form that is more
directly applicable when we need it (in the proof of Proposition 3.8).
###### Lemma 3.4.
Let $\Gamma$, $\sigma$, and $L_{\beta,\gamma}$ as in Lemma 3.1.
Take $B:=B(x,r)$, a ball centered on $\Gamma$, and $X=A_{x,r}$, a corkscrew
point as in (2.3). If $\\{Q_{i}\\}_{i\in\mathcal{I}}$ is a finitely
overlapping collection of Borel subsets of $B\cap\Gamma$, then
(3.5)
$\sum_{i\in\mathcal{I}}\left|\ln\left(\frac{\omega_{\beta,\gamma}^{X}(Q_{i})}{\sigma(Q_{i})}\frac{\sigma(B)}{\omega_{\beta,\gamma}^{X}(B)}\right)\right|\sigma(Q_{i})\leq
C\sigma(B),$
where $C$ depends only on $C_{\sigma}$, $C_{0}$, $\beta$, $\gamma$, $n$, $d$,
and the maximal number of overlaps in the collection
$\\{Q_{i}\\}_{i\in\mathcal{I}}$.
Proof. We introduce for $k\in\mathbb{Z}$,
$\mathcal{I}_{k}:=\left\\{i\in\mathcal{I},\,2^{k}\leq\frac{\omega_{\beta,\gamma}^{X}(Q_{i})}{\sigma(Q_{i})}\frac{\sigma(B)}{\omega_{\beta,\gamma}^{X}(B)}\leq
2^{k+1}\right\\}.$
Then we define
$E_{k}:=\bigcup_{i\in\mathcal{I}_{k}}Q_{i}.$
Since $\\{Q_{i}\\}$ is finitely overlapping, we have
$\sum_{i\in\mathcal{I}_{k}}\sigma(Q_{i})\leq C\sigma(E_{k})$, and we can thus
write
(3.6)
$\sum_{i\in\mathcal{I}}\ln\left(\frac{\omega_{\beta,\gamma}^{X}(Q_{i})}{\sigma(Q_{i})}\frac{\sigma(B)}{\omega_{\beta,\gamma}^{X}(B)}\right)\sigma(Q_{i})\lesssim\sum_{k\in\mathbb{Z}}(|k|+1)\sigma(E_{k}).$
Yet, since $\\{Q_{i}\\}_{i\in\mathcal{I}_{k}}$ is a finitely overlapping
covering of $E_{k}$,
$\omega_{\beta,\gamma}^{X}(E_{k})\approx\sum_{i\in\mathcal{I}_{k}}\omega_{\beta,\gamma}^{X}(Q_{i})\approx
2^{k}\frac{\omega_{\beta,\gamma}^{X}(B)}{\sigma(B)}\sum_{i\in\mathcal{I}_{k}}\sigma(Q_{i})\approx
2^{k}\frac{\omega_{\beta,\gamma}^{X}(B)}{\sigma(B)}\sigma(E_{k}),$
which means that
(3.7)
$\frac{\omega_{\beta,\gamma}^{X}(E_{k})}{\sigma(E_{k})}\frac{\sigma(B)}{\omega_{\beta,\gamma}^{X}(B)}\approx
2^{k}.$
The use of (3.7) in (3.2) leads to
$\sigma(E_{k})\lesssim 2^{-k/(1-\theta)}\sigma(B)$
while the use of (3.7) in (3.3) gives
$\sigma(E_{k})\lesssim 2^{k\theta/(1-\theta)}\sigma(B).$
We use the first of the last two estimate to bound $\sigma(E_{k})$ when $k\geq
0$ and the second one to bound $\sigma(E_{k})$ when $k$ is negative. Combined
with (3.6), we deduce
$\begin{split}\sum_{i\in\mathcal{I}}\ln\left(\frac{\omega_{\beta,\gamma}^{X}(Q_{i})}{\sigma(Q_{i})}\frac{\sigma(B)}{\omega_{\beta,\gamma}^{X}(B)}\right)\sigma(Q_{i})&\lesssim\sum_{k\geq
0}(k+1)2^{-\frac{k}{1-\theta}}\sigma(B)+\sum_{k<0}(1-k)2^{-\frac{k\theta}{1-\theta}}\sigma(B)\\\
&\lesssim\sigma(B).\end{split}$
The lemma follows. $\Box$
###### Proposition 3.8.
Let $\Gamma\subset\mathbb{R}^{n}$ be a $d$-Ahlfors regular uniformly
rectifiable set with $d<n-2$, and let $\sigma$ be an Ahlfors regular measure
that satisfies (2.1). Take $\beta>0$ and $\gamma\in(-1,1)$, define
$L_{\beta,\gamma}$ as in (1.1). Then for any ball $B$ centered on $\Gamma$ and
any non-negative non identically zero weak solution $u$ to
$L_{\beta,\gamma}u=0$ in $3B\cap\Omega$, with $u\in W_{\gamma,0(3B)}$,
(3.9)
$\int_{B}\left|\nabla\ln\left(\frac{u}{D_{\beta}^{1-\gamma}}\right)\right|^{2}\,D_{\beta}^{d+2-n}\leq
C\sigma(B).$
with a constant $C>0$ that depends only on $C_{\sigma}$, $C_{0}$, $\beta$,
$\gamma$, $d$, and $n$.
Observe that Proposition 3.8 is the special case of Theorem 2.21 where we take
$\alpha=\beta$. We shall now check that conversely, Theorem 2.21 follows from
Proposition 3.8 and Lemma 2.17.
Proof of Theorem 2.21 from Proposition 3.8 with the help of Lemma 2.17. Let
$\alpha>0$. Notice that
$\begin{split}\left|\nabla\ln\left(\frac{u}{D_{\alpha}^{1-\gamma}}\right)\right|&=\left|\nabla\left[\ln\left(\frac{u}{D_{\beta}^{1-\gamma}}\right)+(1-\gamma)\ln\left(\frac{D_{\beta}}{D_{\alpha}}\right)\right]\right|\\\
&\leq\left|\nabla\left[\ln\left(\frac{u}{D_{\beta}^{1-\gamma}}\right)\right]\right|+(1-\gamma)\frac{D_{\alpha}}{D_{\beta}}\left|\nabla\left[\dfrac{D_{\beta}}{D_{\alpha}}\right]\right|.\end{split}$
So Proposition 3.8, Lemma 2.17, and (2.2) easily implies (3.9). $\Box$
Proof of Proposition 3.8. The proof of the lemma will be divided into 4 steps.
The core step is step 4, where we use the properties of the solutions. Step 3
treats in advance the most complicated term that we met in Step 4. In this
step, we will compare a solution $u$ to a Green function, then to the harmonic
measure, and finally we use Lemma 3.4. In Step 2, we construct the finitely
overlapping covering $\\{Q_{i}\\}_{i\in\mathcal{I}}$ that will be needed in
order to invoke Lemma 3.4. At last, Step 1 introduces the cut-off function
$\phi_{B,\epsilon}$ used to bound the left-hand side of (2.22) and shows that
$D_{\beta}\nabla\phi_{B,\epsilon}$ satisfies the Carleson measure condition.
Step 1: Introduction of the cut-off function $\phi_{B,\epsilon}$. Let
$B=B(x,r)$ be a ball in $\mathbb{R}^{n}$ centered on the boundary $\Gamma$ and
$\epsilon>0$ small. The proof of Theorem 2.21 is a local one and thus as usual
will involve cut off functions. Take $\psi\in C^{\infty}_{0}(\mathbb{R})$ be
such that $\psi\equiv 1$ on $[-1,1]$, $\psi$ is compactly supported in
$(-2,2)$, $0\leq\psi\leq 1$, and $|\psi^{\prime}|\leq 2$. We define the
function $\phi_{B,\epsilon}$ on $\Omega$ by
(3.10)
$\phi_{B,\epsilon}(X):=\psi\left(\frac{\,\mathrm{dist}(X,B)}{10\,\mathrm{dist}(X,\Gamma)}\right)\psi\left(\frac{2\,\mathrm{dist}(X,B)}{r}\right)\psi\left(\frac{\epsilon}{\,\mathrm{dist}(X,\Gamma)}\right).$
The support of $\phi_{B,\epsilon}$ is thus contained in
(3.11) $E_{0}=\big{\\{}X\in 2B,\,\,\mathrm{dist}(X,B)\leq
20\,\mathrm{dist}(X,\Gamma)\text{ and
}\,\mathrm{dist}(X,\Gamma)\geq\epsilon/2\big{\\}}.$
The gradient of $\phi_{B,\epsilon}$ comes from $3$ regions; the first one
(associated to the first cut-off) is
(3.12) $\displaystyle\big{\\{}X\in E_{0}\,,$ $\displaystyle
10\,\mathrm{dist}(X,\Gamma)\leq\,\mathrm{dist}(X,B)\leq
20\,\mathrm{dist}(X,\Gamma)\big{\\}}$ $\displaystyle\subset
E_{1}:=\big{\\{}X\in
2B,\,10\,\mathrm{dist}(X,\Gamma)\leq\,\mathrm{dist}(X,B)\leq
20\,\mathrm{dist}(X,\Gamma)\big{\\}},$
the second one is
(3.13) $\big{\\{}X\in E_{0}\,,\,r\leq 2\,\mathrm{dist}(X,B)\leq
2r\big{\\}}\subset E_{2}:=\big{\\{}X\in
2B,\,\,r/40\leq\,\mathrm{dist}(X,\Gamma)\leq 2r\big{\\}}$
because $\,\mathrm{dist}(X,\Gamma)\leq 2r$ when $X\in 2B$ and similarly
$\,\mathrm{dist}(X,\Gamma)\geq\frac{1}{20}\,\mathrm{dist}(X,B)\geq\frac{r}{20}$,
and the third region is contained in
(3.14) $E_{3}:=\big{\\{}X\in
2B,\,\epsilon/2\leq\,\mathrm{dist}(X,\Gamma)\leq\epsilon\big{\\}}.$
Then it is easy to check that the gradient of $\phi_{B,\epsilon}$ satisfies
(3.15)
$|\nabla\phi_{B,\epsilon}|\leq\frac{100}{\,\mathrm{dist}(X,\Gamma)}\left[{\mathds{1}}_{E_{1}}+{\mathds{1}}_{E_{2}}+{\mathds{1}}_{E_{3}}\right].$
We claim that ${\mathds{1}}_{E_{1}}$, ${\mathds{1}}_{E_{2}}$, and
${\mathds{1}}_{E_{3}}$ all satisfy the Carleson measure condition. That is, we
have
(3.16)
$\int_{B(y,s)}{\mathds{1}}_{E_{j}}^{2}\,\mathrm{dist}(X,\Gamma)^{d-n}\leq
Cs^{d}$
for $1\leq j\leq 3$, $y\in\Gamma$, and $0<s<+\infty$. Of course the square can
be removed, and it will follows from (3.15), (3.16) (as soon as we prove it),
and (2.2) that for any $\beta>0$, $y\in\Gamma$ and $s\in(0,+\infty)$, we have
(3.17)
$\int_{B(y,s)}|D_{\beta}\nabla\phi_{B,\epsilon}|D_{\beta}^{d-n}+\int_{B(y,s)}|D_{\beta}\nabla\phi_{B,\epsilon}|^{2}D_{\beta}^{d-n}\leq
Cs^{d},$
where the constant $C>0$ depends only on $\beta$, $n$, $d$, and $C_{\sigma}$.
We will prove the claim (3.16) in the end of the next Step.
Step 2: Construction of the collection $\\{Q_{i}\\}_{i\in\mathcal{I}}$. We
would like to have a collection of boundary cubes $Q_{i}\subset\Gamma$ and the
corresponding Whitney cubes $R_{i}\subset\Omega$ satisfying rather usual nice
properties: bounded overlap, control of the size and and the distance to the
boundary for $R_{i}$’s, reasonable placement of the corkscrew points. There
are plenty of papers that present various versions of this construction, but
here, in fact, we need something simpler than usual as we do not need to
control the cones or the related Harnack chains. For these reasons, let us
simply carry out the construction by hands.
We need a family of Whitney cubes $\mathcal{W}$, as constructed in [Ste]. We
record the basic properties of $\mathcal{W}$ that we shall need. The
collection $\mathcal{W}$ is the family of maximal dyadic cubes
$R\subset\Omega$ such that $20R\subset\Omega$, different cubes $R_{i}$ and
$R_{j}$, $i\neq j$, in $\mathcal{W}$ have disjoint interiors (by maximality),
their union covers $\Omega=\mathbb{R}^{n}\setminus\Gamma$, and for
$Q\in{\mathcal{W}}$
(3.18) $20Q\subset\Omega\text{ but }60R\cap\Gamma\neq\emptyset.$
Moreover, given $R\in\mathcal{W}$, the number of cubes
$\widetilde{R}\in\mathcal{W}$ such that
(3.19) $\,\mathrm{dist}(R,\Gamma)\approx\,\mathrm{dist}(\widetilde{R},\Gamma)\
\text{ and
}\,\,\mathrm{dist}(R,\widetilde{R})\lesssim\,\mathrm{dist}(R,\Gamma)$
is uniformly bounded by a constant that depends only on the dimensions and the
constants involved in (3.19). Also, the cubes $\widetilde{R}\in{\mathcal{W}}$
such that $3R\cap 3\widetilde{R}$, all satisfy (3.19).
We will use the cubes of ${\mathcal{W}}$ to cover the region $E_{1}\cup
E_{2}\cup E_{3}$. Consider the subset ${\mathcal{W}}_{0}\subset{\mathcal{W}}$
of cubes $R\in{\mathcal{W}}$ that meet $E_{1}\cup E_{2}\cup E_{3}$, and label
these cubes with a set $I$, so that ${\mathcal{W}}_{0}=\\{R_{i}\,;\,i\in
I\\}$. We also want to associate a boundary ball $Q_{i}=\Gamma\cap
B(x_{i},r_{i})$ to each $R_{i}$, $i\in I$. This is a classical thing to do,
but we shall do it by hand to get a slightly better control. We shall choose
the $Q_{i}$ so that
(3.20) $Q_{i}\subset 3B\cap\Gamma$
(we can do this and this will be helpful because we shall consider solutions
in $3B$),
(3.21) $\\{Q_{i}\\}_{i\in\mathcal{I}}$ has bounded overlap,
where the bound for the overlap depends only on $C_{\sigma}$, $n$, and $d$,
and
(3.22)
$r_{i}\approx\,\mathrm{dist}(R_{i},\Gamma)\approx\,\mathrm{dist}(R_{i},x_{i}),$
also with constants depend only on $C_{\sigma}$, $n$, and $d$. As a
consequence, we will be able to use any point $X_{i}\in R_{i}$ as a corkscrew
point for the pair $(x_{i},Cr_{i})$ (see (2.3)).
So let us construct the $Q_{i}$. First write $I=I_{1}\cup I_{2}\cup I_{3}$, a
disjoint union where $I_{2}=\big{\\{}i\in I\,;\,R_{i}\cap
E_{2}\neq\emptyset\big{\\}}$, then $I_{3}=\big{\\{}i\in I\setminus
I_{2}\,;\,R_{i}\cap E_{3}\neq\emptyset\big{\\}}$, and finally
$I_{1}=\big{\\{}i\in I\setminus(I_{2}\cup I_{3})\,;\,R_{i}\cap
E_{1}\neq\emptyset\big{\\}}$.
We start with $i\in I_{2}$. This is the simplest case because $E_{2}\subset
2B=B(x,2r)$, and $\,\mathrm{dist}(X,\Gamma)\geq r/40$ on $E_{2}$. By
definition of ${\mathcal{W}}$, $I_{2}$ has at most $C$ elements (see near
(3.19)), and for each $i\in I_{2}$ we take $x_{i}=x$ and $r_{i}=3r$ (and hence
$Q_{i}=\Gamma\cap 3B$); the constraints (3.20), (3.21), (3.22) for $I_{2}$ are
easily checked.
Next consider $i\in I_{3}$; thus $R_{i}$ meets $E_{3}$, where
$\varepsilon/2\leq\,\mathrm{dist}(X,\Gamma)\leq\varepsilon$. Pick
$x_{i}\in\Gamma$ such that $\,\mathrm{dist}(x_{i},E_{3}\cap
R_{i})=\,\mathrm{dist}(\Gamma,E_{3}\cap R_{i})$, and take $r_{i}=\varepsilon$.
Then $Q_{i}\subset 3B$ if $\varepsilon$ is small enough, because $E_{3}\subset
2B$ and
$\mathrm{diam}(R_{i})\approx\,\mathrm{dist}(R_{i},\Gamma)\approx\,\mathrm{dist}(E_{3}\cap
R_{i},\Gamma)\approx\varepsilon$, (3.22) holds for the same reasons, and the
$Q_{i}$, $i\in I_{3}$ have bounded overlap by the property near (3.19).
Finally, for $i\in I_{1}$, take $X_{i}\in R_{i}\cap E_{1}$ and
$x_{i}\in\Gamma$ such that $|X_{i}-x_{i}|=\,\mathrm{dist}(R_{i}\cap
E_{1},\Gamma)$, and set $Q_{i}=\Gamma\cap B(x_{i},r_{i})$, with
$r_{i}=|X_{i}-x_{i}|$. Since $X_{i}\in E_{1}$, the definition (3.12) yields
(3.23) $\,\mathrm{dist}(X_{i},B)\leq
20\,\mathrm{dist}(X,\Gamma)=20|X_{i}-x_{i}|\leq 2\,\mathrm{dist}(X_{i},B)\leq
2r,$
and in particular $x_{i}\in\frac{5}{2}B$. Also, $X_{i}\notin E_{2}$ (because
$i\notin I_{2}$), so $r_{i}=\,\mathrm{dist}(R_{i}\cap
E_{1},\Gamma)\leq\,\mathrm{dist}(X_{i},\Gamma)\leq r/40$, and hence
$Q_{i}\subset 3B$, as needed for (3.20). The Whitney property (3.22) holds
essentially by definition (and because $R_{i}$ is a Whitney cube), so we just
need to bound the overlap of the $Q_{i}$.
Assume that $Q_{j}\cap Q_{i}\neq\emptyset$ for two indices
$i,j\in\mathcal{I}_{1}$, and let $X_{i},X_{j},r_{i}$, and $r_{j}$ be as above.
Let us first check that $r_{i}\approx r_{j}$. By (3.23)
$\,\mathrm{dist}(x_{i},B)\leq|X_{i}-x_{i}|+\,\mathrm{dist}(X_{i},B)\leq
21|X_{i}-x_{i}|=21r_{i}$
and
$\,\mathrm{dist}(x_{i},B)\geq\,\mathrm{dist}(X_{i},B)-|X_{i}-x_{i}|\leq
9|X_{i}-x_{i}|=9r_{i},$
which can be summarized as
(3.24) $9r_{i}\leq\,\mathrm{dist}(x_{i},B)\leq 21r_{i}.$
Similarly,
(3.25) $9r_{j}\leq\,\mathrm{dist}(x_{j},B)\leq 21r_{j}.$
We can assume without loss of generality that
$\,\mathrm{dist}(x_{j},B)\leq\,\mathrm{dist}(x_{i},B)$, which, together with
(3.24)-(3.25), leads to $9r_{j}\leq 21r_{i}$. Moreover, since $Q_{i}\cap
Q_{j}\neq\emptyset$,
$r_{i}+r_{j}\geq|x_{j}-x_{i}|\geq\,\mathrm{dist}(x_{i},B)-\,\mathrm{dist}(x_{j},B)\geq
9r_{i}-21r_{j},$
hence $22r_{j}\geq 8r_{i}$. Recall that $9r_{j}\leq 21r_{i}$, so the two radii
are equivalent. The bounded overlap property follows, because
$|x_{j}-x_{i}|\leq r_{i}+r_{j}$ when $Q_{i}\cap Q_{j}\neq\emptyset$.
This completes our construction of Whitney cubes $R_{i}$ and associated
surface balls $Q_{i}$, with the properties (3.20)–(3.22) (notice that the
overlap constant for the $Q_{i}$, $i\in I$, is less than the sum of the
overlap constants for the $I_{j}$).
With this at hand, let us prove (3.16). This is now quite easy as for any
$y\in\Gamma$ and $0<s<+\infty$ we have
(3.26) $\int_{B(y,s)}{\mathds{1}}_{E_{1}\cup E_{2}\cup
E_{3}}^{2}\,\mathrm{dist}(X,\Gamma)^{d-n}\\\ \leq\sum_{R_{i}:\,R_{i}\cap
B(y,s)\neq\emptyset}\int_{R_{i}}\,\mathrm{dist}(X,\Gamma)^{d-n}\lesssim\sum_{R_{i}:\,R_{i}\cap
B(y,s)\neq\emptyset}r_{i}^{d}\lesssim s^{d},$
using the bounded overlap property (3.21).
Step 3: Estimates for the integral $S$. Take $u\in W_{\gamma,0}(3B)$, a weak
solution to $L_{\beta,\gamma}u=0$ in $3B$. Let $X_{0}\in 2B$ be a corkscrew
point for the boundary ball $2B\cap\Gamma$, and also choose another corkscrew
point $X_{1}\in\Omega\cap 8B\setminus 4B$, so that
$\,\mathrm{dist}(X_{0},\Gamma)\geq cr$ and $\,\mathrm{dist}(X_{1},\Gamma)\geq
cr$ for a small constant $c$ that depends only on $C_{\sigma}$, $d$, and $n$.
In this Step 3, we prove that
(3.27) $S:=\int_{E_{1}\cup E_{2}\cup
E_{3}}\left|\ln\left(\frac{u}{D_{\beta}^{1-\gamma}}\frac{D_{\beta}^{1-\gamma}(X_{0})}{u(X_{0})}\right)\right|D_{\beta}^{d-n}\leq
Cr^{d},$
where $C$ depends only on $C_{\sigma}$, $C_{0}$, $\beta$, $\gamma$, $n$ and
$d$.
We shall use the comparison principle to compare $u$ to the Green function,
and then estimate the Green function in terms of harmonic measure. We define
the Green function $g$ on $\Omega\times\Omega$ as in [DFM5, Section 11]. The
precise definition is not relevant for the present proof, and the properties
that we care about are the fact that $X\to g(X,X_{1})$ lies in
$W_{\gamma,0}(3B)$ and is a solution to $L_{\beta,\gamma}u=0$ in $3B$, and
that $g(X,Y)=g(Y,X)$ (which is true because the operator $L_{\beta,\gamma}$ is
selfadjoint). Theorem 15.64 in [DFM5] (the comparison theorem) yields that
(3.28)
$\frac{u(X)}{u(X_{0})}\approx\frac{g(X,X_{1})}{g(X_{0},X_{1})}\qquad\text{ for
}X\in 2B,$
with constants that depend only on $C_{\sigma}$, $C_{0}$, $n$, $d$, $\beta$
and $\gamma$. Actually, Theorem 15.64 in [DFM5] requires the solutions (in our
case $u$ and $g(.,X_{1})$) to be solutions in a larger ball $2KB\cap\Omega$,
and not in only $3B\cap\Omega$. This condition is only needed because [DFM5]
also allows sets $\Gamma$ of codimension $1$ or less, and then we need to
ensure that we can connect every component of $2B\cap\Omega$ by Harnack chains
that stays in $2KB$. Here $\Gamma$ is of dimension $d<n-1$, so $2B\cap\Omega$
is very well connected in the first place (see Lemma 2.1 in [DFM2]) , and
assuming that $u$ and $g(.,X_{1})$ are solutions in $3B$ is enough. Using the
fact that $g(X,Y)$ is symmetric, we deduce that
(3.29)
$\frac{u(X)}{u(X_{0})}\approx\frac{g(X_{1},X)}{g(X_{1},X_{0})}\qquad\text{ for
}X\in 2B.$
Recall that $X_{0}$ is a corkscrew point associated to $2B$ and that if $X\in
R_{i}$, then $\,\mathrm{dist}(X,\Gamma)\geq C^{-1}r_{i}$ and $|X-x_{i}|\leq
Cr_{i}$ by (3.22), so $X$ can be used as a corkscrew point for the boundary
ball $Q_{i}$. As a consequence, Lemma 15.28 in [DFM5], (where here
$m(B\cap\Omega)$ is the mass of $B$ for the measure
$\,\mathrm{dist}(X,\Gamma)^{d+1+\gamma-n}dX$ on $\Omega$, with
$\gamma\in(-1,1)$, so $m(B\cap\Omega)\approx r^{d+1+\gamma}$; see the
discussion in Section 3.2 of [DFM5]), and the doubling property of harmonic
measure (Lemma 15.43 of [DFM5]) give that
$g(X_{1},X_{0})\approx\,\mathrm{dist}(X_{0},\Gamma)^{1-d-\gamma}\omega^{X_{1}}(3B)$
and for each $i\in I$
$g(X_{1},X)\approx\,\mathrm{dist}(X,\Gamma)^{1-d-\gamma}\omega^{X_{1}}(Q_{i})\qquad\text{
for }X\in R_{i},$
where the constants depend only on $C_{\sigma}$, $C_{0}$, $n$, $d$, $\beta$
and $\gamma$. Using the equivalence (2.2) and the Ahlfors regularity of
$\sigma$, we deduce that
(3.30) $g(X_{1},X_{0})\approx
D_{\beta}(X_{0})^{1-\gamma}\,\frac{\omega^{X_{1}}(3B)}{\sigma(3B)}$
and for each $i\in I$,
(3.31) $g(X_{1},X)\approx
D_{\beta}(X)^{1-\gamma}\,\frac{\omega^{X_{1}}(Q_{i})}{\sigma(Q_{i})}\qquad\text{
for }X\in R_{i}.$
We gather (3.29), (3.30), and (3.31) to obtain that, for every $i\in I$ and
every $X\in R_{i}$,
(3.32)
$\frac{u(X)}{D_{\beta}^{1-\gamma}(X)}\frac{D_{\beta}^{1-\gamma}(X_{0})}{u(X_{0})}\approx\frac{\omega^{X_{1}}(Q_{i})}{\sigma(Q_{i})}\frac{\sigma(3B)}{\omega^{X_{1}}(3B)},$
where the constants depend only on $C_{\sigma}$, $C_{0}$, $n$, $d$, $\beta$
and $\gamma$. This immediately implies that for $i\in I$ and $X\in R_{i}$,
$\left|\ln\left(\frac{u(X)}{D_{\beta}^{1-\gamma}(X)}\frac{D_{\beta}^{1-\gamma}(X_{0})}{u(X_{0})}\right)\right|\leq
C+\left|\ln\left(\frac{\omega^{X_{1}}(Q_{i})}{\sigma(Q_{i})}\frac{\sigma(3B)}{\omega^{X_{1}}(3B)}\right)\right|$
and then
(3.33) $\begin{split}S&\leq\sum_{i\in
I}\int_{R_{i}}\left|\ln\left(\frac{u}{D_{\beta}^{1-\gamma}}\frac{D_{\beta}^{1-\gamma}(X_{0})}{u(X_{0})}\right)\right|D_{\beta}^{d-n}\\\
&\leq\sum_{i\in
I}\left[C+\left|\ln\left(\frac{\omega^{X_{1}}(Q_{i})}{\sigma(Q_{i})}\frac{\sigma(3B)}{\omega^{X_{1}}(3B)}\right)\right|\right]\int_{R_{i}}D_{\beta}^{d-n}.\end{split}$
By(2.2) and the fact that $R_{i}\in{\mathcal{W}}$ is a Whitney cube with the
property (3.22) (by construction), we have $D_{\beta}\approx r_{i}$ on $R_{i}$
and $|R_{i}|\lesssim r_{i}^{n}$. Hence
$\int_{R_{i}}D_{\beta}^{d-n}\lesssim r_{i}^{d}\approx\sigma(Q_{i})$
by (2.1). Using this observation in (3.33), we infer that
$\begin{split}S&\leq\sum_{i\in
I}\left[C+\left|\ln\left(\frac{\omega^{X_{1}}(Q_{i})}{\sigma(Q_{i})}\frac{\sigma(3B)}{\omega^{X_{1}}(3B)}\right)\right|\right]\sigma(Q_{i})\\\
&=C\sum_{i\in I}\sigma(Q_{i})+\sum_{i\in
I}\left|\ln\left(\frac{\omega^{X_{1}}(Q_{i})}{\sigma(Q_{i})}\frac{\sigma(3B)}{\omega^{X_{1}}(3B)}\right)\right|\sigma(Q_{i}).\end{split}$
The first term of the right-hand side is bounded by $C\sigma(3B)$ since the
$Q_{i}$, $i\in I$, are contained in $\Gamma\cap 3B$ (by (3.20)) and have
bounded overlap (by (3.21)). The second term is also less than $C\sigma(3B)$,
by Lemma 3.4. We conclude that
$S\lesssim\sigma(3B)\lesssim r^{d}$
by (2.1). The claim 3.27 follows.
Step 4: Core of the proof. Let us turn to the main and last step of the proof.
Set
(3.34)
$T:=\int_{\Omega}\bigg{|}\nabla\ln\Big{(}\frac{u}{D_{\beta}^{1-\gamma}}\Big{)}\bigg{|}^{2}\phi_{B,\epsilon}^{2}D_{\beta}^{d+2-n}.$
We aim to prove that
(3.35) $T\leq Cr^{d}+Cr^{d/2}T^{1/2},$
where $C$ depends only on $C_{\sigma}$, $C_{0}$, $\beta$, $\gamma$, $n$ and
$d$. The solution $u$ and the smooth distance $D_{\beta}$ are uniformly
bounded from above and from below by a positive constant on the support of the
cut-off function $\phi_{B,\epsilon}$. Thanks to this fact, the quantities of
both sides of (3.35) are finite, and (3.35) self-improves into
(3.36)
$\int_{\Omega}\Big{|}\nabla\ln\Big{(}\frac{u}{D_{\beta}^{1-\gamma}}\Big{)}\Big{|}^{2}\phi_{B,\epsilon}^{2}D_{\beta}^{d+2-n}=T\leq
Cr^{d},$
where $C>0$ depends on the same parameters as in (3.35). Once we are there,
the left-hand side is uniformly bounded in $\epsilon$, so taking $\epsilon\to
0$ leads to the desired result (3.9).
Keep in mind that
(3.37) $\nabla\ln\bigg{(}\frac{u}{D_{\beta}^{1-\gamma}}\bigg{)}=\frac{\nabla
u}{u}-\frac{\nabla
D_{\beta}^{1-\gamma}}{D_{\beta}^{1-\gamma}}=\dfrac{D_{\beta}^{1-\gamma}\nabla
u-u\nabla D_{\beta}^{1-\gamma}}{D_{\beta}^{1-\gamma}u}.$
We shall use Lemma 2.10 to obtain the existence of a scalar function $b$ and a
vector function $\mathcal{V}$ such that
(3.38) $H_{n-d-1}:=\int_{\Gamma}|X-y|^{-n}(X-y)d\sigma(y)=(b\nabla
D_{\beta}+\mathcal{V})D_{\beta}^{d+1-n}$
as in (2.11), with the bounds (2.12)-(2.15).
Before we start to bound $T$, let us comment on $H_{n-d-1}$ and $\nabla
D_{\beta}$. First, the vector function $H_{n-d-1}$ is smooth and divergence
free in $\Omega$, and the formulation of this fact in the weak sense is that
for any compactly supported $\varphi\in W^{1,1}(\Omega)$,
(3.39) $\int_{\Omega}H_{n-d-1}\cdot\nabla\varphi=0.$
Set $H_{\alpha}(X):=\int_{\Gamma}|X-y|^{-d-1-\alpha}(X-y)d\sigma(y)$ for
$\alpha>0$. Since
$|H_{\alpha}(X)|\leq\int_{\Gamma}|X-y|^{-d-\alpha}d\sigma(y)=D_{\alpha}(X)^{-\alpha}$,
(2.2) implies that
(3.40) $|H_{\alpha}|\lesssim D_{\alpha}^{-\alpha}\lesssim
D_{\beta}^{-\alpha}.$
Moreover, observe that $\nabla(D_{\beta}^{-\beta})=-(d+\beta)H_{\beta+1}$,
directly by (1.2); since also $\nabla(D_{\beta}^{-\beta})=-\beta
D_{\beta}^{-\beta-1}\nabla D_{\beta}$, we deduce from (3.40) that
(3.41) $|\nabla D_{\beta}|\lesssim 1.$
We turn to the proof of (3.35). Write $\phi$ instead of $\phi_{B,\epsilon}$ to
lighten the notation. By (2.12) and (3.37),
$\begin{split}T\lesssim\int_{\Omega}\bigg{|}\nabla\ln\bigg{(}\frac{u}{D_{\beta}^{1-\gamma}}\bigg{)}\bigg{|}^{2}\phi^{2}b\,D_{\beta}^{d+2-n}&=\int_{\Omega}\frac{b\nabla
u}{u}\cdot\bigg{(}\dfrac{D_{\beta}^{1-\gamma}\nabla u-u\nabla
D_{\beta}^{1-\gamma}}{D_{\beta}^{1-\gamma}u}\bigg{)}\
\phi^{2}D_{\beta}^{d+2-n}\\\ &\hskip 56.9055pt-\int_{\Omega}\frac{b\nabla
D_{\beta}^{1-\gamma}}{D_{\beta}^{1-\gamma}}\cdot\nabla\bigg{[}\ln\bigg{(}\frac{u}{D_{\beta}^{1-\gamma}}\bigg{)}\bigg{]}\phi^{2}D_{\beta}^{d+2-n}\\\
&:=T_{1}-T_{2}.\end{split}$
Let us start with $T_{2}$. We simplify the factors $D_{\beta}$ and we invoke
the relation (3.38) to get
$\begin{split}T_{2}&=(\gamma-1)\int_{\Omega}\frac{b\nabla
D_{\beta}}{D_{\beta}^{n-d-1}}\cdot\nabla\bigg{[}\ln\bigg{(}\frac{u}{D_{\beta}^{1-\gamma}}\bigg{)}\bigg{]}\phi^{2}\\\
&=(\gamma-1)\int_{\Omega}H_{n-d-1}\cdot\nabla\bigg{[}\ln\bigg{(}\frac{u}{D_{\beta}^{1-\gamma}}\bigg{)}\bigg{]}\phi^{2}-(\gamma-1)\int_{\Omega}\mathcal{V}\cdot\nabla\bigg{[}\ln\bigg{(}\frac{u}{D_{\beta}^{1-\gamma}}\bigg{)}\bigg{]}\phi^{2}D_{\beta}^{d+1-n}\\\
&:=T_{21}+T_{22}.\end{split}$
We use the Cauchy-Schwarz inequality to bound $T_{22}$, as follows:
$\begin{split}|T_{22}|&\lesssim\Big{(}\int_{\Omega}|\mathcal{V}|^{2}\phi^{2}D_{\beta}^{d-n}\Big{)}^{\frac{1}{2}}\bigg{(}\int_{\Omega}\bigg{|}\nabla\bigg{[}\ln\bigg{(}\frac{u}{D_{\beta}^{1-\gamma}}\bigg{)}\bigg{]}\bigg{|}^{2}\phi^{2}D_{\beta}^{d+2-n}\bigg{)}^{\frac{1}{2}}\\\
&\lesssim
T^{1/2}\left(\int_{2B}|\mathcal{V}|^{2}D_{\beta}^{d-n}\right)^{\frac{1}{2}}\lesssim
r^{d/2}T^{1/2}\end{split}$
by (2.15). This fits with (3.35). As for $T_{21}$, notice that its value will
not be changed if we replace $u$ by $Ku$, where $K$ is a constant. Hence
$T_{21}=(\gamma-1)\int_{\Omega}H_{n-d-1}\cdot\nabla\bigg{[}\ln\bigg{(}\frac{Ku}{D_{\beta}^{1-\gamma}}\bigg{)}\bigg{]}\phi^{2},$
where $K$ is a constant to be chosen later. We force $\phi^{2}$ into the
gradient, then use the fact that $H_{n-d-1}$ is divergence free (see (3.39)),
and obtain
$\displaystyle T_{21}$
$\displaystyle=(\gamma-1)\int_{\Omega}H_{n-d-1}\cdot\nabla\Big{[}\phi^{2}\ln\Big{(}\frac{Ku}{D_{\beta}^{1-\gamma}}\Big{)}\Big{]}-2(\gamma-1)\int_{\Omega}H_{n-d-1}\cdot\nabla\phi\
\phi\ln\Big{(}\frac{Ku}{D_{\beta}^{1-\gamma}}\Big{)}$
$\displaystyle=0-2(\gamma-1)\int_{\Omega}H_{n-d-1}\cdot\nabla\phi\
\phi\ln\Big{(}\frac{Ku}{D_{\beta}^{1-\gamma}}\Big{)}.$
Recall that $|H_{n-d-1}|\lesssim D_{\beta}^{d+1-n}$ by (3.40), and that
$|\nabla\phi|\lesssim{\mathds{1}}_{E_{1}\cup E_{2}\cup E_{3}}/D_{\beta}$ by
(3.15) and (2.2), so
$|T_{21}|\lesssim\int_{E_{1}\cup E_{2}\cup
E_{3}}\bigg{|}\ln\bigg{(}\frac{Ku}{D_{\beta}^{1-\gamma}}\bigg{)}\bigg{|}D_{\beta}^{d-n}.$
We choose $K=D_{\beta}^{1-\gamma}(X_{0})/u(X_{0})$, so that the right-hand
side above is what we called $S$ in (3.27). Hence $T_{21}\lesssim r^{d}$ by
(3.27), as needed for (3.35).
We switch to the estimation of $T_{1}$. We want to use the fact that $u$ is a
solution, and for this we write
(3.42) $\begin{split}T_{1}&=\int_{\Omega}\nabla
u\cdot\frac{D_{\beta}^{1-\gamma}}{u}\bigg{(}\dfrac{D_{\beta}^{1-\gamma}\nabla
u-u\nabla D_{\beta}^{1-\gamma}}{D_{\beta}^{1-\gamma}u}\bigg{)}\
b\phi^{2}D_{\beta}^{d+1+\gamma-n}\\\ &=-\int_{\Omega}\nabla
u\cdot\nabla\bigg{[}\dfrac{D_{\beta}^{1-\gamma}}{u}\bigg{]}\
b\phi^{2}D_{\beta}^{d+1+\gamma-n}\\\ &=-\int_{\Omega}\nabla
u\cdot\nabla\bigg{[}b\phi^{2}\dfrac{D_{\beta}^{1-\gamma}}{u}\bigg{]}\
D_{\beta}^{d+1+\gamma-n}+\int_{\Omega}\nabla u\cdot\nabla
b\bigg{(}\phi^{2}\dfrac{D_{\beta}^{1-\gamma}}{u}\bigg{)}\
D_{\beta}^{d+1+\gamma-n}\\\ &\qquad+2\int_{\Omega}\nabla
u\cdot\nabla\phi\bigg{(}\phi b\dfrac{D_{\beta}^{1-\gamma}}{u}\bigg{)}\
D_{\beta}^{d+1+\gamma-n}\\\ &:=T_{11}+T_{12}+T_{13}.\end{split}$
Notice that $T_{11}=0$ because $u$ is a weak solution to $L_{\beta,\gamma}u=0$
on $\Omega\cap 2B$ and $\phi^{2}\dfrac{D_{\beta}^{1-\gamma}}{u}$ lies in
$W^{1,2}_{loc}(\Omega)$ and compactly supported in $\Omega$. The terms
$T_{12}$ and $T_{13}$ are morally similar. In both case, we don’t like the
terms with $u$ because we don’t know so much about it, so we use (3.37) to
replace $u$ by the nice function $D_{\beta}^{1-\gamma}$, and the difference
will be controlled with the help of $T^{1/2}$, the square root of our initial
integral. We start with
(3.43) $\begin{split}T_{13}&=2\int_{\Omega}\frac{\nabla
u}{u}\cdot\nabla\phi\,\phi\,b\,D_{\beta}^{d+2-n}\\\
&=2\int_{\Omega}\left(\frac{\nabla u}{u}-\frac{\nabla
D_{\beta}^{1-\gamma}}{D_{\beta}^{1-\gamma}}\right)\cdot(\nabla\phi)\,\phi\,b\,D_{\beta}^{d+2-n}-2(\gamma-1)\int_{\Omega}\frac{b\nabla
D_{\beta}}{D_{\beta}^{n-d-1}}\cdot(\nabla\phi)\phi\\\
&:=T_{131}+T_{132}.\end{split}$
We use (2.12) and the fact that $D_{\beta}\nabla\phi$ satisfies the Carleson
measure property to get
$\begin{split}|T_{131}|&\lesssim\Big{(}\int_{\Omega}|\nabla\phi|^{2}D_{\alpha}^{d+2-n}\Big{)}^{1/2}\bigg{(}\int_{\Omega}\bigg{|}\frac{\nabla
u}{u}-\frac{\nabla
D_{\beta}^{1-\gamma}}{D_{\beta}^{1-\gamma}}\bigg{|}^{2}\phi^{2}D_{\alpha}^{d+2-n}\bigg{)}^{1/2}\\\
&\lesssim
r^{d/2}\bigg{(}\int_{\Omega}\bigg{|}\nabla\ln\bigg{(}\frac{u}{D_{\alpha}^{1-\gamma}}\bigg{)}\bigg{|}^{2}\phi^{2}D_{\alpha}^{d+2-n}\bigg{)}^{1/2}\end{split}$
by (3.17), and then (3.37). The bound of $T_{131}$ that we just obtained
appears in the right-hand side of (3.35), as desired. As for $T_{132}$, we
invoke (3.41), (2.12), and then (3.17) to write
$\begin{split}|T_{132}|&\lesssim\int_{2B}|D_{\beta}\nabla\phi|D_{\beta}^{n-d}\lesssim
r^{d}\end{split}.$
Similarly to $T_{13}$, we treat $T_{12}$ as follows:
(3.44) $\begin{split}T_{12}&=\int_{\Omega}\frac{\nabla u}{u}\cdot\nabla
b\,\phi^{2}D_{\beta}^{d+2-n}\\\ &=\int_{\Omega}\bigg{(}\frac{\nabla
u}{u}-\frac{\nabla
D_{\beta}^{1-\gamma}}{D_{\beta}^{1-\gamma}}\bigg{)}\cdot\nabla
b\,\phi^{2}D_{\beta}^{d+2-n}-(\gamma-1)\int_{\Omega}\nabla
D_{\beta}\cdot\nabla b\,\phi^{2}D_{\beta}^{d+1-n}\\\
&:=T_{121}+T_{122}.\end{split}$
Thanks to the definition (3.34) and then (2.13),
(3.45)
$\begin{split}|T_{121}|&\lesssim\bigg{(}\int_{\Omega}\bigg{|}\frac{\nabla
u}{u}-\frac{\nabla
D_{\beta}^{1-\gamma}}{D_{\beta}^{1-\gamma}}\bigg{|}^{2}\phi^{2}D_{\beta}^{d+2-n}\bigg{)}^{1/2}\Big{(}\int_{\Omega}|\nabla
b|^{2}\phi^{2}D_{\beta}^{d+2-n}\Big{)}^{1/2}\\\ &\lesssim
T^{1/2}\left(\int_{2B}|D_{\beta}\nabla
b|^{2}D_{\beta}^{d-n}\right)^{\frac{1}{2}}\lesssim r^{d/2}T^{1/2}.\end{split}$
Now, we want to use (3.38) again, so we force the function $b$ to appear and
we place all the remaining terms in the second gradient. Then $T_{122}$
becomes
$\begin{split}T_{122}&=-(\gamma-1)\int_{\Omega}b\nabla
D_{\beta}\cdot\frac{\nabla b}{b}\,\phi^{2}D_{\beta}^{d+1-n}\\\
&=-(\gamma-1)\int_{\Omega}b\nabla
D_{\beta}\cdot\nabla[\phi^{2}\ln(b)]\,D_{\beta}^{d+1-n}-2(1-\gamma)\int_{\Omega}b\nabla
D_{\beta}\cdot\nabla\phi\,\phi\ln(b)\,D_{\beta}^{d+1-n}\\\
&:=T_{1221}+T_{1222}.\end{split}$
We start with $T_{1222}$. Since $b\approx 1$, we have that $|b\ln(b)|\lesssim
1$. Besides, $|\nabla D_{\beta}|\lesssim 1$ by (3.41). Therefore,
$|T_{1222}|\lesssim\int_{2B}|D_{\beta}\nabla\phi|D_{\beta}^{d-n},$
and then $|T_{1222}|\lesssim r^{d}$ by (3.17). We use (3.38) to write
$T_{1221}$ as
$\begin{split}T_{1221}&=(1-\gamma)\int_{\Omega}H_{n-d-1}\cdot\nabla[\phi^{2}\ln(b)]+(\gamma-1)\int_{\Omega}\mathcal{V}\cdot\nabla[\phi^{2}\ln(b)]\,D_{\beta}^{d+1-n}\\\
&=0+(\gamma-1)\int_{\Omega}\mathcal{V}\cdot\Big{[}2\phi\nabla\phi\ln(b)+\phi^{2}\frac{\nabla
b}{b}\Big{]}D_{\beta}^{d+1-n}\end{split}$
by (3.39). Hence by (2.12) and then the Cauchy-Schwarz inequality,
$\begin{split}|T_{1221}|&\lesssim\int_{2B}|\mathcal{V}|(|\nabla\phi|+|\nabla
b|)D_{\beta}^{d+1-n}\\\
&\lesssim\left(\int_{2B}|\mathcal{V}|^{2}D_{\beta}^{n-d}\right)^{\frac{1}{2}}\left(\int_{2B}(|D_{\beta}\nabla\phi|^{2}+|D_{\beta}\nabla
b|^{2})D_{\beta}^{d-n}\right)^{\frac{1}{2}}.\end{split}$
But since $\mathcal{V}$, $D_{\beta}\nabla b$, and $D_{\beta}\nabla\phi$ all
satisfies the Carleson measure condition, we conclude that $|T_{1221}|\lesssim
r^{d}$ as desired.
We bounded each term derived from $T$ by either $r^{d}$ or $r^{d/2}T^{1/2}$,
and consequently proved the claim (3.35). As was observed before, (3.9),
Proposition 3.8, and then Theorems 2.21 and 1.8 follow. $\Box$
## References
* [AHMMT] J. Azzam, S. Hofmann, J.M. Martell, M. Mourgouglou, X. Tolsa. Harmonic measure and quantitative connectivity: geometric characterization of the $L^{p}$-solvability of the Dirichlet problem. Preprint, arXiv:1907.07102.
* [AHMNT] J. Azzam, S. Hofmann, J.M. Martell, K. Nyström, T. Toro. A new characterization of chord-arc domains, J. Eur. Math. Soc. 19, no 4, 967–981.
* [CFK] L. Caffarelli, E. Fabes, C. Kenig. Completely singular elliptic-harmonic measures. Indiana Univ. Math. J., 30 (1981), no. 6, 917–924.
* [CS] L. Caffarelli, L. Silvestre, An extension problem related to the fractional Laplacian. Comm. Partial Differential Equations 32 (2007), no. 7-9, 1245–1260.
* [CF] R. R. Coifman, C. Fefferman. Weighted norm inequalities for maximal functions and singular integrals. Studia Math., 51 (1974), 241–250.
* [Da] B. E. J. Dahlberg, _Estimates of harmonic measure_. Arch. Rational Mech. Anal. 65 (1977), no. 3, 275–288.
* [DS1] G. David, S. Semmes. Singular integrals and rectifiable sets in $\mathbb{R}^{n}$: Beyond Lipschitz graphs. Asterisque, 193 (1991).
* [DS2] G. David, S. Semmes. Analysis of and on uniformly rectifiable sets. Mathematical Surveys and Monographs, 38. American Mathematical Society, Providence, RI, 1993.
* [DEM] G. David, M. Engelstein, S. Mayboroda. Square functions, non-tangential limits and harmonic measure in co-dimensions larger than 1. Duke Math. J., to appear.
* [DFM2] G. David, J. Feneuil, S. Mayboroda. Elliptic theory for sets with higher co-dimensional boundaries. Accepted, Mem. Amer. Math. Soc. Available at arXiv:1702.05503.
* [DFM3] G. David, J. Feneuil, S. Mayboroda. Dahlberg’s theorem in higher co-dimension. J. Funct. Anal. 276 (2019), no. 9, 2731–2820.
* [DFM4] G. David, J. Feneuil, S. Mayboroda. A new elliptic measure in lower dimensional sets. Acta Math. Sinica 35 (2019), no 6, 876–902.
* [DFM5] G. David, J. Feneuil, S. Mayboroda. Elliptic theory in domains with boundaries of mixed dimensions. Preprint, arXiv:2003.09037.
* [DM1] G. David, S. Mayboroda. Harmonic measure is absolutely continuous with respect to the Hausdorff measure on all low dimensional uniformly rectifiable sets. Preprint, arXiv:2006.14661.
* [DM2] G. David, S. Mayboroda. Approximation of Green functions and domains with uniformly rectifiable boundaries of all dimensions Preprint, arXiv:2010.09793.
* [FKP] R. Fefferman, C. Kenig and J. Pipher, The theory of weights and the Dirichlet problem for elliptic equations, Ann. of Math. (2) 134 (1991), no. 1, 65–124.
* [Fen1] J. Feneuil. Absolute continuity of the harmonic measure on low doimensional rectifiable sets. Preprint, arXiv:2006.03118.
* [Fen2] J. Feneuil. The Green function with pole at infinity applied to the study of the elliptic measure. Preprint, arXiv:2010.04034.
* [HM1] S. Hofmann, J.M. Martell. Uniform rectifiability and harmonic measure I: uniform rectifiability implies Poisson kernels in $L^{p}$. Ann. Sci. Éc. Norm. Supér. (4), 47 (2014), no. 3, 577–654.
* [HMT] S. Hofmann, J.-M. Martell, T. Toro, $A_{\infty}$ implies NTA for a class of variable coefficient elliptic operators. J. Differential Equations 263 (2017), no. 10, 6147–6188.
* [HMU] S. Hofmann, J.M. Martell, I. Uriarte-Tuero. Uniform rectifiability and harmonic measure, II: Poisson kernels in $L^{p}$ imply uniform rectifiability. Duke Math. J., 163 (2014), no. 8, 1601–1654.
* [Ken] C. E. Kenig. Harmonic analysis techniques for second order elliptic boundary value problems. CBMS Regional Conference Series in Mathematics, 83. Amer. Math. Soc., Providence, RI, 1994.
* [MM] L. Modica, S. Mortola. Construction of a singular elliptic harmonic measure. Manuscripta Math. 33 (1980), no. 1, 81–98.
* [Ste] E. M. Stein. Harmonic analysis: real-variable methods, orthogonality, and oscillatory integrals. Princeton Mathematical Series, 43. Princeton University Press, Princeton, N.J., 1993.
* [Tol] X. Tolsa. Uniform rectifiability, Calderón-Zygmund operators with odd kernel, and quasiorthogonality. Proc. Lond. Math. Soc. (3), 98 (2009), no. 2, 393–426.
|
# Predictive Control and Communication Co-Design via Two-Way Gaussian Process
Regression and AoI-Aware Scheduling
Abanoub M. Girgis, †Jihong Park, Mehdi Bennis, and ‡Mérouane Debbah A
preliminary conference version of this work appeared in the proceedings of
IEEE SPAWC-2020 [1].A. Girgis and M. Bennis are with the Centre for Wireless
Communications, University of Oulu, 90014 Oulu, Finland (e-mail:
<EMAIL_ADDRESS>mehdi.bennis@oulu.fi).†J. Park is with the School of
Information Technology, Deakin University, Geelong, VIC 3220, Australia
(e-mail: jihong.park@deakin.edu.au).‡Mérouane Debbah is with Université Paris-
Saclay, CNRS, CentraleSupélec, 91190, Gif-sur-Yvette, France (e-mail:
<EMAIL_ADDRESS>and the Lagrange Mathematical and
Computing Research Center, 75007, Paris, France.
###### Abstract
This article studies the joint problem of uplink-downlink scheduling and power
allocation for controlling a large number of actuators that upload their
states to remote controllers and download control actions over wireless links.
To overcome the lack of wireless resources, we propose a machine learning-
based solution, where only a fraction of actuators is controlled, while the
rest of the actuators are actuated by locally predicting the missing state
and/or action information using the previous uplink and/or downlink receptions
via a Gaussian process regression (GPR). This GPR prediction credibility is
determined using the age-of-information (AoI) of the latest reception.
Moreover, the successful reception is affected by the transmission power,
mandating a co-design of the communication and control operations. To this
end, we formulate a network-wide minimization problem of the average AoI and
transmission power under communication reliability and control stability
constraints. To solve the problem, we propose a dynamic control algorithm
using the Lyapunov drift-plus-penalty optimization framework. Numerical
results corroborate that the proposed algorithm can stably control $2$x more
number of actuators compared to an event-triggered scheduling baseline with
Kalman filtering and frequency division multiple access, which is $18$x larger
than a round-robin scheduling baseline.
###### Index Terms:
Predictive control, age of information (AoI), Gaussian process regression
(GPR), ultra-reliable and low-latency communication (URLLC), Beyond 5G (B5G),
6G.
## I Introduction
Ultra-reliable and low-latency communication (URLLC) is a key enabler for
ensuring the stability of wirelessly networked control systems in real-time
[2, 3]. By physically decoupling sensors, actuators, and controllers, the
control system can exploit the recent progress in the fifth generation (5G)
connectivity [4], machine learning and edge computing [5], thereby
spearheading many emerging applications ranging from large-scale smart
industrial internet of things (IIoT) [6] to autonomous platooning [7]. The
success of these application relies on addressing several fundamental
challenges emanating from unstable and intermittent wireless connectivity,
which incurs distorted and delayed control information receptions, degrading
control stability. Wireless resource allocation and scheduling are thus
instrumental in not only improving communication efficiency but also in
guaranteeing control stability.
Resource allocation and scheduling of control systems play a key role in the
context of wireless networked control systems (WNCS). Round-robin scheduling,
is a static scheduling in which each sensor/controller periodically transmits
the state/action to a controller/actuator with a fixed transmission power and
a predefined repeating order [8, 9]. The scheduling method in [8, 9] maintains
stability for a small number of control systems, but fails to stabilize a
large number of control systems since the scheduling decisions are not
adaptive to the CSI, which hinder control performance. Dynamic round-robin
scheduling in [10, 11] is a channel-aware scheduling adapting the transmission
power to the CSI to ensure reliability of the transmitted signal and minimize
the system energy. However, it does not guarantee stability for a large number
of control systems, since the waiting period of these control systems to
update the state/action information is proportional to the number of served
control systems. Moreover, the scheduling decisions are not adaptive to the
control state that results in wasting wireless resources. To efficiently
utilize communication resources, the event-triggered scheduling suggested in
[12, 13] is a dynamic scheduling method in which one control system with the
largest state instability is scheduled at each time to transmit the
state/action with fixed transmission power. The previously mentioned
scheduling methods save wireless communication resources since the scheduling
decisions are based on the control state but they are CSI-agnostic which may
result in control performance degradation. In addition, these scheduling
methods can not support a large number of control systems owing to the
outdated control action dependent on the last received state.
The control-aware communication scheduling suggested in [14] incorporates in
the scheduling decisions both control and channel states to minimize the
overall transmission time of the scheduled control systems. Therein, the
scheduling decisions are based on communication reliability that is sensitive
to the control state and system dynamics. However, in the absence of wireless
communication, the outdated applied actions over ideal channels are based on
the last received state, hindering control stability. While the previously
mentioned scheduling methods in [8, 9, 11, 10, 12] are effective in small-
scale control environments, as pointed out in [15], the communication and
control designs in [8, 9, 11, 10, 12] are separated from each other, limiting
their adoption for supporting a large number of control systems. This mandates
a tight co-design of reliable communication and stable control operations for
optimal resource allocation and scheduling.
### I-A Contributions
Spurred by the aforementioned motivation, in this work we consider a scenario
consisting of a large number of control systems connected through wireless
links. To overcome the lack of wireless resources, we propose a GPR based
solution in which only a fraction of control systems is controlled by
communicating with the controller, while the remaining control systems are
controlled by locally predicting the missing state/action via GPR, using the
previous received state/action. For each control system, a controller
calculates the action using a linear quadratic regulator (LQR) whose input,
i.e., either the estimated state using a minimum mean square error (MMSE)
estimator if a sensor-controller pair of a control system is scheduled in the
uplink (UL) or the predicted state using GPR at the controller when it is not
scheduled. Then, to regulate the control system, an actuator applies either
the estimated action using an MMSE estimator if the controller-actuator pair
of a control system is scheduled in the downlink (DL), or the predicted action
using GPR at the actuator if it is not scheduled. The centralized scheduler
shared among all control systems schedules at most one sensor-controller pair
in the UL and one controller-actuator pair in the DL, based on both control
and channel states. Here, control stability depends on the GPR prediction
credibility [16] that is determined by the age of information (AoI) of the
latest received signal, and the reliability of the received signal that is
dictated by transmission power, highlighting the importance of the
communication and control co-design.
Given the proposed predictive control framework, we focus on the problem of
jointly optimizing the UL-DL scheduling and power allocation so as to minimize
the network-wide average AoI and transmission power while guaranteeing
communication reliability and control stability. To solve the formulated non-
convex stochastic optimization problem, we develop a dynamic control algorithm
using Lyapunov optimization. Considering an inverted pendulum, numerical
results demonstrate that the proposed scheduling method can stably support
$2$x more control systems compared to the event-triggered scheduling with
Kalman filtering and frequency division multiple access (FDMA), which is $18$x
larger than a time-triggered scheduling baseline. Furthermore, the results
show that the proposed predictive control algorithm is more communication
efficient while achieving a faster control stability than the time-triggered
and event-triggered control baselines, highlighting the effectiveness of the
UL-DL decoupled scheduling and the use of two-way GPRs at both controller and
actuator sides. The remainder of this paper is organized as follows. In
Section II, we specify the WNCS architecture including the system models of
the control, communication, and GPR based approach. In Section III, we
formulate the communication, control, GPR based co-design optimization problem
and propose the stability-aware scheduling algorithm by leveraging Lyapunov
optimization framework to solve the co-design problem in Section IV. In
Section V and Section VI, we present simulations results and conclude the
paper.
Figure 1: An illustration of $M=2$ WNCSs operated via both state/action
measurement by remote sensors/controllers and state/actions prediction by GPR.
## II System Model
### II-A Wireless Networked Control System Architecture
As depicted in Fig. 1, the WNCS architecture under study consists of a set
$\mathcal{M}$ of $M$ independent linear _control systems_ over a shared
wireless channel. Each control system comprises a _plant_ , a _sensor_ that
measures the plant’s state, and _actuator_ that takes an action to control the
plant’s state. The action is computed by a remote _controller_ based on the
control and channel states. To this end, the plant’s state is received by the
controller in the _uplink_ , and the controller’s action is received by the
actuator in the _downlink_. To avoid interference under limited communication
bandwidth, each UL or DL channel is allocated to a single sensor-controller
pair or controller-actuator pair per unit time, respectively. The rest of the
sensor-controller and controller-actuator pairs without receptions locally
predict their missing control states and actions, respectively, based on their
previously received information, to be detailed in Sec. II-D. To be specific,
the plant’s state of control system $i\in\mathcal{M}$ at discrete control time
$k\in\mathbb{Z}_{+}$ is denoted by $\mathbf{x}^{u}_{i,k}\in\mathbb{R}^{D}$.
For a received action at an actuator $\mathbf{u}^{a}_{i,k}\in\mathbb{R}^{P}$
based on the computed action at controller
$\mathbf{u}^{d}_{i,k}\in\mathbb{R}^{P}$, the state evolution of control system
$i$ at time $k$ is described by the discrete-time linear time-invariant (LTI)
system as follows:
$\mathbf{x}^{u}_{i,k+1}=\mathbf{A}_{i}\mathbf{x}^{u}_{i,k}+\mathbf{B}_{i}\mathbf{u}^{a}_{i,k}+\mathbf{w}_{k},$
(1)
where $\mathbf{A}_{i}\in\mathbb{R}^{D\times D}$ is a fixed state transition
matrix of the $i$-th control system, $\mathbf{B}_{i}\in\mathbb{R}^{D\times P}$
is a fixed control action matrix of the $i$-th control system, and
$\mathbf{w}_{k}\in\mathbb{R}^{D}$ is the plant noise at time $k$ which is
independent and identically distributed (IID) Gaussian noise with zero mean
and covariance matrix $\mathbf{W}$. Here, to avoid a non-trivial problem,
$\mathbf{A}_{i}$ is assumed to be unstable, i.e., $\mathbf{A}_{i}$’s spectral
radius
$\rho(\mathbf{A}_{i})=\text{max}\\{|\lambda_{1}(\mathbf{A}_{i})|,\cdots,|\lambda_{D}(\mathbf{A}_{i})|\\}$
is larger than unity, where $\lambda_{D}(\mathbf{A}_{i})$ is the $D$-th
eigenvalue of $\mathbf{A}_{i}$. This implies that the plant’s state infinitely
grows over time unless a proper control action $\mathbf{u}^{a}_{i,k}$ is
provided. To stabilize such control system, each time $k$, the following four
phase operations are considered.
1. 1.
Sensing and Uplink Transmission (at a sensor): A centralized scheduler located
at the base station (BS) shared among all control systems decides which
sensor-controller pair is scheduled to transmit and close its sensing loop
based on both the channel and control states. Then, the scheduled sensor
transmits its state to its controller over a wireless UL fading channel using
analog uncoded communication to be elaborated in Sec. II-B.
2. 2.
State Reception or Prediction (at a controller): If the sensor-controller pair
is scheduled, the controller obtains the current estimated state using MMSE
estimator, and predicts the next state via GPR to be discussed in Sec. II-D.
Otherwise, the controller directly predicts the current and the next states
based on the state history using GPR. The current predicted state by GPR is
fed to the LQR to calculate the action unless the estimated state by the MMSE
estimator is provided. The future predicted state is fed to the centralized
scheduler to make the scheduling decisions.
3. 3.
Action Computation and Downlink Transmission (at a controller): For a given
control state, the controller computes the optimal action using LQR [17]. The
controller transmits the computed action to the scheduled actuator over a
wireless DL fading channel using analog uncoded communication to be elaborated
in Sec. II-B.
4. 4.
Action Reception or Prediction (at an actuator): If the controller-actuator
pair is scheduled, the actuator obtains the current estimated action using a
MMSE estimator, and predicts the next action via an action GPR to be discussed
in Sec. II-D. Otherwise, the actuator directly predicts the current and next
actions based on the action history using GPR. For a given action, the
actuator takes an action and subsequently the plant’s state is updated
according to the dynamics in (1).
### II-B State and Action Communications
The UL state and DL action communications are elaborated, in terms of the
received signal, signal-to-noise ratio (SNR), scheduling, and age-of-
information (AoI) as follows.
Noisy State and Action Receptions: At the control time slot $k$, the received
signal $\mathbf{y}^{l}_{i,k}$ at the $l$-th communication at the receiver of
control system $i$ is represented as
$\displaystyle\mathbf{y}^{l}_{i,k}=\sqrt{P^{l}_{i,k}}\mathbf{C}_{i}\mathbf{H}^{l}_{i,k}\mathbf{q}^{l}_{i,k}+\mathbf{n}^{l}_{k},$
(2)
where $l\in\\{u,d\\}$ represents a communication indicator between the
transmitter-receiver pair in which $l=u$ refers to the UL state communication
between the sensor-controller pair while $l=d$ refers to the DL action
communication between the controller-actuator pair. The transmitted signal, in
the UL state communication (i.e., $\mathbf{q}=\mathbf{x}$ and $l=u$),
$\mathbf{x}^{u}_{i,k}=[x^{u}_{i,k}(1)\cdots x^{u}_{i,k}(D)]$ is the plant’s
state transmitted by the sensor of a control system $i$ at time $k$ such that
$\mathbb{E}\\{|x^{u}_{i,k}(\iota)|^{2}\\}=1,\forall\iota\in\\{1,\cdots,D\\}$.
The transmitted signal, in the DL action communication (i.e.,
$\mathbf{q}=\mathbf{u}$, $l=d$), $\mathbf{u}^{d}_{i,k}=[u^{d}_{i,k}(1)\cdots
u^{d}_{i,k}(P)]$ is the action transmitted by the controller to an actuator of
a control system $i$ at time $k$ such that
$\mathbb{E}\\{|u^{d}_{i,k}(p)|^{2}\\}=1,\forall p\in\\{1,\cdots,P\\}$. The
matrix $\mathbf{H}^{l}_{i,k}\in\mathbb{R}^{\mathcal{F}\times\mathcal{F}}$
represents the wireless channel of the $l$-th communication between the
transmitter-receiver pair of a control system $i$ at time $k$, and
$\mathcal{F}\in\\{D,P\\}$ represents the dimensions of the transmitted state
or action, respectively. The channel is modeled as a Rayleigh block fading
which is static and flat-fading within either UL or DL transmission time.
Channel statistics are perfectly known at the transmitters and receivers, and
$P^{l}_{i,k}\in\left[0,P^{l}_{max}\right]$ is the transmission power of
control system $i$ at time $k$ with total transmission power $P^{l}_{max}$.
Lastly, $\mathbf{n}^{l}_{k}$ is the additive white Gaussian noise at the
receiver with zero-mean and covariance matrix
$\mathbb{E}\\{\mathbf{n}_{k}^{l^{T}}\mathbf{n}^{l}_{k}\\}=N_{0}\mathbf{I}_{\mathcal{F}}$,
where $N_{0}$ is the measurement noise variance, and
$\mathbf{I}_{\mathcal{F}}$ is the $\mathcal{F}\times\mathcal{F}$ identity
matrix. The matrix
$\mathbf{C}_{i}\in\mathbb{R}^{\mathcal{F}\times\mathcal{F}}$ is the
observation matrix of the control system $i$ that equals the identity matrix
in the DL action communication and equals the identity matrix in the UL state
communication to characterize the full-state observations. The measurement
noise $\mathbf{n}^{l}_{k}$ and the plant’s noise $\mathbf{w}_{k}$ in (1) are
uncorrelated zero-mean Gaussian noise with unit variance. Hence, the SNR at
the receiver of the $l$-th communication of a control system $i$ at time $k$
is given as
$\displaystyle\text{SNR}^{l}_{i,k}=\frac{P^{l}_{i,k}\|\mathbf{H}^{l}_{i,k}\|^{2}}{N_{0}},$
(3)
where SNR in (3) is equivalent to the signal-to-distortion ratio (SDR) as a
result of the channel theoretical limit in which the rate-distortion function
of the source equals to the channel capacity and the number of source samples
is matched to the number of channels under analog uncoded communications [18].
In this work, we consider analog uncoded communications, in which the
discrete-time continuous amplitude source samples are amplified and
transmitted to a receiver over wireless channel. Compared to digital
communications, analog communications are favorable for achieving low latency
thanks to skipping encoding and decoding operations, at the cost of signal
distortion during channel propagation [18]. Furthermore, analog uncoded
communication is robust to channel conditions and performs well at different
SNRs compared to digital communication that is sensitive to any degradation in
channel conditions. To ensure reliable communication for control stability,
the successful decoding of the transmitted signal is described by the
indicator function
$\mathbb{I}_{\\{\text{SNR}^{l}_{i,k}\geq\text{SNR}^{l}_{\text{th}}\\}}$ for a
target SNR threshold $\text{SNR}^{l}_{\text{th}}$.
Scheduling and AoI: At time $k$, the centralized scheduler located at the BS
and shared among all control systems schedules at most one sensor-controller
pair of control system $i$ in the UL state communications, and at most one
controller-actuator pair of the control system $i$ in the DL action
communications. Let $\alpha^{l}_{i,k}\in\\{0,1\\}$ be the scheduling variable
of the $l$-th communication of the control system $i$ at time $k$, where
$\alpha^{l}_{i,k}=1$ when the transmitter-receiver pair of the $l$-th
communication of control system $i$ is scheduled at time $k$ and
$\alpha^{l}_{i,k}=0$ otherwise.
The freshness of the received information is measured using AoI, i.e., the
number of elapsed time since the generation of the latest received information
[19]. AoI is composed of the inter-arrival time that is defined as the time
elapsed between two consecutive update generations and the service time
defined as the transmission time of update information. In analog uncoded
communications, AoI depends only on the inter-arrival time since the service
time is deterministic based on channel bandwidth. Hence, the AoI of the $l$-th
communication of the control system $i$ at the receiver linearly increases
with time if it is not scheduled or its SNR is below a threshold. Formally,
the AoI of the $l$-th communication of control system $i$ at the receiver is:
$\displaystyle\beta^{l}_{i,k}=\left\\{\begin{array}[]{ll}1+\beta^{l}_{i,k-1},&\text{if}\;\xi^{l}_{i,k}=0,\\\
1,&\text{o.w.s.},\end{array}\right.$ (6)
where $\beta^{l}_{i,k}\in\mathbb{Z}_{++}$ is the AoI of the $l$-th
communication of the control system $i$ at time $k$ at the receiver, and
$\xi^{l}_{i,k}=\alpha^{l}_{i,k}\,\mathbb{I}_{\\{\text{SNR}^{l}_{i,k}\geq\text{SNR}^{l}_{th}\\}}$
is the transmission indicator variable of $l$-th communication of control
system $i$ at time $k$ that depends on both scheduling variable and SNR
indicator function.
### II-C State and Action Estimation Over Noisy Communication Links
The UL received states and DL received actions are distorted by Rayleigh
fading channels. The original signals are estimated using the MMSE estimator
as detailed next. When one transmitter-receiver pair of the $l$-th
communication of control system $i$ is scheduled, i.e., $\alpha_{i,k}^{l}=1$,
the receiver applies the MMSE estimator to restore the original signal from
the noisy received signal in (2). The resultant estimated signal
$\bar{\mathbf{q}}^{l}_{i,k}$ is given as:
$\displaystyle\bar{\mathbf{q}}^{l}_{i,k}=\mathbb{E}\\{\mathbf{q}^{l}_{i,k}|\mathbf{y}^{l}_{i,k}\\}=\mathbf{G}^{l}_{i,k}\mathbf{y}^{l}_{i,k}=\mathbf{q}^{l}_{i,k}+\mathbf{v}^{l}_{i,k},$
(7)
where $\mathbf{G}^{l}_{i,k}\in\mathbb{R}^{\mathcal{F}\times\mathcal{F}}$ is
the linear MMSE matrix at the receiver of the $l$-th communication of the
control system $i$ at time $k$ that minimizes the mean-squared error (MSE)
between the original and estimated signals as
$\displaystyle\mathbf{G}^{l}_{i,k}=\sqrt{P^{l}_{i,k}}\mathcal{S}_{q}\mathbf{H}^{l^{T}}_{i,k}\left(P^{l}_{i,k}\mathbf{H}^{l}_{i,k}\mathcal{S}_{q}\mathbf{H}^{l^{T}}_{i,k}+N_{0}\mathbf{I}_{\mathcal{F}}\right)^{-1}$
(8)
The term $\mathbf{v}^{l}_{i,k}$ in (7) is the MMSE estimation error following
a zero-mean Gaussian random vector with the covariance matrix
$\mathbf{V}^{l}_{i,k}\in\mathbb{R}^{\mathcal{F}\times\mathcal{F}}$. Following
[20], we assume that $\mathbf{q}^{l}_{i,k}$ follows a zero-mean Gaussian
distribution with the covariance matrix
$\mathcal{S}_{q}\in\mathbb{R}^{\mathcal{F}\times\mathcal{F}}$, then we have
$\displaystyle\mathbf{V}^{l}_{i,k}$
$\displaystyle=\mathbb{E}\\{\mathbf{v}^{l}_{i,k}\mathbf{v}_{i,k}^{l^{T}}\\}=\mathbb{E}\Big{\\{}\left(\bar{\mathbf{q}}^{l}_{i,k}-\mathbf{q}^{l}_{i,k}\right)\left(\bar{\mathbf{q}}^{l}_{i,k}-\mathbf{q}^{l}_{i,k}\right)^{T}\Big{\\}}=\mathcal{S}_{q}-\mathbf{G}^{l}_{i,k}\sqrt{P^{l}_{i,k}}\mathbf{H}^{l}_{i,k}\mathcal{S}_{q}$
(9)
### II-D State and Action Prediction Without Communication
When one transmitter-receiver pair of the $l$-th communication of the control
system $i$ at time $k$ is not scheduled, i.e., $\alpha^{l}_{i,k}=0$, a
receiver applies a number of parallel GPRs proportional to the missing signal
dimensions to predict both the missing current signal and the next signal
using the previously received signals. Each individual GPR learns the
functional relationship $g\in\mathbb{R}$ between the control discrete-time
$k^{\prime}\in\mathbb{Z}_{+}$ and each output of the received signal. This
means that each output of the MMSE estimated signal
$\bar{\mathbf{q}}^{l}_{i,k^{\prime}}$ in (7) is the state observation
$\bar{\mathbf{x}}^{u}_{i,k^{\prime}}\in\mathbb{R}^{D}$ in the UL and the
action $\bar{\mathbf{u}}^{d}_{i,k^{\prime}}\in\mathbb{R}^{P}$ in the DL. This
is accomplished by learning a latent function of the following regression
model
$\bar{q}^{l}_{i,k^{\prime}}(j)=g_{j}(k^{\prime})+\epsilon,\;j\in\\{1,\cdots,\mathcal{F}\\},\;\forall
i,\,l,\,k^{\prime}$, where $\bar{q}^{l}_{i,k^{\prime}}(j)\in\mathbb{R}$ is the
$j$-th output of the estimated signal of the $l$-th communication of the
control system $i$ at time $k^{\prime}$, $g_{j}$ is a $j$-th output latent
function, and $\epsilon\sim\mathcal{N}\left(0,\sigma^{2}_{n}\right)$ is an IID
Gaussian noise distribution with zero mean and variance $\sigma^{2}_{n}$ that
accounts for the measurements or modeling errors [21]. Specifically, to
predict the missing received signal $\bar{\mathbf{q}}^{l}_{i,k}$ of the $l$-th
communication of the control system $i$ at test time $k$, we exploit
$\mathcal{F}$ individual GPRs, where $\mathcal{F}$ represents the MMSE
estimated signal dimensions, and feeding each individual GPR with a training
set $\mathcal{D}^{l,j}_{i,n_{l}}$ of each output of the previous received
signals associated with its observation time $k^{\prime}$, given as
$\mathcal{D}^{l,j}_{i,n_{l}}=\\{(k^{\prime},\xi^{l}_{i,k^{\prime}}\,\bar{q}^{l}_{i,k^{\prime}}(j))|\,j=1,\cdots,\mathcal{F},\;k^{\prime}=1,\cdots,n_{l},\;i=1,\cdots,M,\;l\in\\{u,d\\}\\}$.
Here, $n_{l}=\sum_{k^{\prime}}\xi^{l}_{i,k^{\prime}}$ counts the number of
received signals of the $l$-th communication until time $k^{\prime}$ in the
training set of the control system $i$. Hence, the last time instant in which
the transmitter of the $l$-th communication of the control system transmitted
its observation to the receiver is given as
$\tilde{n}_{l}=k^{\prime}-\beta^{l}_{i,k^{\prime}}+1$. It is obvious that a
large value of AoI decreases the number of observations at the receiver that
affects the signal prediction credibility at a particular level.
In each individual GPR, according to the Gaussian process (GP) characteristics
where any finite subset of random variables taken from a realization of a GP
follows a joint Gaussian distribution, each $j$-th output latent function
$g_{j}$ of the vector-valued latent function
$\mathbf{g}(k)=\left[g_{1}(k)\cdots g_{\mathcal{F}}(k)\right]$ is assumed to
follow a GP as
$g_{j}\left(k\right)\sim\mathcal{GP}\left(m_{j}\left(k\right),\mathcal{R}_{j}(k,k^{\prime})\right)$,
where $m_{j}\left(k\right)$ is the mean function of the $j$-th output of the
missing received signal which is usually taken as zero without loss of
generality [22], and $\mathcal{R}_{j}(k,k^{\prime})$ is the covariance
function of the $j$-th output of the missing received signal between the
outputs at time $k$ and $k^{\prime}$ that defines the correlation between the
outputs according to the input times. It is noted that the stationary
covariance function between the outputs is based on the difference between
their corresponding input times $|k-k^{\prime}|$ in which the two outputs are
strongly correlated if their corresponding input times are sufficiently close
to each other. Since we focus on time-series data, we utilize information from
previously received signals to describe the current data depending on the past
observations. Hence, we use a squared exponential kernel function coupled with
a periodic kernel function, to model the correlation between the outputs
according to their temporal behaviours, as defined in [22]
$\displaystyle\mathcal{R}(k,k^{\prime})=h_{q}^{2}\exp\left[\frac{-\left(k-k^{\prime}\right)^{2}}{2h^{2}_{k}}\right]+\exp\left\\{-2\sin^{2}\left[\nu\pi\left(k-k^{\prime}\right)\right]\right\\},$
(10)
where the first term represents the stationary covariance function that
depends on when the signal $|k-k^{\prime}|$ was received with $h_{k}$ and
$h_{q}$ being hyperparameters representing the time-scaling and output-scaling
of a squared exponential function, respectively, and the second term gives the
periodicity with hyperparameter $\nu$ representing frequency. For a set of
j-th output observations
$\bar{\mathbf{q}}^{l}_{i}(j)=\\{\bar{q}^{l}_{i,1}(j),\cdots,\bar{q}^{l}_{i,n_{l}}(j)\\}^{T}$
and the associated observation times
$\mathbf{k}^{\prime}=\\{1,\cdots,n_{l}\\}^{T}$, the joint distribution of the
$j$-th output past observations $\bar{\mathbf{q}}^{l}_{i,k^{\prime}}(j)$
together with the $j$-th output $g_{j}(k)$ at test time $k$ is given as
$\displaystyle\left[\begin{array}[]{c}\bar{\mathbf{q}}^{l}_{i}(j)\\\
g_{j}(k)\end{array}\right]\sim\mathcal{N}\left(\left[\begin{array}[]{c}\mathbf{0}\\\
0\end{array}\right],\left[\begin{array}[]{cc}\mathbf{R}_{j}(\mathbf{k}^{\prime},\mathbf{k}^{\prime})&\mathbf{r}_{j}(\mathbf{k}^{\prime},k)\\\
\mathbf{r}_{j}(k,\mathbf{k}^{\prime})&\mathcal{R}_{j}(k,k)\end{array}\right]\right),$
(17)
where $\mathcal{R}_{j}\left(k,k\right)\in\mathbb{R}$ is the prior covariance
function of $j$-th output observation at a test time $k$, and
$\mathbf{R}_{j}(\mathbf{k}^{\prime},\mathbf{k}^{\prime})\in\mathbb{R}^{n_{l}\times
n_{l}}$ is the symmetric and positive semi-definite covariance matrix of
$j$-th output past observations with the elements
$\mathcal{R}_{j}\left({\mathbf{k}^{\prime}(a),\mathbf{k}^{\prime}(b)}\right)$
for $a,b=1,\cdots,n_{l}$. Following [21], we treat the prediction mean as the
$j$-th output predicted signal $\hat{q}^{l}_{i,k}(j)$, the posterior
distribution of $g_{j}\left(k\right)$ at test time $k$ based on the training
set $\mathcal{D}^{l,j}_{i,n_{l}}$ can be analytically derived as
$\displaystyle\text{Pr}\left(g_{j}\left(k\right)|\mathcal{D}^{l,j}_{i,n_{l}},k,\mathbf{\Theta}_{j}\right)$
$\displaystyle\sim\mathcal{N}\left(\hat{q}^{l}_{i,k}(j),\sigma^{2}_{i,k}(j)\right).$
(18)
Following [21], the $j$-th output prediction mean
$\hat{q}^{l}_{i,k}\left(j\right)$, and the $j$-th output prediction variance
$\sigma^{2}_{i,k}(j)$ are respectively given as
$\displaystyle\hat{q}^{l}_{i,k}(j)$
$\displaystyle=\mathbf{r}_{j}(k,\mathbf{k}^{\prime})^{T}\mathbf{R}_{j}(\mathbf{k}^{\prime},\mathbf{k}^{\prime})^{-1}\bar{\mathbf{q}}^{l}_{i}(j)=q^{l}_{i,k}(j)+e^{l}_{i,k}(j),$
(19) $\displaystyle\sigma^{2}_{i,k}(j)$
$\displaystyle=\mathbb{E}\Big{\\{}\left(\hat{q}^{l}_{i,k}(j)-q^{l}_{i,k}(j)\right)\left(\hat{q}^{l}_{i,k}(j)-q^{l}_{i,k}(j)\right)^{T}\Big{\\}}=\mathbb{E}\\{e^{l}_{i,k}(j)e^{l^{T}}_{i,k}(j)\\}$
(20)
$\displaystyle=\mathcal{R}_{j}(k,k)-\mathbf{r}_{j}(k,\mathbf{k}^{\prime})\mathbf{R}_{j}(\mathbf{k}^{\prime},\mathbf{k}^{\prime})^{-1}\mathbf{r}_{j}(k,\mathbf{k}^{\prime})^{T},$
where $\mathbf{r}_{j}(\mathbf{k}^{\prime},k)\in\mathbb{R}^{n_{l}\times 1}$ is
$j$-th output observation covariance between the outputs at the $n_{l}$
observation times and a test time $k$, and the term $e^{l}_{i,k}(j)$ is $j$-th
output prediction error defined as the difference between true and predicted
outputs. Moreover, $\mathbf{\Theta}_{j}$ in (18) is the $j$-th output
hyperparameters of the covariance function $\mathcal{R}$. Finally, the
predicted signal at the receiver of the $l$-th communication of control system
$i$ at time $k$ and its prediction error covariance matrix are
$\displaystyle\hat{\mathbf{q}}^{l}_{i,k}$
$\displaystyle=\\{\hat{q}^{l}_{i,k}(1)\cdots\hat{q}^{l}_{i,k}(\mathcal{F})\\}^{T}=\mathbf{q}^{l}_{i,k}+\mathbf{e}^{l}_{i,k},$
(21) $\displaystyle\mathbf{\mathcal{J}}^{l}_{i,k}$
$\displaystyle=\left[\begin{array}[]{ccc}\sigma^{2}_{i,k}(1)&\cdots&0\\\
\vdots&\ddots&\vdots\\\
0&\cdots&\sigma^{2}_{i,k}(\mathcal{F})\end{array}\right].$ (25)
### II-E Action Computation and Actuation
By feeding the estimated or predicted state, the controller computes the
action using LQR. Then, the actuator applies the estimated or predicted action
to stabilize the state as detailed next.
Action Computation After State Estimation/Prediction: For a given estimated
state (i.e., MMSE output) in (7) or predicted state (i.e., GPR output) in
(21), at time $k$, the state $\mathbf{x}_{i,k}^{c}$ available at the
controller based on the UL transmission indicator variable is given as
$\displaystyle\mathbf{x}_{i,k}^{c}=\xi^{u}_{i,k}\bar{\mathbf{x}}^{u}_{i,k}+\left(1-\xi^{u}_{i,k}\right)\hat{\mathbf{x}}^{u}_{i,k}.$
(26)
The received state $\mathbf{x}_{i,k}^{c}$ is used in the LQR located at the
controller, and the optimal action of a control system $i$ at time $k$ is
given by the following linear feedback control law as
$\displaystyle\mathbf{u}^{d}_{i,k}=-\mathbf{\Phi}_{i}\mathbf{x}^{c}_{i,k},$
(27)
where $\mathbf{u}^{d}_{i,k}\in\mathbb{R}^{P}$ is the computed action at the
controller,
$\mathbf{\Phi}_{i}=\left(\mathbf{Z}^{u}+\mathbf{B}^{T}_{i}\mathbf{P}\mathbf{B}_{i}\right)^{-1}\mathbf{B}_{i}^{T}\mathbf{P}\mathbf{A}_{i}$
is the feedback gain matrix of the control system $i$,
$\mathbf{Z}^{s}\in\mathbb{S}_{+}^{D\times D}$ is a positive semi-definite
weight matrix of the state deviation cost, and
$\mathbf{Z}^{u}\in\mathbb{S}_{++}^{P\times P}$ is a positive definite weight
matrix of the action cost. The term
$\mathbf{P}=\mathbf{A}_{i}^{T}\mathbf{P}\mathbf{A}_{i}-\mathbf{A}_{i}^{T}\mathbf{P}\mathbf{B}_{i}\left(\mathbf{B}^{T}_{i}\mathbf{P}\mathbf{B}_{i}+\mathbf{Z}^{u}\right)^{-1}\mathbf{B}_{i}^{T}\mathbf{P}\mathbf{A}_{i}+\mathbf{Z}^{s}$
is the unique positive definite matrix which satisfies the discrete-time
algebraic Riccati equation (DARE). Then, the controller transmits the computed
action $\mathbf{u}^{d}_{i,k}$ in (27) to an actuator of control system $i$ at
time $k$ in the DL, if $\xi_{i,k}^{d}=1$, as discussed in Sec. II-B.
Actuation After Action Estimation/Prediction: For a given estimated control
action (i.e., MMSE output) in (7) or predicted action (i.e., GPR output) in
(21), at time $k$, the action $\mathbf{u}_{i,k}^{a}$ available at the actuator
based on the DL transmission indicator variable is given as
$\displaystyle\mathbf{u}_{i,k}^{a}=\xi^{d}_{i,k}\,\bar{\mathbf{u}}^{d}_{i,k}+\left(1-\xi^{d}_{i,k}\right)\hat{\mathbf{u}}^{d}_{i,k}.$
(28)
Note that the UL and DL transmission indicator variables are periodically
generated by the centralized scheduler in which, within each unit control time
duration, the UL state communication can be firstly activated for sensing the
plant’s state based on the UL transmission indicator variable. Then, the DL
action communication can be activated for actuation based on the DL
transmission indicator variable. Consequently, for a given pair of UL and DL
transmission indicator variables with (26) and (28), the actuator takes a
control action that changes the plant’s state of the control system $i$ at
time $k$ in (1) into four cases of state evolution as follows:
$\displaystyle\mathbf{x}^{o}_{i,k+1}=\mathbf{A}_{i}\mathbf{x}^{u}_{i,k}-\mathbf{B}_{i}(\mathbf{\Phi}_{i}\hat{\mathbf{x}}^{u}_{i,k}+\mathbf{e}^{d}_{i,k})+\mathbf{w}_{k},\;$
$\displaystyle\text{if}\;\xi^{u}_{i,k}=0\;\text{and}\;\xi^{d}_{i,k}=0,\;\text{
(open-loop)},$ (29)
$\displaystyle\mathbf{x}^{s}_{i,k+1}=\mathbf{A}_{i}\mathbf{x}^{u}_{i,k}-\mathbf{B}_{i}(\mathbf{\Phi}_{i}\bar{\mathbf{x}}^{u}_{i,k}+\mathbf{e}^{d}_{i,k})+\mathbf{w}_{k},\;$
$\displaystyle\text{if}\;\xi^{u}_{i,k}=1\;\text{and}\;\xi^{d}_{i,k}=0,\;\text{
(sensing-loop)},$ (30)
$\displaystyle\mathbf{x}^{a}_{i,k+1}=\mathbf{A}_{i}\mathbf{x}^{u}_{i,k}-\mathbf{B}_{i}(\mathbf{\Phi}_{i}\hat{\mathbf{x}}^{u}_{i,k}+\mathbf{v}^{d}_{i,k})+\mathbf{w}_{k},\;$
$\displaystyle\text{if}\;\xi^{u}_{i,k}=0\;\text{and}\;\xi^{d}_{i,k}=1,\;\text{
(actuating-loop)},$ (31)
$\displaystyle\mathbf{x}^{c}_{i,k+1}=\mathbf{A}_{i}\mathbf{x}^{u}_{i,k}-\mathbf{B}_{i}(\mathbf{\Phi}_{i}\bar{\mathbf{x}}^{u}_{i,k}+\mathbf{v}^{d}_{i,k})+\mathbf{w}_{k},\;$
$\displaystyle\text{if}\;\xi^{u}_{i,k}=1\;\text{and}\;\xi^{d}_{i,k}=1,\;\text{
(closed-loop)}.$ (32)
Figure 2: Timing diagram of the control system. First diagram illustrates the
UL and DL transmission indicator variables generated by scheduler, second
diagram illustrates uniform sampling by a sensor, third diagram illustrates
the received/predicted state and calculated action at a controller. Fourth
diagram illustrates the received/predicted action at an actuator.
Timing diagram: Based on the UL and DL transmission indicator variables within
each unit control time duration, the timing diagram of a control system is
illustrated in Fig. 2. The centralized scheduler shared among all control
systems, within each unit control time duration, primarily transmits the UL
and DL transmission indicator variables to the sensor, controller, and
actuator nodes of all control systems. Then, the state is only transmitted by
the sensor if the control system has a reliable UL communication and has
valuable information affecting the control stability (i.e, $\xi^{u}_{i,k}=1$)
which result in saving wireless communication resources. After that, LQR
located at the controller computes the action based on the state available at
the controller in (26) in which the predicted state, if $\xi^{u}_{i,k}=0$, is
applied to the LQR. Lastly, the action is transmitted by the controller if has
a reliable DL communication and valuable information affecting the control
stability (i.e, $\xi^{d}_{i,k}=1$). This is a result of assuming that the UL
and DL transmission indicator variables as being periodically transmitted by
the centralized scheduler every control time duration, the controller
periodically calculates the action depending on the available state, and the
actuator periodically applies the action depending on the available action.
Moreover the discrete-time control time $k$ equals the continuous-time control
time duration unit $\Delta k$ comprising the UL and DL transmission times
while ignoring the computational delay.
## III Communication Control Co-design
### III-A Control-constrained Problem Formulation
Our objective is to minimize the total communication cost per control system
subject to ensuring communication reliability and control stability. The total
communication cost incorporates the AoI and transmission power since the AoI
indirectly affects the control stability through the GPR prediction stability
and wireless resources consumption, while transmit power affects communication
reliability and energy consumption. Formally speaking, we have:
$\displaystyle\mathcal{C}\left(\\{\bar{\beta}^{l}_{i}\\},\\{\bar{\hat{P}}^{l}_{i}\\}\right)=\omega_{\beta_{l}}\sum_{i=1}^{M}\mathcal{G}_{\beta}(\bar{\beta}^{l}_{i})+\omega_{P_{l}}\sum_{i=1}^{M}\mathcal{G}_{P}(\bar{\hat{P}}^{l}_{i}),\qquad\forall
l\in\\{u,d\\},$ (33)
where the non-decreasing concave functions
$\mathcal{G}_{\beta}(\beta)=\log(1+\beta)$ and
$\mathcal{G}_{P}(\hat{P})=\log(1+\hat{P})$ are proportionally fair cost
functions of the AoI and the transmission power function for each control
system, respectively [23]. The transmission power function that depends on the
scheduling variable is given as
$\hat{P}^{l}_{i,k}=\alpha^{l}_{i,k}P^{l}_{i,k}$, and the given positive
weights $\omega_{\beta_{l}}$ and $\omega_{P_{l}}$ adjust the relative
importance of the corresponding cost functions. Throughout this work, the
following notation for the long-term time-averaged of any quantity $z$ is
defined as
$\bar{z}\triangleq\underset{K\to\infty}{\lim\sup}\frac{1}{K}\sum_{k=1}^{K}z$.
In particular, $\bar{\beta}^{l}_{i}$ and $\bar{\hat{P}}^{l}_{i}$ are the long-
term time-averaged of $\beta^{l}_{i}$ and $\hat{P}^{l}_{i}$, respectively.
To evaluate control stability, we consider the quadratic Lyapunov function
that measures the performance of each control system as a function of the
state expressed as
$\displaystyle\mathcal{L}(\mathbf{x}^{u}_{i,k})=\mathbf{x}^{u^{T}}_{i,k}\,\mathcal{Z}\,\mathbf{x}^{u}_{i,k},\qquad\forall\mathcal{Z}\in\mathbb{S}^{D}_{++},$
(34)
where $\mathcal{Z}\in\mathbb{S}^{D}_{++}$ is a unique positive definite
solution to the discrete Lyapunov equation
$\mathbf{A}^{c^{T}}_{i}\mathcal{Z}+\mathcal{Z}\mathbf{A}^{c}_{i}=-\mathbb{I}_{D}$,
and $\mathbf{A}^{c}_{i}$ is a closed-loop state transition matrix defined as
$\mathbf{A}^{c}_{i}=\mathbf{A}_{i}-\mathbf{B}_{i}\mathbf{\Phi}_{i}$. Because
the centralized scheduler has only access to the predicted state, the expected
current value of $\mathcal{L}(\mathbf{x}^{u}_{i,k})$ is calculated in the
following lemma.
###### Lemma 1.
Given the predicted state $\hat{\mathbf{x}}^{u}_{i,k}$ and the state
prediction error covariance matrix $\mathcal{J}^{u}_{i,k}$ at the controller,
the expected current value of $\mathcal{L}(\mathbf{x}^{u}_{i,k})$ is given as
$\displaystyle\mathbb{E}\left[\mathcal{L}(\mathbf{x}^{u}_{i,k})|\hat{\mathbf{x}}^{u}_{i,k}\right]=\|\hat{\mathbf{x}}^{u}_{i,k}\|_{\mathcal{Z}^{\frac{1}{2}}}^{2}+\text{Tr}\left[\mathcal{Z}\,\mathcal{J}^{u}_{i,k}\right],\qquad\forall\;\mathcal{Z}\in\mathbb{S}^{D}_{++}.$
(35)
###### Proof.
Please refer to Appendix.A ∎
Note that the expected current value of $\mathcal{L}(\mathbf{x}^{u}_{i,k})$ of
the control system $i$ at time $k$ naturally grows as the predicted state and
the prediction error get larger as a result of increasing AoI and/or the
insufficiency of received observations number in the training set. The
requirement of control stability is that the expected future value of
$\mathcal{L}(\mathbf{x}^{u}_{i,k+1})$ should decrease at a given rate
$\zeta_{i}\in(0,1]$ of its expected current value of
$\mathcal{L}(\mathbf{x}^{u}_{i,k})$, which means the state of the control
system is monotonically decreasing along trajectories, as
$\displaystyle\mathbb{E}\left[\mathcal{L}(\mathbf{x}^{u}_{i,k+1})|\hat{\mathbf{x}}^{u}_{i,k},\hat{\mathbf{u}}^{d}_{i,k},\mathbf{H}^{u}_{i,k},\mathbf{H}^{d}_{i,k},P^{u}_{i,k},P^{d}_{i,k}\right]\leq\zeta_{i}\mathbb{E}\left[\mathcal{L}(\mathbf{x}^{u}_{i,k})|\hat{\mathbf{x}}^{u}_{i,k}\right],$
(36)
where the expectation in the right hand side of (36) is with respect to the
plant noise $\mathbf{w}_{k}$ in (1), the signal estimation error
$\mathbf{v}^{l}_{i,k}$ defined in (9), and the signal prediction error
$\mathbf{e}^{l}_{i,k}$ defined in (21). According to the objective function in
(33) and the control stability constraint in (36), the control-constrained
optimization problem can be formulated as follows:
$\displaystyle\underset{\mathbf{a}_{k}^{l},\mathbf{P}_{k}^{l}}{(\mathcal{P}1)\quad\text{Minimize}}\mathcal{C}\left(\\{\bar{\beta}^{l}_{i}\\},\\{\bar{\hat{P}}^{l}_{i}\\}\right)$
(37a) $\displaystyle\hskip 40.0pt\text{subject to:}\quad 0\leq P^{l}_{i,k}\leq
P^{l}_{max},\qquad\qquad\qquad\qquad\quad\;\;\forall\,l\in\\{u,d\\},i\in\mathcal{M},\,k,$
(37b) $\displaystyle\hskip
95.0pt\|\mathbf{H}^{l}_{i,k}\|^{2}\,P^{l}_{i,k}/N_{0}\geq\text{SNR}^{l}_{th},\qquad\qquad\quad\;\;\;\forall\,l\in\\{u,d\\},i\in\mathcal{M},\,k,$
(37c) $\displaystyle\hskip
95.0pt\alpha^{l}_{i,k}\in\\{0,1\\},\qquad\qquad\qquad\qquad\qquad\quad\;\forall\,l\in\\{u,d\\},i\in\mathcal{M},\,k,$
(37d) $\displaystyle\hskip 95.0pt\sum^{M}_{i=1}\alpha^{l}_{i,k}\leq
1,\qquad\qquad\qquad\qquad\qquad\quad\;\forall\,l\in\\{u,d\\},i\in\mathcal{M},\,k,$
(37e) $\displaystyle\hskip
50.0pt\eqref{Future_Lyapunov},\qquad\qquad\qquad\qquad\qquad\qquad\qquad\quad\forall\;i\in\mathcal{M},\,k,$
where $\mathbf{a}_{k}^{l}=\\{\alpha^{l}_{i,k}:\forall
l\in\\{u,d\\},\;i\in\mathcal{M}\\}$ and
$\mathbf{P}_{k}^{l}=\\{P^{l}_{i,k}:\forall l\in\\{u,d\\},\;i\in\mathcal{M}\\}$
are the UL-DL scheduling vector at time $k$, and the UL-DL transmission power
vector at time $k$, respectively. The constraint in (37b) bounds the UL-DL
transmission power allocation of a control system $i$ at time $k$ by the total
available transmission power $P^{l}_{max}$, while the constraint in (37c)
ensures the communication reliability that is based on SNR or SDR in analog
uncoded communications. The constraints in (37d)-(37e) ensure at most one
transmitter-receiver pair of a control system $i$ is scheduled at time $k$.
The constraint in (36) ensures the state is decreasing along trajectories that
satisfies the control stability. It is noted that control stability constraint
in (36) is independent with communication constraints in (37b)-(37e). However,
the control stability constraint is affected and determined by the
communication and scheduling variables, hence the original control-constrained
problem $\mathcal{P}1$ is rewritten after directly reflecting the
communication control relationship of the constraint (36) in the following
lemma.
###### Lemma 2.
Given predicted state $\hat{\mathbf{x}}^{u}_{i,k}$, state prediction error
covariance matrix $\mathcal{J}^{u}_{i,k}$, the predicted action
$\hat{\mathbf{u}}^{d}_{i,k}$, the action prediction error covariance matrix
$\mathcal{J}^{d}_{i,k}$, the channel between the sensor-controller pair
$\mathbf{H}^{u}_{i,k}$, the channel between the controller-actuator pair
$\mathbf{H}^{d}_{i,k}$, the UL transmission power $P^{u}_{i,k}$, and the DL
transmission power $P^{d}_{i,k}$, the control stability constraint in (36) is
satisfied IFF the following conditions on the transmission indicator variables
hold, i.e.,
$\displaystyle\textstyle{\underset{K\to\infty}{\lim\sup}\frac{1}{K}\sum\limits_{k=1}^{K}\xi^{u}_{i,k}\geq\underset{K\to\infty}{\lim\sup}\frac{1}{K}\sum\limits_{k=1}^{K}\frac{\|\left(\mathbf{A}^{c}_{i}-\zeta_{i}\mathbf{I}_{D}\right)\hat{\mathbf{x}}^{u}_{i,k}\|_{\mathcal{Z}^{\frac{1}{2}}}^{2}+\text{Tr}\left[\left(\mathbf{A}^{T}_{i}\mathcal{Z}\mathbf{A}_{i}-\zeta_{i}\mathcal{Z}\right)\mathcal{J}^{u}_{i,k}\right]+\text{Tr}\left[\mathbf{B}^{T}_{i}\mathcal{Z}\mathbf{B}_{i}\mathcal{J}^{d}_{i,k}\right]+\text{Tr}\left[\mathcal{Z}\mathbf{W}\right]}{\text{Tr}\left[\left(\mathbf{B}_{i}\mathbf{\Phi}_{i}\right)^{T}\mathcal{Z}\left(\mathbf{B}_{i}\mathbf{\Phi}_{i}\right)\mathcal{J}^{u}_{i,k}\right]-\text{Tr}\left[\left(\mathbf{B}_{i}\mathbf{\Phi}_{i}\right)^{T}\mathcal{Z}\left(\mathbf{B}_{i}\mathbf{\Phi}_{i}\right)\mathbf{V}^{u}_{i,k}\right]}},$
(38)
$\displaystyle\textstyle{\underset{K\to\infty}{\lim\sup}\frac{1}{K}\sum\limits_{k=1}^{K}\xi^{d}_{i,k}\geq\underset{K\to\infty}{\lim\sup}\frac{1}{K}\sum\limits_{k=1}^{K}\frac{\|\left(\mathbf{A}^{c}_{i}-\zeta_{i}\mathbf{I}_{D}\right)\hat{\mathbf{x}}^{u}_{i,k}\|_{\mathcal{Z}^{\frac{1}{2}}}^{2}+\text{Tr}\left[\left(\mathbf{A}^{T}_{i}\mathcal{Z}\mathbf{A}_{i}-\zeta_{i}\mathcal{Z}\right)\mathcal{J}^{u}_{i,k}\right]+\text{Tr}\left[\mathbf{B}^{T}_{i}\mathcal{Z}\mathbf{B}_{i}\mathcal{J}^{d}_{i,k}\right]+\text{Tr}\left[\mathcal{Z}\mathbf{W}\right]}{\text{Tr}\left[\mathbf{B}_{i}^{T}\mathcal{Z}\mathbf{B}_{i}\mathcal{J}^{d}_{i,k}\right]-\text{Tr}\left[\mathbf{B}_{i}^{T}\mathcal{Z}\mathbf{B}_{i}\mathbf{V}^{d}_{i,k}\right]}},$
(39)
$\displaystyle\textstyle{\underset{K\to\infty}{\lim\sup}\frac{1}{K}\sum\limits_{k=1}^{K}\xi^{u}_{i,k}\xi^{d}_{i,k}\geq\underset{K\to\infty}{\lim\sup}\frac{1}{K}\sum\limits_{k=1}^{K}\frac{\|\left(\mathbf{A}^{c}_{i}-\zeta_{i}\mathbf{I}_{D}\right)\hat{\mathbf{x}}^{u}_{i,k}\|_{\mathcal{Z}^{\frac{1}{2}}}^{2}+\text{Tr}\left[\left(\mathbf{A}^{T}_{i}\mathcal{Z}\mathbf{A}_{i}-\zeta_{i}\mathcal{Z}\right)\mathcal{J}^{u}_{i,k}\right]+\text{Tr}\left[\mathbf{B}^{T}_{i}\mathcal{Z}\mathbf{B}_{i}\mathcal{J}^{d}_{i,k}\right]+\text{Tr}\left[\mathcal{Z}\mathbf{W}\right]}{\text{Tr}\left[\left((\mathbf{B}_{i}\mathbf{\Phi}_{i})^{T}\mathcal{Z}(\mathbf{B}_{i}\mathbf{\Phi}_{i})\right)(\mathcal{J}^{u}_{i,k}-\mathbf{V}^{u}_{i,k})\right]+\text{Tr}\left[\mathbf{B}_{i}^{T}\mathcal{Z}\mathbf{B}_{i}(\mathcal{J}^{d}_{i,k}-\mathbf{V}^{d}_{i,k}\right]}}.$
(40)
###### Proof.
Please refer to Appendix. B ∎
Note that the conditions on the UL and DL transmission indicator variables in
(38) and (39), respectively, ensure the control stability constraint in (36)
in a decoupled scheduling between the UL and DL communications based on the
current predicted control and channel states, while the condition on both the
UL and DL scheduling variables in (40) ensures control stability in a coupled
scheduling between the UL and DL communications. Intuitively, the growth of
AoI at a controller/actuator leads to an increasing state/action prediction
error increasing due to an outdated training set. Therefore, the transmitter-
receiver pair of a control system should be scheduled when it has a reliable
communication link and the state/action prediction error is greater than the
state/action estimation error to ensure control stability.
The actuator is physically decoupled from the centralized scheduler and the
controller which is co-located at BS, and the DL indicator variable at the
centralized scheduler relies on the action prediction error at the actuator.
Hence, the controller leverages another GPR, where the input of this GPR is
the discrete-time associated with the generated action by LQR plus the action
estimation error as
$\bar{\mathbf{u}}^{d_{c}}_{i,k^{\prime}}=\mathbf{u}^{d}_{i,k^{\prime}}+\mathbf{v}^{d}_{i,k^{\prime}}$.
As a result of the applied input to this GPR that yields a training set
similar to the one at the actuator as
$\mathcal{D}^{d_{c}}_{i,n_{d}}=\\{(k^{\prime},\xi^{d}_{i,k^{\prime}}\,\bar{\mathbf{u}}^{d_{c}}_{i,k^{\prime}})|\,k^{\prime}=1,\cdots,n_{d},i=1,\cdots,M\\}$,
we obtain the action prediction error similar to the one generated at actuator
side.
### III-B Joint Communication and Control Problem
According to the UL and DL transmission indicator variables constraints in
(38)-(40) that result from the control stability constraint in (36) in problem
$\mathcal{P}1$, problem $\mathcal{P}1$ is rewritten as
$\displaystyle\underset{\mathbf{a}_{k}^{l},\mathbf{P}_{k}^{l}}{(\mathcal{P}2)\quad\text{Minimize}}\mathcal{C}\left(\\{\bar{\beta}^{l}_{i}\\},\\{\bar{\hat{P}}^{l}_{i}\\}\right)$
(41a) $\displaystyle\text{subject
to:}\quad\bar{\alpha}^{u}_{i}\geq\bar{\mathcal{G}}_{lb}\left(\mathfrak{m}^{u}_{i,k}\right),\;\;\;\forall
i\in\mathcal{M},$ (41b) $\displaystyle\hskip
45.0pt\bar{\alpha}^{d}_{i}\geq\bar{\mathcal{G}}_{lb}\left(\mathfrak{m}^{d}_{i,k}\right),\;\;\;\forall
i\in\mathcal{M},$ (41c) $\displaystyle\hskip
45.0pt\overline{\alpha^{u}_{i}\alpha^{d}_{i}}\geq\bar{\mathcal{G}}_{lb}\left(\mathfrak{m}_{i,k}\right),\;\forall
i\in\mathcal{M},$ (41d)
$\displaystyle~{}\eqref{eqopt1e_1a}~{}-~{}\eqref{eqopt1e_1d},$
where $\bar{\alpha}^{u}_{i}$ and $\bar{\alpha}^{d}_{i}$ are the time-averaged
of the UL and DL scheduling variables, respectively, $\mathfrak{m}^{u}_{i,k}$,
$\mathfrak{m}^{d}_{i,k}$, $\mathfrak{m}_{i,k}$ are the lower-bound stability
of the uplink, downlink, and coupling transmission indicator variables of a
control system $i$ at time $k$ in (38), (39), and (40) respectively.
$\bar{\mathcal{G}}_{lb}$ is the time-averaged of the lower-bound function
$\mathcal{G}_{lb}$ that is defined as
$\mathcal{G}_{lb}\left(.\right)=\max\left[\min\left(.,1\right),0\right]$ to
ensure the feasibility of the scheduling constraints. The transmission
indicator variable in (38)-(40) is only written as a function of the
scheduling variables since the SNR indicator function is satisfied in (37c).
The stochastic problem $\mathcal{P}2$ is a mixed-integer non-convex problem
where the source of stochasticity is due to the observed channel and predicted
state at each time $k$. Moreover, the scheduling decision constraint in (37d)
not only depends on its own decision but on all others scheduling decisions.
Hence, a dynamic control algorithm is proposed in the next section to find the
optimal scheduling vector and the optimal transmission power vector of problem
$\mathcal{P}2$ utilizing Lyapunov optimization framework.
## IV Dynamic Control Algorithm Using Lyapunov Optimization
In this section, we propose a dynamic control algorithm using the stochastic
Lyapunov optimization framework to solve problem $\mathcal{P}2$. However, the
problem involves minimizing a weighted sum of non-decreasing concave functions
of the time-averaged AoI and transmission power. Based on the dynamic
stochastic optimization theory [24], it can be transformed into an equivalent
problem that involves minimizing a time-averaged cost function of
instantaneous AoI and transmission power. This transformation is achieved
through the use of non-negative auxiliary variables $\gamma^{\beta^{l}}_{i,k}$
and $\gamma^{P^{l}}_{i,k}$ and corresponding virtual queues
$Q^{\beta^{l}}_{i,k}$ and $Q^{P^{l}}_{i,k}$ with queue dynamics as
$\displaystyle
Q^{\beta^{l}}_{i,k+1}=\max\Big{\\{}Q^{\beta^{l}}_{i,k}-\gamma^{\beta^{l}}_{i,k},0\Big{\\}}+\beta^{l}_{i,k},\qquad\forall
l\in\\{u,d\\},i\in\mathcal{M},k,$ (42a) $\displaystyle
Q^{P^{l}}_{i,k+1}=\max\Big{\\{}Q^{P^{l}}_{i,k}-\gamma^{P^{l}}_{i,k},0\Big{\\}}+\hat{P}^{l}_{i,k},\qquad\forall
l\in\\{u,d\\},i\in\mathcal{M},k,$ (42b)
where $\hat{P}^{l}_{i,k}$ will be optimized at each time $k$. Then, the
transformed problem is given as
$\displaystyle\underset{\mathbf{a}_{k}^{l},\mathbf{P}_{k}^{l},\mathbf{r}^{\beta^{l}},\mathbf{r}^{P^{l}}}{(\mathcal{P}3)\quad\text{Minimize}}\overline{\mathcal{C}\left(\\{\gamma^{\beta^{l}}_{i,k}\\},\\{\gamma^{P^{l}}_{i,k}\\}\right)}$
(43a) $\displaystyle\text{subject
to:}\qquad\bar{\beta}^{l}_{i}\leq\bar{\gamma}_{i}^{\beta^{l}},\qquad\qquad\quad\forall
l\in\\{u,d\\},i\in\mathcal{M}$ (43b)
$\displaystyle\bar{\hat{P}}^{l}_{i}\leq\bar{\gamma}_{i}^{P^{l}},\qquad\qquad\quad\forall
l\in\\{u,d\\},i\in\mathcal{M}$ (43c) $\displaystyle
1\leq\gamma_{i,k}^{\beta^{l}}\leq B_{max},\qquad\forall
l\in\\{u,d\\},i\in\mathcal{M},k$ (43d) $\displaystyle
0\leq\gamma_{i,k}^{P^{f}}\leq P^{l}_{max},\qquad\forall
l\in\\{u,d\\},i\in\mathcal{M},k$ (43e)
$\displaystyle~{}\eqref{eqopt1e_1a}-\eqref{eqopt1e_1d},~{}\eqref{eqopt1_1a}-\eqref{eqopt1_1c},$
where
$\mathbf{r}_{k}^{\beta^{l}}=\\{\gamma^{\beta^{l}}_{i,k}:l\in\\{u,d\\},i\in\mathcal{M}\\}$
and
$\mathbf{r}_{k}^{P^{l}}=\\{\gamma^{P^{l}}_{i,k}:l\in\\{u,d\\},i\in\mathcal{M}\\}$
are the vectors of the introduced auxiliary variables. The constraints in
(43d) and (43e) are introduced to bound the auxiliary variables. These
constraints can be satisfied by ensuring the stability of their virtual
queues, since the lower-bound of these constraints can be viewed as the
arrival rate of their virtual queues, while the upper-bound can be viewed as
the service rate of such virtual queues [24]. Following [24], the problem
$\mathcal{P}2$ and the transformed problem $\mathcal{P}3$ are equivalent in
which the optimal solution of $\mathcal{P}3$ can be directly turned into an
optimal solution of $\mathcal{P}2$.
To handle UL and DL scheduling variables constraints in (41b)$-$(41d)
associated with the control stability constraint in (36), the virtual queues
$Q^{C^{l}}_{i,k}$ and $Q^{C}_{i,k}$ are introduced for all control systems
whose dynamics are
$\displaystyle
Q^{C^{l}}_{i,k+1}=\max\\{Q^{C^{l}}_{i,k}-\alpha^{l}_{i,k},0\\}+\mathcal{G}_{lb}(m^{l}_{i,k}),\qquad\forall
l\in\\{u,d\\},i\in\mathcal{M},k,$ (44a) $\displaystyle
Q^{C}_{i,k+1}=\max\\{Q^{C}_{i,k}-\alpha^{u}_{i,k}\alpha^{d}_{i,k},0\\}+\mathcal{G}_{lb}(m_{i,k}),\qquad\qquad\;\;\;\forall
i\in\mathcal{M},k,$ (44b)
where $\alpha^{l}_{i,k}$ will be optimized at each time $k$. The constraints
in (41b)$-$(41c) can be satisfied, if their virtual queues are mean-rate
stable, i.e., their time-averaged arrival rate is not larger than its time-
averaged service rate [24]. At this point, the dynamic stochastic optimization
is applied to solve the transformed problem $\mathcal{P}3$, which minimizes a
weighted sum of the time-averaged cost function of instantaneous AoI and
transmission power subject to the virtual queues stability constraints and the
original problem constraints in (37b)$-$(37e). In this regard, we define
$\mathbf{Q}^{\beta^{l}}_{k}$, $\mathbf{Q}^{P^{l}}_{k}$, $\mathbf{Q}^{C}_{k}$,
and $\mathbf{Q}^{C^{l}}_{k}$ as a vector of all virtual queues
$Q^{\beta^{l}}_{i,k}$, $Q^{P^{l}}_{i,k}$, $Q^{C}_{i,k}$, and $Q^{C^{l}}_{i,k}$
for all control systems, respectively. We denote the combined queue vector of
all virtual queues at time $k$ by
${\scriptstyle\mathcal{X}_{k}=\left[\mathbf{Q}^{\beta^{l}}_{k},\mathbf{Q}^{P^{l}}_{k},\mathbf{Q}^{C}_{k},\mathbf{Q}^{C^{l}}_{k}\right]}$,
and express the conditional Lyapunov drift-plus-penalty as
$\small\Delta\left(\mathcal{X}_{k}\right)=\mathbb{E}\left[\mathcal{L}\left(\mathcal{X}_{k+1}\right)-\mathcal{L}\left(\mathcal{X}_{k}\right)+V\mathcal{C}\left(\\{\gamma^{\beta^{l}}_{i,k}\\},\\{\gamma^{P^{l}}_{i,k}\\}\right)\Big{|}\mathcal{X}_{k}\right],$
(45)
where $\mathcal{L}\left(\mathcal{X}_{k}\right)$ is the quadratic Lyapunov
function of $\mathcal{X}_{k}$ that measures the virtual queues congestion in a
scalar metric and is defined as
$\mathcal{L}\left(\mathcal{X}_{k}\right)=\frac{1}{2}\sum^{M}_{i=1}\Big{[}(Q^{\beta^{l}}_{i,k})^{2}+(Q^{P^{l}}_{i,k})^{2}+(Q^{C}_{i,k})^{2}+(Q^{C^{l}}_{i,k})^{2}\Big{]}$.
$V\geq 0$ controls the trade-off between minimizing the objective function and
stabilizing the virtual queues. Subsequently, plugging the inequalities
$\left(\max\left[a-b,0\right]+c\right)^{2}\leq
a^{2}+b^{2}+c^{2}-2a\left(b-c\right)$, $\forall a,b,c\geq 0$,
$\left(\max\left(a,0\right)\right)^{2}\leq a^{2}$, and all virtual queue
dynamics into (45), we derive
$\displaystyle\eqref{eq46}$
$\displaystyle\leq{B}+\mathbb{E}\Big{[}\sum^{M}_{i=1}\left(V\omega_{\beta_{l}}\mathcal{G}_{\beta}\left(\gamma^{\beta^{l}}_{i,k}\right)-Q^{\beta^{l}}_{i,k}\gamma^{\beta^{l}}_{i,k}\right)|\mathcal{X}_{k}\Big{]}+\mathbb{E}\Big{[}\sum^{M}_{i=1}\left(V\omega_{P_{l}}\mathcal{G}_{P}\left(\gamma^{P^{l}}_{i,k}\right)-Q^{P^{l}}_{i,k}\gamma^{P^{l}}_{i,k}\right)|\mathcal{X}_{k}\Big{]}$
(46)
$\displaystyle\qquad\,+\mathbb{E}\Big{[}\sum^{M}_{i=1}Q^{\beta^{l}}_{i,k}\beta^{l}_{i,k}|\mathcal{X}_{k}\Big{]}+\mathbb{E}\Big{[}\sum^{M}_{i=1}Q^{P^{l}}_{i,k}\hat{P}^{l}_{i,k}|\mathcal{X}_{k}\Big{]}-\mathbb{E}\Big{[}\sum^{M}_{i=1}Q^{C^{l}}_{i,k}\left(\alpha^{l}_{i,k}-\mathcal{G}_{lb}(m^{l}_{i,k})\right)|\mathcal{X}_{k}\Big{]}$
$\displaystyle\qquad\,-\mathbb{E}\Big{[}\sum^{M}_{i=1}Q^{C}_{i,k}\left(\alpha^{u}_{i,k}\alpha^{d}_{i,k}-\mathcal{G}_{lb}(m_{i,k})\right)|\mathcal{X}_{k}\Big{]}.$
The constant B details in (46) is omitted since it does not affect the system
performance in the Lyapunov optimization. A solution to $\mathcal{P}3$ can be
obtained by minimizing the upper-bound (46) at each time as
$\displaystyle\underset{\hskip
25.0pt\mathbf{a}^{l},\mathbf{P}^{l},\mathbf{r}^{\beta^{l}},\mathbf{r}^{P^{l}}}{(\mathcal{P}4)\quad\text{Minimize}}\quad\sum^{M}_{i=1}\Bigg{[}\left(V\omega_{\beta_{l}}\mathcal{G}_{\beta}\left(\gamma^{\beta^{l}}_{i,k}\right)-Q^{\beta^{l}}_{i,k}\gamma^{\beta^{l}}_{i,k}\right)+\left(V\omega_{P_{l}}\mathcal{G}_{P}\left(\gamma^{P^{l}}_{i,k}\right)-Q^{P^{l}}_{i,k}\gamma^{P^{l}}_{i,k}\right)+Q^{\beta^{l}}_{i,k}\beta^{l}_{i,k}$
(47a)
$\displaystyle\qquad\qquad\qquad\qquad\qquad+Q^{P^{l}}_{i,k}\hat{P}^{l}_{i,k}-Q^{C^{l}}_{i,k}\left(\alpha^{l}_{i,k}-\mathcal{G}_{lb}(m^{l}_{i,k})\right)-Q^{C}_{i,k}\left(\alpha^{u}_{i,k}\alpha^{d}_{i,k}-\mathcal{G}_{lb}(m_{i,k})\right)\Bigg{]}$
$\displaystyle\text{subject
to:}\qquad~{}\eqref{eqopt1e_1a}-\eqref{eqopt1e_1d}~{}\text{and}~{}\eqref{eqopt2_1c}-\eqref{eqopt2_1d}.$
The optimality of problem $\mathcal{P}4$ is asymptotically approached by
increasing $V$ [24]. Since the problem $\mathcal{P}4$ is of separable
structure, which motivate us to determine the AoI auxiliary vector
$\mathbf{r}^{\beta^{l}}$, transmission power auxiliary vector
$\mathbf{r}^{P^{l}}$, scheduling vector $\mathbf{a}^{l}$, and transmission
power vector $\mathbf{P}^{l}$ in an alternative optimization form. Hence, the
overall minimization problem $\mathcal{P}4$ can be decomposed into two
separate sub-problems that can be solved concurrently with the observation of
the virtual queues, control, and channel states.
#### IV-1 Auxiliary Variable Sub-Problems
The first decomposed sub-problem is the AoI auxiliary sub-problem, while the
second decomposed sub-problem is the transmission power sub-problem. Since the
auxiliary variables of such problems are separated and independent among
different control systems, their minimization sub-problems can be decoupled to
be computed for each control system separately as the following convex
problems
$\displaystyle\underset{\hskip
45.0pt\gamma^{\beta^{l}}_{i,k}}{(\mathcal{P}4.1)\quad\text{Minimize}}\quad
V\omega_{\beta_{l}}\mathcal{G}_{\beta}\left(\gamma^{\beta^{l}}_{i,k}\right)-Q^{\beta^{l}}_{i,k}\gamma^{\beta^{l}}_{i,k}$
(48a) $\displaystyle\hskip 20.0pt\text{subject to:}\quad\quad
1\leq\gamma_{i,k}^{\beta^{l}}\leq B_{max},$ (48b)
$\displaystyle\underset{\hskip
45.0pt\gamma^{P^{l}}_{i,k}}{(\mathcal{P}4.2)\quad\text{Minimize}}\quad
V\omega_{P_{l}}\mathcal{G}_{P}\left(\gamma^{P^{l}}_{i,k}\right)-Q^{P^{l}}_{i,k}\gamma^{P^{l}}_{i,k}$
(49a) $\displaystyle\hskip 20.0pt\text{subject to:}\quad\quad
0\leq\gamma_{i,k}^{P^{l}}\leq P^{l}_{max}.$ (49b)
The optimal AoI auxiliary variables are obtained by differentiating the
objective functions of these problems. Let
$\mathcal{A}(\gamma^{\beta^{l}})=V\omega_{\beta_{l}}\log(1+\gamma^{\beta^{l}})-Q^{\beta^{l}}_{i,k}\gamma^{\beta^{l}}_{i,k}$
and $\gamma^{\beta^{l^{*}}}_{i,k}$ denotes the solution of
$\mathcal{A}(\gamma^{\beta^{l}})$ as
$\acute{\mathcal{A}}(\gamma^{\beta^{l}})=\frac{V\omega_{\beta_{l}}}{\left(1+\gamma^{\beta^{l}}_{i,k}\right)}-Q^{\beta^{l}}_{i,k}=0$,
the optimal AoI auxiliary variable of $\mathcal{P}4.1$ is given as
$\small\gamma^{\beta^{l^{*}}}_{i,k}=\min\left\\{\max\left\\{\frac{V\omega_{\beta_{l}}-Q^{\beta^{l}}_{i,k}}{Q^{\beta^{l}}_{i,k}},1\right\\},B_{max}\right\\},\qquad\forall
l\in\\{u,d\\},i\in\mathcal{M},k.$ (50)
Similarly, by letting
$\mathcal{A}(\gamma^{P^{l}})=V\omega_{P^{l}}\log(1+\gamma^{P^{l}})\,-\,Q^{P^{l}}_{i,k}\gamma^{P^{l}}_{i,k}$
and $\gamma^{P^{l^{*}}}_{i,k}$ denotes the solution of
$\mathcal{A}(\gamma^{P^{l}})$ as
$\acute{\mathcal{A}}(\gamma^{P^{l}})=\frac{V\omega_{P^{l}}}{\left(1+\gamma^{P^{l}}_{i,k}\right)}-Q^{P^{l}}_{i,k}=0$,
the optimal transmission power auxiliary variable of $\mathcal{P}4.2$ is given
as
$\small\gamma^{P^{l^{*}}}_{i,k}=\min\left\\{\max\left\\{\frac{V\omega_{P^{l}}-Q^{P^{l}}_{i,k}}{Q^{P^{l}}_{i,k}},0\right\\},P^{l}_{max}\right\\},\qquad\forall
l\in\\{u,d\\},i\in\mathcal{M},k.$ (51)
#### IV-2 Scheduling Decision and Transmission Power Sub-Problems
The optimal UL-DL scheduling variables and the optimal UL-DL transmission
power variables are obtained by minimizing the remaining terms of the
objective function of problem $\mathcal{P}4$ at each time subject to the
scheduling and transmission power constraints in (37b)$-$(37e), which is
expressed as
$\displaystyle\small\underset{\hskip
40.0pt\mathbf{a}^{l}_{i,k},\mathbf{P}^{l}_{i,k}}{(\mathcal{P}4.3)\quad\text{Minimize}}\;\sum^{M}_{i=1}\left[Q^{\beta^{l}}_{i,k}\beta^{l}_{i,k}+Q^{P^{l}}_{i,k}\,\hat{P}^{l}_{i,k}-Q^{C^{l}}_{i,k}\left(\alpha^{l}_{i,k}-\mathcal{G}_{lb}(m^{l}_{i,k})\right)-Q^{C}_{i,k}\left(\alpha^{u}_{i,k}\alpha^{d}_{i,k}-\mathcal{G}_{lb}(m_{i,k})\right)\right]$
(52a) $\displaystyle\text{subject
to:}\quad~{}\eqref{eqopt1e_1a}-\eqref{eqopt1e_1d},$
which is a mixed-integer non-convex problem. Due to the complexity of
exhaustive search for finding the optimal solution, we propose a low-
complexity two-stage sequential optimization strategy to find a sub-optimal
solution to the joint power allocation and scheduling assignment problem. This
strategy firstly obtains the UL and DL transmission power variables, followed
by the UL and DL scheduling variables. The optimal UL and DL transmission
power for each control system, determined by solving the following power
allocation problem
$\displaystyle\underset{\hskip
15.0pt\mathbf{P}^{u}_{k},\mathbf{P}^{d}_{k}}{(\mathcal{P}4.4)\quad\text{Minimize}}\;\sum^{M}_{i=1}\alpha^{u}_{i,k}\left[\mathcal{Q}^{S_{1}}_{i,k}+Q^{P^{u}}_{i,k}P^{u}_{i,k}\right]+\alpha^{d}_{i,k}\left[\mathcal{Q}^{C_{1}}_{i,k}+Q^{P^{d}}_{i,k}P^{d}_{i,k}\right]-\alpha^{u}_{i,k}\alpha^{d}_{i,k}Q^{C}_{i,k}+\mathcal{Q}^{l_{1}}_{i,k},$
(53a) $\displaystyle\text{subject
to:}\quad\frac{\text{SNR}^{u}_{th}\,{N_{0}}}{\|\mathbf{H}^{u}_{i,k}\|^{2}}\leq
P^{u}_{i,k}\leq P^{u}_{max},\qquad\forall i\in\mathcal{M},k$ (53b)
$\displaystyle\frac{\text{SNR}^{d}_{th}\,{N_{0}}}{\|\mathbf{H}^{d}_{i,k}\|^{2}}\leq
P^{d}_{i,k}\leq P^{d}_{max},\qquad\forall i\in\mathcal{M},k,$ (53c)
where
$\mathcal{Q}^{S_{1}}_{i,k}=-Q^{\beta^{u}}_{i,k}\beta^{u}_{i,k-1}-Q^{C^{u}}_{i,k}$,
$\mathcal{Q}^{C_{1}}_{i,k}=-Q^{\beta^{d}}_{i,k}\beta^{d}_{i,k-1}-Q^{C^{d}}_{i,k}$,
and
$\mathcal{Q}^{l_{1}}_{i,k}=Q^{C^{u}}_{i,k}\mathcal{G}_{lb}(m^{u}_{i,k})+Q^{C^{d}}_{i,k}\mathcal{G}_{lb}(m^{d}_{i,k})+Q^{C}_{i,k}\mathcal{G}_{lb}(m_{i,k})+Q_{i,k}^{\beta^{u}}\left(1+\beta^{u}_{i,k-1}\right)+Q_{i,k}^{\beta^{d}}\left(1+\beta^{d}_{i,k-1}\right)$
are the constant terms defined in the objective function of $\mathcal{P}4.4$.
The above problem $\mathcal{P}4.4$ is a generalized min-weight problem, which
can be decoupled into a series of independent sub-problems for each control
system separately. Hence, the optimal UL-DL transmission power variables are
given as
$\small
P^{l^{*}}_{i,k}=\left\\{\begin{array}[]{cc}\frac{\text{SNR}^{l}_{th}N_{0}}{\|\mathbf{H}^{l}_{i,k}\|^{2}},&\text{if}\;Q^{P^{l}}_{i,k}\geq
0\\\ P^{l}_{max},&\mbox{if}\;Q^{P^{l}}_{i,k}<0.\end{array}\right.$ (54)
Given the optimal UL and DL transmission power variables in (54), the optimal
UL and DL scheduling variables for each control system that has a control
state/action to transmit are obtained by solving the following scheduling
assignment problem
$\displaystyle\underset{\hskip
40.0pt\mathbf{a}^{u}_{k},\mathbf{a}^{d}_{k}}{(\mathcal{P}4.5)\quad\text{Minimize}}\;\sum^{M}_{i=1}\alpha^{u}_{i,k}\left[\mathcal{Q}^{S_{1}}_{i,k}+Q^{P^{u}}_{i,k}P^{u^{*}}_{i,k}\right]+\alpha^{d}_{i,k}\left[\mathcal{Q}^{C_{1}}_{i,k}+Q^{P^{d}}_{i,k}P^{d^{*}}_{i,k}\right]-\alpha^{u}_{i,k}\alpha^{d}_{i,k}Q^{C}_{i,k}+\mathcal{Q}^{l_{1}}_{i,k},$
(55a) $\displaystyle\text{subject
to:}\quad~{}\eqref{eqopt1e_1c}-\eqref{eqopt1e_1d}.$
The optimal UL and DL scheduling variables are obtained as follows:
$\small\alpha^{u^{*}}_{i,k}\,\&\,\alpha^{d^{*}}_{i,k}=\left\\{\begin{array}[]{cc}\alpha^{u}_{j_{1},k}=1,\;\alpha^{d}_{j_{2},k}=1,&\text{if}\;\mathbb{Q}^{1}_{j_{1},k}+\mathbb{Q}^{2}_{j_{2},k}<\mathbb{Q}^{1}_{j_{3},k}+\mathbb{Q}^{2}_{j3,k}+\mathbb{Q}^{3}_{j_{3},k}\\\
\alpha^{u}_{j_{3},k}=1,\;\alpha^{d}_{j_{3},k}=1,&\text{if}\;\mathbb{Q}^{1}_{j_{1},k}+\mathbb{Q}^{2}_{j_{2},k}>\mathbb{Q}^{1}_{j_{3},k}+\mathbb{Q}^{2}_{j_{3},k}+\mathbb{Q}^{3}_{j_{3},k}\\\
\alpha^{u}_{j,k}=0,\;\alpha^{d}_{j,k}=0,&\forall
j\notin\\{j_{1}\,\&\,j_{2}||j_{3}\,\\},\end{array}\right.$ (56)
where
$\mathbb{Q}_{i,k}^{1}=\mathcal{Q}^{S_{1}}_{i,k}+Q^{P^{u}}_{i,k}P^{u^{*}}_{i,k}$,
$\mathbb{Q}_{i,k}^{2}=\mathcal{Q}^{C_{1}}_{i,k}+Q^{P^{d}}_{i,k}P^{d^{*}}_{i,k}$,
and $\mathbb{Q}_{i,k}^{3}=-Q^{C}_{i,k}$ are the terms defined in the objective
function of problem $\mathcal{P}4.5$. Moreover,
$j_{1}=\operatorname*{arg\,min}_{i\in\mathcal{M}}\mathbb{Q}^{1}_{i,k}$,
$j_{2}=\operatorname*{arg\,min}_{i\in\mathcal{M}}\mathbb{Q}^{2}_{i,k}$, and
$j_{3}=\operatorname*{arg\,min}_{i\in\mathcal{M}}(\mathbb{Q}^{1}_{i,k}+\mathbb{Q}^{2}_{i,k}+\mathbb{Q}^{3}_{i,k})$
are control system indices.
## V Simulation Results and Discussions
In this section, the performance of the proposed stability-aware scheduling
algorithm is investigated in an inverted-pendulum on a cart system with $M=2$,
and $M=20$ inverted-pendulums, respectively. Each inverted-pendulum system is
described by a four-dimensional state vector as
$\mathbf{x}^{u}_{i,k}=\left[x_{i,k},\dot{x}_{i,k},\theta_{i,k},\dot{\theta}_{i,k}\right]$,
where $x_{i,k}$ represents the cart’s position along the horizontal axis,
$\dot{x}_{i,k}$ represents the cart’s velocity, $\theta_{i,k}$ represents the
pendulum angle along the vertical axis, and $\dot{\theta}_{i,k}$ represents
the pendulum’s angular velocity. The initial state of the control systems $i$
is $\mathbf{x}^{u}_{i,0}=\left[0\quad 0\quad 0.1\quad 0\right]^{T}$. The
action $\mathbf{u}^{a}_{i,k}$ is the horizontal force applied on the linear
cart. By applying a zeroth-order with a state sampling of $10$ms on the
continuous dynamics of the inverted-pendulum system and linearizing around the
pendulum up-position, i.e., $\theta_{i,k}=0$, we obtain the following
discrete-time linear dynamics matrices [14],
$\displaystyle\mathbf{A}_{i}=\left[\begin{array}[]{cccc}1&0&0&0\\\
0&2.055&-0.722&4.828\\\ 0&0.023&0.91&0.037\\\ 0&0.677&-0.453&2.055\\\
\end{array}\right],\mathbf{B}_{i}=\left[\begin{array}[]{c}0.034\\\ 0.168\\\
0.019\\\ 0.105\\\ \end{array}\right],$ (65)
Since $\mathbf{A}_{i}$’s largest eigenvalues
$\left\\{3.85,0.42,0.92,1.00\right\\}$ is greater than unity, the inverted-
pendulum is unstable without an appropriate control action [20]. To stabilize
the control system, the feedback gain matrix $\mathbf{\Phi}_{i}$ is calculated
at the controller based on LQR in (27). The rest of the simulation parameters
are $P^{l}_{max}=20\,\text{dBm}$, $N_{0}=-20\,\text{dBm}$, $\zeta_{i}=0.01$,
$V=1000$, $\omega_{\beta}=1$, $\omega_{P}=1$, $h_{k}=1$, $h_{q}=1$, $\mu=1$,
and $\sigma_{n}^{2}=0.01$.
The performance of the proposed stability-aware scheduling method is compared
versus five scheduling baselines. In Baseline $1$. (Round-Robin Scheduling),
each sensor/controller periodically transmits its state/action over a wireless
channel with fixed transmission power and a predefined repeating order [8, 9].
In Baseline $2$. (Opportunistic Scheduling), the sensor/controller is
scheduled under favorable channel conditions. Otherwise, the
controller/actuator applies the last received state/action [10, 11]. In
Baseline $3$, (Event-triggered Scheduling without FDMA), one control system
with the largest state discrepancy, i.e., the difference between the current
predicted state using Kalman filtering and the previous received/predicted
state is larger than a predefined threshold, is scheduled at each time to
transmit the state/action with fixed transmission power [12, 13]. In Baseline
$4$. (Event-triggered Scheduling with FDMA), each sensor/controller transmits
its state/action with fixed transmission power based on its stability
condition, i.e., difference between the current and previous states is less
than a predefined threshold, using FDMA [25]. In Baseline $5$. (Ideal Control
Scheduling), all control systems simultaneously transmit their states/actions
with ultra-low latency and high reliability over perfect channels [9]. Results
are collected over ten independent simulation runs, and each simulation is run
for $K=90$ discrete-time steps.
(a) Average Pendulum Angle with $M=2$. (b) Average Pendulum Angle with $M=20$.
Figure 3: Average pendulum angle to the vertical center, i.e., control error
with $M=2$ and $M=20$ pendulum angle systems using the proposed stability-
aware, round-robin, opportunistic, event-triggered with and without FDMA.
Average Control Error Vs. Control Systems. Fig. 3 illustrates the average
pendulum angle to the vertical center of each control system, i.e., the
average control error of each control system, during $90$ control time steps.
As shown in Fig.3(a), the proposed and baseline scheduling methods, assuming a
low number of control systems $\left(M=2\right)$, can keep all the pendulums
upright. Moreover, the proposed stability-Aware with GPR and the event-
triggered with FDMA keep both pendulums close to zero ensuring both pendulums
have the same control performance. This is because the proposed solution
adapts to both channel and control states, i.e., the control system is
scheduled if it has a favorable channel condition and unstable control state
needs to be stabilized. Meanwhile, the event-triggered with FDMA scheduling
has approximately the same performance at the cost of wasting wireless
communication resources by transmitting with a fixed transmission power and a
high communication rate. The opportunistic scheduler without GPR has better
control performance compared to the event-triggered without FDMA and the
round-robin methods leveraging channel state in scheduling compared to the
scheduling baselines. This in turn reflects the connection between the state
estimation stability and control stability.
Fig. 3(b) plots the average control error of each control system for the
proposed and baseline scheduling methods. In large numbers of control systems
the proposed stability-aware with GPR and the event-triggered with FDMA
scheduling methods keep all pendulum upright, unlike the baselines. Unlike the
scheduling baselines except the event-triggered with FDMA scheduling wherein
at most one control system is scheduled each time due to the limited
bandwidth, our proposed scheduling allows all control systems to operate
simultaneously even without receiving either the current state or action,
highlighting the effectiveness of GPRs at the controller and actuator thereby
improving communication efficiency and control stability. Moreover, it
maintains the GPR prediction credibility, achieving control stability.
Meanwhile, the event-triggered with FDMA scheduling keeps some pendulums
upright at the cost of frequent transmissions by equally dividing the
available bandwidth between the control systems, such that each control system
receives a fixed fraction $f_{i}$ of the total capacity $f_{i}=\text{BW}/M$,
affecting transmission latency.
(a) Communication rate of sensing link. (b) Communication rate of actuating
link.
Figure 4: Histogram of achieved communication rate in a large number of
control system $\left(M=20\right)$ of the proposed stability-aware and event-
triggered with FDMA throughout $90$ control time steps.
Communication Rate Vs. Number of Control Systems. Fig.4 presents a histogram
of the achieved communication rates for the sensing and actuating links for
$M=20$ during $90$ control time steps. The sensing/actuating communication
rate is defined as the number of times the sensing/actuating link of a control
system is scheduled divided by the time interval as $n_{l}/K$. It is clear
that the proposed approach achieves sensing communication rates concentrated
in the range from $0$ to $0.4$ and actuating communication rates ranging from
$0$ to $0.2$. On the other hand, the event-triggered with FDMA scheduling
achieves wide sensing and actuating communication rates ranging from $0$ to
$1$. The reason behind this result is that the stability-aware with GPR
scheduling is adapting to both the channel and control states, and GPRs at the
controller and actuator sides compensate for the missing received
observations, hence improving the communication efficiency. Meanwhile, each
control system, in the event-triggered with FDMA scheduling, only transmits
its control state/action based on its control stability condition without
taking into account channel states that result in increasing state/action
estimation uncertainty from the adverse channel states, and in turn affecting
control stability. Hence, each control system in the event-triggered with FDMA
scheduling requires frequent transmissions to ensure control stability by
applying appropriate action based on low communication uncertainty. Note that
the range of the sensing communication rate is larger than that of the
actuating communication rate since some control systems fail to transmit their
states at the beginning affecting prediction credibility. However, the number
of control systems that require frequent scheduling in the actuating link is
larger than that of the sensing link which stems from the fast dynamics
related to the inverted-pendulum system that require a quick appropriate
action.
(a) Single control system out of $M=2$. (b) Single control system out of
$M=20$.
Figure 5: Comparison of the state trajectory, controller AoI, and actuator AoI
of a randomly chosen control system between the proposed stability-aware,
round-robin, Opportunistic, event-triggered with and without FDMA, and ideal
control system.
State Trajectory and Controller/Actuator AoI Vs. Time. To dive deeper into the
benefits of the proposed stability-aware with GPR scheduling, we present in
Fig. 5 the state trajectory, controller AoI, and actuator AoI of a randomly
chosen control system in low and large number control systems regimes. In Fig
5(a), the state trajectory of the proposed stability-aware with GPR scheduling
and event-triggered with FDMA scheduling, in a low number of control systems,
are extremely close to that of the ideal control system where their pendulums
remain upright over time. Meanwhile, the state trajectory of the opportunistic
without GPR scheduling is slightly better than that of the event-triggered
without FDMA scheduling by keeping all pendulums upright over time due to the
scheduled control system with a favorable channel state. Finally, the state
trajectory of the round-robin without GPR scheduling slightly matches the
desired state at the cost of a high communication rate and fixed transmission
power.
As shown in Fig. 5(a), the controller and actuator AoI of the proposed
stability-aware scheduling equals one, i.e., the sensing and actuating links
of a control system are synchronously scheduled, until $15$ control time
steps. This is to guarantee that the received actions to action GPR has a low
state/action estimation uncertainty that affects the GPR state/action
prediction stability and control stability. Then, the actuator AoI starts
increasing compared to the controller AoI that remains at value one until $45$
control time steps, i.e., the sensing link of a control system is scheduled
until $45$ control time steps while the actuating link is not scheduled. The
rationale behind this result is to schedule either the sensing link or/and
actuating link of a control system with favorable channel condition to ensure
the state/action prediction credibility. Hence, the transmitted action, when
the sensing link of a control system is not scheduled, depends on a credible
predicted state. This shows the impact of the decoupled scheduling between the
UL and DL communications compared to the coupled scheduling in terms of
improving communication efficiency by reducing sensing and actuating
communication rates while guaranteeing control stability. For large numbers of
control systems shown in Fig. 5(b), the control error of the stability-aware
with GPR scheduling and event-triggered with FDMA scheduling exponentially
decay over time compared to the scheduling baselines. Meanwhile, the other
scheduling baselines are exponentially growing over time due to the
accumulated control error in the absence of appropriate action.
(a) Single control system out of $M=2$. (b) Single control system out of
$M=20$.
Figure 6: Comparison of the state trajectory, controller AoI, and actuator AoI
of a randomly chosen system between the proposed stability-aware w. GPR using
both direct and recursive approaches.
Time-Series Direct and Recursive GPR Approaches. Fig. 6 illustrates the state
trajectory, controller AoI, and actuator AoI of a randomly chosen control
system in low and large number control systems and for two different time
series GPR approaches. The recursive GPR approach uses the previous predicated
and observed states/actions to predict the next control state/action, while
the direct GPR approach only uses the previous observed states/actions to
predict the next control state/action. It is interesting to observe that the
state trajectory of the direct and recursive GPR approaches at the early phase
are identical since the number of predicted states/actions in the training set
of recursive GPR approach is smaller than the observed states/actions to
predict future states/actions. However, when time increases, the state
trajectory of the direct GPR approach outperforms the recursive GPR approach
since the number of predicted states/actions in training set of the recursive
GPR approach is greater than the observed states/actions. This in turn
deteriorates the predication accuracy hindering control stability [26].
Moreover, the direct and recursive GPR approaches for a low number of control
systems require frequent scheduling at the early phase (until $45$ control
time steps in the sensing link), during which the amount of state observations
ensures the state prediction credibility. Meanwhile, the recursive and direct
GPR approaches for a low number of control systems require only $15$ control
time steps in the actuating link, within which the amount of control actions
ensures the action prediction stability. Fig. 6 highlights that the AoI
increase deteriorates the control performance when increasing prediction
uncertainty until the training set has a sufficient number of observations to
ensure steady-state state/action prediction. Finally, Fig. 6 illustrates that
the state/action prediction of the recursive GPR approach converges faster to
the steady-state compared to the direct GPR approach. This affects the
scheduling decision by increasing the controller and actuator AoIs of the
recursive GPR approach at the cost of a low state/action prediction accuracy
and control performance compared to the direct GPR approach.
Figure 7: Total number of supported control systems using stability-aware
with GPR, event-triggered with FDMA, round-robin, event-triggered without
FDMA, and opportunistic scheduling.
Number of Served Control Systems Vs. Time Interval. Fig. 7 presents the final
capacity of the scheduled control systems over two different time intervals.
We assume that a scheduling method successfully control several controls
systems for all control systems within $\|\theta_{i,k}\|\leq 0.05$ error
region for $10$ independent simulation runs. As observed in Fig. 7, the
proposed approach has a significant impact compared to the baselines in terms
of supporting a large number of control systems over time interval. The
rationale behind this result is due to exploiting two GPRs at the controller
and actuator sides and obtaining a sufficient number of observations in the
GPR training sets increases. This in turn enhances the prediction credibility,
communication efficiency, and control stability performance while supporting a
large number of control systems.
## VI Conclusion
In this work, we proposed a novel stability-aware scheduling algorithm based
on the communication and control co-design exploiting analog uncoded
communication and GPR based to reduce the required communication rate and
ensure control stability through maintaining prediction stability. We
performed extensive simulations for low-latency control systems to demonstrate
the effectiveness of utilizing the GPR based approach, as well as, the
effectiveness of decoupled scheduling between the UL and DL communications to
support a large number of control systems compared to the scheduling
baselines.
## Appendix A Proof of Lemma 1
Given the predicted state $\hat{\mathbf{x}}^{u}_{i,k}$ in (21) and the state
prediction error covariance matrix $\mathcal{J}^{u}_{i,k}$ in (25), it holds
that the expected current Lyapunov value of the system is
$\displaystyle\mathbb{E}\big{[}\mathcal{L}(\mathbf{x}^{u}_{i,k})|\hat{\mathbf{x}}^{u}_{i,k}\big{]}$
$\displaystyle=\mathbb{E}\big{[}\mathbf{x}^{u^{T}}_{i,k}\,\mathcal{Z}\,\mathbf{x}^{u}_{i,k}|\hat{\mathbf{x}}^{u}_{i,k}\big{]}=\mathbb{E}\big{[}\left(\hat{\mathbf{x}}^{u}_{i,k}-\mathbf{e}^{u}_{i,k}\right)^{T}\mathcal{Z}\left(\hat{\mathbf{x}}^{u}_{i,k}-\mathbf{e}^{u}_{i,k}\right)\big{]}$
(66)
$\displaystyle=\mathbb{E}\big{[}\hat{\mathbf{x}}^{u^{T}}_{i,k}\mathcal{Z}\hat{\mathbf{x}}^{u}_{i,k}-\hat{\mathbf{x}}^{u^{T}}_{i,k}\mathcal{Z}\mathbf{e}^{u}_{i,k}-\mathbf{e}^{u^{T}}_{i,k}\mathcal{Z}\hat{\mathbf{x}}^{u}_{i,k}+\mathbf{e}^{u^{T}}_{i,k}\mathcal{Z}\mathbf{e}^{u}_{i,k}\big{]},\qquad\quad\forall\;\mathbb{S}^{D}_{++}.$
(67)
In (67), the first term is a constant as the expectation is taken w.r.t. the
state prediction error $\mathbf{e}^{u}_{i,k}$, while the cross-terms such as
$\mathbb{E}[\hat{\mathbf{x}}^{u^{T}}_{i,k}\mathcal{Z}\mathbf{e}^{u}_{i,k}]$
and
$\mathbb{E}[\mathbf{e}^{u^{T}}_{i,k}\mathcal{Z}\hat{\mathbf{x}}^{u}_{i,k}]$
can be canceled since the predicted state $\hat{\mathbf{x}}^{u}_{i,k}$ and the
state prediction error $\mathbf{e}^{u}_{i,k}$ are uncorrelated. Hence, the
expected current value of $\mathcal{L}(\mathbf{x}_{i,k})$ is given as
$\displaystyle\mathbb{E}\big{[}\mathcal{L}(\mathbf{x}^{u}_{i,k})|\hat{\mathbf{x}}^{u}_{i,k}\big{]}$
$\displaystyle=\hat{\mathbf{x}}^{u^{T}}_{i,k}\mathcal{Z}\hat{\mathbf{x}}^{u}_{i,k}+\text{Tr}\left[\mathcal{Z}\;\mathbb{E}[\mathbf{e}^{u^{T}}_{i,k}\mathbf{e}^{u}_{i,k}]\right]=\|\hat{\mathbf{x}}^{u}_{i,k}\|_{\mathcal{Z}^{\frac{1}{2}}}^{2}+\text{Tr}\left[\mathcal{Z}\;\mathcal{J}^{u}_{i,k}\right],\qquad\qquad\forall\;\mathbb{S}^{D}_{++}$
(68)
where the last term in (68) is obtained via the expectation of the quadratic
form
${\scriptstyle\mathbb{E}[\mathbf{e}^{u^{T}}_{i,k}\mathcal{Z}\mathbf{e}^{u}_{i,k}]}$
w.r.t the state estimation error, i.e,
${\scriptstyle\mathbb{E}[\mathbf{e}^{u^{T}}_{i,k}\mathcal{Z}\mathbf{e}^{u}_{i,k}]=(\mathbb{E}[\mathbf{e}^{u}_{i,k}])^{T}\mathcal{Z}(\mathbb{E}[\mathbf{e}^{u}_{i,k}])+\text{Tr}[\mathcal{Z}\;\mathcal{J}^{u}_{i,k}]}$.
It is observed in (68) that the expected current Lyapunov value grows larger
as the predicted state $\hat{\mathbf{x}}^{u}_{i,k}$ is near instability and/or
the prediction error covariance matrix $\mathcal{J}^{u}_{i,k}$ is larger due
to lack of sufficient observations. $\blacksquare$
## Appendix B Proof of Lemma 2
As a result of the remote sensing-loop state evolution in (30) and the open-
loop state evolution in (29), the expected future value of
$\mathcal{L}(\mathbf{x}^{u}_{i,k+1})$ in (36) of the UL transmission is given
as
$\displaystyle\mathbb{E}\left[\mathcal{L}(\mathbf{x}^{u}_{i,k+1})|\hat{\mathbf{x}}^{u}_{i,k},\hat{\mathbf{u}}^{d}_{i,k},\mathbf{H}^{u}_{i,k},P^{u}_{i,k}\right]=$
$\displaystyle\xi^{u}_{i,k}\;\mathbb{E}\left[\mathcal{L}(\mathbf{x}^{s}_{i,k+1})|\hat{\mathbf{x}}^{u}_{i,k},\hat{\mathbf{u}}^{d}_{i,k},\mathbf{H}^{u}_{i,k},P^{u}_{i,k}\right]$
(69)
$\displaystyle+\left(1-\xi^{u}_{i,k}\right)\;\mathbb{E}\left[\mathcal{L}(\mathbf{x}^{o}_{i,k+1})|\hat{\mathbf{x}}^{u}_{i,k},\hat{\mathbf{u}}^{d}_{i,k},\mathbf{H}^{u}_{i,k},P^{u}_{i,k}\right],$
For a given predicted state $\hat{\mathbf{x}}^{u}_{i,k}$, state prediction
error covariance matrix $\mathcal{J}^{u}_{i,k}$, predicted action
$\hat{\mathbf{u}}^{d}_{i,k}$, action prediction error covariance matrix
$\mathcal{J}^{d}_{i,k}$, wireless UL channel $\mathbf{H}^{u}_{i,k}$, and UL
transmission power $P^{u}_{i,k}$, the expected future value of
$\mathcal{L}(\mathbf{x}^{u}_{i,k+1})$ of the UL transmission in (69) is given
as
$\displaystyle\mathbb{E}$
$\displaystyle\left[\mathcal{L}(\mathbf{x}^{u}_{i,k+1})|\hat{\mathbf{x}}^{u}_{i,k},\hat{\mathbf{u}}^{d}_{i,k},\mathbf{H}^{u}_{i,k},P^{u}_{i,k}\right]$
(70)
$\displaystyle\;\overset{(1)}{=}\xi^{u}_{i,k}\mathbb{E}\left[\mathbf{x}^{s^{T}}_{i,k+1}\mathcal{Z}\mathbf{x}^{s}_{i,k+1}|\hat{\mathbf{x}}^{u}_{i,k},\hat{\mathbf{u}}^{d}_{i,k},\mathbf{H}^{u}_{i,k},P^{u}_{i,k}\right]+\left(1-\xi^{u}_{i,k}\right)\;\mathbb{E}\left[\mathbf{x}^{o^{T}}_{i,k+1}\mathcal{Z}\mathbf{x}^{o}_{i,k+1}|\hat{\mathbf{x}}^{u}_{i,k},\hat{\mathbf{u}}^{d}_{i,k},\mathbf{H}^{u}_{i,k},P^{u}_{i,k}\right]$
$\displaystyle\overset{(2)}{=}\xi^{u}_{i,k}\mathbb{E}\left[\left(\mathbf{A}^{c}_{i}\hat{\mathbf{x}}^{u}_{i,k}-\mathbf{A}^{c}_{i}\mathbf{e}^{u}_{i,k}-\mathbf{B}_{i}\mathbf{\Phi}_{i}\mathbf{v}^{u}_{i,k}-\mathbf{B}_{i}\mathbf{e}^{d}_{i,k}+\mathbf{w}_{k}\right)^{T}\mathcal{Z}\left(\mathbf{A}^{c}_{i}\hat{\mathbf{x}}^{u}_{i,k}-\mathbf{A}^{c}_{i}\mathbf{e}^{u}_{i,k}-\mathbf{B}_{i}\mathbf{\Phi}_{i}\mathbf{v}^{u}_{i,k}-\mathbf{B}_{i}\mathbf{e}^{d}_{i,k}+\mathbf{w}_{k}\right)\right]$
$\displaystyle+(1-\xi^{u}_{i,k})\mathbb{E}\left[\left(\mathbf{A}^{c}_{i}\hat{\mathbf{x}}^{u}_{i,k}-\mathbf{A}^{c}_{i}\mathbf{e}^{u}_{i,k}-\mathbf{B}_{i}\mathbf{\Phi}_{i}\mathbf{e}^{u}_{i,k}-\mathbf{B}_{i}\mathbf{e}^{d}_{i,k}+\mathbf{w}_{k}\right)^{T}\mathcal{Z}\left(\mathbf{A}^{c}_{i}\hat{\mathbf{x}}^{u}_{i,k}-\mathbf{A}^{c}_{i}\mathbf{e}^{u}_{i,k}-\mathbf{B}_{i}\mathbf{\Phi}_{i}\mathbf{e}^{u}_{i,k}-\mathbf{B}_{i}\mathbf{e}^{d}_{i,k}+\mathbf{w}_{k}\right)\right]$
$\displaystyle\overset{(3)}{=}\xi^{u}_{i,k}\left[\|\mathbf{A}^{c}_{i}\hat{\mathbf{x}}^{u}_{i,k}\|_{\mathcal{Z}^{\frac{1}{2}}}^{2}+\text{Tr}\left[\mathbf{A}^{c^{T}}_{i}\mathcal{Z}\mathbf{A}^{c}_{i}\mathcal{J}^{u}_{i,k}\right]+\text{Tr}\left[\left(\mathbf{B}_{i}\mathbf{\Phi}_{i}\right)^{T}\mathcal{Z}\left(\mathbf{B}_{i}\mathbf{\Phi}_{i}\right)\mathbf{V}^{u}_{i,k}\right]+\text{Tr}\left[\mathbf{B}^{T}_{i}\mathcal{Z}\mathbf{B}_{i}\mathcal{J}^{d}_{i,k}\right]+\text{Tr}\left[\mathcal{Z}\mathbf{W}\right]\right]$
$\displaystyle+(1-\xi^{u}_{i,k})\left[\|\mathbf{A}^{c}_{i}\hat{\mathbf{x}}^{u}_{i,k}\|_{\mathcal{Z}^{\frac{1}{2}}}^{2}+\text{Tr}\left[\mathbf{A}^{c^{T}}_{i}\mathcal{Z}\mathbf{A}^{c}_{i}\mathcal{J}^{u}_{i,k}\right]+\text{Tr}\left[\left(\mathbf{B}_{i}\mathbf{\Phi}_{i}\right)^{T}\mathcal{Z}\left(\mathbf{B}_{i}\mathbf{\Phi}_{i}\right)\mathcal{J}^{u}_{i,k}\right]+\text{Tr}\left[\mathbf{B}^{T}_{i}\mathcal{Z}\mathbf{B}_{i}\mathcal{J}^{d}_{i,k}\right]+\text{Tr}\left[\mathcal{Z}\mathbf{W}\right]\right].$
Step $(1)$ is a result of using the quadratic Lyapunov function. The step
$(2)$ holds when applying the remote sensing-loop and open-loop state
evolution in (30) and (29), respectively. The step $(3)$ holds using the
expectation in (70) with respect to the state estimation error
$\mathbf{e}^{u}_{i,k}$, the action prediction error $\mathbf{e}^{d}_{i,k}$,
and the plant noise $\mathbf{w}_{k}$. As a consequence of obtaining the
expected current Lyapunov value in (68) and the expected future Lyapunov value
of the UL transmission in (70), the control stability constraint in (36) for
the UL transmission is
$\displaystyle\xi^{u}_{i,k}\mathbb{E}\left[\mathcal{L}(\mathbf{x}^{s}_{i,k+1})|\hat{\mathbf{x}}^{u}_{i,k},\hat{\mathbf{u}}^{d}_{i,k},\mathbf{H}^{u}_{i,k},P^{u}_{i,k}\right]+(1-\xi^{u}_{i,k})\mathbb{E}\left[\mathcal{L}(\mathbf{x}^{o}_{i,k+1})|\hat{\mathbf{x}}^{u}_{i,k},\hat{\mathbf{u}}^{d}_{i,k},\mathbf{H}^{u}_{i,k},P^{u}_{i,k}\right]\leq\zeta_{i}\mathbb{E}\left[\mathcal{L}(\mathbf{x}^{u}_{i,k})|\hat{\mathbf{x}}^{u}_{i,k}\right]$
(71)
$\displaystyle=\xi^{u}_{i,k}\left[\text{Tr}\left[(\mathbf{B}_{i}\mathbf{\Phi}_{i})^{T}\mathcal{Z}(\mathbf{B}_{i}\mathbf{\Phi}_{i})\mathbf{V}^{u}_{i,k}\right]-\text{Tr}\left[(\mathbf{B}_{i}\mathbf{\Phi}_{i})^{T}\mathcal{Z}(\mathbf{B}_{i}\mathbf{\Phi}_{i})\mathcal{J}^{u}_{i,k}\right]\right]+\|\mathbf{A}^{c}_{i}\hat{\mathbf{x}}^{u}_{i,k}\|_{\mathcal{Z}^{\frac{1}{2}}}^{2}+\text{Tr}\left[\mathbf{A}^{T}_{i}\mathcal{Z}\mathbf{A}_{i}\mathcal{J}^{u}_{i,k}\right]$
$\displaystyle\qquad+\text{Tr}\left[\mathbf{B}^{T}_{i}\mathcal{Z}\mathbf{B}_{i}\mathcal{J}^{d}_{i,k}\right]+\text{Tr}\left[\mathcal{Z}\mathbf{W}\right]\leq\zeta_{i}\|\hat{\mathbf{x}}^{u}_{i,k}\|_{\mathcal{Z}^{\frac{1}{2}}}^{2}+\zeta_{i}\text{Tr}\left[\mathcal{Z}\mathcal{J}^{u}_{i,k}\right]$
After rearranging the terms in (LABEL:lemm2.3), the constraint on the UL
transmission indicator variable is
$\displaystyle\xi^{u}_{i,k}\geq\frac{\|\left(\mathbf{A}^{c}_{i}-\zeta_{i}\mathbf{I}_{D}\right)\hat{\mathbf{x}}^{u}_{i,k}\|_{\mathcal{Z}^{\frac{1}{2}}}^{2}+\text{Tr}\left[\left(\mathbf{A}^{T}_{i}\mathcal{Z}\mathbf{A}_{i}-\zeta_{i}\mathcal{Z}\right)\mathcal{J}^{u}_{i,k}\right]+\text{Tr}\left[\mathbf{B}^{T}_{i}\mathcal{Z}\mathbf{B}_{i}\mathcal{J}^{d}_{i,k}\right]+\text{Tr}\left[\mathcal{Z}\mathbf{W}\right]}{\text{Tr}\left[\left(\mathbf{B}_{i}\mathbf{\Phi}_{i}\right)^{T}\mathcal{Z}\left(\mathbf{B}_{i}\mathbf{\Phi}_{i}\right)\mathcal{J}^{u}_{i,k}\right]-\text{Tr}\left[\left(\mathbf{B}_{i}\mathbf{\Phi}_{i}\right)^{T}\mathcal{Z}\left(\mathbf{B}_{i}\mathbf{\Phi}_{i}\right)\mathbf{V}^{u}_{i,k}\right]}.$
(72)
To capture the overall state evolution of each control system in the UL over
time interval of length $K$, according to the time-averaged Lyapunov [27],
(53) is summed over time ${\scriptstyle k\in\\{0,\cdots,K-1\\}}$, then the
result is divided by $K$ and taking the limit as time tends to infinity. This
yields the UL transmission indicator variable constraint in (38).
Similarly, we obtain the DL transmission indicator variable constraint in (39)
according to the remote actuating-loop state evolution in (31) and open-loop
state evolution in (29). Finally, the expected future Lyapunov value of the
UL-DL coupled transmission is obtained using the closed-loop state evolution
in (32) and the open-loop state evolution in (29) where
$\displaystyle\mathbb{E}\Big{[}\mathcal{L}(\mathbf{x}^{u}_{i,k+1})|\hat{\mathbf{x}}^{u}_{i,k},\hat{\mathbf{u}}^{d}_{i,k},\mathbf{H}^{u}_{i,k},P^{u}_{i,k},$
$\displaystyle\mathbf{H}^{d}_{i,k},P^{d}_{i,k}\Big{]}=\xi^{u}_{i,k}\;\xi^{d}_{i,k}\;\mathbb{E}\left[\mathcal{L}(\mathbf{x}^{c}_{i,k+1})|\hat{\mathbf{x}}^{u}_{i,k},\hat{\mathbf{u}}^{d}_{i,k},\mathbf{H}^{u}_{i,k},P^{u}_{i,k},\mathbf{H}^{d}_{i,k},P^{d}_{i,k}\right]$
(73)
$\displaystyle+\left(1-\xi^{u}_{i,k}\;\xi^{d}_{i,k}\right)\;\mathbb{E}\left[\mathcal{L}(\mathbf{x}^{o}_{i,k+1})|\hat{\mathbf{x}}^{u}_{i,k},\hat{\mathbf{u}}^{d}_{i,k},\mathbf{H}^{u}_{i,k},P^{u}_{i,k},\mathbf{H}^{d}_{i,k},P^{d}_{i,k}\right],$
For a given predicted state $\hat{\mathbf{x}}^{u}_{i,k}$, the state prediction
error covariance matrix $\mathcal{J}^{u}_{i,k}$, the predicted action
$\hat{\mathbf{u}}^{d}_{i,k}$, the action prediction error covariance matrix
$\mathcal{J}^{d}_{i,k}$, the wireless UL channel $\mathbf{H}^{u}_{i,k}$, the
wireless DL channel $\mathbf{H}^{d}_{i,k}$, the UL transmission power
$P^{u}_{i,k}$, and the DL transmission power $P^{d}_{i,k}$, the expected
future value of $\mathcal{L}(\mathbf{x}^{u}_{i,k+1})$ of the UL-DL coupling
transmission in (73) is given as
$\displaystyle\mathbb{E}$
$\displaystyle\left[\mathcal{L}(\mathbf{x}^{u}_{i,k+1})|\hat{\mathbf{x}}^{u}_{i,k},\hat{\mathbf{u}}^{d}_{i,k},\mathbf{H}^{u}_{i,k},P^{u}_{i,k},\mathbf{H}^{d}_{i,k},P^{d}_{i,k}\right]=\xi^{u}_{i,k}\;\xi^{d}_{i,k}\mathbb{E}\left[\mathbf{x}^{c^{T}}_{i,k+1}\mathcal{Z}\mathbf{x}^{c}_{i,k+1}|\hat{\mathbf{x}}^{u}_{i,k},\hat{\mathbf{u}}^{d}_{i,k},\mathbf{H}^{u}_{i,k},P^{u}_{i,k},\mathbf{H}^{d}_{i,k},P^{d}_{i,k}\right]$
(74) $\displaystyle\hskip
160.0pt+\left(1-\xi^{u}_{i,k}\;\xi^{d}_{i,k}\right)\;\mathbb{E}\left[\mathbf{x}^{o^{T}}_{i,k+1}\mathcal{Z}\mathbf{x}^{o}_{i,k+1}|\hat{\mathbf{x}}^{u}_{i,k},\hat{\mathbf{u}}^{d}_{i,k},\mathbf{H}^{u}_{i,k},P^{u}_{i,k}\mathbf{H}^{d}_{i,k},P^{d}_{i,k}\right]$
$\displaystyle=\xi^{u}_{i,k}\;\xi^{d}_{i,k}\mathbb{E}\left[\left(\mathbf{A}^{c}_{i}\hat{\mathbf{x}}^{u}_{i,k}-\mathbf{A}^{c}_{i}\mathbf{e}^{u}_{i,k}-\mathbf{B}_{i}\mathbf{\Phi}_{i}\mathbf{v}^{c}_{i,k}-\mathbf{B}_{i}\mathbf{v}^{d}_{i,k}+\mathbf{w}_{k}\right)^{T}\mathcal{Z}\left(\mathbf{A}^{c}_{i}\hat{\mathbf{x}}^{u}_{i,k}-\mathbf{A}^{c}_{i}\mathbf{e}^{u}_{i,k}-\mathbf{B}_{i}\mathbf{\Phi}_{i}\mathbf{v}^{c}_{i,k}-\mathbf{B}_{i}\mathbf{v}^{d}_{i,k}+\mathbf{w}_{k}\right)\right]$
$\displaystyle+(1-\xi^{u}_{i,k}\;\xi^{d}_{i,k})\mathbb{E}\left[\left(\mathbf{A}^{c}_{i}\hat{\mathbf{x}}^{u}_{i,k}-\mathbf{A}^{c}_{i}\mathbf{e}^{u}_{i,k}-\mathbf{B}_{i}\mathbf{\Phi}_{i}\mathbf{e}^{u}_{i,k}-\mathbf{B}_{i}\mathbf{e}^{d}_{i,k}+\mathbf{w}_{k}\right)^{T}\mathcal{Z}\left(\mathbf{A}^{c}_{i}\hat{\mathbf{x}}^{u}_{i,k}-\mathbf{A}^{c}_{i}\mathbf{e}^{u}_{i,k}-\mathbf{B}_{i}\mathbf{\Phi}_{i}\mathbf{e}^{u}_{i,k}-\mathbf{B}_{i}\mathbf{e}^{d}_{i,k}+\mathbf{w}_{k}\right)\right]$
$\displaystyle=\xi^{u}_{i,k}\;\xi^{d}_{i,k}\left[\|\mathbf{A}^{c}_{i}\hat{\mathbf{x}}^{u}_{i,k}\|_{\mathcal{Z}^{\frac{1}{2}}}^{2}+\text{Tr}\left[\mathbf{A}^{c^{T}}_{i}\mathcal{Z}\mathbf{A}^{c}_{i}\mathcal{J}^{u}_{i,k}\right]+\text{Tr}\left[(\mathbf{B}_{i}\mathbf{\Phi}_{i})^{T}\mathcal{Z}(\mathbf{B}_{i}\mathbf{\Phi}_{i})\mathbf{V}^{u}_{i,k}\right]+\text{Tr}\left[\mathbf{B}^{T}_{i}\mathcal{Z}\mathbf{B}_{i}\mathbf{V}^{d}_{i,k}\right]+\text{Tr}\left[\mathcal{Z}\mathbf{W}\right]\right]$
$\displaystyle+(1-\xi^{u}_{i,k}\;\xi^{d}_{i,k})\left[\|\mathbf{A}^{c}_{i}\hat{\mathbf{x}}^{u}_{i,k}\|_{\mathcal{Z}^{\frac{1}{2}}}^{2}+\text{Tr}\left[\mathbf{A}^{c^{T}}_{i}\mathcal{Z}\mathbf{A}^{c}_{i}\mathcal{J}^{u}_{i,k}\right]+\text{Tr}\left[(\mathbf{B}^{T}_{i}\mathbf{\Phi}_{i})^{T}\mathcal{Z}(\mathbf{B}^{T}_{i}\mathbf{\Phi}_{i})\mathcal{J}^{u}_{i,k}\right]+\text{Tr}\left[\mathbf{B}^{T}_{i}\mathcal{Z}\mathbf{B}_{i}\mathcal{J}^{d}_{i,k}\right]+\text{Tr}\left[\mathcal{Z}\mathbf{W}\right]\right]$
Given the expected current value of $\mathcal{L}(\mathbf{x}^{u}_{i,k})$ in
(68) and future of $\mathcal{L}(\mathbf{x}^{u}_{i,k+1})$ of the UL-DL coupled
transmission in (74), the control stability constraint in (36) for UL-DL
coupled transmission is
$\displaystyle\xi^{u}_{i,k}\;\xi^{d}_{i,k}\mathbb{E}\left[\mathcal{L}(\mathbf{x}^{a}_{i,k+1})|\hat{\mathbf{x}}^{u}_{i,k},\hat{\mathbf{u}}^{d}_{i,k},\mathbf{H}^{u}_{i,k},P^{u}_{i,k}\mathbf{H}^{d}_{i,k},P^{d}_{i,k}\right]+(1-\xi^{u}_{i,k}\;\xi^{d}_{i,k})\mathbb{E}\Big{[}\mathcal{L}(\mathbf{x}^{o}_{i,k+1})|\hat{\mathbf{x}}^{u}_{i,k},\hat{\mathbf{u}}^{d}_{i,k},\mathbf{H}^{u}_{i,k},P^{u}_{i,k}\mathbf{H}^{d}_{i,k},P^{d}_{i,k}\Big{]}$
(75)
$\displaystyle\leq\zeta_{i}\mathbb{E}\left[\mathcal{L}(\mathbf{x}^{u}_{i,k})|\hat{\mathbf{x}}^{u}_{i,k}\right]=\xi^{u}_{i,k}\;\xi^{d}_{i,k}\Big{[}\text{Tr}\left[(\mathbf{B}_{i}\mathbf{\Phi}_{i})^{T}\mathcal{Z}(\mathbf{B}_{i}\mathbf{\Phi}_{i})\mathbf{V}^{u}_{i,k}\right]+\text{Tr}\left[\mathbf{B}_{i}^{T}\mathcal{Z}\mathbf{B}_{i}\mathbf{V}^{d}_{i,k}\right]-\text{Tr}\left[(\mathbf{B}_{i}\mathbf{\Phi}_{i})^{T}\mathcal{Z}(\mathbf{B}_{i}\mathbf{\Phi}_{i})\mathcal{J}^{u}_{i,k}\right]$
$\displaystyle-\text{Tr}\left[\mathbf{B}_{i}^{T}\mathcal{Z}\mathbf{B}_{i}\mathcal{J}^{d}_{i,k}\right]\Big{]}+\|\mathbf{A}^{c}_{i}\hat{\mathbf{x}}^{u}_{i,k}\|_{\mathcal{Z}^{\frac{1}{2}}}^{2}+\text{Tr}\left[\mathbf{A}^{T}_{i}\mathcal{Z}\mathbf{A}_{i}\mathcal{J}^{u}_{i,k}\right]+\text{Tr}\left[\mathbf{B}^{T}_{i}\mathcal{Z}\mathbf{B}_{i}\mathcal{J}^{d}_{i,k}\right]+\text{Tr}\left[\mathcal{Z}\mathbf{W}\right]\leq\zeta_{i}\|\hat{\mathbf{x}}^{u}_{i,k}\|_{\mathcal{Z}^{\frac{1}{2}}}^{2}+\zeta_{i}\text{Tr}\left[\mathcal{Z}\mathcal{J}^{u}_{i,k}\right]$
The UL-DL transmission indicator variable constraint after rearranging the
terms in (LABEL:lemm2.11) is:
$\displaystyle\xi^{u}_{i,k}\;\xi^{d}_{i,k}\geq\frac{\|\left(\mathbf{A}^{c}_{i}-\zeta_{i}\mathbf{I}_{D}\right)\hat{\mathbf{x}}^{u}_{i,k}\|_{\mathcal{Z}^{\frac{1}{2}}}^{2}+\text{Tr}\left[\left(\mathbf{A}^{T}_{i}\mathcal{Z}\mathbf{A}_{i}-\zeta_{i}\mathcal{Z}\right)\mathcal{J}^{u}_{i,k}\right]+\text{Tr}\left[\mathbf{B}^{T}_{i}\mathcal{Z}\mathbf{B}_{i}\mathcal{J}^{d}_{i,k}\right]+\text{Tr}\left[\mathcal{Z}\mathbf{W}\right]}{\text{Tr}\left[\left((\mathbf{B}_{i}\mathbf{\Phi}_{i})^{T}\mathcal{Z}(\mathbf{B}_{i}\mathbf{\Phi}_{i})\right)(\mathcal{J}^{u}_{i,k}-\mathbf{V}^{u}_{i,k})\right]+\text{Tr}\left[\mathbf{B}_{i}^{T}\mathcal{Z}\mathbf{B}_{i}(\mathcal{J}^{d}_{i,k}-\mathbf{V}^{d}_{i,k}\right]}$
(76)
At last, we apply the time-averaged Lyapunov [27] to obtain the overall state
evolution per control system in the UL-DL coupled transmission and UL-DL
coupled transmission indicator variable constraint in (40). $\blacksquare$
## References
* [1] A. M. Girgis, J. Park, C. F. Liu, and M. Bennis, “Predictive control and communication co-design: A gaussian process regression approach,” in _2020 IEEE 21st International Workshop on Signal Processing Advances in Wireless Communications (SPAWC)_ , 2020, pp. 1–5.
* [2] P. Park, S. C. Ergen, C. Fischione, C. Lu, and K. H. Johansson, “Wireless network design for control systems: A survey,” _IEEE Communications Surveys & Tutorials_, vol. 20, no. 2, pp. 978–1013, 2017.
* [3] J. Park, S. Samarakoon, H. Shiri, M. K. Abdel-Aziz, T. Nishio, A. Elgabli, and M. Bennis, “Extreme URLLC: Vision, challenges, and key enablers,” _arXiv preprint arXiv:2001.09683_ , 2020.
* [4] M. Bennis, M. Debbah, and H. V. Poor, “Ultrareliable and low-latency wireless communication: Tail, risk, and scale,” _Proceedings of the IEEE_ , vol. 106, no. 10, pp. 1834–1853, 2018.
* [5] J. Park, S. Samarakoon, M. Bennis, and M. Debbah, “Wireless network intelligence at the edge,” _Proceedings of the IEEE_ , vol. 107, no. 11, pp. 2204–2239, 2019.
* [6] C.-F. Liu and M. Bennis, “Taming the tail of maximal information age in wireless industrial networks,” _IEEE Communications Letters_ , vol. 23, no. 12, pp. 2442–2446, 2019.
* [7] T. Zeng, O. Semiari, W. Saad, and M. Bennis, “Joint communication and control for wireless autonomous vehicular platoon systems,” _IEEE Transactions on Communications_ , vol. 67, no. 11, pp. 7907–7922, 2019.
* [8] J. P. Hespanha, P. Naghshtabrizi, and Y. Xu, “A survey of recent results in networked control systems,” _Proceedings of the IEEE_ , vol. 95, no. 1, pp. 138–162, 2007.
* [9] L. Schenato, B. Sinopoli, M. Franceschetti, K. Poolla, and S. S. Sastry, “Foundations of control and estimation over lossy networks,” _Proceedings of the IEEE_ , vol. 95, no. 1, pp. 163–187, 2007.
* [10] Y. Xu, H. Su, Y.-J. Pan, Z.-G. Wu, and W. Xu, “Stability analysis of networked control systems with round-robin scheduling and packet dropouts,” _Journal of the Franklin Institute_ , vol. 350, no. 8, 2013.
* [11] X. Liu, E. K. Chong, and N. B. Shroff, “A framework for opportunistic scheduling in wireless networks,” _Computer networks_ , vol. 41, no. 4, pp. 451–474, 2003.
* [12] A. Cervin and T. Henningsson, “Scheduling of event-triggered controllers on a shared network,” in _2008 47th IEEE Conference on Decision and Control_. IEEE, 2008, pp. 3601–3606.
* [13] R. Postoyan, P. Tabuada, D. Nešić, and A. Anta, “Event-triggered and self-triggered stabilization of distributed networked control systems,” in _2011 50th IEEE Conference on Decision and Control and European Control Conference_. IEEE, 2011, pp. 2565–2570.
* [14] M. Eisen, M. M. Rashid, K. Gatsis, D. Cavalcanti, N. Himayat, and A. Ribeiro, “Control aware communication design for time sensitive wireless systems,” in _IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)_. IEEE, 2019, pp. 4584–4588.
* [15] G. Zhao, M. A. Imran, Z. Pang, Z. Chen, and L. Li, “Toward real-time control in future wireless networks: communication-control co-design,” _IEEE Communications Magazine_ , vol. 57, no. 2, pp. 138–144, 2019.
* [16] W. Yan, H. Qiu, and Y. Xue, “Gaussian process for long-term time-series forecasting,” in _2009 International Joint Conference on Neural Networks_. IEEE, 2009, pp. 3420–3427.
* [17] K. J. Åström and B. Wittenmark, _Adaptive control_. Courier Corporation, 2013.
* [18] M. Hassanin, O. Fresnedo, J. Garcia-Frias, and L. Castedo, “Analog joint source channel coding for gaussian multiple access channels,” in _International Workshop on Multiple Access Communications_. Springer, 2013, pp. 21–32.
* [19] A. Kosta, N. Pappas, V. Angelakis _et al._ , “Age of information: A new concept, metric, and tool,” _Foundations and Trends® in Networking_ , vol. 12, no. 3, pp. 162–259, 2017.
* [20] A. V. Oppenheim and G. C. Verghese, “Introduction to communication, control, and signal processing,” _Ch8_ , 2010.
* [21] M. A. Alvarez, L. Rosasco, N. D. Lawrence _et al._ , “Kernels for vector-valued functions: A review,” _Foundations and Trends® in Machine Learning_ , vol. 4, no. 3, pp. 195–266, 2012\.
* [22] C. E. Rasmussen, “Gaussian processes in machine learning,” in _Summer School on Machine Learning_. Springer, 2003, pp. 63–71.
* [23] L. Li, M. Pal, and Y. R. Yang, “Proportional fairness in multi-rate wireless lans,” in _IEEE INFOCOM 2008-The 27th Conference on Computer Communications_. IEEE, 2008, pp. 1004–1012.
* [24] M. J. Neely, “Stability and capacity regions or discrete time queueing networks,” _arXiv preprint arXiv:1003.3396_ , 2010.
* [25] M. S. Branicky, S. M. Phillips, and W. Zhang, “Scheduling and feedback co-design for networked control systems,” in _Proceedings of the 41st IEEE Conference on Decision and Control, 2002._ , vol. 2. IEEE, 2002, pp. 1211–1217.
* [26] A. Sorjamaa, J. Hao, N. Reyhani, Y. Ji, and A. Lendasse, “Methodology for long-term prediction of time series,” _Neurocomputing_ , vol. 70, no. 16-18, pp. 2861–2869, 2007.
* [27] A. N. Michel and L. Hou, “Stability results involving time-averaged lyapunov function derivatives,” _Nonlinear Analysis: Hybrid Systems_ , vol. 3, no. 1, pp. 51–64, 2009.
* [28] M. Gastpar and M. Vetterli, “Source-channel communication in sensor networks,” in _Information Processing in Sensor Networks_. Springer, 2003, pp. 162–177.
* [29] M. S. Mahmoud, _Control and estimation methods over communication networks_. Springer, 2014.
|
11institutetext: Berkeley Center for Theoretical Physics and Department of
Physics,
University of California, Berkeley, CA 94720, USA22institutetext: Stanford
Institute for Theoretical Physics,
Stanford University, Stanford, CA 94305, USA
# Island Finder and Entropy Bound
Raphael Bousso 2 and Arvin Shahbazi-Moghaddam<EMAIL_ADDRESS><EMAIL_ADDRESS>
###### Abstract
Identifying an entanglement island requires exquisite control over the entropy
of quantum fields, which is available only in toy models. Here we present a
set of sufficient conditions that guarantee the existence of an island and
place an upper bound on the entropy computed by the island rule. This is
enough to derive the main features of the Page curve for an evaporating black
hole in any spacetime dimension. Our argument makes use of Wall’s maximin
formulation and the Quantum Focusing Conjecture. As a corollary, we derive a
novel entropy bound.
## 1 Introduction
### 1.1 Entanglement Wedges and Islands
The quantum-corrected Engelhardt:2014gca , covariant Hubeny:2007xt Ryu-
Takayanagi Ryu:2006bv (RT) prescription computes the CFT entropy of a
boundary region in terms of a dual asymptotically AdS bulk spacetime.
Originally an ad hoc proposal, it follows under certain assumptions from a
Euclidean gravitational path integral Lewkowycz:2013nqa ; Faulkner:2013ana ;
Dong:2017xht ; Penington:2019kki ; Almheiri:2019qdq . This derivation implies
that the RT prescription is not tied to the AdS/CFT correspondence but can be
evaluated in any spacetime $M$.
Indeed, RT yields the Page curve Page:1993wv for the entropy of the bulk
radiation emitted by a black hole Penington:2019npb ; Almheiri:2019psf . The
bulk state and geometry are treated semiclassically. In this approximation,
the radiation is thermal Hawking:1974sw , and its von Neumann entropy $S(R)$
increases monotonically, implying information loss Hawking:1976ra . Using the
same semiclassical solution, the RT proposal computes the radiation entropy
differently, as the generalized entropy111For a partial Cauchy surface
$X\subset M$, $S_{\rm gen}(X)=\text{Area}[\partial X]/4G\hbar+S(X)$, where
$\partial$ denotes the boundary of a set, and $S(X)$ is the renormalized von
Neumann entropy of the density operator of the quantum field theory state
reduced to $X$. See the appendix in Ref. Bousso:2015mna for a detailed
discussion of this quantity. of the entanglement wedge of the radiation,
$E(R)$:
$S(\mathbf{R})=S_{\rm gen}[E(R)]~{}.$ (1)
The bold-face notation Almheiri:2019yqk distinguishes the (presumably
correct) entropy computed by RT from the entropy $S(R)$ computed directly from
the semiclassical radiation state (See Bousso:2019ykv ; Bousso:2020kmy for a
proposal to reconcile the bold and unbold states.)
The original RT prescription defines an entanglement wedge for regions on the
conformal boundary of AdS. In the present context, $R$ is a bulk system, and
the entanglement wedge must be defined as follows Bousso:2020kmy :
1. 1.
$E(R)=I\cup R$, where $I\subset M$;
2. 2.
$S_{\rm gen}(I\cup R)$ is stationary under any local variations of the
boundary surface $\partial I$;
3. 3.
among all such regions globally, $I$ yields the smallest $S_{\rm gen}(I\cup
R)$.
The above definitions apply if $R$ is a nongravitating system external to $M$;
in asymptotically AdS geometries the radiation can be extracted into such a
system Penington:2019npb ; Almheiri:2019psf . We now turn to the case where
$R$ is a weakly gravitating region inside the spacetime. For example, $R$ may
be a distant region occupied by Hawking radiation in an asymptotically flat or
AdS spacetime (see Ref. Bousso:2019ykv for a detailed setup).
Physically, one expects the RT prescription for a weakly gravitating region to
reduce to that for a nongravitating system, Eq. (1), and we shall assume this
here. $R$ resides in a weakly gravitating region far from any potential island
$I$, so $\partial I\cap\bar{R}=\varnothing$, where an overbar denotes
topological closure. As before, stationarity of $S_{\rm gen}$ is required only
under variations of $\partial I$, not of $\partial R$. (This can be
implemented in a path integral derivation Dong:2020uxp .) Thus, the definition
of $E(R)$ is essentially unchanged.
However, in the presence of gravity, it is simplest to work with the
generalized entropy of the region $\mathbf{R}$ (a cutoff-independent
quantity), so we add its boundary area to both sides of Eq. (1):
$S_{\rm gen}(\mathbf{R})=S_{\rm gen}[E(R)]~{}.$ (2)
It is easy to derive the Page curve, at least approximately, if one ignores
condition 2. We will now summarize this incomplete argument, before discussing
how it can be completed.
Figure 1: Left: evaporating black hole; right: its Page curve. After the Page
time, the semiclassical entropy $S(R)$ of the Hawking radiation in the
asymptotic region $R$ exceeds the Bekenstein-Hawking entropy of the black
hole, $A_{h}/4G\hbar$. The “Hawking partners” in the black hole interior
purify $R$. (Dashed lines indicate entanglement.) Therefore, adjoining
$\hat{I}$ to $R$ decreases the generalized entropy $S_{\rm gen}$. However,
islands must have stationary $S_{\rm gen}$. Solving for this condition exceeds
present analytic control over the entropy. The island finder presented here
sidesteps this obstruction.
The Page time $t_{\rm Page}$ is defined as the time when the black hole and
radiation entropies are equal in the semiclassical analysis:
$S[R(t_{\rm Page})]=\frac{A_{h}(t_{\rm Page})}{4G\hbar}~{}.$ (3)
Let $\hat{I}(t)$ be the black hole interior at time $t$ (see Fig. 1). The
Hawking “partners” in $\hat{I}(t)$ purify the radiation $R(t)$ emitted so far;
hence
$S_{\rm gen}[\hat{I}(t)\cup R]\approx\frac{A_{h}(t)}{4G\hbar}~{}.$ (4)
Before the Page time, this is greater than $S(R)$ by definition, so $\hat{I}$
is not a viable island candidate; one finds that $I(t)=\varnothing$, $E(R)=R$,
and $S(\mathbf{R})=S(R)$. This corresponds to the rising part of the Page
curve, where it agrees with Hawking’s curve. But after the Page time, $S_{\rm
gen}[\hat{I}(t)\cup R]<S(R)$ by Eqs. (3) and (4). Thus, the inclusion of an
island $I(t)\approx\hat{I}(t)\neq\varnothing$ is favored, and we have
$S(\mathbf{R})=S(I\cup R)\approx A_{h}/4G\hbar$. As the black hole evaporates
and its horizon shinks, this yields the decaying part of the Page curve
required by unitarity.
The Page curve result has been extended to asymptotically flat spacetimes
Gautason:2020tmk ; Hartman:2020swn , settings with two layers of holography
Almheiri:2019hni ; Almheiri:2019psy , and eternal black holes Almheiri:2019yqk
. Entanglement islands can also appear in cosmology, where their significance
is less obvious Chen:2020tes ; Hartman:2020khs .
### 1.2 Summary and Outline
Our brief summary of the Page curve result has a major gap. We explained why
condition 3 (global minimization of $S_{\rm gen}$) favors inclusion of the
black hole interior $\hat{I}(t)$ in $E(R)$ after the Page time. However, we
did not show that condition 2 (local extremality) can be satisfied by some
deformation of $\hat{I}(t)$ small enough to preserve condition 3.
One way to fill this gap is to find $I(t)$ exactly, and to verify condition 2.
However, explicit solutions of the quantum extremality condition have been
found only in $1+1$ bulk dimensions Almheiri:2019psf ; Almheiri:2019hni ;
Almheiri:2019yqk , or in toy models of higher-dimensional black holes
Penington:2019npb . The difficulty lies in computing the von Neumann entropy
$S(I\cup R)$. This depends on the detailed state of the dynamics of the
radiation fields, including modes with nonzero angular momentum, and their
interactions. Even free fields scatter nontrivially in a black hole
background, placing an exact calculation out of reach.
In Sec. 2, we develop an alternative way to ensure that condition 2 holds. We
show that the existence of a suitable island $I$ follows from simple
sufficient conditions that are easy to verify:222Ref. Penington:2019npb
presents an elegant existence argument that establishes an island in the
setting of an evaporating black hole. It makes use of properties of the event
horizon and is inequivalent to the more general argument presented here. Let
$I^{\prime}$ be a region that (i) satisfies condition 1, $S_{\rm
gen}(I^{\prime}\cup R)<S(R)$, and suppose that (ii) the generalized entropy of
$I^{\prime}\cup R$ increases under any small outward deformation of $I$, or
decreases under any such deformation. Then there exists an island,
$I\neq\varnothing$; and moreover
$S(\mathbf{R})=S_{\rm gen}(I\cup R)\leq S_{\rm gen}(I^{\prime}\cup R)~{}.$ (5)
We will illustrate in a number of examples that finding a suitable
$I^{\prime}$ is not difficult; in particular, it suffices to understand the
scaling of corrections to simple models of the entropy. Moreover, the upper
bound (5) is powerful enough to establish the main features of the Page curve
for an evaporating black hole.
In Sec. 3, we consider a different but related problem that yields to a
similar analysis. We consider a spacetime and (internal or external) reference
system $R$ in a pure state. We show that the true entropy $S(\mathbf{R})$
cannot exceed the generalized entropy of appropriate bulk regions. For
example, if $R$ is external, and $M$ is an evaporating black hole spacetime,
an upper bound on $S(\mathbf{R})$ is furnished by the generalized entropy of
the bulk region that can be probed by an asymptotic observer (the black hole
exterior).
## 2 Sufficient Conditions for Islands
In this section we identify sufficient conditions for the existence of an
island. In Sec. 2.1 we begin with the case of an external reference system,
$R\cap M=\varnothing$. In Sec. 2.2 we allow $R$ to intersect with $M$. In Sec.
2.3 we consider various examples in which our conditions easily establish the
presence of islands; in particular, we show that they suffice to derive the
Page curve.
### 2.1 External Reference System
Let $M$ be a semiclassical spacetime which together with a reference system
$R$ is in a pure state. We take $R$ to be external to $M$. For definiteness,
we assume that $R$ is nongravitating; otherwise, simply substitute $S\to
S_{\rm gen}$ when the argument contains $R$.
Figure 2: Island finder. Suppose that $I^{\prime}\cup R$ is quantum normal
(top) or anti-normal (bottom). Then the generalized entropy of $I^{\prime}\cup
R$ decreases along the dashed lines to $\tilde{I}^{\prime}\subset\Sigma$. An
island $I$ with even smaller generalized entropy $S_{\rm gen}(I\cup R)$ must
exist on the maximin Cauchy slice $\Sigma$. If $S_{\rm gen}(I^{\prime}\cup
R)<S(R)$, the island cannot be empty.
Suppose there exists a partial Cauchy surface $I^{\prime}\subset M$ satisfying
the following conditions:
$\displaystyle(i)$ $\displaystyle~{}~{}~{}S_{\text{gen}}(I^{\prime}\cup
R)<S(R)~{};$ (6) $\displaystyle(ii)$
$\displaystyle\begin{cases}k^{\mu}\Theta_{\mu}[I^{\prime}\cup R]\geq
0~{},~{}~{}\ell^{\mu}\Theta_{\mu}[I^{\prime}\cup R]\leq 0~{};\\\
\text{~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}or}\\\
k^{\mu}\Theta_{\mu}[I^{\prime}\cup R]\leq
0~{},~{}~{}\ell^{\mu}\Theta_{\mu}[I^{\prime}\cup R]\geq 0~{},\end{cases}$ (7)
where $\Theta$ is the quantum expansion one-form Engelhardt:2014gca ;
Bousso:2015mna and $k$ and $\ell$ are the outward and inward future-directed
null vectors normal to $\partial I^{\prime}$. Thus, for example,
$k^{\mu}\Theta_{\mu}[I^{\prime}\cup R,y]$ is the rate of change of $S_{\rm
gen}$, per unit transverse area and unit affine parameter length, as
$I^{\prime}\cup R$ is deformed outward along the future-directed null geodesic
orthogonal to $I^{\prime}$ at $y$. We drop the argument $y$ when an equation
holds for all $y$.
The first condition states that adjoining $I^{\prime}$ to $R$ decreases the
entropy, even as the Bekenstein-Hawking entropy of $I^{\prime}$ is included.
The second, Eq. (7), says that $I^{\prime}\cup R$ is quantum normal (its
generalized entropy does not decrease under any small outward deformation of
$I^{\prime}$) or quantum anti-normal (the opposite).
We will now show that these conditions imply the existence of a non-empty
island region $I$. Our proof uses the maximin construction of the HRT surface
Wall:2012uf , which we assume extends to a quantum maximin prescription
Akers:2019lzs : on every Cauchy surface of $M$, one finds the region
$I^{\prime\prime}$ that minimizes $S_{\rm gen}(I^{\prime\prime}\cup R)$. (Note
that $I^{\prime\prime}$ may be the empty set.) One then maximizes the same
quantity over all Cauchy surfaces of $M$. The island $I$ is defined to be the
region $I^{\prime\prime}$ that achieves this maximum. (This is expected to
exist Wall:2012uf ; Akers:2019lzs .)
Note that we define $I$ as an achronal region; hence it is non-unique.
Similarly, the maximin Cauchy slice $\Sigma$ is non-unique. The relevant
unique object is the domain of dependence $D(I)$. (We ignore the degenerate
case where there are two islands with identical generalized entropy but
different $D$.) Any other Cauchy slice of $D(I)$ will be an island if $I$ is,
though not every Cauchy slice of $D(I)$ will be part of a maximin slice
$\Sigma$.
Our goal is to show that $I\neq\varnothing$. In the normal case,
$k^{\mu}\Theta_{\mu}[I^{\prime}\cup R]\geq 0$, we define
$\tilde{I}^{\prime}\equiv D(I^{\prime})\cap\Sigma~{},$ (8)
as the representative of $I^{\prime}$ on $\Sigma$. In the anti-normal case,
$k^{\mu}\Theta_{\mu}[I^{\prime}\cup R]\leq 0$, we define the representative
instead as
$\tilde{I}^{\prime}\equiv J(I^{\prime})\cap\Sigma~{},$ (9)
where $J$ denotes all points that can be reached from $I^{\prime}$ along a
causal curve (the future and past of $I^{\prime}$). In either case, note that
$\tilde{I}^{\prime}$ is obtained from $I^{\prime}$ by deforming along an
orthogonal null congruence with initially negative quantum expansion. We
assume the Quantum Focussing Conjecture (QFC) Bousso:2015mna , that the
quantum expansion cannot increase along a null shape deformation. This implies
that
$S_{\rm gen}(\tilde{I}^{\prime}\cup R)\leq S_{\rm gen}(I^{\prime}\cup R)~{}.$
(10)
Since $\Sigma$ is the maximin Cauchy slice,
$S_{\rm gen}(\tilde{I}^{\prime}\cup R)\geq S_{\rm gen}(I\cup R)~{}.$ (11)
Combined with Eq. (10), this implies333This intermediate result is closely
related to corollary 16b of Wall:2012uf . A simple application of our argument
to asymptotically AdS spacetimes yields the following result: given a partial
Cauchy slice $A$ on the asymptotic boundary of AdS, let $X$ be the RT surface
associated to $A$ with homology slice $H$. Now, consider another surface
$X^{\prime}$, homologous to $A$ with homology slice $H^{\prime}$. If
$H^{\prime}$ is a quantum normal or anti-normal region, then $S_{\rm
gen}(H)\leq S_{\rm gen}(H^{\prime})$.
$S_{\rm gen}(I^{\prime}\cup R)\geq S_{\rm gen}(I\cup R)~{}.$ (12)
Using the assumption in Eq. (6), we get
$S_{\rm gen}(I\cup R)<S_{\rm gen}(R)~{}.$ (13)
Therefore, we conclude that $I\neq\varnothing$.
### 2.2 Distant Reference System
In Section 2.1, the reference system $R$ was external to the spacetime $M$. We
will now allow a system that is wholly or partially inside the spacetime:
$R\cap M\neq\varnothing$.
In order to generalize our island finder to this setting, we shall require
that there exists a partial Cauchy slice $I_{0}$ spacelike to $R$ such that
$I_{0}\cup R$ is a quantum normal region with respect to deformations at
$\partial I_{0}$. That is, we require
$\displaystyle\bar{I_{0}}$ $\displaystyle\subset M-\overline{J(R)}~{},$ (14)
$\displaystyle k^{\mu}\Theta_{\mu}[I_{0}\cup R]$ $\displaystyle\geq 0~{},$
(15) $\displaystyle\ell^{\mu}\Theta_{\mu}[I_{0}\cup R]$ $\displaystyle\leq
0~{}.$ (16)
where as before an overbar denotes closure. As before, $\Theta_{\mu}[I_{0}\cup
R]$ is the quantum expansion one-form, and $k^{\mu}$ and $\ell^{\mu}$ are
future-directed null vectors fields in the normal bundle of $\partial I_{0}$
in the outward and inward directions respectively.
For example, these conditions will be satisfied when $I_{0}$ is the interior
of a sufficiently large, approximately round sphere in an asymptotically flat
spacetime; and $R$ is the exterior of a slightly larger sphere concentric with
the first, or any subsystem thereof (such as the Hawking radiation emitted by
a black hole). Note that we do not require that $R$ be weakly gravitating, but
in many examples of interest this will be the case. Also, we do not require
that Eqs. (15) and (16) hold at $\partial R$.
Now, suppose there exists a partial Cauchy slice $I^{\prime}\subset D(I_{0})$
satisfying the conditions (6) and (7). That is, $I^{\prime}\cup R$ is quantum
normal or anti-normal, and adjoining $I^{\prime}$ to $R$ reduces the
generalized entropy of $R$. Then there exists a non-empty quantum extremal
region $\hat{I}\subset D(I_{0})$ satisfying $S_{\rm gen}(\hat{I}\cup R)<S_{\rm
gen}(R)$.
Note that this conclusion is weaker than in Sec. 2.1: $\hat{I}$ need not
globally minimize $S_{\rm gen}(I\cup R)$ among all eligible regions, since we
are restricting our search to a subset of $M$. However, the true entanglement
wedge can only have lower entropy, so $S_{\rm gen}(\hat{I}\cup R)$ provides an
upper bound. Note also that when $I^{\prime}\cup R$ is quantum normal, we can
set $I_{0}=I^{\prime}$, so there is no need to identify a larger $I_{0}$
region.
#### Proof, part I
The proof strategy is similar to that in Section 2.1, except that we wish to
restrict our search to the closed set $D(I_{0})$.444A maximin procedure
restricted to entanglement wedges of AdS without reflecting boundary
conditions was considered in Marolf:2019bgj . See also Appendix B of
Brown:2019rox for a related discussion. We will first assume strict version
of conditions (15) and (16):
$\displaystyle k^{\mu}\Theta_{\mu}[I_{0}\cup R]$ $\displaystyle>0~{},$ (17)
$\displaystyle\ell^{\mu}\Theta_{\mu}[I_{0}\cup R]$ $\displaystyle<0~{},$ (18)
Later, we will demonstrate how to relax this assumption back to conditions
(15) and (16).
Let $\hat{I}$ be the maximin region of $D(I_{0})\cup R$, and let $\Sigma$ be a
Cauchy surface of $D(I_{0})$ on which $\hat{I}$ minimizes the generalized
entropy of $\hat{I}\cup R$ among all subregions of $\Sigma$. Without loss of
generality, we may take $\hat{I}\subset\Sigma$: since any Cauchy surface of
$D(\hat{I})$ is an equally good maximin region, we set $\hat{I}\to\Sigma\cap
D(\hat{I})$. As in Ref. Wall:2012uf , we assume that $\hat{I}$ is stable: any
nearby Cauchy surface $\Sigma^{\prime}$ obtained by infinitesimal deformation
of $\Sigma$ contains a locally minimal region $\hat{I}^{\prime}$
infinitesimally close to $\hat{I}$ with $S_{\rm gen}(\hat{I}^{\prime}\cup
R)\leq S_{\rm gen}(\hat{I}\cup R)$.
It immediately follows from the analysis of Sec. 2.1 that
$\displaystyle S_{\rm gen}(\hat{I}\cup R)<S_{\rm gen}(R)~{}.$ (19)
We will now show that $\partial\hat{I}\cap\partial D(I_{0})=\varnothing$. This
precludes the (unwanted) possibility that $I\cup R$ is maximin but not locally
stationary. It follows that $\hat{I}\cup R$ is a quantum extremal region
Wall:2012uf .
$\partial D(I_{0})$ is the disjoint union of three sets: $\partial I_{0}$, and
two null hypersurfaces $N_{+\ell}$ and $N_{-k}$ that lie in the future and
past of $I_{0}$ respectively. The latter sets are generated by future and
past-directed null geodesics orthogonal to $\partial I_{0}$ in the inward
direction which end at caustics or self-intersections Akers:2017nrr . Let
$\ell^{\mu}$ ($k^{\mu}$) be the normal vector field to $N_{+\ell}$ ($N_{-k}$)
obtained by parallel propagation of $\ell^{\mu}|_{\partial I_{0}}$
($k^{\mu}|_{\partial I_{0}}$).
Let $\Sigma_{M}$ be a Cauchy surface of $M$ that intersects each null
generator of $N_{+\ell}$ and $N_{-k}$ at most at one point. Let
$\Sigma_{N}\equiv\Sigma_{M}\cap D(I_{0})$. By Eqs. (17), (18), and the QFC
Bousso:2015mna ,
$\displaystyle k^{\mu}\Theta_{\mu}[\Sigma_{N}\cup R;p]$
$\displaystyle>0~{}~{}~{}\text{ for all }p\in N_{-k}\cap\partial I_{0}~{};$
(20) $\displaystyle\ell^{\mu}\Theta_{\mu}[\Sigma_{N}\cup R;p]$
$\displaystyle<0~{}~{}~{}\text{ for all }p\in N_{+\ell}\cap\partial I_{0}~{},$
(21)
Suppose for contradiction that there exists a point $q\in\partial
D(I_{0})\cap\partial\hat{I}$. The generator of $\partial D(I_{0})$ that
contains $q$, and hence its tangent vector $\ell^{\mu}$ or $k^{\mu}$, will be
normal to $\partial\hat{I}$ at $q$. (If $q\in\partial I_{0}$, this statement
holds for either generator emanating from $q$.) Since $N_{+\ell}$ ($N_{-k}$)
is nowhere to the past (future) of $\hat{I}$, theorem 1 in Ref. C:2013uza
implies
$\displaystyle k^{\mu}\Theta_{\mu}[\hat{I}\cup R;q]$
$\displaystyle>0~{}~{}~{}\text{ for }q\in N_{-k}\cap\partial I_{0}~{};$ (22)
$\displaystyle\ell^{\mu}\Theta_{\mu}[\hat{I}\cup R;q]$
$\displaystyle<0~{}~{}~{}\text{ for }q\in N_{+\ell}\cap\partial I_{0}~{}.$
(23)
Suppose first that $q\in\partial I_{0}$. In this case the above expansions
imply that a small inward deformation of $\hat{I}$ will decrease the
generalized entropy, in contradiction with the minimality of $S_{\rm
gen}(\hat{I}\cup R)$ among all subregions of the maximin Cauchy surface
$\Sigma$.
We will now demonstrate this rigorously, using the notion of a surface-
orthogonal exponential map Akers:2017nrr ,
$\exp_{K}~{}:~{}NK\to M~{},~{}~{}(p,v)\to c_{p,v}(1)~{},$ (24)
which takes a vector $v$ in the normal bundle $NK$ of a smooth submanifold $K$
at the point $p$ to the point at affine distance 1 along the unique geodesic
through $p$ with tangent vector $v$. We will use $\widetilde{\exp}$ to denote
an exponential map in which the submanifold $\Sigma$ plays the role of $M$ in
the above definition.
Let $\tilde{v}^{\mu}(p)$ be a smooth inward-directed vector field in the
normal bundle to $\partial\hat{I}$, viewed as a submanifold of $\Sigma$. We
define the continuous one-parameter family of inward deformations of $\hat{I}$
as the regions
$\displaystyle\hat{I}(\epsilon)=\text{int}\\{\widetilde{\exp}_{\partial\hat{I}}(p,\epsilon\tilde{v}^{\mu}(p)):p\in\partial\hat{I}\\}~{}.$
(25)
Similarly, in the manifold $M$ we define
$\displaystyle
i(\epsilon)=\hat{I}\cap\text{Int}\\{\exp_{\partial\hat{I}}(p,\epsilon
v^{\mu}(p)):p\in\partial\hat{I}\\}~{},$ (26)
where $v$ is the push-forward of $\tilde{v}$ under the embedding map of
$\Sigma$ into $M$, and $\text{Int}\,X$ denotes the spacetime region spacelike
and interior to $X$. For sufficiently small $\epsilon$, both families are
well-defined. Moreover,
$S_{\rm gen}[\hat{I}(\epsilon)\cup R]=S_{\rm gen}[i(\epsilon)\cup
R]+O(\epsilon^{2})~{}.$ (27)
In $M$, the deformation profile $v^{\mu}\in N\partial\hat{I}$ can be
decomposed as
$\displaystyle v^{\mu}=-ak^{\mu}+b\ell^{\mu}~{},$ (28)
where $a$ and $b$ are positive definite functions. Hence
$\displaystyle\left.\frac{dS_{\rm gen}[\hat{I}(\epsilon)\cup
R]}{d\epsilon}\right|_{\epsilon=0}=\int
dx\sqrt{h}~{}v^{\mu}\Theta_{\mu}[\hat{I}\cup R]~{},$ (29)
where $h$ refers to the intrinsic metric of $\partial\hat{I}$. We now choose
$\tilde{v}\equiv 0$ outside a $\delta$-neighborhood of $q$ in
$\partial\hat{I}$. For small enough $\delta$, $v^{\mu}\Theta_{\mu}[\hat{I}\cup
R]<0$ in the entire $\delta$-neighborhood, by Eqs. (22) and (23) and
continuity. Hence,
$\left.\frac{dS_{\rm gen}[\hat{I}(\epsilon)\cup
R]}{d\epsilon}\right|_{\epsilon=0}=\left.\frac{dS_{\rm gen}[i(\epsilon)\cup
R]}{d\epsilon}\right|_{\epsilon=0}<0~{}.$ (30)
Hence, $\hat{I}$ does not minimize the generalized entropy on $\Sigma$. This
contradicts our assumption that $\hat{I}$ is a maximin region. Therefore, no
such point $q\in\partial\hat{I}\cap\partial I_{0}$ can exist:
$\partial\hat{I}\cap\partial I_{0}=\varnothing~{}.$ (31)
Suppose instead that $q\in\partial D(I_{0})-\partial I_{0}$. In this case,
minimality of $S_{\rm gen}(\hat{I}\cup R)$ on $\Sigma$, together with Eq. (22)
or (23), implies that a small deformation of $\Sigma$ into the interior of
$D(I_{0})$ near $q$ will produce a Cauchy surface of $I_{0}$ whose
minimal-$S_{\rm gen}$ region (together with $R$) has greater generalized
entropy than $\hat{I}\cup R$. But this is impossible if $\hat{I}$ was
constructed by the maximin procedure. Again we will now aim to make this
argument rigorous.
Figure 3: Maximin restricted to the domain of dependence (“wedge”) $D(I_{0})$
returns a region $\hat{I}$ on a maximin slice $\Sigma$. If $I_{0}\cup R$ is
quantum normal, then $\partial\hat{I}$ cannot intersect $\partial D(I_{0})$
(dashed). Left: if $\hat{I}\cap\partial I_{0}\neq\varnothing$, then $S_{\rm
gen}(\hat{I}(\epsilon)\cup R)<S_{\rm gen}(\hat{I}\cup R)$, contradicting the
min of maximin. Right: if $\hat{I}\cap\partial D(I_{0})-\partial I_{0}$ then
$S_{\rm gen}(\hat{I}(\epsilon)\cup R)>S_{\rm gen}(\hat{I}\cup R)$ on the
deformed slice $\Sigma(\epsilon)$ violates the max of maximin.
For definiteness, we assume that $q\in N_{+\ell}$. (If instead $q\in N_{-k}$,
the time reverse of our argument applies.) In any open neighborhood $O(q)$,
$\Sigma$ (and hence $\hat{I}$) must enter the interior,
$O(q)\cap\Sigma\cap\text{int}[D(I_{0})]\neq\varnothing$, or else Eq. (23)
would violate the minimality of $S_{\rm gen}(\hat{I}\cup R)$ on $\Sigma$.
Hence the inward-directed vector field $t^{\mu}$ orthogonal to
$\partial\hat{I}$ and tangent to $\hat{I}$ is spacelike in an open
neighborhood of $q$ on $\partial\hat{I}$. Moreover, $\Sigma$ must contain the
null generator segment of $N_{+\ell}$ connecting $\partial I_{0}$ to $q$, or
$\Sigma$ would fail to be achronal.
Let the achronal hypersurfaces $\Sigma(\epsilon)$ be a smooth one parameter
deformation of $\Sigma$ such that $\Sigma(\epsilon)$ agrees with $\Sigma$
outside a $\delta$-neighborhood of $q$ (and everywhere for $\epsilon=0$). We
also require that $\Sigma(\epsilon_{2})$ is nowhere to the future of
$\Sigma(\epsilon_{1})$ if $\epsilon_{1}<\epsilon_{2}$. The stability
assumption guarantees the existence of a smooth one-parameter family of
regions $\hat{I}(\epsilon)$, each of which locally minimizes $S_{\rm
gen}[\hat{I}(\epsilon)\cup R]$ on the corresponding $\Sigma(\epsilon)$. By the
max step of maximin,
$\frac{dS_{\rm gen}[\hat{I}(\epsilon)\cup R]}{d\epsilon}\leq 0~{}.$ (32)
For small enough $\epsilon$, there exists an infinitesimal vector field
$w^{\mu}$ in the normal bundle of $\partial\hat{I}$ in a neighborhood of $q$
such that
$\partial\hat{I}(\epsilon)=\\{\exp_{\partial\hat{I}}(p,\epsilon
w^{\mu}(p)+O(\epsilon^{2})):p\in\partial\hat{I}\\}~{}.$ (33)
We have
$\frac{dS_{\rm gen}[\hat{I}(\epsilon)\cup R]}{d\epsilon}=\int
dx\sqrt{h}~{}w^{\mu}\Theta_{\mu}{[\hat{I}\cup R]}~{}.$ (34)
Since $t^{\mu}$ is spacelike, it is linearly independent of $\ell^{\mu}$, so
there exists a unique decomposition
$w^{\mu}=ct^{\mu}+d\,(-\ell^{\mu})~{}.$ (35)
with $c$ and $d$ nonnegative functions on $\partial\hat{I}$ that vanish
outside an open neighborhood of $q$. Hence
$\frac{S_{\rm gen}[\hat{I}\cup R]}{d\epsilon}=\int
dx\sqrt{h}~{}[ct^{\mu}\Theta_{\mu}+d\,(-\ell^{\mu})\Theta_{\mu}]~{}.$ (36)
The first term is positive-definite by the minimality of $S_{\rm
gen}[\hat{I}\cup R]$ on $\Sigma$; the second is positive by Eq. (23). Hence
$\frac{S_{\rm gen}[\hat{I}\cup R]}{d\epsilon}>0~{},$ (37)
which contradicts Eq. (32). Therefore,
$\partial\hat{I}\cap N_{+\ell}=\varnothing~{},$ (38)
and a time-reversed argument implies
$\partial\hat{I}\cap N_{-k}=\varnothing~{}.$ (39)
Together with Eq. (31) this establishes that
$\hat{I}\subset\text{int}[D(I_{0})]~{};$ (40)
hence $\hat{I}$ is locally quantum extremal.
#### Proof, part II
We will now discuss what happens if we relax the conditions (17) and (18) to
their non-strict versions (15) and (16). We will argue that while in this case
$\hat{I}$ might not be contained in $\text{int}[D(I_{0})]$, an island
candidate still exists, i.e., there exists $\hat{I}\subset D(I_{0})$ such that
$\hat{I}\cup R$ is quantum extremal and $S_{\rm gen}(\hat{I}\cup R)<S_{\rm
gen}(R)$.
Let us start with the case where there exists at least a point $p\in\partial
I_{0}$ where $k^{\mu}\Theta_{\mu}[I_{0}\cup R;p]>0$ and a point $q$ ($p=q$ is
allowed) where $\ell^{\mu}\Theta_{\mu}[I_{0}\cup R;p]<0$. Then,
$\partial\hat{I}$ cannot be a cross section of $N_{-k}$ ($N_{+\ell}$) because
then by the QFC there would be a point $r$ in the cross section, along the
same generator as $p$ ($q$), where $k^{\mu}\Theta_{\mu}[\hat{I}\cup R;r]>0$
($\ell^{\mu}\Theta_{\mu}[\hat{I}\cup R;r]<0$). As discussed above, this
contradicts maximin.
If $\partial\hat{I}$ only partially coincides with $\partial D(I_{0})$ then
there must exist a point $r$ in the boundary of the intersection set such that
in any sufficiently small neighborhood of it $\partial\hat{I}$ and $N_{-k}$
($N_{+\ell}$) do not coincide. In the $\hbar\to 0$ limit, Lemma B of C:2013uza
implies that there exists a point $s$ in a neighborhood of $r$ such that
$k^{\mu}\theta_{\mu}[\hat{I};s]>0$ ($\ell^{\mu}\theta_{\mu}[\hat{I};s]<0$).
Here, we will assume that the Lemma B of C:2013uza continues to hold when we
replace the classical expansion with the quantum expansion.555Our assumption
is motivated by the semiclassical generalization of many similar conditions on
the classical expansion Bousso:2015mna . Note that if the von Neumann entropy
term in the quantum expansion is $O(G\hbar)$ while the classical expansion is
$O(1)$, Lemma B trivially generalizes to the version with quantum expansions.
We will then conclude that $k^{\mu}\Theta_{\mu}[\partial\hat{I}\cup R;s]>0$
($\ell^{\mu}\Theta_{\mu}[\partial\hat{I}\cup R;s]<0$). However,
$s\in\text{int}[D(I_{0})]$, so a non-zero quantum expansion at $s$ is not
allowed by maximin.
Next, we consider the case where $\ell^{\mu}\Theta_{\mu}[I_{0}\cup R]=0$ (The
case with $k$ and $\ell$ exchanged is similar by time-reflection symmetry). By
the previous arguments, $\partial\hat{I}$ cannot intersect $N_{-k}$. Also, by
the quantum version of Lemma B, $\partial\hat{I}$ cannot intersect $N_{+\ell}$
only in part. We will therefore focus on discussing the case where
$\partial\hat{I}$ is a cross section of $N_{+\ell}$.
Associated with any cross section $L$ of $N_{+\ell}$, we can define a partial
Cauchy slice $\Sigma_{L}$ of $D(I_{0})$ which intersects $N_{+\ell}$ at $L$.
Let $L_{1}$ be the latest cross section of $N_{+\ell}$ such that any other
cross section $L$ in the past of $L_{1}$ satisfies
$\ell^{\mu}\Theta_{\mu}[\Sigma_{L}\cup R]=0$.666If such $L_{1}$ does not
exists, then $N_{+\ell}$ is a semi-infinite stationary null hypersurface
which, at least classically, by the no-hair theorem needs to be a semi-
infinite portion of the horizon of a Kerr-Newman black hole. And since
$k^{\mu}\Theta_{\mu}[I_{0}\cup R]\geq 0$, $\partial I_{0}$ needs to lie fully
in the past of the bifurcation surface which provides us with a classical
extremal surface on $N_{+\ell}$. We then expect to find a quantum extremal
region $\hat{I}\cup R$ such that $\partial\hat{I}$ is either on or near this
classical extremal surface. By the QFC, any cross section $L$ which is partly
in the future of $L_{1}$ needs to contain at least a point $r$ at which
$\ell^{\mu}\Theta_{\mu}[\Sigma_{L}\cup R;r]<0$. Hence, $\partial\hat{I}$ must
be a cross section in the past of $L_{1}$. Furthermore, a maximin Cauchy slice
$\Sigma$ corresponding to $\hat{I}$ cannot intersect the future of $L_{1}$ as
it would violate the minimality of $S_{\rm gen}(\hat{I}\cup R)$. $\Sigma$ then
has to leave $N_{+\ell}$ in the past of $L_{1}$ or on $L_{1}$. Let $L_{2}$ be
the cross section at which $\Sigma$ leaves $N_{+\ell}$. Then,
$k^{\mu}\Theta_{\mu}[\Sigma_{L_{2}}\cup R]\leq 0$ or else the min condition is
violated. Since $k^{\mu}\Theta_{\mu}[I_{0}\cup R]\geq 0$, we expect that there
exists a cross section $L_{3}$ between $\partial I_{0}$ and $L_{2}$ such that
$k^{\mu}\Theta_{\mu}[\Sigma_{L_{3}}\cup R]=0$. In the classical limit in
particular, we expect that the results of Andersson:2007gy showing a similar
existence on spacelike Cauchy surfaces can be applied here by taking a limit
of spatial Cauchy surfaces that approach $N_{+\ell}$. Together with
stationarity along $N_{+\ell}$, this would imply that $\Sigma_{L_{3}}\cup R$
is quantum extremal.
### 2.3 Examples
We will now present some examples where the above sufficient conditions (i)
and (ii) provide an efficient diagnostic for the existence of islands. We will
require no detailed calculations of matter entropy and its derivatives, nor
will we be forced to assume special symmetries, low dimensions of spacetime,
or adopt other toy models. Our sufficient conditions establish the existence
of an island and its key properties, at the cost of not exactly locating the
island.
In a final example, we show that neither of the two sufficient conditions can
be eliminated.
#### Evaporating black hole after the Page time
For concreteness, we pick asymptotically flat boundary conditions, though our
conclusion will not depend on this choice. We furthermore assume that the
black hole mostly radiates massless particles, as will be the case if its
initial mass is sufficiently large.
We consider an evaporating black hole formed in a pure state. At late times,
the spacetime will be approximately spherically symmetric. Let
$r=(A/4\pi)^{1/2}$ be the area radius of spheres. Near the horizon, the metric
is well-approximated by
$ds^{2}=-\left(1-\frac{r_{s}(v)}{r}\right)dv^{2}+2dv\,dr+r^{2}d\Omega^{2}~{},$
(41)
Due to evaporation, $r_{s}(v)$ decreases slowly with retarded time $v$:
$dr_{s}/dv\sim-O(G\hbar/r_{s}^{2})$. For $r\gg r_{s}$, the metric is well-
approximated instead by the outgoing Vaidya metric Abdolrahimi:2016emo , but
this will not be important in our analysis.
Figure 4: Evaporating black hole after the Page time. Hawking radiation has
accumulated in $R$. As shown on the left, the boundary of the past of $R$,
denoted by $r_{0}(v)$, intersects the stretched horizon (shown in purple) at
the sphere $A_{s}$, which together with the $A_{h}$ sphere on the event
horizons reside at retarded time $v_{s}$. We consider candidate regions
$I^{\prime}$ with boundary $\partial I^{\prime}$ on a causal horizon spacelike
to $R$ (grey regions). The generalized second law implies that $S_{\rm
gen}(I^{\prime}\cup R)$ increases under future outward deformations. For
future inward deformations, quantum normalcy follows from the (trivial)
classical normalcy in the dark grey subregion, which is chosen to keep quantum
corrections to the expansion small. The $I^{\prime}$ that minimizes $S_{\rm
gen}(I^{\prime}\cup R)$ subject to these restrictions is shown in pink. Its
boundary is located a $\Delta v=r_{s}\log c_{1}$ to the future of $A_{h}$, as
shown on the right. We show that it provides an extremely tight upper bound on
the true entropy $S(\mathbf{R})=S_{\rm gen}(I\cup R)\leq S_{\rm
gen}(I^{\prime}\cup R)=A_{h}/4G\hbar+O(1)$.
Let $u$ be retarded time on $\mathscr{I}^{+}$, and let $R$ be the portion of
$\mathscr{I}^{+}$ given by $u\leq u_{0}$; see Fig. 4. $R$ is a reservoir that
contains the Hawking radiation emitted until the time $u_{0}$. The boundary of
the past of $R$ is given by $u=u_{0}$. We will be interested in the behavior
of the metric only near the retarded time when this surface intersects the
stretched horizon $r_{s}$, so it will be sufficient to set $r_{s}$ to that
value and neglect its $v$-dependence from here on. Let $A_{s}=4\pi r_{s}^{2}$
be the area of the stretched horizon where it meets the past of $R$. Let
$A_{h}=4\pi r_{h}^{2}$ be the area of the event horizon where it intersects
the future of $A_{s}$; the areas satisfy
$A_{s}=A_{h}+O(G\hbar)~{}.$ (42)
We choose $u_{0}$ late enough so that $S(R)>A_{h}/4G\hbar+\log c_{1}$, where
$\log c_{1}$ will be small in a sense made precise below. That is, $R$ extends
to after the Page time, with a little room to spare. We seek an $I^{\prime}$
that satisfies our sufficient conditions while placing a tight bound on the
entropy $S(\mathbf{R})$.
Let $I^{\prime}(r,v)$ be a Cauchy slice of the interior of the sphere $(r,v)$.
Since $S_{\rm gen}(I^{\prime}\cup R)$ is well-defined only for achronal
$I^{\prime}\cup R$, we require $I^{\prime}(r,v)\subset M-J^{-}(R)$, or
$u>u_{0}$. In the ingoing coordinates of Eq. (41), $u=u_{0}$ corresponds to a
function $r_{0}(v)$, defined implicitly by $v=u_{0}+2r_{*}(r_{0})$, where
$r_{*}(r)=r+r_{h}\log(\frac{r}{r_{h}}-1)$. Near the horizon, this satisfies
$\Delta r_{0}(v)\equiv
r_{0}(v)-r_{h}=r_{h}\exp\left(\frac{v-u_{0}}{2r_{h}}-1\right)~{}.$ (43)
We thus require $r<r_{0}(v)$ for the boundary of $I^{\prime}$.
Quantum normalcy of $I^{\prime}\cup R$ requires that the generalized entropy
grows along both of the null directions away from $I^{\prime}$. Any future
outward light-cone outside the horizon is a null surface of constant $u$ that
reaches $\mathscr{I}^{+}$. Hence it is a causal horizon. The generalized
second law of thermodynamics Bekenstein:1972tm ; Wall:2011hj applies to all
causal horizons. It implies that the future outward quantum expansion at
$I^{\prime}$ is positive if $\partial I^{\prime}$ is outside the horizon. (In
more general settings that are not exactly spherically symmetric, we can
accomplish the same goal by choosing $\partial I^{\prime}$ to be a cut of a
causal horizon.) We thus require $r>r_{h}$.
The past outward classical expansion is trivially positive:
$\partial_{r}A=8\pi r$. Quantum normalcy follows if the quantum correction,
$4G\hbar\,\partial_{r}S(I^{\prime}\cup R)$, is negligible, i.e., if
$\partial_{r}S(I^{\prime}\cup R)\ll r/G\hbar$. Let $\Sigma$ be a global Cauchy
slice containing $I^{\prime}\cup R$ and define
$I^{\prime}_{c}=\Sigma-(I^{\prime}\cup R)$. By purity of the global quantum
state, $S(I^{\prime}\cup R)=S(I^{\prime}_{c})$.
Dimensional analysis dictates that the leading divergence in the renormalized
entropy scales as
$\partial_{\rho}S(I^{\prime}_{c})\sim O(1/\rho)~{},$ (44)
where $\rho=r_{0}-r$ and we may assume $\rho\ll r_{0}$. To see this, suppose
first that the only available scales are $r$ and $\rho$. Terms stronger than
Eq. (44) would be of the form $\partial_{r}S(I^{\prime}_{c})\sim
r^{n}/\rho^{n+1}$ with $n>0$ and positive coefficient. Such a term would imply
that at fixed $\rho$, $dS/dr<0$, which is not physically sensible. In the
presence of an additional mass scale $m\sim\hbar/\lambda$, an enhancement of
Eq. (44) would have to take the form
$\partial_{r}S(I^{\prime}_{c})\sim\lambda^{n}/\rho^{n+1}$, $n>0$. Formally,
this is an enhancement for $\rho\ll\lambda$, but physically, a mass scale
cannot have any physical effect in this UV regime.
Hence, quantum normalcy is assured if we require
$r_{0}-r>c_{1}\frac{G\hbar}{r}~{},~{}~{}c_{1}\gg 1~{}.$ (45)
To summarize, we may consider any $I^{\prime}$ whose boundary is in the range
$r_{h}\leq r\leq r_{0}(v)+c_{1}\frac{G\hbar}{r}~{},~{}~{}c_{1}\gg 1~{}.$ (46)
We now minimize $S_{\rm gen}(I^{\prime}\cup R)$ over this range. By purity,
$S_{\rm gen}(I^{\prime}\cup R)=4\pi r^{2}+4G\hbar S(I^{\prime}_{c})$. Along
any ingoing light-cone, the classical area decreases rapidly and $\Delta r$
only increases as we go to smaller $r$, so we are driven to the smallest $r$
in the search space, the event horizon. Scanning in the other null direction,
along the event horizon, $S_{\rm gen}(I^{\prime}_{c})$ will decrease towards
the past, by the GSL.
Hence we obtain the tightest upper bound on $S(\mathbf{R})$ by choosing
$I^{\prime}$ to be the interior of the event horizon, as early as is possible
while maintaining Eq. (45). With the boundary of $I^{\prime}$ on the event
horizon, $r-r_{0}=\Delta r_{0}\propto\exp(v/2r_{h})$ by Eq. (43). Moreover, at
$A_{h}$, we have $r_{0}=r_{s}$ and hence $\Delta r_{0}\sim O(G\hbar/r_{h})$.
To grow this by the factor of $c_{1}$ demanded in Eq. (45), we must choose
$v=v_{s}+r_{h}\log c_{1}$, where $v_{s}$ is the $v$-coordinate of $A_{s}$ and
$A_{h}$.
To summarize, the optimal choice of $I^{\prime}$ is
$(r,v)=(r_{h},v_{s}+r_{h}\log c_{1})~{},~{}~{}c_{1}\gg 1~{}.$ (47)
The true entropy $S(\mathbf{R})$ is upper bounded by
$S_{\rm gen}(I^{\prime}\cup R)=\frac{\pi
r^{2}}{G\hbar}+S(I^{\prime}_{c})=\frac{A_{h}}{4G\hbar}+O(\log c_{1})~{}.$ (48)
Note that the $O(G\hbar)$ area difference between the event horizon and the
stretched horizon is negligible. The $O(\log c_{1})$ term captures both the
(negative) correction to the horizon area due to evaporation since $A_{h}$,
and the (positive, and larger) correction due to the von Neumann entropy of
$S(I^{\prime}_{c})$.
We stress that this upper bound is quite tight. The correct $S(\mathbf{R})$ is
given by Eq. (48) with $O(\log c_{1})$ replaced by $O(1)$. Recall that $c_{1}$
should be large enough to overcome any $O(1)$ coefficients that might enhance
the von Neumann entropy in an exact calculation. But it is itself $O(1)$ in
that sense, and $\log c_{1}$ is even smaller. In particular, we can always
choose $\log c_{1}\ll\log\log(A_{h}/G\hbar)$ in the semiclassical limit.
We also emphasize that exact spherical symmetry is not crucial; our argument
only relies on the scaling behavior of the relevant terms.
#### Recollapsing flat universe
Our next
Figure 5: Spatially flat radiation-dominated universe with negative
cosmological constant, purified by a reference universe (thermal Minkowski
space, right). If we choose a large enough reference region $R$ at $t_{\rm
Mink}=0$, then the region $I^{\prime}$ at the turnaround time $t=0$ satisfies
our sufficient conditions. Therefore an island $I$ must exist.
example was studied in detail in Ref. Hartman:2020khs .777Ref. Hartman:2020khs
considered a different but related question to ours: given a region $I$ in the
cosmology, can one find a region $R$ in the reference spacetime such that $I$
is an island with respect to $R$. By contrast, we specify a reference region
$R$ and use our sufficient conditions to establish the existence of an island
for it. Consider a radiation-dominated, spatially flat Friedmann-Robertson-
Walker (FRW) universe $M$ with cosmological constant $\Lambda$, purified by a
thermal state on a Minkowski background $M_{R}$ without gravity. The metric of
$M$ and $M_{R}$ is
$\displaystyle ds^{2}$
$\displaystyle=-dt^{2}+a(t)^{2}(dr^{2}+r^{2}\,d\Omega^{2})~{},$ (49)
$\displaystyle ds_{R}^{2}$
$\displaystyle=-dt_{R}^{2}+dr_{R}^{2}+r_{R}^{2}\,d\Omega_{R}^{2}~{}.$ (50)
Without loss of generality, one can set the scale factor $a(0)=1$ at the
turnaround time $t=0$, when $da/dt=0$ and hence $-\Lambda/8\pi G=\rho_{\rm
rad}$. A thermofield double (TFD) state is constructed at $t=0$, $t_{R}=0$.
A simple implication of the TFD state suffices for the purposes of our
analysis. Consider two spatial regions, one in $M$ at $t=0$ and the other in
$M_{R}$ at $t_{R}=0$. The von Neumann entropy of their union vanishes
approximately, if they have the same spatial coordinates. If the regions are
unequal, then the von Neumann entropy of their union will be given by the sum
of the thermal entropy of the nonoverlap portions:
$S=s_{\rm rad}(\hat{V}+\hat{V}_{R})~{},$ (51)
where $\hat{V}$ and $\hat{V}_{R}$ are the volumes of the nonoverlap portions
in $M$ and in $M_{R}$, and the entropy density is $s_{\rm rad}\sim\rho_{\rm
rad}^{3/4}$. These statements receive corrections on scales below the thermal
length scale, $\lambda\sim\rho_{\rm rad}^{-1/4}$.
Now choose $R\subset M_{R}$ to be a ball of radius $r_{R}$ at $t_{R}=0$, and
$I^{\prime}\subset M$ a ball of radius $r<r_{R}$ at $t=0$. By time symmetry
around $t=t_{R}=0$, the region $I^{\prime}\cup R$ will be quantum normal or
anti-normal. We have
$S_{\rm gen}(I^{\prime}\cup R)-S(R)=\frac{\pi r^{2}}{G\hbar}-\frac{4\pi s_{\rm
rad}}{3}r^{3}~{}.$ (52)
To satisfy our second condition, this must be negative, so we require
$r>r_{\rm crit}\equiv\frac{3}{4\pi Gs_{\rm rad}}$ (53)
This condition on $I^{\prime}$ can be satisfied, and hence an island $I\subset
M$ must exist, for a sufficiently large reference region, $r_{R}>r_{\rm
crit}$.
Going beyond spherical symmetry, we can choose $R$ to be any convex reference
region of arbitrary shape in $M_{R}$, and let $I^{\prime}$ be the identical
coordinate region in $M$. Then $I^{\prime}\cup R$ will be normal or anti-
normal by convexity and time symmetry. Moreover,
$S_{\rm gen}(I^{\prime}\cup R)-S(R)=\frac{A[\partial
I^{\prime}]}{4G\hbar}-\left(V_{R}+O(A[\partial R]\lambda)\right)s_{\rm rad}$
(54)
will be negative for any sufficiently large region of fixed shape. Any such
references region must have an island $I$.
Note that $I$ will not be the identical coordinate region to $R$, because of
the $O(A[\partial I^{\prime}]\lambda)$ corrections to the von Neumann entropy.
Moreover, in the nonspherical case, minimization of the area term will favor a
more round shape for $I$ than for $R$.
#### Bag of gold
Figure 6: A time-symmetric Cauchy slice of a bag-of-gold geometry. The bag has
large entropy (grey) compared to the area of its throat, and it purifies the
reference system $R$. Then it is easy to find a (classically and quantum)
anti-normal region $I^{\prime}$ (pink) such that $S_{\rm gen}(I^{\prime}\cup
R)<S(R)$. Hence, there must exist an island $I$. $I$ is expected to be
approximately the interior of the classically minimal surface labelled
$r_{0}$.
Next, we discuss a time-symmetric slice of a “bag-of-gold” geometry
Marolf:2008tx , shown in Fig. 6. Its defining feature is the existence of an
arbitrarily large volume of space behind a throat (a minimal area surface) of
fixed area. To construct it, we glue the interior of a sphere of radius
$r_{1}$ of a closed FRW universe, at the time of recollapse, $da/dt=0$, to the
exterior of a sphere of the same size behind the bifurcation surface in a
maximally extended Schwarzschild spacetime Horowitz:1983vn . The corresponding
spatial metrics are:
$\displaystyle ds_{\rm in}^{2}$ $\displaystyle=a(d\chi^{2}+\sin^{2}\chi
d\Omega^{2})~{},~{}~{}$ $\displaystyle 0\leq\chi\leq\chi_{1}~{};$ (55)
$\displaystyle ds_{\rm out}^{2}$
$\displaystyle=\left(1-\frac{r_{0}}{r}\right)^{-1}dr^{2}+r^{2}d\Omega^{2}~{},~{}~{}$
$\displaystyle r_{0}\leq r\leq r_{1}~{},$ (56)
where in the second line we omitted the portion of the Schwarzschild metric
outside of the bifurcation surface as it will not be needed for the analysis
below.
Let $r_{0}$ be the radius of the throat, and let $\chi$ be the angle at which
the metrics are glued. The gravitational constraints imply $\chi_{1}>\pi/2$
and
$\displaystyle a\sin\chi_{1}$ $\displaystyle=r_{1}~{},$ (57) $\displaystyle
a\sin^{3}\chi_{1}$ $\displaystyle=r_{0}~{}.$ (58)
The Friedmann equation implies that the energy density in the bag is
$\rho=\frac{1}{8\pi Ga^{2}}~{}.$ (59)
Now suppose that the bag contains thermal photon radiation purified by a
external reference system $R$. The entropy density in the bag is
$s\sim\rho^{3/4}$, and hence
$S(R)\sim(G\hbar)^{-3/4}a^{3/2}~{}.$ (60)
Let $I^{\prime}$ be the interior of some sphere $r^{\prime}$ between the edge
of the bag, $r_{1}$, and the throat, $r_{0}$; hence
$S_{\rm gen}(I^{\prime}\cup R)\sim\frac{r^{\prime 2}}{G\hbar}~{}.$ (61)
By time reversal symmetry, $I^{\prime}\cup R$ is quantum normal or anti-
normal. Moreover, we can achieve $S(R)>S_{\rm gen}(I^{\prime}\cup R)$, by an
arbitrarily large margin, by taking $a$ large while holding $r^{\prime}$ and
$r_{0}$ fixed. (This will only increase $r_{1}$.) Hence, our conditions are
satisfied, and a nontrivial island $I\subset M$ must exist.
Importantly, this construction is insensitive to the spherical symmetry that
we assumed for simplicity. It is also insensitive to the addition of
perturbative matter near the throat. Such modifications can affect the precise
position of the island, which may be very hard to determine. But so long as
they are small enough, our sufficient conditions will hold, and they guarantee
the existence of an island.
#### Collapsing star (an example without islands)
Figure 7: Collapse of a spherical star that is maximally entangled with a
distant reservoir $R$. $I^{\prime}$ (pink) is the interior of a sphere
surrounding the star at a time close to the singularity. Hence $\partial
I^{\prime}$ has a small area, and condition (i) can be satisfied by an
arbitrarily large margin. But $I^{\prime}$ is quantum trapped and so fails to
satisfy condition (ii). Indeed, $R$ does not possess any island in this
spacetime.
To illustrate the importance of condition (ii), let us discuss a case for
which condition (i), Eq. (6), is satisfied, but condition (ii), Eq. (7), is
violated. Consider an Oppenheimer–Snyder spacetime: a black hole formed by the
collapse of a “star,” modeled as a spherical, homogeneous ball of dust.
Suppose that the star is in a maximally mixed state with entropy $S_{\rm
star}$ and let $R$ be an early portion of $\mathscr{I}^{+}$ which contains
only a purification of the star (and no Hawking radiation), giving
$S(R)=S_{\rm star}$. We choose $I^{\prime}$ to be the interior of a sphere
just outside of the star and very close to the singularity (see Fig. 7). Then
$S_{\rm gen}(I^{\prime}\cup R)\approx A(\partial I^{\prime})/4G\hbar$. Picking
$S_{\rm star}$ large with $A(\partial I^{\prime})$ held fixed, we can arrange
for
$\displaystyle 1\ll S_{\rm gen}(I^{\prime}\cup R)\ll S(R)~{}.$ (62)
The first inequality ensures semiclassical control at $\partial I^{\prime}$.
The second states that condition (i) is satisfied (by an arbitrarily large
margin).
However, $\partial I^{\prime}$ is a classically trapped surface, i.e.
$\theta_{k}<0,\theta_{\ell}<0$. And since $\partial I^{\prime}$ is not close
to $\partial J^{-}(R)$, we expect quantum corrections to be small:
$\Theta_{\ell}=\theta_{\ell}+O(G\hbar)$. Condition (ii) is therefore violated.
Indeed, there are no islands associated to $R$ in this spacetime. To see this,
note that there are no classically extremal spheres. As in the previous
example, near $\partial J^{-}(R)$, quantum corrections to $\theta_{\ell}$ can
become large; but $\partial J^{-}(R)$ stays far from the horizon and so has
large classical (and quantum) expansion everywhere.
This example shows that condition (ii) is essential. So is condition (i), of
course. For example, suppose we chose $R$ to be a later portion of
$\mathscr{I}^{+}$. As before, $R$ contains only the purification of the star,
but no Hawking radiation. Since $\partial J^{-}(R)$ gets close to the horizon,
where $\Theta_{k}$ can vanish, there will be a quantum extremal region
$\tilde{I}$ with $\Theta_{\ell}$=0. However, this region fails to be an island
because $S_{\rm gen}(\tilde{I}\cup R)>S(R)$.
## 3 New Entropy Bound
In this section, we will show that in a globally pure state, the entropy of a
reference system $R$ cannot exceed the generalized entropy of suitable
asymptotic regions. We consider an external reference system in Sec. 3.1, and
we generalize to $R\subset M$ in Sec. 3.2. We discuss examples in Sec. 3.3.
### 3.1 External Reference System
Given an external reference system $R$, an island $I\subset M$ in a
semiclassical spacetime $M$ is defined as a region that is quantum extremal
and homologous to $R$ (i.e., $\partial I\subset M$), such that $S_{\rm
gen}(I\cup R)$ is minimal among all such regions. In section 2.1 we identified
sufficient conditions for $I\neq\varnothing$.
We will now employ similar techniques to derive an entropy bound on the exact
entropy of the reference system, $S(\mathbf{R})$, assuming that this is
computed by the “island formula”:
$\displaystyle S(\mathbf{R})=S_{\rm gen}(I\cup R)~{}.$ (63)
Following Ref. Almheiri:2019hni , we denote $\mathbf{R}$ in boldface when
referring to the exact (nonperturbatively computed) state of the region. We
write $R$ when referring to the semiclassically computed state. For
simplicity, we will assume that the global quantum state is pure,
$S(R\cup M)=0~{},$ (64)
though generalizations can easily be considered.
Let $I^{\prime}_{c}\subset{M}$ be any partial Cauchy slice of $M$ that is
quantum normal or anti-normal:
$\begin{cases}k^{\mu}\Theta_{\mu}[I^{\prime}_{c}]\geq
0~{},~{}~{}\ell^{\mu}\Theta_{\mu}[I^{\prime}_{c}]\leq 0~{};\\\ \text{or}\\\
k^{\mu}\Theta_{\mu}[I^{\prime}_{c}]\leq
0~{},~{}~{}\ell^{\mu}\Theta_{\mu}[I^{\prime}_{c}]\geq 0~{},\end{cases}$ (65)
where $k$ and $\ell$ are the future-directed null vector fields orthogonal to
$\partial I^{\prime}_{c}$. We also require that $I^{\prime}_{c}$ is
“asymptotic,” though only in the weak sense that in the conformally
compactified spacetime,
$\partial\Sigma\subset\partial I^{\prime}_{c}~{},$ (66)
where $\Sigma$ is a Cauchy slice of $M$. That is, $I^{\prime}_{c}$ must
contain the asymptotic region of $M$, but it may extend deep into the interior
of $M$. A simple example of a region $I^{\prime}_{c}$ that satisfies Eqs. (65)
and (66) is the exterior of a sufficiently large approximately round sphere.
Let $I^{\prime}$ be the complement of $I^{\prime}_{c}$ on some global Cauchy
slice of $M$. By Eqs. (64) and (65), $I^{\prime}\cup R$ will be anti-normal or
normal. By Eq. (66), $I^{\prime}$ is homologous to $R$. Our notation reflects
the fact that $I^{\prime}$ shares these properties with the region denoted
$I^{\prime}$ in Sec. 2.1.
However, here we do not assume the inequality $S_{\rm gen}(I^{\prime}\cup
R)<S(R)$, and hence we will not be guaranteed the existence of an island
$I\neq\varnothing$. This does not affect the maximin analysis performed in
Sec. 2.1: (anti-)normalcy of $I^{\prime}\cup R$ implies that the true island
$I$ satisfies
$S_{\rm gen}(I\cup R)\leq S_{\rm gen}(I^{\prime}\cup R)~{},$ (67)
regardless of whether $I$ is the empty set or not. Using Eqs. (63) and (64),
we thus find the entropy bound
$S(\mathbf{R})\leq S_{\rm gen}(I^{\prime}_{c})~{}.$ (68)
### 3.2 Distant Reference System
The bound (68) generalizes to the case where $R\subset M$, subject to
appropriate modifications. (It is easy to generalize further to the case where
$R$ is partly internal to $M$ and partly an external system.) We shall assume
that gravity is negligible in $R$, so that the notion of an exact state of
$\mathbf{R}$ can be made precise. The island rule can then be adapted to
compute the generalized entropy of $R$:
$\displaystyle S_{\rm gen}(\mathbf{R})=S_{\rm gen}(I\cup R)~{},$ (69)
where $I$ is an island (possibly the empty set), as described above. The
relevant homology rule is $I\subset\mbox{int}[M-\overline{J(R)}]$, where $J$
denotes the union of the causal past and future.
We again assume global purity, $S(M)=0$. To obtain a bound on $S_{\rm
gen}(\mathbf{R})$, we consider a spatial region $I^{\prime}_{c}$ that
satisfies the following conditions (see Fig. 8):
* •
For $I^{\prime}$ to be of the correct homology type, without directly
referring to $I^{\prime}$,888Our goal is to formulate a bound in terms of
quantities that are accessible to an asymptotic observer. we require that
$I^{\prime}_{c}$ is adjacent to $R$ in $M$; and in the conformally
compactified spacetime $\tilde{M}$, $I^{\prime}_{c}$ contains any conformal
boundary portions not covered by $R$:
$\partial I^{\prime}_{c}\supset\partial(\tilde{\Sigma}-\bar{R})~{},$ (70)
where $\tilde{\Sigma}\supset R$ is a Cauchy slice of $\tilde{M}$, and
$\bar{R}$ denotes the closure of $R$ in $\tilde{M}$.
* •
$I^{\prime}_{c}$ is quantum normal or anti-normal under shape deformations of
its inner boundary in $M$, i.e., at $(\partial I^{\prime}_{c}-\partial R)\cap
M$ .
* •
$I^{\prime}_{c}$ contains a region $I_{0,c}$ that is quantum anti-normal at
$\partial I_{0,c}-\partial R$. (Normal is not allowed in this criterion.)
Global purity implies that $I^{\prime}\cup R$ will be quantum normal or anti-
normal at $\partial I^{\prime}$. It also guarantees quantum normalcy of
$I_{0}\cup R$, where $I_{0}\equiv\Sigma-I_{0,c}-R$ and $\Sigma$ is a Cauchy
surface that contains $I_{0,c}$ and $R$. By Sec. 2.2, the maximin procedure
restricted to the wedge $D(I_{0})$ will return a region
$\hat{I}\subset\mbox{int}[D(I_{0})]$ that satisfies the homology rule and has
stationary $S_{\rm gen}(\hat{I}\cup R)$ under shape deformations at
$\partial\hat{I}$.
Figure 8: The entropy $S(\mathbf{R})$ of an external or distant reference
system $R$ must be less than the generalized entropy of any region
$I^{\prime}_{c}$ that is normal or anti-normal (blue).
Note that $\hat{I}$ may be empty; and $S_{\rm gen}(\hat{I}\cup R)$ need not be
globally minimal, since the true island $I$ may not be contained in
$D(I_{0})$. However, we have
$S_{\rm gen}(I\cup R)\leq S_{\rm gen}(\hat{I}\cup R)~{},$ (71)
The quantum (anti-)normalcy of $I^{\prime}$ implies
$S_{\rm gen}(\hat{I}\cup R)\leq S_{\rm gen}(I^{\prime}\cup R)~{}.$ (72)
Using Eqs. (69) and global purity, we thus find the entropy bound
$S_{\rm gen}(\mathbf{R})\leq S_{\rm gen}(I^{\prime}_{c})~{}.$ (73)
Recall that $\partial I^{\prime}_{c}\supset\partial R$, so in any situation
where the generalized entropy can be separated into a regularized entropy and
Bekenstein-Hawking term, the area terms associated with $\partial R$ will
cancel, and Eq. (73) reduces to Eq. (68).
### 3.3 Examples and Discussion
The bound (73), and its external $R$ version (68), are powerful and versatile.
They require knowledge only of $R$ and $I^{\prime}_{c}$, but not of the rest
of the spacetime $M$. The only nontrivial condition, Eq. (65), can be easily
verified. Often the quantum expansion is dominated by the classical expansion,
so that it is easy to check whether $I^{\prime}_{c}$ is quantum (anti-)normal;
yet the bound remains nontrivial.
Figure 9: Not a Schwarzschild black hole. This could be a highly dynamical
spacetime far from a stationary black hole solution. $I^{\prime}_{c}$ is is
quantum anti-normal, so $S(\mathbf{R})\leq S_{\rm gen}(I^{\prime}_{c})$. In
the left example, $R$ is on the conformal boundary. Quantum anti-normalcy
follows from the GSL (future null congruence), and from the classical area
theorem (past null congruence) if quantum corrections are small. In the right
example, $R$ is inside the spacetime. Quantum anti-normalcy follows by the
area theorem and smallness of corrections if $\partial I^{\prime}_{c}$ stays
far enough from the horizon.
For example, by the generalized second law, causal wedges of a boundary region
must be quantum anti-normal.999Note that the causal wedge is not in general a
domain of dependence. The region that is always quantum normal is the maximal
Cauchy evolution of the causal wedge. This follows both for asymptotically
anti-de Sitter and flat spacetimes. When $R$ is disjoint from the conformal
completion of $M$, this ensures the quantum anti-normalcy of $I^{\prime}_{c}$.
The requirement that $R$ and $I^{\prime}_{c}$ be spacelike separated prevents
the application of the GSL when $R$ is part of the conformal completion of
$M$. For instance, suppose that $R$ is a subset of $\mathscr{I}^{+}$. The GSL
can still be applied to any future causal horizon associated to
$\mathscr{I}^{+}$ regions including $R$, guaranteeing $k^{\mu}\Theta_{\mu}\geq
0$. However, all past horizons will intersect the past of $R$, blocking the
application of the GSL to $I^{\prime}_{c}$ along them. To establish that
$\ell^{\mu}\Theta_{\mu}\leq 0$, we can use the classical area law on past
horizons, so long as quantum corrections to $\ell^{\mu}\theta_{\mu}$ is
negligible (see Fig. 9, left).
More generally, when $R$ is a subset of $M$, the classical area law ensures
condition (65) if the quantum corrections to both expansions are suppressed
(see right Fig. 9). This causal wedge method for finding $I^{\prime}_{c}$
suggests a nice physical interpretation of $I^{\prime}_{c}$ as a region that
can be explored geometrically by asymptotic observers.
The bound thus tells us that $S(\mathbf{R})$ cannot be greater than the
generalized entropy of any causal wedge region (subject to quantum effects on
the expansion remaining negligible). If $I^{\prime}_{c}$ is a whole Cauchy
surface $\Sigma$ of $M$, this reduces to the trivial statement that
$S(\mathbf{R})\leq S(\Sigma)$. (In this case, by purity, equality must hold.)
But if $I^{\prime}_{c}$ has a boundary in $M$, the bound is nontrivial.
Indeed, a quantum anti-normal causal wedge can reach very close
($O[(G\hbar)^{1/2}]$ distance) to a black hole horizon. For a black hole after
the Page time, this means that bound becomes nearly saturated. The bound then
implies the nontrivial statement of unitarity.
The bound simplifies if $I^{\prime}_{c}$ contains little matter entropy, so
that
$S_{\rm gen}(I^{\prime}_{c})\approx\frac{A}{4G\hbar}~{},$ (74)
In this case,
$S(\mathbf{R})\lesssim\frac{A}{4G\hbar}~{}.$ (75)
The entropy deliverable to an asymptotic observer by a spacetime causally
explored to an inner boundary of area $A$ cannot be greater than $A/4G\hbar$.
## Acknowledgements
ASM would like to thank Raghu Mahajan for discussions. This work was supported
in part by the Berkeley Center for Theoretical Physics; by the Department of
Energy, Office of Science, Office of High Energy Physics under QuantISED Award
DE-SC0019380 and under contract DE-AC02-05CH11231; and by the National Science
Foundation under Award Numbers 1820912 (RB) and 2014215 (ASM).
## References
* (1) N. Engelhardt and A. C. Wall, _Quantum Extremal Surfaces: Holographic Entanglement Entropy beyond the Classical Regime_ , _JHEP_ 01 (2015) 073 [1408.3203].
* (2) V. E. Hubeny, M. Rangamani and T. Takayanagi, _A Covariant holographic entanglement entropy proposal_ , _JHEP_ 07 (2007) 062 [0705.0016].
* (3) S. Ryu and T. Takayanagi, _Holographic derivation of entanglement entropy from AdS/CFT_ , _Phys. Rev. Lett._ 96 (2006) 181602 [hep-th/0603001].
* (4) A. Lewkowycz and J. Maldacena, _Generalized gravitational entropy_ , _JHEP_ 08 (2013) 090 [1304.4926].
* (5) T. Faulkner, A. Lewkowycz and J. Maldacena, _Quantum corrections to holographic entanglement entropy_ , _JHEP_ 11 (2013) 074 [1307.2892].
* (6) X. Dong and A. Lewkowycz, _Entropy, Extremality, Euclidean Variations, and the Equations of Motion_ , _JHEP_ 01 (2018) 081 [1705.08453].
* (7) G. Penington, S. H. Shenker, D. Stanford and Z. Yang, _Replica wormholes and the black hole interior_ , 1911.11977.
* (8) A. Almheiri, T. Hartman, J. Maldacena, E. Shaghoulian and A. Tajdini, _Replica Wormholes and the Entropy of Hawking Radiation_ , _JHEP_ 05 (2020) 013 [1911.12333].
* (9) D. N. Page, _Information in black hole radiation_ , _Phys. Rev. Lett._ 71 (1993) 3743 [hep-th/9306083].
* (10) G. Penington, _Entanglement Wedge Reconstruction and the Information Paradox_ , _JHEP_ 09 (2020) 002 [1905.08255].
* (11) A. Almheiri, N. Engelhardt, D. Marolf and H. Maxfield, _The entropy of bulk quantum fields and the entanglement wedge of an evaporating black hole_ , _JHEP_ 12 (2019) 063 [1905.08762].
* (12) S. W. Hawking, _Particle Creation by Black Holes_ , _Commun. Math. Phys._ 43 (1975) 199.
* (13) S. W. Hawking, _Breakdown of Predictability in Gravitational Collapse_ , _Phys. Rev. D_ 14 (1976) 2460.
* (14) R. Bousso, Z. Fisher, S. Leichenauer and A. C. Wall, _Quantum focusing conjecture_ , _Phys. Rev._ D93 (2016) 064044 [1506.02669].
* (15) A. Almheiri, R. Mahajan and J. Maldacena, _Islands outside the horizon_ , 1910.11077.
* (16) R. Bousso and M. Tomašević, _Unitarity From a Smooth Horizon?_ , _Phys. Rev. D_ 102 (2020) 106019 [1911.06305].
* (17) R. Bousso and E. Wildenhain, _Gravity/ensemble duality_ , _Phys. Rev. D_ 102 (2020) 066005 [2006.16289].
* (18) X. Dong, X.-L. Qi, Z. Shangnan and Z. Yang, _Effective entropy of quantum fields coupled with gravity_ , _JHEP_ 10 (2020) 052 [2007.02987].
* (19) F. F. Gautason, L. Schneiderbauer, W. Sybesma and L. Thorlacius, _Page Curve for an Evaporating Black Hole_ , _JHEP_ 05 (2020) 091 [2004.00598].
* (20) T. Hartman, E. Shaghoulian and A. Strominger, _Islands in Asymptotically Flat 2D Gravity_ , _JHEP_ 07 (2020) 022 [2004.13857].
* (21) A. Almheiri, R. Mahajan, J. Maldacena and Y. Zhao, _The Page curve of Hawking radiation from semiclassical geometry_ , _JHEP_ 03 (2020) 149 [1908.10996].
* (22) A. Almheiri, R. Mahajan and J. E. Santos, _Entanglement islands in higher dimensions_ , _SciPost Phys._ 9 (2020) 001 [1911.09666].
* (23) Y. Chen, V. Gorbenko and J. Maldacena, _Bra-ket wormholes in gravitationally prepared states_ , 2007.16091.
* (24) T. Hartman, Y. Jiang and E. Shaghoulian, _Islands in cosmology_ , _JHEP_ 11 (2020) 111 [2008.01022].
* (25) A. C. Wall, _Maximin Surfaces, and the Strong Subadditivity of the Covariant Holographic Entanglement Entropy_ , _Class. Quant. Grav._ 31 (2014) 225007 [1211.3494].
* (26) C. Akers, N. Engelhardt, G. Penington and M. Usatyuk, _Quantum Maximin Surfaces_ , _JHEP_ 08 (2020) 140 [1912.02799].
* (27) D. Marolf, A. C. Wall and Z. Wang, _Restricted Maximin surfaces and HRT in generic black hole spacetimes_ , _JHEP_ 05 (2019) 127 [1901.03879].
* (28) A. R. Brown, H. Gharibyan, G. Penington and L. Susskind, _The Python’s Lunch: geometric obstructions to decoding Hawking radiation_ , _JHEP_ 08 (2020) 121 [1912.00228].
* (29) C. Akers, R. Bousso, I. F. Halpern and G. N. Remmen, _Boundary of the future of a surface_ , _Phys. Rev. D_ 97 (2018) 024018 [1711.06689].
* (30) A. C. Wall, _The Generalized Second Law implies a Quantum Singularity Theorem_ , _Class. Quant. Grav._ 30 (2013) 165003 [1010.5513].
* (31) L. Andersson and J. Metzger, _The Area of horizons and the trapped region_ , _Commun. Math. Phys._ 290 (2009) 941 [0708.4252].
* (32) S. Abdolrahimi, D. N. Page and C. Tzounis, _Ingoing Eddington-Finkelstein Metric of an Evaporating Black Hole_ , _Phys. Rev. D_ 100 (2019) 124038 [1607.05280].
* (33) J. D. Bekenstein, _Black holes and the second law_ , _Lett. Nuovo Cim._ 4 (1972) 737.
* (34) A. C. Wall, _A proof of the generalized second law for rapidly changing fields and arbitrary horizon slices_ , _Phys. Rev. D_ 85 (2012) 104049 [1105.3445].
* (35) D. Marolf, _Black Holes, AdS, and CFTs_ , _Gen. Rel. Grav._ 41 (2009) 903 [0810.4886].
* (36) G. T. Horowitz, _The positive energy theorem and its extensions_ , in _Conference on Asymptotic Behavior of Mass and Spacetime Geometry_ , p. 0001, 10, 1983.
|
# Spatio-temporal characterization of ultrashort vector pulses
Apostolos Zdagkas Optoelectronics Research Centre and Centre for Photonic
Metamaterials, University of Southampton, Southampton SO17 1BJ, United Kingdom
Venkatram Nalla Optoelectronics Research Centre and Centre for Photonic
Metamaterials, University of Southampton, Southampton SO17 1BJ, United Kingdom
Nikitas Papasimakis Optoelectronics Research Centre and Centre for Photonic
Metamaterials, University of Southampton, Southampton SO17 1BJ, United Kingdom
Nikolay I. Zheludev Optoelectronics Research Centre and Centre for Photonic
Metamaterials, University of Southampton, Southampton SO17 1BJ, United Kingdom
Centre for Disruptive Photonic Technologies, School of Physical and
Mathematical Sciences and The Photonics Institute, Nanyang Technological
University, Singapore 637378, Singapore
###### Abstract
Ultrafast vectorially polarized pulses have found many applications in
information and energy transfer owing mainly to the presence of strong
longitudinal components and their space-polarization non-separability. Due to
their broad spectrum, such pulses often exhibit space-time couplings, which
significantly affect the pulse propagation dynamics leading to reduced energy
density or utilized to create new effects like a rotating or sliding wavefront
at focus. Here, we present a new method for the spatio-temporal
characterization of ultrashort cylindrical vector pulses based on a
combination of spatially resolved Fourier transform spectroscopy and Mach-
Zehnder interferometry. The method provides access to spatially resolved
spectral amplitudes and phases of all polarization components of the pulse. We
demonstrate the capabilities of the method by completely characterizing a $10$
fs radially polarized pulse from a Ti:sapphire laser at $800$ nm.
## I Introduction
Space-time couplings (STCs) in propagating waves are defined as the dependence
of the temporal properties of the electric field on the transverse spatial
coordinates [1]. Mathematically they are revealed as the non-separability of
the spatial and temporal terms of the electric field of a pulse into a
product, $E(\mathbf{r},t)\neq f(\mathbf{r})g(t)$. They can significantly
affect the energy density of ultrashort pulses at focus [2] since in most
cases pulses possessing STCs are not transform limited. STCs also alter the
propagation dynamics of ultrashort pulses. The latter has been utilized to
create new effects like the lighthouse effect [3] where a tilted wavefront is
transformed to a rotating wavefront at focus and the “sliding” or “flying”
focus [4, 5] effect where STCs create a focal spot that locally travels with
speed greater or lower than the speed of light in free space.
A range of techniques have been demonstrated for the complete spatio-temporal
characterization (retrieval of electric field amplitude and phase) of linearly
polarized pulses [6]. Such approaches are typically based on scanning the
transverse profile of the unknown pulse with a known reference pulse [7, 8, 9]
or utilize concepts from wavefront characterization techniques [10, 11].
Moreover, self-referenced methods have been proposed, where a small part of
the pulse under characterization is used as reference. In the latter, the
reference is interfered with the unknown pulse and the unknown spectral phase
is retrieved through spatially resolved Fourier transform spectroscopy [12]. A
variation of the latter, named TERMITES [13], uses a slightly expanded replica
of the unknown along with an iterative algorithm as a way to reduce the
requirements of creating a very homogeneous reference pulse.
However, none of the aforementioned methods can simultaneously characterize
the spatially dependent polarization, intensity and phase that vector
polarized pulses exhibit. Indeed, methods such as TERMITES are unsuitable for
the characterization of cylindrical vector pulses (CVPs), for example radially
polarized pulses, that exhibit polarization singularities at their center.
Ultrafast radially polarized pulses have been generated in the femtosecond
regime [14] and have been compressed down to few femtoseconds [15, 16]. Due to
their broad spectrum, such pulses often exhibit space-time couplings (STC).
Moreover, space-time couplings have been prescribed to them through a
metamaterial converter [17]. Therefore, their complete characterization
becomes of paramount importance for most applications of CVPs.
The characterization of CVPs has been limited mainly to their spatially
varying polarization profile. The temporal profile of such pulses has been
characterized independently by standard approaches, such as FROG [18] and
SPIDER [19]. However, such measurements are performed at a single spatial
position of the cross section of the pulse, which is far from a complete
characterization of the pulses that may exhibit STCs [1]. For example,
theoretical calculations have shown that the Flying Donut pulse is
isodiffracting [20], a property that leads to the spatial profiles of
intensity for every frequency component of the pulse to scale along the
trajectory of the pulse in the same way. Furthermore, since most ultrafast
systems use pulse compressors based on prisms or gratings, in order to
implement chirped pulse amplification (CPA) [21], the appearance of pulse
front tilt (PFT) is very likely to occur from small misalignments [1, 22]
leading to detrimental effects for the pulse duration and its energy density.
Such space time couplings occur also in the case of ultrafast CVPs. For
example, Fig. 1 illustrates the effect of pulse front tilt and pulse front
curvature on the shape of a radially polarized pulse. PFT appears as a
linearly varying delay of the pulse across a transverse direction while pulse
front curvature is a quadratically varying delay from the center of the pulse
to its edges. Therefore, the complete characterization of CVPs becomes of
paramount importance for most applications.
Only recently a new method has been utilized for the complete spatio-temporal
characterization of a 100 fs vector pulse through a two-fold interferometer
[23]. Although it was shown to successfully characterize spatially varying
polarization gates, it is based on a point scan scheme with an optical fiber,
rendering it impractical for the complete transverse spatial characterization
of the pulse which would require millions of points to be scanned.
Figure 1: Characteristic examples of space-time couplings in vector polarized
pulses. Cross sections of the x component of the instantaneous electric field
of a radially polarized pulse (a) in the absence of space-time couplings, (b)
with pulse front tilt ($3\;\mathrm{fs/mm}$), and (c) with pulse front
curvature ($3\;\mathrm{fs/mm^{2}}$). The insets to (a-c) present corresponding
3D illustrations in the form of isosurfaces. The pulse propagates along the
positive z axis.
In this work, we present a method for the complete spatio-temporal
characterization of CVPs. The method is an extension of the TERMITES technique
by means of a Mach-Zehnder interferometer and is termed TERMITES-MAZE (MAch-
Zehnder Extended). It allows to fully characterize the spatial and temporal
profile of all polarization components of ultrafast CVPs at the spatial
resolution of a camera sensor. We illustrate the capabilities of the TERMITES-
MAZE technique by applying it to the case of a $10$ fs radially polarized
pulse centered at $800$ nm. We show that our approach can reveal the space-
time couplings (e.g. pulse front tilt) across different planes for different
polarization components of the pulse. We present a detailed description of the
experimental implementation of TERMITES-MAZE, while the corresponding
algorithm for the analysis of the experimental data is freely available under
the BSD 3-clause licence as a python module [24].
As few cycle laser pulses with vectorial fields, orbital angular momentum
(OAM) [25] and other forms of structured light [26] become common, the need
for their characterization increases. The presence of STCs become even more
important in such broadband pulses that can eventually alter their propagation
properties [27] and their interaction with matter [28, 29] leading to the
appearance of new effects. The tool provided here has the capability to
characterize all these types of space-time-polarization “entanglement” that
are increasingly studied [30] and thus has the potential to accelerate the
research and emerging applications based on few cycle vector structured light.
## II Cylindrical vector pulses
Cylindrical vector beams are solutions to Maxwell’s equation whose amplitude,
phase and polarization are axially symmetric. They have been extensively
studied and they are routinely generated and characterized [31]. Their unique
properties have found numerous applications such as tight focusing [32],
efficient particle trapping [33, 34], super-resolution microscopy [35], dense
optical 3d data storage [36] and data encoding for optical communications [37]
to name a few. Furthermore, their space polarization non-separability has been
applied to extend the concepts and tools of quantum physics to “classically
entangled” states [38, 39].
Cylindrical vector pulses are the pulsed version of the cylindrical vector
beams. Although they offer more opportunities for applications, they are less
studied due to the difficulty of their generation and characterization. They
share all of the characteristics that make these light structures interesting
with additional important features arising due to their pulsed nature. Their
short duration renders them ideal for microprocessing of materials since they
combine the reduced thermal damage with the increased efficiency and
homogeneity of the radial polarization when used in micro-drilling [40]. In
fact, single-cycle CVPs, termed ”Flying Donuts”, exist [41], for which
applications in particle acceleration have been suggested. Moreover, such
pulses exhibit a finite topological structure with a finite number of
localized singularities [42], while their toroidal topology is ideal for
engaging toroidal and anapole modes in matter [43, 44].
## III Method
A method for the complete characterization of a CVP should provide all the
characteristics of the pulse like the polarization, spectral amplitude and
spectral phase at each spatial position. The TERMITES-MAZE approach is capable
of retrieving all these characteristics through an interference of a linearly
polarized reference pulse with an unknown CVP. Firstly, the TERMITES method is
applied for the characterization of the reference pulse and then a Mach-
Zehnder interferometer is implemented for the characterization of the unknown.
In this way only one polarization of the CVP is characterized at a time.
Finally, since all the spatial information is captured, the complete vectorial
shape of the pulse can be reconstructed.
### III.1 Experimental implementation
The TERMITES-MAZE method is an extension of the TERMITES method and as such it
involves two experiments. The first experiment is the characterization of a
reference linearly polarized pulse using the TERMITES method and the second is
a Mach-Zehnder interferometer in which an unknown pulse, in our case a CVP,
and the reference pulse are interfered. In both cases, Fig. 2a and b, a $10$
fs Ti:Sa laser with peak frequency at $800$ nm (Spectra-Physics, Element PRO)
is used to generate a linearly polarized pulse. The pulse then travels through
a pulse shaper that uses a spatial light modulator (SLM) (Biophotonic
solutions inc, MIIPS Box640). A polarizer after the pulse shaper is used to
select the polarization and to control the intensity of certain frequencies
that are set by the SLM. The pulse shaper is used to compress the ultrafast
pulse to its Fourier limited duration after travelling though the many optical
components that are required for the generation and characterization of the
radially polarized pulse. Then a beam expander, doubles the beam width in
order to fit the radius of a segmented waveplate that transforms a linear
polarized beam to a radial polarized one. After the beam expander, a 50:50
beam splitter (BS1 in Fig.2) is used to generate two replicas of the pulse. Up
to this point the setup is common to both experiments. In the following we
describe the two experiments separately.
Figure 2: Experimental implementation of TERMITES-MAZE. (a) Following the
TERMITES technique, a Michelson interferometer setup with a convex mirror in
one arm and a delay line in the other is used for the spatio-temporal
characterization of a linearly polarized pulse (this pulse will be used as
reference). The pulse compression part is shown in the lower part of the
setup. (b) A Mach-Zehnder interferometer setup with the one arm carrying the
reference pulse (as characterized in (a)) and the other hosting the setup for
the generation of the CVP. A delay line is used to scan the unknown pulse with
the reference. The generated images are then used to extract the spectral
phase and frequency distribution of the unknown pulse. The interference
pattern has the expected form of the interference of a linear and a radially
polarized pulse. The intensity pattern is antisymmetric with respect to the
center of the radially polarized pulse which is a manifestation of the radial
polarization shown with yellow arrows. PS: pulse shaper, P: polarizer, BE:
beam expander, BS: beam splitter, WP: $\lambda/2$ waveplate, DG: dispersive
glass, PM: parabolic mirror, SHG: second harmonic generation crystal, L: lens,
F: filter, FS: fiber spectrometer, CM: convex mirror and SWP: segmented
waveplate.
The TERMITES experimental setup consists of a simple Michelson interferometer
and a digital camera (Thorlabs DCC1545M), see upper part of Fig. 2a. One arm
of the interferometer includes a nanometre precision delay line with a
scanning range of $20\mu m$. The other arm includes a convex mirror, whose
purpose is to expand the unknown pulse. Hence the image on the camera sensor
will be the result of a pulse interfering with an expanding replica of its
central part. This expanding central part is considered smoother, in terms of
spatio-temporal couplings, than the pulse itself since it is just a small part
of the pulse stretched to fit a wider area. A polarizer is finally placed
before the camera to increase the signal to noise ratio.
The TERMITES algorithm reveals the spectral phase differences between every
point of the pulse and its center. A simple temporal characterization of its
central part is the only requirement for a complete spatio-temporal
characterization. In our case the “Multiphoton Intrapulse Interference Phase
Scan” (MIIPS) [45] technique is used for this purpose since it is already a
part of the setup. More specifically, instead of simply characterizing the
temporal shape of the pulse, we compress it. The compression will produce a
Fourier limited pulse that is also required for the generation of a Fourier
limited CVP.
The compression part of the setup can be seen in the lower part of Fig. 2a.
The pulse is focused at a second harmonic generation crystal. The generated
spectral intensities are then recorded by a spectrometer (Ocean Optics
USB2000+XR1-ES) and fed to the pulse shaper’s software where the MIIPS method
is used for the compression. In the schematic of Fig. 2a, the dispersion
compensation glass and the waveplate of the lower arm are used to create the
same amount of dispersion with the waveplate of the upper arm, the two passes
through the beam splitter and the polarizer. The mirrors have low group delay
dispersion and more or less the same number of mirrors is used in the two
arms.
A good spatial resolution to resolve all the fringes and avoid $2\pi$ phase
jumps across neighbouring pixels after the Fourier transform is required.
Hence a careful choice of the convex mirror induced curvature and the camera
sensor is necessary. The same is true for the step of the delay line. Since
both spatial and temporal patterns are periodic effects, the experiments are
designed to capture more than 10 samples per period, much more than the
theoretical minimum according to the sampling theorem [46].
The above produces a reference (known) linearly polarized pulse. A simple
interference of this reference pulse with an unknown more complex structured
pulse, like a CVP, is sufficient to reveal the full spatial and temporal
structure of the latter. Such an experimental setup is given in Fig. 2b. After
the first beam splitter (BS1 in Fig. 2b) two replicas of the pulse are created
that travel in two separate arms of a Mach-Zehnder interferometer. The right
part of the interferometer carries the reference pulse, while the left arm is
used to generate a CVP, which then propagates to the final beam splitter (BS2
in Fig. 2b) where the two pulses recombine and the interference takes place.
In this case, the beam splitter has to be polarization independent, as the
polarization of the unknown pulse is spatially dependent. As an example, here
we characterize a radially polarized pulse that is generated from a segmented
waveplate [32, 47, 16]. Our polarization transformer consist of 8 achromatic
half-wave plates with the relative orientation of the fast axis at $\pm 11,\pm
34,\pm 56$ and $\pm 79$ degrees. The joints between the waveplates scatter
part of the incident light and create diffraction patterns. For that, a $75\mu
m$ pinhole is placed at the Fourier plane to spatially filter the pulse. An
iris placed before the focusing mirror is used to select the size of the
radial pulse that has to be smaller or equal to the reference. Any diffraction
introduced by the iris will be also filtered by the pinhole. A half-waveplate
is placed before the segmented waveplate to rotate the input polarization and
hence select the output polarization that can be either radial or azimuthal.
Finally the filtered pulse travels to a delay line that creates a variable
delay between the CVP and the reference pulse. The Thorlabs TSGNF5 single axis
piezoelectric stage along with a hollow retroreflector were used for the delay
line. The interferogram is then captured by the camera. A polarizer before the
camera is used not only for the increase of the signal to noise ratio but to
characterize the two orthogonal polarizations of the CVP. The latter is
achieved by rotating the polarizer at $\pm 45$ degrees with respect to the
polarization of the reference pulse. Hence two measurements are required for
the complete spatio-temporal characterization of the CVP. The delay between
the polarizations is assumed to be zero because of the working principle of
the achromatic segmented waveplate. Finally, the dispersion of the pulse must
be considered carefully. Both paths need to have the same dispersion as the
TERMITES part of the experiment, where the reference pulse was characterized.
### III.2 Analysis of the recorded data
Figure 3: A schematic description of the required processing of the images
used in the TERMITES algorithm. The data are acquired from the setup described
in Fig. 2a. (a) A sequence of interference images captured by a digital camera
for successive delays $\tau$ of the reference pulse. (b) The cross-correlation
signal, $S(x,y,\tau)$, at a single pixel of the camera sensor. That is the
measured intensity as a function of the delay. (c) The spectrum derived by a
fast Fourier transform ($F_{T}$) performed on the cross-correlation signal of
the pixel. (d) Unwrapped phase data (red dots) and the quadratic fitted
surface (blue) close to the central wavelength ($\simeq 800$ nm). The fitted
phase is removed from the total spatially depended phase to reveal the phase
due to the STC. (e) The retrieved phase at $794$ nm after the application of
the iterative algorithm. (f) The 3d reconstruction of the pulse in time and
space through an inverse Fourier transform ($\tilde{F}_{T}$) of the retrieved
amplitude and phase data.
In this section we briefly discuss the TERMITES algorithm and we present the
extra analysis steps needed for the characterization of CVPs. The TERMITES
method is used for the spatiotemporal characterization of linearly polarized
pulses up to a constant spectral phase and an ambiguity in the pulse front
curvature. Our approach inherits these limitations.
Figure 4: The final step of TERMITES-MAZE method for the characterization of a
radially polarized pulse. (a) A simplified schematic of the experiment where
the reference pulse from the previous part of the method is interfered with
the unknown CVP. (b) A sequence of interference images captured by a digital
camera for successive delays of the reference pulse.(c) The Fourier
transformed cross-correlation signal close to the central wavelength of the
pulse. The retrieved amplitude and phase of the x polarization of the radially
polarized pulse at $794$ nm after the division and subtraction of the
reference pulse’s amplitude and phase from the cross-correlation spectra. (d)
The 3d reconstruction of the x polarization of the radially polarized pulse in
time and space.
In TERMITES, a pulse is interfered with a magnified replica of itself. A delay
line is used to scan the unknown pulse with the reference. The interference
pattern is then recorded by a digital camera and an image for each relative
delay is stored in a computer, resulting in a 3D dataset (see Fig. 3a). For
each pixel across all images an interferogram is captured which is the
collected energy during the exposure period transferred to the sensor by the
two pulses, as a function of the delay. This signal is proportional to the
time integral of the intensity of the total electric field and can be written
as
$\displaystyle S(\mathbf{r},\tau)$
$\displaystyle=\int|E_{\mathrm{R}}(\mathbf{r},t)+E(\mathbf{r},t-\tau)|^{2}dt$
$\displaystyle=I(\mathbf{r})+I_{\mathrm{R}}(\mathbf{r})+\int[E_{\mathrm{R}}(\mathbf{r},t)E^{*}(\mathbf{r},t-\tau)+E^{*}_{\mathrm{R}}(\mathbf{r},t)E(\mathbf{r},t-\tau)]dt,$
(1)
with the subscript “$\mathrm{R}$” denoting the reference (diverging) pulse and
$\tau$ the delay. The signal on a central pixel is shown in Fig. 3b. A
constant signal is observed when there is no overlap of the pulses,
mathematically described by the first two terms of Eq. 1, and an interference
signal when they overlap, mathematically described by the integral. The term
$s(\mathbf{r},\tau)=\int
E_{\mathrm{R}}(\mathbf{r},t)E^{*}(\mathbf{r},t-\tau)dt$ is the cross-
correlation of the two fields. Hence Eq. 1 takes the form
$\displaystyle
S(\mathbf{r},\tau)=I(\mathbf{r})+I_{\mathrm{R}}(\mathbf{r})+s(\mathbf{r},\tau)+s^{*}(\mathbf{r},\tau).$
(2)
Figure 5: (a) and (c) Retrieved amplitude of the horizontally and vertically
polarized areas of a radially polarized pulse at $794$ nm respectively. (b)
and (d) Retrieved phase at the same wavelength for the horizontal (vertical)
polarization according to the TERMITES-MAZE algorithm. The spectral phase has
a difference of $\pi$ between the left and right (top and bottom) areas as it
is expected for a radially polarized pulse.
A Fourier transform is then performed for every pixel and the frequency domain
signal is now described by the equation
$\displaystyle
S(\mathbf{r},\omega)=F_{\mathrm{T}}[S(\mathbf{r},\tau)]=F_{\mathrm{T}}[I(\mathbf{r})+I_{\mathrm{R}}(\mathbf{r})]+s(\mathbf{r},\omega)+s^{*}(\mathbf{r},-\omega),$
(3)
with $F_{\mathrm{T}}$ denoting the time to frequency Fourier transform
operator. The equation describes the presence of a zero frequency term,
derived from the Fourier transform of the time stationary term, and two
frequency signals with equal amplitude and at symmetric frequencies. Fig. 3c
shows the results derived from the above analysis regarding a central pixel of
the camera which corresponds to the central part of the pulse. A zero
frequency term and two symmetric peaks with central wavelength at around $800$
nm, in accordance with the central wavelength of our laser, are formed by the
Fourier transform.
The result of the above procedure leads to two sets of images. One carries the
intensity distribution at each frequency and the other the spectral phase. The
negative frequencies can be derived by the complex conjugate of the positive
and hence are redundant for the analysis. Additionally, the noise free
spectrum is from about 700 to 900 nm. Only a few images are thus needed for
the analysis and hence the processing time is reduced significantly without
any loss of information.
The Fourier transform of the cross-correlation of two signals,
$E(\mathbf{r},t)$ and $E_{\mathrm{R}}(\mathbf{r},t)$, is equal to the product
of their spectra with one of them being complex conjugate, as in the cross-
correlation integrand. In our case we have
$s(\mathbf{r},\omega)=E_{\mathrm{R}}(\mathbf{r},\omega)E^{*}(\mathbf{r},\omega)$.
The amplitude and phase of the signal that we are interested are thus given by
$\displaystyle|s(\mathbf{r},\omega)|$
$\displaystyle=A_{\mathrm{R}}(\mathbf{r},\omega)A(\mathbf{r},\omega)$ (4)
$\displaystyle\mathrm{arg}(s(\mathbf{r},\omega))$
$\displaystyle=\varphi_{\mathrm{R}}(\mathbf{r},\omega)-\varphi(\mathbf{r},\omega)$
(5)
with $A$ being the amplitude of the spectrum and $\varphi$ the spectral phase.
In the TERMITES algorithm, the spectral phase that is created from the convex
mirror is removed by fitting a quadratic surface to the phase data, Fig. 3d
and then an iterative algorithm is applied Fig. 3e until the retrieved
spectral phase can correctly reproduce the known relationship between the
pulse and its expanding replica [13]. An inverse Fourier transform then
reconstructs the linearly polarized pulse in time and space, Fig. 3f.
Following the characterization of the reference pulse the Mach-Zehnder
experiment is performed. The analysis is the same up to Eq. 4 and 5. However,
in this case the reference spectral amplitude and spectral phase are known,
hence a simple division of the spectral amplitudes and a subtraction of the
spectral phases gives the complete frequency domain representation of the
unknown CVP. The procedure is then repeated for the orthogonal polarization
and the results are combined.
Figure 6: (a)-(b) $x-t$ cross-section of the x polarized electric field of a
10 fs radially polarized pulse and its isosurface plot at half of the maximum
electric field value as it is reconstructed with the TERMITES-MAZE method.
(c)-(d) $y-t$ cross-section of the y polarized electric field of the same
pulse and its isosurface plot. Figure 7: (a-b) $x-t$ cross-section of the x
polarized electric field of a 10 fs radially polarized pulse exhibiting PFT
and its isosurface plot at half of the maximum electric field value as it is
reconstructed with the TERMITES-MAZE method. (c-d) $y-t$ cross-section of the
y polarized electric field of the same pulse and its isosurface plot.
Due to the nature of the experiments, the pulses must be very well aligned.
This is particularly important for the alignment between the two experiments.
In order to achieve high precision, we perform the alignment with image
processing. A fit of a two-dimensional elliptical Gaussian function to the
intensity of the same pulse as captured by the sensor at the two different
experiments, gives the position of its center, the FWHM at different
orientations and the rotation angle. Since the pulse is the same, a comparison
of these values provides the exact spatial shift of the pulse on the sensor.
Then by cropping and flipping the corresponding images the two experiments are
aligned with pixel size precision.
Fig. 5 shows the spectral amplitude and spectral phase of the $x$ and $y$
polarizations of a radially polarized pulse at $794$ nm as it is derived by
the analysis. Only areas with significant energy are considered for the phase
maps. In these areas the pulse has a phase difference of $\pi$ between its
left and right (or top-bottom) parts as it is expected for a CVP due to the
polarization singularity in its center. The small deviation of the phase
profile from the cylindrical symmetry is due to the existence of relatively
weak space-time couplings, like pulse front tilt and curvature as it can be
seen in Fig. 6.
An inverse Fourier transform is used to reconstruct the 3 dimensional shape of
the pulse in the time domain. Fig 6a-d show the $x-t$ and $y-t$ cross-sections
for the $x$ and $y$ polarization respectively of a radially polarized pulse
and their 3d reconstruction as an isosurface at half of the maximum amplitude.
The pulse is 10 fs long and it is clear that apart from a small curvature and
a small pulse front tilt in the $x$ direction, it does not have significant
aberrations.
In contrast, in Fig. 7 we show the reconstruction of a pulse exhibiting PFT in
its $x$ direction. This is actually the initial pulse that was used for the
experiment and the PFT indicates a misalignment of the pulse compression
setup. The existence of the PFT offers an opportunity to apply the developed
method for the characterization and correction of the latter. The measured PFT
is found to be about 7 $\mathrm{fs/mm}$ and it was retrieved by fitting a
paraboloid to the spatially depended delay of the pulse which is directly
related to the spatially depended spectral phase provided by the analysis
[48]. A way to quickly correct the tilt is to insert a wedge. The induced PFT
due to a wedge can be calculated from the following equation [6]
$\displaystyle\xi=\frac{\lambda_{0}\tan(\theta)}{c}\frac{\mathrm{d}n}{\mathrm{d}\lambda}$
(6)
with $\lambda_{0}$ the central wavelength of the pulse, $c$ the speed of
light, $n$ the refractive index and $\theta$ the angle of the wedge. Here, we
have used a commercially available $5^{\circ}$ fused silica wedge plate in the
path of the pulse. From Eq. 6 we calculate the PFT introduced by the wedge to
be about 6 $\mathrm{fs/mm}$. The wedge was inserted in such a way that
influenced and corrected the profile of the pulse along the horizontal axis,
$x$ direction. The corrected pulse is shown in the Fig. 6 and exhibits a PFT
of about 0.5 $\mathrm{fs/mm}$, which is close to the expected value from Eq.
6. The fact that the existence of the PFT on the generated CVP was discovered
only during the development of the method highlights its practicality to any
field employing such pulses.
## IV Conclusions
In this paper we have presented a technique for the spatio-temporal
characterization of CVPs and we demonstrated its capabilities with the
complete spatio-temporal characterization of a 10 fs radially polarized pulse.
The technique is a Mach-Zehnder interferometer extension to the TERMITES
method that is necessary to overcome the limitation of the latter to
characterize pulses with polarization singularities. A radially polarized
pulse exhibiting pulse front tilt was also characterized to illustrate the
capabilities of the method.
In the current implementation, the method characterizes CVPs whose relative
phase between the two orthogonal polarizations is a priory known and hence it
is not measured. However, there are cases that this relative phase is not
known and it is the important quantity. Such cases are pulses with dynamic
polarization like polarization gates that are used for the generation of
isolated attosecond pulses [49] or pulses with polarization singularities that
are unknown to exist a priori. The TERMITES-MAZE method can be easily extended
for the characterization of such pulses by altering the setup to use a
reference pulse polarized at 45 degrees and interfere it with the x and y
components of the unknown pulse on a polarizing beam splitter, similar to the
POLLIWOG method [50].
The method is not limited by the bandwidth of the pulse since it is based on
an interferometric setup. Apart from the characterization of CVPs it can be
applied to ultrashort light structures possessing a number of polarization or
phase singularities. The latter is of paramount importance to transfer the
field of topological optics [51] from monochromatic beams to singular pulses
[42]. Additionally, all applications of CVPs such as free space communications
or micro-machining that depend on the propagation dynamics or light matter
interactions can directly benefit from the method acting as a feedback that
allows to tune the pulses to have the optimal shape or exploring the
generation and effects of new light structures to already existing
applications.
Finally, a software package for the analysis of the experimental data has been
developed in the python programming language and is freely available under the
BSD 3-clause licence [24].
## References
* Akturk _et al._ [2010] S. Akturk, X. Gu, P. Bowlan, and R. Trebino, Spatio-temporal couplings in ultrashort laser pulses, Journal of Optics 12, 093001 (2010).
* Pretzler _et al._ [2000a] G. Pretzler, A. Kasper, and K. Witte, Angular chirp and tilted light pulses in cpa lasers, Applied Physics B: Lasers and Optics 70, 1 (2000a).
* Vincenti and Quéré [2012] H. Vincenti and F. Quéré, Attosecond lighthouses: How to use spatiotemporally coupled light fields to generate isolated attosecond pulses, Phys. Rev. Lett. 108, 113904 (2012).
* Froula _et al._ [2018] D. H. Froula, D. Turnbull, A. S. Davies, T. J. Kessler, D. Haberberger, J. P. Palastro, S.-W. Bahk, I. A. Begishev, R. Boni, S. Bucht, _et al._ , Spatiotemporal control of laser intensity, Nature Photonics 12, 262 (2018).
* Sainte-Marie _et al._ [2017] A. Sainte-Marie, O. Gobert, and F. Quere, Controlling the velocity of ultrashort light pulses in vacuum through spatio-temporal couplings, Optica 4, 1298 (2017).
* Dorrer [2019] C. Dorrer, Spatiotemporal metrology of broadband optical pulses, IEEE Journal of Selected Topics in Quantum Electronics 25, 1 (2019).
* Bowlan _et al._ [2006] P. Bowlan, P. Gabolde, A. Shreenath, K. McGresham, R. Trebino, and S. Akturk, Crossed-beam spectral interferometry: a simple, high-spectral-resolution method for completely characterizing complex ultrashort pulses in real time, Opt. Express 14, 11892 (2006).
* Kosik _et al._ [2005] E. M. Kosik, A. S. Radunsky, I. A. Walmsley, and C. Dorrer, Interferometric technique for measuring broadband ultrashort pulses at the sampling limit, Opt. Lett. 30, 326 (2005).
* Wyatt _et al._ [2006] A. S. Wyatt, I. A. Walmsley, G. Stibenz, and G. Steinmeyer, Sub-10 fs pulse characterization using spatially encoded arrangement for spectral phase interferometry for direct electric field reconstruction, Opt. Lett. 31, 1914 (2006).
* Cousin _et al._ [2012] S. L. Cousin, J. M. Bueno, N. Forget, D. R. Austin, and J. Biegert, Three-dimensional spatiotemporal pulse characterization with an acousto-optic pulse shaper and a hartmann shack wavefront sensor, Opt. Lett. 37, 3291 (2012).
* Gabolde and Trebino [2006] P. Gabolde and R. Trebino, Single-shot measurement of the full spatio-temporal field of ultrashort pulses with multi-spectral digital holography, Opt. Express 14, 11460 (2006).
* Miranda _et al._ [2014] M. Miranda, M. Kotur, P. Rudawski, C. Guo, A. Harth, A. L’Huillier, and C. L. Arnold, Spatiotemporal characterization of ultrashort laser pulses using spatially resolved fourier transform spectrometry, Opt. Lett. 39, 5142 (2014).
* Pariente _et al._ [2016] G. Pariente, V. Gallet, A. Borot, O. Gobert, and F. Quéré, Space-time characterization of ultra-intense femtosecond laser beams, Nature Photonics 10, 547 (2016).
* Moh _et al._ [2006] K. Moh, X.-C. Yuan, J. Bu, D. Low, and R. Burge, Direct noninterference cylindrical vector beam generation applied in the femtosecond regime, Applied physics letters 89, 251114 (2006).
* Kong _et al._ [2019] F. Kong, H. Larocque, E. Karimi, P. B. Corkum, and C. Zhang, Generating few-cycle radially polarized pulses, Optica 6, 160 (2019).
* Carbajo _et al._ [2014] S. Carbajo, E. Granados, D. Schimpf, A. Sell, K.-H. Hong, J. Moses, and F. X. Kärtner, Efficient generation of ultra-intense few-cycle radially polarized laser pulses, Opt. Lett. 39, 2487 (2014).
* Papasimakis _et al._ [2018] N. Papasimakis, T. Raybould, V. A. Fedotov, D. P. Tsai, I. Youngs, and N. I. Zheludev, Pulse generation scheme for flying electromagnetic doughnuts, Phys. Rev. B 97, 201409 (2018).
* Trebino and Kane [1993] R. Trebino and D. J. Kane, Using phase retrieval to measure the intensity and phase of ultrashort pulses: frequency-resolved optical gating, J. Opt. Soc. Am. A 10, 1101 (1993).
* Iaconis and Walmsley [1998] C. Iaconis and I. A. Walmsley, Spectral phase interferometry for direct electric-field reconstruction of ultrashort optical pulses, Opt. Lett. 23, 792 (1998).
* Zdagkas _et al._ [2020] A. Zdagkas, N. Papasimakis, V. Savinov, and N. I. Zheludev, Space-time nonseparable pulses: Constructing isodiffracting donut pulses from plane waves and single-cycle pulses, Physical Review A 102, 063512 (2020).
* Strickland and Mourou [1985] D. Strickland and G. Mourou, Compression of amplified chirped optical pulses, Optics Communications 55, 447 (1985).
* Pretzler _et al._ [2000b] G. Pretzler, A. Kasper, and K. Witte, Angular chirp and tilted light pulses in cpa lasers, Applied Physics B: Lasers and Optics 70, 1 (2000b).
* Alonso _et al._ [2020] B. Alonso, I. Lopez-Quintas, W. Holgado, R. Drevinskas, P. G. Kazansky, C. Hernández-García, and Í. J. Sola, Complete spatiotemporal and polarization characterization of ultrafast vector beams, Communications Physics 3, 1 (2020).
* Zdagkas [2021] A. Zdagkas, Termites-maze python module (2021).
* Allen _et al._ [1992] L. Allen, M. W. Beijersbergen, R. J. C. Spreeuw, and J. P. Woerdman, Orbital angular momentum of light and the transformation of laguerre-gaussian laser modes, Phys. Rev. A 45, 8185 (1992).
* Rubinsztein-Dunlop _et al._ [2016] H. Rubinsztein-Dunlop, A. Forbes, M. V. Berry, M. R. Dennis, D. L. Andrews, M. Mansuripur, C. Denz, C. Alpmann, P. Banzer, T. Bauer, E. Karimi, L. Marrucci, M. Padgett, M. Ritsch-Marte, N. M. Litchinitser, N. P. Bigelow, C. Rosales-Guzmán, A. Belmonte, J. P. Torres, T. W. Neely, M. Baker, R. Gordon, A. B. Stilgoe, J. Romero, A. G. White, R. Fickler, A. E. Willner, G. Xie, B. McMorran, and A. M. Weiner, Roadmap on structured light, Journal of Optics 19, 013001 (2016).
* Porras [2002] M. A. Porras, Diffraction effects in few-cycle optical pulses, Phys. Rev. E 65, 026606 (2002).
* Hoff _et al._ [2017] D. Hoff, M. Krüger, L. Maisenbacher, A. Sayler, G. Paulus, and P. Hommelhoff, Tracing the phase of focused broadband laser pulses, Nature Physics 13 (2017).
* Zhang _et al._ [2020] Y. Zhang, D. Zille, D. Hoff, P. Wustelt, D. Würzler, M. Möller, A. M. Sayler, and G. G. Paulus, Observing the importance of the phase-volume effect for few-cycle light-matter interactions, Phys. Rev. Lett. 124, 133202 (2020).
* Forbes _et al._ [2019] A. Forbes, A. Aiello, and B. Ndagano, Classically entangled light, in _Progress in Optics_ , Vol. 64 (Elsevier, 2019) pp. 99–153.
* Zhan [2009] Q. Zhan, Cylindrical vector beams: from mathematical concepts to applications, Adv. Opt. Photon. 1, 1 (2009).
* Dorn _et al._ [2003] R. Dorn, S. Quabis, and G. Leuchs, Sharper focus for a radially polarized light beam, Phys. Rev. Lett. 91, 233901 (2003).
* Michihata _et al._ [2009] M. Michihata, T. Hayashi, and Y. Takaya, Measurement of axial and transverse trapping stiffness of optical tweezers in air using a radially polarized beam, Appl. Opt. 48, 6143 (2009).
* Moradi _et al._ [2019] H. Moradi, V. Shahabadi, E. Madadi, E. Karimi, and F. Hajizadeh, Efficient optical trapping with cylindrical vector beams, Opt. Express 27, 7266 (2019).
* Chen _et al._ [2013] R. Chen, K. Agarwal, C. J. R. Sheppard, and X. Chen, Imaging using cylindrical vector beams in a high-numerical-aperture microscopy system, Opt. Lett. 38, 3111 (2013).
* Li _et al._ [2011] X. Li, Y. Cao, and M. Gu, Superresolution-focal-volume induced 3.0 tbytes/disk capacity by focusing a radially polarized beam, Opt. Lett. 36, 2510 (2011).
* Milione _et al._ [2015] G. Milione, T. A. Nguyen, J. Leach, D. A. Nolan, and R. R. Alfano, Using the nonseparability of vector beams to encode information for optical communication, Opt. Lett. 40, 4887 (2015).
* Konrad and Forbes [2019] T. Konrad and A. Forbes, Quantum mechanics and classical light, Contemporary Physics 60, 1 (2019), https://doi.org/10.1080/00107514.2019.1580433 .
* McLaren _et al._ [2015] M. McLaren, T. Konrad, and A. Forbes, Measuring the nonseparability of vector vortex beams, Physical Review A 92, 023833 (2015).
* Allegre _et al._ [2012] O. J. Allegre, W. Perrie, S. P. Edwardson, G. Dearden, and K. G. Watkins, Laser microprocessing of steel with radially and azimuthally polarized femtosecond vortex pulses, Journal of Optics 14, 085601 (2012).
* Hellwarth and Nouchi [1996] R. W. Hellwarth and P. Nouchi, Focused one-cycle electromagnetic pulses, Phys. Rev. E 54, 889 (1996).
* Zdagkas _et al._ [2019] A. Zdagkas, N. Papasimakis, V. Savinov, M. R. Dennis, and N. I. Zheludev, Singularities in the flying electromagnetic doughnuts, Nanophotonics 8, 1379 (2019).
* Raybould _et al._ [2016] T. Raybould, V. Fedotov, N. Papasimakis, I. Youngs, and N. Zheludev, Focused electromagnetic doughnut pulses and their interaction with interfaces and nanostructures, Opt. Express 24, 3150 (2016).
* Raybould _et al._ [2017] T. Raybould, V. A. Fedotov, N. Papasimakis, I. Youngs, and N. I. Zheludev, Exciting dynamic anapoles with electromagnetic doughnut pulses, Applied Physics Letters 111, 081104 (2017), http://dx.doi.org/10.1063/1.4999368 .
* Lozovoy _et al._ [2004] V. V. Lozovoy, I. Pastirk, and M. Dantus, Multiphoton intrapulse interference. iv. ultrashort laser pulse spectral phase characterization and compensation, Opt. Lett. 29, 775 (2004).
* Shannon [1949] C. E. Shannon, Communication in the presence of noise, Proceedings of the IRE 37, 10 (1949).
* Machavariani _et al._ [2007] G. Machavariani, Y. Lumer, I. Moshe, A. Meir, and S. Jackel, Efficient extracavity generation of radially and azimuthally polarized beams, Opt. Lett. 32, 1468 (2007).
* Jolly _et al._ [2020] S. W. Jolly, O. Gobert, and F. Quéré, Spatio-temporal characterization of ultrashort laser beams: a tutorial, Journal of Optics 22, 103501 (2020).
* Sola _et al._ [2006] I. Sola, E. Mével, L. Elouga, E. Constant, V. Strelkov, L. Poletto, P. Villoresi, E. Benedetti, J.-P. Caumes, S. Stagira, _et al._ , Controlling attosecond electron dynamics by phase-stabilized polarization gating, Nature Physics 2, 319 (2006).
* Walecki _et al._ [1997] W. J. Walecki, D. N. Fittinghoff, A. L. Smirl, and R. Trebino, Characterization of the polarization state of weak ultrashort coherent signals by dual-channel spectral interferometry, Opt. Lett. 22, 81 (1997).
* Dennis _et al._ [2009] M. R. Dennis, K. O’Holleran, and M. J. Padgett, Chapter 5 Singular Optics: Optical Vortices and Polarization Singularities, in _Progress in Optics_, Vol. 53 (Elsevier, 2009) pp. 293–363.
|
# The Evolution of NGC 7465 as Revealed by its Molecular Gas Properties
Lisa M. Young Physics Department, New Mexico Institute of Mining and
Technology, 801 Leroy Place, Socorro, NM 87801, USA Adjunct Astronomer,
National Radio Astronomy Observatory, Socorro, NM 87801, USA David S. Meier
Physics Department, New Mexico Institute of Mining and Technology, 801 Leroy
Place, Socorro, NM 87801, USA Adjunct Astronomer, National Radio Astronomy
Observatory, Socorro, NM 87801, USA Martin Bureau Sub-department of
Astrophysics, Department of Physics, University of Oxford, Denys Wilkinson
Building, Keble Road, Oxford, OX1 3RH, UK Alison Crocker Department of
Physics, Reed College, Portland, OR 97202, USA Timothy A. Davis School of
Physics & Astronomy, Cardiff University, Queens Buildings, The Parade,
Cardiff, CF24 3AA, UK Selçuk Topal Department of Physics, Van Yüzüncü Yıl
University, Van 65080, Turkey
(Received 26 November 2020; Revised 25 January 2021; Accepted 27 January 2021)
###### Abstract
We present ALMA observations of CO isotopologues and high-density molecular
tracers (HCN, HCO+, CN, etc.) in NGC 7465, an unusually gas-rich early-type
galaxy that acquired its cold gas recently. In the inner 300 pc, the molecular
gas kinematics are misaligned with respect to all other galaxy components; as
the gas works its way inward it is torqued into polar orbits about the stellar
kinematically-decoupled core (KDC), indicating that the stellar KDC is not
related to the current gas accretion event. The galaxy also exhibits unusually
high 12CO/13CO line ratios in its nucleus but typical 13CO/C18O ratios. Our
calculations show that this result does not necessarily indicate an unusual
[12CO/13CO] abundance ratio but rather that 12CO (1-0) is optically thin due
to high temperatures and/or large linewidths associated with the inner
decoupled, misaligned molecular structure. Line ratios of the higher-density
tracers suggest that the densest phase of molecular gas in NGC 7465 has a
lower density than is typical for nearby galaxies, possibly as a result of the
recent gas accretion. All of the observed molecular properties of NGC 7465 are
consistent with it having acquired its molecular (and atomic) gas from a
spiral galaxy. Further detailed studies of the CO isotopologues in other
early-type galaxies would be valuable for investigating the histories of those
that may have acquired their gas from dwarfs. Finally, these ALMA data also
show an unidentified line source that is probably a background galaxy similar
to those found at $z=1-3$ in blind CO surveys.
Early-type galaxies (429) – Interstellar molecules (849) – CO line emission
(262) – Molecular gas (1073) – Galaxy evolution (594)
††facilities: ALMA††software: CASA (Emonts et al., 2019), Astropy (Astropy
Collaboration et al., 2013, 2018)
## 1 Introduction
The gas in early-type galaxies (ETGs) offers interesting clues into their
evolution, as there can be a much stronger disconnect between the stars and
the gas than there is in spiral galaxies. In spirals, the continuous star
formation activity implies that there has been a continuous supply of cold gas
over a Hubble time, and the stars and the gas have co-evolved in an
uninterrupted symbiotic relationship. In contrast, in early-type galaxies much
of the cold and warm gas is kinematically misaligned (even retrograde) with
respect to the stars, such that there is little possibility the stars formed
from that gas or the gas was expelled from those stars.
We expect irregular gas kinematics in the outer parts of galaxies where the
cold gas (especially HI) is more vulnerable to interactions and will retain
signatures of disruption for many Gyr. But such misalignments are also present
in the inner few kpc of early-type galaxies, where we find that 20% of the
regular, relaxed CO disks have polar or retrograde rotation with respect to
the stars in the host galaxy. Thus, as much as 30% to 50% of the molecular gas
in early-type galaxies has been acquired recently from some external source
such as an accreted dwarf galaxy (Davis et al., 2011). Similar results are
found for ionized gas (Jin et al., 2016; Bryant et al., 2019). Of the
remainder, some of the relaxed, prograde molecular gas may have been resident
in its host early-type galaxy for a long time, through the galaxy’s transition
to the red sequence (e.g. Davis & Bureau, 2016). Some of the molecular gas
probably also originated in its current host, but made an extended detour
through a hot phase in the galactic halo before condensing back into a cold
phase (e.g. Russell et al., 2019; Davis et al., 2019).
These varied histories also have implications for the physical properties of
the gas. If the cold gas in an early-type galaxy was accreted in a minor
merger or from a relatively pristine cold flow, the metallicity and isotopic
abundance patterns in the molecular gas could reflect a significantly
different star formation history and stellar initial mass function (IMF) from
those that produced the current stellar populations in the galaxy (Davis &
Young, 2019). On the other hand, if the gas has been resident in the galaxy
for a long time, gradually mixing with the mass loss material from evolved
stars in the galaxy, we would expect a much tighter correspondence between the
metallicity and isotopic abundances of the stars and the gas.
In this context we have been undertaking detailed studies of the chemical and
physical properties of the molecular gas in early-type galaxies; these
chemical and physical properties might also serve as indicators of the
galaxies’ histories in a complementary manner to the information revealed by
kinematics. Here we present Atacama Large Millimeter/submillimeter Array
(ALMA) observations of the early-type galaxy NGC 7465, which is known to have
recently accreted its cold gas (Section 2). We present 12CO (1-0) data at
0.′′8 (110 pc) resolution and 12CO, 13CO, C18O, and several other higher-
density molecular tracers at 2′′ (280 pc) resolution (Section 3). We describe
the millimeter continuum emission, compare it to the optical nebular emission
line ratios, and estimate the ionized gas metallicity (Section 4). We also
make comparisons of CO to stellar and ionized gas kinematics at matched
resolution (Section 5). Quantitative analysis of the molecular line ratios,
their spatial variations, and comparisons to other galaxy types (Sections 6 \-
11) reveal clues to the evolution of NGC 7465 and the broader context of
galaxy evolution.
## 2 About NGC 7465
NGC 7465 was originally selected because it is a member of the ATLAS3D sample
of early-type galaxies (Cappellari et al., 2011), so there is a wealth of
information about its stellar and gas content, including two-dimensional maps
of its stellar kinematics, ionized gas distribution and kinematics, stellar
populations, star formation history, atomic gas, and molecular gas (Krajnović
et al., 2011; Young et al., 2011; Serra et al., 2012; McDermid et al., 2015,
and references therein). It is one of the more CO-bright members of the
ATLAS3D sample, so there is extensive information on its 12CO J=1-0 emission.
Its stellar mass is $\log(M_{\star}/M_{\odot})=10.4$ and its adopted distance
is 29.3 Mpc (Cappellari et al., 2011).
Though it is an early-type galaxy, NGC 7465 shows moderate levels of current
star formation. Davis et al. (2014) estimated its total star formation rate
from its 22µm WISE flux, giving 1.37 $\pm\ 0.5$ M⊙ yr-1 and a specific star
formation rate SFR/M⋆ $\approx(5.5\pm 2)\times 10^{-11}$ yr-1. These rates are
low when compared to nearby spirals, placing NGC 7465 below the “star
formation main sequence” or SFR–M⋆ relation (e.g. Cluver et al., 2014) but
high when compared to other early-type galaxies (Davis et al., 2014). Its
depletion time (SFR/Mgas) is relatively long, at 7 Gyr. Its radio and far-IR
continuum emission imply a radio/FIR $q$ ratio that is also consistent with
star formation activity (Nyland et al., 2017).
In addition to being CO-bright, NGC 7465 is notable for signatures of
interactions with near neighbors NGC 7463 and NGC 7464 at projected
separations $<$ 20 kpc. Though it is an early-type galaxy (an S0 galaxy and a
fast rotator), it has faint blue outer arms at radii of 8 to 10 kpc (Figure
1). Its HI emission is highly disturbed and strongly misaligned with respect
to the stellar body of the galaxy (Li & Seaquist, 1994; Serra et al., 2012).
This kinematic evidence strongly suggests that the galaxy’s cold gas was
recently acquired from some external source, and this accretion/interaction
event probably drove the formation of the blue outer arms. The stellar
kinematics show a kinematically-distinct core (Krajnović et al., 2011) which,
as we will show, is probably not related to the most recent gas-transfer
interaction. A deep $u$-band image from Duc et al. (2015) also shows recent
star formation activity in an irregular spiral structure at radii $\approx$
3′′ to 20′′ ($\approx$ 0.4 to 3 kpc).
Figure 1: Top: color composite of the MATLAS $u,g,$ and $i$ images (Duc et
al., 2015). Two neighboring galaxies are visible off the northwest corner of
the image. Bottom: 12CO contours superposed on a $u-i$ color image (bluer
colors are white). Contour levels are 0.0164 $\times$ (1, 3, 9, 27, 81) Jy
bm-1 km s-1, where the lowest contour is the nominal column density
sensitivity equivalent to a 5$\sigma$ signal in one channel of width 15 km
s-1. The synthesized beam (0.′′8 $\times$ 0.′′7) is shown as a blue ellipse in
the bottom left corner. A primary beam correction has been applied; in both
panels the dotted circle indicates the half power point of the ALMA primary
beam at 114 GHz.
NGC 7465 is classified as a barred S0 galaxy by some (e.g. de Vaucouleurs et
al., 1991), and indeed Figure 1 might be interpreted as a nearly face-on
galaxy with a bar $\approx$ 1′ (8.5 kpc) in length. An alternative
interpretation arising from the ATLAS3D dynamical analysis (Cappellari et al.,
2013) is that the inner $r\lesssim 30$′′ of the galaxy is an edge-on
axisymmetric fast rotator and the outer blue arms are transient extraplanar
structures. The primary reason for this alternative interpretation is that the
inner part of NGC 7465 does not show typical kinematic features of strongly
barred early-type galaxies, which are (1) large misalignments between the
photometric major axis of the bar and the kinematic major axis of the stars in
that region, and (2) cylindrical rotation in the stellar kinematics, yielding
parallel iso-velocity contours (Krajnović et al., 2011; Lablanche et al.,
2012). Additional stellar kinematic data covering a larger field of view would
be helpful in distinguishing between these alternatives.
Finally, IRAM 30m observations of molecular lines in the ATLAS3D galaxies have
shown that NGC 7465 has some uncommon integrated molecular line ratios
(Crocker et al., 2012). Relative to 12CO, NGC 7465 has unusually faint 13CO
and HCN, and it has a low HCN/HCO+ ratio. At higher resolution, these line
ratios might reflect the influence of an active galactic nucleus on its
surrounding medium. The galaxy contains a Seyfert or LINER nucleus (e.g.
Gonçalves et al., 1999), a 5 GHz synchrotron point source (Nyland et al.,
2016) and a Swift hard X-ray source (Baumgartner et al., 2013). Our new ALMA
data, with additional lines and higher spatial resolution, provide a better
perspective on how to interpret the molecular line ratios and connect them to
the physical properties of the gas and to galaxy evolution.
## 3 Observations
NGC 7465 has been observed by ALMA in several projects. Here we used the Band
3 observations of 13CO, C18O, and CS in project 2018.1.01253.S (December
2018); 12CO, HCN, C2H, and HCO+ in 2016.1.01119.S (November 2016); and 12CO
and CN in 2018.1.01599.S (October 2018). We used the standard pipeline-
calibrated raw data and carried out continuum subtraction, imaging, and
cleaning ourselves. An additional round of phase self calibration was not
found to offer any improvement in the images because of the relatively modest
signal-to-noise ratios. For continuum subtraction we used zero-order or first-
order fits to the line-free channels in the visibility domain; for imaging, we
made a wide variety of images at varying channel widths and resolutions as
necessary to optimize the resolution or to match resolutions for resolved line
ratios. The times on source, along with fiducial (“natural” weighting) beam
sizes and rms noise levels, are indicated in Table 1 for the detected spectral
lines.
Line emission in the data cubes was cleaned down to about the 1$\sigma$ level
and primary beam corrections were applied. The 12CO integrated intensity
images (e.g. Figure 1) were created by modestly smoothing the data cube in
spatial and velocity dimensions and clipping on the smoothed cubes to create a
velocity-dependent mask defining the volume with emission. These 12CO masks
were then applied to the fainter lines for creating their integrated intensity
images.
Line fluxes reported in this paper include uncertainties which are only based
on the thermal fluctuations in the data, representing signal-to-noise
considerations; they do not include uncertainties in the absolute flux
calibration scale, such as those due to secular variability in the calibrator
sources. The flux calibration of ALMA data that have been processed with
standard calibration procedures is usually accurate to $\approx$ 5% to 10%
(e.g. Martín et al., 2019; Andrews et al., 2018; Remijan et al., 2019). Line
ratios between data taken in different observing setups and different tunings,
such as 12CO/13CO (see Table 1), have an additional uncertainty due to this
absolute flux calibration. However, line ratios between data taken
simultaneously using the same setup (e.g. 13CO/C18O or HCN/HCO+) do not have
this additional uncertainty.
In addition to the spectral lines mentioned above, we also have upper limits
on CH3OH($2_{k}-1_{k}$), SiO(2-1; $v=0$), and HNCO$(4_{0,4}-3_{0,3})$.
Furthermore, there is an unidentified line source in the vicinity of NGC 7465.
It seems to be a galaxy at a much higher redshift, and it is described in more
detail in Appendix A.
Table 1: Fiducial data cube parameters for NGC 7465
Line | Time | Beam size | rms | $\Delta v$
---|---|---|---|---
| (sec) | (′′) | (mJy bm-1) | (km s-1)
12CO(1-0) | 1590 | 1.60$\times$1.18 | 1.59 | 5.1
| 11733 | 0.81$\times$0.74 | 0.53 | 2.6
CN(1-0) | 11733 | 0.84$\times$0.76 | 0.066 | 45.0
13CO(1-0) | 3810 | 1.89$\times$1.81 | 0.38 | 13.3
C18O(1-0) | 3810 | 1.90$\times$1.82 | 0.35 | 13.3
CS(2-1) | 3810 | 2.12$\times$2.04 | 0.29 | 14.9
CH3OH($2_{k}-1_{k}$) | 3810 | 2.18$\times$2.06 | 0.35 | 15.1
HCO+(1-0) | 8890 | 1.70$\times$1.37 | 0.26 | 13.1
HCN(1-0) | 8890 | 1.71$\times$1.38 | 0.26 | 13.2
HNCO $(4_{0,4}-3_{0,3})$ | 8890 | 1.72$\times$1.39 | 0.28 | 13.3
C2H(N=1-0) | 8890 | 1.72$\times$1.38 | 0.29 | 13.4
SiO(2-1, $v=0$) | 8890 | 1.73$\times$1.39 | 0.28 | 13.5
Note. — Beam sizes and rms noise levels refer to images made with “natural”
uv-weighting and the listed channel widths. In the case of the 12CO and CN
observations, the channel widths in the last column are close to the best
velocity resolutions allowed by the data. For the other lines, the listed
channel widths are representative but not the best possible. At a distance of
29.3 Mpc, 1′′ corresponds to 140 pc.
## 4 Radio continuum and ionized gas
As we have observations at similar angular resolutions covering a large range
of frequencies from 86 to 115 GHz (with gaps), the data are suitable for
imaging the continuum intensity and estimating the spectral index of any
emission. NGC 7465 has two continuum point sources, one in the nucleus and one
5′′ south of the nucleus (Figure 2). The nuclear source has a falling spectral
index; its flux densities measured at 86.15 and 107.9 GHz are 742 $\pm$ 12
$\mu$Jy and 616 $\pm$ 18 $\mu$Jy, respectively. Parametrizing the flux density
as $S_{\nu}\propto\nu^{\alpha}$ therefore gives $\alpha=-0.83\pm 0.13.$
Similarly, using all of the available line-free frequencies in CASA’s multi-
term, multi-frequency synthesis deconvolver (Rau & Cornwell, 2011), we find
$\alpha=-0.59\pm 0.12.$ These estimates are consistent with each other and
both clearly indicate synchrotron emission in the nucleus of the galaxy, which
is not surprising given the other active galactic nucleus (AGN) indicators
mentioned in Section 2. For comparison, Domínguez-Fernández et al. (2020) find
that at 230 GHz the nuclear continuum source has a flux density of 900 $\pm$
200 $\mu$Jy, which is broadly consistent with the flux densities measured here
and suggests that in this mm regime the nuclear source’s spectrum is
flattening as dust emission begins to dominate.
The fainter continuum source has flux densities of 155 $\pm$ 12 $\mu$Jy and
245 $\pm$ 18 $\mu$Jy at 86.15 and 107.9 GHz, suggesting instead a rising
spectral index consistent with dust emission. It appears to be associated with
a region of recent star formation activity, as indicated by a bright blue
source in the $u$ image, peaks in [O III] and H$\alpha$ emission (Ferruit et
al., 2000), and a small peak in the molecular surface brightness.
Moustakas & Kennicutt (2006) presented optical spectroscopy and measurements
of the nebular emission line fluxes in NGC 7465. The nebular spectrum (from a
region 2.′′5 square) has log([N II]/H$\alpha$) = -0.28 $\pm$ 0.03 and log([O
III]/H$\beta$) = 0.44 $\pm$ 0.03, suggesting the ionization there is dominated
by the AGN (Moustakas et al., 2010). The spectrum integrated over a much
larger 50′′ $\times$ 60′′ rectangular region has log([N II]/H$\alpha$) = -0.39
$\pm$ 0.03 and log([O III]/H$\beta$) = 0.06 $\pm$ 0.03, consistent with star
formation activity. The ATLAS3D data provide more spatial information at
higher resolution, though only on the [O III]/H$\beta$ ratio; they show an
even higher ratio of log([O III]/H$\beta$) = 0.62 in the nucleus, on the
location of the nuclear radio synchrotron and hard X-ray sources (Figure 2).
Lower ratios of log([O III]/H$\beta$) from $-0.25$ to $-0.3$ are seen in
patches at radii of 3′′ to 10′′, and while this single ratio cannot
conclusively identify the source of ionization (e.g. Sarzi et al., 2010), all
the data are consistent with the interpretation that the current star
formation activity in the interior of the galaxy is particularly concentrated
in this annular region.
The nebular emission line fluxes in Moustakas & Kennicutt (2006) also enable
an estimate of the gas metallicity in NGC 7465. We use line fluxes integrated
over the large 50′′ $\times$ 60′′ region; following the prescriptions outlined
in Moustakas et al. (2010), the emission line ratios suggest a metallicity
consistent with solar. We find 12 + log(O/H) = 8.38 $\pm$ 0.06 using the
calibration of Pilyugin & Thuan (2005) and $\approx$ 9.0 using the calibration
of Kobulnicky & Kewley (2004); each of these is consistent with its respective
median of the SINGS galaxies (mostly spirals) in Moustakas et al. (2010).
Figure 2: The [O III]/H$\beta$ ratio from the ATLAS3D project (Cappellari et
al., 2011). Contours are the 3mm continuum emission, showing the two point
sources. Contour levels are ($-3$, 3, 10, and 25) times the rms noise, which
is 11 $\mu$Jy bm-1; the peak intensity is 590 $\mu$Jy bm-1. The beam size of
the continuum emission is shown as the solid blue ellipse in the lower left
corner.
## 5 Molecular Gas Distribution and Kinematics
### 5.1 Large-scale disk
The 12CO emission from NGC 7465 (Figure 1) shows a disky structure consisting
of irregular, flocculent spiral arms extending to radii of at least 20′′. Some
emission is also detected beyond the half-power point of the ALMA primary
beam, so it seems likely that there is significant molecular gas beyond the
region that we are able to image in these datasets. This irregular disk has an
axis ratio near 1, suggesting a low inclination, but its kinematic major axis
is strongly twisted and asymmetric as might be expected of recently accreted
gas (Figure 3). Although there is a large mismatch between the angular
resolution of the HI and CO data, the CO velocity field on large scales is
consistent with that of the HI (Li & Seaquist, 1994), suggesting that those
two gas phases probably have a common origin.
Section 2 discusses the fact that it is not clear whether NGC 7465 is a face-
on barred galaxy or an edge-on axisymmetric galaxy with extraplanar arm/ring
structures. In either case, steady-state models are likely to be misleading
when trying to interpret this galaxy’s kinematics. The interactions with its
neighbors mean that its potential is rapidly time-varying. Either these
interactions, or the perturbations caused by a bar (if one is present), might
explain the non-circular motions and possible gas inflows discussed in this
section. In addition, we note that dense molecular gas is found over most of
the extent of the elongated structure. In a bar model, this would imply that
either dense gas is present beyond the inner Lindblad resonance (ILR),
contrary to expectation (e.g. Athanassoula, 1992), or alternatively that the
radius of the ILR is much larger than is usual in early-type disk galaxies and
it extends to almost the end of the bar and thus corotation. Overall, we
therefore argue that it is problematic to interpret the gas distribution and
kinematics in NGC 7465 in terms of a relaxed/steady state early-type barred
disk galaxy potential.
### 5.2 Inner kpc and decoupled core
At radii $\lesssim$ 5′′ (where we have most of our molecular line detections)
the ATLAS3D stellar isophotes are flattened and their major axis is aligned to
the kinematically-decoupled stellar core (KDC; Figure 3, center panel). The
KDC is both photometrically and kinematically misaligned by 150° with respect
to stellar structures farther out. In this region the molecular gas forms a
bright ridge about 10′′ long (1.4 kpc), oriented along a northeast-southwest
axis. The ridge is misaligned by about 100° with respect to the stellar KDC
and 120° with respect to the stellar kinematic and photometric axes at larger
radii. The molecular gas shows a strong velocity gradient along the ridge, so
that the most extreme CO velocities are found along the ridge, about 1′′ from
the nucleus as defined by the radio continuum position. 12CO (2-1) emission at
higher resolution (Domínguez-Fernández et al., 2020) shows a patchy spiral
feature that can be followed inward to a radius of 2′′ and a smaller molecular
ridge 2′′ long, with a saddle at the location of the AGN. But the smaller
ridge observed in (2-1) emission maintains the same orientation as the more
extended ridge we observe in the present data; thus, on scales of 1′′ (140 pc)
the molecular ridge is still perpendicular to the stellar KDC.
Davis et al. (2011) discussed the kinematics of ionized and molecular gas in
early-type galaxies, and noted that the two phases agreed in all of the cases
they studied at moderate ($\approx$ 1 kpc) resolution. That is also still true
in NGC 7465 for radii $\gtrsim$ 2′′ (280 pc), but for smaller radii the
kinematics of the ionized and molecular gas diverge. The ionized gas continues
its inward twist such that at small radii the kinematic position angle of the
ionized gas nearly matches that of the stellar KDC but is 45° offset from the
molecular kinematic position angle. The molecular gas, on the other hand,
maintains its fixed orientation nearly perpendicular to the stellar KDC. This
divergence of the kinematics is suggestive of more complex noncircular motions
or perhaps multiple ionization sources in the galaxy’s nucleus, such that the
inner ionized gas is decoupled from the molecular gas. The ionized gas
kinematics in the center of NGC 7465 may even be tracing an AGN-driven ionized
outflow, and Appendix B presents a closer look at that evidence.
The mismatch between molecular and stellar kinematics in the center of NGC
7465 also means that the stellar KDC has not formed out of the molecular gas
currently present in the galaxy. It must represent a previous formation event,
or perhaps a much earlier episode of an extended interaction. More detailed
observations of the stellar populations in the KDC would be required to date
its formation (e.g. Sarzi et al., 2016).
Figure 3: CO, stellar, and ionized-gas velocity fields of NGC 7465. In the
12CO velocity field, the contours show the integrated CO intensity at
0.8′′$\times$ 0.7′′ resolution, with contour levels at 0.03, 0.3, and 1.0 Jy
bm-1 km s-1. The CO beam size is shown as a gray ellipse in the bottom left
corner. These CO data have slightly better angular resolution than the ground-
based optical data, which are from Cappellari et al. (2011). In the stellar
and ionized gas panels, the contours are stellar isophotes.
### 5.3 Other molecular species in the inner kpc
The integrated intensity maps of the other molecular lines are presented in
Figure 4. To avoid biases related to the different signal-to-noise ratios, a
velocity-dependent mask tracing the 12CO emission is used for all of the other
spectral lines. Their distributions can be broadly classified into two major
groups. 12CO, HCN, HCO+, CN, and C2H are centrally concentrated and peaked on
the nucleus of the galaxy, whereas 13CO and C18O show peaks a couple of
arcseconds ($\approx$ 1 beam) north and east of the nucleus rather than on the
nucleus itself. In 13CO, for example, the integrated intensity at the position
of the nucleus is a factor of two lower than farther out along the molecular
ridge. CS emission is relatively weak and while it appears to have a peak off
the nucleus, the spatially-resolved ratios (Section 6) suggest that this
appearance may be due to noise in the integrated intensity. Finally, HNCO,
CH3OH and SiO emission are not detected in these data.
Position-velocity slices through the data cubes (Figure 5) are consistent with
those from Domínguez-Fernández et al. (2020), showing the presence of two
kinematic structures in the central molecular gas of NGC 7465. At radii $<$
1.′′3 we find a nuclear ring or bar (or possibly an outflow) that is most
prominent in HCO+ and CN emission, also visible in HCN and 12CO emission, and
extremely weak or absent in 13CO emission. Exterior to 1.′′3 the primary
structure is the warped disk, where the 13CO emission is at its brightest, and
there the 13CO emission follows the 12CO emission. The appearance of a
Keplerian decline in the 12CO emission in Figure 5 is an artifact of the twist
in the kinematic position angle, so that at large radii this slice is closer
to the kinematic minor axis than the major axis.
Figure 4: Integrated intensity images of multiple molecular species in the
inner few kpc of NGC 7465. The position of the nuclear 3mm continuum source is
marked with a cross, to guide the eye. These integrated intensities are
computed using a velocity-dependent mask that follows the rotation of the gas,
as defined by 12CO emission. Each line’s synthesized beam is indicated by the
ellipse in the lower right corner, and the colors are scaled to the minimum
and maximum of each image. Integrated line fluxes towards the nucleus are
given in Table 2. For the fainter lines, a dotted white contour indicates
pixels where the integrated intensity has a signal-to-noise ratio $>2.5.$ The
corresponding 12CO contour is mostly outside this field of view. Figure 5:
Position-velocity slices at a position angle of +42°, one beam wide. The
slices cut through the nucleus and follow the molecular ridge, tracking the
kinematic major axis at small radii (see Figure 3). Velocities are measured
with respect to the systemic velocity, 1961 km s-1. Dotted lines at $\pm$
1.′′3 demarcate the inner kinematic component with the strong velocity
gradient. 12CO is plotted both at the highest available resolution and at a
resolution matching the 13CO data.
## 6 Line strengths: CO isotopologues
For the bright lines of NGC 7465, we can create line ratio maps from the
integrated intensity images, as in Figure 6 for 12CO/13CO. For the fainter
lines we measure integrated line fluxes and ratios by defining spatial regions
(Figure 7), summing the cubes within the spatial regions to produce integrated
spectra (e.g. Figure 8), and then summing the spectra over velocity ranges
defined by the 12CO emission. This procedure ensures that the same data volume
is used for both lines of a ratio, even if their signal-to-noise ratios are
very different. Spatially-resolved line ratios measured in this manner are
presented in Figures 9 and 10. Reference column densities for the nuclear
spectrum in the optically thin and local thermodynamic equilibrium (LTE)
approximations are listed in Table 2 and line fluxes for all of our defined
regions are presented in Appendix C.
For an alternate method of exploring spatial variations in line ratios, we
also implemented a “shift and stack” technique. In this method all of the
spectra within a given annulus (about the nuclear 3mm continuum peak) can be
aligned in velocity, using the 12CO velocity field at each position to
determine the velocity shift. Thus all of the expected signal can be
concentrated in a narrow range of frequencies. However, because of the
departures from azimuthal symmetry in the center of NGC 7465, ratios produced
by this technique have significantly larger uncertainties than those derived
from individual regions.
Figure 6: 12CO/13CO integrated intensity ratio (computed in units of
brightness temperature). Black contours show the 3mm continuum emission, and
the angular resolution of the 12CO/13CO ratio image is shown by the gray
ellipse in the lower left corner.
Figure 7: Integrated intensities of 12CO (top) and 13CO (bottom). The angular
resolution in each panel is shown with a gray ellipse in the lower left
corner. Black and white polygons show the regions used for summing spectra, to
investigate the spatial variations of the line ratios. The boundary of the
nuclear region is dotted. A few white pixels in the lower portion of the top
panel are locations where 12CO is not detected at this angular resolution.
Numbers in the lower panel identify the regions for reference to the table in
Appendix C. Figure 8: Spectra extracted from the nuclear region, which is 2′′
(1 beam) in diameter and centered on the 3mm continuum point source. The grey
boxes mark the velocity range used for integrating at this position. In the
C2H panel, the two dotted lines mark the centers of the two main fine
structure blends and the velocity scale is calculated for the low frequency
($J=\frac{3}{2}-\frac{1}{2}$) blend. These nuclear spectra exhibit unusually
weak 13CO emission. Figure 9: Line ratios are integrated over the regions
shown in Figure 7; the horizontal error bars indicate the range of radii
probed by each region. Ratios measured towards the nucleus are plotted as
stars, and those measured over the entire region with detected 12CO emission
are plotted at 10′′ and circled. The region with relatively low 12CO/HCN at
5.′′3 is centered on the fainter 3mm continuum source. For purposes of
clarity, only the detections ($>3\sigma$) are plotted. Figure 10: Similar to
Figure 9, for different ratios; symbols have the same meanings. The radial
coordinates for some ratios have been slightly shifted for better visibility.
Table 2: Line fluxes and reference column densities in the nucleus of NGC 7465
Species | Line flux | $\log N$ (20 K) | $\log N$ (90 K)
---|---|---|---
| (Jy km s-1) | (cm-2) | (cm-2)
12CO | 003.52 (0.04) | 17.14 | 17.69
13CO | 000.096 (0.014) | 15.65 | 16.20
C18O | $<$ 0.039 | $<$ 15.26 | 15.81
HCN | 000.132 (0.012) | 12.80 | 13.37
HCO+ | 000.303 (0.013) | 13.39 | 13.96
CN | 000.226 (0.006) | 13.61 | 14.16
CS | 000.041 (0.009) | 13.02 | 13.54
C2H | 000.091 (0.014) | 14.70 | 15.27
CH3OH | $<$ 0.035 | $<$ 13.95 | $<$ 14.57
SiO | $<$ 0.040 | $<$ 12.79 | $<$ 13.33
HNCO | $<$ 0.037 | $<$ 13.49 | $<$ 14.19
Note. — Line fluxes are measured in a circular region 2′′ (280 pc) in
diameter, centered on the nuclear 3mm continuum source; column densities are
averaged over this region. Statistical uncertainties of the line fluxes are
listed in brackets, or the upper limits are quoted as 3 times the statistical
uncertainty of the sum over the adopted velocity range. For 13CO the peak
intensity is offset from the nucleus and a 2′′ region centered on the peak has
a line flux of 0.015 Jy km s-1. Column densities are calculated for two
representative temperatures in the optically thin LTE assumption. Current data
do not constrain the optical depths of the HCN and HCO+ transitions, but CN
and C2H are optically thin (Section 8) and even 12CO might be at this position
(Section 11.1).
NGC 7465 has relatively weak 13CO emission. Measured 12CO(1-0)/13CO(1-0)
ratios are highest at 39 $\pm$ 9 in the nucleus, or 33 $\pm$ 5 when averaged
over a circle of 1′′ radius, decreasing to $\approx$ 15 $\pm$ 2 at a radius of
1 kpc (see Figures 6 and 9).111All line ratios reported in this paper are
computed using integrated fluxes expressed in temperature units (K km s-1).
The large ratios in the center of NGC 7465 are higher than those typically
found in nearby spirals, which tend to be in the range of 8–20 (especially on
scales $>$ 1 kpc, e.g. Crocker et al., 2012; Cao et al., 2017; Cormier et al.,
2018; Israel, 2020). For comparison, Figure 11 shows typical 12CO/13CO line
ratios for galaxies of various types and illustrates that the high ratios in
the nucleus of NGC 7465 are similar to those found in ULIRGs and advanced
mergers such as Arp 220 and NGC 2623 (Brown & Wilson, 2019). Figure 11 also
shows that the range of 12CO/13CO line ratios exhibited by early-type galaxies
is a factor of 10, broader than the range exhibited by any other galaxy type.
Other early-type galaxies known to have high 12CO/13CO ratios include NGC 1266
and UGC 09519 (Crocker et al., 2012); NGC 1266 is remarkable for having a
strong AGN-powered molecular outflow (Alatalo et al., 2014), while UGC 09519
has a kinematically-misaligned HI disk that suggests it was recently acquired
(Serra et al., 2014). NGC 7465 is thus consistent with the pattern that high
12CO/13CO ratios tend to be associated with major disturbances to the ISM.
NGC 7465’s strong radial gradient in 12CO/13CO is also uncommon. It amounts to
a factor of two over an unresolved length scale, such that the ratios at radii
$<$ 140 pc are at least a factor of two higher than those outside and possibly
more. For context, the spirals in Paglione et al. (2001), Cao et al. (2017),
and Cormier et al. (2018) sometimes have enhancements of 13CO in their nuclei
and sometimes deficits, but none of them are as extreme as in NGC 7465. Two
barred lenticular galaxies are in fact similar to spirals in this respect
(Topal et al., 2016), with typical 12CO/13CO ratios about 5 to 15, and with
the higher ratios in the central kpc or so; they only exhibit modest gradients
of a factor of two over kpc scales. Measurements at sub-kpc resolution
sometimes show spatial gradients of the same magnitude as those in NGC 7465.
Meier & Turner (2004) found variations of factors of 5 over $\approx$ 200 pc
in the center of NGC 6946, but unlike NGC 7465, the spiral shows an
enhancement of 13CO in its center. The strong deficit of 13CO in the center of
NGC 7465 is more reminiscent of the nearby early-type galaxy Cen A, where
McCoy et al. (2017) find 12CO/13CO $>$ 20 in the circumnuclear disk at radii
$<$ 200 pc and 12CO/13CO reaching nearly to unity farther out in the arms.
Figure 11: A compilation of observed 12CO/13CO line ratios, in the same spirit
as Figure 1 of Zhang et al. (2018). Symbols in grey are dwarf galaxies,
spirals, and ULIRGs; at the highest luminosities they are sub-mm galaxies at
$z\approx 2-3.$ Data are drawn from Davis (2014), Jiménez-Donaire et al.
(2017), Brown & Wilson (2019), Cormier et al. (2018), Henkel et al. (2014),
Braine et al. (2017), and Heikkilä et al. (1999); see also the citations in
Zhang et al. (2018). The large light error bars for the SMC and the LMC show
the full range of ratios measured by Israel et al. (2003) at 12 pc resolution.
Open grey circles are stacks of multiple galaxies (Méndez-Hernández et al.,
2020). We use an improved FIR luminosity for IC 10 from Rémy-Ruyer et al.
(2015). Line ratios of early-type galaxies (orange symbols) are compiled from
Crocker et al. (2012) and Alatalo et al. (2015); Cen A is from McCoy et al.
(2017). For NGC 7465 and Cen A (green and purple, respectively), the error
bars show the full range of ratios measured in the galaxy.
Measurements of 13CO/C18O in NGC 7465 are limited to a few spatial regions
with detections of C18O (e.g. Figure 12). In a large spatial region defined by
all of the detected 12CO emission, we have a tentative detection of C18O
yielding 13CO/C18O = 3.0 $\pm$ 1.0. The bright molecular ridge has a ratio of
4.0 $\pm$ 1.2. Figure 13 shows that these ratios are similar to those of
galaxies of comparable IR luminosity (e.g. Jiménez-Donaire et al., 2017).
Further interpretation of the CO isotopologue line ratios is in Section 11.
Figure 12: 13CO and C18O in NGC 7465. These spectra are from a region at the
northeast end of the molecular ridge, where the 13CO emission is brightest.
Gaussian fits are overlaid, though our analysis uses line ratios derived from
simple sums rather than from the fits. Since the C18O line is so faint, its
width and center velocity have been constrained to match those of 13CO in the
fits. Figure 13: A compilation of observed 13CO/C18O line ratios, after
Figure 1 of Zhang et al. (2018). For NGC 4526 (Young et al. in prep) and Cen A
(McCoy et al., 2017), the error bars show the full range of ratios measured in
the galaxy, with the symbol at the median ratio. In addition to the citations
noted in Figure 11 and Zhang et al. (2018), this figure includes measurements
by Johansson et al. (1994) and Wang et al. (2009).
## 7 High-Density tracers
Figures 9 and 10 show that, as expected, the emission from HCN, HCO+, and C2H
is enhanced relative to 12CO and 13CO in the center of NGC 7465. This pattern
is common in nearby galaxies and is expected to be driven by higher gas
densities in the centers (e.g. Jiménez-Donaire et al., 2019). We also observe
radial trends in C2H/HCN and C2H/HCO+, in the sense that C2H is more strongly
enhanced just outside of the nucleus than in the nucleus itself. HCN/HCO+ and
12CO/CS are effectively constant wherever we can measure them, but CS/HCN
increases with radius. All of these results are discussed in greater detail
below.
### 7.1 HCN and HCO+
Relative to 12CO, NGC 7465 has fairly typical HCN brightness compared to other
nearby galaxies. Specifically, 12CO/HCN in Figure 9 ranges from 16 $\pm$ 1 to
50 $\pm$ 15\. These ratios are very similar to their analogs measured in the
nearby spirals of the EMPIRE survey by Jiménez-Donaire et al. (2019) and in
two other lenticulars by Topal et al. (2016). The EMPIRE data have somewhat
lower linear resolution (1 to 2 kpc) than our measurements in NGC 7465 (240
pc) but in all of these cases the disks are well resolved.
In contrast, the HCN in NGC 7465 is relatively faint compared to that in a
larger sample of early-type galaxies; in the unresolved single-dish
measurements of Crocker et al. (2012), NGC 7465 has the highest 12CO/HCN ratio
of the sample. The unresolved measurements undoubtedly fold in some radial
gradients, so it is difficult to make detailed comparisons at different scales
and it would be worthwhile to obtain additional resolved 12CO/HCN measurements
in early-type galaxies.
The large difference in critical or effective density between CO and HCN
implies that observed ratios like 12CO/HCN encode information about the
density distribution of the molecular gas, or equivalently the relative
proportions of low-density and high-density gas (e.g. $10^{2}$ – $10^{3}$ cm-3
vs. $10^{5}$ – 106 cm-3; Meier & Turner, 2012; Leroy et al., 2017). Thus, NGC
7465 has a mix of low/high-density gas which is typical of most nearby
spirals, but it is relatively deficient in high-density gas compared to many
other local early-type galaxies. The proportion of high-density gas is
maximized in the nucleus but is also high towards the southern continuum
source, where we find the second-lowest 12CO/HCN and 12CO/HCO+ ratios in the
galaxy as well as some of the lowest reliably-measured 12CO/13CO ratios. This
region must contain a concentration of dense gas with relatively high CO
optical depth.
Relative to 12CO, NGC 7465 is exceptionally bright in HCO+ emission when
compared to nearby galaxies. We measure local 12CO/HCO+ ratios from $6.9\pm
0.3$ in the nucleus to $20.7\pm 2.9$ in the disk. In comparison, the spiral
galaxies of the EMPIRE survey usually have 12CO/HCO+ $>25$ even in their
nuclei (Jiménez-Donaire et al., 2019). Thus, the regions of NGC 7465 with the
faintest HCO+, at radii $\approx 1$ kpc, still have brighter HCO+ than is
common even in the nuclei of the spirals. It is also notable that the HCO+
emission in the nucleus of NGC 7465 has a dramatically different line shape
than 12CO (Figure 8), which probably reflects differing spatial distributions.
### 7.2 HCN/HCO+
Because HCO+ is so bright, the HCN/HCO+ ratios in NGC 7465 are low; they are
significantly $<1$ everywhere we are able to measure them, and this is unusual
for nearby galaxies. We measure HCN/HCO+ = 0.44 $\pm$ 0.04 in the nucleus,
with ratios elsewhere ranging up to 0.60 $\pm$ 0.11. In nearby galaxies it is
much more common to find HCN/HCO+ $>$ 1 (e.g. Jiménez-Donaire et al., 2019;
Topal et al., 2016). Ratios $<$ 1 are seen in restricted parts of a few
spirals (e.g. parts of NGC 4254 and Maffei 2; Meier & Turner, 2012; Gallagher
et al., 2018). Low ratios of HCN/HCO+ $\approx$ 0.5 are occasionally seen in
other early-type galaxies (Crocker et al., 2012) and more commonly in ULIRGs
and starburst galaxies (e.g. Baan et al., 2008; Krips et al., 2008; Privon et
al., 2015; Sliwa & Downes, 2017). Extremely low HCN/HCO+ ratios $\lesssim$ 0.2
have been measured in dwarf galaxies like the LMC and IC 10 (Seale et al.,
2012; Braine et al., 2017; Kepley et al., 2018; Anderson et al., 2014). In
this respect, NGC 7465 has more in common with starburst or dwarf galaxies
than with nearby spirals. The low HCN/HCO+ and HCN/CO ratios in NGC 7465 also
place it firmly amongst the nuclear starburst galaxies and “composite”
Seyferts in the HCN(1-0) diagnostic diagram of Kohno et al. (2001).
Spatial variations in the HCN/HCO+ ratios of NGC 7465 do not significantly
exceed the ratios’ mutual statistical uncertainties, so there is no compelling
evidence for a radial gradient (Figure 10). Even though there is an AGN in NGC
7465, there is definitely no enhancement in HCN/HCO+ towards the nucleus,
unless it is on length scales of tens of pc where it would be obscured by our
spatial resolution. There is certainly nothing like the factor of 2 rise in
HCN/HCO+ seen in the central 0.5 kpc of NGC 1068 (Viti et al., 2014).
### 7.3 Physical properties of the dense molecular gas
Due to the difference in critical densities, the HCN/HCO+ line ratio is most
commonly interpreted as an indicator of density in the high-density ($n\sim
10^{4}-10^{6}$ cm-3) portion of the molecular gas. However, for the lowest
HCN/HCO+ line ratios in the LMC, Anderson et al. (2014) have appealed to a
combination of low density and low gas-phase metallicity, with low metallicity
contributing in two ways. First, lower heavy element abundances will produce
lower column densities of these species and smaller optical depths in their
transitions, which is necessary for the line ratio to deviate from 1. On top
of that, lower metallicities are associated with lower [N/O] abundance ratios
(van Zee & Haynes, 2006), which might manifest as lower abundances of HCN
relative to HCO+ (Braine et al., 2017).
NGC 7465 does not have a low metallicity (Section 4), so there is no reason to
assume a low [N/O] ratio a priori. However, the observed line ratios in NGC
7465 do constrain the abundance ratio [HCN/HCO+] to be significantly smaller
than that inferred for NGC 253 by Meier et al. (2015), which was [HCN/HCO+]
$\approx$ 5\. In LTE, reproducing our observed line ratios of HCN/HCO+ = 0.44
$\pm$ 0.04 requires an abundance ratio [HCN/HCO+] $\leq 0.25$ (Table 2).222We
adopt the convention that ratios written without brackets (e.g. HCN/HCO+)
refer to measured line ratios whereas ratios inside square brackets (e.g.
[HCN/HCO+]) refer to estimated molecular abundances after correcting for
optical depth and excitation effects. In the more likely situation that LTE
does not apply, we make calculations using the RADEX large velocity gradient
code (van der Tak et al., 2007) to account for optical depth and excitation
effects on the HCN/HCO+ line ratio; broad ranges of parameters will reproduce
the data from NGC 7465 but they always require an abundance ratio [HCN/HCO+]
$\lesssim 3,$ and a ratio of 5 is ruled out as it cannot reproduce the line
ratios in NGC 7465. For an assumed [HCN/HCO+] = 3, relatively low densities
$n_{\rm H_{2}}\leq 10^{4.3}$ cm-3 are required to reproduce the observed line
ratios. Thus, even though the metallicity of NGC 7465 is not low enough to
suggest unusually low N abundance, the observed HCN/HCO+ line ratios in NGC
7465 are best explained with a combination of relatively low-density gas
and/or a relatively low HCN abundance (or enhanced HCO+). More specific
estimates of the density and temperature will require additional transitions
or isotopologues, but the inference of a low HCN abundance is robust to other
variations in physical properties.
Viewing the molecular gas as two “phases” – a more diffuse phase with
densities $\sim 10^{2}$ to $10^{3}$ cm-3 traced by CO, and a denser phase
$\gtrsim 10^{5}$ cm-3 traced by HCN and similar species – the strong spatial
trends in diffuse/dense gas tracers suggest that the relative proportions of
the two phases change with radius in NGC 7465. However, the internal
properties of the denser phase do not seem to change, even though the
properties in the more diffuse phase are dramatically varying (Section 11.1).
## 8 C2H and CN: photodissociation regions
The C2H molecule is regarded as a tracer of photon-dominated or
photodissociation regions (PDRs), as its formation is driven by the presence
of C+, and it has been shown to be brighter on the illuminated sides of
molecular clouds (e.g. Meier & Turner, 2005, 2012; Meier et al., 2015; García-
Burillo et al., 2017). The two main fine structure blends in our data are
$N=1-0$, $J=\frac{3}{2}-\frac{1}{2}$, with a rest frequency of 87.3169 GHz,
and $N=1-0$, $J=\frac{1}{2}-\frac{1}{2}$, at 87.40199 GHz. They are separated
by 292 km s-1, which is just greater than the velocity range covered by 12CO
in NGC 7465, so they are not overlapping. The low-frequency component is
detected in NGC 7465 but the high-frequency component is not; as the high-
frequency component is a factor of 2.3 fainter in LTE in the optically thin
limit, its nondetection is not surprising, but we can rule out line ratios
$\approx$ 1 (the optically thick limit). Thus we cannot measure the C2H
optical depth directly but the data are consistent with the transitions being
optically thin. Further, the C2H line fluxes quoted in this paper refer only
to the blend at 87.32 GHz.
The C2H emission in NGC 7465 is relatively bright, when compared to HCN. We
find C2H/HCN ratios ranging from 0.71 $\pm$ 0.13 in the nucleus to 1.0 $\pm$
0.2 in the outer parts of the molecular ridge. These are significantly higher
than the corresponding ratios in the central kpc of NGC 253, which range from
0.17 to 0.47 ($\pm$ 15%; Meier et al., 2015). The C2H/HCN ratios in NGC 7465
are also on the high side when compared to the set compiled by Martín et al.
(2014) for active and starbursting galaxies. On the other hand, ratios of
C2H/HCO+ in NGC 7465 are very similar to those in NGC 253 and other galaxies.
We find 0.31 $\pm$ 0.05 to 0.50 $\pm$ 0.11 in NGC 7465, where Meier et al.
find 0.20 to 0.57 ($\pm$ 15%) in NGC 253 and most of the sample galaxies in
Martín et al. (2014) show a similar range. These results suggest that HCN
emission in NGC 7465 is relatively faint because of moderate densities in the
molecular gas; both C2H and HCO+ are relatively prominent because their
critical densities are a factor of 10 lower than that of HCN.
In addition to being bright, the C2H emission in NGC 7465 is also notable in
that the radial trends of C2H/HCN and C2H/HCO+ in NGC 7465 are opposite to the
trends in NGC 253. In the spiral those ratios peak at the nucleus; in other
words, C2H is more strongly concentrated than HCN and HCO+ are. In NGC 7465 it
is the opposite – C2H is less centrally concentrated than HCN and HCO+. The
radial trend in NGC 7465 thus suggests enhanced C2H emission just outside of
the nucleus, at radii 2′′ to 5′′ or 280 to 700 pc (as far out as we can detect
C2H). If the C+ ionization is indeed dominated by star formation activity,
this radial trend might be consistent with the ionized gas emission line
ratios in Figure 2 (Section 4): the lowest [O III]/H$\beta$ ratios (most
suggestive of star formation activity) occur slightly outside the nucleus at
radii $\approx$ 3′′ to 10′′. Similarly, from the stellar population analysis
of Krajnović et al. (2020), there is a suggestion that the smallest stellar
ages occur at a radius of 2.′′5 rather than at larger or smaller radii.
As another tracer of C+, we note that NGC 7465 is also detected in [C II]
158µm emission (Lapham et al., 2017), though the spatial resolution is not
good enough to compare to the radial trends in our ALMA data. Many spiral
galaxies are known to have significant central deficits of [C II] emission at
radii $<$ 1 kpc, and many potential explanations have been proposed (e.g.
Smith et al., 2017). Possibly, in some cases, the central [C II] deficit
arises because an AGN’s hard radiation field drives the ionization balance
towards C2+; in other cases, a softer radiation field (caused by old stellar
populations) may be driving the ionization balance towards neutral carbon. At
present we simply comment that the central C2H deficit in NGC 7465 (relative
to HCN and HCO+) might be consistent with what is known about the ionization
states of C in other galaxy nuclei.
The CN(1-0) transition is also a PDR tracer (Boger & Sternberg, 2005), though
it is expected to be even more strongly enhanced in X-ray dominated regions
(XDRs) and may thus help to distinguish between PDRs and XDRs (Meijerink et
al., 2007). It has fine structure components that appear in these data as
blends with rest frequencies of about 113.17 GHz and 113.49 GHz; they are
clearly separated in our data and both are clearly detected. The high-
frequency blend is the brighter one and there is no evidence that their line
ratio in NGC 7465 ever deviates from the theoretical optically thin limit of 2
(e.g. Meier et al., 2015). Thus, the CN emission in NGC 7465 is optically
thin. For simplicity, we show only the brighter component in the figures and
the line fluxes we quote refer only to it.
The range of 12CO/CN ratios we measure in NGC 7465 is large, from $15.0\pm
0.4$ in the nucleus to $93\pm 14$ when averaged over the full 12CO emission
region, and it is consistent with the wide range of ratios measured in other
nearby galaxies (Wilson, 2018). CN/HCN ratios are consistent with 1 aside from
one region, and this again is very typical of local galaxies (Ueda et al.,
2017; Cicone et al., 2020).
There is some evidence that CN is even more centrally concentrated in NGC 7465
than the other high-density tracers like HCN. The radial variations of 12CO/CN
exceed those of all other 12CO ratios (Figure 9). The transition in line
profile shapes, from double-peaked 12CO through flat-topped HCO+ to centrally-
peaked CN (Figures 8 and 5), might thus suggest the presence of a high-density
circumnuclear disk at $r\lesssim 100$ pc. CN might be particularly enhanced
close to the AGN because of its X-ray sensitivity (Meijerink et al., 2007).
Higher resolution data, especially for CN, would be valuable for testing this
scenario. As a caveat we note that the current data do not show a measurable
radial trend in CN/HCN (Figure 10), but such a trend could be obscured by the
resolution and sensitivity of our data. We do find a systematic difference in
the spatial distribution of the two PDR tracers, C2H and CN; C2H is less
centrally concentrated than HCN whereas CN is equally or possibly more
centrally concentrated than HCN. The differences between these two species
might be consistent with an interpretation that C2H is a better tracer of star
formation activity whereas CN is a better tracer of XDRs.
## 9 CS
The CS molecule is also a high-density tracer, whose $J=2-1$ transition has a
critical or effective density intermediate between those of HCN and HCO+
$J=1-0$ (e.g. Leroy et al., 2017). There is also some evidence that CS
emission is enhanced in PDRs, where the radiation field is stronger (Lintott
et al., 2005; Meier & Turner, 2005), which has motivated suggestions that CS
emission traces massive star formation even better than HCN emission does
(e.g. Bayet et al., 2009; Davis et al., 2013).
NGC 7465 has 12CO/CS ratios that are fairly typical compared to those of other
nearby galaxies. We find 12CO/CS = 62 $\pm$ 13 in the nucleus, with variations
of no more than 25% in all of the regions where we detect CS. For comparison,
the spirals in Gallagher et al. (2018) have typical ratios of 12CO/CS
$\approx$ 90 in their central kpc, so NGC 7465 is brighter in CS than these.
However, it is faint in CS compared to the other early-type galaxies in Davis
et al. (2013), which have 12CO/CS in the range of 9 to 44. The spatially-
unresolved (IRAM 30m) ratios in Davis et al. (2013) might also be biased high
if the beam-filling factor of 12CO is higher than CS, so that resolved
measurements of CS would probably increase the discrepancy between them and
NGC 7465.
CS/HCN, which should provide better insight (than 12CO/CS) into the properties
of the dense molecular phase, is very similar in NGC 7465 to other nearby
galaxies. We find 0.26 $\pm$ 0.06 in the nucleus of NGC 7465 and 0.57 $\pm$
0.14 for the larger regions that extend to 0.8 kpc. These ratios are
consistent with those in the spirals Maffei 2 and NGC 253 (0.2 to 1; Meier &
Turner, 2012; Meier et al., 2015). The spirals in Gallagher et al. (2018) also
typically have CS/HCN $\approx$ 0.1 to 0.3, with localized ratios up to 0.5 in
the central 200 pc of NGC 3627 or the arms of NGC 4321. CS/HCN in NGC 7465 is
also consistent with the single-dish measurements of some other early-type
galaxies in Davis et al. (2013). It is notable that others of the early-type
galaxies in that sample have highly unusual ratios of CS/HCN as large as 3, so
that bright CS emission might be a feature of some early-type galaxies.
In terms of radial trends, it is striking that the nucleus of NGC 7465 does
not differ from the off-nuclear regions in terms of 12CO/CS; in contrast,
there are strong radial trends of factors of 2–4 in 12CO/HCN, 12CO/HCO+,
12CO/C2H and 12CO/CN. We also find no radial variation of HCN/HCO+ but a
factor of two in CS/HCN. If the density of the molecular gas were the only
factor driving changes in those line ratios, we would expect HCN, HCO+, and CS
to behave in qualitatively the same fashion as the critical density of CS is
intermediate between those of HCO+ and HCN. Evidently, then, we require more
than simple density variations to explain the behavior of these high-density
tracers in NGC 7465.
Davis et al. (2013) have also studied unresolved CS emission in a small sample
of early-type galaxies. They compared it to [O III]/H$\beta$, which can serve
as an indicator of whether the ionization in the optically-emitting ionized
gas is dominated by AGN activity or star formation (see also Section 4). In
their early-type galaxies, the global [O III]/H$\beta$ ratio is correlated
with CS/HCN in the sense that galaxies whose ionization is more dominated by
star formation (smaller [O III]/H$\beta$; softer radiation fields) have higher
CS/HCN. NGC 7465 is consistent with this trend, both in terms of its
integrated line ratios and its internal spatially-resolved behavior, for which
we find higher CS/HCN and lower [O III]/H$\beta$ (and bluer optical colors) at
$r\approx$ 0.5 to 1 kpc than towards the nucleus. NGC 7465 thus provides
circumstantial evidence supporting the interpretation that CS emission is
enhanced by PDR conditions.
For the ionized gas metallicity estimates in Section 4 and the column density
estimates in Table 2, NGC 7465 is also consistent with the suggestions of
Bayet et al. (2012) and Davis et al. (2013) that CS/HCN could be used as an
indicator of the metallicity of the molecular gas. Future work on spatially-
resolved nebular emission line data might also be valuable to test whether the
trend in Davis et al. (2013) also applies within individual galaxies.
## 10 Shock tracers
Emission from SiO, CH3OH, and HNCO is commonly interpreted as tracing shocks
in the ISM (e.g. Meier & Turner, 2012; Meier et al., 2015), as significant
energy input in the form of shocks is required to liberate them from grain
surfaces. These molecules are not detected in NGC 7465, with the most
stringent limits being 12CO/(SiO, CH3OH, or HNCO) $>$ 83 near the center of
NGC 7465. Ratios relative to 12CO are not particularly significant in terms of
a physical interpretation, but here merely serve as an indication of the
relative faintness of the lines compared to detections in other galaxies. For
context, in NGC 253, Meier et al. (2015) found spatially-resolved ratios of
12CO/SiO and 12CO/HNCO $\approx$ 20 to 80; NGC 253 thus has brighter SiO and
HNCO than NGC 7465, by at least a factor of 4 relative to 12CO. The
circumnuclear disk of NGC 1068 has 12CO/SiO = 12.5 $\pm$ 1.5 (García-Burillo
et al., 2010), which again is at least a factor of 7 brighter in SiO than NGC
7465. Finally, Topal et al. (2016) also measured 12CO/HNCO = 66 $\pm$ 5 and 78
$\pm$ 18 in the centers of the lenticulars NGC 4710 and NGC 5866. Thus, these
two also have modestly brighter HNCO than NGC 7465.
For CH3OH, Davis et al. (2013) detected a small sample of early-type galaxies
with unresolved measurements of 12CO/CH3OH in the range 14 to 70. As we
measure 12CO/CH3OH $>$ 83, we conclude that NGC 7465 is also relatively faint
in CH3OH emission. Similarly, unresolved measurements of CH3OH/HCN reach as
high as 1.5 to 2.2 in the early-type galaxies NGC 6014 and NGC 5866 (Davis et
al., 2013), whereas the resolved upper limits in NGC 7465 are 0.2 to 0.8.
In short, SiO, CH3OH, and HNCO are relatively weak in NGC 7465 compared to
other galaxies that have been studied to date. The relatively weak emission
from these shock tracers is a bit surprising, given the obvious disturbances
in the galaxy (Sections 2 and 5). The molecular and HI disks are strongly
warped and their kinematics are significantly misaligned with respect to those
of the stars, so it is clear that the gas was recently acquired from an
external source. We might have expected stronger emission from shock tracers
as the gas disk is still in the process of settling; possibly the shocks are
primarily occurring farther out in the galaxy, where our sensitivity is not
currently good enough to detect the shock tracers.
## 11 Discussion: constraints on abundances and galaxy evolution
### 11.1 CO isotopic ratios in NGC 7465
Figure 11 shows that the center of NGC 7465 boasts an unsually high 12CO/13CO
line ratio; it is reminiscent of the ratios commonly observed in LIRGs and
ULIRGs rather than those in spirals and other early-type galaxies. It also
constrains the [12CO/13CO] abundance in the nucleus of NGC 7465 to be $\geq$
39 $\pm$ 9, which is consistent with most measurements in the disk of the
Milky Way, in other nearby spirals, and in part of the LMC (e.g. Johansson et
al., 1994; Meier et al., 2008; Romano et al., 2019, and references therein).
However, it is incompatible with the very low abundance ratios of [12C/13C] =
9 ($\pm$ 2), 21 ($\pm$ 6), and 24 ($\pm$ 1) estimated in the centers of NGC
4945, NGC 253, and the Milky Way (Langer & Penzias, 1990; Tang et al., 2019;
Martín et al., 2019). Thus, the chemical enrichment pattern in the central
molecular gas of NGC 7465 differs from the pattern in these nearby spirals.
The smaller 13C abundance in the center of NGC 7465 is plausibly connected
with the clear signs of gas accretion, as the gas currently in the center of
NGC 7465 was recently in a different galaxy and possibly in the outskirts of
that galaxy. For context, the stellar dynamical analysis of Cappellari et al.
(2013) implies a maximum circular rotation speed of 163 km s-1, so the orbital
time at the outer edge of the molecular arms in our data (15′′ or 2.1 kpc) is
80 Myr. It may have taken a few orbital times for the gas to make its way
inward to the center of the galaxy, and this timescale is comparable to or
shorter than the timescale for significant 13C production (e.g. Romano et al.,
2019).
Low inferred optical depths in 12CO(1-0) (see below) also imply that both 13CO
and C18O are optically thin and thus the measurements of [13CO/C18O] in NGC
7465 are 2.98 $\pm$ 0.99 and 4.0 $\pm$ 1.2. These measurements are made in
overlapping regions so they are not independent.
The optical depths of the CO transitions also convey information about the
physical properties of the molecular gas, and if the [12CO/13CO] abundance
ratio in the center of NGC 7465 is in the typical range of 40–60, then the
12CO emission is unusually optically thin. We thus consider several possible
explanations for the unusually high 12CO/13CO line ratio in the center of NGC
7465 and its strong radial gradient.
(1) Unusual isotopic abundances due to a burst of star formation? Away from
the nucleus, the 12CO/13CO and 13CO/C18O line ratios in NGC 7465 are similar
to those in nearby spirals. But a very recent burst of star formation in the
center of the galaxy would affect the isotopic abundances, as 13C in
particular comes from low-mass stars with long lifetimes and a region
undergoing a starburst might therefore be deficient in 13C. Indeed, this
effect probably contributes to the exceptionally high ratios of 12CO/13CO
$\approx$ 90 in the advanced merger NGC 2623 (Brown & Wilson, 2019). In NGC
7465, although there is some current star formation activity in the central
kpc (Sections 2 and 4), the star formation rate is modest compared to what we
usually think of as starbursts. The luminosity-weighted mean stellar ages are
younger in the center of the galaxy (Krajnović et al., 2020) but neither the
ages themselves nor the radial gradient in stellar ages are unusual even for
early-type galaxies. Ultimately a full multi-level isotopic abundance study
would be necessary to resolve the question, but with current data the
motivation for assuming unusual isotopic abundances in NGC 7465 is not
compelling.
(2) Unusual molecular abundances due to fractionation or photodissociation of
the rarer isotopes? The photodissociation of CO and H2 molecules usually
proceeds through absorption-driven excitation, meaning that the molecules can
shield themselves from photodissociation if they have sufficient column
densities to make the relevant UV transitions optically thick. 13CO and C18O
molecules should then be found deep within the dark interior of a molecular
cloud, in the same way (but more so) that 12CO molecules should be found
deeper than H2 and there may be a skin of CO-dark H2 around a UV-illuminated
cloud. However, this isotope-specific photodissociation is probably not the
explanation for the high 12CO/13CO ratio in the center of NGC 7465. The C18O
species should be even less abundant than 13CO (e.g. Meier et al., 2008;
Romano et al., 2019). Thus, photodissociation effects predict that unusually
large 12CO/13CO line ratios should also be accompanied by unusually large
13CO/C18O ratios, which we do not find in NGC 7465 (Figure 13). We further
note that the center of NGC 7465 does not exhibit particularly unusual levels
of photodissociation in general, as its [C II]/CO line ratio is typical for
nearby spiral galaxies (Figure 14). Finally, Viti et al. (2020) argue that
fractionation is unlikely to be a significant effect for the CO isotopologues
in typical conditions, though it may be more important for other molecular
species.
Figure 14: Matched-resolution [C II] and 12CO intensities for early-type
galaxies (circles; Lapham et al., 2017; Werner et al., 2014; Wilson et al.,
2013; Mittal et al., 2011, 2012). For comparison, the greyscale shows a two-
dimensional histogram of the same lines in the spiral galaxies of the KINGFISH
and HERACLES surveys (Kennicutt et al., 2011; Leroy et al., 2009). The angular
resolution of all these data is 12′′, which is $\lesssim$ 500 pc in the
spirals but $\approx$ 2 kpc in the early-type galaxies.
(3) Variations in the physical properties of the gas? The spatial variation of
12CO/13CO, with higher ratios in the central 100 pc of the galaxy (Figure 6),
suggests that the optical depth of both molecular species is lower in the
center. For typical abundances and conditions, the optical depth of 12CO (1-0)
must change by a factor of a few such that it is $>1$ at 1 kpc and $<1$,
possibly as low as 0.1, at 100 pc. Thus, while the 12CO column density
estimates in Table 2 are underestimates, they are not as far off as usually
assumed, and they are probably less than a factor of 2 too low. Further, the
inferred variations of optical depth could be plausibly explained by radial
variations of the temperature and/or density of the molecular gas.
Bayet et al. (2013) used the $\mathrm{J}=1-0$ and $2-1$ transitions of 12CO
and 13CO to constrain the densities and temperatures of the molecular gas in
18 early-type galaxies, including NGC 7465. They found the most probable
conditions for NGC 7465 to be in the range $10^{4.5}$ to $10^{5.6}$ cm-3 and
80 to 150 K. Notably, these are the highest temperatures inferred for the
sample of 18. These estimates are based on single-dish spectra, so they are
spatially unresolved and dominated by the conditions at small radii where the
emission is brightest. However, for these densities and temperatures,
calculations using the RADEX code (van der Tak et al., 2007) suggest that
increases in the temperature of a factor of 3 (e.g. from 30 K to 90 K) could
account for the larger 12CO/13CO 1-0 ratio in the central 100 pc of NGC 7465.
Density changes could have a similar effect on 12CO/13CO but would then also
produce a factor of at least 2 variation in HCN/HCO+, which we have ruled out.
Therefore, unless the CO-emitting gas is completely disconnected from the HCN-
emitting gas, temperature variations are more likely than density variations
to explain the 12CO/13CO gradient in NGC 7465. On the other hand, Bayet et al.
(2013) did not make use of resolved kinematic information or the fact that the
molecular line widths all increase dramatically towards the nucleus of NGC
7465, which can also affect the line ratios and is discussed in more detail
below.
(4) Lower opacity due to higher line width? It is striking that the highest
12CO/13CO ratios in NGC 7465 are coincident with the largest linewidths and
strongest local velocity gradients in the galaxy (Figure 5; Appendix B). The
12CO line profiles in the disk at radii of 5′′ to 20′′ (0.7 to 2.8 kpc) have
full widths at half maximum (FWHM) of 23.5 $\pm$ 5 km s-1, whereas the line
profiles in the nucleus are much broader and double-peaked (Figure 8). At
0.′′8 (110 pc) resolution, the line profile toward the nucleus has a FWHM of
230 km s-1; that is 10 times larger than in the disk of the galaxy. And since
the optical depth of a transition depends on the ratio of the total column
density to the line width (e.g. Paglione et al., 2001; van der Tak et al.,
2007), such a dramatic increase in the line width could produce a
corresponding factor of 10 decrease in the optical depth and a rise in the
12CO/13CO ratio.
With the present data it is not possible to distinguish whether the 12CO
optical depths in NGC 7465 are more significantly affected by temperature
changes or increases of the line widths. Some of the line profiles in the
centers of the lenticulars NGC 4710 and NGC 5866 have similarly large
linewidths to NGC 7465 but they do not have unusually large 12CO/13CO ratios
(Topal et al., 2016). Thus, it’s not yet clear what the difference is between
these cases and NGC 7465. Additional data on higher-energy transitions would
help distinguish between the possibilities.
For context, some authors have suggested that mechanical feedback from star
formation increases local velocity dispersions in molecular gas, produces low-
density, diffuse gas, and contributes to increases in 12CO/13CO ratios near
regions of active star formation (e.g. Tan et al., 2011). But in most spirals
the variations of linewidths and 12CO/13CO have smaller magnitude than we
find. Conversely, other studies of early-type galaxies have found that
dynamically regular, relaxed disks in high-density environments like the Virgo
Cluster tend to have low 12CO/13CO ratios, and those could be due to a
combination of relaxed kinematics and ram-pressure stripping of low-density
gas (e.g. Crocker et al., 2012; Alatalo et al., 2015).
AGN-driven outflows might also disturb their surrounding molecular gas,
increasing its linewidths and its 12CO/13CO line ratios (e.g. NGC 1266,
Crocker et al., 2012). Further inspection of the 12CO/13CO ratios in Figure 6
shows that the ratios are high (the optical depths are low) not just towards
the AGN but also in a band stretching along the minor axis of the molecular
ridge. In combination with the unusually misaligned gas kinematics in the
center of the galaxy (Figure 3; Appendix B), this pattern in the 12CO/13CO
ratio image suggests that there might be a modest-velocity ionized gas outflow
from the AGN stirring up the gas along the minor axis of the molecular ridge.
### 11.2 Implications of the CO isotopic abundances for galaxy evolution
Comparisons of the stellar and gas kinematics in early-type galaxies provide
clear evidence that much – perhaps as much as half – of the gas in early-type
galaxies has been acquired from some outside source, after the main stellar
body was in place (Sarzi et al., 2006; Davis et al., 2011; Bryant et al.,
2019). Additional detailed work on the gas/dust ratios and associations with
morphological disturbances suggests that most of the “outside sources” are
minor mergers with mass ratios $\gtrsim 10:1$ (Kaviraj et al., 2012; Davis et
al., 2015). In some cases there are also additional clues about the origin of
the gas from similarities or differences between the metallicities of the gas
and the stars; for gas accreted in a minor merger, one might expect the
metallicity of the accreted gas to be lower than the metallicity of the bulk
of the stars (Griffith et al., 2019; Davis & Young, 2019).
NGC 7465 clearly did acquire its cold gas from an external source, given the
kinematic mismatches between the CO, HI, and the stars. But the gas probably
did not come from a dwarf galaxy in a minor merger. The large HI content
(9.5$\times 10^{9}$ M⊙; Serra et al., 2012) is most consistent with the gas
coming from a large spiral galaxy. Indeed, there are several other similarly
gas-rich galaxies still in its group (Li & Seaquist, 1994; Serra et al.,
2012). Furthermore, the metallicity of the ionized gas in NGC 7465 is
consistent with that of a spiral rather than a dwarf (Section 4). And finally,
the inferred CO isotopic ratios in NGC 7465 are also consistent with it having
captured its molecular gas from a spiral. Its 13CO/C18O ratios are entirely
unremarkable for spirals; its nuclear 12CO/13CO line ratio is a factor of 2 to
3 higher than is common in spirals, but the measurement can be plausibly
attributed to high temperatures and/or large linewidths rather than to
abundance variations.
For other early-type galaxies with smaller gas masses, CO isotopic ratios may
serve as an important complement to metallicity for identifying the original
host of their external-source cold gas. Or, indeed, for early-type galaxies
with relaxed prograde gas, these indicators should help identify externally-
sourced cold gas even when the kinematics are inconclusive. As 13C is a slow-
release element with significant production in low- to intermediate-mass
stars, it may retain signatures of an external origin longer than 12C or 18O,
and possibly also longer than kinematic disturbances (e.g. if the gas is now
located in the interior of the galaxy where dynamical timescales are short).
Figures 11 and 13, which are motivated by the discussions in Taniguchi &
Ohyama (1998), Zhang et al. (2018), and Méndez-Hernández et al. (2020),
illustrate that ULIRGs and sub-mm galaxies tend to have high 12CO/13CO line
ratios $(\gtrsim 20)$ and low 13CO/C18O ratios $(\lesssim 2)$. In fact, there
also appears to be a strong correlation between these line ratios and the
total FIR luminosity of a galaxy. The cause of such a correlation is not
obvious; presumably the total FIR luminosity of a galaxy is a proxy for other
properties that are more fundamental. Zhang et al. (2018) argued it is a proxy
for the IMF, though Romano et al. (2019) cast doubt on that interpretation
given the highly uncertain production rates of 18O.
In any case, a thorough understanding of the chemical evolution of galaxies
requires that we should be able to connect the sub-mm galaxies at $z\approx
2-3,$ at the highest luminosities in Figures 11 and 13, to their descendants
in the local universe. It is usually assumed that the sub-mm galaxies have
actual abundance ratios [13CO/C18O] $\approx$ 1 to 2, just like their observed
line ratios, because their high 12CO/13CO line ratios indicate that 13CO and
C18O should be optically thin (e.g. Danielson et al., 2013; Sliwa et al.,
2017). Over time the 13C abundance should increase due to its production in
low- to intermediate-mass stars. And as the sub-mm galaxies already have
stellar masses typically $10^{10}$ to $10^{11.5}$ M⊙ at $z\approx 2-3$
(Boogaard et al., 2019), their modern-day descendants will be amongst the most
massive local galaxies. But large galaxies in the local universe tend to have
higher ratios of [13CO/C18O] or [13C/12C][16O/18O] around 5 to 6 (e.g. Martín
et al., 2019; Harada et al., 2018; Romano et al., 2019). Whether the chemical
evolution models can reproduce these kinds of variations on the appropriate
timescales depends a great deal on their assumptions, particularly involving
the production of the rare isotopes (Romano et al., 2019). Further isotopic
observations of massive early-type galaxies, particularly those that might
have retained their molecular gas through the cosmic star formation peak and
through their transition to the red sequence, would be useful for testing the
chemical evolution models.
On a related note, Bayet et al. (2012) and Davis et al. (2013) have argued
that the [HCN/CO] abundance ratio might be a useful indicator of
$\alpha$-element enhancement in molecular gas. As this $\alpha$ enhancement is
common in the stellar populations of early-type galaxies, studies of [HCN/CO]
and [HCN/CS] abundance ratios (Section 9) could provide additional chemical
clues to the evolution of early-type galaxies and the origin of their gas.
## 12 Summary
We present spatially-resolved molecular line observations of NGC 7465, an
unusually gas-rich early-type galaxy that recently acquired $\approx 10^{10}$
M⊙ of atomic and molecular gas in an interaction with another galaxy. Its
disturbed kinematics indicate that the gas is still in the process of settling
and migrating inward. We analyze ALMA observations of 12CO (1-0) at 0.′′8
resolution (110 pc) plus the 3mm lines of 13CO, C18O, HCO+, HCN, CN, C2H, and
CS at $\approx$ 2′′ resolution (280 pc) and the continuum emission. Besides
NGC 7465, the data reveal an unidentified line source with a relatively bright
peak line flux density of 2.5 mJy; it is probably a galaxy at $z=0.2$ or 1.4.
We find two 3mm continuum sources in NGC 7465; the brighter one is a nuclear
synchrotron source associated with the AGN that is also detected in low-
frequency radio and hard X-ray emission. The fainter one is associated with
molecular clouds and star formation activity, traced by local peaks in 12CO,
13CO, HCN, HCO+, and CS (though not CN).
The 12CO (1-0) distribution at 0.′′8 resolution (110 pc) shows a low-
inclination flocculent disk with a strong kinematic position angle twist
spiraling inwards to a linear ridge or bar-like feature where the highest
velocities are found. In the inner 300 pc of the galaxy, the molecular gas
kinematics are misaligned by $\approx$ 120° with respect to the large-scale
stellar rotation, by $\approx$ 100° (in the other direction) with respect to
the stellar kinematically-decoupled core, and by about 45° with respect to the
ionized gas kinematics. We conclude that the prominent stellar kinematically-
decoupled core did not form out of the molecular gas present now; it must have
had its origin in a previous event. Furthermore, the complex misalignments may
be signatures of outflows or other non-circular kinematics in ionized gas.
Despite the dramatic kinematic misalignments, there is no evidence of enhanced
emission from the shock tracers SiO, CH3OH, or HNCO.
All of the detected molecules are centrally concentrated except for 13CO (and
possibly C18O and CS). The distribution of 13CO has a prominent central dip of
almost a factor of two in integrated line intensity and it peaks at about 300
pc from the nucleus. The central 13CO dip produces an unusually large nuclear
12CO/13CO line ratio, $39\pm 9$ at this resolution. This ratio is higher than
those found in typical spirals and early-type galaxies; it is comparable to
those measured in ULIRGs and $z\approx 2-3$ sub-mm galaxies. It also
constrains the abundance ratio [12CO/13CO] $\geq 39\pm 9,$ which is typical of
spiral disks but is higher than sometimes found in spiral nuclei. The isotopic
ratios are thus consistent with the gas having been accreted from another
spiral galaxy and transported rapidly (faster than the 13C enrichment
timescale) to the nucleus of NGC 7465. The observed 13CO/C18O line ratio
suggests [13CO/C18O] = 4.0 $\pm$ 1.2, which is also consistent with a spiral
galaxy origin but is significantly higher than corresponding ratios estimated
in ULIRGs/sub-mm galaxies and lower than ratios in Local Group dwarfs.
Modest star formation activity is occurring in the center of NGC 7465 but
there is no compelling reason to assume the intrinsic [12CO/13CO] abundance
ratio is very different from the range of $40-60$ that is usually found in the
disks of spirals. In this case the 12CO (1-0) emission from the nucleus of NGC
7465 must be unusually optically thin, perhaps even having $\tau<1.$ Such low
optical depths are plausible because of (1) the high CO temperatures $\approx
100$ K inferred from single-dish data on multiple J-level transitions, and (2)
the very large linewidth in the center of the galaxy. The nuclear spectrum
($r<140$ pc) has a FWHM of 250 km s-1, a factor of 10 larger than elsewhere in
the galaxy, due to the strong velocity gradient in the central misaligned
molecular structure (possibly an edge-on circumnuclear ring).
HCN emission is relatively faint in NGC 7465 and HCO+ is relatively bright,
yielding HCN/HCO+ ratios in the range of 0.4 to 0.6 everywhere we can measure
– these ratios are a factor of two lower than typical for spiral galaxies. An
assumption of LTE requires an intrinsic abundance ratio [HCN/HCO+] $<0.25$ but
even for densities too low to approach LTE the data still require [HCN/HCO+]
$<3,$ which is smaller than is found in some nearby spirals. We also find
(based on previously published nebular line fluxes) roughly solar metallicity
in the ionized gas outside of the nucleus, where ionization is not dominated
by the AGN. Thus, even though the galaxy’s gas-phase metallicity does not
suggest an intrinsically low N abundance, the observed HCN/HCO+ line ratios do
still seem to require relatively low densities in the dense molecular phase
(e.g. $n_{\rm H_{2}}\leq 10^{4.3}$ cm-3) and/or low HCN abundances.
We find no measurable gradient in the HCN/HCO+ line ratio on scales of 100 pc
to 1 kpc. Thus, even though there is an AGN in NGC 7465, it is not affecting
the J=1-0 line ratios of the dense molecular tracers on those scales (in
marked contrast to some other AGN). On the other hand, the proportion of dense
relative to diffuse molecular gas is clearly changing with radius, as
reflected by ratios like 12CO/HCN, 12CO/HCO+, and 12CO/CN. Thus the physical
properties of the densest phase of the molecular gas do not appear to change
with radius, even though the properties of the more diffuse phase must be
changing to reproduce the 12CO/13CO variation.
The CN emission from NGC 7465 is optically thin and unremarkable in its
intensity. C2H is relatively bright, with C2H/HCN $\approx$ 1\. As C2H and
HCO+ have quite similar critical densities, we infer that the relative
brightness of C2H is also an indicator of relatively low densities in the
“high-density” molecular phase. The intensity and spatial distribution of CS
emission in NGC 7465 are consistent with previous suggestions that CS can be
enhanced in regions of star formation activity, here traced by low [O
III]/H$\beta$ ratios.
All of the gas-phase data gathered about NGC 7465 to date – the gas content,
kinematics, metallicity, and molecular and isotopic abundance patterns – are
consistent with the interpretation that it acquired its gas recently from a
large spiral galaxy rather than a dwarf galaxy. In the broader context,
however, more work is needed on isotopic abundance ratios in early-type
galaxies. Some of them are believed to have accreted their gas recently from a
dwarf galaxy, while others may have retained small quantities of molecular gas
from their previous lives as sub-mm galaxies at $z\approx 2-3,$ through the
Universe’s peak star formation epoch, all the way down to the present day.
More quantitative work is also required on modeling the isotopic abundance
evolution of early-type galaxies, to test whether the current-day properties
of these galaxies can be reproduced and what constraints they may impose on
the evolutionary models.
We thank Davor Krajnović for helpful discussions on the stellar kinematics of
barred galaxies. This paper makes use of the following ALMA data:
ADS/JAO.ALMA#2018.1.01253.S,
ADS/JAO.ALMA#2016.1.01119.S, and
ADS/JAO.ALMA#2018.1.01599.S. ALMA is a partnership of ESO (representing its
member states), NSF (USA) and NINS (Japan), together with NRC (Canada), MOST
and ASIAA (Taiwan), and KASI (Republic of Korea), in cooperation with the
Republic of Chile. The Joint ALMA Observatory is operated by ESO, AUI/NRAO and
NAOJ. The National Radio Astronomy Observatory is a facility of the National
Science Foundation operated under cooperative agreement by Associated
Universities, Inc.
## Appendix A A distant, unidentified source near NGC 7465
We find strong line emission and 3mm continuum emission from an unidentified
source near (in projection) to NGC 7465. The line is centered at 97.67 GHz and
its width is 0.14 GHz. The source is modestly resolved in these data, and its
position is 23h 02m 02.s857, $+$15° 58′ 17.′′8 (ICRS). Figure 15 shows its
location on a deep optical image from the MATLAS survey (Duc et al., 2015) and
Figure 16 shows its spectrum. It also has a 3mm continuum flux density of
(0.11 $\pm$ 0.01) mJy/beam and a slight suggestion of rotation along a
northwest – southeast axis, but the distance between the image centroids in
the extreme channels is only 0.′′35 so the rotation is not well resolved in
these data.
Although the source is superposed on the outer parts of NGC 7465, where there
are bright blue patches associated with recent star formation activity, there
is no optical source coincident with the 3mm emission. It certainly is not a
part of NGC 7465 due to its large linewidth and the fact that its frequency
does not match any known bright line. But as we only have one detected
spectral line in the frequencies covered by these data, it is difficult to
identify the line. If it is 12CO(1-0), it would be at $z=0.18.$ In this case
13CO(1-0) would be around 93.4 GHz, where we have no coverage, and HCN(1-0)
would be beyond the low frequency end of ALMA Band 3. If the line is
12CO(2-1), it would be at $z=1.3604$, 13CO(2-1) would be at 93.3 GHz, and
HCN(3-2) would be at 112.6 GHz. No line is apparent at 112.6 GHz but the data
do not rule out typical HCN/12CO ratios like those seen in local galaxies.
Similarly if the line is 12CO(3-2) at $z=2.5404,$ HCN(4-3) would be at 100.1
GHz but the limits on a nondetection there are not yet useful. Firm
identifications of the redshift in this case will probably require searches
for the corresponding higher-J transitions of 12CO in the higher ALMA bands or
the lower-J transitions at the Jansky Very Large Array.
Table 3: Flux, luminosity and mass estimates for the unidentified source
Line ID | $z$ | Diameter | Line width | Mdyn | Flux | $L^{\prime}_{\mathrm{line}}$ | Mmol | Mmol/Mdyn
---|---|---|---|---|---|---|---|---
| | (kpc) | (km s-1) | (M⊙) | (Jy km s-1) | (K km s-1 pc2) | (M⊙) |
12CO(1-0) | 0.1802 | 1.06 | 409 | 5.2$\times 10^{9}$/$\sin^{2}i$ | 0.87 | 1.4$\times 10^{9}$ | 4.9$\times 10^{9}$ | 0.95 $\sin^{2}i$
12CO(2-1) | 1.3604 | 2.94 | 217 | 4.0$\times 10^{9}$/$\sin^{2}i$ | 0.46 | 1.1$\times 10^{10}$ | 4.0$\times 10^{10}$ | 10.0 $\sin^{2}i$
12CO(3-2) | 2.5404 | 2.81 | 115 | 1.1$\times 10^{9}$/$\sin^{2}i$ | 0.24 | 8.1$\times 10^{9}$ | 2.9$\times 10^{10}$ | 27 $\sin^{2}i$
Note. — Calculations are made assuming a flat universe with H0 = 70 km s-1
Mpc-1, $\Omega_{m}=0.3$, and $\Omega_{\Lambda}=0.7$. The linear diameter is
estimated from the emission centroids in the outermost channels. The line
luminosity $L^{\prime}_{\mathrm{line}}$ is calculated as in Carilli & Walter
(2013). Estimated luminosity conversions from J=2-1 or J=3-2 to J=1-0 are made
using assumed excitations as in Boogaard et al. (2019) and the molecular mass
(with He) is then estimated from the inferred 12CO(1-0) luminosity using a
conversion factor $\alpha=3.6$ M⊙ (K km s-1 pc2)-1 (Boogaard et al., 2019).
Table 3 lists estimated line fluxes, luminosities, and masses for the most
probable line identifications. Assuming the object is in dynamical
equilibrium, so that we can compare its inferred dynamical and molecular
masses, it seems more likely that the line is 12CO(1-0) or 12CO(2-1) than any
higher-J transition. The higher-J levels would require the source to be quite
close to face-on in order to make the inferred dynamical mass larger than the
molecular mass. The line profile is consistent with a rectangular or double-
horned shape, suggesting it comes from a rotating disk with gas extending to
the flat part of the rotation curve, so the dynamical equilibrium assumption
is plausible by this measure.
The source is very similar in integrated line flux and width to the typical
3mm 12CO lines detected by ALMA in a blind survey of the Hubble Ultra-Deep
Field (HUDF; González-López et al., 2019; Walter et al., 2016). Its continuum
is brighter than is typical, as the brightest 3mm continuum source in the HUDF
has a flux density of 46 $\pm$ 7 $\mu$Jy whereas this one has 110 $\pm$ 10
$\mu$Jy. But overall, this source is compatible with being a similar object to
those. Most of them are identified as 12CO(2-1) at redshifts of 1.0 to 1.5,
based on associations with optical counterparts and occasionally detections of
higher-J lines in higher ALMA bands. In fact, all of the 12CO detections in
the HUDF have optical counterparts (González-López et al., 2019), and they are
typically 0.01 to 1 $\mu$Jy at an observed wavelength of 1 $\mu$m (Boogaard et
al., 2019). The ground-based optical data shown in Figure 15 cannot rule out
the faintest of those typical optical counterparts.
Figure 15: Contours show the line emission from the unidentified source;
levels are $(0.2,0.5,0.8)\times 3.12\times 10^{5}$ (Jy bm-1) Hz. The
background is the color composite of the MATLAS $u$, $g,$ and $i$ images (Duc
et al., 2015). Figure 16: 3mm spectrum of the unidentified source. Primary
beam corrections have been applied.
## Appendix B Possible evidence for an AGN-driven ionized outflow
Figure 17 shows a side-by-side comparison of the inner structure of NGC 7465,
in some indicators that might reveal a small AGN-driven ionized outflow.
Ferruit et al. (2000) presented narrowband HST [O III] and H$\alpha$ imaging
which showed enhanced [O III]/H$\alpha$ in a $r\lesssim 2$′′ region with a
southeast-northwest elongation, consistent with the structure in the ATLAS3D
[O III]/H$\beta$ data. This region of more strongly-ionized gas is expected to
trace enhanced shocks and/or harder radiation fields. The velocity of the
ionized gas shows a gradient along a similar kinematic axis of roughly $-$35°,
as indicated by the ellipse in Figure 17, and the velocity dispersion in [O
III] is also enhanced near the ends of the major axis of this ellipse.
However, the velocity resolution of the [O III] kinematic data is not good
enough to show any signatures of double-peaked line profiles.
The 12CO kinematic and photometric position angles in this region show that
the molecular gas predominantly traces a disk or a ring perpendicular to the
suggested outflow axis. Thus, enhanced [O III]/H$\beta$ ratios along that axis
might simply reflect lower opacities on that axis and higher opacities in the
dusty molecular disk, rather than actual outflow. The enhanced 12CO velocity
dispersion along the proposed outflow axis is probably then related to the
kinematics of the molecular disk rather than the outflow, and enhanced
12CO/13CO line ratios along the proposed outflow axis might reflect higher
temperatures or larger linewidths (Section 11.1). In short, the features
observed here could be caused by an ionized outflow but are not definitive
evidence; higher resolution spectrosocopy targeting the optical nebular lines
would be useful in testing for the existence of an outflow.
Figure 17: The inner structure of NGC 7465. The velocities and line ratios are
the same data previously shown in Figures 2, 3, and 6; the panels on the right
side also show the velocity dispersions in 12CO and [O III]. Beam sizes for
the ALMA data are shown as ellipses in the lower right corners. The cross
marks the location of the nuclear 3mm continuum source, and the dashed ellipse
(6′′ $\times$ 4′′, 850 $\times$ 570 pc) roughly indicates the region where
possible outflow signatures can be seen in the optical data.
## Appendix C Line fluxes in NGC 7465.
Table C presents all the line fluxes measured in the regions used for the
analysis of spatial variations in the molecular properties of NGC 7465.
=5.5cm
Table 4: Line fluxes in the defined regions of NGC 7465
Rgn | Dist. | Area | 12CO | 13CO | C18O | HCN | HCO+ | CN | CS | C2H | CH3OH | SiO | HNCO
---|---|---|---|---|---|---|---|---|---|---|---|---|---
ID | (′′) | ($\square$′′) | (Jy km s-1) | (Jy km s-1) | (Jy km s-1) | (Jy km s-1) | (Jy km s-1) | (Jy km s-1) | (Jy km s-1) | (Jy km s-1) | (Jy km s-1) | (Jy km s-1) | (Jy km s-1)
1 | 0.5 | 3.1 | 3.515 (0.040) | 0.096 (0.013) | $<$ 0.039 | 0.132 (0.012) | 0.303 (0.013) | 0.226 (0.006) | 0.041 (0.009) | 0.091 (0.014) | $<$ 0.035 | $<$ 0.040 | $<$ 0.037
2 | 2.3 | 8.8 | 6.877 (0.069) | 0.325 (0.023) | $<$ 0.061 | 0.128 (0.022) | 0.219 (0.018) | 0.122 (0.015) | 0.087 (0.018) | 0.104 (0.022) | $<$ 0.057 | $<$ 0.066 | $<$ 0.061
3 | 1.9 | 5.9 | 4.852 (0.062) | 0.186 (0.020) | $<$ 0.050 | 0.092 (0.019) | 0.222 (0.018) | 0.162 (0.008) | $<$ 0.041 | $<$ 0.065 | $<$ 0.057 | $<$ 0.053 | $<$ 0.058
4 | 3.4 | 7.4 | 3.787 (0.053) | 0.182 (0.018) | $<$ 0.054 | 0.045 (0.014) | 0.109 (0.015) | 0.058 (0.007) | $<$ 0.043 | 0.048 (0.016) | $<$ 0.042 | $<$ 0.047 | $<$ 0.050
5 | 2.9 | 57.2 | 30.888 (0.277) | 1.266 (0.093) | 0.311 (0.091) | 0.477 (0.070) | 1.122 (0.070) | 0.814 (0.035) | 0.337 (0.068) | 0.468 (0.086) | $<$ 0.260 | $<$ 0.234 | $<$ 0.222
6 | 5.8 | 10.0 | 2.163 (0.047) | 0.130 (0.014) | $<$ 0.041 | $<$ 0.041 | $<$ 0.034 | $<$ 0.018 | $<$ 0.034 | $<$ 0.036 | $<$ 0.027 | $<$ 0.039 | $<$ 0.032
7 | 7.9 | 15.6 | 4.312 (0.091) | 0.201 (0.029) | $<$ 0.082 | $<$ 0.057 | $<$ 0.067 | $<$ 0.037 | $<$ 0.053 | $<$ 0.069 | $<$ 0.060 | $<$ 0.064 | $<$ 0.073
8 | 5.3 | 3.1 | 0.598 (0.016) | 0.039 (0.005) | $<$ 0.013 | 0.019 (0.005) | 0.040 (0.005) | $<$ 0.007 | $<$ 0.013 | $<$ 0.014 | $<$ 0.015 | $<$ 0.017 | $<$ 0.018
9 | 10.0 | 759.8 | 80.792 (0.866) | 2.540 (0.279) | 0.845 (0.265) | $<$ 0.668 | 1.328 (0.212) | 0.843 (0.131) | $<$ 0.643 | $<$ 0.602 | $<$ 0.659 | $<$ 0.631 | $<$ 0.745
Note. — Regions are identified by number in Figure 7. The distance in column 2
refers to the typical distance between the nuclear continuum source and points
within the region, and it indicates the radial coordinate used for each region
in Figures 9 and 10. Region 9 is a much larger irregular area roughly 38′′ by
32′′ in diameter, encompassing virtually all of the detected 12CO emission in
these data, and its typical radial coordinate is arbitrarily set to 10′′ for
ease in plotting. Line ratios measured in K units can be calculated from these
table entries by multiplying the ratio in Jy units by the square of the ratio
of the lines’ wavelengths.
## References
* Alatalo et al. (2014) Alatalo, K., Nyland, K., Graves, G., et al. 2014, ApJ, 780, 186
* Alatalo et al. (2015) Alatalo, K., Crocker, A. F., Aalto, S., et al. 2015, MNRAS, 450, 3874
* Anderson et al. (2014) Anderson, C. N., Meier, D. S., Ott, J., et al. 2014, ApJ, 793, 37
* Andrews et al. (2018) Andrews, S. M., Huang, J., Pérez, L. M., et al. 2018, ApJ, 869, L41
* Astropy Collaboration et al. (2013) Astropy Collaboration, Robitaille, T. P., Tollerud, E. J., et al. 2013, A&A, 558, A33
* Astropy Collaboration et al. (2018) Astropy Collaboration, Price-Whelan, A. M., SipHocz, B. M., et al. 2018, aj, 156, 123
* Athanassoula (1992) Athanassoula, E. 1992, MNRAS, 259, 345
* Baan et al. (2008) Baan, W. A., Henkel, C., Loenen, A. F., Baudry, A., & Wiklind, T. 2008, A&A, 477, 747
* Baumgartner et al. (2013) Baumgartner, W. H., Tueller, J., Markwardt, C. B., et al. 2013, ApJS, 207, 19
* Bayet et al. (2009) Bayet, E., Aladro, R., Martín, S., Viti, S., & Martín-Pintado, J. 2009, ApJ, 707, 126
* Bayet et al. (2012) Bayet, E., Davis, T. A., Bell, T. A., & Viti, S. 2012, MNRAS, 424, 2646
* Bayet et al. (2013) Bayet, E., Bureau, M., Davis, T. A., et al. 2013, MNRAS, 432, 1742
* Boger & Sternberg (2005) Boger, G. I., & Sternberg, A. 2005, ApJ, 632, 302
* Boogaard et al. (2019) Boogaard, L. A., Decarli, R., González-López, J., et al. 2019, ApJ, 882, 140
* Braine et al. (2017) Braine, J., Shimajiri, Y., André, P., et al. 2017, A&A, 597, A44
* Brown & Wilson (2019) Brown, T., & Wilson, C. D. 2019, ApJ, 879, 17
* Bryant et al. (2019) Bryant, J. J., Croom, S. M., van de Sande, J., et al. 2019, MNRAS, 483, 458
* Cao et al. (2017) Cao, Y., Wong, T., Xue, R., et al. 2017, ApJ, 847, 33
* Cappellari et al. (2011) Cappellari, M., Emsellem, E., Krajnović, D., et al. 2011, MNRAS, 413, 813
* Cappellari et al. (2013) Cappellari, M., Scott, N., Alatalo, K., et al. 2013, MNRAS, 432, 1709
* Carilli & Walter (2013) Carilli, C. L., & Walter, F. 2013, ARA&A, 51, 105
* Cicone et al. (2020) Cicone, C., Maiolino, R., Aalto, S., Muller, S., & Feruglio, C. 2020, A&A, 633, A163
* Cluver et al. (2014) Cluver, M. E., Jarrett, T. H., Hopkins, A. M., et al. 2014, ApJ, 782, 90
* Cormier et al. (2018) Cormier, D., Bigiel, F., Jiménez-Donaire, M. J., et al. 2018, MNRAS, 475, 3909
* Crocker et al. (2012) Crocker, A., Krips, M., Bureau, M., et al. 2012, MNRAS, 421, 1298
* Danielson et al. (2013) Danielson, A. L. R., Swinbank, A. M., Smail, I., et al. 2013, MNRAS, 436, 2793
* Davis (2014) Davis, T. A. 2014, MNRAS, 445, 2378
* Davis et al. (2013) Davis, T. A., Bayet, E., Crocker, A., Topal, S., & Bureau, M. 2013, MNRAS, 433, 1659
* Davis & Bureau (2016) Davis, T. A., & Bureau, M. 2016, MNRAS, 457, 272
* Davis et al. (2019) Davis, T. A., Greene, J. E., Ma, C.-P., et al. 2019, MNRAS, 486, 1404
* Davis & Young (2019) Davis, T. A., & Young, L. M. 2019, MNRAS, 489, L108
* Davis et al. (2011) Davis, T. A., Alatalo, K., Sarzi, M., et al. 2011, MNRAS, 417, 882
* Davis et al. (2014) Davis, T. A., Young, L. M., Crocker, A. F., et al. 2014, MNRAS, 444, 3427
* Davis et al. (2015) Davis, T. A., Rowlands, K., Allison, J. R., et al. 2015, MNRAS, 449, 3503
* de Vaucouleurs et al. (1991) de Vaucouleurs, G., de Vaucouleurs, A., Corwin, Herold G., J., et al. 1991, Third Reference Catalogue of Bright Galaxies
* Domínguez-Fernández et al. (2020) Domínguez-Fernández, A. J., Alonso-Herrero, A., García-Burillo, S., et al. 2020, arXiv e-prints, arXiv:2003.05663
* Duc et al. (2015) Duc, P.-A., Cuillandre, J.-C., Karabal, E., et al. 2015, MNRAS, 446, 120
* Emonts et al. (2019) Emonts, B., Raba, R., Moellenbrock, G., et al. 2019, arXiv e-prints, arXiv:1912.09437
* Ferruit et al. (2000) Ferruit, P., Wilson, A. S., & Mulchaey, J. 2000, ApJS, 128, 139
* Gallagher et al. (2018) Gallagher, M. J., Leroy, A. K., Bigiel, F., et al. 2018, ApJ, 858, 90
* García-Burillo et al. (2010) García-Burillo, S., Usero, A., Fuente, A., et al. 2010, A&A, 519, A2
* García-Burillo et al. (2017) García-Burillo, S., Viti, S., Combes, F., et al. 2017, A&A, 608, A56
* Gonçalves et al. (1999) Gonçalves, A. C., Véron-Cetty, M. P., & Véron, P. 1999, A&AS, 135, 437
* González-López et al. (2019) González-López, J., Decarli, R., Pavesi, R., et al. 2019, arXiv e-prints, arXiv:1903.09161
* Griffith et al. (2019) Griffith, E., Martini, P., & Conroy, C. 2019, MNRAS, 484, 562
* Harada et al. (2018) Harada, N., Sakamoto, K., Martín, S., et al. 2018, ApJ, 855, 49
* Heikkilä et al. (1999) Heikkilä, A., Johansson, L. E. B., & Olofsson, H. 1999, A&A, 344, 817
* Henkel et al. (2014) Henkel, C., Asiri, H., Ao, Y., et al. 2014, A&A, 565, A3
* Israel (2020) Israel, F. P. 2020, A&A, 635, A131
* Israel et al. (2003) Israel, F. P., Johansson, L. E. B., Rubio, M., et al. 2003, A&A, 406, 817
* Jiménez-Donaire et al. (2017) Jiménez-Donaire, M. J., Cormier, D., Bigiel, F., et al. 2017, ApJ, 836, L29
* Jiménez-Donaire et al. (2019) Jiménez-Donaire, M. J., Bigiel, F., Leroy, A. K., et al. 2019, ApJ, 880, 127
* Jin et al. (2016) Jin, Y., Chen, Y., Shi, Y., et al. 2016, MNRAS, 463, 913
* Johansson et al. (1994) Johansson, L. E. B., Olofsson, H., Hjalmarson, A., Gredel, R., & Black, J. H. 1994, A&A, 291, 89
* Kaviraj et al. (2012) Kaviraj, S., Ting, Y.-S., Bureau, M., et al. 2012, MNRAS, 423, 49
* Kennicutt et al. (2011) Kennicutt, R. C., Calzetti, D., Aniano, G., et al. 2011, PASP, 123, 1347
* Kepley et al. (2018) Kepley, A. A., Bittle, L., Leroy, A. K., et al. 2018, ApJ, 862, 120
* Kohno et al. (2001) Kohno, K., Matsushita, S., Vila-Vilaró, B., et al. 2001, in Astronomical Society of the Pacific Conference Series, Vol. 249, The Central Kiloparsec of Starbursts and AGN: The La Palma Connection, ed. J. H. Knapen, J. E. Beckman, I. Shlosman, & T. J. Mahoney, 672
* Krajnović et al. (2011) Krajnović, D., Emsellem, E., Cappellari, M., et al. 2011, MNRAS, 414, 2923
* Krajnović et al. (2020) Krajnović, D., Ural, U., Kuntschner, H., et al. 2020, A&A, 635, A129
* Krips et al. (2008) Krips, M., Neri, R., García-Burillo, S., et al. 2008, ApJ, 677, 262
* Lablanche et al. (2012) Lablanche, P.-Y., Cappellari, M., Emsellem, E., et al. 2012, MNRAS, 424, 1495
* Langer & Penzias (1990) Langer, W. D., & Penzias, A. A. 1990, ApJ, 357, 477
* Lapham et al. (2017) Lapham, R. C., Young, L. M., & Crocker, A. 2017, ApJ, 840, 51
* Leroy et al. (2009) Leroy, A. K., Walter, F., Bigiel, F., et al. 2009, AJ, 137, 4670
* Leroy et al. (2017) Leroy, A. K., Usero, A., Schruba, A., et al. 2017, ApJ, 835, 217
* Li & Seaquist (1994) Li, J. G., & Seaquist, E. R. 1994, AJ, 107, 1953
* Lintott et al. (2005) Lintott, C. J., Viti, S., Rawlings, J. M. C., et al. 2005, ApJ, 620, 795
* Martín et al. (2019) Martín, S., Muller, S., Henkel, C., et al. 2019, A&A, 624, A125
* Martín et al. (2014) Martín, S., Verdes-Montenegro, L., Aladro, R., et al. 2014, A&A, 563, L6
* McCoy et al. (2017) McCoy, M., Ott, J., Meier, D. S., et al. 2017, ApJ, 851, 76
* McDermid et al. (2015) McDermid, R. M., Alatalo, K., Blitz, L., et al. 2015, MNRAS, 448, 3484
* Meier & Turner (2004) Meier, D. S., & Turner, J. L. 2004, AJ, 127, 2069
* Meier & Turner (2005) —. 2005, ApJ, 618, 259
* Meier & Turner (2012) —. 2012, ApJ, 755, 104
* Meier et al. (2008) Meier, D. S., Turner, J. L., & Hurt, R. L. 2008, ApJ, 675, 281
* Meier et al. (2015) Meier, D. S., Walter, F., Bolatto, A. D., et al. 2015, ApJ, 801, 63
* Meijerink et al. (2007) Meijerink, R., Spaans, M., & Israel, F. P. 2007, A&A, 461, 793
* Méndez-Hernández et al. (2020) Méndez-Hernández, H., Ibar, E., Knudsen, K. K., et al. 2020, MNRAS, 497, 2771
* Mittal et al. (2011) Mittal, R., O’Dea, C. P., Ferland, G., et al. 2011, MNRAS, 418, 2386
* Mittal et al. (2012) Mittal, R., Oonk, J. B. R., Ferland, G. J., et al. 2012, MNRAS, 426, 2957
* Moustakas & Kennicutt (2006) Moustakas, J., & Kennicutt, Robert C., J. 2006, ApJS, 164, 81
* Moustakas et al. (2010) Moustakas, J., Kennicutt, Robert C., J., Tremonti, C. A., et al. 2010, ApJS, 190, 233
* Nyland et al. (2016) Nyland, K., Young, L. M., Wrobel, J. M., et al. 2016, MNRAS, 458, 2221
* Nyland et al. (2017) —. 2017, MNRAS, 464, 1029
* Paglione et al. (2001) Paglione, T. A. D., Wall, W. F., Young, J. S., et al. 2001, ApJS, 135, 183
* Privon et al. (2015) Privon, G. C., Herrero-Illana, R., Evans, A. S., et al. 2015, ApJ, 814, 39
* Rau & Cornwell (2011) Rau, U., & Cornwell, T. J. 2011, A&A, 532, A71
* Remijan et al. (2019) Remijan, A., Biggs, A., Cortes, P., et al. 2019, ALMA Technical Handbook, ALMA Doc. 7.3 version 1.1, , , ISBN 978-3-923524-66-2
* Rémy-Ruyer et al. (2015) Rémy-Ruyer, A., Madden, S. C., Galliano, F., et al. 2015, A&A, 582, A121
* Romano et al. (2019) Romano, D., Matteucci, F., Zhang, Z.-Y., Ivison, R. J., & Ventura, P. 2019, MNRAS, 490, 2838
* Russell et al. (2019) Russell, H. R., McNamara, B. R., Fabian, A. C., et al. 2019, MNRAS, 490, 3025
* Sarzi et al. (2006) Sarzi, M., Falcón-Barroso, J., Davies, R. L., et al. 2006, MNRAS, 366, 1151
* Sarzi et al. (2010) Sarzi, M., Shields, J. C., Schawinski, K., et al. 2010, MNRAS, 402, 2187
* Sarzi et al. (2016) Sarzi, M., Ledo, H. R., Coccato, L., et al. 2016, MNRAS, 457, 1804
* Seale et al. (2012) Seale, J. P., Looney, L. W., Wong, T., et al. 2012, ApJ, 751, 42
* Serra et al. (2012) Serra, P., Oosterloo, T., Morganti, R., et al. 2012, MNRAS, 422, 1835
* Serra et al. (2014) Serra, P., Oser, L., Krajnović, D., et al. 2014, MNRAS, 444, 3388
* Sliwa & Downes (2017) Sliwa, K., & Downes, D. 2017, A&A, 604, A2
* Sliwa et al. (2017) Sliwa, K., Wilson, C. D., Aalto, S., & Privon, G. C. 2017, ApJ, 840, L11
* Smith et al. (2017) Smith, J. D. T., Croxall, K., Draine, B., et al. 2017, ApJ, 834, 5
* Tan et al. (2011) Tan, Q.-H., Gao, Y., Zhang, Z.-Y., & Xia, X.-Y. 2011, Research in Astronomy and Astrophysics, 11, 787
* Tang et al. (2019) Tang, X. D., Henkel, C., Menten, K. M., et al. 2019, A&A, 629, A6
* Taniguchi & Ohyama (1998) Taniguchi, Y., & Ohyama, Y. 1998, ApJ, 507, L121
* Topal et al. (2016) Topal, S., Bureau, M., Davis, T. A., et al. 2016, MNRAS, 463, 4121
* Ueda et al. (2017) Ueda, J., Watanabe, Y., Iono, D., et al. 2017, PASJ, 69, 6
* van der Tak et al. (2007) van der Tak, F. F. S., Black, J. H., Schöier, F. L., Jansen, D. J., & van Dishoeck, E. F. 2007, A&A, 468, 627
* van Zee & Haynes (2006) van Zee, L., & Haynes, M. P. 2006, ApJ, 636, 214
* Viti et al. (2020) Viti, S., Fontani, F., & Jiménez-Serra, I. 2020, MNRAS, 497, 4333
* Viti et al. (2014) Viti, S., García-Burillo, S., Fuente, A., et al. 2014, A&A, 570, A28
* Walter et al. (2016) Walter, F., Decarli, R., Aravena, M., et al. 2016, ApJ, 833, 67
* Wang et al. (2009) Wang, M., Chin, Y. N., Henkel, C., Whiteoak, J. B., & Cunningham, M. 2009, ApJ, 690, 580
* Werner et al. (2014) Werner, N., Oonk, J. B. R., Sun, M., et al. 2014, MNRAS, 439, 2291
* Wilson (2018) Wilson, C. D. 2018, MNRAS, 477, 2926
* Wilson et al. (2013) Wilson, C. D., Cridland, A., Foyle, K., et al. 2013, ApJ, 776, L30
* Young et al. (2011) Young, L. M., Bureau, M., Davis, T. A., et al. 2011, MNRAS, 414, 940
* Zhang et al. (2018) Zhang, Z.-Y., Romano, D., Ivison, R. J., Papadopoulos, P. P., & Matteucci, F. 2018, Nature, 558, 260
|
# Transfer-tensor description of memory effects in open-system dynamics and
multi-time statistics
Stefano Gherardini<EMAIL_ADDRESS>Dipartimento di Fisica e
Astronomia, Università degli Studi di Firenze, via G. Sansone 1, I-50019 Sesto
Fiorentino, Italy LENS, CNR-INO, and QSTAR, via N. Carrara 1, I-50019 Sesto
Fiorentino, Italy INFN Sezione di Firenze, via G. Sansone 1, I-50019 Sesto
Fiorentino, Italy Andrea Smirne<EMAIL_ADDRESS>Institute of
Theoretical Physics, and IQST, Universität Ulm, Albert-Einstein-Allee 11,
89069 Ulm, Germany Dipartimento di Fisica “Aldo Pontremoli", Università degli
Studi di Milano, via Celoria 16, 20133 Milan, Italy Istituto Nazionale di
Fisica Nucleare, Sezione di Milano, via Celoria 16, 20133 Milan, Italy Susana
Huelga<EMAIL_ADDRESS>Institute of Theoretical Physics, and IQST,
Universität Ulm, Albert-Einstein-Allee 11, 89069 Ulm, Germany Filippo Caruso
<EMAIL_ADDRESS>Dipartimento di Fisica e Astronomia, Università degli
Studi di Firenze, via G. Sansone 1, I-50019 Sesto Fiorentino, Italy LENS,
CNR-INO, and QSTAR, via N. Carrara 1, I-50019 Sesto Fiorentino, Italy
###### Abstract
The non-Markovianity of an arbitrary open quantum system is analyzed in
reference to the multi-time statistics given by its monitoring at discrete
times. On the one hand, we exploit the hierarchy of inhomogeneous transfer
tensors, which provides us with relevant information about the role of
correlations between the system and the environment in the dynamics. The
connection between the transfer-tensor hierarchy and the CP-divisibility
property is then investigated, by showing to what extent quantum Markovianity
can be linked to a description of the open-system dynamics by means of the
composition of 1-step transfer tensors only. On the other hand, we introduce
the set of stochastic transfer tensor transformations associated with local
measurements on the open system at different times and conditioned on the
measurement outcomes. The use of the transfer-tensor formalism accounts for
different kinds of memory effects in the multi-time statistics and allows us
to compare them on a similar footing with the memory effects present in non-
monitored non-Markovian dynamics, as we illustrate on a spin-boson case study.
## I Introduction
In a dynamical quantum system the interaction with the external environment
typically leads to non-Markovian behaviours, which, broadly speaking, are
associated with the presence of memory effects in the system evolution [1, 2,
3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22].
Although by now many distinct definitions of quantum non-Markovianity have
been introduced, two main pathways can be identified [23].
In one case, the focus is on the evolution of the open-system state at
different times, as fixed, e.g., by the dynamical maps or the master equations
that characterize the open-system dynamics [24, 25], while in the other case
the focus is on the statistics associated with a sequence of local
measurements (or other active interventions) performed on the open system at
subsequent times. These two approaches to non-Markovianity are indeed
inherently different, as the former addresses the predictions related with
observables at a single time, while the latter concerns the multi-time
statistics. In addition to such a difference, which already appears at the
classical level [26, 27], the quantum nature of the system at hand implies a
further, fundamental distinction between the single-time and the multi-time
notions of non-Markovianity. Contrary to what happens classically, in quantum
systems it is generally not possible, even in principle, to access the
information associated with sequential measurements without disturbing the
system that is being measured. Indeed, quantum measurements alter the current
state of the system, as well as its correlations with the environment, which
will later result in a modified evolution [7, 9, 29, 32, 28, 30, 34, 33, 31,
35]. As a consequence, the presence of intermediate measurements can bring
along specific forms of memory, which combine in a non-trivial way with the
memory strictly due to the system-environment interaction.
In this paper, we investigate the relation between the memory effects
appearing in an open quantum system dynamics and those associated with the
multi-time statistics due to sequential measurements, by means of the transfer
tensor (TT) method [36, 37, 38, 39]. We will show that the latter, which was
introduced to treat efficiently the long-time dynamics of open quantum
systems, also allows one to treat memory effects on the dynamics and the
multi-time statistics on a similar footing. First, we prove to what extent the
divisibility property of the dynamical maps, which is the defining property of
quantum non-Markovianity according to [7, 4], is linked to the hierarchy of
TTs also in the inhomogeneous case, i.e., beyond the time-translational
invariant regime explored in the original paper [36]. Then, we extend the
definition of the TTs to the situation where the open system is measured at
subsequent times, via the corresponding conditional stochastic dynamics. This
allows us to identify different forms of memory in the multi-time statistics,
related with the interplay between the correlations and environmental-state
transformations induced by, respectively, the interaction between the open
system and the environment and the sequential measurements. After introducing
proper quantifiers related with the $L_{2}$-norm of the multi-step (i.e., step
greater than $1$) TTs, we compare in a case study the relevance of such memory
effects for different kinds of measurements, including the case without any
intermediate measurement.
The rest of the paper is organized as follows. In Sec.II, we recall the main
features of the TTs formalism that are relevant for our analysis, focusing in
particular on the recursive construction of the TTs. In Sec.III, we prove that
if only multi-step TTs are different from zero the open-system dynamics is
divisible, while the converse statement does not hold, as we show by means of
an explicit example. In Sec.IV, we first generalize the TTs formalism to the
case of multi-time measurements, by introducing a family of TTs conditioned on
the sequence of measurement outcomes, and then we exploit it to characterize
the different kinds of memory effects associated with the multi-time
statistics. Finally, the conclusions and outlooks of our analysis are
discussed in Sec.V
## II Inhomogeneous transfer tensors hierarchy
We begin by briefly recalling the TT formalism for the dynamics of an open
quantum system [36] and the microscopic characterization of the TTs recently
derived in [38], which allows to build up the complete TTs hierarchy and to
connect it with the system-environment correlations. Such a construction will
be extended to an evolution conditioned on the occurrence of sequential
measurements in the following of the paper.
### II.1 General microscopic definition of TTs
Let us consider a quantum system $\mathbb{S}$ in interaction with an
environment $\mathbb{E}$. The dynamics of the composite system
$\mathbb{S}+\mathbb{E}$ is governed by a possibly time-dependent Hamiltonian
$H(t)=H_{0}(t)+H_{\rm int}$, where $H_{\rm int}$ takes into account the
interaction of the system with the environment, while
$H_{0}(t)=H_{\mathbb{S}}(t)+H_{\mathbb{E}}$ concerns the uncoupled time
evolution of $\mathbb{S}$ and $\mathbb{E}$. The interaction Hamiltonian
$H_{\rm int}$ is fixed but unknown, and $H_{\mathbb{S}}(t)$ can be time-
dependent. Assuming an initial product state between the open system and the
environment and a fixed initial state of $\mathbb{E}$, i.e.,
$\rho_{t_{0}}=\rho_{\mathbb{S},t_{0}}\otimes\rho_{\mathbb{E},t_{0}}$, the
dynamics of $\mathbb{S}$ can be described by means of the formalism of quantum
dynamical maps [24, 25, 8]. In fact, one can define a one parameter family of
completely positive, trace-preserving (CPTP) dynamical maps
$\left\\{\Phi(t,t_{0})\right\\}_{t\geq t_{0}}$, with
$\Phi(t,t_{0}):\mathcal{S}(\mathcal{H}_{\mathbb{S}})\rightarrow\mathcal{S}(\mathcal{H}_{\mathbb{S}})$,
where $\mathcal{S}(\mathcal{H}_{\mathbb{S}})$ denotes the sets of density
operators (non-negative operators with unit trace) acting on
$\mathcal{H}_{\mathbb{S}}$ and $\Phi(t,t_{0})$ is CPTP at any time $t$.
Hereafter, the action of the dynamical maps of the system will be taken into
account at the discrete time instants (not necessarily equally spaced)
$t_{k}$, $k=1,\ldots,m$, and we will use the notation
$\Phi(t_{k},t_{0})\equiv\Phi_{k}$ and
$\rho_{\mathbb{S},k}=\rho_{\mathbb{S}}(t_{k})=\Phi_{k}[\rho_{\mathbb{S},0}].$
(1)
This will make the comparison with the sequential-measurement scenario more
transparent and will further allow us to directly apply the TT method [36].
The latter relies in fact on a family of maps, the TTs $T_{k,j}$, which are
defined by the relation
$T_{k,0}=\Phi_{k}-\sum_{j=1}^{k-1}T_{k,j}\Phi_{j},$ (2)
equivalently expressed by
$\Phi_{k}=\sum_{j=0}^{k-1}T_{k,j}\Phi_{j}.$ (3)
In other terms, the state at the time instant $t_{k}$ is obtained by
propagating the states at $t_{j}$ according to the equation
$\rho_{\mathbb{S},k}=\sum_{j=0}^{k-1}T_{k,j}\rho_{\mathbb{S},j}$. We stress
that, at variance with the original formulation, we are not assuming time-
invariance; thus our analysis will also apply to time-dependent Hamiltonians
and non-stationary initial environmental states [36].
A useful expression for generic, possibly inhomogeneous TTs was recently
derived in [38]. Given the global unitary dynamics of the composite system
$\mathbb{S}+\mathbb{E}$ from $t_{s}$ to $t_{k}$, i.e.,
$\mathcal{U}_{t_{s}:t_{k}}[\rho]\equiv
U_{t_{s}:t_{k}}\rho\,U_{t_{s}:t_{k}}^{\dagger}$ with
$U_{t_{s}:t_{k}}\equiv\mathcal{T}\exp(-i\int_{t_{s}}^{t_{k}}H(\tau)d\tau)$
($\mathcal{T}$ is the time ordering operator), let us introduce the CPTP maps
$\Gamma_{k|k-n}:\mathcal{S}(\mathcal{H}_{\mathbb{S}})\rightarrow\mathcal{S}(\mathcal{H}_{\mathbb{S}})$
defined as
$\Gamma_{k|k-n}[\sigma_{\mathbb{S}}]\equiv{\rm
Tr}_{\mathbb{E}}\left[\mathcal{U}_{t_{k-n}:t_{k}}[\sigma_{\mathbb{S}}\otimes\sigma_{\mathbb{E},k-n}]\right],$
(4)
where ${\rm Tr}_{\mathbb{S}/\mathbb{E}}[\cdot]$ denotes the partial trace
w.r.t. the Hilbert space of $\mathbb{S}$ or $\mathbb{E}$,
$\sigma_{\mathbb{S}}$ denotes a density operator on $\mathcal{H}_{\mathbb{S}}$
and $\sigma_{\mathbb{E},k-n}$ is a time-dependent density operator on
$\mathcal{H}_{\mathbb{E}}$. Here, we will focus on the case where
$\sigma_{\mathbb{E},k-n}$ is the state of the environment at $t_{k-n}$. It can
then be proven that the $n-$step TT $T_{k,k-n}$ is given by the recursive
relation [38]
$T_{k,k-n}=\Gamma_{k|k-n}-\sum_{j=1}^{n-1}T_{k,k-j}\Gamma_{k-j|k-n}$ (5)
that allows one to reconstruct the whole hierarchy of TTs from the CPTP maps
$\Gamma_{k|k-n}$, as shown explicitly in [38]. In the next paragraph, we
repeat the construction for the lowest orders of the hierarchy, both for the
sake of illustration and since it will be useful in the following when we will
move to the stochastic case.
### II.2 Reconstruction of the TTs hierarchy
First, let us combine together the definition of the dynamical maps of the
system at the discrete time instants $t_{k}$, $k=1,\ldots,n$, i.e., Eq. (1),
and the recursive general expression of the TT transformation given by Eq.(3).
In this way, the latter can be written via the following recursive relation:
$\Phi_{k}=\underline{T}_{k}\underline{\Xi}_{k-1}[\rho_{\mathbb{S},0}],$ (6)
where
$\underline{T}_{k}\equiv\left(T_{k,0},T_{k,1},\ldots,T_{k,k-1}\right)$
and
$\underline{\Xi}_{k-1}[\rho_{\mathbb{S},0}]\equiv\begin{pmatrix}\mathbbm{1}_{\mathbb{S}}[\rho_{\mathbb{S},0}]\\\
\Phi_{1}=\underline{T}_{1}\underline{\Xi}_{0}[\rho_{\mathbb{S},0}]\\\
\vdots\\\
\Phi_{k-1}=\underline{T}_{k-1}\underline{\Xi}_{k-2}[\rho_{\mathbb{S},0}]\end{pmatrix},$
which are valid for $n\geq 1$
($\underline{T}_{0}=\underline{\Xi}_{-1}=\mathbbm{1}_{\mathbb{S}}[\rho_{\mathbb{S},0}]$,
identity operator on $\mathbb{S}$). For the sake of clarity, we show the first
terms of the recursive expansion of Eq. (6):
$\begin{cases}\Phi_{0}=\mathbbm{1}_{\mathbb{S}}[\rho_{\mathbb{S},0}]\\\
\Phi_{1}=T_{1,0}\Phi_{0}=T_{1,0}[\rho_{\mathbb{S},0}]\\\
\Phi_{2}=T_{2,0}\Phi_{0}+T_{2,1}\Phi_{1}=\left(T_{2,0}+T_{2,1}T_{1,0}\right)[\rho_{\mathbb{S},0}]\\\
\Phi_{3}=T_{3,0}\Phi_{0}+T_{3,1}\Phi_{1}+T_{3,2}\Phi_{2}\\\
\,\,\,\,\,\,\,\,\,=\left(T_{3,0}+T_{3,1}T_{1,0}+T_{3,2}T_{2,0}+T_{3,2}T_{2,1}T_{1,0}\right)[\rho_{\mathbb{S},0}].\\\
\end{cases}$ (7)
By combining Eqs. (1) and (6), we immediately find that
$T_{1,0}[\rho_{\mathbb{S},0}]=\Phi_{1}[\rho_{\mathbb{S},0}],$ (8)
so that $\rho_{\mathbb{S},1}=T_{1,0}[\rho_{\mathbb{S},0}]$. Then, by
continuing the recursion, one has
$\rho_{\mathbb{S},2}=T_{2,0}[\rho_{\mathbb{S},0}]+T_{2,1}[T_{1,0}[\rho_{\mathbb{S},0}]].$
Here, it is worth observing that $T_{2,1}[T_{1,0}[\rho_{\mathbb{S},0}]]$
returns the component of $\rho_{\mathbb{S},2}$ conditioned on finding the
system in the states $\rho_{\mathbb{S},0}$ and $T_{1,0}[\rho_{\mathbb{S},0}]$
at the time instants $t_{0}$ and $t_{1}$ and no system-environment
correlations were present. In fact, given the maps $\Gamma_{k|k-n}$ defined in
Eq. (4), the 1-step TTs $T_{k,k-1}$ satisfy
$T_{k,k-1}=\Gamma_{k|k-1}$ (9)
as a direct consequence of Eq. (5) (for the trivial case $n=1$). In other
terms, we have
$T_{2,0}[\rho_{\mathbb{S},0}]=\rho_{\mathbb{S},2}-\Gamma_{2|1}[\rho_{\mathbb{S},1}],$
(10)
i.e.,
$T_{2,0}=\Phi_{2}-\Gamma_{2|1}\Phi_{1}.$ (11)
For the sake of brevity, from here on we will remove the subscript
$\mathbb{S}$ from the notation; it will used (together with $\mathbb{E}$) only
if necessary.
Moving on to the level $k=3$ of the hierarchy, by applying Eq. (7) one has
that the reduced density operator of $\mathbb{S}$ at $t_{3}$ is equal to
$\displaystyle\rho_{3}$ $\displaystyle=$ $\displaystyle
T_{3,0}[\rho_{0}]+T_{3,1}[T_{1,0}[\rho_{0}]]+T_{3,2}[T_{2,0}[\rho_{0}]]$
$\displaystyle+$ $\displaystyle T_{3,2}[T_{2,1}[T_{1,0}[\rho_{0}]]].$
Using the identity in Eq. (9) for $T_{3,2}$ and $T_{2,1}$, as well as Eq.
(10), we have
$\begin{cases}T_{3,2}[T_{2,0}[\rho_{0}]]=\Gamma_{3|2}[\rho_{2}-\Gamma_{2|1}[\rho_{1}]]\\\
T_{3,2}[T_{2,1}[T_{1,0}[\rho_{0}]]]=\Gamma_{3|2}\Gamma_{2|1}[\rho_{1}].\end{cases}$
(12)
Moreover, $T_{3,1}$ can be expressed by exploiting Eq. (5) (for $n=2$) that
along with Eq. (9) gives
$T_{3,1}\equiv\Gamma_{3|1}-\Gamma_{3|2}\Gamma_{2|1}.$ (13)
Thus, $T_{3,1}$ can be written as a function of a term involving only one-step
$\Gamma$s, i.e., $\Gamma_{3|2}\Gamma_{2|1}$, and of the $2-$step
$\Gamma_{3|1}$. By combining together all terms of Eqs. (12) and (13), one has
$\rho_{3}=T_{3,0}[\rho_{0}]+\Gamma_{3|1}[\rho_{1}]+\Gamma_{3|2}[\rho_{2}]-\Gamma_{3|2}\Gamma_{2|1}[\rho_{1}],$
i.e.,
$T_{3,0}=\Phi_{3}-\Gamma_{3|2}\Phi_{2}+(\Gamma_{3|2}\Gamma_{2|1}-\Gamma_{3|1})\Phi_{1}.$
(14)
As mentioned, the procedure can be generalized to any $k$ via Eq. (5), so that
the whole hierarchy of TTs can be expressed in terms of the maps
$\Gamma_{k|k-n}$ [38] (indeed, $\Phi_{k}=\Gamma_{k|0}$). Importantly, this
construction reveals the influence of the correlations between the system and
the environment into the dynamics of the open quantum system. In fact, the
CPTP maps (4) generate a hierarchy of $1-$, $2-$,…,$n-$step TTs conditioned to
the fact that the system passes through product states at the different steps
taken into account. As we will show in the next section, this is naturally
linked to the presence of memory effects in the open-system dynamics.
## III Transfer tensors and Markovianity of the dynamics
Here, we derive a definite connection between a property of the TTs – the
possibility to build up the whole hierarchy only in terms of $1-$step TTs –
and the non-Markovianity of the dynamics. In particular, we rely on the
definition introduced in [4], which identifies Markovian dynamics as those
described by a family of CP-divisible dynamical maps. Adapting the original
definition in [4] to the case of a discrete set of times, then the following
definition can be stated:
The (discrete) dynamics $\left\\{\Phi_{k}\right\\}_{k=1,\ldots,m}$ is
Markovian when for any $k\geq j\geq 0$ there is a CPTP map $\mathcal{E}_{k,j}$
such that $\Phi_{k}=\mathcal{E}_{k,j}\Phi_{j}$.
Hence, if we assume that $T_{k,k-n}=0$ for $n\geq 2$, then by applying
recursively Eq. (3) we have
$\Phi_{k}=T_{k,k-1}T_{k-1,k-2}\cdots T_{j+1,j}\Phi_{j}$ (15)
for any $k\geq j\geq 0$. But then Eq. (9) implies
$\Phi_{k}=\Gamma_{k|k-1}\Gamma_{k-1|k-2}...\Gamma_{j+1|j}\Phi_{j}.$ (16)
Any conditional map $\Gamma_{k|j}$ is CPTP by construction, see Eq.(4), so
that we can conclude that any map $\Phi_{k}$ can be decomposed as
$\Phi_{k}=\mathcal{E}_{k,j}\Phi_{j}$, with $\mathcal{E}_{k,j}$ CPTP map:
_If only one-step TTs are different from zero, the resulting (discrete-time)
evolution is CP-divisible_.
We further note that if we restrict to the case of equally-spaced time
instants, $t_{k}=k\Delta$, and translational invariant TTs (i.e.,
$T_{k,k-n}=T_{n,0}$), then $T_{k,k-n}=0$ for $n\geq 2$ implies that the open-
system dynamics is not only Markovian, but also a semigroup, i.e.,
$\Phi_{k}=(\Phi_{1})^{k}$, thus recovering what shown in [36].
The previous result, besides linking a property of the hierarchy of TTs to the
Markovianity of the corresponding dynamics, provides us with an explicit
illustration of one of the physical meanings of such property. In fact, from
Eq.(16) we see that when one-step TTs are the only non-zero TTs, the dynamical
maps can be obtained by using the maps $\Gamma$ only. System-environment
correlations, despite being present as a consequence of the interaction term
within the unitary operators, will not affect the reduced dynamics of the open
system, so that at any time $t_{k-n}$ the actual global state can be
effectively replaced by a product state
$\rho_{\mathbb{S},k-n}\otimes\sigma_{\mathbb{E},k-n}$, see Eq.(4).
Experimentally, Eq. (16) can be validated by independently measuring the maps
$\\{\Phi_{k}\\}$ and the 1-step $\\{\Gamma_{k,k-1}\\}$, obtained by breaking
the system-environment correlations at the time instants $t_{k}$. In
particular, the former can be measured by means of quantum tomography
processes, while the latter, e.g., by preparing a fresh copy of the system in
the state $\sigma_{\mathbb{S},k}$ [13], which is known by virtue of the
previously reconstructed map $\Phi_{k}$.
The fact that only one-step TTs are non-zero is in general a stronger
requirement than the CP-divisibility of the dynamics, as follows from the
analysis of [23] (see also [40]) and the explicit example provided below.
Indeed, having only one-step TTs different from zero is only a _sufficient
criterion_ (and thus not necessary) for CP-divisible discrete-time evolutions.
In this regard, let us consider the simplest case, namely $m=2$, whereby we
recall that $m$ is the last (and greater) value of the index $k$ in the
sequence of time instants $t_{k}$. Hence, the only non-trivial requirement for
CP-divisibility is the existence of a CPTP map $\mathcal{E}_{2,1}$ obeying the
relation
$\mathcal{E}_{2,0}=\Phi_{2}=\mathcal{E}_{2,1}\Phi_{1}=\mathcal{E}_{2,1}T_{1,0}$
since by definition $T_{1,0}\equiv\Phi_{1}$. Now, according to the definition
of the TT transformation, $\Phi_{2}$ is also equal to
$\Phi_{2}=T_{2,0}+T_{2,1}T_{1,0}.$
Therefore, assuming that the inverse of $T_{1,0}$ (i.e., $(T_{1,0})^{-1}$)
does exist, one has
$\mathcal{E}_{2,1}=T_{2,0}\left(T_{1,0}\right)^{-1}+T_{2,1}.$ (17)
What we will show is that $\mathcal{E}_{2,1}$ can be CPTP also if $T_{2,0}\neq
0$ and even if $\left(T_{1,0}\right)^{-1}$ is not be generally CPTP. In fact,
consider the reduced dynamics of a two-level system, fixed by the unitary
dynamics given by the global Hamiltonian
$H=\sigma_{z}+H_{\mathbb{E}}+\sigma_{z}\otimes B$, where the free
environmental Hamiltonian $H_{\mathbb{E}}$ and the interaction term commute:
$[H_{\mathbb{E}},B]=0$. Now, also assume that the global state at the initial
time $t_{0}=0$ is $\rho(0)=\rho_{\mathbb{S}}(0)\otimes\rho_{\mathbb{E}}(0)$,
where the environmental state is stationary with respect to the free dynamics
such that $[H_{\mathbb{E}},\rho_{\mathbb{E}}]=0$. As a consequence, on the one
hand, the reduced dynamics of the open system is a pure dephasing, namely
$\rho_{11}(t)=\rho_{11}(0)$ and $\rho_{10}(t)=k(t)\rho_{10}(0)$, where
$\rho_{ij}(t)=\langle i|\rho_{\mathbb{S}}(t)|j\rangle$ ($\left\\{\lvert
1\rangle,\lvert 0\rangle\right\\}$ are the eigenvectors of $\sigma_{z}$) and
$k(t)=e^{i\omega t}\mbox{tr}_{\mathbb{E}}[e^{2iBt}\rho_{\mathbb{E}}]$. On the
other hand, the state of the environment will not evolve in time, also when
interacting with the system, i.e.,
$\rho_{\mathbb{E}}(t)=\rho_{\mathbb{E}}(0)$. For any two times $t_{2}\geq
t_{1}\geq 0$, we will then have
$T_{2,1}=\Gamma_{2|1}=\Phi_{1},$
where we used Eq.(9) and Eq.(4), respectively. As a result (see also Eq.(11)),
$T_{2,0}=\Phi_{2}-\Phi^{2}_{1}.$
Specifically, using the matrix representation in the basis $\left\\{\lvert
1\rangle,\lvert 0\rangle\right\\}$, for a generic state $\rho_{\mathbb{S}}$ we
have
$\mathcal{E}_{2,1}[\rho_{\mathbb{S}}]=\begin{pmatrix}\rho_{11}&\frac{k(t_{2})}{k(t_{1})}\rho_{10}\\\
\frac{k^{*}(t_{2})}{k^{*}(t_{1})}\rho_{01}&\rho_{00}\end{pmatrix},$
which is positive definite for any $\rho_{\mathbb{S}}$ if and only if
$|k(t_{2})|\leq|k(t_{1})|$. Note that this implies that $\mathcal{E}_{2,1}$ is
also CPTP under the same condition, since for pure dephasing positivity and
complete positivity coincide. On the other hand,
$T_{2,0}[\rho_{\mathbb{S}}]=\begin{pmatrix}0&(k(t_{2})-k(t_{1})^{2})\rho_{10}\\\
(k^{*}(t_{2})-k^{*}(t_{1})^{2})\rho_{01}&0\end{pmatrix}$
that is equal to 0 for any $\rho_{\mathbb{S}}$ if and only if
$k(t_{2})=k(t_{1})^{2}$. The latter condition implies but is not implied by
the former (indeed, $|k(t)|\leq 1$), which proves our claim:
$\mathcal{E}_{2,1}$ can be CPTP, and hence the dynamics CP-divisible, even
though the $2$-step transfer tensor $T_{2,0}$ is different from zero.
## IV Transfer tensors and multi-time statistics
In this paragraph, we move on to the second part of our investigation, where
we consider an open system that, besides interacting with the environment, is
measured at some discrete instants of time. We will first introduce a proper
counterpart of the transfer tensors in such dynamical regime, and will then
show how it can be used to account for the memory present in the multi-time
statistics. Before that, let us specify the framework we refer to.
Assume that the open quantum system $\mathbb{S}$ is monitored at the same $m$
instants of time where the discrete dynamics has been evaluated so far. In
particular, we take a sequence of quantum measurements locally performed on
$\mathbb{S}$ according to the observables $\mathcal{O}_{k}\equiv
F_{\theta_{k}}\otimes I_{\mathbb{E}}$; $\\{\theta_{k}\\}$ is the set of the
possible measurement outcomes, and $\\{F_{\theta_{k}}\\}$ denotes the set of
positive semi-definite operators on $\mathcal{H}_{\mathbb{S}}$ satisfying the
relation $\sum_{\theta_{k}}F_{\theta_{k}}=\mathbbm{1}_{\mathbb{S}}$ $\forall
k$. The probability that the outcome $\theta_{k}$ associated with the
measurement operator $F_{\theta_{k}}$ occurs is equal to ${\rm
Tr}[\rho_{\mathbb{S},k}F_{\theta_{k}}]$, while the post-measurement state of
$\mathbb{S}$ equals to
$\widehat{\rho}_{\mathbb{S},k}\equiv\mathcal{M}_{\theta_{k}}[\rho_{\mathbb{S},k}]/{\rm
Tr}[\mathcal{M}_{\theta_{k}}[\rho_{\mathbb{S},k}]]$, where we introduced the
super-operator $\mathcal{M}_{\theta_{k}}[\rho_{\mathbb{S},k}]\equiv
M_{\theta_{k}}\rho_{\mathbb{S},k}M^{\dagger}_{\theta_{k}}$ and
$M_{\theta_{k}}$ fulfills the identity
$F_{\theta_{k}}=M^{\dagger}_{\theta_{k}}M_{\theta_{k}}$. Note that
$\rho_{\mathbb{S},k}$ ($\widehat{\rho}_{\mathbb{S},k}$) denotes the state of
the system before (after) the measurement at the time step $k$.
Importantly, a measurement of the open system affects not only its current
state, but also its future evolution as a consequence of the change in the
correlations between the system and the environment due to the measurement
itself [9]. The very notion of dynamics of the open system becomes more
subtle, since it cannot be clearly separated from the results of the
sequential measurements. It is thus useful to introduce the notion of
conditional dynamics, whereby the system admits a different dynamics for each
sequence of measurement outcomes. Explicitly, the conditional dynamics of the
system (for a given global evolution $\mathcal{U}_{t_{s}:t_{k}}$ and initial
environmental state $\rho_{\mathbb{E},0}$) is fixed by the CP map
$\displaystyle\widetilde{\Phi}^{\underline{\theta},\underline{t}}_{k}[\rho_{\mathbb{S},0}]$
$\displaystyle\equiv$ $\displaystyle{\rm
Tr}_{\mathbb{E}}\left[\mathcal{U}_{t_{k-1}:t_{k}}(\mathcal{M}_{\theta_{k-1}}\otimes\mathbbm{1}_{\mathbb{E}})\cdots\right.$
(18) $\displaystyle\cdots$
$\displaystyle\left.(\mathcal{M}_{\theta_{1}}\otimes\mathbbm{1}_{\mathbb{E}})\,\mathcal{U}_{t_{0}:t_{1}}[\rho_{\mathbb{S},0}\otimes\rho_{\mathbb{E},0}]\right],$
depending on the time instants $\underline{t}\equiv(t_{1},\ldots,t_{k})$ as
well as on the measurement outcomes
$\underline{\theta}\equiv(\theta_{1},\ldots,\theta_{k-1})$. In other terms,
$\widetilde{\Phi}^{\underline{\theta},\underline{t}}_{k}$ has to be understood
as a stochastic map, so that we could effectively define different
trajectories in the set of CP maps, each of them associated with a different
sequence of measurement outcomes. In general
$\widetilde{\Phi}^{\underline{\theta},\underline{t}}_{k}$ is not trace
preserving.
The joint probability distributions to get the measurement outcomes
$\theta_{1},\theta_{2},\ldots,\theta_{k}$ at the time instants
$t_{1},t_{2},\ldots,t_{k}$ is directly linked to the stochastic map
$\widetilde{\Phi}^{\underline{\theta},\underline{t}}_{k}$ by the following
relation:
$q_{k}\equiv{\rm Prob}(\theta_{k},t_{k};\ldots;\theta_{1},t_{1})={\rm
Tr}[\mathcal{M}_{\theta_{k}}\widetilde{\Phi}^{\underline{\theta},\underline{t}}_{k}[\rho_{\mathbb{S},0}]].$
(19)
These quantities define the multi-time statistics associated to sequential
measurements at different times and, as recalled in the introduction, suitable
notions of Markovianity can be attributed to them [29, 23, 15, 13, 14, 28, 30,
34, 31].
### IV.1 Conditional transfer tensors
As anticipated, in order to characterize the conditional dynamics originated
by different sequences of outcomes, we can introduce a stochastic version of
the TTs. Let us denote them as
$\widetilde{T}^{\underline{\theta},\underline{t}}_{k,j}$, with
$k,j=1,\ldots,m$, explicitly pointing out their dependence on the instants and
outcomes of the repeated measurements.
The basic idea is to express the definition of the conditional dynamical map
in a recursive fashion. As first step, note that the global state after the
first measurement is proportional (via the normalization factor) to
$\widehat{\rho}_{1}\sim(M_{\theta_{1}}\otimes
I_{\mathbb{E}})\mathcal{U}_{t_{0}:t_{1}}[\rho_{0}](M_{\theta_{1}}^{\dagger}\otimes
I_{\mathbb{E}})\equiv\mathcal{V}_{t_{0}:t_{1}}[\rho_{0}]$ (20)
that also provides the formal definition of the map $\mathcal{V}[\rho]$.
Accordingly, the stochastic quantum map of $\mathbb{S}$ at time $t_{1}$, i.e.,
$\widetilde{\Phi}_{1}$, can be implicitly linked to
$\mathcal{V}_{t_{0}:t_{1}}[\rho_{0}]$ through the relation
$\mathcal{M}_{\theta_{1}}\widetilde{\Phi}_{1}[\rho_{\mathbb{S},0}]\equiv{\rm
Tr}_{\mathbb{E}}(\mathcal{V}_{t_{0}:t_{1}}[\rho_{0}]).$ (21)
Analogously, the $k-$th dynamical map of $\mathbb{S}$ after a sequence of
quantum measurements is defined by
$\displaystyle\widehat{\rho}_{k}$ $\displaystyle\sim$
$\displaystyle(M_{\theta_{k}}\otimes
I_{\mathbb{E}})\mathcal{U}_{t_{k-1}:t_{k}}[\widehat{\rho}_{k-1}](M_{\theta_{k}}^{\dagger}\otimes
I_{\mathbb{E}})$ (22) $\displaystyle\equiv$
$\displaystyle\mathcal{V}_{t_{k-1}:t_{k}}[\widehat{\rho}_{k-1}]$
with
$\mathcal{M}_{\theta_{k}}\widetilde{\Phi}_{k}[\rho_{\mathbb{S},0}]={\rm
Tr}_{\mathbb{E}}(\mathcal{V}_{t_{k-1}:t_{k}}[\widehat{\rho}_{k-1}])$ (23)
and $\widehat{\rho}_{0}\equiv\rho_{0}$. For the sake of clarity, we recall
that, by definition, the CP maps $\widetilde{\Phi}_{k}$ are obtained by
tracing the state of the composite system $\mathbb{S}+\mathbb{E}$ (w.r.t. the
environment $\mathbb{E}$) just after applying the evolution superoperator
$\mathcal{U}_{t_{k-1}:t_{k}}$. As done with the definition of transfer tensors
linking together the quantum maps $\Phi_{k}$ of the system at time instants
$t_{k}$, $k=1,\ldots,m$, we can thus introduce the stochastic transformations
$\widetilde{T}^{\underline{\theta},\underline{t}}_{k,j}$ that relate the
conditional dynamics of $\mathbb{S}$ after each measurement of the sequence.
In particular, Eq. (23) has the same structure as the not monitored dynamics,
with $\Phi_{k}$ replaced by $\mathcal{M}_{\theta_{k}}\widetilde{\Phi}_{k}$.
Hence, we define the stochastic TTs via
$\widetilde{\Phi}^{\underline{\theta},\underline{t}}_{k}=\sum_{j=0}^{k-1}\widetilde{T}^{\underline{\theta},\underline{t}}_{k,j}\mathcal{M}_{\theta_{j}}\widetilde{\Phi}^{\underline{\theta},\underline{t}}_{j}\,.$
(24)
Note that if there are no measurements, $\mathcal{M}_{\theta_{k}}=\mathbbm{1}$
and $\widetilde{\Phi}^{\underline{\theta},\underline{t}}_{k}=\Phi_{k}$ (see
Eq. (18)) with the result that the TTs
$\widetilde{T}^{\underline{\theta},\underline{t}}_{k,j}$ are no longer
conditional objects and can be identified with the $T_{k,j}$ since Eq. (24)
reduces to Eq. (3). From now on, we will drop the label
$(\cdot)^{\underline{\theta},\underline{t}}$, which is implied in all the
expressions with a tilde.
As before, by expanding Eq. (24) we can write the expression for the $k-$th
dynamical map of $\mathbb{S}$ as a function of $\rho_{\mathbb{S},0}$, i.e.,
$\widetilde{\Phi}_{k}=\underline{\widetilde{T}}_{k}\underline{\mathcal{M}}_{\theta_{k-1}}\underline{\Upsilon}_{k-1}[\rho_{\mathbb{S},0}],$
(25)
where
$\displaystyle\underline{\widetilde{T}}_{k}$ $\displaystyle\equiv$
$\displaystyle\left(\widetilde{T}_{k,0},\widetilde{T}_{k,1},\ldots,\widetilde{T}_{k,k-1}\right)$
$\displaystyle\underline{\mathcal{M}}_{\theta_{k-1}}$ $\displaystyle\equiv$
$\displaystyle{\rm
diag}\left(\mathcal{M}_{\theta_{0}},\mathcal{M}_{\theta_{1}},\ldots,\mathcal{M}_{\theta_{k-1}}\right)$
(26)
and
$\underline{\Upsilon}_{k-1}[\rho_{\mathbb{S},0}]\equiv\begin{pmatrix}\mathbbm{1}_{\mathbb{S}}[\rho_{\mathbb{S},0}]\\\
\widetilde{\Phi}_{1}=\underline{\widetilde{T}}_{1}\underline{\Upsilon}_{0}[\rho_{\mathbb{S},0}]\\\
\widetilde{\Phi}_{2}=\underline{\widetilde{T}}_{2}\underline{\mathcal{M}}_{\theta_{1}}\underline{\Upsilon}_{1}[\rho_{\mathbb{S},0}]\\\
\vdots\\\
\widetilde{\Phi}_{k-1}=\underline{\widetilde{T}}_{k-1}\underline{\mathcal{M}}_{\theta_{k-2}}\underline{\Upsilon}_{k-2}[\rho_{\mathbb{S},0}]\end{pmatrix}$
(27)
with $\underline{\Upsilon}_{-1}$ and $\mathcal{M}_{\theta_{0}}$ equal to the
identity operator $\mathbbm{1}_{\mathbb{S}}$ (the first measurement of the
sequence, indeed, is not performed at time $t_{0}$, but at $t_{1}$) and
$\underline{\Upsilon}_{0}=\underline{\Xi}_{0}=\Phi_{0}$.
For the sake of clarity, we show also in this case the first terms of the
recursive expansion of Eq. (25):
$\begin{cases}\widetilde{\Phi}_{0}=\Phi_{0}=\mathbbm{1}_{\mathbb{S}}[\rho_{\mathbb{S},0}]\\\
\widetilde{\Phi}_{1}=\widetilde{T}_{1,0}\widetilde{\Phi}_{0}=\widetilde{T}_{1,0}[\rho_{\mathbb{S},0}]\\\
\widetilde{\Phi}_{2}=\widetilde{T}_{2,0}\widetilde{\Phi}_{0}+\widetilde{T}_{2,1}\mathcal{M}_{\theta_{1}}\widetilde{\Phi}_{1}=\left(\widetilde{T}_{2,0}+\widetilde{T}_{2,1}\mathcal{M}_{\theta_{1}}\widetilde{T}_{1,0}\right)[\rho_{\mathbb{S},0}]\\\
\widetilde{\Phi}_{3}=\widetilde{T}_{3,0}\widetilde{\Phi}_{0}+\widetilde{T}_{3,1}\mathcal{M}_{\theta_{1}}\widetilde{\Phi}_{1}+\widetilde{T}_{3,2}\mathcal{M}_{\theta_{2}}\widetilde{\Phi}_{2}\\\
\,\,\,\,\,\,\,\,\,=\left(\widetilde{T}_{3,0}+\widetilde{T}_{3,1}\mathcal{M}_{\theta_{1}}\widetilde{T}_{1,0}+\widetilde{T}_{3,2}\mathcal{M}_{\theta_{2}}\widetilde{T}_{2,0}\right.\\\
\,\,\,\,\,\,\,\,\left.+\,\widetilde{T}_{3,2}\mathcal{M}_{\theta_{2}}\widetilde{T}_{2,1}\mathcal{M}_{\theta_{1}}\widetilde{T}_{1,0}\right)[\rho_{\mathbb{S},0}].\end{cases}$
In addition, let us observe that also the multi-time statistics for
$\mathbb{S}$, given by the joint probability distribution in Eq.(19), can be
expressed by means of these recursive relations. Indeed, by using again Eq.
(25), one has that
$q_{k}={\rm
Tr}\left(\mathcal{M}_{\theta_{k}}\underline{\widetilde{T}}_{k}\underline{\mathcal{M}}_{\theta_{k-1}}\underline{\Upsilon}_{k-1}[\rho_{\mathbb{S},0}]\right),$
(28)
$\forall k=1,\ldots,m$, where the first terms of the recursion for $q_{k}$ are
given by
$\begin{cases}q_{0}={\rm Tr}\left(\rho_{\mathbb{S},0}\right)=1\\\ q_{1}={\rm
Tr}\left(\mathcal{M}_{\theta_{1}}\widetilde{T}_{1,0}[\rho_{\mathbb{S},0}]\right)\\\
q_{2}={\rm
Tr}\left(\mathcal{M}_{\theta_{2}}\widetilde{T}_{2,0}[\rho_{\mathbb{S},0}]\right)+{\rm
Tr}\left(\mathcal{M}_{\theta_{2}}\widetilde{T}_{2,1}\mathcal{M}_{\theta_{1}}\widetilde{T}_{1,0}[\rho_{\mathbb{S},0}]\right)\\\
q_{3}={\rm
Tr}\left(\mathcal{M}_{\theta_{3}}\widetilde{T}_{3,0}[\rho_{\mathbb{S},0}]\right)+{\rm
Tr}\left(\mathcal{M}_{\theta_{3}}\widetilde{T}_{3,1}\mathcal{M}_{\theta_{1}}\widetilde{T}_{1,0}[\rho_{\mathbb{S},0}]\right)\\\
\,\,\,\,\,\,\,+\,\,{\rm
Tr}\left(\mathcal{M}_{\theta_{3}}\widetilde{T}_{3,2}\mathcal{M}_{\theta_{2}}\widetilde{T}_{2,0}[\rho_{\mathbb{S},0}]\right)\\\
\,\,\,\,\,\,\,+\,\,{\rm
Tr}\left(\mathcal{M}_{\theta_{3}}\widetilde{T}_{3,2}\mathcal{M}_{\theta_{2}}\widetilde{T}_{2,1}\mathcal{M}_{\theta_{1}}\widetilde{T}_{1,0}[\rho_{\mathbb{S},0}]\right).\end{cases}$
Finally, also the set $\\{\widetilde{T}\\}$ of stochastic TTs admits a
hierarchic structure, very similar to that of Eq. (5), based on a
generalization of the CPTP maps $\Gamma$s. The main difference is that now we
need a map acting on the reduced dynamics of $\mathbb{S}$ at each $t_{k}$
after that a quantum measurement has been performed on the system. Thus, such
map will depend on the resulting post-measurement environmental state. In
particular, following the same construction as in [38], we get
$\widetilde{T}_{k,k-n}=\widetilde{\Gamma}_{k|k-n}-\sum_{j=1}^{n-1}\widetilde{T}_{k,k-j}\mathcal{M}_{\theta_{k-j}}\widetilde{\Gamma}_{k-j|k-n},$
(29)
where the CPTP maps
$\widetilde{\Gamma}_{k|k-n}[\sigma_{\mathbb{S}}]={\rm
Tr}_{\mathbb{E}}[\mathcal{U}_{t_{k-n}:t_{k}}[\sigma_{\mathbb{S}}\otimes\widetilde{\sigma}_{\mathbb{E},k-n}]]$
(30)
also include a dependence on the whole sequence of outcomes, fixing the
environmental state $\widetilde{\sigma}_{\mathbb{E},k-n}$ (which motivates the
use of the tilde).
### IV.2 Conditional transfer tensors and memory effects
The relation in Eq. (29) allows us to bring forward the connection between TTs
and CP-divisibility to the stochastic level that describes the effects
sequential measurements.
In particular, let us consider the situation where, for any sequence of
outcomes, all the stochastic TTs $\widetilde{T}_{k,k-n}$ are equal to 0 for
$n\geq 2$. The stochastic TTs hierarchy of Eq. (29) implies that each $1$-step
transfer tensor $\widetilde{T}_{k,k-1}$ is equal to the corresponding $1$-step
$\widetilde{\Gamma}$, i.e.,
$\widetilde{T}_{k,k-1}=\widetilde{\Gamma}_{k|k-1}$, which by definition is
completely positive. In addition, recursively applying
$\widetilde{\Phi}_{k}=\sum_{j=0}^{k-1}\widetilde{T}_{k,j}\mathcal{M}_{\theta_{j}}\widetilde{\Phi}_{j}$,
we get
$\widetilde{\Phi}_{k}=\widetilde{T}_{k|k-1}\mathcal{M}_{\theta_{k-1}}\widetilde{T}_{k-1|k-2}\mathcal{M}_{\theta_{k-2}}\cdots\widetilde{T}_{j+1|j}\mathcal{M}_{\theta_{j}}\widetilde{\Phi}_{j}$
(31)
for any $k\geq j\geq 0$. Therefore, we have
$\widetilde{\Phi}_{k}=\widetilde{\Gamma}_{k|k-1}\mathcal{M}_{\theta_{k-1}}\widetilde{\Gamma}_{k-1|k-2}\mathcal{M}_{\theta_{k-2}}\cdots\widetilde{\Gamma}_{j+1|j}\mathcal{M}_{\theta_{j}}\widetilde{\Phi}_{j},$
(32)
which implies that for any $k\geq j\geq 0$ there is a (conditional) CP map
$\widetilde{\mathcal{E}}_{k,j}$ such that
$\widetilde{\Phi}_{k}=\widetilde{\mathcal{E}}_{k,j}\widetilde{\Phi}_{j}$
i.e., the family of maps $\widetilde{\Phi}_{k}$ is CP-divisibile [41]. We
conclude that, if all the stochastic TTs $\widetilde{T}_{k,k-n}$ are equal to
0 for $n\geq 2$, the corresponding conditional dynamics
$\\{\widetilde{\Phi}_{k}\\}_{k=1,\ldots,m}$ is CP-divisible for any fixed
sequence of outcomes.
As one can directly check from the definition in Eq. (18), the condition in
Eq. (32) is satisfied whenever the global state after the measurements is a
product state, e.g., if the set $\\{\mathcal{M}_{\theta_{k}}\\}$ describe
projective measurements of a non-degenerate system’s observables. This means
that in these situations the conditional dynamics $\widetilde{\Phi}_{k}$ will
be automatically CP-divisible, simply due to the kind of employed measurement.
As we will see in the next paragraph by means of an explicit example, CP-
divisibility of the conditional dynamics, and in particular the validity of
Eq. (32), does not imply that only one-step conditional TTs are different from
zero.
The stochastic TTs $\widetilde{T}_{k,k-n}$ with $n\geq 2$ enclose the
influence of the possible correlations resulting both from the system-
environment interaction and the measurement on the open system up to a time
step $j$ on the conditional dynamics at the later time step $k$. In general,
such correlations depend on the whole sequence of measurement outcomes
performed up to the $j-$th time instant, so that their influence on the
subsequent conditional dynamics can be read as a signature of memory in the
multi-time statistics defined by Eq. (19). However, even in the case where Eq.
(32) holds, some memory can be present in the conditional dynamics, and hence
in the multi-time statistics. In fact, the post-measurement environmental
state $\widetilde{\sigma}_{\mathbb{E}}$, which defines the conditional maps
$\widetilde{\Gamma}$ in Eq. (30), might (and in general will) depend on the
sequence of outcomes. As we will see in the next paragraph, the proper
description of such memory effects can require non-zero TTs
$\widetilde{T}_{k,k-n}$ with $n\geq 2$.
Finally, it is worth observing that, if in addition to the validity of Eq.
(32) the dependence of $\widetilde{\sigma}_{\mathbb{E}}$ on the previous
outcomes can be neglected (i.e., no memory of the sequence of measurements is
left), Eq. (32) automatically implies the quantum regression theorem [42, 43,
44, 45, 46]. Essentially, the latter means that the whole multi-time
statistics can be reconstructed only from the initial open-system state and
open-system dynamical maps that are not conditioned on the sequence of
outcomes. Similar considerations have been recently discussed also in [15].
### IV.3 Spin-boson case study
In this paragraph, we show an application of the previously introduced
stochastic transfer tensors, with the aim to illustrate the capability of the
method in giving a quantitative account of the memory effects both in the
conditional dynamics and in the multi-time statistics. In addition, the use of
TTs also allows to compare on a similar footing the memory effects in the
presence of different kinds of measurements, or even no intermediate
measurements at all.
We consider a single spin in interaction with $5$ quantum harmonic oscillators
and monitored by a sequence of $M=5$ quantum measurements at regular intervals
$\Delta$ with $t_{k}=k\Delta$ and $k=1,\ldots,M$. The Hamiltonian $H$ of the
composite system is given by
$H=\frac{1}{2}\sigma^{z}_{\mathbb{S}}+\sum_{k=1}^{5}\omega_{\mathbb{E},k}b_{\mathbb{E},k}^{\dagger}b_{\mathbb{E},k}+\sum_{k=1}^{5}g_{k}(\sigma_{\mathbb{S}}^{-}b_{\mathbb{E},k}^{\dagger}+\sigma_{\mathbb{S}}^{+}b_{\mathbb{E},k})$
(33)
where $\sigma^{z}$ is the Pauli matrix in the $z$ direction,
$b_{\mathbb{E}}^{\dagger}$ and $b_{\mathbb{E}}$ denote respectively the
creation and annihilation operators associated to each harmonic oscillator of
$\mathbb{E}$, and $\sigma_{\mathbb{S}}^{+}$ and $\sigma_{\mathbb{S}}^{-}$ are
the raising and lowering operators of the spin $\mathbb{S}$. Moreover,
regarding the frequencies $\omega_{\mathbb{E},k}$ of the oscillators and the
interaction couplings $g_{k}$, for the numerical simulations here performed we
have chosen the values
$\\{\omega_{\mathbb{E},k}\\}_{k=1}^{5}\approx(1.99,0.73,0.89,2.04,1.58)$ and
$\\{g_{k}\\}_{k=1}^{5}\approx(1.67,1.32,2.15,2.70,1.07)$. All these values are
expressed in units ensuring that $\hbar=1$. The former have been uniformly
sampled from the interval $[0,5]$, while the latter are sampled from the
probability distribution $\alpha g\exp(-g/\beta)$, with $\alpha=1$ and
$\beta=2$. The parameters $\alpha$ and $\beta$ tune the intensity,
respectively, of the interaction between $\mathbb{S}$ and $\mathbb{E}$ and of
the value of $g$ that corresponds to the peak of the distribution. Note that
this model accounts for both decoherence and dissipation in the dynamics of
$\mathbb{S}$ due to the interaction with the external environment, yet with
the form of the interaction preserving the total number of excitations. On the
other hand, the excitations are generally not preserved by the action of the
measurements. In our simulations, each measurement at $t=t_{k}$ is provided by
the following POVM operators [47]:
$F_{\pm}\equiv(1-\lambda)|\pm\rangle\\!\langle\pm|+\frac{\lambda}{2}\mathbbm{1}_{\mathbb{S}}$
(34)
where $|\pm\rangle\equiv(|0\rangle\pm|1\rangle)/\sqrt{2}$ denote the
eigenstates of the Pauli matrix $\sigma^{x}$. This choice is motivated by the
fact that, by changing the parameter $\lambda$, it is possible to analyze the
case without measurements ($\lambda=1$), the one with projective measurements
($\lambda=0$) and all the intermediate cases ($0<\lambda<1$) corresponding to
a partial collapse of the spin wave-function.
The evolution of the composite system has been evaluated starting from an
initial product state of the ground states of both $\mathbb{S}$ and
$\mathbb{E}$, and taking the time interval $\Delta$ between measurements equal
to 1, 2 and 3 (in natural units allowing for $\hbar=1$). Conversely, the
hierarchy of transfer tensors has been derived by numerically implementing Eq.
(29). The latter is based on the computation of the maps
$\widetilde{\Gamma}_{k|k-n}$, which are all obtained by propagating an initial
product state composed by the reduced states of $\mathbb{S}$ and $\mathbb{E}$
after each possible quantum measurement at the time instants $t_{k}$.
Figure 1: Spin-boson dynamics: Ground state population
$\rho_{\mathbb{S}}^{(1,1)}$ (blues solid lines) and coherence terms
$|\rho_{\mathbb{S}}^{(1,2)}|$ (red dash-dotted lines) of the single spin
$\mathbb{S}$ as a function of time and $4$ distinct $\lambda$ values for
$\Delta=1$ in natural units. Depending on the value of $\lambda$ that fixes
the intermediate measurement operator, the coherence and population of the
spin has a different behaviour. In particular, revivals occur for $\lambda=1$,
while a more and more pronounced damping is observed for decreasing values of
$\lambda$. Indeed, sudden changes in the population and coherence terms denote
the action of the measurements, and they are more evident in the case of
projective measurements ($\lambda=0$). Figure 2: Comparison between the
$L_{2}$-norm of the stochastic TTs $\widetilde{T}_{k,0}$ (coloured dots) and
the figure of merit $\mathcal{D}_{k}$ (circles, squares, diamonds and x-marks)
for $k=2,\ldots,5$ (corresponding to the discrete time instants
$t_{2},\ldots,t_{5}$) and $\lambda=0,0.25,0.5,1$ (identified, respectively, by
the colour black, red, blue and magenta). The former quantifier takes into
account the influence of the $k$-step TTs all referring to $t=t_{0}$, while
the latter sums up the impact of all the involved $k-\ell$ steps TTs with
$\ell=0,\ldots,k-2$ and they both account for the non-Markovianity of the
implemented quantum dynamics. If all the stochastic TTs
$\widetilde{T}_{k,k-n}$ are equal to 0 for $n\geq 2$, then both quantifiers
are equal to zero (the opposite does not generally hold).
In Fig. 1 we plot both the population and coherence terms of the single spin
$\mathbb{S}$ as a function of time and $\lambda$ values
($\lambda\in\\{0,0.25,0.5,1\\}$) for $\Delta=1$ (all in natural units). As one
can observe from the figure, depending on the value of $\lambda$ that
corresponds to different intermediate measurement operators, the spin quantum
coherence tends to decrease for smaller values of $\lambda$ or to be
maintained oscillating over time otherwise. In the latter case, repeated
“coherence revivals” are originated by the interaction of $\mathbb{S}$ with
only the environment $\mathbb{E}$ composed of 5 quantum harmonic oscillators.
Instead, in the former case also the external measurement apparatus plays a
crucial role. Hence, from Fig. 1 one can clearly appreciate that monitoring
the open quantum system $\mathbb{S}$ through a sequence of quantum
measurements radically changes its reduced dynamics, as well as the
corresponding non-Markovian behaviors. In particular, if the quantum coherence
of $\mathbb{S}$ is damping over time, then also high-order step memory effects
can be considered negligible. Otherwise, each state of $\mathbb{S}$ also
depends on past contributions, even up to the initial instant in the extreme
case.
Let us now analyze more quantitatively these considerations by introducing
proper quantifiers of non-Markovianity. In this regard, in Fig. 2 we plot two
quantifiers of the impact of the stochastic $n$-step TTs, with $n\geq 2$, on
the conditional dynamics of the system. In each panel, we consider different
POVMs, defined as in Eq. (34) for $\lambda=0,0.25,0.5,1$ and identified
respectively by the colours black, red, blue and magenta, as a function of the
discrete time instants $t_{k}$ with $k=1,\ldots,5$. The three panels
correspond to three different values of the time interval $\Delta$. The two
quantifiers are the $L_{2}$-norm $\|\cdot\|_{2}$ of the stochastic transfer
tensors $\widetilde{T}_{k,0}$ (coloured dots in the figure) and the figure of
merit (circles, squares, diamonds and x-marks)
$\mathcal{D}_{k}\equiv\frac{1}{k-2}\sum_{\ell=0}^{k-2}\|\widetilde{T}_{k,\ell}\mathcal{M}_{\theta_{\ell}}\widetilde{\Phi}_{\ell}\|_{2}.$
(35)
Starting from the decomposition of the conditional map $\widetilde{\Phi}_{k}$
in correspondence of $t=t_{k}$ as given in Eq. (24), one can observe that the
former quantifier focuses on the role of the $k$-step TT for $k\geq 2$ (note,
indeed, that $\widetilde{\Phi}_{0}=\mathbbm{1}$, i.e., no measurement is
performed at the initial time instant), while the latter sums up the influence
of all the involved $k-\ell$ steps TTs with $k\geq 2$ and $\ell=0,\ldots,k-2$
(thus, the $1$-step TT is not considered). If all the stochastic TTs
$\widetilde{T}_{k,k-n}$ are equal to 0 for $n\geq 2$, then both quantifiers
are equal to zero; the opposite does not generally hold.
Overall, from Fig. 2 we can observe that the quantities
$\|\widetilde{T}_{k,0}\|_{2}$ and $\mathcal{D}_{k}$ are comparable; namely, in
the considered parametric regime the $k$-th step TT provides with similar
information on the memory effects influencing $\widetilde{\Phi}_{k}$, compared
to the sum of all the contributions from the TTs with step larger than 1. They
both indicate that non-Markovian behaviours, originated from the interaction
of the spin both with the environment and the observer, become relevant for
$t/\Delta\geq 3$ with all the considered values of $\Delta$. Moreover, both
the non-Markovianity quantifiers reach their maximum values in the case where
no intermediate measurement is performed (magenta dots and x-marks),
corresponding in Fig. 1 to a nearly periodic evolution of the coherence term
$|\rho_{\mathbb{S}}^{(1,2)}|$ over time. Thus, comparing Figs. 1 and 2, we can
deduce that such periodic behaviours are originated by high-order step memory
effects, presumably even from $t_{0}$. While without applying any measurement
both quantifiers typically increase with time, a more pronounced non-monotonic
behavior is observed in the presence of intermediate measurements, which
reduce the memory effects caused by the interaction between $\mathbb{S}$ and
$\mathbb{E}$. Although a decreased impact of multi-step TTs in the presence of
intermediate measurements (corresponding to a decreased impact of system-
environment correlations on the conditional dynamics) might have been
expected, the quantifiers do not necessarily reach their minimum values in the
case of projective measurements ($\lambda=0$). Rather, the minimum values are
reached more often for $\lambda=0.25$ or $\lambda=0.5$ (the behavior for these
two POVMs is quite similar). However, as anticipated, $n$-step TTs for $n\geq
2$ can be non-zero also in the case of projective measurements, i.e., when Eq.
(32) holds. Such TTs account for memory effects due to the dependence of the
post-measurement environmental state on the previous outcomes. Even more, the
latter can exceed the memory effects – due to both system-environment
correlations and changes in the environmental states – occurring in the case
of non-projective measurements, as quantified by $\|\widetilde{T}_{k,0}\|_{2}$
and $\mathcal{D}_{k}$.
Figure 3: $L_{2}$-norm of the difference between the left- and the right-and-
side of equation (32), namely
$\|\widetilde{\Phi}_{k}-\widetilde{\Gamma}_{k|k-1}\mathcal{M}_{\theta_{k-1}}\widetilde{\Gamma}_{k-1|k-2}\mathcal{M}_{\theta_{k-2}}\cdots\widetilde{\Gamma}_{j+1|j}\mathcal{M}_{\theta_{j}}\widetilde{\Phi}_{j}\|_{2}$,
computed for $\Delta=2$ and $\lambda=0,0.25,0.5,1$ (magenta dots, blue
x-marks, red diamonds, black squares, respectively) in correspondence of the
discrete time-instants $t_{k}=k\Delta$ with $k\in\\{2,3,4,5\\}$. This
represents the $L_{2}$-norm of the difference between the actual stochastic
map $\widetilde{\Phi}_{k}$ and the CP-divisible map provided by the product
states of the dynamical evolution of the spin after each measurement. During
the transient of the spin dynamics, Eq. (32) is violated as the value of
$\lambda$ increases (thus, for less invasive intermediate measurements);
however, such tendency can be reversed in case the spin approaches an
equilibrium asymptotic state or a stable periodicity phase.
In order to point out the memory effects originated by system-environment
correlations, we consider the $L_{2}$-norm of the difference between the left-
and the right-and-side of Eq. (32), namely between the actual stochastic map
$\widetilde{\Phi}_{k}$ (in correspondence of the $k$-th time instant $t_{k}$)
and the CP-divisible map
$\widetilde{\Gamma}_{k|k-1}\mathcal{M}_{\theta_{k-1}}\widetilde{\Gamma}_{k-1|k-2}\mathcal{M}_{\theta_{k-2}}\cdots\widetilde{\Gamma}_{j+1|j}\mathcal{M}_{\theta_{j}}\widetilde{\Phi}_{j}$
associated with the product states after each measurement. The results are
plotted in Fig. 3. It is worth noting that Eq. (32) is satisfied only when
projective measurements are performed, so that the composite system is
actually in a product state after each measurement and the conditional
dynamics of $\mathbb{S}$ can be fully described by means of the composition of
$1$-step maps $\widetilde{\Gamma}$. In this case, the computation of the
operators $\widetilde{\Gamma}_{k|k-1}$ is enough by itself. However, as said,
some $n$-step stochastic TTs with $n\geq 2$ can be still different from zero,
despite overall their combination is canceled to ensure the validity of Eq.
(32).
As final remark, from Fig. 3 we can also observe that for $k=2,3,4$ the
violation of Eq. (32) increases with the value of $\lambda$, i.e., when moving
toward less invasive measurements; indeed, the largest violation is reached
for the case without intermediate measurements. Moreover, quite interestingly,
the situation is reversed in the correspondence of the last observed value of
time, $t_{k}/\Delta=5$, for which $\lambda=0.25$ induces the largest
violation, while the unconditional dynamics the smallest (apart, of course,
from the case of projective measurements). The latter behaviour of the
unconditional dynamics is usually originated by a tendency towards an
equilibrium asymptotic state or to a stable periodicity phase. In our case
study, being the environment composed by a finite number (five) of quantum
harmonic oscillators, the simulated unconditional dynamics is practically
periodic, with period approximately of $t=10$ (in natural units).
## V Conclusion and outlook
In this paper, we have studied the non-Markovian dynamics of open quantum
systems and their multi-time statistics by means of the transfer-tensor
approach, which allowed us to treat the memory effects in the two different
scenarios on a similar footing. After showing the connection between the
hierarchy of the TTs and the divisibility of the dynamics, we have extended
the definition of TTs to the case where the open system of interest undergoes
quantum measurements at subsequent instants of time. We have thus introduced a
stochastic family of TTs, depending on the sequence of measurement outcomes,
and a related hierarchy that captures how the multi-time statistics is
influenced by both the system-environment correlations and the dependence of
the current environmental state on the previous outcomes. Finally, we have
defined two quantifiers of multi-time memory effects, which rely directly on
the various contributions to the hierarchy of TTs, and we have investigated
their behavior in a paradigmatic case study, comparing different kinds of
intermediate measurements, as well as the case with no monitoring at all.
The precise relation between the memory effects present, respectively, in non-
Markovian quantum dynamics and in their multi-time statistics remains still to
be addressed. In particular, more quantitative and general results
characterizing different kinds of multi-time statistics, such as those obeying
the quantum-regression theorem, seem to be needed in order to distinguish the
different sources of memory and to compare them with those leading to non-
Markovian quantum dynamics. Hopefully, our results will provide useful
insights to look for such rigorous connections. More in general, our analysis
suggests the usefulness of the TT approach also to deal with multi-time
statistics; an example in this direction might be to use the hierarchy of
stochastic TTs to evaluate the non-Markovianity along the so-called most
probable trajectory of the system, namely the trajectory originated with
higher probability by a sequence of quantum measurements [48, 49]. In
addition, the TT formalism addressed in this paper is also expected to provide
a powerful tool to evaluate experimentally the degree of non-Markovianity [50]
in monitored quantum dynamics [51, 52], as well as to test the validity of the
quantum regression theorem or possible generalizations [53, 54, 11]. One
further advantage of the presented approach would be to track memory effects
also in the reduced state evolution of the environment $\mathbb{E}$. This
feature might be useful in those quantum technology devices (e.g., even the
commercial ones now available online as IBM or Rigetti [55, 56, 57]) in which
the number of the environmental degrees of freedom is not much larger than the
ones of the system $\mathbb{S}$ under analysis. Accordingly, in such a case,
also the state of $\mathbb{E}$ is subjected to frequent and relevant changes
due to the presence of $\mathbb{S}$, and the bi-directional exchange of
information between the system and the environment likely leads to departures
from Markovian dynamics and the quantum regression theorem. Finally, it is
worth also mentioning how stochastic transfer tensors may find application in
modeling the interaction between a quantum system $\mathbb{S}$ and one or more
external thermal baths, giving rise to thermalization processes in the large-
time limit. During the transient of the dynamics, indeed, $\mathbb{S}$ may
exhibit non-Markovian behaviours and non-trivial memory effects. In this
regard, if for example we resort to quantum collision models [58, 59], a
thermal baths is assumed of being composed by a large number of small
subsystems, such that, at discrete time instants, the dynamics between the
quantum system and the bath take place through successive “collisions”
provided by pairwise short interactions. Similar thermalization effects are
also obtained by monitoring $\mathbb{S}$ through a sequence of projective
measurements, in the limit of many measurements [60]. Accordingly, in such a
contexts, the TT formalism is expected to shed light in quantitative terms on
the effective influence of transient non-Markovian dynamics, due to the
interaction with thermal baths, within the quantum system under analysis in
allowing for specific thermalization quantum processes.
#### Note added:
During the completion of this work, the related paper [61] appeared, in which
a generalization of the transfer-tensor formalism to multi-time measurements
is introduced by means of the process-tensor description of the open-system
evolution.
###### Acknowledgements.
S.G. and F.C. were financially supported from PATHOS EU H2020 FET-OPEN grant
No. 828946, the Fondazione CR Firenze through the project Quantum-AI and UNIFI
grant Q-CODYCES.
## References
* [1] M.M. Wolf, and J.I. Cirac. Dividing Quantum Channels. Commun. Math. Phys. 279, 147-168 (2008).
* [2] H.-P. Breuer, E.-M. Laine, and J. Piilo. Measure for the Degree of Non-Markovian Behavior of Quantum Processes in Open Systems. Phys. Rev. Lett. 103, 210401 (2009).
* [3] E.-M. Laine, J. Piilo, and H.-P. Breuer. Measure for the non-Markovianity of quantum processes. Phys. Rev. A 81, 062115 (2010).
* [4] A. Rivas, S.F. Huelga, and M.B. Plenio. Entanglement and non-Markovianity of quantum evolutions. Phys. Rev. Lett. 105, 050403 (2010).
* [5] B.-H. Liu, L. Li, Y.-F. Huang, C.-F. Li, G.-C. Guo, E.-M. Laine, H.-P. Breuer, and J. Piilo. Experimental control of the transition from Markovian to non-Markovian dynamics of open quantum systems. Nat. Phys. 7, 931-934 (2011).
* [6] D. Chruscinski, and S. Maniscalco. Degree of Non-Markovianity of Quantum Evolution. Phys. Rev. Lett. 112, 120404 (2014).
* [7] A. Rivas, S.F. Huelga, and M.B. Plenio. Quantum non-Markovianity: characterization, quantification and detection. Rep. Prog. Phys. 77, 094001 (2014).
* [8] F. Caruso, V. Giovannetti, C. Lupo, and S. Mancini. Quantum channels and memory effects. Rev. Mod. Phys. 86, 1203 (2014).
* [9] H.-P. Breuer, E.-M. Laine, J. Piilo, and B. Vacchini. Colloquium: Non-Markovian dynamics in open quantum systems. Rev. Mod. Phys. 88, 021002 (2016).
* [10] M.M. Müller, S. Gherardini, and F. Caruso. Stochastic quantum Zeno-based detection of noise correlations. Scientific Reports 6, 38650 (2016).
* [11] I. de Vega, and D. Alonso. Dynamics of non-Markovian open quantum systems. Rev. Mod. Phys. 89, 015001 (2017).
* [12] S. Cialdi, M.A.C. Rossi, C. Benedetti, B. Vacchini, D. Tamascelli, S. Olivares, and M.G.A. Paris. All-optical quantum simulator of qubit noisy channels. Appl. Phys. Lett. 110, 081107 (2017).
* [13] F.A. Pollock, C. Rodríguez-Rosario, T. Frauenheim, M. Paternostro, and K. Modi. Operational Markov condition for quantum processes. Phys. Rev. Lett. 120, 040405 (2018).
* [14] F.A. Pollock, C. Rodríguez-Rosario, T. Frauenheim, M. Paternostro, and K. Modi, Non-Markovian quantum processes: Complete framework and efficient characterization. Phys. Rev. A 97, 012127 (2018).
* [15] A.A. Budini. Quantum Non-Markovian Processes Break Conditional Past-Future Independence. Phys. Rev. Lett. 121, 240401 (2018).
* [16] S. Campbell, F. Ciccarello, G.M. Palma, and B. Vacchini. System-environment correlations and Markovian embedding of quantum non-Markovian dynamics. Phys. Rev. A 98, 012142 (2018).
* [17] P. Taranto, F.A. Pollock, S. Milz, M. Tomamichel, and K. Modi. Quantum Markov Order. Phys. Rev. Lett. 122, 140401 (2019).
* [18] H.-V. Do, C. Lovecchio, I. Mastroserio, N. Fabbri, F.S. Cataliotti, S. Gherardini, M.M. Müller, N. Dalla Pozza, and F. Caruso. Experimental proof of Quantum Zeno-assisted Noise Sensing. New J. Phys. 21, 113056 (2019).
* [19] S. Campbell, M. Popovic, D. Tamascelli, and B. Vacchini. Precursors of non-Markovianity. New J. Phys. 21, 053036 (2019).
* [20] S. Cialdi, C. Benedetti, D. Tamascelli, S. Olivares, M. G. A. Paris, and B. Vacchini. Experimental investigation of the effect of classical noise on quantum non-Markovian dynamics. Phys. Rev. A 100, 052104 (2019).
* [21] S. Gherardini, S. Marcantoni, and F. Caruso. Irreversibility mitigation in unital non-Markovian quantum evolutions. Phys. Rev. Res. 2, 033250 (2020).
* [22] S. Martina, S. Gherardini, and F. Caruso. Machine learning approach for quantum non-Markovian noise classification. arXiv:2101.03221.
* [23] L. Li, M.J. W. Hall, H.M. Wiseman. Concepts of quantum non-Markovianity: a hierarchy. Phys. Rep. 759, 1-51 (2018).
* [24] H. Breuer, and F. Petruccione. The Theory of Open Quantum Systems (Oxford University Press, Oxford, 2002).
* [25] Á. Rivas, and S.F. Huelga. Open Quantum Systems: An Introduction (Springer, Berlin, 2012).
* [26] B. Vacchini, A. Smirne, E.-M. Laine, J. Piilo, and H.-P. Breuer. Markovian and non-Markovian dynamics in quantum and classical systems. New J. Phys. 13, 093004 (2011).
* [27] A. Smirne, A. Stabile, and B. Vacchini. Signatures of non-Markovianity in classical single-time probability distributions. Phys. Scr. T153, 014057 (2013).
* [28] S. Milz, M. S. Kim, F. A. Pollock, K. Modi. Completely Positive Divisibility Does Not Mean Markovianity. Phys. Rev. Lett. 123, 040401 (2019).
* [29] A. Smirne, D. Egloff, M.G. Díaz, M.B. Plenio, and S.F. Huelga. Coherence and non-classicality of quantum Markov processes. Quantum Sci. Technol. 4, 01LT01 (2019).
* [30] P. Strasberg and M.G. Díaz. Classical quantum stochastic processes. Phys. Rev. A. 100, 022120 (2019).
* [31] S. Milz, F. Sakuldee, F.A. Pollock, and K. Modi. Kolmogorov extension theorem for (quantum) causal modelling and general probabilistic theories. Quantum 4, 255 (2020).
* [32] A. Smirne, T. Nitsche, D. Egloff, S. Barkhofen, S. De, I. Dhand, C. Silberhorn, S.F. Huelga, and M.B. Plenio. Experimental Control of the Degree of Non-Classicality via Quantum Coherence. Quantum Sci. Technol. 5 04LT01 (2020).
* [33] M.G. Díaz, B. Desef, M. Rosati, D. Egloff, J. Calsamiglia, A. Smirne, M. Skotiniotis, and S. F. Huelga. Accessible coherence in open quantum system dynamics. Quantum 4, 249 (2020).
* [34] S. Milz, D. Egloff, P. Taranto, T. Theurer, M.B. Plenio, A. Smirne, S.F. Huelga. When is a non-Markovian quantum process classical? Phys. Rev. X 10, 041049 (2020).
* [35] S. Gherardini, A. Smirne, M.M. Müller, and F. Caruso. Advances in sequential measurements and control of open quantum systems. Proceedings 12 (1), 11 (2019).
* [36] J. Cerrillo, and J. Cao. Non-Markovian dynamical maps: Numerical processing of open quantum trajectories. Phys. Rev. Lett. 112, 110401 (2014).
* [37] R. Rosenbach, J. Cerrillo, S.F. Huelga, J. Cao, and M.B. Plenio. Efficient simulation of non-Markovian system-environment interaction. New J. Phys. 18, 023035 (2016).
* [38] F.A. Pollock, and K. Modi. Tomographically reconstructed master equations for any open quantum dynamics. Quantum 2, 76 (2018).
* [39] Y.-Q. Chen, K.-L. Ma, Y.-C. Zheng, J. Allcock, S. Zhang, and C.-Y. Hsieh. Non-Markovian Noise Characterization with the Transfer Tensor Method. Phys. Rev. Applied 13, 034045 (2020).
* [40] In fact, Eq.(16) implies the property named No Information Backflow in [23], which is sufficient but not necessary for CP-divisibility, so that the latter cannot imply $T_{k,k-n}=0$ for $n\geq 2$.
* [41] We use here the term CP-divisible referred to a family of maps that are not necessarily trace-preserving. Hence, we also relax the requirement that the intermediate maps $\widetilde{\mathcal{E}}_{k,j}$ are trace preserving.
* [42] M. Lax. Quantum Noise. XI. Multitime Correspondence between Quantum and Classical Stochastic Processes. Phys. Rev. 172, 350 (1968).
* [43] N.G. van Kampen, Stochastic Processes in Physics and Chemistry (North-Holland, Amsterdam,1992).
* [44] H. Carmichael, An Open Systems Approach to Quantum Optics (Springer-Verlag, Berlin, 1993).
* [45] C. W. Gardiner and P. Zoller, Quantum Noise: A Handbook of Markovian and Non-Markovian Quantum Stochastic Methods with Applications to Quantum Optics (Springer, Berlin, 2004).
* [46] G. Guarnieri, A. Smirne, and B. Vacchini. Quantum regression theorem and non-Markovianity of quantum dynamics. Phys. Rev. A 90, 022110 (2014).
* [47] C. Addis, T. Heinosaari, J. Kiukas, E.-M. Laine, and S. Maniscalco. Dynamics of incompatibility of quantum measurements in open systems. Phys. Rev. A 93, 022114 (2016).
* [48] S.J. Weber, A. Chantasri, J. Dressel, A.N. Jordan, K.W. Murch, and I. Siddiqi. Mapping the optimal route between two quantum states. Nature (London) 511, 570 (2014).
* [49] S. Gherardini. Exact nonequilibrium quantum observable statistics: A large-deviation approach. Phys. Rev. A 99 (6), 062105 (2019).
* [50] C.-F. Li, G.-C. Guo, J. Piilo. Non-Markovian quantum dynamics: What is it good for? EPL 128, 30001 (2019).
* [51] G.A.L. White, C.D. Hill, F.A. Pollock, L.C.L. Hollenberg, and K. Modi. Demonstration of non-Markovian process characterisation and control on a quantum processor. Nat. Comm. 11, 6301 (2020).
* [52] L. Xiang, Z. Zong, Z. Zhan, Y. Fei, C. Run, Y. Wu, W. Jin, C. Xiao, Z. Jia, P. Duan, J. Wu, Y. Yin, G. Guo. Quantify the non-Markovian process with intermediate projections in a superconducting processor. arXiv:2105.03333.
* [53] H.-S. Goan, P.-W. Chen, and C.-C. Jian. Non-Markovian finite-temperature two-time correlation functions of system operators: Beyond the quantum regression theorem. J. Chem. Phys. 12, 124112 (2011).
* [54] D.P.S. McCutcheon. Optical signatures of non-Markovian behavior in open quantum systems. Phys. Rev. A 93, 22119 (2016).
* [55] J. Morris, F.A. Pollock, and K. Modi. Non-Markovian memory in IBMQX4. arXiv:1902.07980.
* [56] G. García-Pérez, M.A.C. Rossi, and S. Maniscalco IBM Q Experience as a versatile experimental testbed for simulating open quantum systems. npj Quantum Inf. 6, 1 (2020).
* [57] S. Martina, et al. In preparation.
* [58] F. Ciccarello, S. Lorenzo, V. Giovannetti, and G.M. Palma. Quantum collision models: open system dynamics from repeated interactions. arXiv:2106.11974.
* [59] O. Arisoy, S. Campbell, and Ö.E. Müstecaplioğlu. Thermalization of finite many-body systems by a collision model. Entropy 21, 1182 (2019).
* [60] S. Gherardini, G. Giachetti, S. Ruffo, and A. Trombettoni. Thermalization processes induced by quantum monitoring in multi-level systems. arXiv:2012.15216.
* [61] M. R. Jørgensen and F. A. Pollock. A discrete memory-kernel for multi-time correlations in non-Markovian quantum processes. Phys. Rev. A 102, 052206 (2020).
|
# De Giorgi’s inequality for the thresholding scheme with arbitrary mobilities
and surface tensions
Tim Laux Institut für angewandte Mathematik, Universität Bonn, Endenicher
Allee 60, 53115 Bonn, Germany<EMAIL_ADDRESS>and Jona Lelmi
Institut für angewandte Mathematik, Universität Bonn, Endenicher Allee 60,
53115 Bonn, Germany<EMAIL_ADDRESS>
###### Abstract.
We provide a new convergence proof of the celebrated Merriman-Bence-Osher
scheme for multiphase mean curvature flow. Our proof applies to the new
variant incorporating a general class of surface tensions and mobilities,
including typical choices for modeling grain growth. The basis of the proof
are the minimizing movements interpretation of Esedoğlu and Otto and De
Giorgi’s general theory of gradient flows. Under a typical energy convergence
assumption we show that the limit satisfies a sharp energy-dissipation
relation.
## 1\. Introduction
The thresholding scheme is a highly efficient computational scheme for
multiphase mean curvature flow (MCF) which was originally introduced by
Merriman, Bence, and Osher [25], [26]. The main motivation for MCF comes from
metallurgy where it models the slow relaxation of grain boundaries in
polycrystals [29]. Each ”phase” in our mathematical jargon corresponds to a
grain, i.e., a region of homogeneous crystallographic orientation. The
effective surface tension $\sigma_{ij}(\nu)$ and the mobility $\mu_{ij}(\nu)$
of a grain boundary depend on the mismatch between the lattices of the two
adjacent grains $\Omega_{i}$ and $\Omega_{j}$ and on the relative orientation
of the grain boundary, given by its normal vector $\nu$. It is well known that
for small mismatch angles, the dependence on the normal can be neglected [30].
The effective evolution equations then read
(1) $\displaystyle V_{ij}=-\mu_{ij}\sigma_{ij}H_{ij}\quad\text{along the grain
boundary }\Sigma_{ij},$
where $V_{ij}$ and $H_{ij}$ denote the normal velocity and mean curvature of
the grain boundary $\Sigma_{ij}=\partial\Omega_{i}\cap\partial\Omega_{j}$,
respectively. These equations are coupled by the Herring angle condition
(2)
$\displaystyle\sigma_{ij}\nu_{ij}+\sigma_{jk}\nu_{jk}+\sigma_{ki}\nu_{ki}=0\quad\text{along
triple junctions }\Sigma_{ij}\cap\Sigma_{jk},$
which is a balance-of-forces condition and simply states that triple junctions
are in local equilibrium. Efficient numerical schemes allow to carry out
large-scale simulations to give insight into relevant statistics like the
average grain size or the grain boundary character distribution, as an
alternative to studying corresponding mean field limits as in [4], [16]. The
main obstruction to directly discretize the dynamics (1)–(2) are ubiquitous
topological changes in the network of grain boundaries like for example the
vanishing of grains. Thresholding instead naturally handles such topological
changes. The scheme is a time discretization which alternates between the
following two operations: _(i)_ convolution with a smooth kernel; _(ii)_
thresholding. The second step is a simple pointwise operation and also the
first step can be implemented efficiently using the Fast Fourier Transform.
One of the main objectives of our analysis is to rigorously justify this
intriguingly simple scheme in the presence of such topological changes.
The basis of our analysis is the underlying gradient-flow structure of
(1)–(2), which means that the solution follows the steepest descent in an
energy landscape. More precisely, the energy is the total interfacial area
weighted by the surface tensions $\sigma_{ij}$, and the metric tensor is the
$L^{2}$-product on normal velocities, weighted by the inverse mobilities
$\frac{1}{\mu_{ij}}$. One can read off this structure from the inequality
$\displaystyle\frac{d}{dt}\sum_{i,j=1}^{N}\sigma_{ij}\textup{Area}(\Sigma_{ij})=-\sum_{i,j=1}^{N}\frac{1}{\mu_{ij}}\int_{\Sigma_{ij}}V_{ij}^{2}\,dS\leq
0,$
which is valid for sufficiently regular solutions to (1)–(2). In the seminal
work [8], Esedoğlu and Otto showed that the efficient thresholding scheme
respects this gradient-flow structure as it may be viewed as a minimizing
movements scheme in the sense of De Giorgi. More precisely, they show that
each step in the scheme is equivalent to solving a variational problem of the
form
(3)
$\displaystyle\min_{\chi}\left\\{\frac{1}{2h}d_{h}^{2}(\Sigma,\Sigma^{n-1})+E_{h}(\Sigma)\right\\},$
where $E_{h}(\Sigma)$ and $d_{h}(\Sigma,\Sigma^{n-1})$ are proxies for the
total interfacial energy of the configuration $\Sigma$ and the distance of the
configuration $\Sigma$ to the one at the previous time step $\Sigma^{n-1}$,
respectively. Since the work of Jordan, Kinderlehrer, and Otto [14], the
importance of the formerly often neglected metric in such gradient-flow
structures has been widely appreciated. Also in the present work, the focus
lies on the metric, which in the case of MCF is well-known to be completely
degenerate [27]. This explains the proxy for the metric appearing in the
related well-known minimizing movements scheme for MCF by Almgren, Taylor, and
Wang, [1], and Luckhaus and Sturzenhecker [23]. This remarkable connection
between the numerical scheme and the theory of gradient flows has the
practical implication that it made clear how to generalize the algorithm to
arbitrary surface tensions $\sigma_{ij}$. From the point of view of numerical
analysis, (3) means that thresholding behaves like the _implicit_ Euler scheme
and is therefore unconditionally stability. The variational interpretation of
the thresholding scheme has of course implications for the analysis of the
algorithm as well. It allowed Otto and one of the authors to prove convergence
results in the multiphase setting [17], [18], which lies beyond the reach of
the more classical viscosity approach based on the comparison principle
implemented in [9], [3], [12]. Also in different frameworks, this variational
viewpoint turned out to be useful, such as MCF in higher codimension [21] or
the Muskat problem [13]. The only downside of the generalization [8] are the
somewhat unnatural effective mobilities $\mu_{ij}=\frac{1}{\sigma_{ij}}$. Only
recently, Salvador and Esedoğlu [31] have presented a strikingly simple way to
incorporate a wide class of mobilities $\mu_{ij}$ as well. Their algorithm is
based on the fact that although the same kernel appears in the energy and the
metric, each term only uses certain properties of the kernel, which can be
tuned independently: Starting from two Gaussian kernels $G_{\gamma}$ and
$G_{\beta}$ of different width, they find a positive linear combination
$K_{ij}=a_{ij}G_{\gamma}+b_{ij}G_{\beta}$, whose effective mobility and
surface tension match the given $\mu_{ij}$ and $\sigma_{ij}$, respectively. It
is remarkable that this algorithm retains the same simplicity and structure as
the previous ones [26], [8]. We refer to Section 2 for the precise statement
of the algorithm.
In the present work, we prove the first convergence result for this new
general scheme. We exploit the gradient-flow structure and show that under the
natural assumption of energy convergence, any limit of thresholding satisfies
De Giorgi’s inequality, a weak notion of multiphase mean curvature flow. This
assumption is inspired by the fundamental work of Luckhaus-Sturzenhecker [23]
and has appeared in the context of thresholding in [17], [18]. We expected it
to hold true before the onset of singularities such as the vanishing of
grains. Furthermore, at least in the simpler two-phase case, it can be
verified for certain singularities [6], [5]. We would in fact expect this
assumption to be true generically, which however seems to be a difficult
problem in the multiphase case.
The present work fits into the theory of general gradient flows even better
than the two previous ones [17], [18] and crucially depends on De Giorgi’s
abstract framework, cf. [2]. This research direction was initiated by Otto and
the first author and appeared in the lecture notes [19]. There, De Giorgi’s
inequality is derived for the simple model case of two phases. Here, we
complete these ideas and use a careful localization argument to generalize
this result to the multiphase case. A further particular novelty of our work
is that for the first time, we prove the convergence of the new scheme for
arbitrary mobilities [31].
Our proof rests on the fact that thresholding, like any minimizing movements
scheme, satisfies a sharp energy-dissipation inequality of the form
(4) $\displaystyle
E_{h}(\Sigma^{h}(T))+\frac{1}{2}\int_{0}^{T}\left(\frac{1}{h^{2}}d_{h}^{2}(\Sigma^{h}(t),\Sigma^{h}(t-h))+|\partial
E_{h}|^{2}(\tilde{\Sigma}^{h}(t))\right)dt\leq E_{h}(\Sigma(0)),$
where $\Sigma^{h}(t)$ denotes the piecewise constant interpolation in time of
our approximation, $\tilde{\Sigma}^{h}(t)$ denotes another, intrinsic
interpolation in terms of the variational scheme, cf. Lemma 3, and $|\partial
E_{h}|$ is the metric slope of $E_{h}$, cf. (33).
Our main goal is to pass to the limit in (4) and obtain the sharp energy-
dissipation relation for the limit, which in the simple two-phase case
formally reads
(5)
$\displaystyle\sigma\mathrm{Area}(\Sigma(T))+\frac{1}{2}\int_{0}^{T}\int_{\Sigma(t)}\left(\frac{1}{\mu}V^{2}+\sigma^{2}\mu
H^{2}\right)dS\,dt\leq\sigma\mathrm{Area}(\Sigma(0)).$
To this end, one needs sharp lower bounds for the terms on the left-hand side
of (4). While the proof of the lower bound on the metric slope of the energy
(6) $\displaystyle\liminf_{h\downarrow 0}\int_{0}^{T}|\partial
E_{h}|^{2}(\tilde{\Sigma}^{h}(t))\,dt\geq\sigma^{2}\mu\int_{0}^{T}\int_{\Sigma(t)}H^{2}dS\,dt$
is a straight-forward generalization of the argument in [19], the main novelty
of the present work lies in the sharp lower bound for the distance-term of the
form
(7) $\displaystyle\liminf_{h\downarrow
0}\int_{0}^{T}\frac{1}{h^{2}}d_{h}^{2}(\Sigma^{h}(t),\Sigma^{h}(t-h))\,dt\geq\frac{1}{\mu}\int_{0}^{T}\int_{\Sigma(t)}V^{2}\,dS\,dt.$
This requires us to work on a mesoscopic time scale $\tau\sim\sqrt{h}$, which
is much larger than the microscopic time-step size $h$ and which is natural in
view of the parabolic nature of our problem. It is remarkable that De Giorgi’s
inequality (5) in fact characterizes the solution of MCF under additional
regularity assumptions. Indeed, if $\Sigma(t)$ evolves smoothly, this
inequality can be rewritten as
(8)
$\displaystyle\frac{1}{2}\int_{0}^{T}\int_{\Sigma(t)}\sigma\Big{(}\frac{1}{\sqrt{\mu\sigma}}V+\sqrt{\sigma\mu}H\Big{)}^{2}dS\,dt\leq
0,$
and therefore $V=-\mu\sigma H$. For expository purpose, we focused here on the
vanilla two-phase case. In the multiphase case, the resulting inequality
implies both the PDEs (1) and the balance-of-forces conditions (2), cf. Remark
1. An optimal energy-dissipation relation like the one here also plays a
crucial role in the recent weak-strong uniqueness result for multiphase mean
curvature flow by Fischer, Hensel, Simon, and one of the authors [10]. There,
a new dynamic analogue of calibrations is introduced and uniqueness is
established in the following two steps: _(i)_ any strong solution is a
calibrated flow and _(ii)_ every calibrated flow is unique in the class of
weak solutions. De Giorgi’s general strategy we are implementing here is also
related to the approaches by Sandier and Serfaty [32] and Mielke [28]. They
provide sufficient conditions for gradient flows to converge in the same
spirit as $\Gamma$-convergence of energy functionals, implies the convergence
of minimizers. In the dynamic situation it is clear that one needs conditions
on both energy and metric in order to verify such a convergence.
There has been continuous interest in MCF in the mathematics literature, so we
only point out some of the most relevant recent advances. We refer the
interested reader to the introductions of [17] and [20] for further related
references. The existence of global solutions to multiphase MCF has only been
established recently by Kim and Tonegawa [15] who carefully adapt Brakke’s
original construction and show in addition that phases do not vanish
spontaneously. For the reader who wants to familiarize themselves with this
topic, we recommend the recent notes [34]. Another approach to understanding
the long-time behavior of MCF flow is to restart strong solutions after
singular times. This amounts to solving the Cauchy problem with non-regular
initial data, such as planar networks of curves with quadruple junctions. In
this two-dimensional setting, this has been achieved by Ilmanen, Neves, and
Schulze [11] by gluing in self-similarly expanding solutions for which it is
possible to show that the initial condition is attained in some measure
theoretic way. Most recently, using a similar approach of gluing in self-
similar solutions, but also relying on blow-ups from geometric microlocal
analysis, Lira, Mazzeo, Pluda, Saez [22] were able to construct such strong
solutions, prove stronger convergence towards the initial (irregular) network
of curves, and classify all such strong solutions.
The rest of the paper is structured as follows. In Section 2 we recall the
thresholding scheme for arbitrary mobilitites introduced in [31], show its
connection to the abstract framework of gradient flows, and record the direct
implications of this theory. We state and discuss our main results in Section
3. Section 4 contains the localization argument in space, which will play a
crucial role in the proofs which are gathered in Section 5. Finally, in the
short Appendix, we record some basic facts about thresholding.
## 2\. Setup and the modified thresholding scheme
### 2.1. The modified algorithm
We start by describing the algorithm proposed by Salvador and Esedoğlu in
[31]. Let the symmetric matrix
$\mathbb{\sigma}=(\sigma_{ij})_{ij}\in\mathbf{R}^{N\times N}$ of surface
tensions and the symmetric matrix $\mathbb{\mu}=(\mu_{ij})_{ij}$ of mobilities
be given. In this work we define for notational convenience
$\sigma_{ii}=\mu_{ii}=0$. Let $\gamma>\beta>0$ be given. Define the matrices
$\mathbb{A}=(-a_{ij})_{ij}\in\mathbf{R}^{N\times N}$ and
$\mathbb{B}=(-b_{ij})_{ij}\in\mathbf{R}^{N\times N}$ by
(9) $\displaystyle
a_{ij}=\frac{\sqrt{\pi}\sqrt{\gamma}}{\gamma-\beta}(\sigma_{ij}-\beta\mu_{ij}^{-1}),$
(10) $\displaystyle
b_{ij}=\frac{\sqrt{\pi}\sqrt{\beta}}{\gamma-\beta}(-\sigma_{ij}+\gamma\mu_{ij}^{-1}),$
for $i\neq j$ and $a_{ii}=b_{ii}=0$. Then $a_{ij},b_{ij}$ are uniquely
determined as solutions of the following linear system
(11)
$\begin{cases}\sigma_{ij}=\frac{a_{ij}\sqrt{\gamma}}{\sqrt{\pi}}+\frac{b_{ij}\sqrt{\beta}}{\sqrt{\pi}},\\\
\mu_{ij}^{-1}=\frac{a_{ij}}{\sqrt{\pi}\sqrt{\gamma}}+\frac{b_{ij}}{\sqrt{\pi}\sqrt{\beta}}.\end{cases}$
The algorithm introduced by Salvador and Esedoğlu is as follows. Let the time
step size $h>0$ be fixed. Hereafter $G_{\gamma}^{h}:=G_{\gamma h}^{(d)}$
denotes the $d$-dimensional heat kernel (17) at time $\gamma h$.
###### Algorithm 1 (Modified thresholding scheme).
Given the initial partition $\Omega_{1}^{0},...,\Omega_{N}^{0}$, to obtain the
partition $\Omega_{1}^{n+1},...,\Omega_{N}^{n+1}$ at time $t=h(n+1)$ from the
partition $\Omega_{1}^{n},...,\Omega_{N}^{n}$ at time $t=hn$
1. (1)
For any $i=1,...,N$ form the convolutions
$\phi^{n}_{1,i}=G_{\gamma}^{h}*\mathbf{1}_{\Omega_{i}^{n}},\
\phi^{n}_{2,i}=G_{\beta}^{h}*\mathbf{1}_{\Omega_{i}^{n}}$
2. (2)
For any $i=1,...,N$ form the comparison functions
$\psi^{n}_{i}=\sum_{j\neq i}a_{ij}\phi^{n}_{1,j}+b_{ij}\phi^{n}_{2,j}.$
3. (3)
Thresholding step, define
$\Omega_{i}^{n+1}:=\left\\{x:\psi_{i}^{n}(x)<\min_{j\neq
i}\psi_{j}^{n}(x)\right\\}.$
We will assume the following:
(12) $\displaystyle\text{The coefficients}\ a_{ij},b_{ij}\ \text{satisfy the
strict triangle inequality.}$ (13) $\displaystyle\text{The matrices}\
\mathbb{A}\ \text{and}\ \mathbb{B}\ \text{are positive definite on}\
(1,...,1)^{\perp}.$
In particular, for $v\in(1,...,1)^{\perp}$ we can define norms
$\displaystyle|v|_{\mathbb{A}}^{2}=v\cdot\mathbb{A}v,\
|v|_{\mathbb{B}}^{2}=v\cdot\mathbb{B}v.$
Observe that condition (12) is always satisfied if we choose $\gamma$ large
and $\beta$ small provided the surface tensions and the inverse of the
mobilities satisfy the strict triangle inequality. Indeed define
$\displaystyle
m_{\sigma}=\min_{i,j,k}\\{\sigma_{ik}+\sigma_{kj}-\sigma_{ij}\\}\ \text{and}\
M_{\sigma}=\max_{i,j,k}\\{\sigma_{ik}+\sigma_{kj}-\sigma_{ij}\\},$
where $i,j,k$ range over all triples of distinct indices $1\leq i,j,k,\leq N$.
Define $m_{\frac{1}{\mu}}$ and $M_{\frac{1}{\mu}}$ in a similar way. Then a
computation shows that $a_{ij}$ and $b_{ij}$ satisfy the (strict) triangle
inequality if
(14) $\beta<\frac{m_{\sigma}}{M_{\frac{1}{\mu}}}\ \text{and}\
\gamma>\frac{M_{\sigma}}{m_{\frac{1}{\mu}}},$
which can always be achieved since $\gamma>\beta>0$ are arbitrary. For the
second condition (13), we have the following result of Salvador and Esedoğlu
[31].
###### Lemma 1.
Let the matrix $\sigma$ of the surface tensions and the matrix $\frac{1}{\mu}$
of the inverse mobilities (for the diagonal we set inverses to be zeros) be
negative definite on $(1,...,1)^{\perp}$. Let $\gamma>\beta$ be such that
(15) $\gamma>\frac{\min_{i=1,...,N-1}s_{i}}{\max_{i=1,...,N-1}m_{i}},\
\beta<\frac{\max_{i=1,...,N-1}s_{i}}{\min_{i=1,...,N-1}m_{i}}$
where $s_{i}$ and $m_{i}$ are the nonzero eigenvalues of $J\sigma J$ and
$J\frac{1}{\mu}J$ respectively, where the matrix $J$ has components
$J_{ij}=\delta_{ij}-\frac{1}{N}$. Then $\mathbb{A}$ and $\mathbb{B}$ are
positive definite on $(1,...,1)^{\perp}$.
In particular, if we choose $\gamma$ large enough and $\beta$ small enough,
condition (13) on the matrices $\mathbb{A},\mathbb{B}$ is satisfied provided
the matrices $\sigma$ and $\frac{1}{\mu}$ are negative definite on
$1,...,1)^{\perp}$. By a classical result of Schoenberg [33] this is the case
if and only if $\sqrt{\sigma_{ij}}$ and $1/\sqrt{\mu_{ij}}$ are $\ell^{2}$
embeddable. In particular, this holds for the choice of Read-Schockley surface
tensions and equal mobilities.
For $1\leq i\neq j\leq N$ define the kernels
(16) $K_{ij}(z)=a_{ij}G_{\gamma}(z)+b_{ij}G_{\beta}(z)$
where, for a given $t>0$, we define $G^{(d)}_{t}$ as the heat kernel in
$\mathbf{R}^{d}$, i.e.,
(17) $G^{(d)}_{t}(z)=\frac{e^{-\frac{|z|^{2}}{4t}}}{\sqrt{4\pi t}^{d}}.$
If the dimension $d$ is clear from the context, we suppress the superscript
$(d)$ in (17). We recall here some basic properties of the heat kernel.
(18) $\displaystyle G_{t}(z)>0\ \text{(non-negativity)},$ (19) $\displaystyle
G_{t}(z)=G_{t}(Rz)\ \forall R\in O(d)\ \text{(symmetry)},$ (20) $\displaystyle
G_{t}(z)=\frac{1}{\sqrt{t}^{d}}G_{1}\left(\frac{z}{\sqrt{t}}\right)\
\text{(scaling)},$ (21) $\displaystyle G_{t}*G_{s}=G_{t+s}\ \text{(semigroup
property)},$ (22) $\displaystyle
G_{t}^{(d)}(z)=\prod_{i=1}^{d}G_{t}^{(1)}(z_{i})\ \text{(factorization
property)}.$
We observe that the kernels $K_{ij}$ are positive, with positive Fourier
transform $\hat{K}_{ij}$ provided $\gamma>\max_{i,j}\sigma_{i,j}\mu_{i,j}$ and
$\beta<\min_{i,j}\sigma_{i,j}\mu_{i,j}$. In particular assuming
1. (1)
$\sigma_{ij}$ and $\frac{1}{\mu_{ij}}$ satisfy the strict triangle inequality,
2. (2)
$\sigma$ and $\frac{1}{\mu}$ are negative definite on $(1,...,1)^{\perp}$,
we can always achieve the conditions posed on $\mathbb{A},\mathbb{B}$ and the
positivity of the kernels $K_{ij}$ by choosing $\gamma$ large and $\beta$
small.
Given any $h>0$ we define the scaled kernels
(23) $K^{h}_{ij}(z)=\frac{1}{\sqrt{h}^{d}}K_{ij}(\frac{z}{\sqrt{h}}),$
then the first and the second step in Algorithm 1 may be compactly rewritten
as follows
$\displaystyle\psi_{i}^{n}=\sum_{j\neq
i}K_{ij}^{h}*\mathbf{1}_{\Omega_{j}^{n}}.$
For later use, we also introduce the kernel
(24) $K(z)=\frac{1}{2}G_{\gamma}(z)+\frac{1}{2}G_{\beta}(z).$
### 2.2. Connection to De Giorgi’s minimizing movements
The first observation is that Algorithm 1 has a minimizing movements
interpretation. To explain this, let us introduce the class
$\mathcal{A}:=\left\\{\chi:[0,1)^{d}\to\\{0,1\\}^{N}\biggr{\rvert}\
\sum_{k=1}^{N}\chi_{k}=1\right\\}$
and its relaxation
$\mathcal{M}:=\left\\{u:[0,1)^{d}\to[0,1]^{N}\biggr{\rvert}\
\sum_{k=1}^{N}u_{k}=1\right\\}.$
If $\chi\in\mathcal{A}\cap BV([0,1)^{d})^{N}$, then each of the sets
$\Omega_{i}:=\\{\chi_{i}=1\\}$ is a set of finite perimeter. We denote by
$\partial^{*}\Omega_{i}$ the reduced boundary of the set $\Omega_{i}$, and for
any pair $1\leq i\neq j\leq N$ we denote by
$\Sigma_{ij}:=\partial^{*}\Omega_{i}\cap\partial^{*}\Omega_{j}$ the interface
between the sets. For $u\in\mathcal{M}$ we define
(25) $E(u):=\begin{cases}\sum_{i,j}\sigma_{ij}\mathcal{H}^{d-1}(\Sigma_{ij})\
&\text{if}\ u\in\mathcal{A}\cap BV([0,1)^{d})^{N}\\\ +\infty\
&\text{otherwise}.\end{cases}$
For $h>0$ fixed we define the approximate energy $E_{h}$ for $u\in\mathcal{M}$
(26)
$E_{h}(u)=\sum_{i,j}\frac{1}{\sqrt{h}}\int_{[0,1)^{d}}u_{i}K_{ij}^{h}*u_{j}dx.$
For $u,v\in\mathcal{M}$ and $h>0$ we also define the distance
(27) $\displaystyle d_{h}^{2}(u,v)$
$\displaystyle:=-2hE_{h}(u-v)=-2\sqrt{h}\sum_{i,j}\int(u_{i}-v_{i})K_{ij}^{h}*(u_{j}-v_{j})dx$
$\displaystyle=2\sqrt{h}\int|G_{\gamma}^{h/2}*(u-v)|_{\mathbb{A}}^{2}+|G_{\beta}^{h/2}*(u-v)|_{\mathbb{B}}^{2}\
dx,$
where we used the semigroup property (21) and the symmetry (19) to derive the
last equality.
###### Lemma 2.
The pair $(\mathcal{M},d_{h})$ is a compact metric space. The function $E_{h}$
is continuous with respect to $d_{h}$. For every $1\leq i\leq N$ and
$n\in\mathbf{N}$ define $\chi_{i}^{n}=\mathbf{1}_{\Omega^{n}_{i}}$, where
$\Omega_{1}^{n},...,\Omega_{N}^{n}$ are obtained from
$\Omega_{1}^{n-1},...,\Omega_{N}^{n-1}$ by the thresholding scheme. Then
$\chi^{n}$ minimizes
(28) $\frac{1}{2h}d_{h}^{2}(u,\chi^{n-1})+E_{h}(u)\ \text{among all}\
u\in\mathcal{M}.$
###### Proof.
For $u,v\in\mathcal{M}$ definition (27) and the fact that $\mathbb{A}$ and
$\mathbb{B}$ are positive definite imply that $d_{h}$ is a distance on
$\mathcal{M}$. The fact that $(\mathcal{M},d_{h})$ is compact and $E_{h}$ is
continuous is just a consequence of the fact that $d_{h}$ metrizes the weak
convergence in $L^{2}$ on $\mathcal{M}$, the interested reader may find the
details of the reasoning in [19]. We are thus left with showing that
$\chi^{n}$ satisfies (28). For $u,v\in L^{2}([0,1)^{d})$ define
$(u,v)=\frac{1}{\sqrt{h}}\sum_{i,j}\int u_{i}K^{h}_{ij}*v_{j}dx,$
then by the symmetry (19) of the Gaussian kernel and by the symmetry of both
matrices $\mathbb{A},\mathbb{B}$ it is not hard to show that $(\cdot,\cdot)$
is symmetric. In particular we can write for any $u\in\mathcal{M}$
$\displaystyle\frac{1}{2h}d_{h}^{2}(u,\chi^{n-1})+E_{h}(u)$
$\displaystyle=-E_{h}(u-\chi^{n-1})+E_{h}(u)$
$\displaystyle=-(u-\chi^{n-1},u-\chi^{n-1})+(u,u)$
$\displaystyle=2(\chi^{n-1},u)-(\chi^{n-1},\chi^{n-1}).$
Thus (28) is equivalent to the fact that $\chi^{n}$ minimizes $(\chi^{n-1},u)$
among all $u\in\mathcal{M}$. Since by (2)
$(\chi^{n-1},u)=\int\sum_{i}u_{i}\psi^{n}_{i}dx,$
we see that $\chi^{n}$ minimizes the integrand pointwise, and thus it is a
minimizer for the functional.∎
The previous lemma allows us to apply the general theory of gradient flows in
[2] to this particular problem. We record the key statement for our purposes
in the following lemma.
###### Lemma 3.
Let $(\mathcal{M},d)$ be a compact metric space and
$E:\mathcal{M}\to\mathbf{R}$ be continuous. Given $\chi^{0}\in\mathcal{M}$ and
$h>0$ consider a sequence $\\{\chi^{n}\\}_{n\in\mathbf{N}}$ satisfying
(29) $\chi^{n}\ \text{minimizes}\ \frac{1}{2h}d^{2}(u,\chi^{n-1})+E(u)\
\text{among all}\ u\in\mathcal{M}.$
Then we have for all $t\in\mathbf{N}h$
(30) $\displaystyle\begin{aligned}
E(\chi(t))+\frac{1}{2}\int_{0}^{t}\left(\frac{1}{h^{2}}d^{2}(\chi(s+h),\chi(s))+|\partial
E|^{2}(u(s))\right)ds\leq E(\chi^{0}).\end{aligned}$
Here $\chi(t)$ is the piecewise constant interpolation, $u(t)$ is another
interpolation satisfying
(31) $\displaystyle\begin{aligned}
\int_{0}^{\infty}\frac{1}{2h^{2}}d^{2}(u(t),\chi(t))dt\leq
E(\chi^{0}),\end{aligned}$ (32) $\displaystyle\begin{aligned} \\\ &E(u(t))\leq
E(\chi(t))\ \text{for all}\ t\geq 0,\end{aligned}$
and $|\partial E|(u)$ is the metric slope defined by
(33) $|\partial E|(u):=\lim_{d(u,v)\to
0}\frac{(E(u)-E(v))_{+}}{d(u,v)}\in[0,\infty].$
## 3\. Statement of results
Our main result is the convergence of the modified thresholding scheme to a
weak notion of multiphase mean curvature flow. More precisely, given an
initial partition $\\{\Omega_{1}^{0},...,\Omega_{N}^{0}\\}$ of $[0,1)^{d}$
encoded by $\chi^{0}:[0,1)^{d}\to\\{0,1\\}^{N}$ such that
$\sum_{i}\chi_{i}^{0}=1$, define
$\chi^{h}:[0,1)^{d}\times\mathbf{R}\to\\{0,1\\}^{N}$ by setting
(34) $\displaystyle\begin{aligned} &\chi^{h}(t,x)=\chi^{0}(x)\ \text{for}\
t<h,\\\ &\chi^{h}(t,x)=\chi^{n}(x)\ \text{for}\ t\in[nh,(n+1)h)\ \text{for}\
n\in\mathbf{N}.\end{aligned}$
If $\chi^{0}$ is a function of bounded variation, we denote by
$\Sigma_{ij}^{0}:=\partial^{*}\Omega_{i}^{0}\cap\partial^{*}\Omega_{j}^{0}$.
Our main result is contained in the following theorem.
###### Theorem 1.
Given $\chi^{0}\in\mathcal{A}$ and such that $\nabla\chi^{0}$ is a bounded
measure and a sequence $h\downarrow 0$; let $\chi^{h}$ be defined by (34).
Assume that there exists $\chi:[0,1)^{d}\times(0,T)\to[0,1]^{N}$ such that
(35) $\chi^{h}\rightharpoonup\chi\ \text{in}\ L^{1}([0,1)^{d}\times(0,T)).$
Then $\chi\in\\{0,1\\}^{N}$ almost everywhere, $\sum_{i}\chi_{i}=1$ and
$\chi\in L^{1}((0,T),BV([0,1)^{d}))^{N}$.
If we assume that
(36) $\limsup_{h\downarrow
0}\int_{0}^{T}E_{h}(\chi^{h}(t))dt\leq\sum_{ij}\sigma_{ij}\int_{0}^{T}\mathcal{H}^{d-1}(\Sigma_{ij}(t))dt,$
then $\chi$ is a De Giorgi solution in the sense of Definition 1 below.
###### Definition 1.
Given $\chi^{0}\in\mathcal{A}$ and such that $\nabla\chi^{0}$ is a bounded
measure, a map $\chi:[0,1)^{d}\times(0,T)\to\\{0,1\\}^{N}$ such that
$\sum_{i}\chi_{i}=1$ and $\chi\in L^{1}((0,T),BV([0,1)^{d}))^{N}$ is called a
De Giorgi solution to the multiphase mean curvature flow with surface tensions
$\sigma_{ij}$ and mobilities $\mu_{ij}$ provided the following three facts
hold:
1. (1)
There exist $H_{ij}\in L^{2}(\mathcal{H}^{d-1}_{|\Sigma_{ij}}(dx)dt)$ which
are mean curvatures in the weak sense, i.e., such that for any test vector
field $\xi\in C^{\infty}_{c}([0,1)^{d}\times(0,T))^{d}$
(37)
$\displaystyle\sum_{i,j}\sigma_{ij}\int_{[0,1)^{d}\times(0,T)}(\nabla\cdot\xi-\nu_{ij}\cdot\nabla\xi\nu_{ij})\mathcal{H}^{d-1}_{|\Sigma_{ij}(t)}(dx)dt$
$\displaystyle=-\sum_{i,j}\sigma_{ij}\int_{[0,1)^{d}\times(0,T)}H_{ij}\nu_{ij}\cdot\xi\mathcal{H}^{d-1}_{|\Sigma_{ij}(t)}(dx)dt.$
2. (2)
There exist normal velocities $V_{ij}\in
L^{2}(\mathcal{H}^{d-1}_{|\Sigma_{ij}(t)}(dx)dt)$ with
$\displaystyle\begin{aligned}
\int_{[0,1)^{d}}\eta(t=0)\chi_{i}^{0}dx&+\int_{[0,1)^{d}\times(0,T)}\partial_{t}\eta\
\chi_{i}\ dxdt\\\ &+\sum_{k\neq i}\int_{[0,1)^{d}\times(0,T)}\eta V_{ik}\
\mathcal{H}^{d-1}_{|\Sigma_{ik}(t)}(dx)dt=0\end{aligned}$
for all $\eta\in C^{\infty}_{c}([0,1)^{d}\times[0,T))$.
3. (3)
De Giorgi’s inequality is satisfied, i.e.
(38) $\displaystyle\begin{aligned} &\limsup_{\tau\downarrow
0}\frac{1}{\tau}\sum_{i,j}\sigma_{ij}\int_{(T-\tau,T)}\mathcal{H}^{d-1}(\Sigma_{ij}(t))dt\\\
&+\frac{1}{2}\sum_{i,j}\int_{[0,1)^{d}\times(0,T)}\left(\frac{V_{ij}^{2}}{\mu_{ij}}+\sigma_{ij}^{2}\mu_{ij}H_{ij}^{2}\right)\mathcal{H}^{d-1}_{|\Sigma_{ij}(t)}(dx)dt\leq\sum_{i,j}\sigma_{ij}\mathcal{H}^{d-1}(\Sigma^{0}_{ij}).\end{aligned}$
###### Remark 1.
Observe that inequality (38) together with the definition of the weak mean
curvatures gives a notion of weak solution for the multiphase mean curvature
flow incorporating both the dynamics $V_{ij}=-\sigma_{ij}\mu_{ij}H_{ij}$ and
the Herring angle condition at triple junctions. Indeed if
$\chi:[0,1)^{d}\times(0,T)\to\\{0,1\\}^{N}$ with $\sum_{i}\chi_{i}(t)=1$ is
such that the sets $\Omega_{i}(t)=\\{\chi_{i}(\cdot,t)=1\\}$ meet along smooth
interfaces $\Sigma_{ij}:=\partial\Omega_{i}\cap\partial\Omega_{j}$ which
evolve smoothly and satisfy (37), (38) then
1. (1)
The Herring angle condition at triple junctions is satisfied. Indeed by the
divergence theorem on surfaces (see Theorem 11.8 and Remark 11.42 in [24]) for
any $\xi\in C^{\infty}_{c}(\times[0,1)^{d})^{d}$
$\displaystyle\int_{\Sigma_{ij}(t)}(\nabla\cdot\xi-\nu_{ij}\cdot\nabla\xi\nu_{ij})\
\mathcal{H}^{d-1}_{|\Sigma_{ij}(t)}(dx)=$
$\displaystyle-\int_{\Sigma_{ij}(t)}H_{ij}\nu_{ij}\mathcal{H}^{d-1}_{|\Sigma_{ij}(t)}(dx)$
$\displaystyle+\int_{\partial\Sigma_{ij}(t)}\xi\cdot\nu_{ij}\mathcal{H}^{d-2}(dx).$
Thus (37) and $H_{ij}\in L^{2}(\mathcal{H}^{d-1}_{|\Sigma_{ij}(t)}(dx)dt)$
imply that
$\displaystyle\sigma_{i_{1}i_{2}}\int_{\partial\Sigma_{i_{1}i_{2}}(t)}\xi\cdot\nu_{i_{1}i_{2}}\mathcal{H}^{d-2}(dx)$
$\displaystyle+\sigma_{i_{2}i_{3}}\int_{\partial\Sigma_{i_{2}i_{3}}(t)}\xi\cdot\nu_{i_{2}i_{3}}\mathcal{H}^{d-2}(dx)$
$\displaystyle+\sigma_{i_{3}i_{1}}\int_{\partial\Sigma_{i_{3}i_{1}}(t)}\xi\cdot\nu_{i_{3}i_{1}}\mathcal{H}^{d-2}(dx)=0,$
which forces
$\sigma_{i_{1}i_{2}}\nu_{i_{1}i_{2}}+\sigma_{i_{2}i_{3}}\nu_{i_{2}i_{3}}+\sigma_{i_{3}i_{1}}\nu_{i_{3}i_{1}}=0$
at triple junctions.
2. (2)
We have $V_{ij}=-\sigma_{ij}\mu_{ij}H_{ij}$ on $\Sigma_{ij}(t)$. Indeed in the
smooth case inequality (38) reduces to
$\displaystyle\sum_{i,j}\sigma_{ij}\int_{(0,T)}\frac{d}{dt}\mathcal{H}^{d-1}(\Sigma_{ij}(t))dt$
$\displaystyle+\frac{1}{2}\sum_{i,j}\int_{[0,1)^{d}\times(0,T)}\left(\frac{V_{ij}^{2}}{\mu_{ij}}+\sigma_{ij}^{2}\mu_{ij}H_{ij}^{2}\right)\mathcal{H}^{d-1}_{|\Sigma_{ij}(t)}(dx)dt\leq
0.$
If one recalls that
$\frac{d}{dt}\mathcal{H}^{d-1}(\Sigma_{ij}(t))=\int_{[0,1)^{d}}V_{ij}H_{ij}\mathcal{H}^{d-1}_{|\Sigma_{ij}(t)}(dx)$,
after completing the square we arrive at
$\displaystyle\sum_{i,j}\int_{[0,1)^{d}\times(0,T)}\left(\frac{V_{ij}}{\sqrt{\mu_{ij}}}+\sigma_{ij}\mu_{ij}H_{ij}\right)^{2}\mathcal{H}^{d-1}_{|\Sigma_{ij}(t)}(dx)dt\leq
0,$
which implies $V_{ij}=-\sigma_{ij}\mu_{ij}H_{ij}$.
The following lemma establishes, next to a compactness statement, that our
convergence can be localized in the space and time variables $x$ and $t$, but
also in the variable $z$ appearing in the convolution.
###### Lemma 4.
1. (1)
Let $\\{\chi^{h}\\}_{h\downarrow 0}$ be a sequence of $\\{0,1\\}^{N}$-valued
functions on $(0,T)\times[0,1)^{d}$ that satisfies
(39) $\limsup_{h\downarrow
0}\left(\operatorname{esssup}_{t\in(0,T)}E_{h}(\chi_{h}(t))+\int_{0}^{T}\frac{1}{2h^{2}}d_{h}^{2}(\chi_{h}(t),\chi_{h}(t-h))dt\right)<\infty$
and that is piecewise constant in time in the sense of (34). Such a sequence
is compact in $L^{1}([0,1)^{d}\times(0,T))^{N}$ and any weak limit $\chi$ is
such that $\chi\in L^{1}((0,T),BV([0,1)^{d}))^{N}$ with
(40)
$\sum_{i,j}\sigma_{ij}\int_{0}^{T}\mathcal{H}^{d-1}(\Sigma_{ij}(t))dt\leq\liminf_{h\downarrow
0}\int_{0}^{T}E_{h}(\chi_{h}(t))dt.$
2. (2)
Assume that $u^{h}$ is a sequence of $[0,1]^{N}$-valued functions with
$\sum_{i}u_{i}^{h}=1$ such that (36) holds (with $\chi^{h}$ replaced by
$u^{h}$) and such that $u^{h}\to\chi$ in $L^{1}([0,1)^{d}\times(0,T))^{N}$
holds. Assume also that
(41) $\limsup_{h\downarrow
0}\operatorname{esssup}_{t\in(0,T)}E_{h}(u^{h}(t))<\infty.$
Then as measures on $\mathbf{R}^{d}\times[0,1)^{d}\times(0,T)$ we have the
following weak convergences for any $i\neq j$
(42) $\displaystyle\begin{aligned}
\frac{K_{ij}(z)}{\sqrt{h}}u_{i}^{h}(x,t)&u_{j}^{h}(x-\sqrt{h}z,t)dxdtdz\\\
&\rightharpoonup K_{ij}(z)(\nu_{ij}(x,t)\cdot
z)_{+}\mathcal{H}^{d-1}_{|\Sigma_{ij}(t)}(dx)dtdz.\end{aligned}$ (43)
$\displaystyle\begin{aligned}
\frac{K_{ij}(z)}{\sqrt{h}}u_{i}^{h}(x-\sqrt{h}z,t)&u_{j}^{h}(x,t)dxdtdz\\\
&\rightharpoonup K_{ij}(z)(\nu_{ij}(x,t)\cdot
z)_{-}\mathcal{H}^{d-1}_{|\Sigma_{ij}(t)}(dx)dtdz.\end{aligned}$
Here the convergence may be tested also with continuous functions which have
polynomial growth in $z\in\mathbf{R}^{d}$.
The next proposition is the main ingredient in the proof of Theorem 1. It
establishes the sharp lower bound on the distance-term.
###### Proposition 1.
Suppose that (35) and the conclusion of Lemma 4 (2) hold. Assume also that the
left hand side of (45) is finite. Then for every $1\leq k\leq N$ there exists
$V_{k}\in L^{2}(|\nabla\chi_{k}|dt)$ such that
(44) $\partial_{t}\chi_{k}=V_{k}|\nabla\chi_{k}|dt$
in the sense of distributions. Given $i\neq j$, it holds that
$V_{i}(x,t)=-V_{j}(x,t)$ on $\Sigma_{ij}(t)$ and if we define
$V_{ij}(x,t):=V_{i}(x,t)\nu_{ij}(x,t)|_{\Sigma_{ij}(t)}$ then we have
(45) $\liminf_{h\downarrow
0}\int_{0}^{T}\frac{1}{h^{2}}d_{h}^{2}(\chi^{h}(t),\chi^{h}(t-h))dt\geq\sum_{i,j}\frac{1}{\mu_{ij}}\int_{0}^{T}\int_{\Sigma_{ij}(t)}|V_{ij}(x,t)|^{2}\mathcal{H}^{d-1}(dx)dt.$
The final ingredient is the analogous sharp lower bound for the metric slope.
###### Proposition 2.
Suppose that the conclusion of Lemma 4 (2) holds and that (35) holds with
$\chi^{h}$ replaced by $u^{h}$. Then for any $i\neq j$ there exists a mean
curvature $H_{ij}\in L^{2}(\mathcal{H}^{d-1}_{\Sigma_{ij}(t)}(dx)dt)$ in the
sense of (37). Moreover the following inequality is true:
$\liminf_{h\downarrow 0}\int_{0}^{T}|\partial
E_{h}|^{2}(u_{h}(t))dt\geq\sum_{i,j}\mu_{ij}\sigma_{ij}^{2}\int_{0}^{T}\int_{\Sigma_{ij}(t)}|H_{ij}(x,t)|^{2}\mathcal{H}^{d-1}(dx)dt.$
## 4\. Construction of suitable partitions of unity
In the sequel we will frequently want to localize on one of the interfaces. To
do so, we need to construct a suitable family of balls on wich the behavior of
the flow is split into two majority phases and several minority phases.
Hereafter we will ignore the time variable and consider a map
$\chi:[0,1)^{d}\to\\{0,1\\}^{N}$ such that $\chi\in
BV([0,1)^{d},\mathbf{R}^{N})$, $\sum_{k}\chi_{k}=1$. Given $1\leq i<j\leq N$
we denote by $\partial^{*}\Omega_{i}$ the reduced boundary of the set
$\\{\chi_{i}=1\\}$ and by
$\Sigma_{ij}=\partial^{*}\Omega_{i}\cap\partial^{*}\Omega_{j}$ the interface
between phase $i$ and phase $j$. Given a real number $r>0$ and a natural
number $n\in\mathbf{N}$ we define
(46) $\mathcal{F}_{n}^{r}:=\left\\{B(x,nr\sqrt{d}):\ x\in
r\mathbf{Z}^{d}\right\\}$
where the balls appearing in the definition are intended to be open. Observe
that for any $n\geq 2$ and any $r>0$ the collection of balls in
$\mathcal{F}_{n}^{r}$ is a covering of $\mathbf{R}^{d}$ with the property that
any point $x\in\mathbf{R}^{d}$ lies in at most $c(n,d)$ distinct balls
belonging to $\mathcal{F}^{r}_{n}$, where $0<c(n,d)\leq(2n)^{d}$ is a constant
that depends on $n,d$ but not on $r$. Given numbers $1\leq l\neq p\leq N$ we
define
(47) $\mathcal{E}^{r}:=\left\\{B\in\mathcal{F}^{r}_{2}:\
B\cap\Sigma_{lp}\neq\emptyset,\ \frac{\mathcal{H}^{d-1}(\Sigma_{ij}\cap
2B)}{\omega_{d-1}(4r)^{d-1}}\leq\frac{1}{2^{d}},\
\\{i,j\\}\neq\\{l,p\\}\right\\}.$
Here $2B$ denotes the ball with center given by the center of $B$ and twice
its radius. Given $l,p$ as above, denote by $\\{B_{m}^{r}\\}_{m\in\mathbf{N}}$
an enumeration of $\mathcal{E}^{r}$ and by $\\{\rho_{m}\\}_{m\in\mathbf{N}}$ a
smooth partition of unity subordinate to $\\{B_{m}^{r}\\}_{m\in\mathbf{N}}$.
###### Lemma 5.
Fix $1\leq l\neq p\leq N$. With the above construction the following two
properties hold.
1. (1)
For any $1\leq i\neq j\leq N$, $\\{i,j\\}\neq\\{l,p\\}$ and any $\eta\in
L^{1}(\mathcal{H}^{d-1}_{|\Sigma_{ij}})$
(48) $\lim_{r\downarrow
0}\sum_{m\in\mathbf{N}}\int_{B_{m}^{r}}\eta\mathcal{H}^{d-1}_{|\Sigma_{ij}}(dx)=0.$
2. (2)
For any $\eta\in L^{1}(\mathcal{H}^{d-1}_{|\Sigma_{lp}})$
(49) $\lim_{r\downarrow
0}\sum_{m\in\mathbf{N}}\int\rho_{m}\eta\mathcal{H}^{d-1}_{|\Sigma_{lp}}(dx)=\int\eta\mathcal{H}^{d-1}_{|\Sigma_{lp}}(dx).$
## 5\. Proofs
###### Proof of Theorem 1.
By Lemma 2, we can apply Lemma 3 on the metric space $(\mathcal{M},d_{h})$ so
that we get inequality (30) with $(E,d,\chi,u)=(E_{h},d_{h},\chi^{h},u^{h})$.
Our first observation is that
(50) $\lim_{h\downarrow
0}E_{h}(\chi^{0})=\sum_{i,j}\sigma_{ij}\mathcal{H}^{d-1}(\Sigma_{ij}^{0}),$
which follows from the consistency, cf. Lemma 7 in the Appendix. Inequality
(30) then yields that the sequence $\chi^{h}$ satisfies (39), so that Lemma 4
(1) applies to get that $\chi\in L^{1}((0,T),BV([0,1)^{d}))^{N}$,
$\chi\in\\{0,1\\}^{N}$ a.e., $\sum_{i}\chi_{i}=1$ and, after extracting a
subsequence, $\chi^{h}\to\chi$ in $L^{2}([0,1)^{d}\times(0,T))^{N}$. We claim
that this implies $u^{h}\to\chi$ in $L^{2}([0,1)^{d}\times(0,T))^{N}$. To see
this, observe that (31) implies
(51) $\displaystyle\begin{aligned}
hE_{h}(\chi^{0})&\geq-\int_{0}^{T}E_{h}(u_{h}(t)-\chi_{h}(t))dt\\\ &\geq
C\frac{1}{\sqrt{h}}\sum_{i=1}^{N}\left(\int|G_{\gamma}^{h/2}*(u_{i}^{h}-\chi_{i}^{h})|^{2}dxdt+\int|G_{\beta}^{h/2}*(u_{i}^{h}-\chi_{i}^{h})|^{2}dxdt\right)\end{aligned}$
where $C$ is a constant which depends on $N,\mathbb{A},\mathbb{B}$ but not on
$h$ and comes from the fact that all norms on $(1,...,1)^{\perp}$ are
comparable. Inequality (51) clearly implies that $K^{h}*u_{h}-K^{h}*\chi_{h}$
converges to zero in $L^{2}$. Observe that inequality (LABEL:varEnergy) in
particular yields (41) with $\chi^{h}$ replaced by $u^{h}$. Recalling (152) in
the Appendix, we learn that $u^{h}-\chi^{h}$ converges to zero in $L^{2}$.
This implies that we can apply Lemma 4 (2) both to the sequence $u^{h}$ and
the sequence $\chi^{h}$. In particular, we may apply Proposition 1 for
$\chi^{h}$ and Proposition 2 for $u^{h}$. Now the proof follows the same
strategy as the one in the two-phase case in [19]. For the sake of
completeness, we sketch the argument here. First of all, Lemma 3 gives
inequality (30) for $(E_{h},d_{h},\chi^{h},u^{h})$, namely for
$n\in\mathbf{N}$
(52) $\displaystyle\rho(nh)\leq E_{h}(\chi^{0}),$
where we set
$\rho(t)=E_{h}(\chi^{h}(t))+\frac{1}{2}\int_{0}^{t}\left(\frac{1}{h^{2}}d_{h}^{2}(\chi^{h}(s+h),\chi^{h}(s))+|\partial
E_{h}(u^{h}(s))|^{s}\right)ds$. Multiplying (52) by $\eta(nh)-\eta((n+1)h)$
for some non-increasing function $\eta\in C_{c}([0,T))$ we get
$-\int\frac{d\eta}{dt}\rho
dt\leq(\eta(0)+h\sup\left|\frac{d\eta}{dt}\right|)E_{h}(\chi^{0})$. As test
function $\eta$, we now choose
$\eta(t)=\max\\{\min\\{\frac{T-t}{\tau},1\\},0\\}$ and obtain
(53) $\displaystyle\frac{1}{\tau}\int_{T-\tau}^{T}E_{h}(\chi^{h}(t))dt$
$\displaystyle+\frac{1}{2}\int_{0}^{T-\tau}\left(\frac{1}{h^{2}}d_{h}^{2}(\chi^{h}(t),\chi^{h}(t-h))+|\partial
E_{h}(u^{h}(t))|^{2}\right)dt\leq(1+\frac{h}{\tau})E_{h}(\chi^{0}).$
Now it remains to pass to the limit as $h\downarrow 0$: to get (38) from
inequality (53) one uses the lower semicontinuity (40) for the first left hand
side term, the sharp bound (45) for the second left hand side term, the bound
(2) for the last left hand side term and finally one uses the consistency
Lemma 7 in the Appendix to treat the right hand side term. To get (38) it
remains to pass to the limit in $\tau\downarrow 0$.
∎
###### Proof of Lemma 4.
Argument for (1). For the compactness, the arguments in [19] adapt to this
setting with minor changes. The first observation is that, by inequality (152)
in the Appendix, one needs to prove compactness in
$L^{2}([0,1)^{d}\times(0,T))^{N}$ of $\\{K^{h}*\chi^{h}\\}_{h\downarrow 0}$.
For this, one just needs a modulus of continuity in time. I.e. it is
sufficient to prove that there exists a constant $C>0$ independent of $h$ such
that $I_{h}(s)\leq C\sqrt{s}$, where
$I_{h}(s)=\int_{(s,T)\times[0,1)^{d}}|\chi_{h}(x,t)-\chi_{h}(x,t-s)|^{2}dxdt.$
This is can be done applying word by word the argument in [19] once we show
the following: for any pair $\chi,\chi^{\prime}\in\mathcal{A}$ of admissible
functions, we have
(54)
$\int|\chi-\chi^{\prime}|dx\leq\frac{C}{\sqrt{h}}d_{h}^{2}(\chi,\chi^{\prime})+C\sqrt{h}\left(E_{h}(\chi)+E_{h}(\chi^{\prime})\right).$
Here the constant $C$ depends on $N,\mathbb{A},\mathbb{B}$ but not on $h$.
To prove (54) we proceed as follows: let $\mathbb{S}\in\mathbf{R}^{N\times N}$
be a symmetric matrix which is positive definite on $(1,...,1)^{\perp}$. Since
any two norms on a finite dimensional space are comparable, there exists a
constant $C>0$ depending on $\mathbb{S}$ and $N$ such that
(55) $|\chi-\chi^{\prime}|\leq|\chi-\chi^{\prime}|^{2}\leq
C|\chi-\chi^{\prime}|^{2}_{\mathbb{S}}$
where $|\cdot|_{\mathbb{S}}$ denotes the norm induced by $\mathbb{S}$. For a
function $u\in\mathcal{M}$ write $(\tilde{K}^{h}*)u_{h}$ for the function
$\left((\tilde{K}^{h}*)u_{h}\right)_{i}=\sum_{j\neq i}K_{ij}^{h}*u_{h}^{j}.$
Then we calculate
(56) $\displaystyle\begin{aligned}
|\chi-\chi^{\prime}|^{2}_{\mathbb{S}}=-(\chi-\chi^{\prime})\cdot(\tilde{K}^{h}*)(\chi-\chi^{\prime})+(\chi-\chi^{\prime})(\mathbb{S}+(\tilde{K}^{h}*))(\chi-\chi^{\prime}).\end{aligned}$
Select $\mathbb{S}=(s_{ij})$ where $s_{ij}=-\int K_{ij}(z)dz$. Then, by our
assumption (13) $\mathbb{S}$ is positive definite on $(1,...,1)^{\perp}$ and
after integration on $[0,1)^{d}$ identity (56) becomes
$\int|\chi-\chi^{\prime}|^{2}_{\mathbb{S}}dx=\frac{1}{2\sqrt{h}}d_{h}^{2}(\chi,\chi^{\prime})+\int(\chi-\chi^{\prime})(\mathbb{S}+(\tilde{K}^{h}*))(\chi-\chi^{\prime})dx.$
We now proceed to estimate the integral on the right hand side. By the choice
of $\mathbb{S}$ and Jensen’s inequality we have
(57)
$\displaystyle\int(\chi-\chi^{\prime})(\mathbb{S}+(\tilde{K}^{h}*))(\chi-\chi^{\prime})dx\leq
C\int|\mathbb{S}+(\tilde{K}^{h}*))(\chi-\chi^{\prime})|dx$ $\displaystyle\leq
C\sum_{i,j}\int
K^{h}_{ij}(z)|(\chi_{j}-\chi_{j}^{\prime})(x-z)-(\chi_{j}-\chi_{j}^{\prime})(x)|dxdz.$
Using the triangle inequality and (150) in the Appendix we can estimate the
right hand side to obtain the following inequality
(58) $\displaystyle\begin{aligned}
\int(\chi-\chi^{\prime})(\mathbb{S}+(\tilde{K}^{h}*))(\chi-\chi^{\prime})dx\end{aligned}$
$\displaystyle\begin{aligned} \leq C\sum_{i,j}&\left(\sum_{k\neq j}\int
K^{h}_{ij}(z)\chi_{j}(x-z)\chi_{k}(x)dxdz\right.\\\ &\left.+\sum_{k\neq j}\int
K^{h}_{ij}(z)\chi_{j}(x)\chi_{k}(x-z)dxdz\right.\\\ &\left.+\sum_{k\neq j}\int
K^{h}_{ij}(z)\chi^{\prime}_{j}(x-z)\chi^{\prime}_{k}(x)dxdz\right.\\\
&\left.+\sum_{k\neq j}\int
K^{h}_{ij}(z)\chi^{\prime}_{j}(x)\chi^{\prime}_{k}(x-z)dxdz\right).\end{aligned}$
Observing that there is a constant $C>0$ such that $K_{ij}\leq CK_{jk}$ we
conclude that
$\int(\chi-\chi^{\prime})(\mathbb{S}+(\tilde{K}^{h}*))(\chi-\chi^{\prime})dx\leq
C\sqrt{h}\left(E_{h}(\chi)+E_{h}(\chi^{\prime})\right).$
This proves (54) and closes the argument for the compactness.
We also have to prove (40), but this follows from (42) with $u^{h}$ replaced
by $\chi^{h}$ once we have shown that the limit $\chi$ is such that
$|\nabla\chi|$ is a bounded measure, equiintegrable in time. This can be done
with an argument similar to the one used in [19] for the two-phase case.
Observe that this only requires the weaker assumption (41).
Argument for (2). As mentioned in the previous paragraph, we already know that
the limit $\chi$ is such that $|\nabla\chi|$ is a bounded measure,
equiintegrable in time. We will prove (42). Then (43) easily follows by
recalling that $\nu_{ij}=-\nu_{ji}$. A standard argument (to be found in [19])
which relies on the exponential decay of the kernel yields the fact that we
can test convergences (42) with functions with at most polynomial growth in
$z$ provided we already have the result for bounded and continuous test
functions, thus we focus on this case.
Let $\xi\in C_{b}(\mathbf{R}^{d}\times[0,1)^{d}\times(0,T))$ be a bounded and
continuous function. To show (42) we aim at showing that
(59) $\displaystyle\begin{aligned} \lim_{h\downarrow
0}\int\xi(z,x,t)\frac{K_{ij}(z)}{\sqrt{h}}&u_{i}^{h}(x,t)u_{j}^{h}(x-\sqrt{h}z,t)dxdtdz\\\
&=\int\xi(z,x,t)K_{ij}(z)(\nu_{ij}(x,t)\cdot
z)_{+}\mathcal{H}^{d-1}_{\Sigma_{ij}(t)}(dx)dtdz.\end{aligned}$
Upon splitting $\xi$ into the positive and the negative part, by linearity we
may assume that $0\leq\xi\leq 1$. We can split (59) into the local lower bound
(60) $\displaystyle\begin{aligned} \liminf_{h\downarrow
0}\int\xi(z,x,t)\frac{K_{ij}(z)}{\sqrt{h}}&u_{i}^{h}(x,t)u_{j}^{h}(x-\sqrt{h}z,t)dzdxdt\\\
&\geq\int\xi(z,x,t)K_{ij}(z)(\nu_{ij}(x,t)\cdot
z)_{+}\mathcal{H}^{d-1}_{\Sigma_{ij}(t)}(dx)dtdz.\end{aligned}$
and the global upper bound
(61) $\displaystyle\begin{aligned} \liminf_{h\downarrow
0}\int\frac{K_{ij}(z)}{\sqrt{h}}&u_{i}^{h}(x,t)u_{j}^{h}(x-\sqrt{h}z,t)dzdxdt\\\
&\leq\int K_{ij}(z)(\nu_{ij}(x,t)\cdot
z)_{+}\mathcal{H}^{d-1}_{\Sigma_{ij}(t)}(dx)dtdz.\end{aligned}$
Indeed we can recover the limsup inequality in (59) by splitting
$\xi=1-(1-\xi)$ and applying the local lower bound (60) to $1-\xi$.
We first concentrate on the local lower bounds in the case where $u^{h}=\chi$,
namely we will show
(62) $\displaystyle\begin{aligned} \liminf_{h\downarrow
0}\int\xi(z,x,t)\frac{K_{ij}(z)}{\sqrt{h}}&\chi_{i}(x,t)\chi_{j}(x-\sqrt{h}z,t)dzdxdt\\\
&\geq\int\xi(z,x,t)K_{ij}(z)(\nu_{ij}(x,t)\cdot
z)_{+}\mathcal{H}^{d-1}_{\Sigma_{ij}(t)}(dx)dtdz.\end{aligned}$
By Fatou’s lemma the claim is reduced to showing that for a.e. point $t$ in
time and every $z\in\mathbf{R}^{d}$
(63) $\displaystyle\liminf_{h\downarrow
0}\int\xi(z,x,t)\frac{K_{ij}(z)}{\sqrt{h}}$
$\displaystyle\chi^{i}(x,t)\chi^{j}(x-\sqrt{h}z,t)dx$
$\displaystyle\geq\int\xi(z,x,t)K_{ij}(z)(\nu_{ij}(x,t)\cdot
z)_{+}\mathcal{H}^{d-1}_{\Sigma_{ij}(t)}(dx).$
Fix a point $t$ such that $\chi(\cdot,t)\in BV([0,1)^{d},\\{0,1\\}^{N})$ and
any $z\in\mathbf{R}^{d}$. In the sequel, we will drop those variables, so
$\chi(x)=\chi(x,t)$, $\xi(x)=\xi(z,x,t)$. By approximation we may assume that
$\xi\in C^{\infty}([0,1)^{d})$. Let $\rho_{m}$ be a partition of unity
obtained by applying the construction of Section 4 to the function $\chi(x)$
on the interface $\Sigma_{ij}$. Then by Lemma 5 we have
$\displaystyle\begin{aligned} &\int\xi(x)(\nu_{ij}(x)\cdot
z)_{+}\mathcal{H}^{d-1}_{|\Sigma_{ij}}(dx)\\\ &=\lim_{r\downarrow
0}\left(\sum_{m\in\mathbf{N}}\int\rho_{mij}(x)\xi(x)(\nu_{ij}(x)\cdot
z)_{+}\mathcal{H}^{d-1}_{|\Sigma_{ij}}(dx)\right)\end{aligned}$
$\displaystyle\begin{aligned} =\lim_{r\downarrow
0}\sum_{m\in\mathbf{N}}&\left(\int\rho_{mij}(x)\xi(x)(\nu_{i}(x)\cdot
z)_{+}\mathcal{H}^{d-1}_{|\partial^{*}\Omega_{i}}(dx)\right.\\\
&\left.-\sum_{k\neq i,j}\int\rho_{mij}(x)\xi(x)(\nu_{ij}(x)\cdot
z)_{+}\mathcal{H}^{d-1}_{|\Sigma_{ik}}(dx)\right)\end{aligned}$ (64)
$\displaystyle\begin{aligned} =\lim_{r\downarrow
0}\sum_{m\in\mathbf{N}}\int\rho_{mij}(x)\xi(x)(\nu_{i}(x)\cdot
z)_{+}\mathcal{H}^{d-1}_{|\partial^{*}\Omega_{i}}(dx).\end{aligned}$
We now focus on estimating the argument of the last limit. Observe that
$(\nu_{ij}(x)\cdot
z)_{+}\mathcal{H}^{d-1}_{|\partial^{*}\Omega_{i}}(dx)=(\partial_{z}\chi_{i})_{+}$,
thus by definition of positive part of a measure, given $\epsilon>0$ we can
select, for any $m\in\mathbf{N}$, a function $\tilde{\xi}_{m}\in
C^{1}_{c}(B_{m})$ such that $0\leq\tilde{\xi}_{m}\leq 1$ and such that
(65)
$\int\rho_{mij}\xi\tilde{\xi}_{m}\partial_{z}\chi_{i}+2^{-m}\epsilon\geq\int\rho_{mij}\xi(\nu_{i}\cdot
z)_{+}\mathcal{H}^{d-1}_{|\partial^{*}\Omega_{i}}(dx).$
Let $\eta_{m}:=\rho_{mij}\xi\tilde{\xi}_{m}\in C^{1}_{c}(B_{m})$, then
$\displaystyle\begin{aligned}
\int\eta_{m}\partial_{z}\chi_{i}=-\int\partial_{z}\eta_{m}\chi_{i}dx\end{aligned}$
$\displaystyle\begin{aligned} =\lim_{h\downarrow
0}\int\frac{\eta_{m}(x+\sqrt{h}z)-\eta(x)}{\sqrt{h}}\chi_{i}(x)dx\end{aligned}$
$\displaystyle\begin{aligned} =\lim_{h\downarrow
0}\int\eta_{m}(x)\frac{\chi_{i}(x)-\chi_{i}(x-\sqrt{h}z)}{\sqrt{h}}dx\end{aligned}$
$\displaystyle\begin{aligned} \leq\liminf_{h\downarrow 0}\sum_{k\neq
i}\int\eta_{m}(x)\frac{\chi_{i}(x)\chi_{k}(x-\sqrt{h}z)}{\sqrt{h}}dx\end{aligned}$
$\displaystyle\begin{aligned} \leq\liminf_{h\downarrow
0}&\int\eta_{m}(x)\frac{\chi_{i}(x)\chi_{j}(x-\sqrt{h}z)}{\sqrt{h}}dx\\\
&+\limsup_{h\downarrow 0}\sum_{k\neq
i,j}\int\eta_{m}(x)\frac{\chi_{i}(x)\chi_{k}(x-\sqrt{h}z)}{\sqrt{h}}dx\end{aligned}$
$\displaystyle\begin{aligned} \leq\liminf_{h\downarrow
0}&\int\eta_{m}(x)\frac{\chi_{i}(x)\chi_{j}(x-\sqrt{h}z)}{\sqrt{h}}dx\\\
&+\sum_{k\neq i,j}\limsup_{h\downarrow
0}\int\eta_{m}(x)\frac{\chi_{i}(x)\chi_{k}(x-\sqrt{h}z)}{\sqrt{h}}dx.\end{aligned}$
Observe that for each $m\in\mathbf{N}$, using also the consistency Lemma 7
$\displaystyle\begin{aligned} \limsup_{h\downarrow
0}\int\eta_{m}(x)\frac{\chi_{i}(x)\chi_{k}(x-\sqrt{h}z)}{\sqrt{h}}dx\end{aligned}$
$\displaystyle\begin{aligned} \leq\limsup_{h\downarrow
0}\int\eta_{m}(x)\frac{\chi_{i}(x)\chi_{k}(x-\sqrt{h}z)+\chi_{i}(x-\sqrt{h}z)\chi_{j}(x)}{\sqrt{h}}dx\end{aligned}$
$\displaystyle\begin{aligned} =\int\eta_{m}(x)|\nu_{ik}(x)\cdot
z|\mathcal{H}^{d-1}_{|\Sigma_{ik}}(dx)\end{aligned}$
$\displaystyle\begin{aligned}
\leq|z|\mathcal{H}^{d-1}(B_{mij}^{r}\cap\Sigma_{ik})\end{aligned}$
Thus we obtain
$\displaystyle\begin{aligned}
\int\eta_{m}\partial_{z}\chi_{i}\leq\liminf_{h\downarrow
0}&\int\eta_{m}(x)\frac{\chi_{i}(x)\chi_{j}(x-\sqrt{h}z)}{\sqrt{h}}dx\\\
&+\sum_{k\neq
i,j}|z|\mathcal{H}^{d-1}(B_{mij}^{r}\cap\Sigma_{ik})\end{aligned}$
Inserting back into (64), recalling also Lemma 5 and the inequality (65),
using Fatou’s lemma, the fact that $\rho_{mij}$ is a partition of unity and
that $0\leq\tilde{\xi}_{m}\leq 1$ we obtain that
$\displaystyle\int\xi(x)(\nu_{ij}(x)\cdot
z)_{+}\mathcal{H}^{d-1}_{|\Sigma_{ij}}(dx)\leq\liminf_{h\downarrow 0}$
$\displaystyle\int\xi(x)\frac{\chi_{i}(x)\chi_{j}(x-\sqrt{h}z)}{\sqrt{h}}dx+\epsilon$
and (62) follows letting $\epsilon$ go to zero. To derive inequality (60) we
just apply Lemma 9 in the Appendix.
To get the upper bound (61) we argue as follows. First of all recall
Assumption (36) which says
(66) $\int_{0}^{T}E_{h}(u^{h}(t))dt\to\int_{0}^{T}E(\chi(t))dt.$
Now, if we define
$e^{ij}_{h}(u^{h})=\frac{1}{\sqrt{h}}\int_{0}^{T}\int
u_{i}^{h}(t)K_{ij}^{h}*u_{j}^{h}(t)dxdt$
we have that by (60) $\liminf_{h\downarrow 0}e_{h}^{ij}(u_{h})\geq
e^{ij}(\chi)$, where $e^{ij}(\chi)$ is defined in the obvious way. Assume that
there exists a pair $i,j$ such that $\limsup_{h\downarrow
0}e_{h}^{ij}(u^{h})>e^{ij}(\chi)$, then
$\displaystyle\begin{aligned} \int_{0}^{T}E(\chi(t))dt&=\lim_{h\downarrow
0}\int_{0}^{T}E_{h}(u^{h}(t))dt\\\ &=\limsup_{h\downarrow
0}\int_{0}^{T}E_{h}(u^{h}(t))dt\\\
&\geq\sum_{(l,p)\neq(i,j)}\liminf_{h\downarrow
0}e_{h}^{lp}(u^{h})+\limsup_{h\downarrow 0}e_{h}^{ij}(u^{h})\\\
&>\int_{0}^{T}E(\chi(t))dt\end{aligned}$
which is a contradiction. Thus we have proved (61). ∎
###### Proof of Proposition 1.
Since we assume that the left hand side of (45) is finite, in view of (27),
upon passing to a subsequence we may assume that, in the sense of
distributions, the limit
(67) $\lim_{h\downarrow
0}\frac{1}{h\sqrt{h}}\left(\left|G^{h/2}_{\gamma}*(\chi-\chi(\cdot-h))\right|_{\mathbb{A}}^{2}+\left|G^{h/2}_{\beta}*(\chi-\chi(\cdot-h))\right|_{\mathbb{B}}^{2}\right)=\omega$
exists as a finite positive measure on $[0,1)^{d}\times(0,T)$. Here we
indicated with $\chi_{l}^{h}(\cdot-h)$ the time shift of function
$\chi_{l}^{h}$. We denote by $\tau$ a small fraction of the characteristic
spatial scale, namely $\tau=\alpha\sqrt{h}$ for some $\alpha>0$, which we
think as a small number. Given $1\leq l\leq N$ we define
(68) $\delta\chi^{h}_{l}:=\chi^{h}_{l}-\chi^{h}_{l}(\cdot-\tau).$
We divide the proof into two parts: first we show that the normal velocities
exist, and afterwards we prove the sharp bound. But first, let us state two
distributional inequalities that will be used later. Namely
* •
In a distributional sense it holds that
(69) $\limsup_{h\downarrow 0}-\frac{1}{\sqrt{h}}\sum_{i\neq
j}\delta\chi_{i}K_{ij}^{h}*\delta\chi_{j}\leq\alpha^{2}\omega.$
* •
There exists a constant $C>0$ such that for any $1\leq i\leq N$ and any
$\theta\in\\{\gamma,\beta\\}$ in a distributional sense it holds that
(70) $\limsup_{h\downarrow
0}\frac{1}{\sqrt{h}}(\chi_{i}-\chi_{i}(\cdot-\tau))G_{\theta}^{h}*(\chi_{i}-\chi_{i}(\cdot-\tau))\leq
C\alpha^{2}\omega.$
We observe that it suffices to prove (69), then (70) follows immediately.
Indeed recall that $\mathbb{A}$ and $\mathbb{B}$ are positive definite on
$(1,...,1)^{\perp}$. In particular there exists a constant $C>0$ such that for
any $v\in(1,...,1)^{\perp}$ one has
$|v|^{2}_{\mathbb{A}}+|v|_{\mathbb{B}}^{2}\geq C|v|^{2}\geq Cv_{i}^{2}$ for
any $i\in\\{1,...,N\\}$. Applying this to the vector
$v=G_{\theta}^{h/2}*\delta\chi_{i}$ one gets
(71)
$|G_{\theta}^{h/2}*\delta\chi_{i}|^{2}\leq\frac{1}{C}|G_{\theta}^{h/2}*\delta\chi|_{\mathbb{A}}^{2}+|G_{\theta}^{h/2}*\delta\chi|_{\mathbb{B}}^{2}.$
The claim then follows from the definition of $\omega$, (69), the symmetry
(19) and the semigroup property (21). Indeed it is sufficient to check that,
in the sense of distributions
(72) $\lim_{h\downarrow 0}\frac{1}{\sqrt{h}}\sum_{i\neq
j}\delta\chi_{i}K_{ij}^{h}*\delta\chi_{j}+\frac{1}{\sqrt{h}}\left(|G_{\gamma}^{h/2}*\delta\chi|_{\mathbb{A}}^{2}+|G_{\beta}^{h/2}*\delta\chi|_{\mathbb{B}}^{2}\right)=0.$
To this aim, pick a test function $\eta\in
C^{\infty}_{c}([0,1)^{d}\times(0,T))$. Spelling out the definition of the
norms $|\cdot|_{\mathbb{A}}$ and $|\cdot|_{\mathbb{B}}$, the claim is proved
once we show that
(73) $\displaystyle\lim_{h\downarrow
0}\frac{1}{\sqrt{h}}\sum_{i,j}a_{ij}\int\xi(\delta\chi_{i}G_{\gamma}^{h}*\delta\chi_{j}-G_{\gamma}^{h/2}*\delta\chi_{i}G_{\gamma}^{h/2}*\delta\chi_{j})dxdt=0,$
and the same claim with $a_{ij}$, $\gamma$ replaced by $b_{ij},\beta$
respectively.
We concentrate on (73). Clearly, we are done once we show that for any $i\neq
j$
(74) $\displaystyle\lim_{h\downarrow
0}\frac{1}{\sqrt{h}}\int\xi(\delta\chi_{i}G_{\gamma}^{h}*\delta\chi_{j}-G_{\gamma}^{h/2}*\delta\chi_{i}G_{\gamma}^{h/2}*\delta\chi_{j})dxdt=0.$
To show this, using the semigroup property (21) we rewrite the argument of the
limit as
(75)
$-\frac{1}{\sqrt{h}}\int[\xi,G_{\gamma}^{h/2}*](\delta\chi_{i})G_{\gamma}^{h/2}*\delta\chi_{j}dxdt,$
and we observe that by the boundedness of the measures
$\frac{1}{\sqrt{h}}|G_{\gamma}^{h/2}*\delta\chi|^{2}_{\mathbb{A}}$ it suffices
to show
(76) $\lim_{h\downarrow
0}\frac{1}{\sqrt{h}}\int|[\xi,G_{\gamma}^{h/2}*](\delta\chi_{i})|^{2}dxdt=0.$
To prove this, spelling out the integrand, using the Cauchy-Schwarz inequality
and recalling the scaling (20) we observe that
$\displaystyle\begin{aligned}
\int|[\xi,G_{\gamma}^{h/2}*](\delta\chi_{i})|^{2}dxdt\end{aligned}$ (77)
$\displaystyle\begin{aligned}
&\leq\int\left(\int|\xi(x,t)-\xi(x-z,t)|^{2}G_{\gamma}^{h/2}(z)dz\right)G_{\gamma}^{h/2}*|\delta\chi_{i}(x,t)|^{2}dxdt\\\
&\leq\frac{h}{2}\sup|\nabla\xi|^{2}\int
G_{\gamma}(z)|z|^{2}dz\int_{0}^{T}\int|\delta\chi_{i}(x,t)|^{2}dxdt.\end{aligned}$
Observe that by the compactness of $\chi^{h}$ in
$L^{2}([0,1)^{d}\times(0,T))$, (LABEL:boundOh) is of order $h$, thus (76)
indeed holds true.
The proof of (69) is essentially already contained in the paper [19]. For the
convenience of the reader we sketch the main ideas here. One reduces the claim
to proving the following facts.
(78) $\lim_{h\downarrow
0}\frac{1}{\sqrt{h}}\sum_{ij}\delta\chi_{i}K^{h}_{ij}*\delta\chi_{j}-\frac{1}{\sqrt{h}}\left(\left|G_{\gamma}^{h/2}*\delta\chi\right|_{\mathbb{A}}^{2}+\left|G_{\beta}^{h/2}*\delta\chi\right|_{\mathbb{B}}^{2}\right)=0.$
(79) $\limsup_{h\downarrow
0}\frac{1}{\sqrt{h}}\left|G_{\gamma}^{h/2}*\delta\chi\right|_{\mathbb{A}}^{2}-\alpha^{2}\frac{1}{h\sqrt{h}}\left|G_{\gamma}^{h/2}*(\chi-\chi(\cdot-h))\right|_{\mathbb{A}}^{2}\leq
0.$
(80) $\limsup_{h\downarrow
0}\frac{1}{\sqrt{h}}\left|G_{\beta}^{h/2}*\delta\chi\right|_{\mathbb{B}}^{2}-\alpha^{2}\frac{1}{h\sqrt{h}}\left|G_{\beta}^{h/2}*(\chi-\chi(\cdot-h))\right|_{\mathbb{B}}^{2}\leq
0.$
Claim (78) was proved in the previous paragraph, while (79) and (80) are
consequences of Jensen’s inequality in the time variable for the convex
functions $|\cdot|_{\mathbb{A}}^{2}$ and $|\cdot|_{\mathbb{B}}^{2}$
respectively. More precisely, assume without loss of generality that $\tau=Nh$
for some $N\in\mathbf{N}$, then by a telescoping argument and Jensen’s
inequality for $|\cdot|_{\mathbb{A}}^{2}$ we get
$\displaystyle\begin{aligned}
&\frac{1}{\sqrt{h}}|G_{\gamma}^{h/2}*\delta\chi|_{\mathbb{A}}^{2}\\\ &\leq
N\sum_{n=0}^{N-1}\frac{1}{\sqrt{h}}|G_{\gamma}^{h/2}*(\chi^{h}(\cdot-
nh)-\chi^{h}(\cdot-(n+1)h))|_{\mathbb{A}}^{2}.\end{aligned}$
Recalling that $N=\alpha/\sqrt{h}$ we can rewrite the right hand side as
(81)
$\frac{\alpha^{2}}{N}\sum_{n=0}^{N-1}\frac{1}{h\sqrt{h}}|G_{\gamma}^{h/2}*(\chi^{h}(\cdot-
nh)-\chi^{h}(\cdot-(n+1)h))|^{2}_{\mathbb{A}}.$
This is an average of time shifts of
$\alpha^{2}\frac{1}{h\sqrt{h}}|G_{\gamma}^{h/2}*(\chi^{h}-\chi^{h}(\cdot-h))|^{2}_{\mathbb{A}}$.
Since $Nh=o(1)$ all these time shifts are small, thus the average has the same
distributional limit as
$\alpha^{2}\frac{1}{h\sqrt{h}}|G_{\gamma}^{h/2}*(\chi^{h}-\chi^{h}(\cdot-h))|^{2}_{\mathbb{A}}$.
This proves (79). The argument for (80) is similar.
### Existence of the normal velocities
We now prove the existence of the normal velocities. Fix $1\leq i\leq N$ and
observe that for $w\in\\{\gamma,\beta\\}$ we have
(82)
$\begin{split}|\chi_{i}-\chi_{i}(-\tau)|\leq&(\chi_{i}-\chi_{i}(-\tau))G_{w}^{h}*(\chi_{i}-\chi_{i}(-\tau))+|\chi_{i}-G_{w}^{h}*\chi_{i}|\\\
&+|\chi_{i}(-\tau)-G_{w}^{h}*\chi_{i}(-\tau)|,\end{split}$
which follows simply by observing that
$|\chi_{i}-\chi_{i}(\cdot-\tau)|=|\chi_{i}-\chi_{i}(\cdot-\tau)|^{2}=(\chi_{i}-\chi_{i}(\cdot-\tau)G_{w}^{h}*(\chi_{i}-\chi_{i}(\cdot-\tau))+(\chi_{i}-\chi_{i}(\cdot-\tau))(\chi-
G_{w}^{h}*\chi)+(\chi_{i}(\cdot-\tau)-\chi_{i})(\chi_{i}(\cdot-\tau)-G_{w}^{h}*\chi_{i}(\cdot-\tau))$.
Using Jensen’s inequality and the elementary identity (150) in the Appendix we
have
(83) $\displaystyle\begin{aligned} |\chi_{i}-G_{w}^{h}*\chi_{i}|&\leq\int
G_{w}^{h}(z)|\chi_{i}(x)-\chi_{i}(x-z)|dz\\\ &=\int
G_{w}^{h}(z)\chi_{i}(x)(1-\chi_{i}(x-z))dz+\int
G_{w}^{h}(z)(1-\chi_{i}(x))\chi_{i}(x-z)dz\\\ &=\sum_{k\neq i}\int
G_{w}^{h}(z)\chi_{i}(x)\chi_{k}(x-z)dz+\sum_{k\neq i}\int
G_{w}^{h}(z)\chi_{k}(x)\chi_{i}(x-z)|dz.\end{aligned}$
Now observe that by testing (42) with $G_{w}/K_{ij}$ (which is bounded, and
thus admissible), we learn that
(84) $\lim_{h\downarrow 0}\frac{1}{\sqrt{h}}\int
G_{w}^{h}(z)\chi_{i}(x)\chi_{k}(x-z)dz=\int G_{w}(z)(\nu_{ik}(x,t)\cdot
z)_{+}dz\mathcal{H}^{d-1}_{|\Sigma_{ik}(t)}(dx)dt.$
Thus, if we divide (83) by $\sqrt{h}$ and let $h\downarrow 0$, using also (70)
we obtain
(85) $\begin{split}\alpha|\partial_{t}\chi_{i}|&\leq\liminf_{h\downarrow
0}\frac{|\delta\chi_{i}|}{\sqrt{h}}\\\ &\leq\limsup_{h\downarrow
0}\frac{|\delta\chi_{i}|}{\sqrt{h}}\\\ &\leq
C\alpha^{2}\omega+C\mathcal{H}^{d-1}_{\partial^{*}\Omega_{i}(t)}(dx)dt,\end{split}$
where $C$ is a constant which depends on $\gamma,\beta,N$, the mobilities and
the surface tensions. If we divide by $\alpha$ and then let $\alpha\to 0$ we
learn that $|\partial_{t}\chi_{i}|$ is absolutely continuous with respect to
$\mathcal{H}^{d-1}_{|\partial^{*}\Omega_{i}(t)}(dx)dt$. In particular, there
exists $V_{i}\in L^{1}(\mathcal{H}^{d-1}_{|\partial^{*}\Omega_{i}(t)}(dx)dt)$
which is the normal velocity of $\chi_{i}$ in the sense that
$\partial_{t}\chi_{i}=V_{i}|\nabla\chi_{i}|$ in the sense of distributions.
The optimal integrability $V_{i}\in
L^{2}(\mathcal{H}^{d-1}_{|\partial^{*}\Omega_{i}(t)}(dx)dt)$ will be shown in
the second part of the proof. Let us record for later use that with a similar
reasoning we actually obtain that
$\limsup_{h}\frac{|\delta\chi_{i}|}{\sqrt{h}}$ is absolutely continuous with
respect to $\mathcal{H}^{d-1}_{|\partial^{*}\Omega_{i}(t)}(dx)dt$.Thus in
particular inequality (85) holds with $\omega$ replaced by its absolutely
continuous part with respect to
$\mathcal{H}^{d-1}_{|\partial^{*}\Omega_{i}(t)}(dx)dt$; calling this
$\omega^{ac}_{i}$, it means
(86) $\limsup_{h\downarrow 0}\frac{|\delta\chi_{i}|}{\sqrt{h}}\leq
C\alpha^{2}\omega_{i}^{ac}+C\mathcal{H}^{d-1}_{|\partial^{*}\Omega_{i}(t)}(dx)dt.$
### Sharp Bound
Before entering into the proof of the sharp bound, we need to prove the
following property. For any $i\neq j$ we have that, in a distributional sense,
the following holds
(87) $\lim_{h\downarrow
0}\frac{1}{\sqrt{h}}\delta\chi_{i}^{+}K_{ij}^{h}*\delta\chi_{j}^{+}=0=\lim_{h\downarrow
0}\frac{1}{\sqrt{h}}\delta\chi_{i}^{-}K_{ij}^{h}*\delta\chi_{j}^{-}.$
We focus on the first limit, the second one being analogous. The first
observation is that the limit
(88) $\lambda:=\lim_{h\downarrow
0}\frac{1}{\sqrt{h}}\delta\chi_{i}^{+}K_{ij}^{h}*\delta\chi_{j}^{+}$
is a nonnegative bounded measure, which is absolutely continous with respect
to $\mathcal{H}^{d-1}_{|\Sigma_{ij}(t)}(dx)dt$. Indeed, spelling out the $z$
integral and using the fact that
$\delta\chi_{i}^{+}=\chi_{i}(1-\chi_{i}(-\tau))$ we obtain
(89) $\displaystyle\begin{aligned}
\frac{1}{\sqrt{h}}\delta\chi_{i}^{+}K_{ij}^{h}*\delta\chi_{j}^{+}=\frac{1}{\sqrt{h}}\int
K_{ij}^{h}(z)\delta\chi_{i}^{+}(x,t)*\delta\chi_{j}^{+}(x-z,t)dz\end{aligned}$
(90) $\displaystyle\begin{aligned} \leq\frac{1}{\sqrt{h}}\int
K_{ij}^{h}(z)\chi_{i}(x,t)\chi_{j}(x-z,t)dz\end{aligned}$
which by (42) in Lemma 4, as $h\downarrow 0$, converges to
(91) $\int K_{ij}(z)(\nu_{ij}(x,t)\cdot
z)_{+}\mathcal{H}^{d-1}_{|\Sigma_{ij}(t)}(dx)dt$
which is absolutely continous with respect to
$\mathcal{H}^{d-1}_{|\Sigma_{ij}(t)}(dx)dt$.
Now, given $\nu_{0}\in\mathbf{S}^{d-1}$ we claim that
(92) $\displaystyle\begin{aligned} \lambda\leq&\int_{\nu_{0}\cdot z\leq
0}K_{ij}(z)(\nu_{ij}\cdot z)_{+}\mathcal{H}^{d-1}_{|\Sigma_{ij}(t)}(dx)dt\\\
&+\int_{\nu_{0}\cdot z\geq 0}K_{ij}(z)(\nu_{ij}\cdot
z)_{-}\mathcal{H}^{d-1}_{|\Sigma_{ij}(t)}(dx)dt.\end{aligned}$
To see this, let us denote momentarily the right-hand side of (88)
(disintegrated in the $z$-variable) as
$\lambda_{h}:=\chi_{i}(x,t)(1-\chi_{i})(x,t-\tau)K_{ij}^{h}(z)\chi_{i}(x-z,t)(1-\chi_{i})(x-z,t-\tau)$.
Using the fact that $0\leq\chi_{i},\chi_{j}\leq 1$ and $\sum_{l}\chi_{l}=1$ we
obtain the following inequalities
(93) $\displaystyle\begin{aligned}
\lambda_{h}\leq\chi_{i}(x,t)K_{ij}^{h}(z)\chi_{i}(x-z,t).\end{aligned}$ (94)
$\displaystyle\begin{aligned}
\lambda_{h}\leq&\chi_{j}(x,t-\tau)K_{ij}^{h}(z)\chi_{i}(x-z,t-\tau)\\\
&+C\sum_{k\neq
i,j}K_{ij}^{h}(z)\left(|\delta\chi_{k}|(x,t)+|\delta\chi_{k}|(x-z,t)\right).\end{aligned}$
Here $C$ is a constant that does not depend on $h$. Using inequality (93) on
the domain $\\{\nu_{0}\cdot z\leq 0\\}$ and inequality (94) on the domain
$\\{\nu_{0}\cdot z\geq 0\\}$ we obtain
$\displaystyle\begin{aligned} \mathllap{\lambda}\leq&\limsup_{h\downarrow
0}\frac{1}{\sqrt{h}}\int_{\nu_{0}\cdot z\leq
0}\chi_{i}(x,t)K_{ij}^{h}(z)\chi_{i}(x-z,t)dz\\\ &+\limsup_{h\downarrow
0}\frac{1}{\sqrt{h}}\int_{\nu_{0}\cdot z\geq
0}\chi_{j}(x,t-\tau)K_{ij}^{h}(z)\chi_{i}(x-z,t-\tau)dz\\\ &+C\sum_{k\neq
i,j}\limsup_{h\downarrow 0}\left(\frac{1}{\sqrt{h}}\int
K_{ij}^{h}(z)|\delta\chi_{k}|(x,t)dz+\frac{1}{\sqrt{h}}\int
K_{ij}^{h}(z)|\delta\chi_{k}|(x-z,t)dz\right)\end{aligned}.$
Observe that for any $1\leq k\leq N$ we have
(95) $\limsup_{h\downarrow 0}\frac{1}{\sqrt{h}}\int
K_{ij}^{h}(z)|\delta\chi_{k}|(x,t)dz=\frac{1}{\sqrt{h}}\int
K_{ij}^{h}(z)|\delta\chi_{k}|(x-z,t)dz.$
This can be seen by showing that
(96) $\lim_{h\downarrow 0}\frac{1}{\sqrt{h}}\int
K_{ij}^{h}(z)\left(|\delta\chi_{k}|(x,t)-|\delta\chi_{k}|(x-z,t)\right)dz$
which can be shown to be true by testing with an admissible test function, and
putting the spatial shift $z$ on it. Thus recalling (42) and (86), we obtain
that
(97) $\displaystyle\begin{aligned} \lambda\leq&\int_{\nu_{0}\cdot z\leq
0}K_{ij}(z)(\nu_{ij}\cdot z)_{+}\mathcal{H}^{d-1}_{|\Sigma_{ij}(t)}(dx)dt\\\
&+\int_{\nu_{0}\cdot z\geq 0}K_{ij}(z)(\nu_{ij}\cdot
z)_{-}\mathcal{H}^{d-1}_{|\Sigma_{ij}(t)}(dx)dt\\\ &+C\sum_{k\neq
i,j}\alpha^{2}\omega_{k}^{ac}+\mathcal{H}^{d-1}_{|\partial^{*}\Omega_{k}(t)}(dx)dt.\end{aligned}$
Since we already know that $\lambda$ is absolutely continous with respect to
$\mathcal{H}^{d-1}_{|\Sigma_{ij}(t)}(dx)dt$, the same bound holds true if we
replace the right hand side with its absolutely continuous part with respect
to $\mathcal{H}^{d-1}_{|\Sigma_{ij}(t)}(dx)dt$. Observing that for $k\neq i,j$
by Lemma 6 in the Appendix the measures
$\mathcal{H}^{d-1}_{|\partial^{*}\Omega_{k}(t)}(dx)dt$ and
$\mathcal{H}^{d-1}_{|\partial^{*}\Sigma_{ij}(t)}(dx)dt$ are mutually singular
, this yields (92).
Writing $\lambda=\theta(x,t)\mathcal{H}^{d-1}_{|\Sigma_{ij}(t)}(dx)dt$ for
some $L^{1}(\mathcal{H}^{d-1}_{|\Sigma_{ij}(t)}(dx)dt)$-function $\theta$ we
obtain that inequality (92) yields
(98) $\displaystyle\begin{aligned} \theta(x,t)\leq\int_{\nu_{0}\cdot z\leq
0}&K_{ij}(z)(\nu_{ij}(x,t)\cdot z)_{+}dz\\\ &\int_{\nu_{0}\cdot z\geq
0}K_{ij}(z)(\nu_{ij}(x,t)\cdot z)_{-}dz\end{aligned}$
for every $\nu_{0}\in\mathbf{S}^{d-1}$ and
$\mathcal{H}^{d-1}_{|\Sigma_{ij}(t)}(dx)dt$-a.e.
$(x,t)\in[0,1)^{d}\times(0,T)$. By a separability argument, we see that the
null set on which (98) does not hold can be chosen so that it is independent
of the choice of $\nu_{0}$. If we select $\nu_{0}=\nu_{ij}(x,t)$ this yields
$\theta\leq 0$ almost everywhere with respect to
$\mathcal{H}^{d-1}_{|\Sigma_{ij}(t)}(dx)dt$. Since we already know that
$\lambda$ is nonnegative this gives $\lambda=0$.
Before getting the sharp bound, we also need to check that $V_{ij}$ is well
defined, i.e. we need to prove that for any $i\neq j$ we have $V_{i}=-V_{j}$
a.e. with respect to $\mathcal{H}^{d-1}_{|\Sigma_{ij}(t)}(dx)dt$. To see this,
we start by observing that if $\xi\in C^{\infty}_{c}([0,1)^{d}\times(0,T))$,
thanks to the fact that $\sum_{k\neq i}\chi_{k}=1-\chi_{i}$, we get
(99) $\displaystyle\begin{aligned} \int\xi
V_{i}\mathcal{H}^{d-1}_{|\partial^{*}\Omega_{i}(t)}(dx)dt&=-\int\partial_{t}\xi\chi_{i}dxdt\\\
&=\sum_{k\neq i}\int\partial_{t}\xi\chi_{k}dxdt\\\ &=-\sum_{k\neq i}\int\xi
V_{k}\mathcal{H}^{d-1}_{|\partial^{*}\Omega_{k}(t)}(dx)dt.\end{aligned}$
Choosing $\xi=f(t)g(x)$ for some $f\in C^{\infty}_{c}((0,T))$ and $g\in
C^{\infty}([0,1)^{d})$, by a separability argument, we obtain that for a.e.
$t$ and every $g\in C^{\infty}([0,1)^{d})$
(100) $\displaystyle\begin{aligned} \int
gV_{i}\mathcal{H}^{d-1}_{|\partial^{*}\Omega_{i}(t)}(dx)&=-\sum_{k\neq i}\int
gV_{k}\mathcal{H}^{d-1}_{|\partial^{*}\Omega_{k}(t)}(dx).\end{aligned}$
Pick $t$ such that (100) holds. Let $g\in C^{\infty}([0,1)^{d})$ and let
$\rho_{m}$ be a partition of unity obtained by the construction of Section 4
applied to the function $\chi(\cdot,t)$ on the interface $\Sigma_{ij}(t)$.
Then
(101) $\displaystyle\begin{aligned}
\sum_{m\in\mathbf{N}}\int\rho_{m}gV_{i}\mathcal{H}^{d-1}_{|\partial^{*}\Omega_{i}(t)}(dx)&=-\sum_{m\in\mathbf{N}}\sum_{k\neq
i}\int\rho_{m}gV_{k}\mathcal{H}^{d-1}_{\partial^{*}\Omega_{k}(t)}(dx).\end{aligned}$
Passing to the limit $r\downarrow 0$ in (101) we get by Lemma 5 that
(102) $\displaystyle\begin{aligned} \int
gV_{i}\mathcal{H}^{d-1}_{|\Sigma_{ij}(t)}(dx)&=-\int
gV_{j}\mathcal{H}^{d-1}_{\Sigma_{ij}(t)}(dx).\end{aligned}$
Since this identity holds for any $g\in C^{\infty}([0,1)^{d})$, a density
argument gives $V_{i}(x,t)=-V_{j}(x,t)$ for
$\mathcal{H}^{d-1}_{|\Sigma_{ij}(t)}$-a.e. $x$. In other words
(103) $\int|V_{i}(x,t)+V_{j}(x,t)|\mathcal{H}^{d-1}_{|\Sigma_{ij}(t)}(dx)=0.$
Integrating in time yields that $V_{i}=-V_{j}$ a.e. with respect to
$\mathcal{H}^{d-1}_{|\Sigma_{ij}(t)}(dx)dt$.
We now proceed with the derivation of the sharp lower bound. Define
$c_{ij}:=\int K_{ij}(z)dz$. Then we have
(104) $\displaystyle\begin{aligned}
c_{ij}(|\delta\chi_{i}|+|\delta\chi_{j}|)=c_{ij}(\delta\chi_{i}^{+}+\delta\chi_{j}^{-}+\delta\chi_{i}^{-}+\delta\chi_{j}^{+})\end{aligned}$
$\displaystyle\begin{aligned}
=\frac{1}{2}&\left(\delta\chi_{i}^{+}K_{ij}^{h}*(1-\delta\chi_{j}^{-})+(1-\delta\chi_{j}^{-})K_{ij}^{h}*\delta\chi_{i}^{+}+\delta\chi_{j}^{-}K_{ij}^{h}*(1-\delta\chi_{i}^{+})\right.\\\
&\left.+(1-\delta\chi_{i}^{+})K_{ij}^{h}*\delta\chi_{j}^{-}+\delta\chi_{i}^{-}K_{ij}^{h}*(1-\delta\chi_{j}^{+})+(1-\delta\chi_{j}^{+})K_{ij}^{h}*\delta\chi_{i}^{-}\right.\\\
&\left.+\delta\chi_{j}^{+}K_{ij}^{h}*(1-\delta\chi_{i}^{-})+(1-\delta\chi_{i}^{-})K_{ij}^{h}*\delta\chi_{j}^{+}\right)+\left(\delta\chi_{i}^{+}K_{ij}^{h}*\delta\chi_{j}^{-}\right.\\\
&\left.+\delta\chi_{j}^{-}K_{ij}^{h}*\delta\chi_{i}^{+}+\delta\chi_{i}^{-}K_{ij}^{h}*\delta\chi_{j}^{+}+\delta\chi_{j}^{+}K_{ij}^{h}*\delta\chi_{i}^{-}\right)\end{aligned}$
Now we rewrite the terms in the second parenthesis using
$-ab=a_{+}b_{-}+a_{-}b_{+}-a_{+}b_{+}-a_{-}b_{-}$ and then adding and
subtracting the contributions of the minority phases we obtain
(105) $\displaystyle\begin{aligned}
c_{ij}(|\delta\chi_{i}|+|\delta\chi_{j}|)\leq\frac{1}{2}&\left(\delta\chi_{i}^{+}K_{ij}^{h}*(1-\delta\chi_{j}^{-})+(1-\delta\chi_{j}^{-})K_{ij}^{h}*\delta\chi_{i}^{+}+\delta\chi_{j}^{-}K_{ij}^{h}*(1-\delta\chi_{i}^{+})\right.\\\
&\left.+(1-\delta\chi_{i}^{+})K_{ij}^{h}*\delta\chi_{j}^{-}+\delta\chi_{i}^{-}K_{ij}^{h}*(1-\delta\chi_{j}^{+})+(1-\delta\chi_{j}^{+})K_{ij}^{h}*\delta\chi_{i}^{-}\right.\\\
&\left.+\delta\chi_{j}^{+}K_{ij}^{h}*(1-\delta\chi_{i}^{-})+(1-\delta\chi_{i}^{-})K_{ij}^{h}*\delta\chi_{j}^{+}\right)-\sum_{l,p}\delta\chi_{l}K_{lp}^{h}*\delta\chi_{p}\\\
&+\delta\chi_{i}^{+}K_{ij}^{h}*\delta\chi_{j}^{+}+\delta\chi_{i}^{-}K_{ij}^{h}*\delta\chi_{j}^{-}+\delta\chi_{j}^{+}K_{ij}^{h}*\delta\chi_{i}^{+}\\\
&+\delta\chi_{j}^{-}K_{ij}^{h}*\delta\chi_{i}^{-}+\sum_{\\{l,p\\}\neq\\{i,j\\},\\{l,p\\}}\delta\chi_{l}K_{lp}^{h}*\delta\chi_{p}.\end{aligned}$
Now the main idea is to split the integral of $K_{ij}$ in the definition of
$c_{ij}$ into two parts. More precisely, by the symmetry (19), for any
$\nu_{0}\in\mathbf{S}^{d-1}$ and any $V_{0}>0$ we have
(106) $c_{ij}=2\int_{0\leq\nu_{0}\cdot z\leq\alpha
V_{0}}K_{ij}(z)dz+2\int_{\nu_{0}\cdot z>\alpha V_{0}}K_{ij}(z)dz.$
Substituting into (105) and dividing by $\sqrt{h}$ we obtain
(107) $\displaystyle\begin{aligned} 2\int_{0\leq\nu_{0}\cdot z\leq\alpha
V_{0}}K_{ij}(z)dz\frac{(|\delta\chi_{i}|+|\delta\chi_{j}|)}{\sqrt{h}}\end{aligned}$
$\displaystyle\begin{aligned}
=\frac{1}{2\sqrt{h}}&\bigg{(}\delta\chi_{i}^{+}K_{ij}^{h}*(1-\delta\chi_{j}^{-})+(1-\delta\chi_{j}^{-})K_{ij}^{h}*\delta\chi_{i}^{+}+\delta\chi_{j}^{-}K_{ij}^{h}*(1-\delta\chi_{i}^{+})\\\
&+(1-\delta\chi_{i}^{+})K_{ij}^{h}*\delta\chi_{j}^{-}+\delta\chi_{i}^{-}K_{ij}^{h}*(1-\delta\chi_{j}^{+})+(1-\delta\chi_{j}^{+})K_{ij}^{h}*\delta\chi_{i}^{-}\\\
&+\delta\chi_{j}^{+}K_{ij}^{h}*(1-\delta\chi_{i}^{-})+(1-\delta\chi_{i}^{-})K_{ij}^{h}*\delta\chi_{j}^{+}\\\
&-4\int_{\nu_{0}\cdot z>\alpha
V_{0}}K_{ij}(z)dz(|\delta\chi_{i}|+|\delta\chi_{j}|)\\\
&-2\sum_{lp}\delta\chi_{l}K_{lp}^{h}*\delta\chi_{p}+2\delta\chi_{i}^{+}K_{ij}^{h}*\delta\chi_{j}^{+}+2\delta\chi_{i}^{-}K_{ij}^{h}*\delta\chi_{j}^{-}\\\
&+2\delta\chi_{j}^{+}K_{ij}^{h}*\delta\chi_{i}^{+}+2\delta\chi_{j}^{-}K_{ij}^{h}*\delta\chi_{i}^{-}\\\
&+\sum_{(l,p)\neq(i,j),(l,p)\neq(j,i)}\delta\chi_{l}K_{lp}^{h}*\delta\chi_{p}\bigg{)}.\end{aligned}$
We will be interested in bounding the $\liminf$ of the left hand side. Observe
that the distributional limit of the last five terms is non-positive. Indeed,
the limit of first four terms vanish distributionally by property (87), while
the last term is bounded from above by
$\sum_{(l,p)\neq(i,j),(l,p)\neq(j,i)}\delta\chi_{l}^{+}K_{lp}^{h}*\delta\chi_{p}^{+}+\delta\chi_{l}^{-}K_{lp}^{h}*\delta\chi_{p}^{-},$
which vanish distributionally by property (87). We thus obtain that the
$\liminf$ of the left hand side of (107) is bounded from above by
(108) $\displaystyle\begin{aligned} \liminf_{h\downarrow
0}\frac{1}{2\sqrt{h}}&\bigg{(}\delta\chi_{i}^{+}K_{ij}^{h}*(1-\delta\chi_{j}^{-})+(1-\delta\chi_{j}^{-})K_{ij}^{h}*\delta\chi_{i}^{+}+\delta\chi_{j}^{-}K_{ij}^{h}*(1-\delta\chi_{i}^{+})\\\
&+(1-\delta\chi_{i}^{+})K_{ij}^{h}*\delta\chi_{j}^{-}+\delta\chi_{i}^{-}K_{ij}^{h}*(1-\delta\chi_{j}^{+})+(1-\delta\chi_{j}^{+})K_{ij}^{h}*\delta\chi_{i}^{-}\\\
&+\delta\chi_{j}^{+}K_{ij}^{h}*(1-\delta\chi_{i}^{-})+(1-\delta\chi_{i}^{-})K_{ij}^{h}*\delta\chi_{j}^{+}\\\
&-4\int_{\nu_{0}\cdot z>\alpha
V_{0}}K_{ij}(z)dz(|\delta\chi_{i}|+|\delta\chi_{j}|)-2\sum_{lp}\delta\chi_{l}K_{lp}^{h}*\delta\chi_{p}\bigg{)}.\end{aligned}$
For the last term we use the sharp bound (69), relating this term to our
dissipation measure $\omega$. We would like to get a good bound for the other
terms. This cannot be done naively as before, since we want the bound to be
sharp. We claim that
(109) $\displaystyle\begin{aligned} \limsup_{h\downarrow
0}\frac{1}{\sqrt{h}}&\left(\delta\chi_{i}^{+}K_{ij}^{h}*(1-\delta\chi_{j}^{-})+(1-\delta\chi_{j}^{-})K_{ij}^{h}*\delta\chi_{i}^{+}\delta\chi_{j}^{-}K_{ij}^{h}*(1-\delta\chi_{i}^{+})\right.\\\
&\left.+(1-\delta\chi_{i}^{+})K_{ij}^{h}*\delta\chi_{j}^{-}+\delta\chi_{i}^{-}K_{ij}^{h}*(1-\delta\chi_{j}^{+})+(1-\delta\chi_{j}^{+})K_{ij}^{h}*\delta\chi_{i}^{-}\right.\\\
&\left.+\delta\chi_{j}^{+}K_{ij}^{h}*(1-\delta\chi_{i}^{-})+(1-\delta\chi_{i}^{-})K_{ij}^{h}*\delta\chi_{j}^{+}\right.\\\
&\left.-4\int_{\nu_{0}\cdot z>\alpha
V_{0}}K_{ij}(z)dz(|\delta\chi_{i}|+|\delta\chi_{j}|)\right)\end{aligned}$
$\displaystyle\begin{aligned} \leq 8\int_{0\leq\nu_{0}\cdot z\leq\alpha
V_{0}}K_{ij}(z)|\nu_{ij}(x)\cdot
z|&dz\mathcal{H}^{d-1}_{|\Sigma_{ij}(t)}(dx)dt\\\ &+C\sum_{k\neq
i,j}(\alpha^{2}\omega^{ac}_{k}+\mathcal{H}^{d-1}_{|\partial^{*}\Omega_{k}(t)}(dx)dt.\end{aligned}$
Here $C$ is a constant that depends on $\gamma,\beta,\mathbb{A},\mathbb{B}$,
but not on $h$. Assume for the moment that (109) is true and let us conclude
the argument in this case. Using (109) and (69) we obtain
(110) $\displaystyle\begin{aligned} 2\liminf_{h\downarrow
0}\int_{0\leq\nu_{0}\cdot z\leq\alpha
V_{0}}K_{ij}(z)dz\frac{(|\delta\chi_{i}|+|\delta\chi_{j}|)}{\sqrt{h}}\end{aligned}$
$\displaystyle\begin{aligned} \leq\alpha^{2}\omega+4\int_{0\leq\nu_{0}\cdot
z\leq\alpha V_{0}}K_{ij}(z)|\nu_{ij}(x)\cdot
z|&dz\mathcal{H}^{d-1}_{|\Sigma_{ij}(t)}(dx)dt\\\ &+C\sum_{k\neq
i,j}(\alpha^{2}\omega^{ac}_{k}+\mathcal{H}^{d-1}_{|\partial^{*}\Omega_{k}(t)}(dx)dt\end{aligned}$
in the sense of distributions on $[0,1)^{d}\times(0,T)$. Observe also that the
left hand side of (110) is an upper bound for $\int_{0\leq\nu_{0}\cdot
z\leq\alpha V_{0}}K_{ij}(z)dz(|\partial_{t}\chi_{i}|+|\partial_{t}\chi_{j}|)$,
thus the inequality still holds true if the left hand side is replaced by this
term. Remember that $\omega_{k}^{ac}$ is absolutely continuous with respect to
$\mathcal{H}^{d-1}_{|\partial^{*}\Omega_{k}(t)}(dx)dt$, thus there exist
functions $W_{k}\in
L^{1}(\mathcal{H}^{d-1}_{|\partial^{*}\Omega_{k}(t)}(dx)dt)$ such that
$\omega_{k}^{ac}=W_{k}(x,t)\mathcal{H}^{d-1}_{|\partial^{*}\Omega_{k}(t)}(dx)dt$.
We now disintegrate the measure $\omega$, i.e. we find a Borel family
$\omega_{t},t\in(0,T)$, of positive measures on $[0,1)^{d}$ such that
$\omega=\omega_{t}\otimes dt$. Having said this, it is not hard to see that
(110) holds in a disintegrated version, i.e. we have for Lebesgue a.e.
$t\in(0,T)$
(111) $\displaystyle\begin{aligned} 2\int_{0\leq\nu_{0}\cdot z\leq\alpha
V_{0}}K_{ij}(z)dz(|V_{i}(x,t)|\mathcal{H}^{d-1}_{|\partial^{*}\Omega_{i}(t)}(dx)+|V_{j}(x,t)|\mathcal{H}^{d-1}_{|\partial^{*}\Omega_{j}(t)}(dx))\end{aligned}$
$\displaystyle\begin{aligned}
\leq\alpha^{2}\omega_{t}+4\int_{0\leq\nu_{0}\cdot z\leq\alpha
V_{0}}K_{ij}(z)|\nu_{ij}(x)\cdot
z|&dz\mathcal{H}^{d-1}_{|\Sigma_{ij}(t)}(dx)\\\ &+C\sum_{k\neq
i,j}(\alpha^{2}W_{k}(x,t)+1)\mathcal{H}^{d-1}_{|\partial^{*}\Omega_{k}(t)}(dx).\end{aligned}$
Here $\nu_{0}\in\mathbf{S}^{d-1}$ and $V_{0}\in(0,\infty)$ are arbitrary:
indeed even if the set of points in time for which (111) holds is a priori
dependent on $\nu_{0}$ and $V_{0}$, a standard separability argument allows us
to conclude that we can get rid of this dependence.
Fix a point $t$ in time such that (111) holds. In what follows, we drop the
time variable $t$ which is fixed, so for example $V_{i}(x)=V_{i}(x,t)$,
$\Sigma_{ij}=\Sigma_{ij}(t)$ and so on. Fix $\xi\in C([0,1)^{d})$, observe
that by definition of $V_{ij}$ and by using the fact that
$\Sigma_{ij}\subset\partial^{*}\Omega_{i}\cap\partial^{*}\Omega_{j}$ we have
(112) $\displaystyle\begin{aligned} 4\alpha\int_{0\leq\nu_{0}\cdot z\leq\alpha
V_{0}}K_{ij}(z)dz\int_{[0,1)^{d}}\xi(x)|V_{ij}(x)|\mathcal{H}^{d-1}_{|\Sigma_{ij}}(dx)\end{aligned}$
$\displaystyle\begin{aligned}
\leq\alpha^{2}\int_{[0,1)^{d}}\xi(x)\omega_{t}(dx)&+4\int_{[0,1)^{d}}\int_{0\leq\nu_{0}\cdot
z\leq\alpha V_{0}}K_{ij}(z)|\nu_{ij}(x)\cdot
z|dz\xi(x)\mathcal{H}^{d-1}_{|\Sigma_{ij}(t)}(dx)\\\ &+C\sum_{k\neq
i,j}\int_{[0,1)^{d}}\xi(x)(\alpha^{2}W_{k}(x,t)+1)\mathcal{H}^{d-1}_{|\partial^{*}\Omega_{k}(t)}(dx).\end{aligned}$
Let us relabel $\nu_{0}$, $V_{0}$ and $\xi$ to make clear that they may depend
on the pair $i,j$. Thus $\nu_{0}^{ij}\in\mathbf{S}^{d-1}$,
$V_{0}^{ij}\in(0,\infty)$ and $\xi_{ij}\in C([0,1)^{d})$ are arbitrary, and it
holds
(113) $\displaystyle\begin{aligned} 2\alpha\int_{0\leq\nu_{0}^{ij}\cdot
z\leq\alpha
V^{ij}_{0}}K_{ij}(z)dz\int_{[0,1)^{d}}\xi_{ij}(x)|V_{ij}(x)|\mathcal{H}^{d-1}_{|\Sigma_{ij}}(dx)\end{aligned}$
$\displaystyle\begin{aligned}
\leq\alpha^{2}\int_{[0,1)^{d}}\xi_{ij}(x)\omega_{t}(dx)&+4\int_{[0,1)^{d}}\int_{0\leq\nu_{0}^{ij}\cdot
z\leq\alpha V_{0}^{ij}}K_{ij}(z)|\nu_{ij}(x)\cdot
z|dz\xi_{ij}(x)\mathcal{H}^{d-1}_{|\Sigma_{ij}(t)}(dx)\\\ &+C\sum_{k\neq
i,j}\int_{[0,1)^{d}}\xi_{ij}(x)(\alpha^{2}W_{k}(x,t)+1)\mathcal{H}^{d-1}_{|\partial^{*}\Omega_{k}(t)}(dx).\end{aligned}$
Let $\\{\rho_{m}\\}$ be a partition of unity obtained using the construction
of Section 4 applied to the function $\chi(\cdot,t)$ on the inferface
$\Sigma_{ij}(t)$. Use the above inequality with $\xi_{ij}$ replaced by
$\rho_{m}\xi_{ij}$ and sum over $m$ and $i,j$ to get
(114) $\displaystyle\begin{aligned}
\sum_{i<j}\sum_{m\in\mathbf{N}}\mathbf{LH}_{m}^{ij}\leq\sum_{i<j}\sum_{m\in\mathbf{N}}(\mathbf{I}_{m}^{ij}+\mathbf{II}_{m}^{ij}+\mathbf{III}_{m}^{ij})\end{aligned}$
where we have set
$\displaystyle\begin{aligned}
\mathllap{\mathbf{LH}_{m}^{ij}}=2\alpha\int_{0\leq\nu_{0}^{ij}\cdot
z\leq\alpha
V^{ij}_{0}}K_{ij}(z)dz\int_{[0,1)^{d}}\rho_{mij}(x)\xi_{ij}(x)|V_{ij}(x)|\mathcal{H}^{d-1}_{|\Sigma_{ij}}(dx),\end{aligned}$
$\displaystyle\begin{aligned}
\mathllap{\mathbf{I}^{ij}_{m}}=\alpha^{2}\int_{[0,1)^{d}}\xi_{ij}(x)\rho_{mij}(x)\omega_{t}(dx),\end{aligned}$
$\displaystyle\begin{aligned}
\mathllap{\mathbf{II}^{ij}_{m}}=4\int_{[0,1)^{d}}\int_{0\leq\nu_{0}^{ij}\cdot
z\leq\alpha V_{0}^{ij}}K_{ij}(z)|\nu_{ij}(x)\cdot
z|dz\rho_{mij}(x)\xi_{ij}(x)\mathcal{H}^{d-1}_{|\Sigma_{ij}(t)}(dx),\end{aligned}$
$\displaystyle\begin{aligned} \mathllap{\mathbf{III}^{ij}_{m}}=C\sum_{k\neq
i,j}\int_{[0,1)^{d}}\rho_{mij}(x)\xi_{ij}(x)(\alpha^{2}W_{k}(x,t)+1)\mathcal{H}^{d-1}_{|\partial^{*}\Omega_{k}(t)}(dx).\end{aligned}$
Observe that
(115)
$\sum_{i<j}\sum_{m\in\mathbf{N}}\mathbf{I}_{m}^{ij}\leq\sum_{i<j}\alpha^{2}\int_{[0,1)^{d}}\xi_{ij}(x)\omega_{t}(dx)$
because $\rho_{m}$ is a partition of unity. Moreover by Lemma 5 we get
$\displaystyle\begin{aligned} \mathllap{\lim_{r\downarrow
0}\sum_{i<j}\sum_{m\in\mathbf{N}}\mathbf{LH}_{m}^{ij}}=2\alpha\int_{0\leq\nu_{0}^{ij}\cdot
z\leq\alpha
V^{ij}_{0}}K_{ij}(z)dz\int_{[0,1)^{d}}\xi_{ij}(x)|V_{ij}(x)|\mathcal{H}^{d-1}_{|\Sigma_{ij}}(dx).\end{aligned}$
$\displaystyle\begin{aligned} \mathllap{\lim_{r\downarrow
0}\sum_{i<j}\sum_{m\in\mathbf{N}}\mathbf{II}_{m}^{ij}}=\sum_{i<j}4\int_{[0,1)^{d}}\int_{0\leq\nu_{0}^{ij}\cdot
z\leq\alpha V_{0}^{ij}}K_{ij}(z)|\nu_{ij}(x)\cdot
z|dz\xi_{ij}(x)\mathcal{H}^{d-1}_{|\Sigma_{ij}(t)}(dx).\end{aligned}$
$\displaystyle\begin{aligned} \mathllap{\lim_{r\downarrow
0}\sum_{i<j}\sum_{m\in\mathbf{N}}\mathbf{III}_{m}^{ij}}=0.\end{aligned}$
Putting things together we obtain that for any
$\nu_{0}^{ij}\in\mathbf{S}^{d-1}$, any $V_{0}^{ij}\in(0,\infty)$, and any
$\xi\in C([0,1)^{d})$
(116) $\displaystyle\begin{aligned}
2\alpha\sum_{i<j}\int_{0\leq\nu_{0}^{ij}\cdot z\leq\alpha
V^{ij}_{0}}K_{ij}(z)dz\int_{[0,1)^{d}}\xi_{ij}(x)|V_{ij}(x)|\mathcal{H}^{d-1}_{|\Sigma_{ij}}(dx)\end{aligned}$
$\displaystyle\begin{aligned}
\leq&\sum_{i<j}\alpha^{2}\int_{[0,1)^{d}}\xi_{ij}(x)\omega_{t}(dx)\\\
&+4\sum_{i<j}\int_{[0,1)^{d}}\int_{0\leq\nu_{0}^{ij}\cdot z\leq\alpha
V_{0}^{ij}}K_{ij}(z)|\nu_{ij}(x)\cdot
z|dz\xi_{ij}(x)\mathcal{H}^{d-1}_{|\Sigma_{ij}(t)}(dx).\end{aligned}$
We now claim that by approximation the above inequality is valid for any
simple function $\xi_{ij}\geq 0$. To see this, it is clear that we can
concentrate on $\xi_{ij}=w_{ij}\mathbf{1}_{B_{ij}}$, where
$B_{ij}\subset[0,1)^{d}$ are Borel and $w_{ij}\geq 0$. Observe that by the
dominated convergence theorem, the family
(117) $\mathcal{F}:=\left\\{B=\prod_{i<j}B_{ij}:\
B_{ij}\in\mathcal{B}([0,1)^{d})\ \text{s.t.}\ \forall w_{ij}\geq 0\ \text{
(\ref{inequalityToApr}) holds with}\
\xi_{ij}=w_{ij}\mathbf{1}_{B_{ij}}\right\\}$
is a monotone class. Thus by the monotone class theorem we just need to show
that it contains all the products of open sets. But this is easy because given
$B_{ij}\subset[0,1)^{d}$ open sets, we can always find sequences
$\eta_{k}^{ij}$ of continuous functions with compact support such that
$0\leq\eta_{k}^{ij}\leq\mathbf{1}_{B_{ij}}$ and such that
$\eta_{k}^{ij}\to\mathbf{1}_{B_{ij}}$, thus the claim follows by the monotone
convergence theorem.
With this in place one can use an approximation argument to replace the vector
$\nu_{0}^{ij}$ with the $\mathcal{H}^{d-1}$-measurable vector valued function
$\nu_{ij}$ obtaining the following inequality:
(118) $\displaystyle\begin{aligned}
2\alpha\sum_{i<j}\int_{[0,1)^{d}}\int_{0<\nu_{ij}(x)\cdot z<\alpha
V^{ij}_{0}}K_{ij}(z)dz\xi_{ij}(x)|V_{ij}(x)|\mathcal{H}^{d-1}_{|\Sigma_{ij}}(dx)\leq\end{aligned}$
$\displaystyle\begin{aligned}
\leq&\sum_{i<j}\alpha^{2}\int_{[0,1)^{d}}\eta_{ij}(x)\omega_{t}(dx)\\\
&+4\sum_{i<j}\int_{[0,1)^{d}}\int_{0\leq\nu_{ij}\cdot z\leq\alpha
V_{0}^{ij}}K_{ij}(z)|\nu_{ij}(x)\cdot
z|dz\xi_{ij}(x)\mathcal{H}^{d-1}_{|\Sigma_{ij}}(dx).\end{aligned}$
Now divide by $\alpha^{2}$ and send $\alpha$ to zero. Record the following
limits, which can be computed spelling out the definition of $K_{ij}$, and
recalling the symmetry property (19) and the factorization property (22) for
the heat kernel
(119) $\displaystyle\begin{aligned} \lim_{\alpha\downarrow
0}\frac{1}{\alpha}\int_{0<\nu_{ij}(x)\cdot z<\alpha
V_{0}^{ij}}K_{ij}(z)dz=\frac{V_{0}^{ij}}{\mu_{ij}}.\end{aligned}$
$\displaystyle\begin{aligned} \lim_{\alpha\downarrow
0}\frac{1}{\alpha^{2}}\int_{0\leq\nu_{ij}(x)\cdot z\leq\alpha
V_{0}^{ij}}K_{ij}(z)|\nu_{ij}(x)\cdot
z|dz=\frac{(V_{0}^{ij})^{2}}{4\mu_{ij}}.\end{aligned}$
Then if we insert back into (118) we obtain
(120)
$\displaystyle\sum_{i<j}\frac{2}{\mu_{ij}}\int_{[0,1)^{d}}V_{0}^{ij}\xi_{ij}(x)|V_{ij}(x)|\mathcal{H}^{d-1}_{|\Sigma_{ij}}(dx)$
(121)
$\displaystyle\leq\int_{[0,1)^{d}}\omega_{t}(dx)+\sum_{i<j}\int_{[0,1)^{d}}\frac{(V_{0}^{ij})^{2}}{\mu_{ij}}\xi_{ij}(x)\mathcal{H}^{d-1}_{|\Sigma_{ij}(t)}(dx).$
Now given $M>0$ take sequences of simple functions
(122) $s_{m}^{ij}=\sum_{k=1}^{p_{m}}w_{k}^{ij}\mathbf{1}_{B^{k}_{ij}}$
such that $s_{m}^{ij}\to|V_{ij}|\mathbf{1}_{\\{|V_{ij}|\leq M\\}}$ as
$m\to+\infty$ monotonically almost everywhere with respect to
$\mathcal{H}^{d-1}_{|\Sigma_{ij}(t)}$. We are assuming that
$\\{B_{ij}^{k}\\}_{k=1,...,p_{m}}$ are disjoint and
$\mathcal{H}^{d-1}$-measurable, with the property that $B_{ij}^{k_{1}}\cap
B_{lr}^{k_{2}}=\emptyset$ if $\\{i,j\\}\neq\\{l,r\\}$. Choosing
$V_{0}^{ij}=w_{k}^{ij}$, $\xi_{ij}=\mathbf{1}_{B^{k}_{ij}}$ in (120) and
summing over $k$ we obtain
(123)
$\displaystyle\sum_{i<j}\frac{2}{\mu_{ij}}\int_{[0,1)^{d}}s_{m}^{ij}|V_{ij}(x)|\mathcal{H}^{d-1}_{|\Sigma_{ij}}(dx)$
$\displaystyle\leq\int_{[0,1)^{d}}\omega_{t}(dx)+\sum_{i<j}\int_{[0,1)^{d}}\frac{(s_{m}^{ij})^{2}}{\mu_{ij}}\mathcal{H}^{d-1}_{|\Sigma_{ij}(t)}(dx).$
Taking the limit $m\to+\infty$, using the monotone convergence theorem we
obtain
(124)
$\displaystyle\sum_{i<j}\frac{2}{\mu_{ij}}\int_{[0,1)^{d}}|V_{ij}(x)|^{2}\mathbf{1}_{\\{|V_{ij}|\leq
M\\}}\mathcal{H}^{d-1}_{|\Sigma_{ij}}(dx)$
$\displaystyle\leq\int_{[0,1)^{d}}\omega_{t}(dx)+\sum_{i<j}\int_{[0,1)^{d}}\frac{|V_{ij}(x)|^{2}}{\mu_{ij}}\mathbf{1}_{\\{|V_{ij}|\leq
M\\}}\mathcal{H}^{d-1}_{|\Sigma_{ij}}(dx)$
or, in other words,
$\displaystyle\sum_{i<j}\frac{1}{\mu_{ij}}\int_{[0,1)^{d}}|V_{ij}(x)|^{2}\mathbf{1}_{\\{|V_{ij}|\leq
M\\}}\mathcal{H}^{d-1}_{|\Sigma_{ij}}(dx)\leq\int_{[0,1)^{d}}\omega_{t}(dx).$
Recall that $\mu_{ij}=\mu_{ji}$, thus the inequality above may be rewritten as
$\displaystyle\sum_{i,j}\frac{1}{2\mu_{ij}}\int_{[0,1)^{d}}|V_{ij}(x)|^{2}\mathbf{1}_{\\{|V_{ij}|\leq
M\\}}\mathcal{H}^{d-1}_{|\Sigma_{ij}}(dx)\leq\int_{[0,1)^{d}}\omega_{t}(dx).$
If we now integrate in time we learn by the monotone convergence theorem that
$V_{ij}\in L^{2}(\mathcal{H}^{d-1}_{|\Sigma_{ij}(t)}(dx)dt)$ and that the
sharp bound (45) is satisfied.
### Proof of (109)
To prove (109) we proceed in several steps.
First of all, we claim that the first eight terms may be substituted by
(125) $\displaystyle 2\int_{\nu_{0}\cdot z\geq 0}K_{ij}^{h}(z)$
$\displaystyle\left(|\delta\chi_{i}^{+}-\delta\chi_{j}^{-}(-z)|+|\delta\chi_{i}^{+}(-z)-\delta\chi_{j}^{-}|\right.$
$\displaystyle\left.|\delta\chi_{i}^{-}-\delta\chi_{j}^{+}(-z)|+|\delta\chi_{i}^{-}(-z)-\delta\chi_{j}^{+}|\right)dz.$
To show this, observe that we may replace the implicit $z$-integrals in the
convolution in the first eight terms by twice the integrals over the half
space $\\{\nu_{0}\cdot z\geq 0\\}$ instead of $\mathbf{R}^{d}$. This is
clearly true once we observe that
(126) $\displaystyle\begin{aligned} \lim_{h\downarrow
0}\frac{1}{\sqrt{h}}&\left(\delta\chi_{i}^{+}\int_{\nu_{0}\cdot z\geq
0}K_{ij}^{h}(z)(1-\delta\chi_{j}^{-}(\cdot-z))dz\right.\\\
&\left.+(1-\delta\chi_{j}^{-})\int_{\nu_{0}\cdot z\geq
0}K_{ij}^{h}(z)\delta\chi_{i}^{+}(\cdot-z)dz\right)\end{aligned}$
$\displaystyle\begin{aligned} \mathllap{=}\lim_{h\downarrow
0}\frac{1}{\sqrt{h}}&\left(\delta\chi_{i}^{+}\int_{\nu_{0}\cdot z\leq
0}K_{ij}^{h}(z)(1-\delta\chi_{j}^{-}(\cdot-z))dz\right.\\\
&\left.+(1-\delta\chi_{j}^{-})\int_{\nu_{0}\cdot z\leq
0}K_{ij}^{h}(z)\delta\chi_{i}^{+}(\cdot-z)dz\right)\end{aligned}$
and that similar identities hold exchanging the roles of $i,j$ and $+,-$
respectively. That (126) holds is not difficult to show. Indeed taking into
account the fact that the kernel is even, the argument of the second limit is
just a spatial shift of $z$ of the first one. The spatial shift may be put
onto the test function, and thanks to the scaling of the kernel one can get
the claim. We may thus substitute the first eight terms of the left hand side
of (109) with twice the same terms with the integration with respect to $z$ on
the half space $\\{\nu_{0}\cdot z\geq 0\\}$. If we rely again on the fact that
$\delta\chi_{i}^{+}\in\\{0,1\\}$, by identity (150) in the Appendix we obtain
(125), as claimed.
Now we need two inequalities for the integrand. First note that the integrand
is a second-order finite difference, we claim that
(127) $\displaystyle\begin{aligned}
|\delta\chi_{i}^{+}-\delta\chi_{j}^{-}(\cdot-z)|+|\delta\chi_{i}^{+}(\cdot-z)-\delta\chi_{j}^{-}|+|\delta\chi_{i}^{-}-\delta\chi_{j}^{+}(\cdot-z)|+|\delta\chi_{i}^{-}(\cdot-z)-\delta\chi_{j}^{+}|\end{aligned}$
$\displaystyle\begin{aligned}
\leq\begin{cases}|\delta\chi_{i}^{+}-\delta\chi_{i}^{+}(\cdot-z)|+|\delta\chi_{i}^{-}-\delta\chi_{i}^{-}(\cdot-z)|+|\delta\chi_{j}^{+}-\delta\chi_{j}^{+}(\cdot-z)|+|\delta\chi_{j}^{-}-\delta\chi_{j}^{-}(\cdot-z)|\\\
+4\sum_{k\neq i,j}(|\delta\chi_{k}|+\delta\chi_{k}(\cdot-z)|).\\\
\mathllap{}\\\
|\delta\chi_{i}|+|\delta\chi_{i}(\cdot-z)|+|\delta\chi_{j}|+|\delta\chi_{j}(\cdot-z)|.\end{cases}\end{aligned}$
The second one follows from the triangle inequality. To show the first one,
observe that
(128) $\displaystyle\begin{aligned}
\mathllap{|\delta\chi_{i}^{+}-\delta\chi_{j}^{-}(\cdot-z)|}=&(1-\delta\chi_{i}^{+})\delta\chi_{j}^{-}(\cdot-z)+\delta\chi_{i}^{+}(1-\delta\chi_{j}^{-}(\cdot-z))\end{aligned}$
$\displaystyle\begin{aligned}
\mathllap{}\leq&(1-\delta\chi_{i}^{+})\delta\chi_{i}^{+}(\cdot-z)+\sum_{k\neq
i,j}|\delta\chi_{k}(\cdot-z)|+\delta\chi_{j}^{+}(1-\delta\chi_{j}^{+}(\cdot-z))\\\
&+\sum_{k\neq i,j}|\delta\chi_{k}|\end{aligned}$
and that similarly
(129) $\displaystyle\begin{aligned}
\mathllap{|\delta\chi_{i}^{+}(\cdot-z)-\delta\chi_{j}^{-}|}=&(1-\delta\chi_{i}(\cdot-z)^{+})\delta\chi_{j}^{-}+\delta\chi_{i}(\cdot-z)^{+}(1-\delta\chi_{j})\end{aligned}$
$\displaystyle\begin{aligned}
\mathllap{}\leq&(1-\delta\chi_{i}(\cdot-z)^{+})\delta\chi_{i}^{-}+\sum_{k\neq
i,j}|\delta\chi_{k}|+\delta\chi_{j}(\cdot-z)^{+}(1-\delta\chi_{j}^{-})\\\
&+\sum_{k\neq i,j}|\delta\chi_{k}(\cdot-z)|.\end{aligned}$
Summing up the two inequalities we get
(130)
$\displaystyle|\delta\chi_{i}^{+}-\delta\chi_{j}^{-}(\cdot-z)|+|\delta\chi_{i}^{+}(\cdot-z)-\delta\chi_{j}^{-}|\leq$
$\displaystyle\leq|\delta\chi_{i}^{+}-\delta\chi_{i}^{+}(\cdot-z)|+|\delta\chi_{j}^{-}-\delta\chi_{j}^{-}(\cdot-z)|+2\sum_{k\neq
i,j}|\delta\chi_{k}|+|\delta\chi_{k}(\cdot-z)|.$
Similar bounds hold for the remaining terms in (127).
We now split the integral (125) into the domains of integration
$\\{0\leq\nu_{0}\cdot z\leq\alpha V_{0}\\}$ and $\\{\nu_{0}\cdot z>\alpha
V_{0}\\}$. On the first one we use the first inequality in (127) for the
integrand. Recalling identity (150) and inequality (151) in the Appendix we
obtain, and using the fact that $\sum_{k}\chi_{k}=1$
$\displaystyle\begin{aligned} 2\int_{0\leq\nu_{0}\cdot z\leq\alpha
V_{0}}K_{ij}^{h}(z)&\left(|\delta\chi_{i}^{+}-\delta\chi_{j}^{-}(\cdot-z)|+|\delta\chi_{i}^{+}(\cdot-z)-\delta\chi_{j}^{-}|\right.\\\
&\left.+|\delta\chi_{i}^{-}-\delta\chi_{j}^{+}(\cdot-z)|+|\delta\chi_{i}^{-}(\cdot-z)-\delta\chi_{j}^{+}|\right)dz\end{aligned}$
$\displaystyle\begin{aligned} \leq 2\int_{0\leq\nu_{0}\cdot z\leq\alpha
V_{0}}K_{ij}^{h}(z)&\left(|\chi_{i}-\chi_{i}(\cdot-z)|+|\chi_{i}(-\tau)-\chi_{i}(\cdot-\tau,\cdot-z)|\right.\\\
&\left.|\chi_{j}-\chi_{j}(\cdot-z)|+|\chi_{j}(\cdot-\tau)-\chi_{j}(\cdot-\tau,\cdot-z)|\right.\\\
&\left.+8\sum_{k\neq
i,j}|\delta\chi_{k}|+|\delta\chi_{k}(\cdot-z)|\right)dz\end{aligned}$ (131)
$\displaystyle\begin{aligned} \leq 2\int_{0\leq\nu_{0}\cdot z\leq\alpha
V_{0}}K_{ij}^{h}(z)&\left(\chi_{i}\chi_{j}(\cdot-z)+\chi_{i}(\cdot-z)\chi_{j}+\sum_{k\neq
i,j}\chi_{i}\chi_{k}(\cdot-z)+\chi_{i}(\cdot-z)\chi_{k}\right.\\\
&\left.\chi_{i}(\cdot-\tau)\chi_{j}(\cdot-\tau,\cdot-z)+\chi_{i}(\cdot-\tau,\cdot-z)\chi_{j}(\cdot-\tau)\right.\\\
&\left.+\sum_{k\neq
i,j}\chi_{i}(\cdot-\tau)\chi_{k}(\cdot-\tau,\cdot-z)+\chi_{i}(\cdot-\tau,\cdot-z)\chi_{k}(\cdot-\tau)\right.\\\
&\left.\chi_{j}\chi_{i}(\cdot-z)+\chi_{j}(\cdot-z)\chi_{i}+\sum_{k\neq
i,j}\chi_{j}\chi_{k}(\cdot-z)+\chi_{j}(\cdot-z)\chi_{k}\right.\\\
&\left.\chi_{j}(\cdot-\tau)\chi_{i}(\cdot-\tau,\cdot-z)+\chi_{j}(\cdot-\tau,\cdot-z)\chi_{i}(\cdot-\tau)\right.\\\
&\left.+\sum_{k\neq
i,j}\chi_{j}(\cdot-\tau)\chi_{k}(\cdot-\tau,\cdot-z)+\chi_{j}(\cdot-\tau,\cdot-z)\chi_{k}(\cdot-\tau)\right.\\\
&\left.+8\sum_{k\neq
i,j}|\delta\chi_{k}|+|\delta\chi_{k}(\cdot-z)|\right)dz.\end{aligned}$
On the set $\\{\nu_{0}\cdot z>\alpha V_{0}\\}$ we use the second inequality in
(127), obtaining
(132) $\displaystyle\begin{aligned} 2\int_{\nu_{0}\cdot z>\alpha
V_{0}}K_{ij}^{h}(z)&\left(|\delta\chi_{i}^{+}-\delta\chi_{j}^{-}(\cdot-z)|+|\delta\chi_{i}^{+}(\cdot-z)-\delta\chi_{j}^{-}|\right.\\\
&\left.+|\delta\chi_{i}^{-}-\delta\chi_{j}^{+}(\cdot-z)|+|\delta\chi_{i}^{-}(\cdot-z)-\delta\chi_{j}^{+}|\right)dz\end{aligned}$
$\displaystyle\begin{aligned} \leq 2\int_{\nu_{0}\cdot z>\alpha
V_{0}}K_{ij}^{h}(z)(|\delta\chi_{i}|+|\delta\chi_{i}(\cdot-z)|+|\delta\chi_{j}|+|\delta\chi_{j}(\cdot-z)|)dz.\end{aligned}$
We now observe that for any $1\leq k\leq N$ we have, as we already observed in
(96)
(133) $\lim_{h\downarrow 0}\frac{1}{\sqrt{h}}\int_{0\leq\nu_{0}\cdot
z\leq\alpha
V_{0}}K_{ij}^{h}(z)(|\delta\chi_{k}(\cdot-z)|-|\delta\chi_{k}|)dz=0,$
thus in particular
(134) $\displaystyle\limsup_{h\downarrow
0}\frac{1}{\sqrt{h}}\int_{0\leq\nu_{0}\cdot z\leq\alpha
V_{0}}K_{ij}^{h}(z)|\delta\chi_{k}(\cdot-z)|dz$
$\displaystyle=\limsup_{h\downarrow
0}\frac{1}{\sqrt{h}}\int_{0\leq\nu_{0}\cdot z\leq\alpha
V_{0}}K_{ij}^{h}(z)|\delta\chi_{k}|dz.$
By putting the time shift $\tau$ on the test function if is easy to check that
the distributional limit of the terms of (131) which involve the shift $\tau$
have the same limit as the corresponding terms without the time shift. Thus
recalling (86) and relying on (133) and (42) we obtain that inserting (131)
and (132) into (125), the left hand side of (109) is bounded by
$\displaystyle\begin{aligned} &8\int_{0\leq\nu_{0}\cdot z\leq\alpha
V_{0}}K_{ij}(z)((\nu_{ij}(x,t)\cdot z)_{+}+(\nu_{ij}(x,t)\cdot
z)_{-})\mathcal{H}^{d-1}_{|\Sigma_{ij}(t)}(dx)dt\\\ &+C\sum_{k\neq
i,j}\int_{0\leq\nu_{0}\cdot z\leq\alpha V_{0}}K_{ik}(z)((\nu_{ik}(x,t)\cdot
z)_{+}+(\nu_{ik}(x,t)\cdot z)_{-})\mathcal{H}^{d-1}_{|\Sigma_{ik}(t)}(dx)dt\\\
&+C\sum_{k\neq
i,j}(\alpha^{2}\omega_{k}^{ac}+\mathcal{H}^{d-1}_{|\partial^{*}\Omega_{k}(t)}(dx)dt),\end{aligned}$
which clearly gives the claim once we realize that
$\displaystyle\begin{aligned} &\int_{0\leq\nu_{0}\cdot z\leq\alpha
V_{0}}K_{ik}(z)((\nu_{ik}(x)\cdot z)_{+}+(\nu_{ik}(x)\cdot
z)_{-})\mathcal{H}^{d-1}_{|\Sigma_{ik}(t)}(dx)dt\\\ &\leq
2\int_{\mathbf{R}^{d}}K_{ik}(z)|z|dz\leq C.\end{aligned}$
∎
###### Proof of Proposition 2.
The proof is along the same lines as Proposition 2 in [19], where the claim is
analized in the case of two phases. For the convenience of the reader, we
outline the strategy of the full proof, providing details only for the
required changes. The proof is split into several steps.
step 1. The first observation is that for any $h>0$, any admissible
$u\in\mathcal{M}$ and any smooth vector field $\xi$ we have the following
lower bound for the metric slope, cf. (33)
$\frac{1}{2}|\partial E_{h}|(u)\geq\delta
E_{h}(u)_{\bullet}\xi-\frac{1}{2}\left(\delta
d_{h}(\cdot,u)_{\bullet}\xi\right)^{2}.$
Here $\delta$ denotes the first variation, which is computed considering the
curve $s\to u_{s}$ of configurations which solve the transport equations
(135) $\begin{cases}\partial_{s}u_{i}^{s}+\xi\cdot\nabla u_{i}^{s}=0,\\\
u_{i}^{s}(\cdot,0)=u_{i}(\cdot),\end{cases}$
and by setting
(136) $\delta E_{h}(u)_{\bullet}\xi:=\frac{d}{ds}_{|{s=0}}E_{h}(u^{s})\
\text{and}\ \delta
d_{h}(\cdot,u)_{\bullet}\xi:=\frac{d}{ds}_{|{s=0}}d(u,u^{s}).$
step 2. The second observation is a representation formula for $\delta
E_{h}(u)_{\bullet}\xi$. Namely
(137) $\displaystyle\begin{aligned} \delta
E_{h}(u)_{\bullet}\xi=\sum_{i,j}\frac{1}{\sqrt{h}}&\left(\int\nabla\cdot\xi
u_{i}K_{ij}^{h}*u_{j}dx+\int\nabla\cdot\xi u_{j}K_{ij}^{h}*u_{i}dxdt\right.\\\
&\left.+\int[\xi,\nabla K_{ij}^{h}*](u_{j})u_{i}dx\right).\end{aligned}$
Here $[\xi,\nabla K_{ij}^{h}*]$ denotes the commutator obtained taking the
convolution with $\nabla K_{ij}^{h}$ and multiplying by $\xi$. To check this
formula one starts by assuming $u$ to be smooth and then an approximation
argument gives the result for a general $u\in\mathcal{M}$.
step 3. Representation for $\delta d_{h}(\cdot,u)_{\bullet}\xi$. One checks
that
$\displaystyle\begin{aligned} &\frac{1}{2}\left(\delta
d_{h}(\cdot,u)_{\bullet}\xi\right)^{2}\\\
&=\frac{\sqrt{h}}{2}\sum_{i,j}\left(\int
u_{i}\xi\cdot\nabla^{2}K_{ij}^{h}*(\xi u_{j})dx+\int
u_{j}\xi\cdot\nabla^{2}K_{ij}^{h}*(\xi u_{i})dx\right.\\\ &\left.+\int
u_{i}\nabla\cdot\xi\nabla K_{ij}^{h}*(\xi u_{j})dx++\int
u_{j}\nabla\cdot\xi\nabla K_{ij}^{h}*(\xi u_{i})dx\right.\\\ &\left.-\int
u_{i}\nabla\cdot\xi K_{ij}^{h}*(u_{j}\nabla\cdot\xi)dx-\int
u_{j}\nabla\cdot\xi K_{ij}^{h}*(u_{i}\nabla\cdot\xi)dx\right.\\\
&\left.-\int\xi u_{i}\nabla K_{ij}^{h}*(u_{j}\nabla\cdot\xi)dx-\int\xi
u_{j}\nabla K_{ij}^{h}*(u_{i}\nabla\cdot\xi)dx\right).\end{aligned}$
Once again this formula can be easily checked when $u$ is smooth, an
approximation argument then gives the extension to the case $u\in\mathcal{M}$.
step 4. Passage to the limit in $\delta E_{h}$. We claim that
(138) $\lim_{h\downarrow 0}\int_{0}^{T}\delta E_{h}(u^{h}(t))_{\bullet}\xi
dt=\sum_{i,j}\sigma_{ij}\int\left(\nabla\cdot\xi-\nu_{ij}\cdot\nabla\xi\nu_{ij}\right)\mathcal{H}^{d-1}_{|\Sigma_{ij}(t)}(dx)dt.$
The proof is very similar to the two phases case, and relies on the weak
convergence (42). Firstly, testing (42) with $\nabla\cdot\xi$ we get
$\displaystyle\begin{aligned} &\lim_{h\downarrow
0}\sum_{i,j}\frac{1}{\sqrt{h}}\int\left(\nabla\cdot\xi
u_{i}^{h}K_{ij}^{h}*u_{j}^{h}+\nabla\cdot\xi
u_{j}^{h}K_{ij}^{h}*u_{i}^{h}\right)dxdt\\\
&=\sum_{i,j}2\sigma_{ij}\int\nabla\cdot\xi\mathcal{H}^{d-1}_{|\Sigma_{ij}(t)}(dx)dt.\end{aligned}$
For the term involving the commutator, one checks that
$\displaystyle\lim_{h\downarrow 0}\left(\int[\xi,\nabla
K_{ij}^{h}*](u_{j}^{h})u_{i}^{h}dxdt-\int\nabla\xi z\cdot\nabla
K_{ij}^{h}(z)u_{j}^{h}(x-z,t)u_{i}^{h}(x,t)dzdxdt\right)=0.$
With this in place, we observe that
$\displaystyle\begin{aligned} &\int\nabla\xi z\cdot\nabla
K_{ij}^{h}(z)u_{j}^{h}(x-z,t)u_{i}^{h}(x,t)dzdxdt\\\
&=\int\nabla\xi(x,t)z\cdot\nabla K_{ij}(z)(\nu_{ij}(x,t)\cdot
z)_{+}\mathcal{H}^{d-1}_{|\Sigma_{ij}(t)}(dx)dt\end{aligned}$
which can be seen by testing (42) with $\frac{\nabla\xi z\cdot\nabla
K_{ij}(z)}{K_{ij}(z)}$ which is of polynomial growth in $z$. To conclude (138)
we just need to show that for any symmetric matrix $A\in\mathbf{R}^{d\times
d}$ and any unit vector $\nu$ we have
$\int Az\cdot\nabla K_{ij}(z)(\nu\cdot
z)_{+}dz=-\sigma_{ij}\left(\operatorname{tr}A+\nu\cdot A\nu\right).$
Using the definition of the kernel $K_{ij}$ it suffices to show that
$\int Az\cdot\nabla G_{w}(z)(\nu\cdot
z)_{+}dz=-\frac{\sqrt{w}}{\sqrt{\pi}}\left(\operatorname{tr}A+\nu\cdot
A\nu\right)\ w\in\\{\gamma,\beta\\}.$
step 5. Passage to the limit in $\delta d_{h}(\cdot,u)\xi$. We claim that
(139) $\displaystyle\begin{aligned} \lim_{h\downarrow
0}\frac{1}{2}\left(\delta
d_{h}(\cdot,u^{h})_{\bullet}\xi\right)=\sum_{i,j}\frac{1}{2\mu_{ij}}\int(\xi\cdot\nu_{ij})^{2}\mathcal{H}^{d-1}_{|\Sigma_{ij}(t)}(dx)dt.\end{aligned}$
To prove this, we observe that the terms which do not involve the Hessian
$\nabla^{2}K_{ij}^{h}$ are all $O(\sqrt{h})$. For example, to prove that
(140) $\sqrt{h}\int u_{i}^{h}\nabla\cdot\xi\nabla K_{ij}^{h}*(\xi
u_{j}^{h})dxdt=O(\sqrt{h}),$
spell out the integral in the convolution, use the fact that $\nabla
K_{ij}^{h}=\frac{1}{\sqrt{h}^{d+1}}\nabla K_{ij}(\frac{z}{\sqrt{h}})$, use the
fact that $\nabla\xi(x,t)\xi(x-\sqrt{h}z,t)$ is bounded and test (42) with
$\nabla K_{ij}/K_{ij}$. The other terms can be treated similarly. For the
terms involving the Hessian of the kernel, we split the claim into
(141) $\displaystyle\begin{aligned} \lim_{h\downarrow 0}\sqrt{h}\int
u_{i}^{h}(\xi\cdot\nabla^{2}K_{ij}^{h}*u_{j})\xi
dxdt=\frac{1}{2\mu_{ij}}\int(\xi\cdot\nu_{ij}(x,t))^{2}\mathcal{H}^{d-1}_{|\Sigma_{ij}(t)}(dx)dt,\end{aligned}$
(142) $\displaystyle\begin{aligned} \sqrt{h}\int
u_{i}^{h}\xi\cdot[\xi,\nabla^{2}K_{ij}^{h}*](u_{j}^{h})dxdt=O(\sqrt{h}).\end{aligned}$
The proof of (142) is similar to the argument for (140). To prove identity
(141) observe that by spelling out the $z$-integral, a change of variable and
by testing (42) with
$\frac{\xi(x,t)\cdot\nabla^{2}K_{ij}(z)\xi(x,t)}{K_{ij}(z)}$ we obtain
$\displaystyle\begin{aligned} &\lim_{h\downarrow 0}\sqrt{h}\int
u_{i}^{h}(\xi\cdot\nabla^{2}K_{ij}^{h}*u^{j})\xi dxdt\\\
&=\int\xi\cdot\nabla^{2}K_{ij}(z)\xi(\nu_{ij}(x,t)\cdot
z)_{+}\mathcal{H}^{d-1}_{|\Sigma_{ij}(t)}(dx)dt.\end{aligned}$
Now identity (139) follows from the following formula: for any two vectors
$\xi\in\mathbf{R}^{d}$ and $\nu\in\mathbf{S}^{d-1}$ we have
(143) $\int\xi\cdot\nabla^{2}K_{ij}(z)\xi(\nu\cdot
z)_{+}dz=\frac{1}{2\mu_{ij}}(\xi\cdot\nu)^{2}.$
To check (143), by relying on the definition of the kernels, we just need to
show that for $w\in\\{\gamma,\beta\\}$
$\int\xi\cdot\nabla^{2}G_{w}(x)\xi(\nu\cdot z)_{+}dz=\frac{1}{2\sqrt{\pi
w}}(\xi\cdot\nu)^{2}.$
Since the kernel is isotropic, we can reduce to the case $\xi=e_{1}$, thus we
need to prove
$\int\partial^{2}_{1}G_{w}(x)(\nu\cdot z)_{+}dz=\frac{1}{2\sqrt{\pi
w}}\nu_{1}^{2}.$
This can be done after two integration by parts and observing that
$\int_{\nu\cdot z=0}G_{w}(z)dz=\frac{1}{2\sqrt{\pi w}}.$
conclusion. By step 1 we have
$\frac{1}{2}\int_{0}^{T}|\partial E_{h}|^{2}(u^{h})\ dt\geq\int_{0}^{T}\delta
E_{h}(u^{h})_{\bullet}\xi dt-\frac{1}{2}\int_{0}^{T}\left(\delta
d_{h}(\cdot,u^{h})_{\bullet}\xi\right)^{2}dt.$
Taking the liminf on the left hand side, using step 4 and step 5 we get that
for any smooth vector field $\xi$
$\displaystyle\begin{aligned} \liminf_{h\downarrow
0}\frac{1}{2}\int_{0}^{T}|\partial
E_{h}|^{2}(u^{h})dt\geq&\sum_{i,j}\left[\sigma_{ij}\int\left(\nabla\cdot\xi-\nu_{ij}\cdot\nabla\xi\nu_{ij}\right)\mathcal{H}^{d-1}_{|\Sigma_{ij}(t)}(dx)dt\right.\\\
&\left.-\frac{1}{2\mu_{ij}}\int(\xi\cdot\nu_{ij})^{2}\
\mathcal{H}^{d-1}_{|\Sigma_{ij}(t)}(dx)dt\right].\end{aligned}$
Since the left hand side is bounded, the Riesz representation theorem for
$L^{2}$ yields functions $H_{ij}\in
L^{2}(\mathcal{H}^{d-1}_{|\Sigma_{ij}(t)}(dx)dt)$ such that
$\sum_{i,j}\sigma_{ij}\int\left(\nabla\cdot\xi-\nu_{ij}\cdot\nabla\xi\nu_{ij}\right)\
\mathcal{H}^{d-1}_{|\Sigma_{ij}(t)}(dx)dt=-\sum_{i,j}\sigma_{ij}\int
H_{ij}\nu_{ij}\cdot\xi\ \mathcal{H}^{d-1}_{|\Sigma_{ij}(t)}(dx)dt$
and such that for any $\xi\in
L^{2}(\mathcal{H}^{d-1}_{|\bigcup_{i,j}\Sigma_{ij}(t)}(dx)dt)$
$\displaystyle\begin{aligned} \liminf_{h\downarrow
0}\frac{1}{2}\int_{0}^{T}|\partial E_{h}|(u_{h})\
dt\geq\sum_{i,j}\bigg{(}&-\sigma_{ij}\int
H_{ij}\nu_{ij}\cdot\xi\mathcal{H}^{d-1}_{|\Sigma_{ij}(t)}(dx)dt\\\
&-\frac{1}{2\mu_{ij}}\int(\xi\cdot\nu_{ij})^{2}\mathcal{H}^{d-1}_{|\Sigma_{ij}(t)}(dx)dt\bigg{)}.\end{aligned}$
Since the integration measures are mutually singular we can test with $\xi\in
L^{2}(\mathcal{H}^{d-1}_{|\bigcup_{i,j}\Sigma_{ij}(t)}(dx)dt)$ such that
$\xi_{|\Sigma_{ij}(t)}=-\mu_{ij}\sigma_{ij}H_{ij}\nu_{ij}$. This yields
$\displaystyle\begin{aligned} \liminf_{h\downarrow
0}\frac{1}{2}\int_{0}^{T}|\partial E_{h}|^{2}(u_{h})\
dt\geq&\sum_{i,j}\frac{\sigma_{ij}^{2}\mu_{ij}}{2}\int H_{ij}^{2}\
\mathcal{H}^{d-1}_{|\Sigma_{ij}(t)}(dx)dt.\end{aligned}$
∎
## 6\. Appendix
### 6.1. Proof of Lemma 5
Before giving the proof of this result, we need a simple technical lemma.
###### Lemma 6.
Fix $1\leq l\neq p\leq N$. Then for any $1\leq i\neq j\leq N$ such that
$\\{i,j\\}\neq\\{l,p\\}$ the interfaces $\Sigma_{ij}$ and $\Sigma_{lp}$ are
disjoint. In particular for $\mathcal{H}^{d-1}$-a.e. $x\in\Sigma_{lp}$ we have
that
(144) $\lim_{r\to 0}\mathcal{H}^{d-1}(\Sigma_{ij}\cap B(x,r))=0$
###### Proof.
We first show that the interfaces $\Sigma_{ij}$ and $\Sigma_{lp}$ are
disjoint. This follows immediately once we recall that every point in the
reduced boundary of a set of finite perimeter has density $1/2$ (see [24],
Corollary 15.8). Assume for example that $i\neq l,p$. Thus if
$y\in\Sigma_{lp}$ we have
(145)
$\begin{split}&1\geq\limsup_{r}\frac{|(\Omega_{l}\cup\Omega_{p}\cup\Omega_{i})\cap
B(y,r)|}{\omega_{d}r^{d}}\\\ &=\lim_{r}\frac{|\Omega_{l}\cap
B(y,r)|}{\omega_{d}r^{d}}+\lim_{r}\frac{|\Omega_{p}\cap
B(y,r)|}{\omega_{d}r^{d}}+\limsup_{r}\frac{|\Omega_{i}\cap
B(y,r)|}{\omega_{d}r^{d}}\\\ &=1+\limsup_{r}\frac{|\Omega_{i}\cap
B(y,r)|}{\omega_{d}r^{d}}\end{split}$
which says that $y$ has density zero in $\Omega_{i}$.
The fact that (144) holds is now a consequence of the geneneral fact
$\limsup_{r\downarrow 0}\frac{\mathcal{H}^{d-1}(\Sigma_{ij}\cap
B(x,r))}{\omega_{d-1}r^{d-1}}=0$
for $\mathcal{H}^{d-1}$-a.e. $x\in\Sigma_{ij}^{c}$. ∎
###### Proof of Lemma 5.
The argument for (1) can be found in [17] in the case of two phases and
without localization, i.e. with $\eta=1$ and $N=2$. For the sake of
completeness, we provide the proof in our case. Upon splitting into the
negative and positive part, we may assume $\eta\geq 0$. Clearly the only
nonzero terms in the sum are those for wich
$B_{m}^{r}\cap\Sigma_{ij}\neq\emptyset$. Fix such a ball: by definition there
exists $y\in r\mathbf{Z}^{d}$ such that $B_{m}^{r}=B(y,2r\sqrt{d})$. If
$x\in\Sigma_{ij}\cap B_{m}^{r}$ then we have that $B(x,2r\sqrt{d})\subset
B(y,4r\sqrt{d})$, and by definition of $\mathcal{E}^{r}$ this yields
$\mathcal{H}^{d-1}(B(x,2r\sqrt{d})\cap\Sigma_{ij})\leq\frac{\omega_{d-1}}{2^{d}}(4r)^{d-1}\sqrt{d}^{d-1}=\frac{\omega_{d-1}}{2}(2r)^{d-1}\sqrt{d}^{d-1}.$
Thus $x$ belongs to the set of points in $\Sigma_{ij}\cap B_{m}^{r}$ such that
(146)
$\frac{\mathcal{H}^{d-1}(B(x,2r\sqrt{d})\cap\Sigma_{ij})}{\omega_{d-1}(2r\sqrt{d})^{d-1}}\leq\frac{1}{2}.$
By De Giorgi’s structure theorem the approximate tangent plane exists at every
point $x\in\Sigma_{ij}$, thus (146) cannot hold when $r$ is small enough:
moreover every point $x\in\Sigma_{ij}$ is contained in at most $c(2,d)$ balls,
this means that
(147) $\sum_{m}\mathbf{1}_{\left\\{z\in B_{m}^{r}\cap\Sigma_{ij}:\
\frac{\mathcal{H}^{d-1}(B(x,2r\sqrt{d})\cap\Sigma_{ij})}{\omega_{d-1}(2r\sqrt{d})^{d-1}}\leq\frac{1}{2}\right\\}}(x)\eta(x)\leq
c(2,d)\eta(x)$
and that the left hand side of (147) converges to zero pointwise. By the
dominated convergence theorem we get our claim.
Proof of (2). Upon splitting into the negative and positive part, we may
assume $\eta\geq 0$. Given a point $x\in\Sigma_{lp}$, if $y\in
r\mathbf{Z}^{d}$ is such that $x\in B(y,2r\sqrt{d})$, then
$B(y,4r\sqrt{d})\subset B(x,6r\sqrt{d})$. Thus for any $1\leq i<j\leq N$ with
$(i,j)\neq(l,p)$ we have
$\displaystyle\mathcal{H}^{d-1}(B(y,4r\sqrt{d})\cap\Sigma_{ij})$
$\displaystyle\leq\mathcal{H}^{d-1}(B(x,6r\sqrt{d})\cap\Sigma_{ij})$
$\displaystyle\leq\frac{\omega_{d-1}}{2^{d}}(4r)^{d-1}\sqrt{d}^{d-1}$
provided $r$ is small enough, this follows from Lemma 6. Since
$\mathcal{F}_{2}^{r}$ covers $\mathbf{R}^{d}$ we obtain that
$x\in\bigcup_{m\in\mathbf{N}}B_{m}^{r}$
for all $r$ small enough. In other words
$\lim_{r\downarrow 0}\sum_{m}\rho_{m}(x)\eta(x)=\eta(x)$
pointwise on $\Sigma_{lp}$, and the argument of the limit on the right hand
side is dominated by $\eta$. Thus we may once again appeal to the dominated
convergence theorem and conclude the proof.
∎
### 6.2. Consistency and Monotonicity
The following results are essetially contained in [8] and [17], indeed the
proofs may be adapted because we are assuming that $a_{ij}$ and $b_{ij}$
satisfy the triangle inequality.
###### Lemma 7.
Let $\chi\in L^{1}((0,T),BV([0,1)^{d})^{N})$ such that
$\chi(\cdot,t)\in\mathcal{A}$ for a.e. $t$. Then
$\lim_{h\downarrow 0}\int_{0}^{T}E_{h}(\chi)dt=\int_{0}^{T}E(\chi)dt.$
Even more is true: for any $g\in C^{\infty}([0,1)^{d})$ and any pair $1\leq
i\neq j\leq N$ we have
$\displaystyle\begin{aligned} &\lim_{h\downarrow
0}\frac{1}{\sqrt{h}}\int_{0}^{T}\int
g(x)(\chi_{i}(x,t)K_{ij}^{h}*\chi_{j}(x,t)+\chi_{j}(x,t)K_{ij}^{h}*\chi_{j}(x,t))dxdt\\\
&=\int g(x)K_{ij}(z)|\nu_{ij}\cdot z|dzdxdt.\end{aligned}$
###### Lemma 8.
For any $0<h\leq h_{0}$ we have
$E_{h}(u)\geq\left(\frac{\sqrt{h_{0}}}{\sqrt{h}+\sqrt{h_{0}}}\right)^{d+1}E_{h_{0}}(u).$
### 6.3. Improved convergence of the energies
The following Lemma is an improvement of the convergence of the energies, the
proof of this result is contained, with minor modifications, in the paper
[17], Corollary 3.7.
###### Lemma 9.
Let $u^{h}$ be a sequence of $[0,1]$-valued functions such that $u^{h}\to\chi$
in $L^{1}([0,1)^{d}\times(0,T))$ and
(148) $\lim_{h\downarrow
0}\int_{0}^{T}E_{h}(u^{h}(t))dt=\int_{0}^{T}E(\chi(t))dt.$
Then we have that
(149) $\displaystyle\begin{aligned} \lim_{h\downarrow 0}\frac{1}{\sqrt{h}}\int
G_{\gamma}^{h}(z)|f^{\gamma}_{h}(z)-f^{\gamma}(z)|dz=0,\end{aligned}$
$\displaystyle\begin{aligned} \lim_{h\downarrow 0}\frac{1}{\sqrt{h}}\int
G_{\beta}^{h}(z)|f^{\beta}_{h}(z)-f^{\beta}(z)|dz=0.\end{aligned}$
Where we set
$\displaystyle\begin{aligned} f_{h}^{\gamma}(z)=\sum_{i,j}a_{ij}\int
u_{i}^{h}(x,t)u_{j}^{h}(x-z,t)dxdt,\quad
f^{\gamma}(z)=\sum_{i,j}a_{ij}\int\chi_{i}(x,t)\chi_{j}(x-z,t)dxdt,\end{aligned}$
$\displaystyle\begin{aligned} f_{h}^{\beta}(z)=\sum_{i,j}b_{ij}\int
u_{i}^{h}(x,t)u_{j}^{h}(x-z,t)dxdt,\quad
f^{\beta}(z)=\sum_{i,j}b_{ij}\int\chi_{i}(x,t)\chi_{j}(x-z,t)dxdt.\end{aligned}$
### 6.4. Some inequalities
Here we gather some elementary inequalities which are used frequently.
###### Lemma 10.
Let $a,b,a^{\prime},b^{\prime}\in\\{0,1\\}$, then the following inequalities
hold.
(150) $\displaystyle\begin{aligned} |a-b|=a(1-b)+b(1-a)\end{aligned}$ (151)
$\displaystyle\begin{aligned}
|(a-a^{\prime})_{+}-(b-b^{\prime})_{+}|&+|(a-a^{\prime})_{-}-(b-b^{\prime})_{-}|\\\
&\leq|a-b|+|a^{\prime}-b^{\prime}|\end{aligned}$
###### Proof.
The first identity follows by expanding $|a-b|=|a-b|^{2}$. The second one is
proved in [19]. For the sake of completeness, we reproduce the proof here.
There are two cases. In the first one we have
$(a-a^{\prime})(b-b^{\prime})\geq 0$ and we may assume upon replacing
$(a,a^{\prime},b,b^{\prime})$ with $(-a,-a^{\prime},-b,-b^{\prime})$ that
$(a-a^{\prime})$ and $(b-b^{\prime})$ are non-negative. Then (151) reduces to
$|(a-a^{\prime})-(b-b^{\prime})|\leq|a-b|+|a^{\prime}-b^{\prime}|.$
The second case is given by $(a-a^{\prime})(b-b^{\prime})\leq 0$. By an
argument as before we may assume $(a-a^{\prime})\geq 0\geq(b-b^{\prime})$,
thus (151) reduces to
$(a-a^{\prime})+(b-b^{\prime})\leq|a-b|+|a^{\prime}-b^{\prime}|.$
∎
###### Lemma 11.
There exists a constant $C>0$ depending only on $N,\mathbb{A},\mathbb{B}$ such
that for any $v\in\mathcal{M}$
(152) $\int|v-K^{h_{0}}*v|dx\leq C\sqrt{h_{0}}E_{h}(v)\ \text{for all}\
h_{0}\geq h.$
###### Proof.
The proof of (152) is aontained in the proof of Lemma 3 in [19] for the two
phases case when $K^{h}$ is the scaled version of the Gaussian with variance
$1$. The same proof may be adapted to our setting because we still have
monotonicity of the energy (Lemma 8) and we can prove essentially by the use
of Jensen’s inequality that
$\int|v-K^{h}*v|dx\leq C\sqrt{h}E_{h}(v).$
∎
## Acknowledgements
This project has received funding from the Deutsche Forschungsgemeinschaft
(DFG, German Research Foundation) under Germany’s Excellence Strategy –
EXC-2047/1 – 390685813.
## References
* [1] Fred Almgren, Jean E. Taylor, and Lihe Wang. Curvature-driven flows: a variational approach. SIAM Journal on Control and Optimization, 31(2):387–438, 1993.
* [2] Luigi Ambrosio, Nicola Gigli, and Giuseppe Savaré. Gradient flows in metric spaces and in the space of probability measures. Lectures in Mathematics ETH Zürich. Birkhäuser Verlag, Basel, 2005\.
* [3] Guy Barles and Christine Georgelin. A simple proof of convergence for an approximation scheme for computing motions by mean curvature. SIAM Journal on Numerical Analysis, 32(2):484–500, 1995.
* [4] K. Barmak, Eva Eggeling, M. Emelianenko, Y. Epshteyn, D. Kinderlehrer, R. Sharp, and S. Ta’asan. Critical events, entropy, and the grain boundary character distribution. Physical Review B, 83(13):134117, 2011.
* [5] Antonin Chambolle and Matteo Novaga. Convergence of an algorithm for the anisotropic and crystalline mean curvature flow. SIAM Journal on Mathematical Analysis, 37(6):1978–1987, 2006.
* [6] Guido De Philippis and Tim Laux. Implicit time discretization for the mean curvature flow of mean convex sets. Ann. Sc. Norm. Super. Pisa Cl. Sci. (5), 21:911-930, 2020.
* [7] Matt Elsey and Selim Esedoḡlu. Threshold dynamics for anisotropic surface energies. Mathematics of Computation, 87(312):1721–1756, 2018.
* [8] Selim Esedoḡlu and Felix Otto. Threshold dynamics for networks with arbitrary surface tensions. Communications on Pure and Applied Mathematics, 68(5):808–864, 2015\.
* [9] Lawrence C. Evans. Convergence of an algorithm for mean curvature motion. Indiana University Mathematics Journal, 42(2):533–557, 1993.
* [10] Julian Fischer, Sebastian Hensel, Tim Laux, and Theresa Simon. The local structure of the energy landscape in multiphase mean curvature flow: Weak-strong uniqueness and stability of evolutions. arXiv e-prints, March 2020. arXiv:2003.05478 [math.AP].
* [11] Tom Ilmanen, André Neves, and Felix Schulze. On short time existence for the planar network flow. Journal of Differential Geometry, 111(1):39–89, 2019.
* [12] Hitoshi Ishii, Gabriel E. Pires, and Panagiotis E. Souganidis. Threshold dynamics type approximation schemes for propagating fronts. Journal of the Mathematical Society of Japan, 51(2):267–308, 1999\.
* [13] Matt Jacobs, Inwon Kim, and Alpár R. Mészáros. Weak Solutions to the Muskat Problem with Surface Tension Via Optimal Transport. Archive for Rational Mechanics and Analysis, 239(1):389–430, 2021\.
* [14] Richard Jordan, David Kinderlehrer, and Felix Otto. The variational formulation of the Fokker-Planck equation. SIAM Journal on Mathematical Analysis, 29(1):1–17, 1998.
* [15] Lami Kim and Yoshihiro Tonegawa. On the mean curvature flow of grain boundaries. Université de Grenoble. Annales de l’Institut Fourier, 67(1):43–142, 2017.
* [16] Joe Klobusicky, Govind Menon, and Robert L. Pego. Two-dimensional grain boundary networks: stochastic particle models and kinetic limits. Archive for Rational Mechanics and Analysis, pages 1–55, 2020.
* [17] Tim Laux and Felix Otto. Convergence of the thresholding scheme for multi-phase mean-curvature flow. Calculus of Variations and Partial Differential Equations, 55(5):Art. 129, 74, 2016.
* [18] Tim Laux and Felix Otto. Brakke’s inequality for the thresholding scheme. Calculus of Variations and Partial Differential Equations, 59(1):Paper No. 39, 26, 2020.
* [19] Tim Laux and Felix Otto. The thresholding scheme for mean curvature flow and De Giorgi’s ideas for minimizing movements. Advanced Studies in Pure Mathematics 85, The Role of Metrics in the Theory of Partial Differential Equations, pages 63–93, 2020.
* [20] Tim Laux and Theresa Simon. Convergence of the Allen-Cahn equation to multiphase mean curvature flow. Communications on Pure and Applied Mathematics, 71(8):1597–1647, 2018.
* [21] Tim Laux and Nung Kwan Yip. Analysis of diffusion generated motion for mean curvature flow in codimension two: a gradient-flow approach. Archive for Rational Mechanics and Analysis, 232(2):1113–1163, 2019\.
* [22] Jorge Lira, Rafe Mazzeo, Alessandra Pluda, and Mariel Saez. Short-time existence for the network flow, January 2021. arXiv: 2101.04302 [math.DG].
* [23] Stephan Luckhaus and Thomas Sturzenhecker. Implicit time discretization for the mean curvature flow equation. Calculus of Variations and Partial Differential Equations, 3(2):253–271, 1995.
* [24] Francesco Maggi. Sets of finite perimeter and geometric variational problems, volume 135 of Cambridge Studies in Advanced Mathematics. Cambridge University Press, Cambridge, 2012.
* [25] Barry Merriman, James K. Bence, and Stanley J. Osher. Diffusion generated motion by mean curvature. CAM Reports, Department of Mathematics, University of California, Los Angeles, 1992.
* [26] Barry Merriman, James K. Bence, and Stanley J. Osher. Motion of multiple functions: a level set approach. Journal of Computational Physics, 112(2):334–363, 1994.
* [27] Peter W. Michor and David Mumford. Riemannian geometries on spaces of plane curves. Journal of the European Mathematical Society (JEMS), 8(1):1–48, 2006.
* [28] Alexander Mielke. On evolutionary $\varGamma$-convergence for gradient systems. In Macroscopic and large scale phenomena: coarse graining, mean field limits and ergodicity, volume 3 of Lect. Notes Appl. Math. Mech., pages 187–249. Springer, [Cham], 2016.
* [29] William W Mullins. Two-dimensional motion of idealized grain boundaries. Journal of Applied Physics, 27(8):900–904, 1956.
* [30] W. T. Read and W. Shockley. Dislocation models of crystal grain boundaries. Physical Review, 78(3):275, 1950.
* [31] Tiago Salvador and Selim Esedoḡlu. A simplified threshold dynamics algorithm for isotropic surface energies. Journal of Scientific Computing, 79(1):648–669, 2019.
* [32] Etienne Sandier and Sylvia Serfaty. Gamma-convergence of gradient flows with applications to Ginzburg-Landau. Communications on Pure and Applied Mathematics, 57(12):1627–1672, 2004.
* [33] I. J. Schoenberg. Metric spaces and positive definite functions. Transactions of the American Mathematical Society, 44(3):522–536, 1938.
* [34] Yoshihiro Tonegawa. Brakke’s mean curvature flow. SpringerBriefs in Mathematics. Springer, Singapore, 2019.
|
# On the Cauchy problems associated to a ZK-KP-type family equations with a
transversal fractional dispersion
Jorge Morales P<EMAIL_ADDRESS>Departamento de Matemáticas,
Universidad Nacional de Colombia, Sede–Bogotá and Félix H. Soriano M
<EMAIL_ADDRESS>Departamento de Matemáticas, Universidad Nacional de
Colombia, Sede–Bogotá
###### Abstract.
In this paper we examine the well-posedness and ill-posedeness of the Cauchy
problems associated with a family of equations of ZK-KP-type
$\begin{cases}u_{t}=u_{xxx}-\mathscr{H}D_{x}^{\alpha}u_{yy}+uu_{x},\cr
u(0)=\psi\in Z\end{cases}$
in anisotropic Sobolev spaces, where $1\leq\alpha\leq 1$, $\mathscr{H}$ is the
Hilbert transform and $D_{x}^{\alpha}$ is the fractional derivative, both with
respect to $x$.
###### Key words and phrases:
Cauchy Problem, Dispersive equations, Kadomtsev-Petviashvili equation,
Zakharov-Kuznetsov equation, Local well-posedness, Ill-posedness, Anisotropic
Sobolev spaces, Kato theory
###### 2010 Mathematics Subject Classification:
35Q53, 35Q35, 35A01
## Introduction
Nonlinear evolution equations play an important role in different areas of
science and engineering. Some of them are worth mentioning: fluid mechanics,
plasma physics, fiber optics, solid state physics, chemical kinetics, chemical
physics and geochemistry, among others. From the study of their solutions, an
attempt is made to understand the effects of dispersion, diffusion, reaction
and convection associated with the models described by them. For example, the
Korteweg-de Vries (KdV) equation
$u_{t}=u_{xxx}+uu_{x}\qquad(x,t)\in\mathbb{R}^{2},$ (1)
which models the behavior of water waves in shallow channels, has solitary
waves as solutions that behave like particles, which is why Kruskal and
Zabusky called them _solitons_ in their 1965 work ([37]). These solitons are
stable, in the sense that if a solution of the KdV equation (equation (1))
that differs very little in shape from soliton-type solutions, at the
beginning, its shape will maintain an aspect that will differ very little from
the shape of a soliton-type solution through time (see [4] and [6]); in fact,
these solutions eventually take the form of solitons (see [31]). From a
practical point of view, the notion of soliton stability guarantees that, with
meticulous care, in the laboratory we will be able to reproduce these
phenomena, first observed by J. Scott Russell in 1834.
The Benjamin-Bona-Mahony (BBM) equation
$u_{t}+u_{x}+uu_{x}-u_{xxt}=0,$ (2)
was introduced in [5] with the intention to modeling the propagation of long
waves of small amplitude, where the dispersion effect is purely nonlinear. The
way in which this was obtained, it was pursued to arrive at an equation
equivalent to the equation KdV (1). It is interesting to note that despite
this intention, from a purely mathematical point of view, these equations
present significant and interesting differences.
Other one-dimensional equations are, one, introduced independently by Benjamin
in [3] and Ono in [30],
$u_{t}+\mathscr{H}u_{xx}+uu_{x}=0.$ (3)
which models the internal waves into deep stratified fluids, where
$\mathscr{H}$ is the Hilbert transform. The another, is the regularized
Benjamin-Ono (rBO)
$u_{t}+u_{x}+uu_{x}+\mathscr{H}u_{xt}=0$ (4)
where $u=u(x,t)$ is a real function, with $x,t\in\mathbb{R}$. This equation is
a model for wave time evolution with large ridges at the interface between two
immissible fluids.
There are two-dimensional versions that extend the equations previously
mentioned. In the case of the KdV equation we have the Kadomtsev-Petviashvilli
(KP) equation, see [10],
$(u_{t}+auu_{x}+u_{xxx})_{x}\pm u_{yy}=0,$ (5)
that describes waves in thin films of high surface tension. Another one is the
Zakharov-Kuznetsov (ZK) equation, see [38],
$u_{t}=(u_{xx}+u_{yy})_{x}+uu_{x},$ (6)
which arises in the study of the dynamics of geophysical fluids in isotropic
sets (means in which the characteristics of the bodies do not depend on
direction) and ionic acoustic waves in magnetic plasmas.
As a two-dimensional extension of the Benjamin-Ono equation we consider the
following family of equations
$\left(u_{t}+u^{p}u_{x}+\mathscr{H}(u_{xx}+\alpha u_{yy})\right)_{x}-\gamma
u_{yy}=0\quad p\in\mathbb{N}.$ (7)
This is the model of weakly nonlinear dispersive long wave motion in a two-
fluid system, where the interface is capillarized and the bottom fluid is
infinitely deep (see [1], [2] and [18]). Another one version that has received
some attention in recent literature is the ZK-BO equation,
$u_{t}=(\mathscr{H}u_{x}+u_{yy})_{x}+uu_{x}.$ (8)
To finish this introduction we will mention the equations that present a more
general type of dispersion that include as particular cases those mentioned
above. A case of these is the integro differential equation of Whitham
$u_{t}+\alpha uu_{x}+k*u_{x}=0.$ (9)
This was introduced by Whitman in [36] to model the breakage of dispersive
waves in water. It is clear that if $k=\delta+\delta^{\prime\prime}$ or
$k=\mathrm{v.p.}\dfrac{1}{x}$ we have the aforementioned KdV and BO equations.
There is also what Lannes and Saut called the fractional dispersion equation
KdV, fKdV (see [22])
$u_{t}+\alpha uu_{x}+\partial_{x}D^{\alpha}u=0,$ (10)
where $D=\sqrt{-\partial_{x}^{2}}$. In the papers [8], [19], [24] and [25] it
is established the relationship between the parameter $\alpha$ and the arising
of singularities or the global existence of solutions in their energy space,
as well as how this behavior is related to the existence and stability of the
solitary waves associated with these equations. It should be said that in this
direction there are interesting open problems, although they can really be
very difficult.
At this time we also think it is important to mention the work of Kenig,
Martel and Robbiano ([16]) that demonstrates the blow-up of solutions in the
energy space of the generalized scattering equation BO (a slightly different
version of the fKdV)
$u_{t}+|u|^{2\alpha}u_{x}+\partial_{x}D^{\alpha}u=0,$ (11)
when the initial condition is “larger” than the “shape” of the solitary wave
associated with this equation.
In the two-dimensional case we have the equation introduced by Lannes in [21]
to model long waves of small amplitude in weakly transverse regimes
$u_{t}+\mu\frac{3}{2}uu_{x}+c_{ww}(\sqrt{\mu}|D^{\mu}|)\left(1+\frac{D_{x}^{2}}{D_{y}^{2}}\right)^{\frac{1}{2}}u=0,$
(12)
where
$c_{ww}(k)=\left((1+\beta k^{2})\frac{\tanh(k)}{k}\right)^{\frac{1}{2}}\
\text{ and}\ |D^{\mu}|=\sqrt{D_{x}^{2}+\mu D_{y}^{2}}.$
In [22] the relation between the equations KP and FDKP is established. In
fact, they highlight the difference in the dispersive character with respect
to the parameter $\beta$. Furthermore, by limiting the parameter $\mu$, the
equations KPII or KPI are obtained when $\beta=0$ or $\beta>0$ respectively.
An interesting conjecture here is to regard the existence of solitary waves
for values $\beta$ greater than $1/3$.
In this paper we consider the ZK-KP type Cauchy problem
$\begin{cases}u_{t}=u_{xxx}-\mathscr{H}D_{x}^{\alpha}u_{yy}+uu_{x},\cr
u(0)=\psi\in Z\end{cases}$ (13)
for $-1\leq\alpha\leq 1$, where $\mathscr{H}$ denotes the Hilbert transform in
the $x$ variable defined by
$\mathscr{H}(f)=\mathrm{p.v.}\frac{1}{\pi}\int\frac{f(y)}{x-y}\,dy\quad f\in
H^{s}(\mathbb{R}),$
for each $f\in\mathcal{S}$, $D_{x}^{\alpha}$ is the homogeneous fractional
derivative in $x$ variable defined by
$\widehat{D_{x}^{\alpha}f}(\xi,\eta)=|\xi|^{\alpha}\hat{f}(\xi,\eta),$
and $Z$ is one of the Sobolev spaces $X^{s_{1},s_{2}}$,
$\widehat{X}^{s_{1},s_{2}}$, $Y^{s_{1},s_{2}}$ and
$\widehat{Y}^{s_{1},s_{2}}$, that we will specify in the notations.
The cases $\alpha=1$ and $\alpha=-1$ are the Cauchy problems corresponding to
the very popular ZK and KPI equations, respectively, that were mentioned
earlier. If $u(x,y,t)$ is a solution to (13), then $u_{\lambda}$ given by
$u_{\lambda}(x,y,t)=\lambda^{2}u(\lambda
x,\lambda^{\frac{3-\alpha}{2}}y,\lambda^{3}t)$
is also a solution to (13) and
$\left\|u_{\lambda}(t)\right\|_{\dot{H}^{s_{1},s_{2}}(\mathbb{R}^{2})}=\lambda^{s_{1}+\frac{3+\alpha}{4}}\left\|u(t)\right\|_{\dot{H}^{s_{1},s_{2}}(\mathbb{R}^{2})}$
with $s_{2}=\frac{2s_{1}}{3-\alpha}$. This suggests that the local well-
posedness could be guaranteed in
$H^{s_{1},\frac{2s_{1}}{3-\alpha}}(\mathbb{R}^{2})$ for
$s_{1}\geq-\frac{3+\alpha}{4}$.
In this work we propose to show the local well-posedness of the Cauchy problem
(13) in the Sobolev spaces $Z$ mentioned above. For this purpose, we use
Kato’s theory for quasilinear equations and the ideas introduced by Kenig [15]
for the KP-I equation and developed by Linares, Pilod and Saut in [26] for the
f-KPI and f-KPII equations. Specifically, it is making the use of the
Strichartz estimate provided by the group generated by the homogeneous linear
equation associated with (13), making use of energy estimates. We will also
make some ill-posedness observations of this equation for $-1\leq\alpha<0$.
For this we will use the ideas developed by Molinet, Saut and Tzvetkov in
[29]. More precisely, we will show that the flow associated with the solutions
of (13) is not of class $C^{2}$. This, in particular, implies that it cannot
be applied the Picard iteration method to get a solution to the integral
equation obtained by the Duhamel principle applied to (13).
## Notation
1. (1)
$\mathcal{S}(\mathbb{R}^{2})=\mathcal{S}$ denotes the Schwartz space and
$\mathcal{S}^{\prime}(\mathbb{R}^{2})=\mathcal{S}^{\prime}$ denotes its
topological dual vector space, the tempered distributions.
2. (2)
$H^{s}(\mathbb{R}^{2})=H^{s}$ is the $s^{th}$-order Sobolev space.
3. (3)
For a variable or an operator $u$, we denote by $\langle u\rangle$ the
expression $(1+u^{2})^{\frac{1}{2}}$.
4. (4)
For $s_{1},s_{2}\in\mathbb{R}$, the anisotropic Sobolev space
$H^{s_{1},s_{2}}(\mathbb{R}^{2})$ is defined by
$H^{s_{1},s_{2}}(\mathbb{R}^{2})=\left\\{f\in\mathcal{S}^{\prime}\;\left|\;\int_{\mathbb{R}^{2}}{(\langle\xi\rangle^{2s_{1}}+\langle\eta\rangle^{2s_{2}})|\widehat{f}(\xi,\eta)|^{2}d\xi
d\eta}<\infty\right.\right\\}.$
The norm in this space is given by
$\|f\|_{H^{s_{1},s_{2}}(\mathbb{R}^{2})}=\sqrt{\int_{\mathbb{R}^{2}}(\langle\xi\rangle^{2s_{1}}+\langle\eta\rangle^{2s_{2}})|\widehat{f}(\xi,\eta)|^{2}d\xi
d\eta},$
for all $f$ in this space. When there is no risk of confusion, we denote this
space by $H^{s_{1},s_{2}}$. Observe that $H^{s,s}=H^{s}$, and the immediately
above norm is equivalent to the usually given in the literature.
5. (5)
$D_{x}^{s}$, $D_{y}^{s}$, $J_{x}^{s}$, $J_{y}^{s}$ and $J^{s}$ denotes the
operators defined, via Fourier transform, by
$\begin{split}\widehat{D_{x}^{s}f}&=|\xi|^{s}\widehat{f},\\\
\widehat{D_{y}^{s}f}&=|\eta|^{s}\widehat{f},\\\
\widehat{J_{x}^{s}f}&=(1+|\xi|^{2})^{\frac{s}{2}}\widehat{f},\\\
\widehat{J_{y}^{s}f}&=(1+|\eta|^{2})^{\frac{s}{2}}\widehat{f}\quad\text{and}\\\
\widehat{J^{s}f}&=(1+|\xi|^{2}+|\eta|^{2})^{\frac{s}{2}}\widehat{f},\end{split}$
for any $f\in\mathcal{S}^{\prime}(\mathbb{R}^{2})$.
6. (6)
For $s_{1},s_{2}\geq 0$, we denote by
$X^{s_{1},s_{2}}(\mathbb{R}^{2})=X^{s_{1},s_{2}}$ the space
$X^{s_{1},s_{2}}(\mathbb{R}^{2})=\\{f\in
H^{s_{1},s_{2}}\;|\;\partial_{x}^{-1}f\in H^{s_{1},s_{2}}\\}.$
The norm in this space is given by
$\|f\|_{X^{s_{1},s_{2}}}^{2}=\|f\|_{H^{s_{1},s_{2}}}^{2}+\|\partial^{-1}_{x}f\|^{2}_{H^{s_{1},s_{2}}}.$
7. (7)
For $s_{1},s_{2}\geq 0$, we denote by
$\widehat{X}^{s_{1},s_{2}}(\mathbb{R}^{2})=\widehat{X}^{s_{1},s_{2}}$ the
space
$\widehat{X}^{s_{1},s_{2}}(\mathbb{R}^{2})=\\{f\in
H^{s_{1},s_{2}}\;|\;\partial_{x}^{-1}f\in L^{2}\\}.$
The norm in this space is
$\|f\|^{2}_{\widehat{X}^{s_{1},s_{2}}}=\|f\|^{2}_{H^{s_{1},s_{2}}}+\|\partial^{-1}_{x}f\|^{2}_{L^{2}}.$
8. (8)
For $s_{1},s_{2}\geq 0$, we denote by
$X_{\alpha}^{s_{1},s_{2}}(\mathbb{R}^{2})=X_{\alpha}^{s_{1},s_{2}}$ the space
$X_{\alpha}^{s_{1},s_{2}}(\mathbb{R}^{2})=\\{f\in
H^{s_{1},s_{2}}\;|\;\partial_{x}^{\frac{\alpha-1}{2}}f\in L^{2}\\}.$
The norm in this space is given by
$\|f\|_{X^{s_{1},s_{2}}}^{2}=\|f\|_{H^{s_{1},s_{2}}}^{2}+\|\partial^{-1}_{x}f\|^{2}_{H^{s_{1},s_{2}}}.$
9. (9)
For $s_{1},s_{2}\geq 0$, we denote by
$Y^{s_{1},s_{2}}(\mathbb{R}^{2})=Y^{s_{1},s_{2}}$ the space
$Y^{s_{1},s_{2}}(\mathbb{R}^{2})=\\{f\in
H^{s_{1},s_{2}}\;|\;\partial_{x}^{-1}\partial_{y}f\in H^{s_{1},s_{2}}\\}.$
The norm in this space is
$\|f\|^{2}_{Y^{s_{1},s_{2}}}=\|f\|^{2}_{H^{s_{1},s_{2}}}+\|\partial^{-1}_{x}\partial_{y}f\|^{2}_{H^{s_{1},s_{2}}}$
10. (10)
For $s_{1},s_{2}\geq 0$, we denote by
$\widehat{Y}^{s_{1},s_{2}}(\mathbb{R}^{2})=\widehat{Y}^{s_{1},s_{2}}$ the
space
$\widehat{Y}^{s_{1},s_{2}}(\mathbb{R}^{2})=\\{f\in
H^{s_{1},s_{2}}\;|\;\partial_{x}^{-1}\partial_{y}f\in L^{2}\\}.$
The norm in this space is given by
$\|f\|_{\widehat{Y}^{s_{1},s_{2}}}^{2}=\|f\|_{H^{s_{1},s_{2}}}^{2}+\|\partial^{-1}_{x}\partial_{y}f\|^{2}_{L^{2}}.$
11. (11)
We define $H^{\infty}(\mathbb{R}^{2})=\bigcap_{s_{1},s_{2}\geq
0}H^{s_{1},s_{2}}(\mathbb{R}^{2})$. Analogously $X^{\infty}$,
$\widehat{X}^{\infty}$, $Y^{\infty}$ and $\widehat{Y}^{\infty}$.
## 1\. Preliminaries
We start this section of preliminaries by observing that the spaces
$X^{s_{1},s_{2}}$, $\widehat{X}^{s_{1},s_{2}}$, $X_{\alpha}^{s_{1},s_{2}}$,
$Y^{s_{1},s_{2}}$ and $\widehat{Y}^{s_{1},s_{2}}$ are Hilbert spaces. Thanks
to the next lemma, each of these spaces is dense in $L^{2}$
###### Lemma 1.1.
The space $\partial_{x}\mathcal{S}$ is dense in $L^{2}$. In general, it is
dense also in $H^{s_{1},s_{2}}$, $X^{s_{1},s_{2}}$,
$\widehat{X}^{s_{1},s_{2}}$, $X_{\alpha}^{s_{1},s_{2}}$, $Y^{s_{1},s_{2}}$ and
$\widehat{Y}^{s_{1},s_{2}}$.
###### Proof.
Take a non negative function $\phi\in C^{\infty}$ defined on the real numbers,
identically zero on the interval $[-1/2,1/2]$ and identically $1$ out of
$[-1,1]$. For any $\psi\in\mathcal{S}$ we define $\psi_{\lambda}$ using the
equation
$\widehat{\psi}_{\lambda}(\xi,\eta)=\phi(\lambda\xi)\widehat{\psi}(\xi,\eta),$
for all $(\xi,\eta)\in\mathbb{R}^{2}$. The Plancherel theorem allows us show
that $\psi_{\lambda}$ converges to $\psi$ in $L^{2}$, as $\lambda\to\infty$.
The same argument allows us to show that $\partial_{x}\mathcal{S}$ is dense in
any of the spaces mentioned in the lemma statement. ∎
By a duality argument, we can conclude that $L^{2}$ is densely contained in
the dual spaces of any of the spaces mentioned in the lemma.
As consequence of the previously discussed, we can extend the operator
$\partial^{3}_{x}-\mathscr{H}D_{x}^{\alpha}\partial^{2}_{y}$ to the entire
$L^{2}$, with image in the $X^{3}$ dual, $(X^{3})^{*}$. Indeed,
$-\partial^{3}_{x}+\mathscr{H}D_{x}^{\alpha}\partial^{2}_{y}$ is bounded from
$X^{3}$ to $L^{2}$. So its adjoint operator is bounded from $L^{2}$ to
$(X^{3})^{*}$. Thanks to the Fourier transform, it can be seen that this
adjoint operator is an extension of
$\partial^{3}_{x}-\mathscr{H}D_{x}^{\alpha}\partial^{2}_{y}$ to all $L^{2}$.
As a corollary of above, we have the following lemma.
###### Lemma 1.2.
Let $W_{\alpha}$ be the unitary group of operators generated by the operator
$\partial^{3}_{x}-\mathscr{H}D_{x}^{\alpha}\partial^{2}_{yy}$ and let $f$ and
$u$ be continuous functions from an open interval $I$ to $L^{2}$. Then
$u=W_{\alpha}(t)\psi+\int_{0}^{t}W_{\alpha}(t-t^{\prime})f(t^{\prime})\,dt^{\prime}$
if, and only if $u$ has continuous derivative on $I$ with values in
$(X^{3})^{*}$, the dual space $X^{3}$, and
$\partial_{t}u=\partial^{3}_{x}u-\mathscr{H}D_{x}^{\alpha}\partial^{2}_{y}u+f.$
(14)
This lemma gives sense to the well-posedness results in $H^{s_{1},s_{2}}$,
that we shall enunciate later, when $\alpha$ is negative.
The next result is a technical lemma that we use later.
###### Lemma 1.3.
Let us assume that $u$ and $f$ are continuous functions in the $L^{2}$ space
and satisfy (14) in the sense described there. Then,
$\frac{1}{2}\frac{d}{dt}\|u\|^{2}=(u,f).$ (15)
###### Proof.
Assume that $u_{\lambda}=e^{\lambda(\bigtriangleup-D_{x}^{-1})}(u)$ and
$f_{\lambda}$ is defined in the same way. $u_{\lambda}$ and $f_{\lambda}$ are
in $X^{\infty}$ and they are uniformly convergent, as $\lambda$ tends to 0, on
closed intervals to $u$ and $f$ in $L^{2}$. Also, they satisfy the equation
(14). From the antisimmetry of the operator
$\partial^{3}_{x}-\mathscr{H}D_{x}^{\alpha}\partial^{2}_{y}$, it is easy to
see that $u_{\lambda}$ and $f_{\lambda}$ satisfy (15). When we make $\lambda$
tend to 0, we get the lemma. ∎
Next we shall state a set of results about the properties of the spaces with
we work in this paper. Maybe one the most known of these results is the
Sobolev lemma. Here we present a version for the
$H^{s_{1},s_{2}}(\mathbb{R}^{2})$ spaces.
###### Lemma 1.4 (Sobolev).
Let $s_{1}$ and $s_{2}$ be positive real numbers such that
$\frac{1}{s_{1}}+\frac{1}{s_{2}}<2.$ Then,
$H^{s_{1},s_{2}}(\mathbb{R}^{2})\subset C_{\infty}(\mathbb{R}^{2})$ (the set
of continuous functions on $\mathbb{R}^{2}$ vanishing at infinity), with
continuous embedding.
###### Proof.
See [33] ∎
###### Lemma 1.5.
Let $1\leq p<q\leq\infty$ and $f\in L^{p}\cap L^{q}$. Then $f\in L^{r}$ for
$r=\theta p+(1-\theta)q$, where $\theta\in(0,1)$, and we have
$\left\|f\right\|_{L^{r}}^{r}\leq\left\|f\right\|_{L^{p}}^{\theta
p}\left\|f\right\|_{L^{q}}^{(1-\theta)q}$
###### Proof.
The proof is immediate consequence of the Hölder inequality. ∎
###### Lemma 1.6.
If $s\in(0,n/2)$, then $H^{s}(\mathbb{R}^{n})$ is a continuous embedding in
$L^{p}(\mathbb{R}^{n})$, with $p=\frac{2n}{n-2s}$, i.e.
$s=n(\frac{1}{2}-\frac{1}{p})$. Furthermore, for $f\in H^{s}(\mathbb{R}^{n})$,
$s\in(0,n/2)$
$\left\|f\right\|_{L^{p}(\mathbb{R}^{n})}\leq
c_{n,s}\left\|D^{s}f\right\|_{L^{2}(\mathbb{R}^{n})}\leq
c\left\|f\right\|_{H^{s}(\mathbb{R}^{n})}$
where
$D^{l}f=(-\Delta)^{l/2}=((|\xi|)^{l}\hat{f})^{\vee}$
###### Proof.
See the Linares and Ponce book [27] page 48. ∎
###### Lemma 1.7.
Let $s_{1},s_{2}\in\mathbb{R}$ and assume that $D^{s_{1}}f\in
L^{p}(\mathbb{R})$ and $D^{s_{2}}f\in L^{q}(\mathbb{R})$. Then, for all
$\theta\in[0,1]$, $D^{s}f\in L^{r}$, and
$\|D^{s}f\|_{L^{r}(\mathbb{R})}\leq
C_{s}\|D^{s_{1}}f\|_{L^{p}(\mathbb{R})}^{\theta}\|D^{s_{2}}f\|_{L^{q}(\mathbb{R})}^{1-\theta}$
where $\theta=\dfrac{s_{2}-s}{s_{2}-s_{1}}$ and
$\dfrac{1}{r}=\dfrac{\theta}{p}+\dfrac{1-\theta}{q}$.
###### Proof.
A proof can be found in [28] ∎
The next estimate was proved by Kato and Ponce in [14] and it will be very
useful later in this work.
###### Lemma 1.8.
For $s>0$ and $1<p<\infty$, we have
$\|[J^{s},f]g\|_{L^{p}(\mathbb{R}^{n})}\lesssim\|\partial
f\|_{L^{\infty}(\mathbb{R}^{n})}\|J^{s-1}g\|_{L^{p}(\mathbb{R}^{n})}+\|J^{s}f\|_{L^{p}(\mathbb{R}^{n})}\|g\|_{L^{\infty}(\mathbb{R}^{n})},$
(16)
for all $f$ and $g\in\mathcal{S}(\mathbb{R}^{n})$
###### Corollary 1.9.
For $s>0$ and $p\in(1\,\infty)$, $L_{s}^{p}\cap L^{\infty}$ is an algebra,
also
$\|fg\|_{L_{s}^{p}}\leq
c(\|f\|_{\infty}\|g\|_{L_{s}^{p}}+\|f\|_{L_{s}^{p}}\|g\|_{\infty})$ (17)
The following estimate for the commutator operator can be seen in [27] page
51.
###### Lemma 1.10.
For $s>0$, we have
$\left\|\left[\partial_{x}^{s},g\right]f\right\|_{L^{2}(\mathbb{R})}\lesssim\left\|\partial_{x}g\right\|_{L^{\infty}(\mathbb{R})}\left\|\partial_{x}^{s-1}f\right\|_{L^{2}(\mathbb{R})}+\left\|\partial_{x}^{s}g\right\|_{L^{2}(\mathbb{R})}\left\|f\right\|_{L^{\infty}(\mathbb{R})}$
(18)
The next result is the Leibniz rule for fractional derivatives and it was
proved by Kenig, Ponce and Vega in [17].
###### Lemma 1.11.
For $\alpha\in(0,1)$, we have
$\left\|D_{x}^{\alpha}(fg)\right\|_{L^{p}(\mathbb{R})}\leq\left\|D_{x}^{\alpha}(f)\right\|_{L^{p_{1}}(\mathbb{R})}\left\|g\right\|_{L^{q_{1}}(\mathbb{R})}+\left\|D_{x}^{\alpha}g\right\|_{L^{p_{2}}(\mathbb{R})}\left\|f\right\|_{L^{q_{2}}(\mathbb{R})}$
(19)
where $1<p_{1},p_{2},q_{1},q_{2}\leq\infty$ and satisfy
$\frac{1}{p_{1}}+\frac{1}{q_{1}}=\frac{1}{p_{2}}+\frac{1}{q_{2}}=\frac{1}{p}$.
###### Lemma 1.12.
Let $-1\leq\alpha\leq 1$ and $0\leq p\leq\frac{8}{1-\alpha}$
$\|f\|_{L^{p+2}(\mathbb{R}^{2})}^{p+2}\lesssim\|f\|_{L^{2}(\mathbb{R}^{2})}^{2-\frac{p(1-\alpha)}{4}}\|\partial_{x}f\|_{L^{2}(\mathbb{R}^{2})}^{\frac{p(3-\alpha)}{4}}\|D^{\frac{\alpha-1}{2}}_{x}\partial_{y}f\|_{L^{2}(\mathbb{R}^{2})}^{\frac{p}{2}}$
###### Proof.
First, let us prove the lemma for $p=p*=\frac{8}{1-\alpha}$. From Lemmas 1.6
and 1.7 we have
$\begin{split}\left\|f\right\|_{L^{p^{*}+2}(\mathbb{R}^{2})}^{p^{*}+2}&=\int_{\mathbb{R}^{2}}{\left|f(x,y)\right|^{p^{*}+2}}dxdy\\\
&=\int_{\mathbb{R}}{\left\|f(\cdot,y)\right\|_{L^{p^{*}+2}_{x}}^{p^{*}+2}}dy\\\
&\leq
C\int_{\mathbb{R}}{\left\|D^{\frac{p^{*}}{p^{*}+2}}_{x}f(\cdot,y)\right\|_{L^{2}_{x}}^{p^{*}+2}}dy\\\
&\leq
C\int_{\mathbb{R}}{\left\|\partial_{x}f(\cdot,y)\right\|_{L^{2}_{x}}^{2}\left\|D^{\frac{p^{*}-4}{2p^{*}}}_{x}f(\cdot,y)\right\|_{L^{2}_{x}}^{p^{*}}}dy\\\
&\leq
C\left\|\partial_{x}f\right\|_{L^{2}(\mathbb{R}^{2})}^{2}\sup_{y\in\mathbb{R}}{\left\|D^{\frac{p^{*}-4}{2p^{*}}}_{x}f(x,y)\right\|_{L^{2}_{x}}^{p^{*}}}.\end{split}$
(20)
On the other hand, for all $y\in\mathbb{R}$,
$\begin{split}\left\|D^{\frac{p^{*}-4}{2p^{*}}}_{x}f(\cdot,y)\right\|_{L^{2}_{x}}^{2}&=\int_{\mathbb{R}}{\left|D_{x}^{\frac{p^{*}-4}{2p^{*}}}f(x,y)\right|^{2}}dx\\\
&=2\int_{\mathbb{R}}{\int_{-\infty}^{y}{D_{x}^{\frac{1+\alpha}{4}}f(x,\eta)D_{x}^{\frac{1+\alpha}{4}}\partial_{y}f(x,\eta)}}d\eta
dx\\\
&=2\int_{-\infty}^{y}{\int_{\mathbb{R}}{D_{x}f(x,\eta)D_{x}^{\frac{\alpha-1}{2}}}\partial_{y}f(x,\eta)}dxd\eta\\\
&\leq
2\int_{-\infty}^{y}{\left\|\partial_{x}f(\cdot,\eta)\right\|_{L^{2}_{x}}\left\|D^{\frac{\alpha-1}{2}}_{x}\partial_{y}f(\cdot,\eta)\right\|_{L^{2}_{x}}}d\eta\\\
&\leq
2\left\|\partial_{x}f\right\|_{L^{2}(\mathbb{R}^{2})}\left\|D^{\frac{\alpha-1}{2}}_{x}\partial_{y}f\right\|_{L^{2}(\mathbb{R}^{2})}.\end{split}$
(21)
Therefore,
$\|f\|_{L^{p^{*}+2}(\mathbb{R}^{2})}^{p^{*}+2}\leq
c\|\partial_{x}f\|_{L^{2}(\mathbb{R}^{2})}^{\frac{p^{*}(3-\alpha)}{4}}\|D^{\frac{\alpha-1}{2}}_{x}\partial_{y}f\|_{L^{2}(\mathbb{R}^{2})}^{\frac{p^{*}}{2}}$
(22)
The inequality for $0<p<p^{*}$ follows immediately from Lemma 1.5 and the last
inequality. ∎
### 1.1. Kato’s theory
We will make a brief presentation of Kato’s theory described in [11]. With
this it can be showed the well-posedness of the Cauchy problems associated to
linear and quasilinear evolution equations.
#### 1.1.1. Linear case
Suppose that $X$ and $Y$ are reflexives Banach spaces with $Y\subseteq{X}$ in
a dense and continuous way, and let $\\{A(t)\\}_{t\in[0,T]}$ be a operators
family such that
1. (1)
$A(t)\in{G}(X,1,\beta)$. In other words, $-A(t)$ generates a $C_{0}$-semigroup
such that
$\|e^{-sA(t)}\|\leq{e}^{\beta{s}},$
for all $s\in[0,\infty).$
2. (2)
There exists an isomorphism $S:Y\to{X}$ such that $SA(t)S^{-1}=A(t)+B(t)$,
where $B(t)\in{B}(X),$ for $0\leq{t}\leq{T},$ $t\to{B}(t)x$ is strongly
measurable, for each $x\in{X}$, and $t\to\|B(t)\|_{X}$ is integrable in
$[0,T]$.
3. (3)
$Y\subseteq{D}(A(t)),$ for $0\leq{t}\leq{T}$, and $t\rightarrow{A}(t)$ is
strongly continuous from $[0,T]$ to $B(Y,X)$.
###### Theorem 1.13.
Under the above conditions, there exists a operators family
$\\{U(t,s)\\}_{0\leq{s}\leq{t}\leq{T}}$ such that:
1. (1)
$U$ is strongly continuous from $\Delta\to{B}(X)$, where
$\Delta=\\{(t,s):0\leq{s}\leq{t}\leq{T}\\}$.
2. (2)
$U(t,s)U(s,r)=U(t,r)$ for $(t,s)$ and $(s,r)\in\Delta$, and $U(s,s)=I$.
3. (3)
$U(t,s)Y\subset{Y}$ and $U$ is strongly continuous from $\Delta\to{B}(Y)$.
4. (4)
$\dfrac{dU(t,s)}{dt}=-A(t)U(t,s),$ $\dfrac{dU(t,s)}{ds}=U(t,s)A(s),$ in the
strong sense in $B(X,Y)$ space and are strongly continuous from
$\Delta\to{B}(X,Y)$.
The operators family $\\{U(t,s)\\}_{0\leq{s}\leq{t}\leq{T}}$ in the previous
theorem is called _the evolution operators_ associated to ${A(t)}$. An
immediate consequence from the last theorem is, for $\varphi\in{Y}$,
$u(t)=U(t,s)\varphi$ is solution to the Cauchy problem
$\displaystyle\frac{du}{dt}+A(t)u=0\qquad\text{ for }\quad s\leq{t}\leq{T},$
$\displaystyle u(s)=\varphi.$
Moreover, if $f\in{C}([0,T];X)\cap{L^{1}}([0,T];Y)$, then
$u(t)=U(t,0)\varphi+\int_{0}^{t}U(t,s)f(s)ds$
if and only if $u\in{C}([0,T];Y)\cap{C}^{1}((0,T);X)$ and
$\displaystyle\frac{du}{dt}+A(t)u$ $\displaystyle=f(t)\quad\text{ for }\quad
0\leq{t}\leq{T},$ $\displaystyle u(0)$ $\displaystyle=\varphi.$
#### 1.1.2. Quasilinear Case
Let $X$ and $Y$ be reflexives Banach spaces, $Y\subseteq X$, with dense and
continuous embedding. Let us consider the following problem
$\begin{array}[]{ll}&\partial_{t}u+A(t,u)u=f(t,u)\in X,\ 0<t,\\\
&u(0)=u_{0}\in Y,\end{array}$ (23)
where, for each $t$, $A(t,u)$ is a linear operator from $Y$ to $X$ and
$f(t,u)$ is a function from $\mathbb{R}\times Y$ in $X$. Let us also consider
the next conditions:
$(X)$ There exists an isometric isomorphism $S$ from $Y$ to $X$.
There exist $T_{0}>0$ and $W$ an open ball with $w_{0}$ as center such that:
$(A_{1})$ For each $(t,y)\in[0,T_{0}]\times W$, the linear operator $A(t,y)$
belongs to $G(X,1,\beta)$, where $\beta$ is a positive real number. In the
other words, $-A(t,y)$ generate a $C_{0}$ semigroup such that
$\|e^{-sA(t,y)}\|_{\mathcal{B}(X)}\leq e^{\beta s},\ \text{for}\
s\in[0,\infty).$
Note that if $X$ is a Hilbert space, $A\in G(X,1,\beta)$ if, and only if,
1. a)
$\langle Ay,y\rangle_{X}\geq-\beta\|y\|_{X}^{2}$ for all $y\in D(A)$,
2. b)
$(A+\lambda)$ is onto for all $\lambda>\beta.$
(See [13] or [32])
$(A_{2})$ For each $(t,y)\in[0,T_{0}]\times W$ the operator
$B(t,y)=[S,A(t,y)]S^{-1}\in\mathcal{B}(X)$ and is uniformly bounded, i.e.,
there exists $\lambda_{1}>0$ such that
$\displaystyle\|B(t,y)\|_{\mathcal{B}(X)}\leq\lambda_{1}\ \ \text{for all}\
(t,y)\in[0,T_{0}]\times W,$ Furthermore, for some $\mu_{1}>0$, we have that,
for all $y$ and $z\in W$,
$\displaystyle\|B(t,y)-B(t,z)\|_{\mathcal{B}(X)}\leq\mu_{1}\|y-z\|_{Y}.$
$(A_{3})$ $Y\subseteq D(A(t,y))$, for each $(t,y)\in[0,T_{0}]\times W,$ (the
restriction of $A(t,y)$ to $Y$ belongs to $\mathcal{B}(Y,X)$) and, for each
$y\in W$ fix, $t\to A(t,y)$ is strongly continuous. Besides, for all
$t\in[0,T_{0}]$ fix, it is satisfied the following Lipschitz condition,
$\|A(t,y)-A(t,z)\|_{\mathcal{B}(Y,X)}\leq\mu_{2}\|y-z\|_{X},$
where $\mu_{2}\geq 0$ is constant.
$(A_{4})$ $A(t,y)w_{0}\in Y$ for all $(t,y)\in[0,T]\times W$. Also, there
exists a constant $\lambda_{2}$ such that
$\|A(t,y)w_{0}\|_{Y}\leq\lambda_{2},\ \text{for all}\ (t,y)\in[0,T_{0}]\times
W.$
$(f_{1})$ $f$ is a bounded function in $[0,T_{0}]\times W$ to $Y$, i.e., there
exists $\lambda_{3}$ such that
$\|f(t,y)\|_{Y}\leq\lambda_{3},\ \text{for all}\ (t,y)\in[0,T_{0}]\times W,$
Also, the function $t\in[0,T_{0}]\mapsto f(t,y)\in Y$ is continuous with
respect to the topology of $X$ and for all $y$ and $z\in Y$ we have that
$\displaystyle\|f(t,y)-f(t,z)\|_{X}\leq\mu_{3}\|y-z\|_{X},$
where $\mu_{3}\geq 0$ is a constant.
###### Theorem 1.14 (Kato).
Assume that the conditions $(X),$ $(A_{1})-(A_{4})$ and $(f_{1})$ are
satisfied. Given $u_{0}\in Y$, there exist $0<T<T_{0}$ and a unique $u\in
C([0,T];Y)\cap C^{1}((0,T);X)$ solution to (23). Furthermore, the map
$u_{0}\to u$ is continuous in the following sense: consider the sequence of
Cauchy problems,
$\displaystyle\partial_{t}u_{n}+A_{n}(t,u_{n})u_{n}=f_{n}(t,u_{n})\ \ t>0$
(24) $\displaystyle u_{n}(0)=u_{n_{0}}\ n\in\mathbb{N}.$
Suppose that the conditions $(X)$, $(A_{1})$–$(A_{4})$ and $(f_{1})$ are also
satisfied for all $n\geq 0$ in (LABEL:dependencia), with the same $X,\ Y$ and
$S$, and the correspondents $\beta$, $\lambda_{1}$–$\lambda_{3}$,
$\mu_{2}$–$\mu_{3}$ can be chosen independent of $n$. Also, let us suppose
that
$\displaystyle\mathop{s\text{-}\lim}_{n\to\infty}A_{n}(t,w)$
$\displaystyle=A(t,w)\ \text{in }\ B(X,Y),$
$\displaystyle\mathop{s\text{-}\lim}_{n\to\infty}B_{n}(t,w)$
$\displaystyle=B(t,w)\ \text{in }\ B(X),$
$\displaystyle\lim_{n\to\infty}f_{n}(t,w)$ $\displaystyle=f(t,w)\ \text{in }\
Y,$ $\displaystyle\lim_{n\to\infty}u_{n_{0}}$ $\displaystyle=u_{0}\ \text{in
}\ Y,$
where $s$-$\lim$ denotes the strong limit. Then, $T$ can be taken in such a
way that $u_{n}\in C([0,T],Y)\cap C^{1}((0,T),X)$ and
$\lim_{n\to\infty}\sup_{[0,T]}\|u_{n}(t)-u(t)\|_{Y}=0.$
A proof of this theorem can be found in [11] and [20].
### 1.2. Other results
###### Proposition 1.15 (Kato’s inequality).
Let $f\in H^{s}$, $s>2$, $\Lambda=(1-\Delta)^{1/2}$ and $M_{f}$ be the
multiplication operator by $f$. Then, for $|\tilde{t}|,|\tilde{s}|\leq s-1$,
$\Lambda^{-\tilde{s}}[\Lambda^{\tilde{s}+\tilde{t}+1},M_{f}]\Lambda^{-\tilde{t}}\in
B(L^{2}(\mathbb{R}^{2}))$ and
$\left\|\Lambda^{-\tilde{s}}[\Lambda^{\tilde{s}+\tilde{t}+1},M_{f}]\Lambda^{-\tilde{t}}\right\|_{{B}\left(L^{2}\left(\mathbb{R}^{2}\right)\right)}\leq
c\left\|\nabla f\right\|_{H^{s-1}}.$ (25)
###### Proposition 1.16.
Let $f:\mathbb{R}^{2}\to\mathbb{R}$ be a bounded continuous function such that
$\partial_{x}f$ exists and is continuous and bounded. Then, if
$A=f\partial_{x}$,
$\langle
A(u),u\rangle_{L^{2}}\geq-\frac{1}{2}||\partial_{x}f||_{L^{\infty}}||u||_{L^{2}}^{2},$
(26)
for all $u\in D(A)$, $A+\lambda$ is onto, for all
$\lambda>\frac{1}{2}||f||_{L^{\infty}}$. In particular, $A\in
G\left(L^{2}\left(\mathbb{R}^{2}\right),1,\frac{1}{2}||f||_{L^{\infty}}\right)$.
###### Proof.
The inequality (26) is obtained immediately after using the integration by
parts. Let us see that $A+\lambda$ is onto, if
$\lambda>\frac{1}{2}||f||_{L^{\infty}}$. Suppose that $\psi$ is such that
$\langle(A+\lambda)(u),u\rangle_{L^{2}}=0$, for all $u\in D(A)$. Then $\psi\in
D(A^{*})\subseteq D(A)$. From (26), it follows that
$0\geq\langle(a+\lambda)(u),u\rangle_{L^{2}}\geq(\lambda-\frac{1}{2}||f||_{L^{\infty}})||\psi||_{L^{2}}^{2}.$
Hence, $\psi=0$ and, therefore, $A+\lambda$ is onto. ∎
## 2\. Local well-posedness in Sobolev spaces of $s^{th}$ order with $s>2$
In this section we examine the local well-posedness of the problem (13) in the
Sobolev spaces $H^{s}$, $X^{s}$, $\widehat{X}_{\alpha}^{s}$, $Y^{s}$ and
$\widehat{Y}^{s}$, for $s>2$.
### 2.1. Local well-posedness in $H^{s}(\mathbb{R}^{2})$
In this section, we will make use of Kato’s theory to show the local well-
posedness of (13) in the $H^{s}$ spaces. More precisely we have the following
theorem.
###### Theorem 2.1.
Let $s$ and $\alpha$ be real numbers such that $s>2$ and $-1\leq\alpha\leq 1$.
For $\psi\in H^{s}(\mathbb{R}^{2})$, there exist $T>0$, that depends only on
$\|\psi\|_{H^{s}}$, and a unique $u\in C([0,T],H^{s}(\mathbb{R}^{2}))\cap$
$C^{1}([0,T],H^{s-3}(\mathbb{R}^{2})\cap(X^{3})^{*})$ solution to the Cauchy
problem (13)
Moreover, the map $\psi\to u$ from $H^{s}$ to $C([0,T],H^{s}(\mathbb{R}^{2}))$
is continuous.
###### Proof.
Let $W_{\alpha}(t)$ be the operators group defined by
$W_{\alpha}(t)\psi=e^{t(\partial_{x}^{3}-\mathscr{H}\partial_{y}^{2})}\psi=\left(e^{-it(\xi^{3}+\operatorname{sgn}(\xi)|\xi|^{\alpha}\eta^{2})}\widehat{\psi}\right)^{\vee},$
for all $\psi\in H^{s}$. $u$ is solution to the problem (13) if and only if
$v=W_{\alpha}(t)u$ is solution to the problem
$\begin{cases}v_{t}+A(t,v)v=0\\\ v(0)=\psi,\end{cases}$ (27)
where $A(t,v)=W_{\alpha}(t)(W_{\alpha}(-t)v)\partial_{x}W_{\alpha}(-t)$. Let
us see that this last problem satisfies each condition of Kato’s theorem
(Theorem 1.14).
Let $X=L^{2}(\mathbb{R}^{2})$, $Y=H^{s}(\mathbb{R}^{2})$ and
$S=\Lambda_{s}=J^{s}$. From Plancherel’s theorem, it is evident that $S$ is an
isomorphism between $X$ and $Y$.
With the following lemmas we show that the conditions $(A_{1})$-$(A_{4})$ are
satisfied.
###### Lemma 2.2.
$A(t,v)\in G(X,1,\beta(v))$, where
$\beta(v)=\frac{1}{2}\sup\limits_{t}\|\partial_{x}W_{\alpha}(t)v\|_{L^{\infty}}$
###### Proof.
Since $\\{W_{\alpha}(-t)\\}$ is an strongly continuous unitary operators group
and $u\in H^{s}(\mathbb{R}^{2})$, from Proposition 1.16, we obtain the result.
∎
###### Lemma 2.3.
For $S$ given as above,
$SA(t,v)S^{-1}=A(t,v)+B(t,v),$
where $B(t,v)$ is a bounded operator in $L^{2}$, for all $t\in\mathbb{R}$ and
all $v\in H^{s}$, and satisfies the inequalities
$\displaystyle||B(t,v)||_{\mathcal{B}(L^{2})}$ $\displaystyle\leq\lambda(v)$
(28) $\displaystyle||B(t,v)-B(t,v^{\prime})||_{\mathcal{B}(L^{2})}$
$\displaystyle\leq\mu||v-v^{\prime}||_{H^{s}}$ (29)
for $t\in\mathbb{R}$ and all $v$ and $v^{\prime}\in H^{s}(\mathbb{R}^{2})$.
Where $\mu$ is a positive real number and
$\lambda(v)=\sup\limits_{t}C_{s}\|v\|_{H^{s}}.$
###### Proof.
From Lemma 1.15, it follows that
$[S,W_{\alpha}(-t)v]S^{-1}\in\mathcal{B}(L^{2})$ y
$\|[S,W_{\alpha}(-t)v]S^{-1}\|_{\mathcal{B}(L^{2})}\leq C_{s}\|v\|_{H^{s}}.$
Therefore $B(t,v)\in\mathcal{B}(L^{2})$ and satisfies (28)
Proceeding as before (29) can be shown. ∎
###### Lemma 2.4.
$H^{s}(\mathbb{R}^{2})\subset D(A(t,v))$ and $A(t,v)$ is a bounded operator
from $Y=H^{s}(\mathbb{R}^{2})$ to $X=L^{2}(\mathbb{R}^{2})$ with
$\|A(t,v)\|_{\mathcal{B}(Y,X)}\leq\|v\|_{H^{s}},$
for all $v\in Y$. Furthermore, the function $t\mapsto A(t,v)$ is strongly
continuous from $\mathbb{R}$ to $\mathcal{B}(Y,X)$, for all $v\in H^{s}$. On
the other hand, the function $v\mapsto A(t,v)$ satisfies the following
Lipschitz condition
$\|A(t,v)-A(t,v^{\prime})\|_{\mathcal{B}(Y,X)}\leq\|v-v^{\prime}\|_{X},$
where $\mu$ is as the lemma above.
###### Proof.
Inasmuch as $\\{W_{\alpha}(t)\\}$ is a unitary group in $L^{2}$, from the
definition of $A(t,v)$, it follows that
$H^{s}\left(\mathbb{R}^{2}\right)\subset D(A(t,v))$. In fact,
$\displaystyle\|A(t,v)f\|_{L^{2}}$
$\displaystyle=\|W_{\alpha}(-t)v\partial_{x}W_{\alpha}(-t)f\|_{L^{2}}$
$\displaystyle\leq C_{s}\|v\|_{H^{s}}\|\partial_{x}f\|_{L^{2}}$
$\displaystyle\leq\|v\|_{H^{s}}\|f\|_{H^{s}},$
for all $f\in H^{s}$.
Now, for all $t,t^{\prime}\in\mathbb{R}$ and $f,v\in H^{s}$, we have
$\displaystyle\|A(t,v)f-A(t^{\prime},v)f\|_{L^{2}}$
$\displaystyle\leq\|\left(W_{\alpha}(t)-W_{\alpha}(t^{\prime})\right)W_{\alpha}(-t)v\partial_{x}W_{\alpha}(-t)f\|_{L^{2}}+$
$\displaystyle+\|(W_{\alpha}(-t)-W_{\alpha}(-t^{\prime}))v\partial_{x}W_{\alpha}(-t)f\|_{L^{2}}+$
$\displaystyle+\|W_{\alpha}(-t^{\prime})v\partial_{x}(W_{\alpha}(-t)-W_{\alpha}(-t^{\prime}))f\|_{L^{2}}.$
Since the group $\\{W_{\alpha}(t)\\}_{t\in\mathbb{R}}$ is strongly continuous,
$t\mapsto A(t,v)$ is strongly continuous from $\mathbb{R}$ to
$\mathcal{B}(H^{s},L^{2})$. Finally, for any $t\in\mathbb{R}$, we have
$\displaystyle\|A(t,v)f-A(t,v^{\prime})f\|_{L^{2}}$
$\displaystyle\leq\mu\|W_{\alpha}(-t)v-W_{\alpha}(-t)v^{\prime}\|_{L^{2}}\|\partial_{x}W_{\alpha}(-t)f\|_{L^{\infty}}$
$\displaystyle\leq\mu\|v-v^{\prime}\|\,\|f\|_{H^{s}},$
this ends the lemma proof. ∎
If we take the open ball $W$ of $v\in H^{s}$ such that
$\|v\|_{H^{s}(\mathbb{R}^{2})}<R$, the preceding lemmas show that the Cauchy
problem (13) satisfies the conditions of Theorem 1.14. Therefore, for each
$\psi\in H^{s}(\mathbb{R}^{2})$, with $s>2$, there exists $T>0$, that depends
on $\|\psi\|_{H^{s}}$, and a unique $v\in C([0,T],H^{s}(\mathbb{R}^{2}))\cap
C^{1}([0,T],H^{s-1}(\mathbb{R}^{2}))$ solution to the problem (27). Moreover,
the map $\psi\to v$ is continuous from $H^{s}(\mathbb{R}^{2})$ to
$C([0,T],H^{s}(\mathbb{R}^{2}))$. Now, from the group $W_{\alpha}(t)$
properties it can be verified that $u(t)=W_{\alpha}(-t)v$ is solution to
Cauchy problem (13) and satisfies all properties stated in the theorem. ∎
###### Theorem 2.5.
The existence time of the solution of the Cauchy problem (13) can be chosen
independently of $s$ in the following sense: if $u\in
C([0,T],H^{s}(\mathbb{R}^{2}))$ is the solution to (13) with $\psi\in
H^{r}(\mathbb{R}^{2})$, for some $r>s$, then $u\in
C([0,T],H^{r}(\mathbb{R}^{2}))$.
In particular, if $\psi\in H^{\infty}(\mathbb{R}^{2})$, $u\in
C([0,T],H^{\infty}(\mathbb{R}^{2}))$
###### Proof.
Let $r>s$, $u\in C([0,T],H^{r}(\mathbb{R}^{2}))$ solution to (13) and
$v=W_{\alpha}(-t)u$. Suppose that $r\leq s+1$. If we apply $\partial_{x}^{2}$
on both sides of the differential equation (27), we arrive to the following
linear evolution equation for $w(t)=\partial_{x}^{2}v(t)$
$\frac{dw}{dt}+A(t)w+B(t)w=0.$ (30)
where
$\displaystyle A(t)$
$\displaystyle=\partial_{x}W_{\alpha}(t)u(t)W_{\alpha}(-t)$ (31) and
$\displaystyle B(t)$ $\displaystyle=2W_{\alpha}(t)u_{x}(t)W_{\alpha}(-t).$
(32)
Since $v\in C\left([0,T];H^{s}\left(\mathbb{R}^{2}\right)\right)$, then $w\in
C\left([0,T];H^{s-2}\left(\mathbb{R}^{2}\right)\right)$. Besides
$w(0)=\psi_{xx}\in H^{r-2}\left(\mathbb{R}^{2}\right)$, because $\psi\in
H^{r}\left(\mathbb{R}^{2}\right)$. It is needed to see that $w\in
C\left([0,T];H^{r-2}\left(\mathbb{R}^{2}\right)\right)$. For this we shall
prove that the Cauchy problem associated to the linear equation (30) is
locally well-posed for $1-s\leq k\leq s-1$, for which we have the following
lemma whose proof is similar to that of Lemma 3.1 in [12].
###### Lemma 2.6.
The family $\\{A(t)\\}_{0\leq t\leq T}$ has a unique family of evolution
operators associated, $\\{U(t,\tau)\\}_{0\leq t\leq\tau\leq T}$, in the spaces
$X=H^{h}$, $Y=H^{k}$, where
$-s\leq h\leq s-2\quad 1-s\leq k\leq s-1\quad k+1\leq h.$ (33)
In particular, $U(t,\tau):H^{r}\to H^{r}$ for $-s\leq s\leq s-1.$
Then, $w$ satisfies the equation
$w(t)=U(t,0)\psi_{xx}+\displaystyle\int_{0}^{t}U(t,\tau)[-B(\tau)w(\tau)+f(\tau)]\,d\tau.$
(34)
Since $\psi_{xx}\in H^{r-2},$ $B(t)$, given by (32), is an operators family in
$H^{r-2}$ that is strongly continuous for $t$ in the interval $[0,T]$. From
Lemma 2.6, the solution to (34) belongs to
$C\left([0,T];H^{r-2}\left(\mathbb{R}^{2}\right)\right)$. In other words,
$\partial_{x}^{2}u\in C\left([0,T];H^{r-2}\left(\mathbb{R}^{2}\right)\right)$.
If $w_{1}(t)=\partial_{x}\partial_{y}v(t)$, we have
$\frac{dw_{1}}{dt}+A(t)w_{1}+B_{1}(t)w_{1}=f_{1}(t),$ (35)
where
$\displaystyle B_{1}(t)$
$\displaystyle=\mathcal{W}(t)u_{x}(t)\mathcal{W}(-t)=\frac{1}{2}B(t),$ (36)
and $\displaystyle f_{1}(t)$
$\displaystyle=-\mathcal{W}(t)\left(u_{xx}(t)u_{y}(t)\right).$ (37)
As before, we have
$w_{1}(t)=U(t,0)\psi_{xy}+\displaystyle\int_{0}^{t}U(t,\tau)(-B_{1}(\tau)w_{1}(\tau)+f_{1}(\tau))\,d\tau.$
(38)
Since $u_{xx}\in C\left([0,T];H^{r-2}\left(\mathbb{R}^{2}\right)\right)$,
$f_{1}\in C\left([0,T];H^{r-2}\left(\mathbb{R}^{2}\right)\right)$. Inasmuch
as, also, $B_{1}(t)\in\mathbf{(}H^{r-2}\left(\mathbb{R}^{2}\right))$ is
strongly continuous on the interval $[0,T]$. Arguing as before we have that
$w_{1}\in C\left([0,T];H^{r-2}\left(\mathbb{R}^{2}\right)\right)$, or
equivalently $u_{xy}\in
C\left([0,T];H^{r-2}\left(\mathbb{R}^{2}\right)\right)$.
Analogously, if $w_{2}(t)=\partial_{y}^{2}v(t)$, we have that
$\frac{dw_{2}}{dt}+A(t)w_{2}=f_{2}(t),$ (39)
where
$f_{2}(t)=-2\mathcal{W}(t)(u_{xy}u_{y}(t).$ (40)
Therefore,
$\displaystyle
w_{2}(t)=U(t,0)\psi_{yy}+\int_{0}^{t}U(t,\tau)f_{2}(\tau)\,d\tau.$ (41)
Since $u_{xy}\in C\left([0,T];H^{r-2}\left(\mathbb{R}^{2}\right)\right)$,
$f_{2}\in C\left([0,T];H^{r-2}\left(\mathbb{R}^{2}\right)\right)$. Repeating
the argument above, we can conclude that $w_{2}\in
C\left([0,T];H^{r-2}\left(\mathbb{R}^{2}\right)\right)$, or equivalently,
$\partial_{y}^{2}u\in C\left([0,T];H^{r-2}\left(\mathbb{R}^{2}\right)\right)$.
Hence, we have showed that if $s<r\leq s+1$ and $\psi\in H^{r}$, $u\in
C\left([0,T];H^{r}\left(\mathbb{R}^{2}\right)\right)$. To see the case
$r>s+1$, since $\psi\in H^{s^{\prime}}$, for $s^{\prime}<r$, we can use over
and over again what we have proved so far, to obtain that $u\in
C\left([0,T];H^{r}\left(\mathbb{R}^{2}\right)\right)$ ∎
### 2.2. Local well-posedness in $X^{s}$, $\widehat{X}_{\alpha}^{s}$ and
$Y^{s}$
Let us finish this section showing the local well-posedness of (13) in the
spaces $X^{s}$, $\widehat{X}_{\alpha}^{s}$ and $Y^{s}$.
###### Theorem 2.7.
Let $s$ and $\alpha$ be as in Theorem 2.1. Let $Z$ also be any of the spaces
$X^{s}$, $\widehat{X}^{s}$, $\widehat{X}_{\alpha}^{s}$, $Y^{s}$ and
$\widehat{Y}^{s}$. Then, if $\psi\in Z$ and $u\in C([0,T],H^{s})$ is solution
to (13) with $u(0)=\psi$, $u\in C([0,T],Z)$. Moreover, $\psi\mapsto u$ is
continuous from $Z$ to $C([0,T],Z)$
###### Proof.
Suppose that $\psi$ and $u$ are as in the hypothesis of the theorem. Then,
from the fundamental calculus theorem, we have that
$u(t)=W_{\alpha}(-t)\psi+\int_{0}^{t}W_{\alpha}(-t+\tau)\partial_{x}\left(\frac{u^{2}(\tau)}{2}\right)\,d\tau.$
So that,
$\partial_{x}^{-1}u(t)=W_{\alpha}(-t)\partial_{x}^{-1}\psi+\int_{0}^{t}W_{\alpha}(-t+\tau)\left(\frac{u^{2}(\tau)}{2}\right)\,d\tau.$
From here it follows that (13) is locally well-posed in the spaces
$\widehat{X}^{s}$ and $X^{s}$.
On the other hand,
$\partial_{x}^{-1}\partial_{y}u(t)=W_{\alpha}(-t)\partial_{x}^{-1}\partial_{y}\psi+\int_{0}^{t}W_{\alpha}(-t+\tau)(u\partial_{y}u)(\tau)\,d\tau.$
Then, (13) is locally well-posed in the space $\widehat{Y}^{s}$.
To see the local well-posedness in $Y^{s}$ is slightly more complicated. Let
$v$ be as in (27). Hence, $w=\partial_{x}^{-1}\partial_{y}v$ satisfies
$\begin{cases}w_{t}+A(t,v)w=0\\\
w(0)=\partial_{x}^{-1}\partial_{y}\psi,\end{cases}$ (42)
where $A(t,v)$ is as in (27). From linear case of Kato’s theory, more
specifically from Theorem 1.14, when taking $X=L^{2}$ and $Y=H^{s}$, we have
that, if $\partial_{x}^{-1}\partial_{y}\psi\in H^{s}$, $w$ is continuous on
$t$ and depends continuously on this data. This shows the local well-posedness
of the problem (13) in $Y^{s}$.
Analogously we can show that (13) is local well-posedness in space
$X_{\alpha}^{s}$. ∎
## 3\. Local well-posedness in low regularity spaces
In this section we examine the local well-posedness of equation (13) in
$H^{s_{1},s_{2}}(\mathbb{R}^{2})$, for $-1\leq\alpha\leq 1$,
$s_{1}>\frac{17}{12}-\frac{\alpha}{4}$ and $1<s_{2}\leq s_{1}$. We will use
the dispersive properties of the group generated by the linear equation
associated to (13). This is the same strategy used by Kenig in [15] for the
KP-I equation and by Linares, Pilod and Saut in [26] for the f-KPI and f-KPII
equations.
### 3.1. Linear estimates
The linear Cauchy problem associated to the problem (13) is
$\begin{cases}u_{t}=u_{xxx}-\mathscr{H}D_{x}^{\alpha}u_{yy},\cr
u(0)=\psi.\end{cases}$ (43)
where $-1\leq\alpha\leq 1$. The unitary group generated by this problem is
$W_{\alpha}(t)\psi(x,y)=\left(e^{-it(\xi^{3}+\operatorname{sgn}(\xi)\left|\xi\right|^{\alpha}\eta^{2})}\widehat{\psi}\right)^{\vee}=S_{t}^{\alpha}*\psi(x,y),$
(44)
where
$S_{t}^{\alpha}(x,y)=\int_{\mathbb{R}^{2}}{e^{-it(\xi^{3}+\operatorname{sgn}(\xi)\left|\xi\right|^{\alpha}\eta^{2})+ix\xi+iy\eta}}d\xi
d\eta$ (45)
We shall examine this group properties.
###### Lemma 3.1.
Let $-1\leq\alpha\leq 1$ and
$\frac{\alpha}{2}-1<\operatorname{Re}\beta<\frac{\alpha}{2}$. Then,
$\left\|D_{x}^{\beta}W_{\alpha}(t)\psi\right\|_{L^{\infty}(\mathbb{R}^{2})}\lesssim\left|t\right|^{-\frac{5+2\beta-\alpha}{6}}\left\|\psi\right\|_{L^{1}(\mathbb{R}^{2})}.$
(46)
###### Proof.
Effectively,
$\begin{split}D_{x}^{\beta}S_{t}^{\alpha}(x,y)&=\int_{\mathbb{R}^{2}}{\left|\xi\right|^{\beta}e^{-it(\xi^{3}+\operatorname{sgn}(\xi)\left|\xi\right|^{\alpha}\eta^{2})+ix\xi+iy\eta}}d\xi
d\eta\\\
&=\int_{\mathbb{R}_{\xi}}{\left|\xi\right|^{\beta}e^{-it\xi^{3}+ix\xi}\int_{\mathbb{R}_{\eta}}{e^{-it\operatorname{sgn}(\xi)\left|\xi\right|^{\alpha}\eta^{2}+iy\eta}}}d\eta
d\xi\\\
&=\int_{\mathbb{R}_{\xi}}{\left|\xi\right|^{\beta}e^{-it\xi^{3}+ix\xi+\frac{i\operatorname{sgn}(\xi)y^{2}}{4t\left|\xi\right|^{\alpha}}}\int_{\mathbb{R}_{\eta}}{e^{-i\left|t\right|\operatorname{sgn}(t\xi)\left|\xi\right|^{\alpha}\left[\eta-\frac{y}{2t\operatorname{sgn}(\xi)\left|\xi\right|^{\alpha}}\right]^{2}}}}d\eta
d\xi\\\
&=\frac{\pi^{\frac{1}{2}}}{\left|t\right|^{\frac{1}{2}}}\int_{\mathbb{R}}{e^{-it\xi^{3}+ix\xi+\frac{i\operatorname{sgn}(\xi)y^{2}}{4t\left|\xi\right|^{\alpha}}-i\operatorname{sgn}(t\xi)\frac{\pi}{4}}}\left|\xi\right|^{\beta-\frac{\alpha}{2}}d\xi\\\
&=\frac{\pi^{\frac{1}{2}}}{|t|^{\frac{5+2\beta-\alpha}{6}}}\int_{\mathbb{R}}{e^{i\left(\theta^{3}-x\theta
t^{-\frac{1}{3}}-\frac{\operatorname{sgn}(\theta)y^{2}}{4|\theta|^{\alpha}}|t|^{\frac{\beta}{3}-1}\right)}|\theta|^{\beta-\frac{\alpha}{2}}}d\theta.\end{split}$
(47)
Let us see that the last integral is bounded. For this let us take a function
$\chi$ defined on all $\mathbb{R}$, infinitely differentiable, with support in
the interval $[-2,2]$ such that $\chi\equiv 1$ in the interval $[-1,1]$. So,
$\int_{\mathbb{R}}{e^{i\left(\theta^{3}-x\theta
t^{-\frac{1}{3}}-\frac{\operatorname{sgn}(\theta)y^{2}}{4|\theta|^{\alpha}}|t|^{\frac{\beta}{3}-1}\right)}|\theta|^{\beta-\frac{\alpha}{2}}}d\theta=\\\
=\int_{\mathbb{R}}{e^{i\left(\theta^{3}-x\theta
t^{-\frac{1}{3}}-\frac{\operatorname{sgn}(\theta)y^{2}}{4|\theta|^{\alpha}}|t|^{\frac{\beta}{3}-1}\right)}|\theta|^{\beta-\frac{\alpha}{2}}}\chi(\theta)d\theta+\\\
+\int_{\mathbb{R}}{e^{i\left(\theta^{3}-x\theta
t^{-\frac{1}{3}}-\frac{\operatorname{sgn}(\theta)y^{2}}{4|\theta|^{\alpha}}|t|^{\frac{\beta}{3}-1}\right)}|\theta|^{\beta-\frac{\alpha}{2}}}(1-\chi(\theta))d\theta.$
(48)
Clearly the first integral on the right hand side is bounded. To see that the
second integral on the right hand side is bounded, we will make use the Van
der Corput lemma (see [27], Corollary 1.1). The second and third derivatives
of phase function $\vartheta_{\alpha}(\theta)=\theta^{3}-x\theta
t^{-\frac{1}{3}}-\frac{\operatorname{sgn}(\theta)y^{2}}{4|\theta|^{\alpha}}|t|^{\frac{\beta}{3}-1}$
in the integral are
$\displaystyle\vartheta_{\alpha}^{\prime\prime}(\theta)=6\theta+\operatorname{sgn}(\theta)\alpha(\alpha+1)\frac{y^{2}\left|t\right|^{\frac{\beta}{3}-1}}{4}\left|\theta\right|^{-\alpha-2}$
and
$\displaystyle\vartheta_{\alpha}^{\prime\prime\prime}(\theta)=6+\alpha(\alpha+1)(\alpha+2)\frac{y^{2}\left|t\right|^{\frac{\beta}{3}-1}}{4}\left|\theta\right|^{-\alpha-3}.$
It is easily verified that, for $|\theta|\geq 1$,
$|\vartheta_{\alpha}^{\prime\prime}(\theta)|\geq 6$ when $-1\leq\alpha\leq 0$
and that $|\vartheta_{\alpha}^{\prime\prime\prime}(\theta)|\geq 6$ when
$0\leq\alpha\leq 1$. Since the function
$\theta\mapsto|\theta|^{\beta-\frac{\alpha}{2}}(1-\chi(\theta))$ is uniformly
bounded and integrable on the set $|\theta|\geq 1$, it is verified that the
second integral on the right side of (48) is bounded. Which verifies that the
last integral in (47) is uniformly bounded. Therefore,
$\|D_{x}^{\beta}S_{t}^{\alpha}(x,y)\|_{L^{\infty}(\mathbb{R}^{2})}\lesssim|t|^{\frac{5+2\beta-\alpha}{6}}$
The theorem follows immediately from Young’s inequality for convolution. ∎
###### Remark 1.
The last result coincides with that proved by Linares and Pastor in [23] for
the ZK equation (case $\alpha=1$). The results of Saut in [34], for the KPI
equation (case $\alpha=-1$), and of Lizarazo in his PhD thesis (see [28])
(case $\alpha=0$) are better estimates.
###### Corollary 3.2.
Let $-1\leq\alpha\leq 1$, $0<\epsilon<1$ and $0\leq\theta\leq 1$. Then,
$\left\|D_{x}^{\theta\left(\frac{\alpha}{2}-\epsilon\right)}W_{\alpha}(t)\psi\right\|_{L^{p}(\mathbb{R}^{2})}\leq\left|t\right|^{-\frac{\theta(5-2\epsilon)}{6}}\left\|\psi\right\|_{L^{p^{\prime}}(\mathbb{R}^{2})}$
(49)
where $\frac{1}{p}+\frac{1}{p^{\prime}}=1$ and $p=\frac{2}{1-\theta}$.
###### Proof.
Let $\psi\in L^{1}(\mathbb{R}^{2})\cap L^{2}(\mathbb{R}^{2})$. The corollary
follows from Stein’s interpolation theorem for analytical operator families
(see [35]). Let us set $z=\theta+i\gamma\in\mathbb{C}$. For each $z$ we define
the operators $T^{z}$ by
$T_{z}\psi=D_{x}^{z(\frac{\alpha}{2}-\epsilon)}W_{\alpha}(t)\psi.$
This family $\left\\{T_{z}\right\\}$ is an admissible family of operators. By
Lemma 3.1, we have
$\left\|T_{1+i\gamma}\psi\right\|_{L^{\infty}(\mathbb{R}^{2})}=\left\|D_{x}^{(\frac{\alpha}{2}-\epsilon)(1+i\gamma)}W_{\alpha}(t)\psi\right\|_{L^{\infty}(\mathbb{R}^{2})}\leq
c\left|t\right|^{-(\frac{5-2\epsilon}{6})}\left\|\psi\right\|_{L^{1}(\mathbb{R}^{2})}.$
(50)
Also, since $\left\\{W_{\alpha}(t)\right\\}_{t\in\mathbb{R}}$ is an strongly
continuous group on the parameter $t$ in $L^{2}(\mathbb{R}^{2})$, we have
$\left\|T_{i\gamma}\psi\right\|_{L^{2}(\mathbb{R}^{2})}=\left\|D_{x}^{(\frac{\alpha}{2}-\epsilon)(i\gamma)}W_{\alpha}(t)\psi\right\|_{L^{2}(\mathbb{R}^{2})}=\left\|\psi\right\|_{L^{2}(\mathbb{R}^{2})}.$
(51)
From Stein’s interpolation theorem we obtain
$\left\|T_{\theta}\psi\right\|_{L^{p}(\mathbb{R}^{2})}\leq
c\left|t\right|^{-\theta(\frac{5-2\epsilon}{6})}\left\|\psi\right\|_{L^{p^{\prime}}(\mathbb{R}^{2})},$
(52)
what was we wanted to prove. ∎
###### Corollary 3.3.
Let $-1\leq\alpha\leq 1$, $0<\epsilon<1$ and $0<\theta\leq 1$. Then,
$\left\|D_{x}^{\frac{12}{q\left(5-2\epsilon\right)}\left(\frac{\alpha}{2}-\epsilon\right)}\int_{-\infty}^{\infty}{W_{\alpha}(t-t^{\prime})F(t^{\prime})}dt^{\prime}\right\|_{L_{T}^{q}L_{xy}^{p}}\lesssim\left\|F\right\|_{L_{T}^{q^{\prime}}L^{p^{\prime}}_{xy}},$
(53)
where $\frac{1}{q}+\frac{1}{q^{\prime}}=\frac{1}{p}+\frac{1}{p^{\prime}}=1$,
$p=\frac{2}{1-\theta}$, $\frac{2}{q}=\frac{\theta(5-2\epsilon)}{6}$ and
$\frac{1}{p}+\frac{1}{q}=\frac{1}{2}-\frac{\theta(1+2\epsilon)}{12}$.
###### Proof.
From Minkowski’s inequality, Corollary 3.2 and from the Hardy-Littlewood-
Sobolev theorem, we have
$\left\|D_{x}^{\theta\left(\frac{\alpha}{2}-\epsilon\right)}\int_{-\infty}^{\infty}{W_{\alpha}(t-t^{\prime})F(t^{\prime})}dt^{\prime}\right\|_{L_{t}^{q}L_{xy}^{p}}\\\
=\left\|\left\|\int_{-\infty}^{\infty}D_{x}^{\theta\left(\frac{\alpha}{2}-\epsilon\right)}W_{\alpha}(t-t^{\prime})F(\cdot,\cdot,t^{\prime})dt^{\prime}\right\|_{L_{xy}^{p}}\right\|_{L_{t}^{q}}\\\
\leq\left\|\int_{-\infty}^{\infty}\left\|D_{x}^{\theta\left(\frac{\alpha}{2}-\epsilon\right)}W_{\alpha}(t-t^{\prime})F(\cdot,\cdot,t^{\prime})\right\|_{L_{xy}^{p}}dt^{\prime}\right\|_{L_{t}^{q}}\\\
\lesssim\left\|\int_{-\infty}^{\infty}{\left|t-t^{\prime}\right|^{-\frac{\theta(5-2\epsilon)}{6}}\left\|F(\cdot,\cdot,t^{\prime})\right\|_{L_{xy}^{p^{\prime}}}}dt^{\prime}\right\|_{L_{t^{q}}}\\\
\lesssim\left\|\left\|F(\cdot,\cdot,t^{\prime})\right\|_{L_{xy}^{p^{\prime}}}\right\|_{L_{t^{q^{\prime}}}}=\left\|F\right\|_{L_{t}^{q^{\prime}}L_{xy}^{p^{\prime}}}.$
∎
Using the Stein-Thomas argument (see [9]) we have.
###### Proposition 3.4.
Let $\alpha$, $\epsilon$, $\theta$, $p$ and $q$ be as in the previous
corollary. Then,
$\left\|D_{x}^{\frac{6}{q\left(5-2\epsilon\right)}\left(\frac{\alpha}{2}-\epsilon\right)}W_{\alpha}(t)\psi\right\|_{L_{T}^{q}L_{xy}^{p}}\lesssim\left\|\psi\right\|_{L_{xy}^{2}}.$
(54)
From the above we have the following two very useful corollaries in the proof
of the local well-posedness in this section.
###### Corollary 3.5.
For each $T>0$ and $0<\epsilon<1$, we have that
$\left\|W_{\alpha}(t)\psi\right\|_{L_{T}^{2}L_{xy}^{\infty}}\lesssim
T^{\frac{1+2\epsilon}{12}}\left\|D_{x}^{\frac{1}{2}(\epsilon-\frac{\alpha}{2})}\psi\right\|_{L_{xy}^{2}}.$
(55)
###### Proof.
From Hölder’s inequality and Proposition 3.4 we have that
$\begin{split}\left\|W_{\alpha}(t)\psi\right\|_{L_{T}^{2}L_{xy}^{\infty}}&\leq
T^{\frac{1+2\epsilon}{12}}\left\|W_{\alpha}(t)\psi\right\|_{L_{T}^{\frac{12}{5-2\epsilon}}L_{xy}^{\infty}}\lesssim
T^{\frac{1+2\epsilon}{12}}\left\|D_{x}^{\frac{1}{2}(\epsilon-\frac{\alpha}{2})}\psi\right\|_{L_{xy}^{2}}.\end{split}$
∎
Just like in [15] and [26], we prove a refined Strichartz estimate for the
solution to non-homogeneous linear problem
$\partial_{t}w=w_{xxx}-\mathscr{H}D_{x}^{\alpha}w_{yy}+F.$ (56)
So, we have the next lemma.
###### Lemma 3.6.
Let $-1\leq\alpha\leq 1,0<\epsilon<1$ and $T>0$. If $w$ is solution to (56),
then there exists $c_{\epsilon}>0$ such that
$\left\|\partial_{x}w\right\|_{L_{T}^{1}L_{xy}^{\infty}}\leq
c_{\epsilon}T^{\frac{7+2\epsilon}{12}}\left(\sup_{t\in[0,T]}{\left\|J_{x}^{\frac{17}{12}-\frac{\alpha}{4}+\frac{2\epsilon}{3}}w\right\|_{L_{xy}^{2}}}+\int_{0}^{T}\left\|D_{x}^{\frac{5}{12}-\frac{\alpha}{4}+\frac{2\epsilon}{3}}F(t)\right\|_{L_{xy}^{2}}dt\right).$
(57)
###### Proof.
We will make use of a Littlewood-Paley type decomposition of $w$ on $x$
variable. For this, let $\varphi_{0},\varphi\in C^{\infty}_{0}$ with
$\operatorname{supp}(\varphi_{0})=\left\\{\left|\xi\right|<2\right\\}$ and
$\operatorname{supp}(\varphi)=\left\\{\frac{1}{2}<\left|\xi\right|<2\right\\}$
such that $\varphi_{0}(\xi)+\sum_{k=1}^{\infty}{\varphi(2^{-k}\xi)}=1$ for all
$\xi\in\mathbb{R}$. Let us define $P_{k}w$ via Fourier transform by
$\widehat{P_{0}w}(\xi,\eta)=\varphi_{0}(\xi)\widehat{w}(\xi,\eta)$ and
$\widehat{P_{k}w}(\xi,\eta)=\varphi(2^{-k}\xi)\widehat{w}(\xi,\eta)$ for
$k\geq 1$, in such a way that
$w=\sum_{k=0}^{\infty}{P_{k}w}.$
Let us first make an estimate for
$\left\|\partial_{x}P_{0}w\right\|_{L_{T}^{1}L_{xy}^{\infty}}$. Since $w$
satisfies (56), then $P_{0}w$ satisfies the integral equation when apply
$P_{0}$ on both sides of the equation. So, from the Cauchy-Schwarz inequality
and (55), it follows that
$\begin{split}&\left\|\partial_{x}P_{0}w\right\|_{L_{T}^{1}L_{xy}^{\infty}}\leq\\\
&\leq\left\|W_{\alpha}(t)\partial_{x}P_{0}w(0)\right\|_{L_{T}^{1}L_{xy}^{\infty}}+\int_{0}^{t}\left\|W_{\alpha}(t-t^{\prime})\partial_{x}P_{0}F(t^{\prime})\right\|_{L_{T}^{1}L_{xy}^{\infty}}dt^{\prime}\\\
&\leq
T^{\frac{1}{2}}\left(\left\|W_{\alpha}(t)\partial_{x}P_{0}w(0)\right\|_{L_{T}^{2}L_{xy}^{\infty}}+\int_{0}^{t}\left\|W_{\alpha}(t-t^{\prime})\partial_{x}P_{0}F(t^{\prime})\right\|_{L_{T}^{2}L_{xy}^{\infty}}dt^{\prime}\right)\\\
&\lesssim
T^{\frac{1}{2}+\frac{1+2\epsilon}{12}}\left(\left\|D_{x}^{\frac{1}{2}(\epsilon-\frac{\alpha}{2})}\partial_{x}P_{0}w(0)\right\|_{L_{xy}^{2}}+\int_{0}^{t}\left\|D_{x}^{\frac{1}{2}(\epsilon-\frac{\alpha}{2})}\partial_{x}P_{0}F(t^{\prime})\right\|_{L_{xy}^{2}}dt^{\prime}\right)\\\
&\lesssim
c_{\epsilon}T^{\frac{7+2\epsilon}{12}}\left(\left\|J_{x}^{\frac{17}{12}-\frac{\alpha}{4}+\frac{2\epsilon}{3}}P_{0}w(0)\right\|_{L_{xy}^{2}}+\int_{0}^{T}\left\|D_{x}^{\frac{5}{12}-\frac{\alpha}{4}+\frac{2\epsilon}{3}}P_{0}F(t)\right\|_{L_{xy}^{2}}dt\right).\end{split}$
(58)
Now, let us estimate
$\left\|\partial_{x}P_{k}w\right\|_{L_{T}^{1}L_{xy}^{\infty}}$, when $k\geq
1$. For this, we will make a suitable partition in time in such a way that
allow us to control the localized frequencies associated to $x$ variable. So,
let $\mathcal{P}=\left\\{a_{0},a_{1},...,a_{2^{k}}\right\\}$ be a partition
from interval $[0,T]$ with $a_{j}=jT2^{-k}$, $j=0,1,\cdots,2^{k}$. We denote
by $I_{j}$ the interval $[a_{j-1},a_{j}]$.
Then, thanks to Young’s inequality, Cauchy-Schwarz’s inequality and the
inequality
$\left\|(\chi_{\left\\{2^{k-1}<\left|\xi\right|<2^{k}\right\\}}\xi)^{\vee}\right\|_{L^{1}(\mathbb{R})}\lesssim
2^{k},$ we have
$\begin{split}\left\|\partial_{x}P_{k}w\right\|_{L_{T}^{1}L_{xy}^{\infty}}&\leq\left\|(\chi_{\left\\{2^{k-1}<\left|\xi\right|<2^{k}\right\\}}(\xi)\left|\xi\right|)^{\vee}\right\|_{L^{1}(\mathbb{R})}\left\|P_{k}w\right\|_{L_{T}^{1}L_{xy}^{\infty}}\\\
&\lesssim
2^{k}\sum_{j=1}^{2^{k}}\left\|P_{k}w\right\|_{L_{I_{j}}^{1}L_{xy}^{\infty}}\\\
&\lesssim(2^{k})\sum_{j=1}^{2^{k}}(a_{j}-a_{j-1})^{\frac{1}{2}}\left\|P_{k}w\right\|_{L_{I_{j}}^{2}L_{xy}^{\infty}}\\\
&=(2^{k}T)^{\frac{1}{2}}\sum_{j=1}^{2^{k}}\left\|P_{k}w\right\|_{L_{I_{j}}^{2}L_{xy}^{\infty}}.\end{split}$
(59)
From Duhamel’s principle on each interval $[a_{j-1},a_{j}]$, we have that, for
each $t\in[a_{j-1},a_{j}]$,
$P_{k}w(t)=W_{\alpha}(t-a_{j})P_{k}w(\cdot,a_{j})+\int_{a_{j-1}}^{t}W_{\alpha}(t-t^{\prime})P_{k}F(t^{\prime})dt^{\prime}.$
By (59) and (55), we get
$\begin{split}&\left\|\partial_{x}P_{k}w\right\|_{L_{T}^{1}L_{xy}^{\infty}}\lesssim(2^{k}T)^{\frac{1}{2}}\sum_{j=1}^{2^{k}}\left\|P_{k}w\right\|_{L_{[a_{j},b_{j}]}^{2}L_{xy}^{\infty}}\\\
&\lesssim(2^{k}T)^{\frac{1}{2}}\sum_{j=1}^{2^{k}}{\left(\left\|W_{\alpha}(t-a_{j})P_{k}w(\cdot,a_{j})\right\|_{L_{I_{j}}^{2}L_{xy}^{\infty}}+\int_{a_{j-1}}^{t}{\left\|W_{\alpha}(t-t^{\prime})P_{k}F(t^{\prime})\right\|_{L_{I_{j}}^{2}L_{xy}^{\infty}}}dt^{\prime}\right)}\\\
&\lesssim
2^{\frac{k(5+2\epsilon)}{12}}T^{\frac{7+2\epsilon}{12}}\sum_{j=1}^{2^{k}}{\left(\left\|D_{x}^{\frac{1}{2}\left(\epsilon-\frac{\alpha}{2}\right)}P_{k}w(\cdot,a_{j})\right\|_{L_{xy}^{2}}+\int_{I_{j}}{\left\|D_{x}^{\frac{1}{2}\left(\epsilon-\frac{\alpha}{2}\right)}P_{k}F(t^{\prime})\right\|_{L_{xy}^{2}}}dt^{\prime}\right)}\\\
&\lesssim
T^{\frac{7+2\epsilon}{12}}\Bigg{\\{}2^{\frac{k(5+2\epsilon)}{12}}\sum_{j=1}^{2^{k}}{\left\|D_{x}^{\frac{1}{2}\left(\epsilon-\frac{\alpha}{2}\right)}P_{k}w(\cdot,a_{j})\right\|_{L_{xy}^{2}}}+\\\
&\phantom{T^{\frac{7+2\epsilon}{12}}\Bigg{\\{}2^{\frac{k(5+2\epsilon)}{12}}\sum_{j=1}^{2^{k}}\left\|D_{x}^{\frac{1}{2}\left(\epsilon-\frac{\alpha}{2}\right)}P_{k}w\right.}+2^{\frac{k(5+2\epsilon)}{12}}\int_{0}^{T}{\left\|D_{x}^{\frac{1}{2}\left(\epsilon-\frac{\alpha}{2}\right)}P_{k}F(t^{\prime})\right\|_{L_{xy}^{2}}}dt^{\prime}\Bigg{\\}}\\\
&\lesssim
c_{\epsilon}T^{\frac{7+2\epsilon}{12}}\left(\sup_{t\in[0,T]}{\left\|D_{x}^{\frac{17}{12}-\frac{\alpha}{4}+\frac{2\epsilon}{3}}P_{k}w(t)\right\|_{L_{xy}^{2}}}+\int_{0}^{T}{\left\|D_{x}^{\frac{5}{12}-\frac{\alpha}{4}+\frac{2\epsilon}{3}}P_{k}F(t^{\prime})\right\|_{L_{xy}^{2}}}dt^{\prime}\right).\end{split}$
(60)
For the sake of the completeness, let us explain the inequality
$2^{\frac{k(5+2\epsilon)}{12}}\sum_{j=1}^{2^{k}}{\left\|D_{x}^{\frac{1}{2}\left(\epsilon-\frac{\alpha}{2}\right)}P_{k}w(\cdot,a_{j})\right\|_{L_{xy}^{2}}}\leq\sup_{t\in[0,T]}{\left\|D_{x}^{\frac{17}{12}-\frac{\alpha}{4}+\frac{2\epsilon}{3}}P_{k}w(t)\right\|_{L_{xy}^{2}}},$
(61)
what was used to prove the inequality above. Observe that, for each integer
$j\leq 2^{k}$, we have
$\left\|D_{x}^{\frac{1}{2}\left(\epsilon-\frac{\alpha}{2}\right)}P_{k}w(\cdot,a_{j})\right\|_{L_{xy}^{2}}\leq
j^{-\frac{k(17+2\epsilon)}{12}}\left\|D_{x}^{\frac{17}{12}-\frac{\alpha}{4}+\frac{2\epsilon}{3}}P_{k}w(\cdot,a_{j})\right\|_{L_{xy}^{2}}.$
Summing on $j$ from $1$ to $2^{k}$ and considering
$\sum_{j=1}^{2^{k}}j^{-\frac{k(17+2\epsilon)}{12}}\sim
2^{-\frac{k(5+2\epsilon)}{12}}$
it follows the inequality (61).
(60) together with (58) show the present lemma. ∎
### 3.2. Energy Estimate
###### Lemma 3.7.
Let $-1\leq\alpha\leq 1$, $T>0$ and assume that $u\in
C([0,T];H^{\infty}(\mathbb{R}^{2}))$ is a solution to the Cauchy problem (13).
Then, there exists a positive constant $C$ such that, for $1\leq s_{2}\leq
s_{1}$,
$\left\|u\right\|_{L_{T}^{\infty}H_{xy}^{s_{1},s_{2}}}\leq\left\|\psi\right\|_{H^{s_{1},s_{2}}_{xy}}e^{{C(\left\|\partial_{x}u\right\|_{L_{T}^{1}L_{xy}^{\infty}}+\left\|\partial_{y}u\right\|_{L_{T}^{1}L_{xy}^{\infty}})}}.$
(62)
###### Proof.
Let $u$ as in the statement in lemma. Let us estimate first
$\left\|J_{x}^{s_{1}}u\right\|_{L^{2}_{xy}}$. Operating with $J_{x}^{s_{1}}$
and then multiplying by $J_{x}^{s_{1}}u$ on both sides of the equation (13),
and integrating with respect to $x,y$, we obtain
$\begin{split}\frac{1}{2}\frac{d}{dt}\int_{\mathbb{R}^{2}}{{\left(J_{x}^{s_{1}}u\right)}^{2}dxdy}&=\int_{\mathbb{R}^{2}}{J_{x}^{s_{1}}(u\partial_{x}u)J_{x}^{s_{1}}u\,dxdy}\\\
&=\int_{\mathbb{R}^{2}}\left[J_{x}^{s_{1}},u\right]\partial_{x}uJ_{x}^{s_{1}}udxdy+\int_{\mathbb{R}^{2}}uJ_{x}^{s_{1}}\partial_{x}uJ_{x}^{s_{1}}u\,dxdy.\end{split}$
(63)
From inequalities of Cauchy-Schwarz and Kato-Ponce (Lemma 1.8), applied in the
$x$ variable, we have
$\begin{split}\int_{\mathbb{R}^{2}}\left[J_{x}^{s_{1}},u\right]\partial_{x}uJ_{x}^{s_{1}}u\,dxdy&\leq\int_{\mathbb{R}_{y}}\left\|\left[J_{x}^{s_{1}},u\right]\partial_{x}u\right\|_{L_{x}^{2}}\left\|J_{x}^{s_{1}}u\right\|_{L_{x}^{2}}dy\\\
&\lesssim\int_{\mathbb{R}_{y}}\left\|\partial_{x}u\right\|_{L_{x}^{\infty}}\left\|J_{x}^{s_{1}}u\right\|^{2}_{L_{x}^{2}}dy\\\
&\leq\left\|\partial_{x}u\right\|_{L_{xy}^{\infty}}\left\|J_{x}^{s_{1}}u\right\|^{2}_{L_{xy}^{2}}\leq\left\|\partial_{x}u\right\|_{L_{xy}^{\infty}}\left\|u\right\|^{2}_{H_{xy}^{s_{1},s_{2}}}.\end{split}$
(64)
On the other hand,
$\int_{\mathbb{R}^{2}}uJ_{x}^{s_{1}}\partial_{x}uJ_{x}^{s_{1}}\,u\,dxdy=-\frac{1}{2}\int_{\mathbb{R}^{2}}\partial_{x}u\left(J_{x}^{s_{1}}u\right)^{2}dxdy\lesssim\left\|\partial_{x}u\right\|_{L_{xy}^{\infty}}\left\|J_{x}^{s_{1}}u\right\|^{2}_{L_{xy}^{2}}.$
(65)
By (63), (64) and (65), we have
$\frac{d}{dt}\left\|J_{x}^{s_{1}}u\right\|^{2}_{L^{2}_{xy}}\lesssim\left\|\partial_{x}u\right\|_{L_{xy}^{\infty}}\left\|J_{x}^{s_{1}}u\right\|^{2}_{L_{xy}^{2}}\lesssim\left\|\partial_{x}u\right\|_{L_{xy}^{\infty}}\left\|u\right\|^{2}_{H_{xy}^{s_{1},s_{2}}}.$
(66)
In a totally analogous way we shall estimate
$\left\|J_{y}^{s_{2}}u\right\|_{L^{2}_{xy}}$. So,
$\frac{1}{2}\frac{d}{dt}\left\|J_{y}^{s_{2}}u\right\|^{2}_{L^{2}_{xy}}=\int_{\mathbb{R}^{2}}\left[J_{y}^{s_{2}},u\right]\partial_{x}uJ_{y}^{s_{2}}u\,dxdy+\int_{\mathbb{R}^{2}}uJ_{y}^{s_{2}}\partial_{x}uJ_{y}^{s_{2}}u\,dxdy.$
(67)
Proceeding as before, we have
$\begin{split}&\int_{\mathbb{R}^{2}}\left[J_{y}^{s_{2}},u\right]\partial_{x}uJ_{y}^{s_{2}}u\,dxdy\leq\int_{\mathbb{R}_{x}}\left\|\left[J_{y}^{s_{2}},u\right]\partial_{x}u\right\|_{L_{y}^{2}}\left\|J_{y}^{s_{2}}u\right\|_{L_{y}^{2}}dx\lesssim\\\
&\lesssim\int_{\mathbb{R}_{x}}\left(\left\|\partial_{y}u\right\|_{L_{y}^{\infty}}\left\|J_{y}^{s_{2}-1}\partial_{x}u\right\|_{L^{2}_{y}}+\left\|\partial_{x}u\right\|_{L_{y}^{\infty}}\left\|J_{y}^{s_{2}}\partial_{x}u\right\|_{L^{2}_{y}}\right)\left\|J_{y}^{s_{2}}u\right\|_{L^{2}_{y}}dx\\\
&\leq\left\|\partial_{y}u\right\|_{L_{xy}^{\infty}}\left\|J_{y}^{s_{2}-1}\partial_{x}u\right\|_{L^{2}_{xy}}\left\|J_{y}^{s_{2}}u\right\|_{L^{2}_{xy}}+\left\|\partial_{x}u\right\|_{L_{xy}^{\infty}}\left\|J_{y}^{s_{2}}u\right\|_{L^{2}_{xy}}^{2}\\\
&\leq\left\|\partial_{y}u\right\|_{L_{xy}^{\infty}}\left\|J_{x}^{s_{2}}u\right\|_{L^{2}_{xy}}^{\frac{1}{s_{2}}}\left\|J_{y}^{s_{2}}u\right\|^{\frac{2s_{2}-1}{s_{2}}}_{L^{2}_{xy}}+\left\|\partial_{x}u\right\|_{L_{xy}^{\infty}}\left\|J_{y}^{s_{2}}u\right\|_{L^{2}_{xy}}^{2}\\\
&\lesssim\left(\left\|\partial_{x}u\right\|_{L_{xy}^{\infty}}+\left\|\partial_{y}u\right\|_{L_{xy}^{\infty}}\right)\left\|u\right\|_{H^{s_{1},s_{2}}_{xy}}^{2}.\end{split}$
(68)
So, integrating by parts in the second term on the right side of the
inequality (67) and from last inequality, we have
$\frac{d}{dt}\left\|J_{y}^{s_{2}}u\right\|^{2}_{L^{2}_{xy}}\lesssim\left(\left\|\partial_{x}u\right\|_{L_{xy}^{\infty}}+\left\|\partial_{y}u\right\|_{L_{xy}^{\infty}}\right)\left\|u\right\|_{H^{s_{1},s_{2}}_{xy}}^{2}.$
(69)
Gathering (66) and (69), we obtain
$\frac{d}{dt}\left\|u\right\|^{2}_{H^{s_{1},s_{2}}_{xy}}\lesssim\left(\left\|\partial_{x}u\right\|_{L_{xy}^{\infty}}+\left\|\partial_{y}u\right\|_{L_{xy}^{\infty}}\right)\left\|u\right\|_{H^{s_{1},s_{2}}_{xy}}^{2}.$
(70)
From Gronwall inequality we get the lemma. ∎
### 3.3. A Strichartz estimate type
The energy estimate suggests that we have to control the norms
$\left\|\partial_{x}u\right\|_{L_{T}^{1}L_{xy}^{\infty}}$ and
$\left\|\partial_{y}u\right\|_{L_{T}^{1}L_{xy}^{\infty}}$ in order to prove
the local well-posedness of the Cauchy problem (13). The next lemma shows how
we can control these norms.
###### Lemma 3.8.
Let $-1\leq\alpha\leq 1$, $T>0$ and assume that $u\in
C([0,T];H^{\infty}(\mathbb{R}^{2}))$ is a solution to initial value problem
(13) with initial condition $\psi$. Then, for any
$s_{1}>\frac{17}{12}-\frac{\alpha}{4}$ such that
$\begin{cases}s_{2}>1,&\mbox{if }0\leq\alpha\leq 1,\\\
\frac{1}{s_{2}}-\frac{\alpha}{4s_{1}}<1,&\mbox{if }-1\leq\alpha\leq
0\end{cases}$
and $s_{1}\geq s_{2}$, there exist constants $C_{s_{1},s_{2}}$ and
$k_{s_{1},s_{2}}\in(7/12,1)$ such that
$f(T)=\left\|u\right\|_{L_{T}^{1}L_{xy}^{\infty}}+\left\|\partial_{x}u\right\|_{L_{T}^{1}L_{xy}^{\infty}}+\left\|\partial_{y}u\right\|_{L_{T}^{1}L_{xy}^{\infty}}$
(71)
satisfies
$f(T)\leq
C_{s_{1},s_{2}}T^{k_{s_{1},s_{2}}}(1+f(T))\left\|u\right\|_{L_{T}^{\infty}H^{s_{1},s_{2}}_{xy}}.$
(72)
###### Proof.
First let us examine an estimate for
$\left\|\partial_{x}u\right\|_{L^{1}_{T}L_{xy}^{\infty}}$. From the refined
Strichartz estimate (57), we have
$\left\|\partial_{x}u\right\|_{L^{1}_{T}L_{xy}^{\infty}}\leq
c_{\epsilon}T^{\frac{7+2\epsilon}{12}}\left(\sup_{t\in[0,T]}{\left\|J_{x}^{\frac{17}{12}-\frac{\alpha}{4}+\frac{2\epsilon}{3}}u\right\|_{L^{2}_{xy}}}+\int_{0}^{T}\left\|D_{x}^{\frac{5}{12}-\frac{\alpha}{4}+\frac{2\epsilon}{3}}(u\partial_{x}u)\right\|_{L^{2}_{xy}}dt\right).$
(73)
Let us choose $0<\epsilon_{1}<1$ in such a way that
$\frac{5}{12}-\frac{\alpha}{4}+\frac{2\epsilon_{1}}{3}<1$ and
${\frac{17}{12}-\frac{\alpha}{4}+\frac{2\epsilon_{1}}{3}}<s_{1}$. So, the
first term from the right side of (74) we have
$\sup_{t\in[0,T]}{\left\|J_{x}^{\frac{17}{12}-\frac{\alpha}{4}+\frac{2\epsilon_{1}}{3}}u\right\|_{L^{2}_{xy}}}\leq\sup_{t\in[0,T]}{\left\|J_{x}^{s_{1}}u\right\|_{L^{2}_{xy}}}\leq\left\|u\right\|_{L^{\infty}_{T}H^{s_{1},s_{2}}_{xy}}.$
(74)
For the second term on the right side of (74), from the Leibniz rule (equation
(19)) in $x$, we have
$\begin{split}&\int_{0}^{T}\left\|D_{x}^{\frac{5}{12}-\frac{\alpha}{4}+\frac{2\epsilon_{1}}{3}}(u\partial_{x}u)\right\|_{L^{2}_{xy}}dt\lesssim\\\
&\lesssim\int_{0}^{T}{\left\|{{\left\|D_{x}^{\frac{5}{12}-\frac{\alpha}{4}+\frac{2\epsilon_{1}}{3}}u\right\|_{L^{2}_{x}}}{\left\|\partial_{x}u\right\|_{L_{x}^{\infty}}}+{\left\|D_{x}^{\frac{5}{12}-\frac{\alpha}{4}+\frac{2\epsilon_{1}}{3}}\partial_{x}u\right\|_{L^{2}_{x}}}{\left\|u\right\|_{L_{x}^{\infty}}}}\right\|_{L^{2}_{y}}}dt\\\
&\lesssim\int_{0}^{T}{\left\|{{\left\|J_{x}^{s_{1}}u\right\|_{L^{2}_{x}}}{\left\|\partial_{x}u\right\|_{L_{x}^{\infty}}}+{\left\|J_{x}^{s_{1}}u\right\|_{L^{2}_{x}}}{\left\|u\right\|_{L_{x}^{\infty}}}}\right\|_{L^{2}_{y}}}dt\\\
&\lesssim\int_{0}^{T}\left(\left\|u\right\|_{L^{\infty}_{xy}}+\left\|\partial_{x}u\right\|_{L^{\infty}_{xy}}\right)\left\|J_{x}^{s_{1}}u\right\|_{L^{2}_{xy}}dt\\\
&\lesssim\left\|u\right\|_{L_{T}^{\infty}H^{s_{1},s_{2}}_{xy}}f(T)\end{split}$
(75)
Then, from (75), (74) and (73)
$\left\|\partial_{x}u\right\|_{L^{1}_{T}L_{xy}^{\infty}}\leq
C_{1}(s_{1})T^{\frac{7+2\epsilon_{1}}{12}}(1+f(T))\left\|u\right\|_{L_{T}^{\infty}H^{s_{1},s_{2}}_{xy}}.$
(76)
Let us estimate now $\left\|u\right\|_{L_{T}^{1}L_{xy}^{\infty}}$. From
Duhamel’s principle, Cauchy-Schwarz inequality in $t$ and Corollary 3.5, we
have that
$\begin{split}\left\|u\right\|_{L_{T}^{1}L_{xy}^{\infty}}\leq\left\|W_{\alpha}(t)\psi\right\|_{L_{T}^{1}L_{xy}^{\infty}}+\int_{0}^{T}\left\|W_{\alpha}(t-t^{\prime})(u\partial_{x}u)(t^{\prime})\right\|_{L_{T}^{1}L_{xy}^{\infty}}dt^{\prime}\\\
\leq
T^{\frac{1}{2}}\left(\left\|W_{\alpha}(t)\psi\right\|_{L_{T}^{2}L_{xy}^{\infty}}+\int_{0}^{T}\left\|W_{\alpha}(t-t^{\prime})(u\partial_{x}u)(t^{\prime})\right\|_{L_{T}^{2}L_{xy}^{\infty}}dt^{\prime}\right)\\\
\leq
C_{2}T^{\frac{7+2\epsilon_{2}}{12}}\left(\left\|D_{x}^{\frac{1}{2}(\epsilon_{2}-\frac{\alpha}{2})}\psi\right\|_{L_{xy}^{2}}+\int_{0}^{T}\left\|D_{x}^{\frac{1}{2}(\epsilon_{2}-\frac{\alpha}{2})}(u\partial_{x}u)(t^{\prime})\right\|_{L_{xy}^{2}}dt^{\prime}\right),\end{split}$
(77)
for some $\epsilon_{2}=\epsilon_{2}(s_{1})$ such that
$0<\frac{1}{2}(\epsilon_{2}-\frac{\alpha}{2})$ and
$\frac{1}{2}(\epsilon_{2}-\frac{\alpha}{2})+1<s_{1}$. The first term in the
last line of equation above, thanks to the choice of $\epsilon_{2}$, is less
than $\left\|J_{x}^{s_{1}}\psi\right\|_{L^{2}_{xy}}$. For the term within the
integral, from the Leibniz rule in the $x$ variable, we have
$\begin{split}\left\|D_{x}^{{\frac{1}{2}(\epsilon_{2}-\frac{\alpha}{2})}}(u\partial_{x}u)\right\|_{L_{xy}^{2}}&\leq\left\|D_{x}^{\frac{1}{2}(\epsilon_{2}-\frac{\alpha}{2})}u\right\|_{L_{xy}^{2}}\left\|\partial_{x}u\right\|_{L_{xy}^{\infty}}+\left\|D_{x}^{\frac{1}{2}(\epsilon_{2}-\frac{\alpha}{2})}\partial_{x}u\right\|_{L_{xy}^{2}}\left\|u\right\|_{L_{xy}^{\infty}}\\\
&\leq\left\|J_{x}^{s_{1}}u\right\|_{L_{xy}^{2}}\left\|\partial_{x}u\right\|_{L_{xy}^{\infty}}+\left\|J_{x}^{s_{1}}u\right\|_{L_{xy}^{2}}\left\|u\right\|_{L_{xy}^{\infty}}.\end{split}$
(78)
So that
$\begin{split}\left\|u\right\|_{L_{T}^{1}L_{xy}^{\infty}}&\leq
C_{2}T^{\frac{7+2\epsilon_{2}}{12}}\left(\left\|J_{x}^{s_{1}}\psi\right\|_{L^{2}_{xy}}+\int_{0}^{T}\left(\left\|u\right\|_{L_{xy}^{\infty}}+\left\|\partial_{x}u\right\|_{L_{xy}^{\infty}}\right)\left\|J_{x}^{s_{1}}u\right\|_{L_{xy}^{2}}dt^{\prime}\right)\\\
&\leq
C_{2}T^{\frac{7+2\epsilon_{2}}{12}}\left(\left\|J_{x}^{s_{1}}\psi\right\|_{L^{2}_{xy}}+\int_{0}^{T}\left(\left\|u\right\|_{L_{xy}^{\infty}}+\left\|\partial_{x}u\right\|_{L_{xy}^{\infty}}\right)\left\|J_{x}^{s_{1}}u\right\|_{L_{xy}^{2}}dt^{\prime}\right)\\\
&\leq
C_{2}T^{\frac{7+2\epsilon_{2}}{12}}(1+f(T))\left\|u\right\|_{L_{T}^{\infty}H^{s_{1},s_{2}}_{xy}}\end{split}$
(79)
where $C_{2}=C_{2}(s_{1})$. Lastly, let us estimate
$\left\|\partial_{y}u\right\|_{L_{T}^{1}L_{xy}^{\infty}}$. In the same way as
we estimate the other two terms, we have that
$\begin{split}\left\|\partial_{y}u\right\|_{L_{T}^{1}L_{xy}^{\infty}}\leq
C_{3}T^{\frac{7+2\epsilon_{3}}{12}}\left(\left\|D_{x}^{\frac{1}{2}(\epsilon_{3}-\frac{\alpha}{2})}\partial_{y}\psi\right\|_{L_{xy}^{2}}+\int_{0}^{T}\left\|D_{x}^{\frac{1}{2}(\epsilon_{3}-\frac{\alpha}{2})}\partial_{y}(u\partial_{x}u)\right\|_{L_{xy}^{2}}dt\right),\end{split}$
(80)
where we will choose $\epsilon_{3}$ later. For the term within the integral,
from the commutator estimate (Lemma 1.10) in the $x$ variable, we obtain
$\begin{split}\left\|D_{x}^{\frac{1}{2}(\epsilon_{3}-\frac{\alpha}{2})+1}(u\partial_{y}u)\right\|_{L_{xy}^{2}}\leq\left\|\left[D_{x}^{\frac{1}{2}(\epsilon_{3}-\frac{\alpha}{2})+1},u\right]\partial_{y}u\right\|_{L_{xy}^{2}}+\left\|uD_{x}^{\frac{1}{2}(\epsilon_{3}-\frac{\alpha}{2})}\partial_{y}u\right\|_{L_{xy}^{2}}\\\
\leq\left(\left\|u\right\|_{L_{xy}^{\infty}}+\left\|\partial_{x}u\right\|_{L_{xy}^{\infty}}\right)\left\|D_{x}^{\frac{1}{2}(\epsilon_{3}-\frac{\alpha}{2})}\partial_{y}u\right\|_{L_{xy}^{2}}+\left\|\partial_{y}u\right\|_{L_{xy}^{\infty}}\left\|D_{x}^{\frac{1}{2}(\epsilon_{3}-\frac{\alpha}{2})+1}u\right\|_{L_{xy}^{2}}.\end{split}$
(81)
Let us, then, examine
$\left\|D_{x}^{\frac{1}{2}(\epsilon_{3}-\frac{\alpha}{2})+1}u\right\|_{L_{xy}^{2}}$
and
$\left\|D_{x}^{\frac{1}{2}(\epsilon_{3}-\frac{\alpha}{2})}\partial_{y}u\right\|_{L_{xy}^{2}}$.
We will do this for two cases in $\alpha$. The first case is when
$-1\leq\alpha\leq 0$. In this case
$\frac{1}{2}(\epsilon_{3}-\frac{\alpha}{2})>0$ for $0<\epsilon_{3}<1$. Choose
$\epsilon_{3}$ in such a way that
$\frac{1}{2}(\epsilon_{3}-\frac{\alpha}{2})+1<s_{1}$ and
$\frac{\epsilon_{3}}{2s_{1}}-\frac{\alpha}{4s_{1}}+\frac{1}{s_{2}}<1$, or
equivalently
$\frac{1}{2}(\epsilon_{3}-\frac{\alpha}{2})\frac{s_{2}}{s_{2}-1}<s_{1}$. So,
$\left\|D_{x}^{\frac{1}{2}(\epsilon_{3}-\frac{\alpha}{2})+1}u\right\|_{L_{xy}^{2}}\leq\|J_{x}^{s_{1}}u\|_{L^{2}_{x,y}}\leq\|u\|_{H^{x_{1}.x_{2}}_{x,y}}$
(82)
and
$\begin{split}\left\|D_{x}^{\frac{1}{2}(\epsilon_{3}-\frac{\alpha}{2})}\partial_{y}u\right\|_{L_{xy}^{2}}&\leq\left\|J_{x}^{\frac{1}{2}(\epsilon_{3}-\frac{\alpha}{2})}J_{y}u\right\|_{L_{xy}^{2}}\leq\left\|J_{x}^{\frac{1}{2}(\epsilon_{3}-\frac{\alpha}{2})\frac{s_{2}}{s_{2}-1}}u\right\|_{L_{xy}^{2}}^{\frac{s_{2}-1}{s_{2}}}\left\|J_{y}^{s_{2}}u\right\|_{L_{xy}^{2}}^{\frac{1}{s_{2}}}\\\
&\leq\left\|J_{x}^{s_{1}}u\right\|_{L_{xy}^{2}}^{\frac{s_{2}-1}{s_{2}}}\left\|J_{y}^{s_{2}}u\right\|_{L_{xy}^{2}}^{\frac{1}{s_{2}}}\leq\|u\|_{H_{x,y}^{s_{1},s_{2}}}\end{split}$
(83)
The second case is when $0\leq\alpha\leq 1$. Choose, then,
$\epsilon_{3}>\frac{\alpha}{2}$ such that also
$\frac{1}{2}(\epsilon_{3}-\frac{\alpha}{2})+1<s_{1}$ and
$\frac{\epsilon_{3}}{2s_{1}}-\frac{\alpha}{4s_{1}}+\frac{1}{s_{2}}<1$. So we
are in the same situation of the first case, what leads us to the same two
inequalities (82) and (83). This inequalities together to (81) and (80) allow
us show that
$\begin{split}\left\|\partial_{y}u\right\|_{L_{T}^{1}L_{xy}^{\infty}}\leq
C_{3}T^{\frac{7+2\epsilon_{3}}{12}}\left(\left\|\psi\right\|_{H_{xy}^{s_{1},s_{2}}}+(1+f(T))\left\|u\right\|_{L_{T}^{\infty}H_{xy}^{s_{1},s_{2}}}\right).\end{split}$
(84)
The theorem follows from (76), (79) and (84), taking
$C_{s_{1},s_{2}}=C_{1}+C_{2}+C_{3}$ and
$k_{s_{1},s_{2}}=\max_{i=1,2,3}{\frac{7+2\epsilon_{i}}{12}}$. ∎
### 3.4. Local well-posedness
We already have all the necessary ingredients to prove the following theorem.
###### Theorem 3.9.
Let $-1<\alpha<1$ and $0<\epsilon<1$ be such that
$s_{1}>\frac{17+4\epsilon-3\alpha}{12},\quad s_{2}\leq s_{1}\text{
and}\quad\begin{cases}s_{2}>1,&\text{ if }\alpha>0,\\\
\frac{5+4\epsilon-3\alpha}{12s_{1}}+\frac{1}{s_{2}}<1,&\text{ if }\alpha\leq
0.\end{cases}$
Then, for any $\psi\in H^{s_{1},s_{2}}(\mathbb{R}^{2})$, there exist a time
$T=T(\left\|\psi\right\|_{H^{s_{1},s_{2}}})$ and a unique solution $u\in
C([0,T]:H^{s_{1},s_{2}})$ to the Cauchy problem (13) such that
$u,\partial_{x}u,\partial_{y}u\in L_{T}^{1}L_{xy}^{\infty}$. Furthermore, if
$0<T^{\prime}<T$, there exists a neighborhood $\mathcal{V}$ of $\psi$ in
$H^{s_{1},s_{2}}(\mathbb{R}^{2})$ such that $\psi\mapsto u(t)$ is continuous
From the local well-posedness of (13) in $H^{\sigma}$ for $\sigma>2$ ($\sigma$
fix), for $\psi\in H^{\infty}(\mathbb{R}^{2})$, there exists a unique solution
$u\in C([0,T^{*}]:H^{\sigma}(\mathbb{R}^{2}))$, where $T^{*}$ is the maximal
existence time of the solution, such that if $T^{*}<\infty$, then
$\lim_{t\nearrow T^{*}}\left\|u(t)\right\|_{H^{\sigma}}=\infty.$ (85)
Thanks to the above we have the following a priori estimate that will be
useful in the proof of the existence of the solution to problem (13).
###### Lemma 3.10.
Let $2<\sigma$, $-1\leq\alpha\leq 1$ and $s_{1}$, $s_{2}$ be as in Theorem
(3.9) such that $s_{i}\leq\sigma$, $i=1,2$. Let, also, $\psi\in H^{\sigma}$
and $u\in C([0,T^{*});H^{\sigma}(\mathbb{R}^{2}))$ such that $u(0)=\psi$,
where $T^{*}$ is the maximal time of existence of $u$. Then, there exist
$K_{0}=K_{0}(s_{1},s_{2})>0$ and $L_{s_{1},s_{2}}>0$ such that $T^{*}>T$,
where
$(L_{s_{1},s_{2}}\left\|\psi\right\|_{H^{s_{1},s_{2}}}+1)^{\frac{12}{7}}T=1$,
and
$\begin{split}\left\|u\right\|_{L_{T}^{\infty}H^{s_{1},s_{2}}}&\leq
2\left\|\psi\right\|_{H^{s_{1},s_{2}}}\\\
f(T)=\left\|u\right\|_{L_{T}^{1}L_{xy}^{\infty}}+\left\|\partial_{x}u\right\|&{}_{L_{T}^{1}L_{xy}^{\infty}}+\left\|\partial_{y}u\right\|_{L_{T}^{1}L_{xy}^{\infty}}\leq
K_{0}\end{split}$ (86)
###### Proof.
For $s_{1}$ and $s_{2}$ as in the statement of lemma, let
$T_{0}=\sup_{\tilde{T}\in(0,T^{*})}\left\\{\tilde{T}\,|\,\left\|u\right\|_{L_{\tilde{T}}^{\infty}H^{s_{1},s_{2}}}\leq
2\left\|\psi\right\|_{H^{s_{1},s_{2}}}\right\\}$ (87)
From the local well-posedness, this set is not empty. Let $C_{s_{1},s_{2}}$ be
as in Lemma 3.8, $C$ be as in Lemma 3.7,
$L_{s_{1},s_{2}}=2(2C-1)C_{s_{1},s_{2}}$ and
$T=\frac{1}{(L_{s_{1},s_{2}}\|\psi\|_{H^{s_{1},s_{2}}}+1)^{\frac{12}{7}}}.$
Let us see that $T\leq T_{0}$. Suppose not. Thanks to Lemma 3.8
$f(T_{0})\leq
C_{s_{1},s_{2}}T_{0}^{k_{s_{1},s_{2}}}(1+f(T_{0}))\left\|u\right\|_{L_{T_{0}}^{\infty}H^{s_{1},s_{2}}_{xy}}\leq
C_{s_{1},s_{2}}T_{0}^{\frac{7}{12}}(1+f(T_{0}))\left\|u\right\|_{L_{T_{0}}^{\infty}H^{s_{1},s_{2}}_{xy}}$
(88)
Since $\left\|u\right\|_{L_{T_{0}}^{\infty}H^{s_{1},s_{2}}}\leq
2\left\|\psi\right\|_{H^{s_{1},s_{2}}}$ and $T_{0}<T$ ,
$f(T_{0})\leq
2C_{s_{1},s_{2}}\frac{(1+f(T_{0}))}{L_{s}\left\|\psi\right\|_{H^{s_{1},s_{2}}}+1}\left\|\psi\right\|_{H^{s_{1},s_{2}}}$
or equivalently
$(4CC_{s_{1},s_{2}}\left\|\psi\right\|_{H^{s_{1},s_{2}}}+1)f(T_{0})\leq
2C_{s_{1},s_{2}}\left\|\psi\right\|_{H^{s_{1},s_{2}}}$
Therefore,
$f(T_{0})\leq\frac{1}{2C}$
By Lemma 3.7 we would have
$\|u(T_{0})\|_{H^{s_{1},s_{2}}}\leq
e^{\frac{1}{2}}\|\psi\|_{H^{s_{1},s_{2}}}<2\|\psi\|_{H^{s_{1},s_{2}}}.$
From the continuity of $u$, it follows that there exists $\tilde{T}>T_{0}$
such that
$\|u\|_{L^{\infty}_{\tilde{T}}H^{s_{1},s_{2}}}\leq
2\|\psi\|_{H^{s_{1},s_{2}}},$
which contradicts the choice of $T_{0}$. Then, $T\leq T_{0}$. In particular,
$\|u\|_{L^{\infty}_{T}H^{s_{1},s_{2}}}\leq 2\|\psi\|_{H^{s_{1},s_{2}}},$
and, repeating the above reasoning, from inequality (88), taking $T$ instead
of $T_{0}$ we have that
$f(T)\leq\frac{1}{2C}.$
This proves the lemma. ∎
###### Corollary 3.11.
Let $\psi$ and $T$ be as in the last lemma. If $\psi\in H^{\infty}$, then the
solution $u$ to the Cauchy problem (13), with $u(0)=\psi$, belongs to the set
$C([0,T];H^{\infty})$.
###### Proof.
Let $T$ be as in the lemma above. From that lemma, for any number $\sigma$
that satisfies the condition established there, we have $u\in
C([0,T];H^{\sigma})$. From here it follows the corollary. ∎
###### Corollary 3.12.
Let $R>0$ and $\psi\in H^{\infty}$ be such that
$\|\psi\|_{H^{s_{1},s_{2}}}\leq R$. Then, there exists $T_{0}$ that depends on
$R$ and $M$, constant that depends only on $s_{1},s_{2}$, such that $u\in
C([0,T_{0}];H^{\infty})$ and
$f(T_{0})=\left\|u\right\|_{L_{T_{0}}^{1}L_{xy}^{\infty}}+\left\|\partial_{x}u\right\|_{L_{T_{0}}^{1}L_{xy}^{\infty}}+\left\|\partial_{y}u\right\|_{L_{T_{0}}^{1}L_{xy}^{\infty}}\leq
M.$
###### Proof.
If we set $T_{0}=1/{(L_{s_{1},s_{2}}R+1)^{\frac{12}{7}}}$, we have that
$T_{0}\leq T$, $T$ as in the last lemma. From the proof of that lemma it
follows that $f(T_{0})\leq 1/2C$, that does not depend on initial data, only
on $s_{1}$ and $s_{2}$. Making $M=1/2C$ shows the corollary. ∎
#### 3.4.1. Proof of Theorem 3.9
###### Lemma 3.13.
Suppose that $\psi$ and $\phi\in H^{\infty}$ and that $u$ and $v\in
C([0,T];H^{\infty})$ are the solutions to the problem (13) with initial
conditions $\psi$ and $\phi$, respectively. Then,
$\|u-v\|_{L^{2}}(T_{0}(R))\leq\|\psi-\phi\|_{L^{2}}e^{CM},$
where $T_{0}(R)$ and $M$ are as in Corollary 3.12, $C$ is a constant that
depends only on $s_{1}$ and $s_{2}$, and $R$ is the maximum between the norm
of $\phi$ and $\psi$ in the space $H^{s_{1},s_{2}}$.
###### Proof.
The proof is analogous to obtaining the energy estimate. Let us see. Let $u$
and $v$ be as in statement of lemma. Then,
$\partial_{t}(u-v)=\partial_{x}^{3}(u-v)-\mathcal{H}D^{\alpha}\partial_{y}^{2}(u-v)+\frac{1}{2}(u+v)\partial_{x}(u-v)+\frac{1}{2}(u-v)\partial_{x}(u+v).$
Multiplying by $u-v$ on both sides the equation and integrating by parts we
have that
$\frac{1}{2}\frac{d}{dt}\|u-v\|_{2}^{2}\leq\frac{1}{4}(\|u_{x}\|_{L^{\infty}}+\|v_{x}\|_{L^{\infty}})\|u-v\|_{2}^{2}.$
From the Gronwall Lemma and Corollary 3.12 it follows the lemma. ∎
Now, let $\psi\in H^{s_{1},s_{2}}$ and suppose that $\psi_{n}$ is a sequence
of functions in $H^{\infty}$ that converges to $\psi$ in $H^{s_{1},s_{2}}$ .
Taking $R=\sup_{n}\|\psi_{n}\|$, from lemma above, the solutions to (13)
$u_{n}\in C([0,T_{0}];H^{\infty}(\mathbb{R}^{2}))$, with initial condition
$\psi_{n}$, converge uniformly to a function $u$ in
$C([0,T_{0}];L^{2}(\mathbb{R}^{2}))$. Moreover, from Corollary 3.12, the
functions $u_{n}$ are uniformly bounded in $H^{s_{1},s_{2}}$. Therefore, from
Banach-Alaoglu theorem, $u_{n}(t)$ have a subsequence that converges weakly in
$H^{s_{1},s_{2}}$. From the uniform convergence of $u_{n}(t)$ to $u(t)$ in
$L^{2}$, it follows that $u(t)\in H^{s_{1},s_{2}}$, for each $t\in[0,T_{0}]$.
From the continuity of $u$ from $[0,T_{0}]$ to $L^{2}$ follows the weak
continuity of $u$ from $[0,T_{0}]$ to $H^{s_{1},s_{2}}$. In particular, from
uniform boundedness of the sequence $u_{n}(t)$ and Lemma 1.7, it follows that
$u_{n}$ converges strong and uniformly in $H^{s^{\prime}_{1},s^{\prime}_{2}}$
to $u$, for any pair of non negative real numbers
$s^{\prime}_{1},s^{\prime}_{2}$ strictly less than $s_{1},s_{2}$,
respectively. Since each one of the functions $u_{n}$ satisfies the integral
equation associated to (13),
$w=W_{\alpha}(t)w(0)+\frac{1}{2}\int_{0}^{t}W_{\alpha}(t-t^{\prime})\partial_{x}(w^{2}(t^{\prime}))\,dt^{\prime},$
(89)
in $H^{s^{\prime}_{1}-1,s^{\prime}_{2}-1}$, it follows that function $u$
satisfies this same integral equation in
$H^{s^{\prime}_{1}-1,s^{\prime}_{2}-1}$, $1<s^{\prime}_{2}<s_{2}$. From Lemma
3.6, considering $17/12-\alpha/4<s_{1}^{\prime}<s_{1}$, follows that
$\partial_{x}u\in{L^{1}_{T}L^{\infty}_{x,y}}$. In the same way, thanks to the
inequalities (77) and (80), we obtain that $u$ and
$\partial_{y}u\in{L^{1}_{T}L^{\infty}_{x,y}}$. So,
###### Lemma 3.14.
Let $\psi\in H^{s_{1},s_{2}}(\mathbb{R}^{2})$. Then, there exists $T>0$ and a
unique $u\in C([0,T];L^{2}(\mathbb{R}^{2}))\cap
C_{w}([0,T];H^{s_{1},s_{2}}(\mathbb{R}^{2}))\cap
C^{1}([0,T];H^{-3}\cap(X^{3})*)$ solution to problem (13). Furthermore, $u$,
$\partial_{x}u$, $\partial_{y}u\in{L^{1}_{T}L^{\infty}_{x,y}}$ and the map
$\psi\mapsto u$ is Lipschitz continuous from $L^{2}(\mathbb{R}^{2})$ to
$C([0,T];L^{2}(\mathbb{R}^{2}))$.
###### Proof.
It only remains to prove that $u$ is unique and that the map $\psi\mapsto u$
is Lipschitz continuous from $L^{2}(\mathbb{R}^{2})$ to
$C([0,T];L^{2}(\mathbb{R}^{2}))$. Let $u$ and $v$ be as in the statement of
lemma. Since $u$ and $v$ are solutions to the integral equation, from Lemma
3.6, $u_{x}$ and $v_{x}\in{L^{1}_{T}L^{\infty}_{x,y}}$. Now proceed as in the
lemma above, but we need Lemma (1.3), to obtain that
$\frac{1}{2}\frac{d}{dt}\|u-v\|_{L^{2}}^{2}=\frac{1}{4}(\partial_{x}u+\partial_{x}v,(u-v)^{2})_{L^{2}}\leq(\|u_{x}\|_{L_{x,y}^{\infty}}+\|v_{x}\|_{L_{x,y}^{\infty}})\|u-v\|_{L^{2}}^{2}.$
Since $\|u_{x}\|_{L_{x,y}^{\infty}}+\|v_{x}\|_{L_{x,y}^{\infty}}$ is
integrable and $\|u-v\|_{L^{2}}^{2}$ is continuous, from Gronwall lemma, it
follows the theorem. ∎
###### Lemma 3.15.
For $\psi\in H^{s_{1},s_{2}}$ as in the lemma above, the solution $u$ to (13)
described there is strongly continuous in $H^{s_{1},s_{2}}$
###### Proof.
Let $\psi$, $\psi_{n}$, $u_{n}$, $n=1,2,3,\cdots$, and $u$ be as before. Let,
also, $M$, $T_{0}$ be as in Corollary 3.12 and $T\leq T_{0}$. It is clear that
$f_{n}(T)=\|u_{n}\|_{L_{x,y}^{\infty}}+\|\partial_{x}u_{n}\|_{L_{x,y}^{\infty}}+\|\partial_{y}u_{n}\|_{L_{x,y}^{\infty}}$
satisfies that
$f_{n}(T)\leq M,$
for all $n$. From Lemma 3.8 and Corollary 3.12, it follows that
$f_{n}(T)\leq 2C_{s_{1},s_{2}}T^{k_{s_{1},s_{2}}}+(1+M)R.$
From Lemma 3.7,
$\|u_{n}(T)\|_{H^{s_{1},s_{2}}}\leq\|\psi_{n}\|_{H^{s_{1},s_{2}}}e^{2C_{s_{1},s_{2}}T^{k_{s_{1},s_{2}}}(1+M)R},$
for all $n$. From the weak convergence of $u_{n}(T)$ to $u(T)$ in
$H^{s_{1},s_{2}}$ and since $\psi_{n}$ converges to $\psi$ in
$H^{s_{1},s_{2}}$, it follows that
$\|u(T)\|_{H^{s_{1},s_{2}}}\leq\|\psi\|_{H^{s_{1},s_{2}}}e^{2C_{s_{1},s_{2}}T^{k_{s_{1},s_{2}}}(1+M)R}.$
From this last inequality and the weak continuity of $u$ in $H^{s_{1},s_{2}}$,
it follows that
$\|\psi\|_{H^{s_{1},s_{2}}}\leq\liminf_{T\to
0+}\|u(T)\|_{H^{s_{1},s_{2}}}\leq\limsup_{T\to
0+}\|u(T)\|_{H^{s_{1},s_{2}}}\leq\|\psi\|_{H^{s_{1},s_{2}}}.$
Then, $u$ is right strongly continuous at $0$ in the space $H^{s_{1},s_{2}}$.
Since $u(-t,-x,-y)$ is also solution to the equation (13), we have that $u$ is
also left strongly continuous at $0$ in the space $H^{s_{1},s_{2}}$. Now, for
any $t^{*}\in[0,T]$, $u(t+t^{*})$ is also solution to the equation (13) with
initial condition $u(t^{*})$. Then from the unicity of the solution, $u$ also
is strongly continuous at $t^{*}$ in the space $H^{s_{1},s_{2}}$. This ends
the proof. ∎
Now examine the continuity of the solutions to (13) with respect to initial
data. For this purpose we will use a technique very useful an recurrently used
in the literature related to the well-posedness of evolution equations. This
technique is the Bona-Smith method of approximation introduced in [7]. Let us
see.
###### Lemma 3.16.
Let $\phi\in H^{s_{1},s_{2}}$, $s_{1}$ and $s_{2}$ be positive real numbers.
For each $\tau>0$ define $\phi^{\tau}$ by
$\phi^{\tau}(x,y)=\left(\widehat{\phi}(\xi,\eta)\exp\left(-\tau\left((1+\lvert\xi\rvert^{2})^{\frac{s_{1}}{2}}+(1+\lvert\eta\rvert^{2})^{\frac{s_{2}}{2}}\right)\right)\right)\text{\huge$\check{\
}$}(x,y).$ (90)
Then
$\lim_{\tau\to 0+}\|\phi^{\tau}-\phi\|_{H^{s_{1},s_{2}}}=0$
and there exists a constant $C=C(s)$ such that
$\displaystyle\|\phi^{\tau}\|_{H^{s_{1}+1,s_{2}}}$ $\displaystyle\leq
C\left(\frac{1}{\tau
s_{1}}\right)^{\frac{1}{s_{1}}}\|\phi\|_{H^{s_{1},s_{2}}},$ (91)
$\displaystyle\|\phi^{\tau}\|_{H^{s_{1},s_{2}+1}}$ $\displaystyle\leq
C\left(\frac{1}{\tau
s_{2}}\right)^{\frac{1}{s_{2}}}\|\phi\|_{H^{s_{1},s_{2}}}$ (92)
and
$\|\phi^{\tau}-\phi^{\theta}\|_{L^{2}}\leq
C\lvert\tau-\theta\rvert\|\phi\|_{H^{s_{1},s_{2}}}$ (93)
###### Proposition 3.17.
Let $R>0$, and assume that $\Lambda$ is a set and $\psi_{\lambda\in\Lambda}$
is a collection of functions in $H^{\infty}$ such that
$\|\psi_{\lambda}\|_{H^{s_{1},s_{2}}}\leq R$, for all $\lambda\in\Lambda$.
Also, let $\psi_{\lambda}^{\tau}$ be the approximations defined from
$\psi_{\lambda}$ as in (90), and assume that $u_{\lambda}^{\tau}$ is the
solution to (13) with initial condition $\psi_{\lambda}^{\tau}$, for all
$\lambda\in\Lambda$. Suppose that $0\leq\theta<\tau$. Then, for $\nu>0$
$\left\|u_{\lambda}^{\tau}-u_{\lambda}^{\theta}\right\|^{2}_{H^{s_{1},s_{2}}}\leq
C\left(\left\|\psi_{\lambda}^{\tau}-\psi_{\lambda}^{\theta}\right\|^{2}_{H^{s_{1},s_{2}}}+\tau^{\nu}\right).$
for all $\lambda\in\Lambda$.
###### Proof.
It is evident that $u_{\lambda}^{\tau}(t)$ also is defined on $[0,T_{0}]$, for
all $n$ and $\tau$. We will proceed as in the proof of Lemma 3.7. Then,
$\displaystyle\frac{1}{2}\frac{d}{dt}\|J_{x}^{s_{1}}(u_{\lambda}^{\tau}$
$\displaystyle-
u_{\lambda}^{\theta})\|_{L^{2}}^{2}=\int_{\mathbb{R}^{2}}J_{x}^{s_{1}}(u_{\lambda}^{\tau}\partial_{x}u_{\lambda}^{\tau}-u_{\lambda}^{\theta}\partial_{x}u_{\lambda}^{\theta})J_{x}^{s_{1}}(u_{\lambda}^{\tau}-u_{\lambda}^{\theta})\,dxdy$
(94) $\displaystyle=$
$\displaystyle\frac{1}{2}\int_{\mathbb{R}^{2}}J^{s_{1}}_{x}\partial_{x}\left((u_{\lambda}^{\tau}+u_{\lambda}^{\theta})(u_{\lambda}^{\tau}-u_{\lambda}^{\theta})\right)J_{x}^{s_{1}}(u_{\lambda}^{\tau}-u_{\lambda}^{\theta})\,dxdy$
$\displaystyle=$
$\displaystyle\frac{1}{2}\int_{\mathbb{R}^{2}}J^{s_{1}}_{x}\left((u_{\lambda}^{\tau}+u_{\lambda}^{\theta})\partial_{x}(u_{\lambda}^{\tau}-u_{\lambda}^{\theta})\right)J^{s_{1}}_{x}(u_{\lambda}^{\tau}-u_{\lambda}^{\theta})\,dxdy+$
$\displaystyle+\frac{1}{2}\int_{\mathbb{R}^{2}}J_{x}^{s_{1}}\left(\partial_{x}(u_{\lambda}^{\tau}+u_{\lambda}^{\theta})(u_{\lambda}^{\tau}-u_{\lambda}^{\theta})\right)J^{s_{1}}(u_{\lambda}^{\tau}-u_{\lambda}^{\theta})\,dxdy.$
We estimate the last two terms that appear in the inequality above. From Lemma
1.8, for the first term, we have
$\displaystyle\frac{1}{2}\int_{\mathbb{R}^{2}}J^{s_{1}}_{x}\big{(}(u_{\lambda}^{\tau}$
$\displaystyle+u_{\lambda}^{\theta})\partial_{x}(u_{\lambda}^{\tau}-u_{\lambda}^{\theta})\big{)}J^{s_{1}}_{x}(u_{\lambda}^{\tau}-u_{\lambda}^{\theta})\,dxdy\leq$
(95) $\displaystyle\leq$
$\displaystyle\frac{1}{2}\int_{\mathbb{R}^{2}}\left([J^{s_{1}}_{x},u_{\lambda}^{\tau}+u_{\lambda}^{\theta}]\partial_{x}(u_{\lambda}^{\tau}-u_{\lambda}^{\theta})\right)J_{x}^{s_{1}}(u_{\lambda}^{\tau}-u_{\lambda}^{\theta})\,dxdy+$
$\displaystyle+\frac{1}{2}\int_{\mathbb{R}^{2}}(u_{\lambda}^{\tau}+u_{\lambda}^{\theta})\partial_{x}J_{x}^{s_{1}}(u_{\lambda}^{\tau}-u_{\lambda}^{\theta})J_{x}^{s_{1}}(u_{\lambda}^{\tau}-u_{\lambda}^{\theta})\,dxdy$
$\displaystyle\leq$ $\displaystyle
C(\|\partial_{x}(u_{\lambda}^{\tau}+u_{\lambda}^{\theta})\|_{L^{\infty}}\|J^{s_{1}-1}_{x}(u_{\lambda}^{\tau}-u_{\lambda}^{\theta})\|_{L^{2}}\|J^{s_{1}}_{x}(u_{\lambda}^{\tau}-u_{\lambda}^{\theta})\|_{L^{2}}+$
$\displaystyle+\|J^{s_{1}}_{x}(u_{\lambda}^{\tau}+u_{\lambda}^{\theta})\|_{L^{2}}\|\partial_{x}(u_{\lambda}^{\tau}-u_{\lambda}^{\theta})\|_{L^{\infty}}\|J^{s_{1}}_{x}(u_{\lambda}^{\tau}-u_{\lambda}^{\theta})\|_{L^{2}})-$
$\displaystyle-\frac{1}{2}\int_{\mathbb{R}^{2}}\partial_{x}(u_{\lambda}^{\tau}+u_{\lambda}^{\theta})(J^{s_{1}}_{x}(u_{\lambda}^{\tau}-u_{\lambda}^{\theta}))^{2}\,dxdy$
$\displaystyle\leq$ $\displaystyle
C(\|\partial_{x}(u_{\lambda}^{\tau}+u_{\lambda}^{\theta})\|_{L^{\infty}}\|J_{x}^{s_{1}}(u_{\lambda}^{\tau}-u_{\lambda}^{\theta})\|_{L^{2}}^{2}+$
$\displaystyle+\|\partial_{x}(u_{\lambda}^{\tau}-u_{\lambda}^{\theta})\|_{L^{\infty}}\|J_{x}^{s_{1}}(u_{\lambda}^{\tau}+u_{\lambda}^{\theta})\|_{L^{2}}\|J_{x}^{s_{1}}(u_{\lambda}^{\tau}-u_{\lambda}^{\theta})\|_{L^{2}})$
With the second term we proceed in the same way to obtain the following
inequality
$\frac{1}{2}\int_{\mathbb{R}^{2}}J^{s_{1}}_{x}\big{(}\partial_{x}(u_{\lambda}^{\tau}+u_{\lambda}^{\theta})(u_{\lambda}^{\tau}-u_{\lambda}^{\theta})\big{)}J^{s_{1}}\partial_{x}(u_{\lambda}^{\tau}-u_{\lambda}^{\theta})\,dxdy\leq\\\
\leq
C((\|\partial_{x}u_{\lambda}^{\tau}\|_{L^{\infty}}+\|\partial_{x}u_{\lambda}^{\theta}\|_{L^{\infty}})\|J_{x}^{s_{1}}(u_{\lambda}^{\tau}-u_{\lambda}^{\theta})\|_{L^{2}}^{2}+\\\
+\|J_{x}^{s_{1}}(u_{\lambda}^{\tau}+u_{\lambda}^{\theta})\|_{L^{2}}\|\partial_{x}(u_{\lambda}^{\tau}-u_{\lambda}^{\theta})\|_{L^{\infty}}\|J_{x}^{s_{1}}(u_{\lambda}^{\tau}-u_{\lambda}^{\theta})\|_{L^{2}})$
(96)
So, $\|\psi_{\lambda}^{\tau}\|_{H^{s_{1},s_{2}}}\leq R$, for all $\lambda$,
from (94), (95), (96) and Lemma 3.10 we have
$\frac{1}{2}\frac{d}{dt}\|J_{x}^{s_{1}}(u_{\lambda}^{\tau}-u_{\lambda}^{\theta})\|_{L^{2}}^{2}\leq
C\big{(}(\|\partial_{x}u_{\lambda}^{\tau}\|_{L^{\infty}}+\|\partial_{x}u_{\lambda}^{\theta}\|_{L^{\infty}})\|J_{x}^{s_{1}}(u_{\lambda}^{\tau}-u_{\lambda}^{\theta})\|_{L^{2}}^{2}+\\\
+\|\partial_{x}(u_{\lambda}^{\tau}-u_{\lambda}^{\theta})\|_{L^{\infty}}\big{)}$
(97)
On the other hand,
$\displaystyle\frac{1}{2}\frac{d}{dt}\|J_{y}^{s_{2}}(u_{\lambda}^{\tau}$
$\displaystyle-
u_{\lambda}^{\theta})\|_{L^{2}}^{2}=\int_{\mathbb{R}^{2}}J_{y}^{s_{2}}(u_{\lambda}^{\tau}\partial_{x}u_{\lambda}^{\tau}-u_{\lambda}^{\theta}\partial_{x}u_{\lambda}^{\theta})J_{y}^{s_{2}}(u_{\lambda}^{\tau}-u_{\lambda}^{\theta})\,dxdy$
(98) $\displaystyle=$
$\displaystyle\frac{1}{2}\int_{\mathbb{R}^{2}}J^{s_{2}}_{y}\partial_{x}\left((u_{\lambda}^{\tau}+u_{\lambda}^{\theta})(u_{\lambda}^{\tau}-u_{\lambda}^{\theta})\right)J_{y}^{s_{2}}\partial_{(}u_{\lambda}^{\tau}-u_{\lambda}^{\theta})\,dxdy$
$\displaystyle=$
$\displaystyle\frac{1}{2}\int_{\mathbb{R}^{2}}J^{s_{2}}_{y}\left((u_{\lambda}^{\tau}+u_{\lambda}^{\tau})\partial_{x}(u_{\lambda}^{\theta}-u_{\lambda}^{\theta})\right)J^{s_{2}}_{y}(u_{\lambda}^{\tau}-u_{\lambda}^{\theta})\,dxdy+$
$\displaystyle+\frac{1}{2}\int_{\mathbb{R}^{2}}\left(J_{y}^{s_{2}}\partial_{x}(u_{\lambda}^{\tau}+u_{\lambda}^{\tau})(u_{\lambda}^{\theta}-u_{\lambda}^{\theta})\right)J^{s_{2}}(u_{\lambda}^{\tau}-u_{\lambda}^{\theta})\,dxdy.$
Proceeding in the same way that allows us to obtain (95) and (96), it follows
that
$\displaystyle\frac{1}{2}\int_{\mathbb{R}^{2}}$ $\displaystyle
J^{s_{2}}_{y}\big{(}(u_{\lambda}^{\tau}+u_{\lambda}^{\theta})\partial_{x}(u_{\lambda}^{\tau}-u_{\lambda}^{\theta})\big{)}J^{s_{2}}\partial_{x}(u_{\lambda}^{\tau}-u_{\lambda}^{\theta})\,dxdy\leq$
(99) $\displaystyle\leq$ $\displaystyle
C\big{(}(\|\partial_{y}u_{n}^{\tau}\|_{L^{\infty}}+\|\partial_{y}u_{n}^{\theta}\|_{L^{\infty}})\times$
$\displaystyle\times(\|J_{y}^{s_{2}-1}\partial_{x}(u_{\lambda}^{\tau}-u_{\lambda}^{\theta})\|_{L^{2}}\|J_{y}^{s_{2}}(u_{\lambda}^{\tau}-u_{\lambda}^{\theta})\|_{L^{2}}+\|J_{y}^{s_{2}}(u_{\lambda}^{\tau}-u_{\lambda}^{\theta})\|_{L^{2}}^{2})+$
$\displaystyle+\|J_{y}^{s_{2}}(u_{n}^{\tau}+u_{\lambda}^{\theta})\|_{L^{2}}\|\partial_{x}(u_{\lambda}^{\tau}-u_{\lambda}^{\theta})\|_{L^{\infty}}\|J_{y}^{s_{2}}(u_{\lambda}^{\tau}-u_{\lambda}^{\theta})\|_{L^{2}}\big{)}$
$\displaystyle\leq$ $\displaystyle
C\big{(}(\|\partial_{y}u_{\lambda}^{\tau}\|_{L^{\infty}}+\|\partial_{y}u_{\lambda}^{\theta}\|_{L^{\infty}})(\|J_{x}^{s_{1}}(u_{\lambda}^{\tau}-u_{\lambda}^{\theta})\|_{L^{2}}^{2}+\|J_{y}^{s_{2}}(u_{\lambda}^{\tau}-u_{\lambda}^{\theta})\|_{L^{2}}^{2})+$
$\displaystyle+\|\partial_{x}(u_{\lambda}^{\tau}-u_{\lambda}^{\theta})\|_{L^{\infty}}\big{)}$
and
$\displaystyle\frac{1}{2}\int_{\mathbb{R}^{2}}J^{s_{2}}_{y}\big{(}$
$\displaystyle\partial_{x}(u_{\lambda}^{\tau}+u_{\lambda}^{\theta})(u_{\lambda}^{\tau}-u_{\lambda}^{\theta})\big{)}J^{s_{2}}\partial_{x}(u_{\lambda}^{\tau}-u_{\lambda}^{\theta})\,dxdy\leq$
(100) $\displaystyle\leq$ $\displaystyle
C\big{(}(\|\partial_{x}u_{\lambda}^{\tau}\|_{L^{\infty}}+\|\partial_{x}u_{\lambda}^{\theta}\|_{L^{\infty}})\|J_{y}^{s_{2}}(u_{\lambda}^{\tau}-u_{\lambda}^{\theta})\|_{L^{2}}^{2}+$
$\displaystyle+\|J_{y}^{s_{2}-1}\partial_{x}(u_{n}^{\tau}+u_{\lambda}^{\theta})\|_{L^{2}}\|J_{y}^{s_{2}}(u_{\lambda}^{\tau}-u_{\lambda}^{\theta})\|_{L^{2}}\|\partial_{y}(u_{\lambda}^{\tau}-u_{\lambda}^{\theta})\|_{L^{\infty}})$
$\displaystyle\leq$ $\displaystyle
C(\|\partial_{x}u_{\lambda}^{\tau}\|_{L^{\infty}}+\|\partial_{x}u_{\lambda}^{\theta}\|_{L^{\infty}})(\|J_{x}^{s_{1}}(u_{\lambda}^{\tau}-u_{\lambda}^{\theta})\|_{L^{2}}^{2}+\|J_{y}^{s_{2}}(u_{\lambda}^{\tau}-u_{\lambda}^{\theta})\|_{L^{2}}^{2})+$
$\displaystyle+\|\partial_{y}(u_{\lambda}^{\tau}-u_{\lambda}^{\theta})\|_{L^{\infty}})$
Therefore, from (98), (99) and (100), we have
$\frac{1}{2}\frac{d}{dt}\|J_{y}^{s_{2}}(u_{\lambda}^{\tau}-u_{\lambda}^{\theta})\|_{L^{2}}^{2}\leq\\\
\leq
C(\|\partial_{x}u_{\lambda}^{\tau}\|_{L^{\infty}}+\|\partial_{x}u_{\lambda}^{\theta}\|_{L^{\infty}}+\|\partial_{y}u_{\lambda}^{\tau}\|_{L^{\infty}}+\|\partial_{y}u_{\lambda}^{\theta}\|_{L^{\infty}})\times\\\
\times(\|J_{x}^{s_{1}}(u_{\lambda}^{\tau}-u_{\lambda}^{\theta})\|_{L^{2}}^{2}+\|J_{y}^{s_{2}}(u_{\lambda}^{\tau}-u_{\lambda}^{\theta})\|_{L^{2}}^{2})+\\\
+\|\partial_{x}(u_{\lambda}^{\tau}-u_{\lambda}^{\theta})\|_{L^{\infty}}+\|\partial_{y}(u_{\lambda}^{\tau}-u_{\lambda}^{\theta})\|_{L^{\infty}})$
(101)
Gathering (97) and (101), we obtain
$\frac{1}{2}\frac{d}{dt}\|u_{\lambda}^{\tau}-u_{\lambda}^{\theta}\|_{H^{s_{1},s_{2}}}^{2}\leq\\\
\leq
C\big{(}(\|\partial_{x}u_{\lambda}^{\tau}\|_{L^{\infty}}+\|\partial_{x}u_{\lambda}^{\theta}\|_{L^{\infty}}+\|\partial_{y}u_{\lambda}^{\tau}\|_{L^{\infty}}+\|\partial_{y}u_{\lambda}^{\theta}\|_{L^{\infty}})\|u_{\lambda}^{\tau}-u_{\lambda}^{\theta}\|_{H^{s_{1},s_{2}}}^{2}+\\\
+\|\partial_{x}(u_{\lambda}^{\tau}-u_{\lambda}^{\theta})\|_{L^{\infty}}+\|\partial_{y}(u_{\lambda}^{\tau}-u_{\lambda}^{\theta})\|_{L^{\infty}}\big{)}$
(102)
Integrating, it follows that
$\|u_{\lambda}^{\tau}-u_{\lambda}^{\theta}\|_{H^{s_{1},s_{2}}}^{2}\leq\|\psi_{\lambda}^{\tau}-\psi_{\lambda}^{\theta}\|_{H^{s_{1},s_{2}}}^{2}+\\\
+C\int_{0}^{t}(\|\partial_{x}u_{\lambda}^{\tau}\|_{L^{\infty}}+\|\partial_{x}u_{\lambda}^{\theta}\|_{L^{\infty}})\|u_{\lambda}^{\tau}-u_{\lambda}^{\theta}\|_{H^{s_{1},s_{2}}}^{2}\,dt^{\prime}+\\\
+\|\partial_{x}(u_{\lambda}^{\tau}-u_{\lambda}^{\theta})\|_{L_{T_{0}}^{1}L_{x,y}^{\infty}}+\|\partial_{x}(u_{\lambda}^{\tau}-u_{\lambda}^{\theta})\|_{L_{T_{0}}^{1}L_{x,y}^{\infty}}$
(103)
Let us estimate the last two terms of the last inequality. Observe that
$u_{\lambda}^{\tau}-u_{\lambda}^{\theta}=W_{\alpha}(t)(\psi_{\lambda}^{\tau}-\psi_{\lambda}^{\theta})+\int_{0}^{t}W_{\alpha}(t-t^{\prime})\partial_{x}((u_{\lambda}^{\tau}+u_{\lambda}^{\theta})(u_{\lambda}^{\tau}-u_{\lambda}^{\theta}))(t^{\prime})\,dt^{\prime}.$
Therefore, proceeding as in the proof of Lemma 3.8, from refined the
Strichartz estimate (57) and Lemma 3.13, we have
$\displaystyle\|\partial_{x}(u_{\lambda}^{\tau}-u_{\lambda}^{\theta})\|_{L_{T_{0}}^{1}L_{x,y}^{\infty}}\leq$
$\displaystyle
C(\|u_{\lambda}^{\tau}-u_{\lambda}^{\theta}\|_{L_{T_{0}}^{\infty}H_{x,y}^{s_{1}^{\prime},s_{2}^{\prime}}}+\|(u_{\lambda}^{\tau}+u_{\lambda}^{\theta})(u_{\lambda}^{\tau}-u_{\lambda}^{\theta})\|_{L_{T_{0}}^{\infty}H_{x,y}^{s_{1}^{\prime},s_{2}^{\prime}}})$
$\displaystyle\leq$ $\displaystyle
C(\|u_{\lambda}^{\tau}-u_{\lambda}^{\theta}\|_{L_{T_{0}}^{\infty}H_{x,y}^{s_{1}^{\prime},s_{2}^{\prime}}}+\|u_{\lambda}^{\tau}+u_{\lambda}^{\theta}\|_{L_{T_{0}}^{\infty}H_{x,y}^{s_{1}^{\prime},s_{2}^{\prime}}}\|u_{\lambda}^{\tau}-u_{\lambda}^{\theta})\|_{L_{T_{0}}^{\infty}H_{x,y}^{s_{1}^{\prime},s_{2}^{\prime}}})$
$\displaystyle\leq$ $\displaystyle
C\|u_{\lambda}^{\tau}-u_{\lambda}^{\theta}\|_{L_{T_{0}}^{\infty}H_{x,y}^{s_{1}^{\prime},s_{2}^{\prime}}}$
$\displaystyle\leq$ $\displaystyle
C\|\psi_{\lambda}^{\tau}-\psi_{\lambda}^{\theta}\|_{L^{2}}^{\nu}\|u_{\lambda}^{\tau}-u_{\lambda}^{\theta}\|_{L_{T_{0}}^{\infty}H_{x,y}^{s_{1},s_{2}}}^{1-\nu}$
$\displaystyle\leq$ $\displaystyle C\tau^{\nu},$
where $s^{\prime}_{1}$, $s^{\prime}_{2}$ and $\nu$ are such that
$1<s^{\prime}_{i}=\nu s_{i}<s_{i}$, $i=1,2$. In the same way, again, if we
proceed as in the proof of Lemma 3.8, more precisely, in the way that we got
(84), we also have
$\|\partial_{y}(u_{\lambda}^{\tau}-u_{\lambda}^{\theta})\|_{L_{T_{0}}^{1}L_{x,y}^{\infty}}\leq
C\tau^{\nu}$
From these last two inequalities with (103) and the Gronwall Lemma follows the
proposition. ∎
###### Corollary 3.18.
Let $\psi_{\lambda}^{\tau}$ and $u_{\lambda}^{\tau}$ be as in the result
above. Then, $\left\\{u_{\lambda}^{\tau}\right\\}_{\tau>0}$ converges
uniformly in $t$ to $u_{\lambda}$, when $\tau\to 0+$. In other words,
$\lim_{\tau\to
0+}\sup_{t\in[0,T]}\left\|u_{\lambda}^{\tau}(t)-u_{\lambda}(t)\right\|_{H^{s_{1},s_{2}}}=0.$
###### Proof.
Take $\theta=0$ in Proposition 3.17. ∎
###### Theorem 3.19.
In $H^{s_{1},s_{2}}$ the map $\psi\mapsto u$, where $u$ is solution to (13)
with initial condition $\psi$, is continuous. More precisely, if
$(\psi_{n})_{n\in\mathbb{N}}$ is a sequence such that $\psi_{n}\to\psi$ in
$H^{s_{1},s_{2}}$ and if $u_{n}\in C([0,T_{0}];H^{s_{1},s_{2}})$ are the
corresponding solutions to (13) with initial condition $\psi_{n}$, then
$\lim_{n\to\infty}\sup_{t\in[0,T_{0}]}\left\|u_{n}(t)-u(t)\right\|_{s}=0$
###### Proof.
Let $\psi\in H^{s_{1},s_{2}}$ and $(\psi_{n})$ be a sequence in
$H^{s_{1},s_{2}}$ that converges strongly to $\psi$ in this space. Let, also,
$R=\max(\sup_{n}\|\psi_{n}\|_{H^{s_{1},s_{2}}},\|\psi\|_{H^{s_{1},s_{2}}})$.
Now, let us take $\psi_{m,n}=e^{\frac{1}{m}\bigtriangleup}\psi_{n}$ and
$\psi_{m}=e^{\frac{1}{m}\bigtriangleup}\psi$. Let $u_{n}$, $u$, $u_{m,n}$,
$u_{m}$, $u^{\tau}_{n}$, $u^{\tau}$, $u^{\tau}_{m,n}$, $u^{\tau}_{m}$ be the
corresponding solutions to (13) in $[0,T_{0}]$ with conditions $\psi_{n}$,
$\psi$, $\psi_{m,n}$, $\psi_{m}$, $\psi^{\tau}_{n}$, $\psi^{\tau}$,
$\psi^{\tau}_{m,n}$, $\psi^{\tau}_{m}$, respectively. Also, observe that
$u_{n,m}$, $u_{m}$, $u^{\tau}_{m,n}$ and $u^{\tau}_{m}$ converge uniformly to
$u_{n}$, $u$, $u^{\tau}_{n}$ and $u^{\tau}$ in the weak sense, as
$m\to\infty$. Therefore,
$\displaystyle\langle u_{n}-u,\varphi\rangle_{H^{s_{1},s_{2}}}=$
$\displaystyle\lim_{m\to\infty}\langle
u_{m,n}-u_{m,n}^{\tau},\varphi\rangle_{H^{s_{1},s_{2}}}+\langle
u_{m,n}^{\tau}-u_{m}^{\tau},\varphi\rangle_{H^{s_{1},s_{2}}}+$
$\displaystyle+\langle u_{m}^{\tau}-u_{m},\varphi\rangle_{H^{s_{1},s_{2}}}$
$\displaystyle=\lim_{m\to\infty}$ $\displaystyle\left[\langle
u_{m,n}-u_{m,n}^{\tau},\varphi\rangle_{H^{s_{1},s_{2}}}+\langle
u_{m,n}^{\tau}-u_{m},\varphi\rangle_{H^{s_{1},s_{2}}}\right]+$
$\displaystyle+\langle
u_{n}^{\tau}-u^{\tau},\varphi\rangle_{H^{s_{1},s_{2}}}.$
On the other hand, Corollary 3.18 implies that, given $\epsilon>0$, there
exists $\tau_{0}$ such that for $0<\tau\leq\tau_{0}$
$\lvert\langle u_{m,n}-u_{m,n}^{\tau},\varphi\rangle_{H^{s_{1},s_{2}}}+\langle
u_{m,n}^{\tau}-u_{m},\varphi\rangle_{H^{s_{1},s_{2}}}\rvert\leq\epsilon\left\|\varphi\right\|_{H^{s_{1},s_{2}}},$
for all $m>0$. Then,
$\displaystyle\lvert\langle
u_{n}-u,\varphi\rangle_{H^{s_{1},s_{2}}}\rvert\leq\epsilon\left\|\varphi\right\|_{H^{s_{1},s_{2}}}+\left\|u_{n}^{\tau}-u^{\tau}\right\|_{H^{s_{1},s_{2}}}\left\|\varphi\right\|_{H^{s_{1},s_{2}}},$
for all $\varphi\in H^{s_{1},s_{2}}$. Therefore,
$\left\|u_{n}-u\right\|_{H^{s_{1},s_{2}}}\leq\epsilon+\left\|u_{n}^{\tau}-u^{\tau}\right\|_{H^{s_{1},s_{2}}}.$
(104)
Arguments similar to those used in Proposition 3.17 allow us show that, for
$\tau$ small enough,
$\left\|u_{n}^{\tau}-u^{\tau}\right\|_{H^{s_{1},s_{2}}}\leq
C\left\|\psi_{n}^{\tau}-\psi^{\tau}\right\|_{H^{s_{1},s_{2}}}\tau^{-\frac{1}{s}}\leq\left\|\psi_{n}-\psi\right\|_{H^{s_{1},s_{2}}}\tau^{-\frac{1}{s}}.$
Then, fixing $\tau$ small enough, we can conclude from (104) that
$\left\|u_{n}-u\right\|_{H^{s_{1},s_{2}}}\leq 2\epsilon,$
for $n$ large enough. ∎
Let us see now that the problem is locally well-posed in the spaces
$X^{s_{1},s_{2}}$, $\widehat{X}^{s_{1},s_{2}}$ and $Y^{s_{1},s_{2}}$. Since
the solution to (13) satisfy the integral equation
$u=W_{\alpha}(t)\psi+\int_{0}^{t}W_{\alpha}(t-t^{\prime})(u\partial_{x}u)(t^{\prime})\,dt^{\prime},$
we have that $\partial_{x}^{-1}u$ and $\partial_{x}^{-1}\partial_{y}u$ satisfy
the equations
$\partial_{x}^{-1}u=W_{\alpha}(t)\partial_{x}^{-1}\psi+\int_{0}^{t}W_{\alpha}(t-t^{\prime})\left(\frac{u^{2}}{2}\right)(t^{\prime})\,dt^{\prime}$
and
$\partial_{x}^{-1}\partial_{y}u=W_{\alpha}(t)\partial_{x}^{-1}\partial_{y}\psi+\int_{0}^{t}W_{\alpha}(t-t^{\prime})(u\partial_{y}u)(t^{\prime})\,dt^{\prime},$
respectively. From here and Theorem 3.9, for $s_{1}$ and $s_{2}$ as in that
theorem, it follows that (13) is locally well-posed in the spaces
$\widehat{X}^{s_{1},s_{2}}$, $X^{s_{1},s_{2}}$ and
$\widehat{Y}^{s_{1},s_{2}}$. The case $Y^{s_{1},s_{2}}$ requires some
additional effort. For $\psi\in Y^{\infty}$, the solution $u$ to (13), with
initial condition $\psi$, belongs to $C([0,T_{0}];Y^{\infty})$. We can proceed
as in the proof Lemma 3.7, to show that
$\frac{1}{2}\frac{d}{dt}\|\partial_{x}^{-1}\partial_{y}u\|^{2}\leq
C(\|u_{x}\|+\|u_{y}\|)\|u\|_{Y^{s_{1},s_{2}}}^{2}.$
This inequality with (70) allow us prove that
$\frac{1}{2}\frac{d}{dt}\|u\|_{Y^{s_{1},s_{2}}}^{2}\leq
C(\|u_{x}\|_{\infty}+\|u_{y}\|_{\infty})\|u\|_{Y^{s_{1},s_{2}}}^{2}.$ (105)
Thanks to the Gronwall lemma we have the following generalization of Lemma
3.7,
$\|u\|_{Y^{s}}\leq\|\psi\|_{Y^{s_{1},s_{2}}}e^{C(\|u_{x}\|_{L^{1}_{t}L_{x,y}^{\infty}}+\|u_{y}\|_{L^{1}_{t}L_{x,y}^{\infty}})}.$
Now, if $\psi$ is arbitrary, in the same way that we prove for the space
$H^{s_{1},s_{2}}$, the solution $u\in C([0,T];Y^{s_{1},s_{2}})$. To see that
the map $\psi\mapsto u$ is continuous from ${Y^{s_{1},s_{2}}}$ to
$C([0,T];Y^{s_{1},s_{2}})$ we repeat the same argument of Bona-Smith that we
use before. So, summarizing, we paraphrase Theorem (2.7) for the current
situation.
###### Theorem 3.20.
Let $s_{1},s_{2}$ and $\alpha$ be as in Theorem 3.9. Let, also, $Z$ any of the
spaces $X^{s_{1},s_{2}}$, $\widehat{X}^{s_{1},s_{2}}$, $Y^{s_{1},s_{2}}$ and
$\widehat{Y}^{s_{1},s_{2}}$. Then, if $\psi\in Z$ and $u\in
C([0,T];H^{s_{1},s_{2}})$ is solution to (13) with $u(0)=\psi$, then $u\in
C([0,T];Z)$. Moreover, $\psi\mapsto u$ is continuous from $Z$ to $C([0,T];Z)$
## 4\. Remarks on ill-posedness of the equation (13)
In this section we prove that the flow associated to the equation (13) is not
of class $C^{2}$ for $-1\leq\alpha<0$. In particular, we have that we cannot
apply Picard iterative process to solve the Duhamel equation associated to
this equation.
For this we will use the ideas given in [29] to prove that the flow associated
with the KP-I equation is not of class $C^{2}$.
### 4.1. The flow associated to the problem (13) is not $C^{2}$
###### Theorem 4.1.
Let $(s_{1},s_{2})\in\mathbb{R}^{2}$ and $-1\leq\alpha<0$. Then, there does
not exist $T>0$ such that (13) has a unique solution $u$ for all $\phi\in
H^{s_{1}.s_{2}}$ and that the flow $S_{t}:\phi\mapsto u$ is not of class
$C^{2}$ at $0$ from $H^{s_{1}.s_{2}}$ to $H^{s_{1}.s_{2}}$
###### Proof.
Let’s see first what we should show. For this we consider
$u(\lambda,t)=S_{t}(\lambda\phi)$, $S_{t}$ the flow associated to the problem
(13). So this solution satisfies the Duhamel equation associated to the
equation (13), i. e.,
$u(\lambda,t)=\lambda
W_{\alpha}(t)\phi-\int_{0}^{t}W_{\alpha}(t-t^{\prime})u(t^{\prime})u_{x}(t^{\prime})\,dt^{\prime}.$
(106)
If the flow is twice differentiable around $0$ in $H^{s_{1}.s_{2}}$ then,
thanks to the chain rule,
$\partial_{\lambda}u(0,t)=W_{\alpha}(t)\phi,$
and
$\partial_{\lambda}^{2}u(0,t)=-2\int_{0}^{t}W_{\alpha}(t-t^{\prime})W_{\alpha}(t^{\prime})\phi
W_{\alpha}(t^{\prime})\phi_{x}\,dt^{\prime}.$
Which would imply that the map
$\phi\mapsto\int_{0}^{t}W_{\alpha}(t-t^{\prime})W_{\alpha}(t^{\prime})\phi
W_{\alpha}(t^{\prime})\phi_{x}\,dt^{\prime}$
would be a quadratic form coming from a continuous symmetric bilinear
transformation in $H^{s_{1}.s_{2}}$, and that, in particular, for some fixed
$C$
$\left\|\int_{0}^{t}W_{\alpha}(t-t^{\prime})W_{\alpha}(t^{\prime})\phi
W_{\alpha}(t^{\prime})\phi_{x}\,dt^{\prime}\right\|_{H^{s_{1}.s_{2}}}\leq
C\|\phi\|_{H^{s_{1}.s_{2}}}^{2}$
for all $\phi\in H^{s_{1}.s_{2}}$. So let’s show that this is not the case. To
do this, suppose that this inequality is valid and consider the function
$\phi$ defined via the Fourier transform by
$\widehat{\phi}=\gamma^{-3/2}\mathbbm{1}_{D_{1}}+\gamma^{-3/2}N^{-s_{1}-\frac{3-\alpha}{2}s_{2}}\mathbbm{1}_{D_{2}},$
where $\gamma$ and $N$ are positive numbers such that $\gamma\ll 1$ and $N\gg
1$, $D_{1}$ and $D_{2}$ are the sets
$D_{1}=\left[\frac{\gamma}{2},\gamma\right]\times\left[-\frac{\gamma^{2}}{6},\frac{\gamma^{2}}{6}\right]\
\text{and}\
D_{2}=\left[N,N+\gamma\right]\times\left[\sqrt{-\frac{3}{\alpha}}N^{\frac{3-\alpha}{2}},\sqrt{-\frac{3}{\alpha}}N^{\frac{3-\alpha}{2}}+\gamma^{2}\right]$
and $\mathbbm{1}_{D_{i}}$ are the characteristic functions of the sets
$D_{i}$, $i=1,2$. $\|\phi\|_{H^{s_{1}.s_{2}}}\sim 1$ for whatever values that
we take for $\gamma$ and $N$. Let us see that for a convenient choice in
parameters of $\gamma$ and $N$,
$\left\|\int_{0}^{t}W_{\alpha}(t-t^{\prime})W_{\alpha}(t^{\prime})\phi
W_{\alpha}(t^{\prime})\phi_{x}\,dt^{\prime}\right\|_{H^{s_{1}.s_{2}}}$ (107)
can be as large as we want. Calculating its Fourier transform, it follows that
$\int_{0}^{t}W_{\alpha}(t-t^{\prime})W_{\alpha}(t^{\prime})\phi
W_{\alpha}(t^{\prime})\phi_{x}\,dt^{\prime}$ (108)
is $f_{1}+f_{2}+f_{3}$ where
$\widehat{f}_{1}(t,\xi,\eta)=\frac{i\xi
e^{it(\xi^{3}+\operatorname{sgn}(\xi)|\xi|^{\alpha}\eta^{2})}}{2\gamma^{3}}\int_{\begin{subarray}{c}(\xi_{1},\eta_{1})\in
D_{1}\\\ (\xi-\xi_{1},\eta-\eta_{1})\in
D_{1}\end{subarray}}\frac{e^{-it\chi(\xi,\xi_{1},\eta,\eta_{1})}-1}{\chi(\xi,\xi_{1},\eta,\eta_{1})}\,d\xi_{1}d\eta_{1},$
$\widehat{f}_{2}(t,\xi,\eta)=\frac{i\xi
e^{it(\xi^{3}+\operatorname{sgn}(\xi)|\xi|^{\alpha}\eta^{2})}}{2\gamma^{3}N^{2(s_{1}+\frac{3-\alpha}{2}s_{2})}}\int_{\begin{subarray}{c}(\xi_{1},\eta_{1})\in
D_{2}\\\ (\xi-\xi_{1},\eta-\eta_{1})\in
D_{2}\end{subarray}}\frac{e^{-it\chi(\xi,\xi_{1},\eta,\eta_{1})}-1}{\chi(\xi,\xi_{1},\eta,\eta_{1})}\,d\xi_{1}d\eta_{1}$
and
$\displaystyle\widehat{f}_{3}(t,\xi,\eta)=$ $\displaystyle\frac{i\xi
e^{it(\xi^{3}+\operatorname{sgn}(\xi)|\xi|^{\alpha}\eta^{2})}}{2\gamma^{3}N^{s_{1}+\frac{3-\alpha}{2}s_{2}}}\int_{\begin{subarray}{c}(\xi_{1},\eta_{1})\in
D_{2}\\\ (\xi-\xi_{1},\eta-\eta_{1})\in
D_{1}\end{subarray}}\frac{e^{-it\chi(\xi,\xi_{1},\eta,\eta_{1})}-1}{\chi(\xi,\xi_{1},\eta,\eta_{1})}\,d\xi_{1}d\eta_{1}+$
$\displaystyle+\frac{i\xi
e^{it(\xi^{3}+\operatorname{sgn}(\xi)|\xi|^{\alpha}\eta^{2})}}{2\gamma^{3}N^{s_{1}+\frac{3-\alpha}{2}s_{2}}}\int_{\begin{subarray}{c}(\xi_{1},\eta_{1})\in
D_{1}\\\ (\xi-\xi_{1},\eta-\eta_{1})\in
D_{2}\end{subarray}}\frac{e^{-it\chi(\xi,\xi_{1},\eta,\eta_{1})}-1}{\chi(\xi,\xi_{1},\eta,\eta_{1})}\,d\xi_{1}d\eta_{1},$
where $\chi$ is the resonant function
$\begin{split}\chi&=\chi(\xi,\xi_{1},\eta,\eta_{1})=\vartheta(\xi,\eta)-\vartheta(\xi_{1},\eta_{1})-\vartheta(\xi-\xi_{1},\eta-\eta_{1})\\\
&=3\xi\xi_{1}(\xi-\xi_{1})+\operatorname{sgn}(\xi)\frac{\eta^{2}}{|\xi|^{\theta}}-\operatorname{sgn}(\xi_{1})\frac{\eta_{1}^{2}}{|\xi_{1}|^{\theta}}-\operatorname{sgn}(\xi-\xi_{1})\frac{(\eta-\eta_{1})^{2}}{|\xi-\xi_{1}|^{\theta}},\end{split}$
(109)
where
$\vartheta(\xi,\eta)=\xi^{3}+\operatorname{sgn}(\xi)\frac{\eta^{2}}{|\xi|^{\theta}},$
(110)
is the phase function and $\theta=-\alpha$, which is between $0$ and $1$.
Since in our case $\xi,\xi_{1}$ and $\xi-\xi_{1}$ are positive, we have
$\chi=3\xi\xi_{1}(\xi-\xi_{1})+\frac{\eta^{2}}{\xi^{\theta}}-\frac{\eta_{1}^{2}}{\xi_{1}^{\theta}}-\frac{(\eta-\eta_{1})^{2}}{(\xi-\xi_{1})^{\theta}}$
(111)
Observe that to estimate (107) it is enough estimate
$\|f_{3}(t)\|_{H^{s_{1},s_{2}}}$, in fact,
$\left\|\int_{0}^{t}W_{\alpha}(t-t^{\prime})W_{\alpha}(t^{\prime})\phi
W_{\alpha}(t^{\prime})\phi_{x}\,dt^{\prime}\right\|_{H^{s_{1}.s_{2}}}\geq\|f_{3}(t)\|_{H^{s_{1},s_{2}}}.$
To continue we need the following lemma.
###### Lemma 4.2.
Suppose that
$(\xi_{1},\eta_{1})\in
D_{1}\qquad\text{and}\qquad(\xi-\xi_{1},\eta-\eta_{1})\in D_{2}$
or
$(\xi_{1},\eta_{1})\in
D_{2}\qquad\text{and}\qquad(\xi-\xi_{1},\eta-\eta_{1})\in D_{1}.$
Then,
$|\chi(\xi,\xi_{1},\eta,\eta_{1})|\lesssim\gamma^{2}N.$
###### Proof.
First let us calculate the $\eta$ such that
$\chi(\xi,\xi_{1},\eta,\eta_{1})=0$. So, we have
$[\xi^{\theta}-(\xi-\xi_{1})^{\theta}]\xi_{1}^{\theta}\eta^{2}-2\xi^{\theta}\xi_{1}^{\theta}\eta_{1}\eta+[\xi_{1}^{\theta}+(\xi-\xi_{1})^{\theta}]\xi^{\theta}\eta_{1}^{2}-3\xi^{1+\theta}\xi_{1}^{1+\theta}(\xi-\xi_{1})^{1+\theta}=0$
(112)
Then, we get
$\eta=\frac{\xi^{\theta}\eta_{1}}{[\xi^{\theta}-(\xi-\xi_{1})^{\theta}]}\pm\sqrt{\frac{3\xi^{1+\theta}\xi_{1}(\xi-\xi_{1})^{1+\theta}}{[\xi^{\theta}-(\xi-\xi_{1})^{\theta}]}+\frac{\xi^{\theta}(\xi-\xi_{1})^{\theta}[\xi_{1}^{\theta}+(\xi-\xi_{1})^{\theta}-\xi^{\theta}]\eta_{1}^{2}}{[\xi^{\theta}-(\xi-\xi_{1})^{\theta}]^{2}\xi_{1}^{\theta}}}$
(113)
Let $\eta*$ be the smallest zero between the two calculated above. So,
$\eta^{*}-\eta_{1}=\frac{(\xi-\xi_{1})^{\theta}\eta_{1}}{[\xi^{\theta}-(\xi-\xi_{1})^{\theta}]}-\sqrt{\frac{3\xi^{1+\theta}\xi_{1}(\xi-\xi_{1})^{1+\theta}}{[\xi^{\theta}-(\xi-\xi_{1})^{\theta}]}+\frac{\xi^{\theta}(\xi-\xi_{1})^{\theta}[\xi_{1}^{\theta}+(\xi-\xi_{1})^{\theta}-\xi^{\theta}]\eta_{1}^{2}}{[\xi^{\theta}-(\xi-\xi_{1})^{\theta}]^{2}\xi_{1}^{\theta}}}.$
(114)
Let
$R=\frac{(\xi-\xi_{1})^{\theta}\eta_{1}}{[\xi^{\theta}-(\xi-\xi_{1})^{\theta}]}+\sqrt{\frac{3\xi^{1+\theta}\xi_{1}(\xi-\xi_{1})^{1+\theta}}{[\xi^{\theta}-(\xi-\xi_{1})^{\theta}]}+\frac{\xi^{\theta}(\xi-\xi_{1})^{\theta}[\xi_{1}^{\theta}+(\xi-\xi_{1})^{\theta}-\xi^{\theta}]\eta_{1}^{2}}{[\xi^{\theta}-(\xi-\xi_{1})^{\theta}]^{2}\xi_{1}^{\theta}}}.$
Then
$|\eta^{*}-\eta_{1}|=\dfrac{\frac{(\xi^{\theta}-\xi_{1}^{\theta})(\xi-\xi_{1})^{\theta}}{(\xi^{\theta}-(\xi-\xi_{1})^{\theta})\xi_{1}^{\theta}}{\left|\eta_{1}^{2}-3\xi^{1+\theta}\xi_{1}^{1+\theta}g(\xi,\xi_{1})\right|}}{R},$
(115)
where
$g(\xi,\xi_{1})=\begin{cases}\dfrac{\xi-\xi_{1}}{\xi^{\theta}-\xi_{1}^{\theta}}&\text{
if }\xi\neq\xi_{1}\\\\[14.22636pt] \frac{1}{\theta}\xi_{1}^{1-\theta}&\text{
in another case.}\end{cases}$
Now, since
$\frac{1}{\theta}\xi_{1}^{1-\theta}\leq
g(\xi,\xi_{1})\leq\frac{1}{\theta}\xi^{1-\theta}$ (116)
and
$R\geq\frac{(\xi-\xi_{1})^{\theta}\eta_{1}}{(\xi^{\theta}-(\xi-\xi_{1})^{\theta})},$
we have
$\displaystyle|\eta^{*}-\eta_{1}|$
$\displaystyle=\frac{\xi^{\theta}-\xi_{1}^{\theta}}{\xi_{1}^{\theta}}\frac{\left|\eta_{1}^{2}-3\xi^{1+\theta}\xi_{1}^{1+\theta}g(\xi,\xi_{1})\right|}{\eta_{1}}$
$\displaystyle\leq
3\frac{\xi^{\theta}-\xi_{1}^{\theta}}{\xi_{1}^{\theta}}\left|\eta_{1}-\sqrt{3}\xi^{\frac{1+\theta}{2}}\xi_{1}^{\frac{1+\theta}{2}}\sqrt{g(\xi,\xi_{1})}\right|$
$\displaystyle\leq
3\theta\frac{\xi-\xi_{1}}{\xi_{1}}\left|\eta_{1}-\sqrt{\frac{3}{\theta}}\xi_{1}^{\frac{3+\theta}{2}}-\xi_{1}^{\frac{1+\theta}{2}}\left(\xi^{\frac{1+\theta}{2}}\sqrt{g(\xi,\xi_{1})}-\frac{1}{\sqrt{\theta}}\xi_{1}\right)\right|,$
Remember that $\eta_{1}$ take values in
$\left[\sqrt{\frac{3}{\theta}}N^{\frac{3+\theta}{2}},\sqrt{\frac{3}{\theta}}N^{\frac{3+\theta}{2}}+\gamma^{2}\right]$
and that $\xi_{1}$ in $[N,N+\gamma]$. Whence, thanks to the Taylor formula
with remainder applied to the function $x\mapsto x^{\frac{3+\theta}{2}}$,
$\sqrt{3}\xi_{1}^{\frac{3+\theta}{2}}\in\left[\sqrt{3}N^{\frac{3+\theta}{2}},\sqrt{3}N^{\frac{3+\theta}{2}}+\sqrt{3}\frac{3+\theta}{2}N^{\frac{1+\theta}{2}}\gamma+\sqrt{3}\frac{(3+\theta)(1+\theta)}{8}\gamma^{2}\right]$
and, moreover
$\left|\eta_{1}-\sqrt{3}\xi_{1}^{\frac{3+\theta}{2}}\right|\leq
2\sqrt{3}N^{\frac{1+\theta}{2}}\gamma+\sqrt{3}\gamma^{2}.$
On the other hand from (116)
$\xi^{\frac{1+\theta}{2}}\sqrt{g(\xi,\xi_{1})}-\frac{1}{\sqrt{\theta}}\xi_{1}\leq\frac{1}{\sqrt{\theta}}(\xi-\xi_{1})\leq\frac{1}{\sqrt{\theta}}\gamma.$
Therefore,
$|\eta^{*}-\eta_{1}|\leq
3\theta\frac{\gamma}{N}\left(2\sqrt{\frac{3}{\theta}}N^{\frac{1+\theta}{2}}\gamma+\sqrt{\frac{3}{\theta}}\gamma^{2}+\sqrt{\frac{3}{\theta}}(N^{\frac{1+\theta}{2}}+\gamma)\gamma\right)\leq
18\sqrt{\theta}\frac{\gamma^{2}}{N^{\frac{1-\theta}{2}}}.$
By the mean value theorem, there exists $\bar{\eta}\in[\eta,\eta^{*}]$ such
that
$\displaystyle\chi(\xi,\xi_{1},\eta,\eta_{1})$
$\displaystyle=\chi(\xi,\xi_{1},\eta^{*},\eta_{1})+(\eta-\eta^{*})\partial_{\eta}\chi(\xi,\xi_{1},\bar{\eta},\eta_{1})$
$\displaystyle=-(\eta-\eta^{*})\frac{2(\bar{\eta}(\xi^{\theta}-(\xi-\xi_{1})^{\theta})-\eta_{1}\xi^{\theta})}{\xi^{\theta}(\xi-\xi_{1})^{\theta}}$
$\displaystyle=-(\eta-\eta^{*})\frac{2((\bar{\eta}-\eta_{1})\xi^{\theta}-(\xi-\xi_{1})^{\theta}\bar{\eta})}{\xi^{\theta}(\xi-\xi_{1})^{\theta}}$
$\displaystyle=-(\eta-\eta^{*})2\left(\frac{\bar{\eta}-\eta_{1}}{(\xi-\xi_{1})^{\theta}}-\frac{\bar{\eta}}{\xi^{\theta}}\right).$
So,
$\displaystyle|\chi(\xi,\xi_{1},\eta,\eta_{1})|$
$\displaystyle\lesssim\frac{\gamma^{2}}{N^{\frac{1-\theta}{2}}}\left(\frac{\gamma^{2}}{N^{\frac{1-\theta}{2}}\gamma^{\theta}}+\frac{N^{\frac{3+\theta}{2}}}{N^{\theta}}\right)$
$\displaystyle\lesssim\gamma^{2}N$
The lemma follows immediately observing that
$\chi(\xi,\xi_{1},\eta,\eta_{1})=\chi(\xi,\xi-\xi_{1},\eta,\eta-\eta_{1}).$
∎
Let us finish the proof of the theorem. Let us choose $\gamma$ and $N$ in such
a way that $\gamma^{2}N=N^{-\varepsilon}$ for $\varepsilon\ll 1$. Thanks to
the previous lemma we have that
$\left|\frac{e^{it\xi}-1}{\xi}\right|=|t|+O(N^{-\epsilon})$
for $(\xi_{1},\eta_{1})\in D_{1}$ and $(\xi-\xi_{1},\eta-\eta_{1})\in D_{2}$
or $(\xi_{1},\eta_{1})\in D_{2}$ and $(\xi-\xi_{1},\eta-\eta_{1})\in D_{1}$.
So
$\|f_{3}(t,\cdot,\cdot)\|_{H^{s}}\gtrsim\frac{NN^{\frac{3+\theta}{2}s}\gamma^{3}\gamma^{\frac{3}{2}}}{N^{\frac{3+\theta}{2}s}\gamma^{3}}=\gamma^{\frac{3}{2}}N=N^{(1-3\varepsilon)/4}.$
This leads to a contradiction since
$1\sim\|\phi\|_{H^{s}}^{2}\gtrsim\|f_{3}(t,\cdot,\cdot)\|_{H^{s}}.$
∎
An immediate consequence of the previous theorem is the following theorem.
###### Theorem 4.3.
For $(s_{1},s_{2})\in\mathbb{R}^{2}$, $-1\leq\alpha<0$ and a positive real
number $T$, there does not exists a space $X_{T}$ continuously embedded in
$C([-T,T];H^{s_{1},s_{2}})$ such that, for a fix constant $C$,
$\|W_{\alpha}(\cdot)\phi\|_{X_{T}}\leq C\|\phi\|_{H^{s_{1},s_{2}}},$ (117)
for all $\phi\in H^{s_{1},s_{2}}$, and
$\left\|\int_{0}^{t}W_{\alpha}(t-t^{\prime})(u(t^{\prime})u_{x}(t^{\prime}))\,dt^{\prime}\right\|_{X_{T}}\leq
C\|u\|_{X_{T}}^{2},$ (118)
for all $u\in X_{T}$.
Note that the estimates given in the theorem statement are necessary to prove
the contraction properties of the operator $\Phi$ defined by
$\Phi(u)=W_{\alpha}(t)\phi+\int_{0}^{t}W_{\alpha}(t-t^{\prime})(u(t^{\prime})u_{x}(t^{\prime}))\,dt^{\prime}.$
###### Proof.
Suppose that we have (117) and (118) for all $\phi\in H^{s_{1},s_{2}}$ and for
all $u\in X_{T}$. Let $\phi\in H^{s_{1},s_{2}}$ and set
$u(t)=W_{\alpha}(t)\phi$. Then,
$\left\|\int_{0}^{t}W_{\alpha}(t-t^{\prime})(W_{\alpha}(t^{\prime})\phi
W_{\alpha}(t^{\prime})\phi_{x})\,dt^{\prime}\right\|_{X_{T}}\leq
C\|W_{\alpha}(t)\phi\|_{X_{T}}^{2}\leq\|\phi\|_{H^{s_{1},s_{2}}}^{2}.$
Since $X_{T}$ is continuously embedded in $C([-T,T];H^{s_{1},s_{2}})$
$\left\|\int_{0}^{t}W_{\alpha}(t-t^{\prime})(W_{\alpha}(t^{\prime})\phi
W_{\alpha}(t^{\prime})\phi_{x})\,dt^{\prime}\right\|_{H^{s_{1},s_{2}}}\leq
C\|\phi\|_{H^{s_{1},s_{2}}}^{2},$
which is contradictory with the previous theorem. This ends the proof. ∎
Another immediate corollary is the following theorem.
###### Theorem 4.4.
The flow associated to the problem (13), for $-1\leq\alpha<0$, whose well-
posedness was proved in the previous section, is not of class $C^{2}$.
## References
* [1] Ablowitz, M. J., and Clarkson, P. A. Solitons, nonlinear evolution equations and inverse scattering, vol. 149 of London Mathematical Society Lecture Note Series. Cambridge University Press, Cambridge, 1991.
* [2] Ablowitz, M. J., and Segur, H. Long internal waves in fluids of great depth. Stud. Appl. Math. 62, 3 (1980), 249–262.
* [3] Benjamin, T. B. Internal waves of permanent form in fluids of great depth. J. Fluid Mech. 29 (1967), 559–592.
* [4] Benjamin, T. B. The stability of solitary waves. Proc. Roy. Soc. (London) Ser. A 328 (1972), 153–183.
* [5] Benjamin, T. B., Bona, J. L., and Mahony, J. J. Model equations for long waves in nonlinear dispersive systems. Philos. Trans. Roy. Soc. London Ser. A 272, 1220 (1972), 47–78.
* [6] Bona, J. On the stability theory of solitary waves. Proc. Roy. Soc. London Ser. A 344, 1638 (1975), 363–374.
* [7] Bona, J. L., and Smith, R. The initial-value problem for the Korteweg-de Vries equation. Philos. Trans. Roy. Soc. London Ser. A 278, 1287 (1975), 555–601.
* [8] Castro, A., Córdoba, D., and Gancedo, F. Singularity formations for a surface wave model. Nonlinearity 23, 11 (2010), 2835–2847.
* [9] Ginibre, J., and Velo, G. Smoothing properties and retarded estimates for some dispersive evolut ion equations. Comm. Math. Phys. 144, 1 (1992), 163–188.
* [10] Kadomtsev, B. B., and Petviashvili, V. I. On the Stability of Solitary Waves in Weakly Dispersing Media. Soviet Physics Doklady 15 (Dec. 1970), 539.
* [11] Kato, T. Quasi-linear equations of evolution, with applications to partial differential equations. In Spectral theory and differential equations (Proc. Sympos., Dundee, 1974; dedicated to Konrad Jörgens), vol. 448 of Lecture Notes in Math. Springer, Berlin, 1975, pp. 25–70.
* [12] Kato, T. On the Korteweg-de Vries equation. Manuscripta Math. 28, 1-3 (1979), 89–99.
* [13] Kato, T. Perturbation theory for linear operators. Classics in Mathematics. Springer-Verlag, Berlin, 1995. Reprint of the 1980 edition.
* [14] Kato, T., and Ponce, G. Commutator estimates and the Euler and Navier-Stokes equations. Comm. Pure Appl. Math. 41, 7 (1988), 891–907.
* [15] Kenig, C. E. On the local and global well-posedness theory for the KP-I equation. Ann. Inst. H. Poincaré Anal. Non Linéaire 21, 6 (2004), 827–838.
* [16] Kenig, C. E., Martel, Y., and Robbiano, L. Local well-posedness and blow-up in the energy space for a class of $L^{2}$ critical dispersion generalized Benjamin-Ono equations. Ann. Inst. H. Poincaré Anal. Non Linéaire 28, 6 (2011), 853–887.
* [17] Kenig, C. E., Ponce, G., and Vega, L. Well-posedness and scattering results for the generalized Korteweg-de Vries equation via the contraction principle. Comm. Pure Appl. Math. 46, 4 (1993), 527–620.
* [18] Kim, B. Three-dimensional solitary waves in dispersive wave systems. PhD thesis, Massachusetts Institute of Technology. Dept. of Mathematics., Cambridge, MA, 2006.
* [19] Klein, C., and Saut, J.-C. A numerical approach to blow-up issues for dispersive perturbations of Burgers’ equation. Phys. D 295/296 (2015), 46–65.
* [20] Kobayasi, K. On a theorem for linear evolution equations of hyperbolic type. J. Math. Soc. Japan 31, 4 (1979), 647–654.
* [21] Lannes, D. The water waves problem, vol. 188 of Mathematical Surveys and Monographs. American Mathematical Society, Providence, RI, 2013. Mathematical analysis and asymptotics.
* [22] Lannes, D., and Saut, J.-C. Remarks on the full dispersion Kadomtsev-Petviashvli equation. Kinet. Relat. Models 6, 4 (2013), 989–1009.
* [23] Linares, F., and Pastor, A. Well-posedness for the two-dimensional modified Zakharov-Kuznetsov equation. SIAM J. Math. Anal. 41, 4 (2009), 1323–1339.
* [24] Linares, F., Pilod, D., and Saut, J.-C. Well-posedness of strongly dispersive two-dimensional surface wave Boussinesq systems. SIAM J. Math. Anal. 44, 6 (2012), 4195–4221.
* [25] Linares, F., Pilod, D., and Saut, J.-C. Remarks on the orbital stability of ground state solutions of fKdV and related equations. Adv. Differential Equations 20, 9/10 (09 2015), 835–858.
* [26] Linares, F., Pilod, D., and Saut, J.-C. The Cauchy problem for the fractional Kadomtsev-Petviashvili equations. SIAM Journal on Mathematical Analysis 50, 3 (2018), 3172–3209.
* [27] Linares, F., and Ponce, G. Introduction to nonlinear dispersive equations. Universitext (1979). Springer, 2009.
* [28] Lizarazo Osorio, J. d. C. El problema de Cauchy de la clase de ecuaciones de dispersión generalizada de Benjamin-Ono bidimensionales. PhD thesis, Universidad Nacional de Colombia, Bogotá, Colombia, Junio 2018.
* [29] Molinet, L., Saut, J.-C., and Tzvetkov, N. Well-posedness and ill-posedness results for the Kadomtsev-Petviashvili-I equation. Duke Math. J. 115, 2 (2002), 353–384.
* [30] Ono, H. Algebraic solitary waves in stratified fluids. J. Phys. Soc. Japan 39, 4 (1975), 1082–1091.
* [31] Pego, R. L., and Weinstein, M. I. Asymptotic stability of solitary waves. Comm. Math. Phys. 164, 2 (1994), 305–349.
* [32] Reed, M., and Simon, B. Methods of modern mathematical physics. II. Fourier analysis, self-adjointness. Academic Press [Harcourt Brace Jovanovich Publishers], New York, 1975\.
* [33] Sánchez Salazar, F. El problema de Cauchy asociado a una ecuación del tipo rBO-ZK. PhD thesis, Universidad Nacional de Colombia, Junio 2015.
* [34] Saut, J.-C. Remarks on the generalized kadomtsev-petviashvili equations. Indiana University Mathematics Journal 42, 3 (1993), 1011–1026.
* [35] Stein, E. M., and Weiss, G. L. Introduction to Fourier Analysis on Euclidean Spaces. Princeton Mathematical Series, No. 32. Princeton University Press, Princeton, N.J., 1971.
* [36] Whitham, G. B. Variational methods and applications to water waves. Proceedings of the Royal Society of London A: Mathematical, Physical and Engineering Sciences 299, 1456 (1967), 6–25.
* [37] Zabusky, N., and Kruskal, M. Interaction of “Solitons” in a Collisionless Plasma and the Recurrence of Initial States. Phys. Rev. Lett. 15 (Aug 1965), 240–243.
* [38] Zakharov, V., and Kuznetsov, E. Three-dimensional solitons. Soviet Physics JETP 29 (01 1974), 594–597.
|
# Bounds on mutual information of mixture data for classification tasks
Yijun Ding James C. Wyant College of Optical Sciences
University of Arizona
Tucson, AZ, USA
<EMAIL_ADDRESS>Amit Ashok James C. Wyant College of Optical
Sciences
and Department of Electrical Engineering
University of Arizona
Tucson, AZ, USA
<EMAIL_ADDRESS>
###### Abstract
The data for many classification problems, such as pattern and speech
recognition, follow mixture distributions. To quantify the optimum performance
for classification tasks, the Shannon mutual information is a natural
information-theoretic metric, as it is directly related to the probability of
error. The mutual information between mixture data and the class label does
not have an analytical expression, nor any efficient computational algorithms.
We introduce a variational upper bound, a lower bound, and three estimators,
all employing pair-wise divergences between mixture components. We compare the
new bounds and estimators with Monte Carlo stochastic sampling and bounds
derived from entropy bounds. To conclude, we evaluate the performance of the
bounds and estimators through numerical simulations.
###### Index Terms:
Mixture distribution, classification, Shannon mutual information, bounds,
estimation, mixed-pair
## I Introduction
### I-A Motivation
We study the performance of classification tasks, where the goal is to infer
the class label $C$ from sample data $\bm{x}$. The Shannon mutual information
$\mathrm{I}(\bm{x};C)$ characterizes the reduction in the uncertainty of the
class label $C$ with the knowledge of data $\bm{x}$ and provides a way to
quantify the relevance of the data $\bm{x}$ with respect to the class label
$C$. As the mutual information is related to probability of classification
error ($\mathrm{P_{e}}$) through Fano’s inequality and other bounds [1, 2, 3],
it has been widely used for feature selection [4, 5], learning [6], and
quantifying task-specific information [7] for classification.
Statistical mixture distributions such as Poisson, Wishart or Gaussian
mixtures are frequently used in the fields of speech recognition [8], image
retrieval [9], system evaluation [10], compressive sensing [11], distributed
state estimation [12], hierarchical clustering [13] etc. As a practical
example, consider a scenario in which the data $\bm{x}$ is measured with a
noisy system, eg. Poisson noise in an photon-starved imaging system or
Gaussian noise in a thermometer reading. If the actual scene (or temperature)
has a class label, e.g. target present or not (the temperature is below
freezing or not), then the mutual information $\mathrm{I}(\bm{x};C)$ describes
with what confidence one can assign a class label $C$ to the noisy measurement
data $\bm{x}$.
The goal of this paper is to develop efficient methods to quantify the optimum
performance of classification tasks, when the distribution of the data
$\bm{x}$ for each given class label, $\mathrm{pr}(\bm{x}|C)$, follows a known
mixture distribution. As the mutual information $\mathrm{I}(\bm{x};C)$, which
is commonly used to quantify task-specific information, does not admit an
analytical expression for mixture data, we provide analytical expressions for
bounds and estimators of $\mathrm{I}(\bm{x};C)$.
### I-B Problem Statement and Contributions
We consider the data as a continuous random variable $\bm{x}$ and the class
label as a discrete random variable $C$, where $C$ can be any integer in
$[1,\Pi]$ and $\Pi$ is the number of classes. The bold symbol $\bm{x}$
emphasizes that $\bm{x}$ is a vector, which can be high-dimensional. We assume
that, when restricted to any of the classes, the conditional differential
entropy of $\bm{x}$ is well-defined, or in other words, $(\bm{x},C)$ is a good
mixed-pair vector [14]. The mutual information between the data $\bm{x}$ and
the class label $C$ can be defined as [15]
$\begin{split}\mathrm{I}(\bm{x};C)&=\mathrm{KL}(\,\mathrm{pr}(\bm{x},C)||\,\mathrm{Pr}(C)\cdot\mathrm{pr}(\bm{x})\,)\\\
&=\sum_{C}\int\mathrm{dx}\,\mathrm{pr}(\bm{x},C)\ln{\frac{\mathrm{pr}(\bm{x},C)}{\mathrm{Pr}(C)\cdot\mathrm{pr}(\bm{x})}}.\end{split}$
(1)
When $\mathrm{pr}(\bm{x})$ is a mixture distribution with $N$ components,
$\mathrm{pr}(\bm{x})=\sum_{i=1}^{N}w_{i}\,\mathrm{pr}_{i}(\bm{x}),$ (2)
where $w_{i}$ is the weight of component $i$ ($w_{i}\geq 0$ and
$\sum_{i}w_{i}=1$), and $\mathrm{pr}_{i}$ is the probability density of
component $i$. The conditional distribution of the data, when the class label
is $c$, also follows a mixture distribution.
In this work, we propose new bounds and estimators of the Shannon mutual
information between a mixture distribution $\bm{x}$ and its class label $C$.
We provide a lower bound, a variational upper bound and three estimators of
$\mathrm{I}(\bm{x};C)$, all based on pair-wise distances. We present closed-
form expressions for the bounds and estimators. Furthermore, we use numerical
simulations to compare the bounds and estimators to Monte Carlo (MC)
simulations and a set of bounds derived from entropy bounds.
### I-C Related works
Although estimation of conditional entropy and mutual information has been
extensively studied [16, 17, 18, 19], research has focused on purely discrete
or continuous data. Nair et al. [14] extended the definition of the joint
entropy to mixed-pairs, which consists of one discrete variable and one
continuous variable. Ross [20], Moon et al. [21] and Beknazaryan et al. [15]
provided methods for estimating mutual information from samples of mixed-pairs
based on nearest-neighbor or kernel estimator. Gao et al. [22] extended the
definition of mutual information to the case that each random variable can
have both discrete and continuous components through the Radon-Nikodym
derivative. Here our goal is to study mutual information for mixed-pairs,
where the data $\bm{x}$ is continuous and the class label $C$ is discrete.
When the underlying distribution of the data is unknown, the mutual
information can be approximated from samples with a number of density or
likelihood-ratio estimators based on binning [23, 24], kernel methods[25, 26,
27], k-nearest-neighbor (kNN) distances [28, 29], or approximated Gaussianity
(Edgeworth expansion [30]). To accommodate high dimensional data (such as
image and text) or large datasets, Gao et al. [31] improved the kNN estimator
with a local non-uniformity correction term; Jiao et al. [32] proposed a
minimax estimator of entropy that achieves the optimal sample complexity;
Belghazi et al. [33] presented a general purpose neural-network estimator;
Poole et al. [34] provided a thorough review and several new bounds on mutual
information that is capable to trade off bias for variance.
However, when the underlying data distribution is known, the exact computation
of mutual information is tractable only for a limited family of distributions
[35, 36]. The mutual information for mixture distributions has no known
closed-from expression [37, 38, 39]; hence MC sampling and numerical
integration are often employed as unbiased estimators. MC sampling of
sufficient accuracy is computationally intensive [40]. Numerical integration
is limited to low-dimensional problems [41]. To reduce the computational
requirement, deterministic approximations have been developed using merged
Gaussian [8, 42], component-wise Taylor-series expansion [43], unscented
transform [44] and pair-wise KL divergence between matched components [9]. The
merged Gaussian and unscented transform estimators are biased, while the
Taylor expansion method provides a trade-off between computational demands and
accuracy.
Two papers that have deeply inspired our work are [8] and [45]. Hersey et al.
[8] proposed a variational upper bound and an estimator of the KL divergence
between two Gaussian mixtures by pair-wise KL divergence. Hersey et al. [8]
has shown empirically that the variational upper bound and estimator perform
better than other deterministic approximations, such as merged Gaussian,
unscented transform and matched components. Kolchinsky et al. [45] has bounded
entropy of mixture distributions with pair-wise KL divergence and
Chernoff-$\alpha$ ($C_{\alpha}$) divergence and demonstrated through numerical
simulations that these bounds are tighter than other well-known existing
bounds, such as the kernel-density estimator [41, 46] and the expected-
likelihood-kernel estimator [47, 48, 49]. Our results are not obvious from
either paper, as the calculation of $\mathrm{I}(\bm{x};C)$ involves a
summation of multiple entropies or KL divergences. Instead of providing bounds
for each term (entropy) in the summation, we directly bound and estimate the
mutual information.
## II Main Results
In this section, we provide three estimators of $\mathrm{I}(\bm{x};C)$ and a
pair of lower and upper bounds. All bounds and estimators are based on pair-
wise KL divergence and $C_{\alpha}$ divergence. Furthermore, we provide proofs
of the lower and upper bounds. Before presenting our main results, we start
with a few definitions. The marginal distribution on the class label $C$ is
$\mathrm{Pr}(C=c)=P_{c}=\sum_{i\in\\{c\\}}w_{i}.$ (3)
Note that $\sum_{c=1}^{\Pi}P_{c}=1$ and $\\{c\\}$ is the set of the components
that have class label $C=c$. The conditional distribution of the data, when
the class label is $c$, is given by
$\mathrm{pr}(\bm{x}|c)=\sum_{i\in\\{c\\}}\frac{w_{i}}{P_{c}}\,\mathrm{pr}_{i}(\bm{x}).$
(4)
Expressing the marginal distribution in terms of the conditional distribution,
we have
$\mathrm{pr}(\bm{x})=\sum_{c}P_{c}\cdot\mathrm{pr}(\bm{x}|c).$ (5)
The joint distribution of the data and class label is
$\mathrm{pr}(\bm{x},c)=P_{c}\cdot\mathrm{pr}(\bm{x}|c)=\sum_{i\in\\{c\\}}w_{i}\mathrm{pr}_{i}(\bm{x})$
(6)
### II-A Pair-wise distances
The Kullback-Leibler (KL) divergence is defined as
$\mathrm{KL}(\mathrm{pr}_{i}||\mathrm{pr}_{j})=\int\mathrm{d\bm{x}}~{}\mathrm{pr}_{i}(\bm{x})\ln\frac{\mathrm{pr}_{i}(\bm{x})}{\mathrm{pr}_{j}(\bm{x})}.$
(7)
The $C_{\alpha}$ divergence [50] between the two distribution
$\mathrm{pr}_{i}(\bm{x})$ and $\mathrm{pr}_{j}(\bm{x})$ is defined as
$\mathrm{C_{\alpha}}(\mathrm{pr}_{i}||\mathrm{pr}_{j})=-\ln\int\mathrm{d\bm{x}}\,\mathrm{pr}_{i}^{\alpha}(\bm{x})\,\mathrm{pr}_{j}^{1-\alpha}(\bm{x}),$
(8)
for real-valued $\alpha\in[0,1]$. More specifically, when $\alpha=1/2$, the
Chernoff divergence is Bhattacharyaa distance.
### II-B Bounds and estimates of the mutual information
We adopt the convention that $\ln 0=0$ and $\ln(0/0)=0$. An exact expression
of the mutual information is
$\mathrm{I}(\bm{x};C)=\mathrm{H}(C)-\sum_{i=1}^{N}w_{i}\mathrm{E}_{\mathrm{pr}_{i}}\left[\ln{\frac{\sum_{j=1}^{N}w_{j}\mathrm{pr}_{j}}{\sum_{k\in\\{C_{i}\\}}w_{k}\mathrm{pr}_{k}}}\right],$
(9)
where $\\{C_{i}\\}$ is the set of component index that is in the same class
with component $i$ and
$\mathrm{E}_{\mathrm{pr}_{i}}[f]=\int\mathrm{dx}\,\mathrm{pr}_{i}(\bm{x})f(\bm{x})$
is the expectation of $f$ with respect to the probability density function
$\mathrm{pr}_{i}$.
Two approximations of $\mathrm{I}(\bm{x};C)$ are
$\begin{split}\mathrm{{\hat{I}}_{C_{\alpha}}}(\bm{x};C)&=\mathrm{H}(C)-\sum_{i=1}^{N}w_{i}\ln{\frac{\sum_{j=1}^{N}w_{j}e^{-\mathrm{C_{\alpha}}(\mathrm{pr}_{i}||\mathrm{pr}_{j})}}{\sum_{k\in\\{C_{i}\\}}w_{k}e^{-\mathrm{C_{\alpha}}(\mathrm{pr}_{i}||\mathrm{pr}_{k})}}},\\\
\mathrm{{\hat{I}}_{KL}}(\bm{x};C)&=\mathrm{H}(C)-\sum_{i=1}^{N}w_{i}\ln{\frac{\sum_{j=1}^{N}w_{j}e^{-\mathrm{KL}(\mathrm{pr}_{i}||\mathrm{pr}_{j})}}{\sum_{k\in\\{C_{i}\\}}w_{k}e^{-\mathrm{KL}(\mathrm{pr}_{i}||\mathrm{pr}_{k})}}}.\end{split}$
(10)
Another approximation of $\mathrm{I}(\bm{x};C)$ is
$\begin{split}&\mathrm{{\hat{I}}_{KL\&C_{\alpha}}}(\bm{x};C)=\mathrm{H}(C)-\sum_{i=1}^{N}w_{i}\ln{\frac{\sum_{j=1}^{N}w_{j}e^{-D_{ij}}}{\sum_{k\in\\{C_{i}\\}}w_{k}e^{-D_{ik}}}},\\\
&\text{where}\quad\frac{1}{D_{ij}}=\frac{1}{2}\left(\frac{1}{\mathrm{KL}(\mathrm{pr}_{i}||\mathrm{pr}_{j})}+\frac{1}{{\mathrm{C_{\alpha}}(\mathrm{pr}_{i}||\mathrm{pr}_{j})}}\right).\end{split}$
(11)
As $D$ is a function of both KL and $C_{\alpha}$ divergences, we denote this
estimator with the subscript ‘$KL\&C_{\alpha}$’.
A lower bound on $\mathrm{I}(\bm{x};C)$ based on pair-wise $C_{\alpha}$ is
$\begin{split}&\mathrm{{I}_{lb\\_{C_{\alpha}}}}=-\sum_{c=1}^{\Pi}P_{c}\ln\left[\sum_{c^{\prime}=1}^{\Pi}P_{c^{\prime}}\cdot\text{min}(1,Q_{cc^{\prime}})\right],\,\,where\\\
&Q_{cc^{\prime}}=\sum_{i\in\\{c\\}}\sum_{j\in\\{c^{\prime}\\}}\left(\frac{w_{i}}{P_{c}}\right)^{\alpha_{c}}\left(\frac{w_{j}}{P_{c^{\prime}}}\right)^{1-\alpha_{c}}e^{-C_{\alpha_{c}}(\mathrm{pr}_{i}||\mathrm{pr}_{j})},\\\
\end{split}$ (12)
and min$(\cdot)$ is the minimum value function.
A variational upper bound on $\mathrm{I}(\bm{x},C)$ based on pair-wise KL is
$\mathrm{{I}_{ub\\_KL}}=\mathrm{H}(C)-\sum_{m}\sum_{c=1}^{\Pi}\phi_{m,c}\ln\frac{\sum_{c^{\prime}=1}^{\Pi}\phi_{m,c^{\prime}}e^{-KL(\mathrm{pr}_{m,c},\mathrm{pr}_{m,c^{\prime}})}}{\phi_{m,c}},$
(13)
where $\phi_{m,c}$ are the variational parameters and detailed explanation of
the notations is given in Section II-D.
### II-C Proof of the lower bound
We propose a lower bound on $\mathrm{I}(\bm{x};C)$ based on pair-wise
$C_{\alpha}$ divergences. For ease of notation, we denote the conditional
distribution $\mathrm{pr}(\bm{x}|C=c)$ as $\mathrm{pr}_{c}$. We first make use
of a derivation from [45] and [51] to bound $\mathrm{I}(\bm{x};C)$ with class-
wise divergence $C_{\alpha}(\mathrm{pr}_{c},\mathrm{pr}_{c^{\prime}})$:
$\begin{split}\mathrm{I}(\bm{x};C)=&\sum_{c}P_{c}\int\mathrm{dx}\,\mathrm{pr}_{c}\cdot\ln{\frac{\mathrm{pr}_{c}}{\mathrm{pr}(\bm{x})}}\\\
=&-\sum_{c}P_{c}\int\mathrm{dx}\,\mathrm{pr}_{c}\cdot\ln{\frac{\sum_{c^{\prime}}P_{c^{\prime}}\mathrm{pr}_{c^{\prime}}^{1-\alpha_{c}}}{\mathrm{pr}_{c}^{1-\alpha_{c}}}}\\\
&-\sum_{c}P_{c}\int\mathrm{dx}\,\mathrm{pr}_{c}\cdot\ln\frac{\mathrm{pr}(\bm{x})}{\mathrm{pr}_{c}^{\alpha_{c}}\sum_{c^{\prime}}P_{c^{\prime}}\mathrm{pr}_{c^{\prime}}^{1-\alpha_{c}}}\\\
\geq&-\sum_{c}P_{c}\ln{\sum_{c^{\prime}}P_{c^{\prime}}\int\mathrm{dx}\,\mathrm{pr}_{c}^{\alpha_{c}}\cdot\mathrm{pr}_{c^{\prime}}^{1-\alpha_{c}}}\\\
&-\ln\int\mathrm{dx}\,\sum_{c}P_{c}\mathrm{pr}_{c}^{1-\alpha_{c}}\cdot\frac{\mathrm{pr}(\bm{x})}{\sum_{c^{\prime}}P_{c^{\prime}}\mathrm{pr}_{c^{\prime}}^{1-\alpha_{c}}}\\\
\\\
=&-\sum_{c}P_{c}\ln\left[\sum_{c^{\prime}}P_{c^{\prime}}e^{-C_{\alpha_{c}}(\mathrm{pr}_{c}||\mathrm{pr}_{c^{\prime}})}\right]\\\
\end{split}$ (14)
This inequality follows from Jensen’s inequality and the convexity of function
$\ln(x)$. The parameter $\alpha_{c}$, which is specific for a class $c$, can
be any value in $[0,1]$.
The class-wise $C_{\alpha}$ divergence has a minimum value of zero and the
minimum is achieved when the two class has the same distribution. Furthermore,
we can bound the $C_{\alpha}$ divergence through the subadditivity of the
function $f(x)=x^{\alpha}$ when $0\leq\alpha\leq 1$. In other words, as
$f(a+b)\leq f(a)+f(b)$ for $a\geq 0$ and $b\geq 0$, the $C_{\alpha}$
divergence between the conditional distributions $\mathrm{pr}_{c}$ and
$\mathrm{pr}_{c^{\prime}}$ can be bounded by:
$\begin{split}e&{}^{-C_{\alpha}(\mathrm{pr}_{c}||\mathrm{pr}_{c^{\prime}})}=\int\mathrm{dx}\,\left[\sum_{i\in\\{c\\}}\frac{w_{i}}{P_{c}}\mathrm{pr}_{i}\right]^{\alpha}\left[\sum_{j\in\\{c^{\prime}\\}}\frac{w_{j}}{P_{c^{\prime}}}\mathrm{pr}_{j}\right]^{1-\alpha}\\\
&\leq\text{min}\left[1,\sum_{i\in\\{c\\}}\sum_{j\in\\{c^{\prime}\\}}\left(\frac{w_{i}}{P_{c}}\right)^{\alpha}\left(\frac{w_{j}}{P_{c^{\prime}}}\right)^{1-\alpha}e^{-C_{\alpha}(\mathrm{pr}_{i}||\mathrm{pr}_{j})}\right],\\\
&=\text{min}(1,Q_{cc^{\prime}}).\\\ \end{split}$ (15)
Therefore, Equation 12 is a lower bound on $\mathrm{I}(\bm{x};C)$.
The best possible lower bound can be obtained by finding the parameters
$\alpha_{c}$ that maximize $\mathrm{{I}_{lb\\_{C_{\alpha}}}}$, which is
equivalent to minimize
$\sum_{c^{\prime}}P_{c^{\prime}}\text{min}(1,Q_{cc^{\prime}})$. In a special
case when all components are symmetric and identical except the center
location, eg. homoscedastic Gaussian mixture,
$e^{-C_{\alpha}(\mathrm{pr}_{i}||\mathrm{pr}_{j})}$ achieves minimum value at
$\alpha=1/2$ [45].
### II-D Proof of the variational upper bound
Here we propose a direct upper bound on the mutual information
$\mathrm{I}(\bm{x};C)$ using a variational approach. The underlying idea is to
match components from different classes. To pick one component from each
class, there are $N_{1}\times N_{2}...\times N_{\Pi}$ combinations, where
$N_{c}$ is the number of components in class $c$. Denote an integer
$M=\prod_{c=1}^{\Pi}N_{c}$. A component $i$ in class $c$ can be split into
$M/N_{c}$ components with each component corresponding to a component-
combination in the other $\Pi-1$ classes. Mathematically speaking, we
introduce the variational parameters $\phi_{ij}\geq 0$ satisfying the
constraints $\sum_{j=1}^{M/N_{c}}\phi_{ij}=w_{i}$. Using the variational
parameters, we can write the joint distribution as
$\begin{split}\mathrm{pr}(\bm{x},c)=\sum_{i\in\\{c\\}}w_{i}\mathrm{pr}_{i}=\sum_{i\in\\{c\\}}\sum_{j=1}^{M/N_{c}}\phi_{ij}\mathrm{pr}_{i}.\end{split}$
(16)
Note that the set $\\{c\\}$ has $N_{c}$ components. By rearranging indices
$(i,j)$ into a vector $m$ of length $M$, we can simplify the joint
distribution to
$\mathrm{pr}(\bm{x},c)=\sum_{m=1}^{M}\phi_{m,c}\mathrm{pr}_{m,c}(\bm{x}),$
(17)
where the subscript $c$ emphasizes that each class has a unique mapping from
$(i,j)$ to $m$ and $\mathrm{pr}_{m,c}(\bm{x})$ equals to the corresponding
$\mathrm{pr}_{i}(\bm{x})$.
With this notation, the marginal distribution of the data $\bm{x}$ is
$\mathrm{pr}(\bm{x})=\sum_{c=1}^{\Pi}\mathrm{pr}(\bm{x},c)=\sum_{c=1}^{\Pi}\sum_{m=1}^{M}\phi_{m,c}\mathrm{pr}_{m,c}(\bm{x}).$
(18)
We further define a mini-batch $m$ as
$b_{m}(\bm{x})=\sum_{c=1}^{\Pi}\phi_{m,c}\mathrm{pr}_{m,c}(\bm{x}).$ (19)
Each mini-batch contains $\Pi$ components with one component from each class.
With this definition, the marginal distribution of the data can be written as
$\mathrm{pr}(\bm{x})=\sum_{m=1}^{M}b_{m}(\bm{x})$. The probability of a
component in the $m^{th}$ batch is $P_{m}=\sum_{c}\phi_{m,c}$. The probability
density function of the $m^{th}$ batch is
$\mathrm{pr}(\bm{x}|m)=b_{m}(\bm{x})/P_{m}$.
Now we use Jensen’s inequality, or more specificly log-sum inequality [52], to
bound $\mathrm{I}(\bm{x};C)$ by batch-conditional entropy,
$\begin{split}&\mathrm{I}(\bm{x};C)=\mathrm{H}(C)+\sum_{c}\int\mathrm{dx}\,\mathrm{pr}(\bm{x},c)\cdot\ln{\frac{\mathrm{pr}(\bm{x},c)}{\mathrm{pr}(\bm{x})}}\\\
&=\mathrm{H}(C)+\sum_{c}\int\mathrm{dx}\left(\sum_{m}\phi_{m,c}\mathrm{pr}_{m,c}\right)\,\ln\frac{\sum_{m}\phi_{m,c}\mathrm{pr}_{m,c}}{\sum_{m}b_{m}}\\\
&\leq\mathrm{H}(C)+\sum_{c}\int\mathrm{dx}\sum_{m}\left(\phi_{m,c}\mathrm{pr}_{m,c}\ln\frac{\phi_{m,c}\mathrm{pr}_{m,c}}{b_{m}}\right)\\\
&=\mathrm{H}(C)+\mathrm{H}(\bm{x}|m)-\mathrm{H}(C|m)-\sum_{i=1}^{N}w_{i}\mathrm{H}_{i}(\bm{x})\\\
\end{split}$ (20)
where $\mathrm{H}(C)=-\sum_{c}P_{c}\ln P_{c}$ is the entropy of the class
label; $\mathrm{H}(\bm{x}|m)=\sum_{m}P_{m}\mathrm{H}(\mathrm{pr}(\bm{x}|m))$
is the batch-conditional entropy of the data;
$\mathrm{H}(C|m)=\sum_{m}P_{m}\mathrm{H}_{m}(C)$ is the batch-conditional
entropy of the label, where $\mathrm{H}_{m}(C)$ is the entropy of the class
label for batch $m$, and
$\mathrm{H}_{i}(\bm{x})=\mathrm{H}(\mathrm{pr}_{i}(\bm{x}))$ is the entropy of
the $i^{th}$ component.
We can further bound the batch-conditional entropy with pair-wise KL
divergence as
$\begin{split}&\mathrm{I}(\bm{x};C)\leq\mathrm{H}(C)+\hat{\mathrm{H}}_{KL}(\bm{x}|m)-\mathrm{H}(C|m)-\sum_{i=1}^{N}w_{i}\mathrm{H}_{i}(\bm{x})\\\
&=\mathrm{H}(C)-\sum_{m}\sum_{c}\phi_{m,c}\ln\frac{\sum_{c^{\prime}}\phi_{m,c^{\prime}}e^{-KL(\mathrm{pr}_{m,c},\mathrm{pr}_{m,c^{\prime}})}}{\phi_{m,c}}\\\
&:=\mathrm{{I}_{ub\\_{KL}}}\end{split}$ (21)
where $\hat{\mathrm{H}}_{KL}(\bm{x}|m)$ is an upper bound of the batch-
conditional entropy and the inequality has been proved in [45].
The tightest upper bound attainable through this method can be found by
varying parameters $\phi_{m,c}$ to minimize $\mathrm{{I}_{ub\\_{KL}}}$. The
minimization problem has been proved to be convex (see Appendix A). The upper
bound $\mathrm{{I}_{ub\\_{KL}}}$ can be minimized iteratively by fixing the
parameters $\phi_{m,c^{\prime}}$ (where $c^{\prime}\neq c$) and optimizing
parameters $\phi_{m,c}$ under linear constraints. At each iteration step
$\mathrm{{I}_{ub\\_{KL}}}$ is lowered, and the convergent is the tightest
variational upper bound on the mutual information.
Non-optimum variational parameters still provide upper bounds on
$\mathrm{I}(\bm{x};C)$. There are $M\times\Pi$ variational parameters,
$M\times\Pi$ non-equality constraints and $N$ equality constraints. When the
number of classes or components is large, the minimization problem will be
computationally intensive. A non-optimum solution that is similar to the
matched bound [8, 53] can be obtained by dividing all components into
max($N_{c}$) mini-batches by matching each component $i$ to one component in
each class. Mathematically speaking, $\phi_{ij}=w_{i}$ for one pair of matched
$(i,j)$ and $\phi_{ij}=0$ otherwise. To find the mini-batches, the Hungarian
method [54, 55] for assignment problems can be applied.
## III Numerical simulations
(a)
(b)
Figure 1: (a) The locations of the center of the components and the mixture
distribution $\mathrm{pr}(\bm{x}|c=2)$ when $\sigma=0.5$ (insert). (b)
Estimates of $\mathrm{I}(\bm{x};C)$.
In this section, we run numerical simulations and compare estimators on mutual
information between mixture data and class labels. We consider a simple
example of binary classification of mixture data, where the mixture components
are two-dimensional homoscedastic Gaussians. The component centers are close
to the class boundary and uniformly distributed along the boundary. The
location of the component centers are plotted in the Figure 1(a), where the
component centers are represented by a red star (class 1) or a yellow circle
(class 2). Each class consists 100 two-dimensional Gaussian components with
equal weights. The components have the same covariance matrix $\sigma^{2}I$,
where $I$ is the identity matrix and $\sigma$ represents the size of the
Gaussian components. The conditional distribution $\mathrm{pr}(\bm{x}|c=2)$ is
plotted in the insert of Figure 1(a) for $\sigma=0.5$. When $\sigma$ is
larger, the components of the mixtures distribution are more connected; when
$\sigma$ is smaller, the components are more isolated. Estimates of
$\mathrm{I}(\bm{x};C)$ are calculated for varying $\sigma$.
A pair of obvious bounds of $\mathrm{I}(\bm{x};C)$ are $[0,\mathrm{H}(C)]$,
where $\mathrm{H}(C)$ is the entropy of the class label $\mathrm{Pr}(C)$.
Another pair of upper and lower bounds of $\mathrm{I}(\bm{x};C)$ can be
derived from bounds on mixture entropy as
$\begin{split}\mathrm{{I}_{lb\\_2H}}&=\mathrm{H}_{lb}(\bm{x})-\mathrm{H}_{ub}(\bm{x}|C)\\\
\mathrm{{I}_{ub\\_2H}}&=\mathrm{H}_{ub}(\bm{x})-\mathrm{H}_{lb}(\bm{x}|C),\end{split}$
(22)
where the upper and lower bound of entropy based on pair-wise KL and
$C_{\alpha}$ divergences have been provided by [45]. These bounds on
$\mathrm{I}(\bm{x},C)$ are based on two entropy bounds, hence the subscript
‘2H’.
We evaluate the following estimates of $\mathrm{I}(\bm{x},C)$:
1. 1.
The new variational upper bound and the new lower bound,
$\mathrm{{I}_{ub\\_{KL}}}$ and $\mathrm{{I}_{lb\\_{C_{\alpha}}}}$, are plotted
in dark red and blue solid lines, respectively.
2. 2.
The estimates based on the pair-wise KL, $C_{\alpha}$ or D (a function of both
KL and $C_{\alpha}$ divergences) are plotted in yellow, light blue and black
dashed lines, respectively.
3. 3.
The true mutual information, $\mathrm{I}(\bm{x},C)$, as estimated by MC
sampling of the mixture model (grey solid line).
4. 4.
The upper and lower bounds $\mathrm{{I}_{lb\\_2H}}$ and
$\mathrm{{I}_{ub\\_2H}}$ are plotted in orange and green dot-dashed lines,
respectively.
The obvious bounds on $\mathrm{I}(\bm{x},C)$, which are $[0,H(C)]$, are also
presented by an area in grey. The Monte-Carlo simulation results, which can
serve as the benchmark, are calculated with $10^{6}$ samples. We use
$\alpha=1/2$ in the calculation of $C_{\alpha}$ divergences, as it provides
the optimum bounds for our example. We also present details of our
implementation and results of two other scenarios in Appendix B.
Our new upper bound and lower bound appear to be tighter than the bounds
derived from entropy bounds over the range of $\sigma$ considered in our
simulation. In Figure 1(b), where the estimates of $\mathrm{I}(\bm{x},C)$ are
plotted, the results show that the blue and dark red solid lines are almost
always within the area covered by the green and orange dot-dashed lines. The
three estimates, $\mathrm{{\hat{I}}_{KL}}$, $\mathrm{{\hat{I}}_{C_{\alpha}}}$
and $\mathrm{{\hat{I}}_{KL\&C_{\alpha}}}$, all follow the trend of
$\mathrm{I}(\bm{x},C)$. More specifically, $\mathrm{{\hat{I}}_{KL}}$ (yellow
dashed line) follows the new variational upper bound
$\mathrm{{I}_{ub\\_{KL}}}$ (deep red solid line) closely;
$\mathrm{{\hat{I}}_{C_{\alpha}}}$ (blue dashed line) is a good estimator of
$\mathrm{I}(\bm{x},C)$ (grey solid line);
$\mathrm{{\hat{I}}_{KL\&C_{\alpha}}}$ (black dashed line) is another good
estimator of $\mathrm{I}(\bm{x},C)$, as the black dashed line tracks the grey
solid line closely. The differences between the three estimates and the
$\mathrm{I}(\bm{x},C)$ calculated from MC simulation are plotted in Appendix
B.
## IV Conclusion
We provide closed-form bounds and approximations of mutual information between
mixture data and class labels. The closed-form expressions are based on pair-
wise distances, which are feasible to compute even for high-dimensional data.
Based on numerical results, the new bounds we proposed are tighter than the
bounds derived from bounds on entropy and the approximations serve as good
surrogates for the true mutual information.
## Appendix A The Minimization of $\mathrm{{I}_{ub\\_KL}}$
The minimization problem of $\mathrm{{I}_{ub\\_KL}}$ by varying $\phi_{mc}$ is
convex, which we prove in this section. The convexity of the minimization
problem can be checked through the first and second-order derivatives. For
ease of notation, we define
$S_{m,c}=\sum_{c^{\prime}}\phi_{m,c^{\prime}}e^{-\mathrm{KL}(\mathrm{pr}_{m,c},\mathrm{pr}_{m,c^{\prime}})}$
and
$E_{m,cc^{\prime}}=\exp({-\mathrm{KL}(\mathrm{pr}_{m,c},\mathrm{pr}_{m,c^{\prime}})})$.
The first derivative of $\mathrm{{I}_{ub\\_{KL}}}$ is:
$\frac{\partial{I}_{ub\\_{KL}}}{\partial\phi_{m,c}}=-\ln\left(\frac{S_{m,c}}{\phi_{m,c}}\right)-\frac{\phi_{m,c}}{S_{m,c}}-\sum_{c^{\prime}\neq
c}\frac{\phi_{m,c^{\prime}}E_{m,c^{\prime}c}}{S_{m,c^{\prime}}}+1$ (23)
The second derivative is:
$\begin{split}H_{cc}&=\frac{\partial^{2}{I}_{ub\\_{KL}}}{(\partial\phi_{m,c})^{2}}\\\
&=\frac{(S_{m,c}-\phi_{m,c})^{2}}{(S_{m,c})^{2}\phi_{m,c}}+\sum_{c^{\prime}\neq
c}\frac{\phi_{m,c^{\prime}}(E_{m,c^{\prime}c})^{2}}{(S_{m,c^{\prime}})^{2}}\\\
\end{split}$ (24)
for the diagonal terms and
$\begin{split}H_{cc^{\prime}}&=\frac{\partial^{2}{I}_{ub\\_{KL}}}{\partial\phi_{m,c}\partial\phi_{m,c^{\prime}}}\\\
&=\frac{\phi_{m,c}-S_{m,c}}{(S_{m,c})^{2}}E_{m,cc^{\prime}}+\frac{\phi_{m,c^{\prime}}-S_{m,c^{\prime}}}{(S_{m,c^{\prime}})^{2}}E_{m,c^{\prime}c},\\\
\end{split}$ (25)
for $c^{\prime}\neq c$. For any given vector $\bm{\theta}$ of length $\Pi$,
$\begin{split}\bm{\theta}^{T}H\bm{\theta}=&\sum_{c}\left(\theta_{c}^{2}H_{cc}+\sum_{c^{\prime}\neq
c}\theta_{c}\theta_{c^{\prime}}H_{cc^{\prime}}\right)\\\
=&\sum_{c}\Bigg{[}\theta_{c}^{2}\frac{(S_{m,c}-\phi_{m,c})^{2}}{(S_{m,c})^{2}\phi_{m,c}}+\sum_{c^{\prime}\neq
c}\theta_{c^{\prime}}^{2}\frac{\phi_{m,c}(E_{m,cc^{\prime}})^{2}}{(S_{m,c})^{2}}\\\
&\quad\quad+\sum_{c^{\prime}\neq
c}2\theta_{c}\theta_{c^{\prime}}\frac{\phi_{m,c}-S_{m,c}}{(S_{m,c})^{2}}E_{m,cc^{\prime}}\Bigg{]}\\\
=&\sum_{c}\left[\frac{(S_{m,c}-\phi_{m,c})\theta_{c}}{S_{m,c}\sqrt{\phi_{m,c}}}-\sum_{c^{\prime}\neq
c}\frac{\sqrt{\phi_{m,c}}E_{m,cc^{\prime}}\theta_{c^{\prime}}}{S_{m,c}}\right]^{2}\\\
\geq&0.\end{split}$ (26)
Therefore, $\mathrm{{I}_{ub\\_{KL}}}$ is convex when $\phi_{m,c}$ are
considered as the variables.
## Appendix B On the Numerical simulations
This appendix provides more simulation results and the detailed expressions
used in the numerical simulations. The additional results consider two
different distributions of the component-center locations. We further present
the difference between the the estimated and the true mutual information. Last
but not least, the closed form expressions include the KL and $C_{\alpha}$
divergences between Gaussian components, the bounds on $\mathrm{I}(\bm{x};C)$
derived from entropy bounds, and the relation between Shannon mutual
information and bounds on binary classification error ($\mathrm{P_{e}}$).
### B-A Numerical simulation results
We consider three scenarios: (1) the component centers are uniformly
distributed along the class boundary, (2) the component centers are bunched
into one group, and (3) the component centers are bunched into several groups.
Results on the first scenario has been presented in the Section III. In this
section, we report on Scenarios 2 and 3. Illustration of the two scenarios are
shown in Figure 2(a) and 3(a), respectively.
(a)
(b)
Figure 2: Scenario 2, where the center of the components are bunched into one
group, illustration (a), the mixture distribution $\mathrm{pr}(\bm{x}|c=2)$
when $\sigma=0.5$ (insert), and estimates of $\mathrm{I}(\bm{x};C)$ (b).
(a)
(b)
Figure 3: Scenario 3, where the center of the components are bunched into
multiple groups, illustration (a), the mixture distribution
$\mathrm{pr}(\bm{x}|c=2)$ when $\sigma=0.5$ (insert), and estimators of
$\mathrm{I}(\bm{x};C)$ (b).
The results demonstrated in these two scenarios are similar to that of
Scenario 1. To further demonstrate that our estimators are good surrogates for
the true mutual information, we plot the difference between the three
estimates and the true mutual information calculated from MC sampling in
Figure 4.
(a)
(b)
(c)
Figure 4: $\mathrm{\hat{I}(\bm{x};C)-I(\bm{x};C)}$ for (a) Scenario 1, (b)
Scenario 2 and (c) Scenario 3. The $\mathrm{I(\bm{;}C)}$ is calculated from MC
simulations.
### B-B Closed form expressions for Gaussian mixtures
Gaussian functions are often used as components in mixture distributions and
have closed form expressions for pair-wise KL and $C_{\alpha}$ divergences.
Denoting the difference in the means of two components as
$\bm{\mu}_{ij}=\bm{\mu}_{i}-\bm{\mu}_{j}$ and
$\Sigma_{\alpha,ij}=(1-\alpha)\Sigma_{i}+\alpha\Sigma_{j}$, the $C_{\alpha}$
divergence between two Gaussian components are
$\begin{split}&\mathrm{C_{\alpha}}(\mathrm{pr}_{i}||\mathrm{pr}_{j})=\frac{\alpha(1-\alpha)}{2}{\bm{\mu}^{T}_{ij}}\Sigma_{\alpha,ij}^{-1}\bm{\mu}_{ij}+\frac{1}{2}\ln\frac{|\Sigma_{\alpha,ij}|}{|\Sigma_{i}|^{{1-\alpha}}|\Sigma_{j}|^{{\alpha}}},\end{split}$
(27)
where $|\cdot|$ is the determinant. The KL divergence between the same two
Gaussian components are
$\mathrm{KL}(\mathrm{pr}_{i}||\mathrm{pr}_{j})=\frac{1}{2}\left[{\bm{\mu}^{T}_{ij}}\Sigma_{j}^{-1}\bm{\mu}_{ij}+\ln\frac{|\Sigma_{j}|}{|\Sigma_{i}|}\right]+\frac{\text{tr}(\Sigma_{j}^{-1}\Sigma_{i})-d}{2},$
(28)
where tr$(\cdot)$ is the trace of the matrix in the parenthesis and $d$ is the
dimension of the data.
When all mixture components have equal covariance matrices
$\Sigma_{i}=\Sigma_{j}=\Sigma$, we can denote
$\lambda_{ij}={\bm{\mu}^{T}_{ij}}\Sigma^{-1}\bm{\mu}_{ij}$ and have
$\begin{split}\mathrm{C_{\alpha}}(\mathrm{pr}_{i}||\mathrm{pr}_{j})&=\alpha(1-\alpha)\lambda_{ij}/2,\\\
\mathrm{KL}(\mathrm{pr}_{i}||\mathrm{pr}_{j})&=\lambda_{ij}/2.\end{split}$
(29)
With these expressions, the bounds and estimates of $\mathrm{I}(\bm{x};C)$
have simple forms.
### B-C Expressions of $\mathrm{{I}_{lb\\_2H}}$ and $\mathrm{{I}_{ub\\_2H}}$
The detailed expression for the mutual information bounds derived from entropy
bounds are:
$\begin{split}\mathrm{{I}_{lb\\_2H}}&=\mathrm{H}(C)-\sum_{i=1}^{N}w_{i}\ln{\frac{\sum_{j=1}^{N}w_{j}e^{-\mathrm{C_{\alpha}}(\mathrm{pr}_{i}||\mathrm{pr}_{j})}}{\sum_{k\in\\{C_{i}\\}}w_{k}e^{-\mathrm{KL}(\mathrm{pr}_{i}||\mathrm{pr}_{k})}}},\\\
\mathrm{{I}_{ub\\_2H}}&=\mathrm{H}(C)-\sum_{i=1}^{N}w_{i}\ln{\frac{\sum_{j=1}^{N}w_{j}e^{-\mathrm{KL}(\mathrm{pr}_{i}||\mathrm{pr}_{j})}}{\sum_{k\in\\{C_{i}\\}}w_{k}e^{-\mathrm{C_{\alpha}}(\mathrm{pr}_{i}||\mathrm{pr}_{k})}}}.\end{split}$
(30)
### B-D Bounds on $\mathrm{P_{e}}$ for binary classification
Bounds on $\mathrm{I}(\bm{x};C)$ can be used to calculate bounds on
$\mathrm{P_{e}}$. The Fano’s inequality [1] provides a lower bound on
$\mathrm{P_{e}}$ for binary classification, as following
$\mathrm{P_{e}}\geq h_{b}^{-1}[\mathrm{H}(C)-\mathrm{I}(\mathbf{x};C)],$ (31)
where $h_{b}(x)=-x\log_{2}(x)-(1-x)\log_{2}(1-x)$ is the binary entropy
function, $h_{b}^{-1}(\cdot)$ is the inverse function of $h_{b}(\cdot)$. More
specifically, one can calculate $P_{e}$ by placing the value
$\mathrm{H}(C)-\mathrm{I}(\mathbf{x};C)$ on the left side of the binary
entropy function and solving for $x$.
A tight upper bound on binary classification error $\mathrm{P_{e}}$ has been
reported recently [3],
$\mathrm{P_{e}}\leq{\text{min}}\left\\{P_{min},\,f^{-1}[\mathrm{H}(C)-\mathrm{I}(\mathbf{x};C)]\right\\}:=\mathrm{\hat{P}_{e\\_ub}},$
(32)
where $P_{min}$ is min$\\{P_{1},P_{2}\\}$, and $f(x)$ is a function defined by
$f(x)=-P_{min}\log_{2}\frac{P_{min}}{x+P_{min}}-x\log_{2}\frac{x}{x+P_{min}},$
(33)
and $f^{-1}(\cdot)$ is the inverse function of $f(\cdot)$.
When $\mathrm{P_{e}}\ll 1$,
$-\mathrm{P_{e}}(\log{\mathrm{P_{e}}}-\log{P_{min}})\lesssim\mathrm{H}(C)-\mathrm{I}(\mathbf{x};C)\lesssim-\mathrm{P_{e}}\log{\mathrm{P_{e}}}$.
Therefore, $\mathrm{P_{e}}$ is on the same order of magnitude as
$\mathrm{H}(C)-\mathrm{I}(\mathbf{x};C)$, when $\mathrm{P_{e}}\ll 1$.
### B-E $\mathrm{P_{e}}$ estimates
When $\sigma$ is small, all estimates of $\mathrm{I}(\bm{x},C)$ converges to
1. To compare the estimates for $0.1>\sigma>0.01$, we calculate and present an
estimate of $\mathrm{P_{e}}$ in this section. This estimate of
$\mathrm{P_{e}}$ is the upper bound on $\mathrm{P_{e}}$ presented in the
previous section. When a lower bound of $\mathrm{I}(\bm{x},C)$ is used in the
calculation (blue solid line and green dashed line), the estimates of
$\mathrm{P_{e}}$ are upper bounds on $\mathrm{P_{e}}$. The other lines in the
$\mathrm{P_{e}}$ plots are neither upper bound nor lower bound. The black
lines in the plots, which is the estimate of $\mathrm{P_{e}}$ calculated from
$\mathrm{I}_{MC}$, is also not the true $\mathrm{P_{e}}$ but can serve as a
benchmark for an upper bound on $\mathrm{P_{e}}$.
(a)
(b)
(c)
Figure 5: Estimates of $\mathrm{P_{e}}$ calculated from a number of estimators
of $\mathrm{I(\bm{x};C)}$.
In Figure 5, the blue and dark red solid lines are significantly closer to the
grey line than the green and orange dashed lines. The results demonstrated
that our new bounds is tighter than the bounds derived from entropy bounds.
Furthermore, the three estimators (blue, black and yellow dashed lines) are
good surrogates for the true mutual information.
## References
* [1] R. M. Fano and D. Hawkins, “Transmission of information: A statistical theory of communications,” _American Journal of Physics_ , vol. 29, pp. 793–794, 1961.
* [2] V. Kovalevskij, “The problem of character recognition from the point of view of mathematical statistics,” 1967.
* [3] B.-G. Hu and H.-J. Xing, “An optimization approach of deriving bounds between entropy and error from joint distribution: Case study for binary classifications,” _Entropy_ , vol. 18, no. 2, p. 59, 2016.
* [4] J. R. Vergara and P. A. Estévez, “A review of feature selection methods based on mutual information,” _Neural computing and applications_ , vol. 24, no. 1, pp. 175–186, 2014.
* [5] R. Battiti, “Using mutual information for selecting features in supervised neural net learning,” _IEEE Transactions on neural networks_ , vol. 5, no. 4, pp. 537–550, 1994.
* [6] N. Tishby, F. C. Pereira, and W. Bialek, “The information bottleneck method,” _arXiv preprint physics/0004057_ , 2000.
* [7] M. A. Neifeld, A. Ashok, and P. K. Baheti, “Task-specific information for imaging system analysis,” _JOSA A_ , vol. 24, no. 12, pp. B25–B41, 2007\.
* [8] J. R. Hershey and P. A. Olsen, “Approximating the Kullback Leibler divergence between Gaussian mixture models,” in _2007 IEEE International Conference on Acoustics, Speech and Signal Processing-ICASSP’07_ , vol. 4. IEEE, 2007, pp. IV–317.
* [9] J. Goldberger, S. Gordon, and H. Greenspan, “An efficient image similarity measure based on approximations of kl-divergence between two Gaussian mixtures,” in _null_. IEEE, 2003, p. 487.
* [10] Y. Ding and A. Ashok, “X-ray measurement model incorporating energy-correlated material variability and its application in information-theoretic system analysis,” _arXiv preprint arXiv:2002.11046_ , 2020.
* [11] J. M. Duarte-Carvajalino, G. Yu, L. Carin, and G. Sapiro, “Task-driven adaptive statistical compressive sensing of Gaussian mixture models,” _IEEE Transactions on Signal Processing_ , vol. 61, no. 3, pp. 585–600, 2012\.
* [12] B. Noack, M. Reinhardt, and U. D. Hanebeck, “On nonlinear track-to-track fusion with Gaussian mixtures,” in _17th International Conference on Information Fusion (FUSION)_. IEEE, 2014, pp. 1–8.
* [13] J. Goldberger and S. T. Roweis, “Hierarchical clustering of a mixture model,” in _Advances in Neural Information Processing Systems_ , 2005, pp. 505–512.
* [14] C. Nair, B. Prabhakar, and D. Shah, “On entropy for mixtures of discrete and continuous variables,” _arXiv preprint cs/0607075_ , 2006.
* [15] A. Beknazaryan, X. Dang, and H. Sang, “On mutual information estimation for mixed-pair random variables,” _Statistics & Probability Letters_, vol. 148, pp. 9–16, 2019.
* [16] L. Kozachenko and N. N. Leonenko, “Sample estimate of the entropy of a random vector,” _Problemy Peredachi Informatsii_ , vol. 23, no. 2, pp. 9–16, 1987\.
* [17] I. Ahmad and P.-E. Lin, “A nonparametric estimation of the entropy for absolutely continuous distributions (corresp.),” _IEEE Transactions on Information Theory_ , vol. 22, no. 3, pp. 372–375, 1976.
* [18] B. Laurent _et al._ , “Efficient estimation of integral functionals of a density,” _The Annals of Statistics_ , vol. 24, no. 2, pp. 659–681, 1996\.
* [19] G. P. Basharin, “On a statistical estimate for the entropy of a sequence of independent random variables,” _Theory of Probability & Its Applications_, vol. 4, no. 3, pp. 333–336, 1959.
* [20] B. C. Ross, “Mutual information between discrete and continuous data sets,” _PloS one_ , vol. 9, no. 2, p. e87357, 2014.
* [21] K. R. Moon, K. Sricharan, and A. O. Hero, “Ensemble estimation of mutual information,” in _2017 IEEE International Symposium on Information Theory (ISIT)_. IEEE, 2017, pp. 3030–3034.
* [22] W. Gao, S. Kannan, S. Oh, and P. Viswanath, “Estimating mutual information for discrete-continuous mixtures,” in _Advances in neural information processing systems_ , 2017, pp. 5986–5997.
* [23] G. A. Darbellay and I. Vajda, “Estimation of the information by an adaptive partitioning of the observation space,” _IEEE Transactions on Information Theory_ , vol. 45, no. 4, pp. 1315–1321, 1999.
* [24] R. Moddemeijer, “On estimation of entropy and mutual information of continuous distributions,” _Signal processing_ , vol. 16, no. 3, pp. 233–248, 1989\.
* [25] A. M. Fraser and H. L. Swinney, “Independent coordinates for strange attractors from mutual information,” _Physical review A_ , vol. 33, no. 2, p. 1134, 1986.
* [26] Y.-I. Moon, B. Rajagopalan, and U. Lall, “Estimation of mutual information using kernel density estimators,” _Physical Review E_ , vol. 52, no. 3, p. 2318, 1995.
* [27] K. Kandasamy, A. Krishnamurthy, B. Poczos, L. Wasserman _et al._ , “Nonparametric von mises estimators for entropies, divergences and mutual informations,” _Advances in Neural Information Processing Systems_ , vol. 28, pp. 397–405, 2015.
* [28] A. Kraskov, H. Stögbauer, and P. Grassberger, “Estimating mutual information,” _Physical review E_ , vol. 69, no. 6, p. 066138, 2004.
* [29] S. Singh and B. Póczos, “Finite-sample analysis of fixed-k nearest neighbor density functional estimators,” _Advances in neural information processing systems_ , vol. 29, pp. 1217–1225, 2016.
* [30] M. M. V. Hulle, “Edgeworth approximation of multivariate differential entropy,” _Neural computation_ , vol. 17, no. 9, pp. 1903–1910, 2005.
* [31] S. Gao, G. Ver Steeg, and A. Galstyan, “Efficient estimation of mutual information for strongly dependent variables,” in _Artificial intelligence and statistics_ , 2015, pp. 277–286.
* [32] J. Jiao, K. Venkat, Y. Han, and T. Weissman, “Minimax estimation of functionals of discrete distributions,” _IEEE Transactions on Information Theory_ , vol. 61, no. 5, pp. 2835–2885, 2015.
* [33] M. I. Belghazi, A. Baratin, S. Rajeshwar, S. Ozair, Y. Bengio, A. Courville, and D. Hjelm, “Mutual information neural estimation,” in _International Conference on Machine Learning_ , 2018, pp. 531–540.
* [34] B. Poole, S. Ozair, A. v. d. Oord, A. A. Alemi, and G. Tucker, “On variational bounds of mutual information,” _arXiv preprint arXiv:1905.06922_ , 2019.
* [35] J. V. Michalowicz, J. M. Nichols, and F. Bucholtz, _Handbook of differential entropy_. Crc Press, 2013\.
* [36] F. Nielsen and R. Nock, “Entropies and cross-entropies of exponential families,” in _2010 IEEE International Conference on Image Processing_. IEEE, 2010, pp. 3621–3624.
* [37] M. A. Carreira-Perpinan, “Mode-finding for mixtures of Gaussian distributions,” _IEEE Transactions on Pattern Analysis and Machine Intelligence_ , vol. 22, no. 11, pp. 1318–1323, 2000.
* [38] J. V. Michalowicz, J. M. Nichols, and F. Bucholtz, “Calculation of differential entropy for a mixed Gaussian distribution,” _Entropy_ , vol. 10, no. 3, pp. 200–206, 2008.
* [39] O. Zobay _et al._ , “Variational bayesian inference with Gaussian-mixture approximations,” _Electronic Journal of Statistics_ , vol. 8, no. 1, pp. 355–389, 2014.
* [40] J.-Y. Chen, J. R. Hershey, P. A. Olsen, and E. Yashchin, “Accelerated monte carlo for Kullback-Leibler divergence between Gaussian mixture models,” in _2008 IEEE International Conference on Acoustics, Speech and Signal Processing_. IEEE, 2008, pp. 4553–4556.
* [41] H. Joe, “Estimation of entropy and other functionals of a multivariate density,” _Annals of the Institute of Statistical Mathematics_ , vol. 41, no. 4, pp. 683–697, 1989.
* [42] F. Nielsen and R. Nock, “Maxent upper bounds for the differential entropy of univariate continuous distributions,” _IEEE Signal Processing Letters_ , vol. 24, no. 4, pp. 402–406, 2017.
* [43] M. F. Huber, T. Bailey, H. Durrant-Whyte, and U. D. Hanebeck, “On entropy approximation for Gaussian mixture random vectors,” in _2008 IEEE International Conference on Multisensor Fusion and Integration for Intelligent Systems_. IEEE, 2008, pp. 181–188.
* [44] S. J. Julier and J. K. Uhlmann, “A general method for approximating nonlinear transformations of probability distributions,” Technical report, Robotics Research Group, Department of Engineering Science, Tech. Rep., 1996.
* [45] A. Kolchinsky and B. D. Tracey, “Estimating mixture entropy with pairwise distances,” _Entropy_ , vol. 19, no. 7, p. 361, 2017.
* [46] P. Hall and S. C. Morton, “On the estimation of entropy,” _Annals of the Institute of Statistical Mathematics_ , vol. 45, no. 1, pp. 69–88, 1993.
* [47] T. Jebara and R. Kondor, “Bhattacharyya and expected likelihood kernels,” in _Learning theory and kernel machines_. Springer, 2003, pp. 57–71.
* [48] T. Jebara, R. Kondor, and A. Howard, “Probability product kernels,” _Journal of Machine Learning Research_ , vol. 5, no. Jul, pp. 819–844, 2004\.
* [49] J. E. Contreras-Reyes and D. D. Cortés, “Bounds on rényi and shannon entropies for finite mixtures of multivariate skew-normal distributions: Application to swordfish (xiphias gladius linnaeus),” _Entropy_ , vol. 18, no. 11, p. 382, 2016.
* [50] H. Chernoff _et al._ , “A measure of asymptotic efficiency for tests of a hypothesis based on the sum of observations,” _The Annals of Mathematical Statistics_ , vol. 23, no. 4, pp. 493–507, 1952.
* [51] D. Haussler, M. Opper _et al._ , “Mutual information, metric entropy and cumulative relative entropy risk,” _The Annals of Statistics_ , vol. 25, no. 6, pp. 2451–2492, 1997.
* [52] T. M. Cover and J. A. Thomas, _Elements of information theory_. John Wiley & Sons, 2012.
* [53] M. N. Do, “Fast approximation of Kullback-Leibler distance for dependence trees and hidden Markov models,” _IEEE signal processing letters_ , vol. 10, no. 4, pp. 115–118, 2003.
* [54] J. Munkres, “Algorithms for the assignment and transportation problems,” _Journal of the society for industrial and applied mathematics_ , vol. 5, no. 1, pp. 32–38, 1957.
* [55] H. W. Kuhn, “The hungarian method for the assignment problem,” _Naval research logistics quarterly_ , vol. 2, no. 1-2, pp. 83–97, 1955.
|
§ INTRODUCTION
Supersymmetric field theories are invariant under a set of transformations taking bosons to fermions and vice versa. By definition, these supersymmetry transformations form a Lie superalgebra extending the algebra of Poincaré transformations to a super-Poincaré algebra. By construction, the fields in the theory are linear representations of the bosonic Poincaré subalgebra. Ideally, this supermultiplet of fields furnishes a (possibly reducible) linear representation of the full supersymmetry algebra. For this to be true, it is necessary for the commutator of two supersymmetries|as realized on a component field|to close onto the translations on that field (up to other bosonic symmetries and possibly gauge transformations).
On- and Off-shell Supersymmetry
However, it is generally the case that this requirement fails and—strictly speaking—the fields do not furnish a linear representation of the supersymmetry algebra.
If the action is invariant under the fermionic transformations that extend the Poincaré symmetries to a consistent superalgebra, how can the fields fail to be a representation? An important observation in this context is that the failure of the supersymmetries to close on the fields is by terms that vanish if the equations of motion are imposed.[Such terms are called trivial symmetries. In a canonical treatment of symmetries, in which there is a Poisson bracket on the space of fields, these symmetries are canonical transformations generated by the action $S$ itself. They are trivial, in the sense that $\delta S \sim \{ S, S\}_{\mathrm {PB}} \equiv 0$.]
In this context, the supersymmetry is said to close on-shell only, and the set is referred to as an on-shell supermultiplet.
This should not be interpreted to mean that the component fields in the theory are required to be on-shell; only that the supersymmetries do not close on the off-shell realization.
In the ideal case, this complication does not arise: Successive supersymmetry transformations close properly on the fields irrespective of any field equations, and the collection of fields furnishes a linear representation. When this fortuitous situation is realized, it is emphasized by referring to it as off-shell supersymmetry. When we compare these off-shell theories to the on-shell ones described previously, we notice that they avoid the obstruction to closure with extra fields: Roughly, the terms in the transformations of components that give rise to obstructions are replaced with new fields.
Auxiliary Fields and the Off-shell Problem
Under supersymmetry transformations, the new component fields transform into the obstruction.
Such fields are called auxiliary fields.
In two-derivative actions, they have algebraic equations of motion (no derivatives on them), so they do not contribute dynamical degrees of freedom. In principle, they can be integrated out of the path integral (or set to their classical values), but removing them changes the supersymmetry transformations which now close only on-shell.
The problem of extending on-shell formulations of supersymmetric theories to off-shell supersymmetry by adding auxiliary fields is known as the off-shell problem. It is considered important for various reasons ranging from the purely mathematical representation theory of supersymmetry algebras to phenomenological supersymmetric model building. (In the case of most interest to us in this paper, we intend to use off-shell supersymmetry to study self-interactions in the form of higher-derivative corrections to the M-theory effective action.)
A considerable amount of effort has been devoted to solving the off-shell problem. An immediate indication of the inadequacy of any step-wise approach attempting to add suitable component fields and transformations was provided by Siegel and Roček [1] (see also sec. 2.4.3 of [2]): For any non-real representation in theories with eight or more supersymmetries (e.g. hypermuliplets), any potential solution to this problem requires an infinite number of fields. The proof is essentially a counting argument comparing the dimensions of off-shell Fock representations of the supertranslation algebra to those of superfield representations.
Manifest Supersymmetry
Just as the translation part of the Poincaré algebra is represented on fields by coordinate derivatives, so too can the supersymmetries be geometrized to translations in fermionic directions. Introducing coordinates for these directions, we can consider superfields as functions of both the even and odd coordinates. The component fields of a supermultiplet then appear as coefficients in the Taylor expansion of the superfields in the odd variables.
By construction, superfields are field representations of the super-Poincaré algebra (just as the component fields are so for the Poincaré subalgebra), and the supersymmetry is said to be manifest. However, these are generally (highly-)reducible representations, so we must reduce them in a way that is compatible with manifest supersymmetry.
To this end, we can impose combinations of reality conditions, gauge symmetry, and multiplet-reducing constraints compatible with superspace. In the latter, we use the supertranslations generated by fermionic covariant derivatives $D$ to impose covariant constraints such as $D\Phi = 0$, $D^2 \Phi =0$, etc. These are often called shortening conditions, because they imply that $\Phi$ does not depend on some of the odd coordinates thereby shortening its Taylor expansion.
The problem of finding irreducible representations of the super-Poincaré algebra thus becomes one of understanding the possible shortening conditions. This is non-trivial because the superalgebra implies that successive $D$s can generate bosonic derivatives $\{D, D\}\sim -2i \partial$, so imprudently-chosen constraints generate kinetic operators generalizing those of d'Alembert and Dirac. In this event, the multiplet is on-shell again in the sense described above.
But note that in superspace it further implies the superfield equations of motion directly on the representation. In other words, we cannot take such a manifestly supersymmetric representation off-shell in any sense; the construction of an action principle using such a representation is moot.
Much work in the superspace literature has been devoted to finding clever shortening conditions by enlarging the superspace with appropriate bosonic spaces. These naturally circumvent the counting arguments by introducing field dependence on bosonic coordinates geometrizing the R-symmetry transformations [3, 4, 2] or on bosonic ghost-like variables (e.g. Lorentz harmonics [5, 6] or pure spinor [7, 8, 9]). This approach can be generalized and systemized using supercosets [10].
Alternatively, we can attempt to work around the problem using a superspace with a smaller structure group than the maximal one. Examples of this include light-cone superspace [11] and its generalizations (e.g. [12, 13]). In these reductions, new shortening conditions can be defined that are not manifestly covariant under the original symmetry group, but also do not imply any dynamical equations. Such off-shell representations may then be combined into a supermultiplet representing the larger bosonic group linearly, albeit not manifestly.
11D, $N=1/8$ Superspace
An attempt to classify the possible structure group reductions appropriate to 16 supercharges was made in [14].
In previous work, we applied a similar approach to eleven-dimensional supergravity based on the reduction $Spin(10,1)\to SL(2, \mathbf C) \times G_2$ [15, 16, 17, 18, 19, 20].
Besides being the most familiar and phenomenologically relevant, this superspace has the significant advantage of requiring only finitely-many auxiliary fields (and has the largest super-Poincaré symmetry with this property [1]).
In this formulation the component fields of eleven-dimensional supergravity are embedded into a set of so-called prepotential superfields satisfying no constraints other than the reality and shortening conditions (i.e. chiral superfields) allowed off-shell.
Besides the 11D frame $e_{\bm m}{}^{\bm a}$, 3-form $C_{\bm {mnp}}$, and gravitino $\psi_{\bm m}{}^{\bm \alpha}$, these superfields contribute a finite set of auxiliary fields guaranteeing that 4 of the 32 supersymmetries close off-shell on the set of physical + auxiliary fields.
In this paper, we will construct superfields whose $\theta$-independent components are precisely the physical 11D gauge connections (i.e. the frame, 3-form, and gravitino).
The manifest supersymmetry transformations of the component fields are effected by acting with the $SL(2, \mathbf C)$ superspace derivatives $D$ on these superfields.
Besides mixing the bosonic gauge fields, this action generates the auxiliary fields in a form familiar from the standard component “tensor calculus”.
We use this to explicitly identify the spectrum of auxiliary fields (or, more precisely, the superfields that have the auxiliary components as their $\theta$-independent term).
The supergeometrical tensors of the eleven-dimensional theory (i.e. 4-form field strength, torsion, and curvature 2-form) may then be defined in the naïve way by taking the curls of the gauge connections.
By construction, this spectrum of superfields forms a(n off-shell) representation of the reduced structure supergroup.
As a result of this analysis, we derive the set of torsion constraints for the supergeometry with this reduced structure group.
As it is traditionally done, this set is the starting point for solving the curved superspace Bianchi identities [21, 22, 23].
[It is known that imposing just the dimension zero torsion condition
$T_{\bm\alpha \bm\beta}{}^{\bm c} \propto (\Gamma^{\bm c})_{\bm \alpha \bm \beta}$
is sufficient to put 11D superspace on-shell [24]. This observation allows one to postulate an “off-shell” formulation of 11D supergravity (or at least parametrize higher derivative corrections) by relaxing this constraint and re-solving the Bianchi identities [25]. However, there is some disagreement about this point
[26, 27, 28].
Our result does not speak directly to this issue, because we attempt
to keep only 4 of the 32 supersymmetries off-shell.]
At the end of this lengthy and technical process (which we do not carry out in this paper), one finally arrives at the solution to the constraints which are the off-shell prepotentials.
Amusingly, then, in the situation we described above and will follow in this paper, we have essentially started with the solution to a set of constraints we do not yet know, and can then derive these constraints by acting repeatedly with the superspace derivatives on the unconstrained fields.
Since the prepotentials with which we start are unconstrained, this process is guaranteed to terminate without ever imposing a dynamical constraint.
Along the way, all the physical potentials, field strengths, and Bianchi identities are derived.
§.§ Results
The component fields of eleven-dimensional supergravity consist of a frame $e_{\bm m}{}^{\bm a}$, a gauge 3-form $C_{\bm {mnp}}$, and a gravitino $\psi_{\bm m}^{\bm \alpha}$.
Under linearized supersymmetry, they transform into each other as
\begin{align}
\label{E:LinSUSY}
\delta_{\bm \epsilon} e_{\bm a }{}^{\bm b} &= \bar {\bm \epsilon} \Gamma^{\bm b} \psi_{\bm a}
\cr
\delta_{\bm \epsilon} C_{\bm {abc}} &= 3 \bar {\bm \epsilon} \Gamma_{[\bm {ab} }\psi_{\bm c]}
\cr
\delta_{\bm \epsilon} \psi_{\bm a}{}^{\bm \beta} &= -\tfrac12 \omega_{\bm a}{}^{\bm{bc}}
(\Gamma_{\bm {bc}} \bm \epsilon)^{\bm \beta}
- \tfrac1{6\cdot 4!} \left( 3 \Gamma^{\bm {bcde}}\Gamma_{\bm a }\bm \epsilon
- \Gamma_{\bm a } \Gamma^{\bm {bcde}}\bm \epsilon \right) \bm G_{\bm {bcde}}
\end{align}
where $\bm G := d C$ is the 4-form field strength, and we have flattened all the indices with the background frame $\delta_{\bm m}^{\bm a}$.
As a realization of supersymmetry, these transformations can only close on-shell:
The degrees of freedom needed for the off-shell counting to match are missing (cf. table <ref>).
\begin{align*}
{\renewcommand{\arraystretch}{1.7} %adds some padding
\begin{array}{|c|c|c|c|c|c|c|c|}
\hline
\textrm{gauge field}
& ~~~~\textrm{on-shell}~~~~
& \textrm{off-shell}\\
\hline
\textrm{frame}
& ~~\tfrac12 ({\textrm{D}}-1)({\textrm{D}}-2) -1 ~~
& \tfrac12 {\textrm{D}}({\textrm{D}}-1) \\
& 55 \\
\hline
\textrm{gravitino}
& 2^{\lfloor{\tfrac {\textrm{D}}2}\rfloor -1} \cdot ({\textrm{D}}-3)
& ~~2^{\lfloor{\tfrac {\textrm{D}}2}\rfloor} \cdot ({\textrm{D}}-1)~~ \\
& 128
& 320 \\
\hline
& {{\textrm{D}}-2\choose p}
& {{\textrm{D}}-1\choose p} \\
& 120 \\
\hline
\hline
~~\textrm{total (D=11)}~~
& {\color{PineGreen}128+128 }
& {\color{Red}175+ 320} \\
\hline
\end{array}
\end{align*}
Counting of degrees of freedom in D dimensions relevant to Poincaré supergravity. (For conformal supergravity, some (gamma-)traces should be subtracted.)
In this paper, we will explain how to fix this mismatch and give explicitly the modifications to these supersymmetry transformations (cf. eqs. <ref>, <ref>, and <ref>).
To get a representation on which 4 of the 32 supersymmetries close off-shell, we decompose the 11D component fields under
\begin{align}
\label{E:StructureGroupReduction}
\to Spin(3,1) \times Spin(7)
\to Spin(3,1) \times G_2
= SL(2, \mathbf C)\times G_2
\end{align}
For notational convenience, we will label spacetime, polarizations, indices, et cetera by $M$ for 11D, $X$ for the 4D part, and $Y$ for the 7D part.
Under the decomposition of the tangent space,
\begin{align}
\begin{array}{cccclcc}
e_{\bm m}{}^{\bm a} &\to &
\frameXX_{m}{}^{a} ~,~
\frameYX_{i}{}^{a} ~,~
\frameXY_{m}{}^{j} ~,~
\frameYY_i{}^j
&\textrm{with }& a, m = 0,1,2,3,
\\
\psi_{\bm m}^{\bm \alpha}&\to &
\grinoXX_{m}^{~\alpha} ~,~
\grinoXY_{m}^{~\alpha i} ~,~
\grinoYX_{j}^{~\alpha } ~,~
\grinoYY_{j}^{~\alpha i} ~,~
&\textrm{and }& \alpha = 1,2,
\\
C_{\bm {mnp}} &\to &
\CXXX_{mnp}~,~
\CXXY_{mn \, i} ~,~
\CXYY_{m \, ij}~,~
\CYYY_{ijk}
&\textrm{and }& i,j = 1,\dots, 7
\end{array}
\end{align}
To avoid introducing too much notation, the $i, j$, …indices will be doing quadruple duty as 7D coordinate indices, 7D tangent space indices, $G_2$ indices, and as a label for 7 additional gravitino fields. (We collect the index definitions in table <ref>).
\begin{align*}
{\renewcommand{\arraystretch}{1.5} %adds some padding
\begin{array}{|c|c|c|c|c|c|c|c|}
\hline
\textrm{index} & ~~~~\textrm{range}~~~~ & \textrm{description} &~~~\textrm{label}~~~\\
\hline
\bm m, \bm n, \dots & 0, \cdots , 10 & \textrm{11D coordinate} &\\
\bm a, \bm b, \dots & 0, \cdots , 10 & \textrm{11D tangent} &M\\
\bm \alpha, \bm \beta, \dots & 1, \cdots , 32 & \textrm{11D spinor} &\\
\hline
m, n, \dots & 0, 1,2 , 3 & \textrm{4D coordinate} &\\
a, b, \dots & 0, 1,2 , 3 & \textrm{4D tangent} &X \\
~~\alpha, \beta , \dots , \dt \alpha, \dt \beta \dots & 1, 2 & \textrm{4D spinor} &\\
\hline
i,j,\dots & 1,\cdots , 7 & ~~\textrm{7-component label} ~~&Y\\
\hline
\end{array}
\end{align*}
Legend of indices used in this work. To avoid having even more indices, those in the various $\bm7$-dimensional representations ($GL(7)$ coordinate, $SO(7)$ tangent, $G_2$ representation, and label for seven gravitini), have all been identified.
Using the $G_2$ structure to trade the 7D graviton polarizations $\gravitonYY_{ij}$ for a stable 3-form $F_{ijk}$, these components embed into a set of superfields called prepotentials [16], as summarized in table <ref>.
The prepotential superfields are unconstrained superfields or constrained only to be real or chiral.[ Such constraints are innocuous: Besides factors of $\bar D^2$ arising from the variation of chiral fields, these fields are unconstrained as integration variables in a path integral. (For this reason, we will sometimes sloppily refer to them as unconstrained.)]
The components that arise in the Taylor expansion of the fields are listed under the superfield. They have been separated into three groups:
* Gauge fields transform canonically. Their (double) curls are field strength, torsion, and curvature 2-form components. Using table <ref>, they add up to $128|128$ components on-shell, but off-shell (modulo gauge transformation), we find a mismatch between $175$ bosonic and $320$ fermionic components.
* Auxiliary fields are gauge invariant, but do not propagate. Due to this, it is easily verified that they contribute $201$ bosonic and $56$ fermionic degrees of freedom all of which are off-shell.
* Compensators suffer Stückelberg shifts under “pregauge” transformations. They do not appear in the spectrum of gauge(-invariant) superfields, so they do not contribute any degrees of freedom at all.
The compensating fields are artifacts of the reduction of the structure supergroup implied by (<ref>).
This reduction is crucial to the representation of the off-shell supersymmetries, but it has no geometrical eleven-dimensional meaning. (Presumably it has some pregeometrical meaning, but the off-shell pregeometry for 11D is precisely the thing we do not know.)
This situation is precisely analogous to 4D, $N=1$ Poincaré supergravity itself, which is described in terms of conformal supergravity (which is not a physical symmetry) coupled to a scalar superfield called the conformal compensator and denoted $\G$ (<ref>) in this paper.
Relatedly, the auxiliary fields that appear in the Taylor expansion of the prepotentials are—strictly speaking—not invariant under all of the pregeometrical symmetries.
However, the parts of the pregeometrical transformations under which they are not covariant are again artifacts of the structure group reduction having no eleven-dimensional interpretation.
Therefore, it is possible to covariantize these auxiliary components by mixing in some parts of the other prepotentials.
This is done explicitly in section <ref>, but the details are not important here.
\begin{align*}
{\renewcommand{\arraystretch}{1.5} %adds some padding
\begin{array}{|c|c|c|c|c|c|c|c|c|}
\hline
& U^a & \Psi_i^\alpha & \mathcal V^i & X & \Sigma_{\alpha i} & V_{ij} & \Phi_{ijk} &~~\textrm{sdim}_{\mathbf R} ~~\\
~\textrm{prepotential}~& \textrm{real} & \textrm{spinor} & ~~\textrm{real}~~ & ~~\textrm{real}~~ & ~~\textrm{chiral}~~ & \textrm{real} & \textrm{chiral}&\\
\hline
\physical{\textrm{physical }}
& ~~\gravitonXX_{ab} , \grinoXX_a{}^\beta ~~
& \grinoXY_a{}^{\beta j} & \gravitonXY_a^i & \CXXX_{abc}
& \CXXY_{ab\,i} & ~~\CXYY_{a\, ij}, \spinorYY_{\alpha\, ij} ~~
& ~~\CYYY_{ijk}, \gravitonYY_{ij}, \spinorYYY_{\alpha\, ijk}~~&\\
% & 2|2 & \\
\auxiliary{\textrm{auxiliary }} & \dX^a
&~~ \l^i_\alpha, \y_i^a, \t_i{}^{\alpha \beta} ,\r_i^\alpha ~~
& \dY^i & \d & & \dYY_{ij} & \f_{ijk} &\\
&&&\surfeit^i_\alpha &\compX,\surfeit'_\alpha
&\compY_i,\surfeit''_{\alpha i}&&&
\\
\hline
\textrm{on-shell} & 2|2 & 0|14 & 14|0 & 0|0 & 7|0 & 42|42 & 63|70 & 128|128 \\
%\textrm{off-shell} & 10|12 & 98|140 & 28|0 & 2|0 & 21|0 & 84|84 & 133|140 & 376|376\\
\textrm{off-shell} & (\physical{6}+\auxiliary{4})|\physical{12} & \auxiliary{98}|(\physical{84}+\auxiliary{56}) & (\physical{21}+\auxiliary{7})|0 & (\physical{1}+\auxiliary{1})|0 & \physical{21}|0 & (\physical{63}+\auxiliary{21})|\physical{84} & (\physical{63}+\auxiliary{70})|\physical{140} & 376|376\\
\hline
\end{array}
\end{align*}
The unconstrained “prepotential” superfields of the 11D, $N=1/8$ supergravity multiplet and their physical, auxiliary, and compensating component content. The components of the gravitino of the type $\psi_i{}^\beta$ and $\psi_i{}^{\beta j}$ are encoded in $\chi_{\alpha ij}$ and $\chi_{\alpha ijk}$ (<ref>). The real superdimension contributed by each superfield is computed modulo gauge transformation. The superdimension of the auxiliary field space adds up to $201|56$.
Supersymmety Transfomations
Once the eleven-dimensional superfields corresponding to the components in table <ref> have been defined, we are in a position to compute their supersymmetry transformations.
In superspace, supersymmetry transformations of component fields result from infinitesimal translations in the odd directions—in other words—from acting with the fermionic covariant derivatives.
The result of this calculation is that the graviton transformation is unchanged from the on-shell rule:
\begin{align}
\label{E:SupersymmetryTransformationsGraviton}
\delta_\epsilon \gravitonXX_{ab} &=
2i \epsilon \sigma_{(a} \bar \grinoXX_{b)}
\cr
\delta_\epsilon \gravitonXY_{a i} &=
- i \epsilon\grinoXX_{a\,i} +i \epsilon \sigma_{a} \bar \grinoYX_{i}
\cr
\delta_\epsilon \gravitonYY_{ij} &=
-2i \epsilon\grinoYY_{(i, j)}
\end{align}
($SL(2, \mathbf C)$ spinor index contractions are implied.)
The supersymmetry transformation of the 3-form acquires a correction by the dimension-1/2 auxiliary field $\l^i_\alpha$ of the gravitino multiplet (cf. table <ref>, or eq. <ref> for the explicit expression):[Some of the gravitino components can be shifted by the gauge-invariant dimension-1/2 auxiliary field $\l^i_\alpha$, and this type of correction could also have appeared in the graviton transformation. Because this field vanishes on-shell, this leads to an ambiguity in the definition of the off-shell gravitini. The definition of the gravitino we make below (see eq. <ref>) ensures that $\l^i_\alpha$ appears in only one place in the off-shell transformations (<ref>) and
\begin{align}
\label{E:SupersymmetryTransformationsThreeform}
\delta_\epsilon \CXXX_{abc} &= -6 \epsilon \sigma_{[ab} \grinoXX_{c]}
\cr
\delta_\epsilon \CXXY_{ab\, i} &= - 4i \epsilon \sigma_{[a} \bar \grinoXY_{b] i}
-2 \epsilon \sigma_{ab} \grinoYX_{i}
\cr
\delta_\epsilon \CXYY_{a\, ij} &= \varphi_{ijk} \epsilon \grinoXY_{a}{}^k
-4i\delta_{k[i} \epsilon \sigma_{a} \bar \grinoYY_{j]}{}^k
-\tfrac16 \varphi_{ijk} \epsilon \sigma_{a} \bar \l_{}^k
\cr
\delta_\epsilon \CYYY_{ijk} &= \varphi_{l[ij} \epsilon \grinoYY_{k]}{}^l
\end{align}
In these transformations, we define $\delta_\epsilon$ to include a Wess-Zumino gauge transformation|that is, a field-dependent gauge transformation that involves unphysical terms in the superfield multiplet.
Finally, we present the gravitino supersymmetry transformations.
These allow us to identify the dimension-1 auxiliary fields needed to close the supersymmetry algebra.
For legibility, we suppress the ubiquitous spin connection terms in addition to including compensating Wess-Zumino gauge transformations.
(The complete expressions are presented in appendix <ref>.)
\begin{align}
\label{E:SupersymmetryTransformationsGravitino}
\delta_\epsilon \grinoXX_a^\beta &\simeq
- \tfrac i2 (\epsilon\eta_{ab} + \epsilon \sigma_{ab})^\beta \dX^b
-\tfrac i2 (\bar \epsilon \bar \sigma_a )^{\beta} \R
\\
\delta_\epsilon \grinoYX_i^\beta &\simeq
-\tfrac i{12} (\epsilon \sigma^{ab})^\beta \t_{ab\, i}
-\tfrac i2 \epsilon^\beta \dY_i
- \tfrac i{12} (\bar \epsilon \bar \sigma_a )^\beta \Ga_i^{a}
\cr
\delta_\epsilon \grinoXY_{a}{}^{\beta j}
-\tfrac{5i}{12} (\epsilon\eta_{ab} + \tfrac 85 \epsilon \sigma_{ab} )^\beta\, \bar \Ga^{bj}
-i (\bar \epsilon\bar \sigma_a )^\beta \dY^j - \tfrac i6
\big[ 4 \eta_{ab} \bar \epsilon \bar \sigma_c + i \varepsilon_{abcd} \bar \epsilon \bar \sigma^d
\big]^\beta \t^{jbc}
\cr&
-\tfrac14 \epsilon^\beta \big[
i \Ga^j_a
-\y^j_a - \bar \y^j_a
\big]
\cr
\delta_\epsilon \grinoYY_{i}{}^{\beta j} &\simeq
-i(\epsilon \sigma^{ab})^\beta \pi_{\bm 14} \FSXXYY_{ab\,i}{}^j
-i (\bar \epsilon \bar \sigma^a)^\beta \left[
\hmap_i^j(\FSXYYY_{a\, klm})
+\tfrac i{6}\varphi_i{}^{jk} (
i \Ga_{a k}
-\y_{a k}
\right]
- \tfrac 12 \epsilon^\beta \Ztorsion_i{}^j
\nonumber
\end{align}
In these expressions, only certain $G_2$ representations of the 4-form field strength $\FS G_{\bm{abcd}}$ and dimension-1 auxiliary fields appear explicitly.
For the former, these are the $\bf{14}$-dimensional part of the $[2,2]$-form component $\FSXXYY_{ab\, ij}$, and the $\bf{27}+\bf{1}$ parts of the $[1,3]$-form component $\FSXYYY_{a\, ijk}$.
(The explicit definitions of these projectors are given in appendix <ref>; cf. eqs. <ref> and <ref>.)
For the latter, the invariants $\R$, $\Ga_i^a$, and $\Ztorsion_i{}^j$ appearing here are combinations of the more basic auxiliary fields $\d$, $_i^a$, $_ij$, $_ijk$ fields in table \ref{T:Components} and components of the 4-form field strength.
These relations are given explicitly in section \ref{S:Auxiliary} (cf.\ eqs.\ \ref{E:R}, \ref{E:Ga}, and \ref{E:Ztorsion}).
\paragraph{Outline}
This concludes our description of the $376|376$-component multiplet of eleven-dimensional supergravity furnishing a linear representation of the superalgebra with four real supersymmetries.
In the remainder of the paper, we will explain the derivation of this result.
In the process, we will provide explicit expressions for the gauge superfields, their field strengths, and relations between these under the action by the fermionic superspace derivatives.
The results are organized as follows:
In the next section, we give the steps starting from the basic prepotentials and arriving at the connection and auxiliary superfields.
In section \ref{S:Dynamics}, we describe the dynamical equations that result from the off-shell action derived in [16].
These equations relate auxiliary fields to physical components, thereby demonstrating the relation between on- and off-shell field content.
In section \ref{S:Supergeometry}, we consolidate the results for the 4-form, torsion, and curvature field strengths in terms of supergeometry, and conclude the body of the paper.
What follows are four appendices with conventions beginning with basic results for 4D, $N=1$ superspace (app.\ \ref{S:Superspace}) and
the essentials of $G_2$ geometry (app.\ \ref{S:G2}).
This is followed by an appendix \ref{S:Pregauge} of intermediate calculations combining the prepotentials of section \ref{S:Prepotentials} into building blocks with simplified transformations that are used to define all the other fields throughout this work.
Finally, we present the result of acting with fermionic superspace derivatives in all possible ways on the gravitino components in appendix \ref{S:GrinoDescendants}.
This demonstrates the consistency of the scheme and generates the supersymmetry transformations of the eleven-dimensional gravitino components reported above in \eqref{E:SupersymmetryTransformationsGravitino}.
\section{Prepotentials}
\label{S:Prepotentials}
We now begin the task of explicitly identifying all the elements described in the previous sections.
As implied in table \ref{T:Components}, we embed the components into prepotential superfields
\begin{align}
\label{E:Prepotentials}
U^a &\sim \cdots
+ (\theta \sigma^m \bar \theta) \frameXX_m{}^a
+ \theta^2 (\sigma^m \bar \theta)_\alpha \grinoXX_m{}^\alpha
+ \bar \theta^2 (\theta \sigma^m )_{\dt \alpha} \bar \grinoXX_m{}^{\dt \alpha }
+ \theta^2 \bar \theta^2 \dX^a
\cr
\Psi^{\alpha i} &\sim \cdots
+ (\theta \sigma^m \bar \theta) \grinoXY_m{}^{\alpha i}
+ \theta^2 (\bar \theta \bar \sigma_a)^\alpha \y_i^a
+ \bar \theta^2 (\theta \sigma^{ab})^\alpha \t_{ab i}
+ \theta^2 \bar \theta^2 \r^i_\alpha
\cr
\mathcal V^i &\sim \cdots
+ (\theta \sigma^m \bar \theta) \frameXY_m^i
+ \bar \theta^2 \theta^\alpha \surfeit_\alpha^i
+ \theta^2 \bar \theta_{\dt \alpha}\bar \surfeit^{\dt \alpha i}
+ \theta^2\bar \theta^2 \dY^i
\cr
X &\sim \cdots
+ \theta^2 \bar \compX
+ \bar \theta^2 \compX
+ \theta \sigma_m \bar \theta \mathcal \, \varepsilon^{mnpq} \CXXX_{npq}
+ \bar \theta^2 \theta^\alpha \surfeit_\alpha
+ \theta^2 \bar \theta_{\dt \alpha}\bar \surfeit^{\dt \alpha}
+ \theta^2\bar \theta^2 \d
\cr
\Sigma_i^\alpha &\sim \cdots
+ [\theta^\alpha \compY_i + (\theta \sigma^{mn})^\alpha \CXXY_{mn \, i} ]
+ \theta^2 \tilde \surfeit_i^\alpha
\cr
V_{ij} &\sim \cdots
+ (\theta \sigma^m \bar \theta) \CXYY_{m\, ij}
+ \bar \theta^2 \theta^\alpha \spinorYY_{\alpha \, ij}
+ \theta^2 \bar \theta_{\dt \alpha}\bar \spinorYY_{ij}^{\dt \alpha}
+ \theta^2\bar \theta^2 \dYY_{ij}
\cr
\Phi_{ijk} &\sim \CYYY_{ijk} + i \calibration_{ijk}
+ \theta^\alpha \spinorYYY_{\alpha\, ijk}
+ \theta^2 \f_{ijk}
\end{align}
We recall that, in addition to the {\color{PineGreen} gauge fields} transforming canonically, there are {\color{Orange} auxiliary fields} that do not propagate, and {\color{Red} compensating components} that shift under pregauge transformations so they do not appear in the spectrum.
Our main result is a kind of inversion of this embedding: We will present explicit expressions for 11-dimensional superfields containing these gauge fields and auxiliary fields as the leading term in their Taylor expansions.
(To avoid introducing yet more notation, we will use the same symbols to denote the component field and the superfield having that component as the leading term in its $θ$-expansion.
This should cause no confusion, as henceforth we will always mean the superfield unless explicitly stated otherwise.)
There are no analogous superfields for the compensating components, precisely because these have no lift to 11D.
These superfields are complicated but explicit combinations of the prepotentials that transform only under the physical part of the pregauge transformations (by which we mean those corresponding to 11D diffeomorphisms, supersymmetry, Lorentz transformations, and 3-form gauge transformations). Concretely, the gauge fields must transform as 11D supergravity gauge fields, whereas the auxiliary fields must be invariant.
The procedure to construct the eleven-dimensional superfields is the following:
\begin{enumerate}
\item We write the linearized transformations of the prepotentials \eqref{E:Prepotentials}. These are collected in appendix \ref{S:Pregauge}. They contain abelian transformations of the M-theory 3-form, a non-abelian gauging thereof under 7D diffeomorphisms, local 4D, $N=1$ superconformal transformations, and extensions thereof ({\it e.g.}\ extended supersymmetry transformations).
\item The field strengths of the non-abelian tensor hierarchy (\ref{E:ATHXf}, \ref{E:KK}) are invariant under the linearized hierarchy transformations but not under the superconformal ones. So we make combinations that are invariant under one or two of these additional parameters using as few superspace derivatives as possible.
\item These partial invariants are used to covariantize (with respect to the superconformal parameters) the 3-form components we already had from the non-abelian tensor hierarchy.
This results in the eleven-dimensional 3-form superfield $\physical C_{\bm{abc}}$ \eqref{E:gauge3form} with leading Taylor component the physical 3-form in the on-shell spectrum.
\item In the same way, we construct the eleven-dimensional frame superfield $\physical e_{\bm a}{}^{\bm b}$ \eqref{E:11Dframe} and gravitino superfield $\physical \psi_{\bm a}{}^{\bm \beta}$ \eqref{E:grino} from the partial invariants so that they transform as they should (\ref{E:translationParameter} and \ref{E:localSupersymmetry}). Taking curls defines the dimension-3/2 torsion $\FS{T}_{\bm {ab}}{}^{\bm \gamma}$ ({\it i.e.}\ the gravitino curvature)
and dimension-2 curvature $\FS{R}_{\bm {ab}}{}^{\bm {cd}}$.
\item Finally, we use the partial invariants to covariantize the auxiliary fields in table \ref{T:Components}. The resulting invariants are listed in equation \ref{E:AuxiliarySuperfields}.
\end{enumerate}
These steps are summarized in table \ref{T:Summary}.
%%%% Tabel 2
\begin{table}[t]
\begin{align*}
{\renewcommand{\arraystretch}{1.5} %adds some padding
\begin{array}{|ccccc|}
\hline
~~~\textrm{prepotential} ~~~
& ~~~\textrm{partial invariant} ~~~
& ~~~ \textrm{gauge field} ~~~
& ~~~ \textrm{auxiliary} ~~~
& ~~~\textrm{field strength} ~~~\\
\hline
& \ast
& \frameXX_a{}^b , \grinoXX_a{}^\beta
& \dX^a
& \Torsion_{ab}{}^{\gamma}, \Curvature_{ab}{}^{cd} \\
\Psi_i^\alpha
& \X_i^a , \T_i
& \frameYX_i{}^a, \grinoXY_b{}^{\alpha i}
&\l_i^\alpha , \y_i^a , \Ga_i^a,\t_{ab\,i},\r_i^\alpha
& \Torsion_{ab}{}^{\gamma k} ,
\Curvature_{aj}{}^{cd} \\
\mathcal V^i
& \KK_\alpha^i
& \frameXY_a{}^i, \grinoYX_i{}^{\beta}
& \dY^i
& \Curvature_{ab}{}^{ij} \\
X & \G , \S, \P
& \CXXX_{abc}
& \d, \R
& \FSXXXX_{abcd}\\
\Sigma_i^\alpha
& \T_i
& \CXXY_{ab\, i}
& \textrm{none}
&\FSXYYY_{abc\, i}\\
& \hatW^\alpha_{ij}
& \CXYY_{a\, ij}, \spinorYY^\alpha_{ij}
& \dYY_{ij}
&\FSXYYY_{ab\, ij} , \Curvature_{ab\, ij}\\
\Phi_{ijk}
& \Z_{ijk} , \E_{ijkl}
& \CYYY_{ijk}, \frameXX_i{}^j , \spinorYYY^\alpha_{ijk}
& \m_{ijk} , \f_{ijk}
& \FSXYYY_{a\, ijk}, \FSYYYY_{ijkl}, \Curvature_{aj\, kl}, \Curvature_{ij\, kl}\\
\hline
\end{array}
\end{align*}
\caption{\scriptsize Prepotentials and their derived quantities. All derived quantities contain admixtures of other prepotentials, so these identifications are approximate.
(The assignment of curvature tensor components to any one field is similarly ambiguous.)
Explicit formul\ae{} for partial invariants in terms of the prepotentials are deferred to appendix \ref{S:Pregauge}.
(Hatted quantities are $\Psi$-corrected versions of the na\"ive hierarchy field strengths.)
The dimension-$1, \tfrac32, 2$ components of the field strength $\FS{G}$, torsion $\Torsion$, and curvature $\Curvature$ are the curls of the gauge fields as described in section \ref{S:Connections}.
\label{T:Summary}
} %end scriptsize
\end{table}
\subsection{Eleven-dimensional Superfields}
\label{S:11DSuperfields}
In this section, we define superfields that have the property that their leading component corresponds to a field in eleven dimensions.
These are either physical connections (and their curls) or auxiliary fields that do not propagate.
\subsubsection{Connection Superfields}
\label{S:Connections}
We construct the superfields carrying the eleven-dimensional 3-form, frame, and gravitino as their leading component by combining the prepotentials \eqref{E:Prepotentials} and their derivatives into superfields that transform as connections.
Much work goes into this construction in which successive combinations are made to have progressively simpler transformations under the gauge and local superconformal parameters.
This is outlined in appendix \ref{S:Pregauge}, with the results presented here.
\paragraph{M-theory 3-form}
Under dimensional reduction, the eleven-dimensional 3-form splits up into a collection of $[p,q]$ forms embedded into superfields in [20]. From this, we may extract the superfields for these components, and covariantize them with respect to the local superconformal transformations.
This results in the following set of superfields:
\begin{subequations}
\label{E:gauge3form}
\begin{align}
\hyperlabel{E:CXXX}
\CXXX_{abc}&:= -\tfrac14\varepsilon_{abc\,{\delta \dt \delta} } \left(
[D^\delta, \bar D^{\dt \delta}] X
- (D^2 +\bar D^2) U^{\delta \dt \delta}
\right)
\\
\hyperlabel{E:CXXY}
\CXXY_{\alpha \beta \, i} &:= -\tfrac i2 D_{(\alpha} \left[ \,
\Sigma_{\beta) i}
+\Psi_{\beta) i}\,
\right]
\\
\hyperlabel{E:CXYY}
\CXYY_{\alpha \dt \alpha \, ij}&:=
\tfrac12 [D_\alpha, \bar D_{\dt \alpha}] V_{ij}
-\varphi_{ijk} \partial_{\alpha \dt \alpha} \mathcal V^k
+ \varphi_{ijk} \X_{\alpha \dt \alpha}^k
\\
\hyperlabel{E:CYYY}
\CYYY_{ijk} &:= \tfrac12(\Phi_{ijk} + \bar \Phi_{ijk}) + \tfrac12 \psi_{ijkl} \T^l
\end{align}
\end{subequations}
The combinations $_i^a$ and $_i$ may be thought of as partial covariantizations of the $D̅Ψ$ and $DΨ$ parts of the gravitino superfields (by the derivatives $∂_i U^a$ and $∂_i X$, respectively).
They are defined explicitly in \eqref{E:X} and \eqref{E:T}, but we will not need that result.
Note, however that the corrections by ${U^a, Ψ_i^α, X}$ to all the na\"ive components of the gauge 3-form are needed to cancel the local superconformal transformations of the compensating fields.
With this, the corrected superfields transform as gauge $[p,q]$-forms, so we may immediately define the field strength superfield $G_abcd = 4 ∂_[a C_bcd]$ corresponding to the M-theory 4-form invariant.
(An explicit set of expressions for this in terms of tensor hierarchy invariants is given in \eqref{E:4formInvariants}.)
The construction given here is typical of everything that follows. We will now present the analogous results for the frame and gravitino.
\paragraph{Eleven-dimensional Frame}
The eleven-dimensional frame is determined by starting with the na\"ive term in the $θ$-expansion \eqref{E:Prepotentials} of the prepotentials and adding terms so that the resulting superfields transform as $δe_a^b = ∂_a ξ^b -λ_a^b$.
Up to an $SO(3,1) ×SO(7)$ gauge choice and a normalization we fix by matching to reference [23], this fixes the components of the frame to be
\begin{subequations}
\label{E:11Dframe}
\begin{align}
\hyperlabel{E:frameXX}
\frameXX_{\un a}{}^{\un b} &:=
-\tfrac12 [D_\alpha, \bar D_{\dt \alpha} ] \, U^{\un b}
-\tfrac23 \delta_\alpha^\beta \delta_{\dt \alpha}^{\dt \beta} \S
\\
\hyperlabel{E:frameXY}
\frameXY_{\un a}{}^{j} &:= \tfrac12 [D_\alpha, \bar D_{\dt \alpha} ] \, \mathcal V^j
\\
\hyperlabel{E:frameYX}
\frameYX_i{}^{\un b} &:= - \tfrac12 [
\bar D^{\dt \beta} \Psi_i^\beta - D^\beta \bar \Psi_i^{\dt \beta}
\\
\hyperlabel{E:frameYY}
\frameYY_{i}{}^{j} &:= \tfrac14 \varphi^{j kl} \F_{ikl}
- \tfrac1{36} \delta_i^j \varphi^{klm} \F_{klm}
\end{align}
\end{subequations}
The translation and local Lorentz parameters are shared among these components, so they are also fixed. For the translations, this gives
\begin{align}
\label{E:translationParameter}
\xi^{\un a} :=-i ( \bar D^{\dt \alpha} L^\alpha + D^\alpha \bar L^{\dt \alpha} )
\xi^i &:=
\tfrac1{2} (\tau^i +\bar \tau^i )
-\tfrac1{2i} (\Omega^i - \bar \Omega^i)
\end{align}
We give the explicit expressions for the local Lorentz parameter \eqref{E:localLorentz} in appendix \ref{S:GrinoDescendants}.
We see explicitly from these expressions that the 4D component is as expected, and the Kaluza-Klein component is the na\"ive one.
We also see that the real part of the complex vector in $D̅ Ψ$ is the ``lower-left block'' of the Kaluza-Klein decomposition.
This block is usually taken to be zero, but in superspace that is only true to lowest order in the $θ$-expansion. (The lowest component of $_i^b $ vanishes in Wess-Zumino gauge. See [29] for a complete discussion of this component.)
We also see explicitly that the part $_i^j$ of the linearized frame lying along the seven-dimensional space is a particular component of the partial field strength $_ijk$ for the ``scalars'' of the tensor hierarchy. (This is a partial covariantization of the imaginary part of the chiral field $Φ_ijk$, see eq.\ \ref{E:F}.)
The symmetrization of this frame defines the linearized eleven-dimensional metric $h_ab : = 2e_(a^cδ_b)c$.
The Riemann tensor $R_ab^cd := ∂_[a ∂^[c h_b]^d]$ is the double curl of this.
The linearized curvature 2-form $R_ab^cd := 2 ∂_[a ω_b]^cd$ is the curl of the spin connection.
These curvatures agree component-wise when the spin connection is taken to be torsion-free, since $T_ab^c = 0$ implies $ω_c^ab := 12 ∂^[a h_c^b]$.
\paragraph{Eleven-dimensional Gravitino}
The definition of the eleven-dimensional gravitino suffers from various ambiguities.
Below, we present what we consider the simplest one.
To do so, it is helpful to first define the components (which appear also in table \ref{T:Components})
\begin{subequations}
\label{E:spinor}
\begin{align}
\hyperlabel{E:spinorYY}
\spinorYY_{ij}^\alpha &:=
\W_{ij}^\alpha +2\partial_{[i} \Psi_{j]}^\alpha
+i\varphi_{ijk} (\KK^{\alpha k} -\tfrac14 \bar D^2 \Psi^{\alpha k})
\\
\hyperlabel{E:spinorYYY}
\spinorYYY_{ijk}^\alpha
&:= D^\alpha \left( i\F_{ijk}
+\tfrac 12 \psi_{ijkl} \T^l
+3 \varphi_{l[ij} \partial_{k]} \mathcal V^l\right)
\end{align}
\end{subequations}
The first of these may be understood as the covariantization (canceling the local superconformal shifts) of the 21 gaugini in the abelian vector multiplet prepotential $V_ij$ by the Kaluza-Klein gauge field $𝒱^i$ and the gravitino superfield $Ψ_i^α$.
The second is the analogous thing for the 35 spin-1/2 fields in the chiral multiplet $Φ_ijk$. (The part of $Ψ_i^α$ that enters
requires the further covariantization by $Σ_i^α$ into $$.)
Together, they comprise 56 spin-1/2 fields from the 4D perspective [30].
Using various $G_2$ projectors given in appendix \ref{S:G2}, these components can be reorganized into the $56 = 7+ 49$ components $ ψ_i^β, ψ_i^βj$ of the linearized eleven-dimensional gravitino. Using this, we explicitly define the eleven-dimensional gravitino components as
\begin{subequations}
\label{E:grino}
\begin{align}
\hyperlabel{E:grinoXX}
\grinoXX_{\un a}{}^\beta &:=-\tfrac{i}8 \bar D^2 D^{\beta} U_{\un a}
+\tfrac i{3} \delta_\alpha^\beta
\bar D_{\dt \alpha} \S
\\
\hyperlabel{E:grinoYX}
\grinoYX_i{}^\beta
&:= \tfrac1{12}\l_i^\beta-\tfrac1{12}\left[
\varphi_i{}^{jk} \spinorYY^\beta_{jk}
-\tfrac 1{6} \psi_i{}^{jkl} \spinorYYY^\beta_{jkl}
\right]
\\
\hyperlabel{E:grinoXY}
\grinoXY_{\un a}{}^{\beta j}
&:= 2 D^\beta \X_{\un a}^j
-\tfrac23 \delta_\alpha^\beta \left[
2D^\gamma \X^j_{\gamma \dt \alpha}
+ \bar D_{\dt \alpha} \T^j
+2i \bar \KK^j_{\dt \alpha}
\right]
\\
\hyperlabel{E:grinoYY}
\grinoYY_i{}^{\beta j}
\hmap_i{}^j(\spinorYYY^\beta_{klm})
-\tfrac 1{36} \varphi_i{}^{jk} \psi_k{}^{ lmn} \spinorYYY_{lmn}^\beta
-\pi_{\bm{14}} ( \spinorYY^\beta_i{}^j )
\end{align}
\end{subequations}
(In the last component, we see explicitly the $49 = 28 + 7 + 14$ structure (recall eq.\ \ref{E:hmap}) and how the original $21$ and $35$ are mixed together through $G_2$ representations to form the $7$ and $49$.)
These components are defined by their gauge transformations
\begin{align}
\label{E:localSupersymmetry}
\delta \physical \psi_{\bm a}{}^{\bm \beta} = \partial_{\bm a} \epsilon^{\bm \beta}
\epsilon^\beta := -\tfrac14 \bar D^2 L^\beta
\epsilon^{\beta j} := - 2 D^\beta \Omega^j
\end{align}
so that their curls $T_ab^γ$ are invariant; they are the dimension-3/2 torsion components.
But this condition is ambiguous: As we will explain presently (cf.\ eq.\ \ref{E:grinEoM}), there is an invariant spinor superfield $ł^i_α$ of dimension 1/2 that can be added to three of the four components defined above.
The ones we have defined here are as we originally derived them. Curiously, they also turn out to be the simplest ones in the sense that they imply that all the dimension-1/2 torsions vanish, as we will see in section \ref{S:Supergeometry}.
\subsubsection{Auxiliary Superfields}
\label{S:Auxiliary}
We will now define superfields containing the auxiliary components as the leading term of their $θ$-expansion.
These ``auxiliary superfields'' will be given the same name as their leading component to avoid a further proliferation of new notation.
This should cause no confusion since, throughout this paper, we will always mean the superfield unless we explicitly state otherwise.
%\paragraph{Auxiliary Superfields}
The definition of the dimension-1/2 gravitini was ambiguous because there is a dimension-1/2 invariant
\begin{subequations}
\label{E:AuxiliarySuperfields}
\begin{align}
\hyperlabel{E:grinEoM}
\l_i^\alpha
&:= 2\bar D_{\dt \alpha} \X_i^{ \un a}
-D^\alpha \T_i
+2i \J_i^\alpha
\end{align}
This field is present in any Lagrangian theory of the gravitino multiplet $\Psi_i^\alpha$ where, as a component field, it forms a Lagrange multiplier pair with a dimension-3/2 spinor $\r_i^\alpha$ to be defined below.
Because of this, and as we will see explicitly in section \ref{S:Dynamics}, this superfield is proportional to the $\Psi_i^\alpha$ equation of motion.
There are many more dimension-1 auxiliary fields since most 4D, $N=1$ superfields contain one.
The vector multiplets all contain a real auxiliary field we denote by $\auxiliary d$. There are 4 such real prepotentials: the conformal graviton $U^a$, the Kaluza-Klein field $\mathcal V^i$, the 3-form $X$, and the abelian vectors $V_{ij}$. Their $\auxiliary d$-components are covariantized as
\begin{align}
\hyperlabel{E:dX}
\dX^{\un a} &:=\tfrac18 D^\beta \bar D^2 D_\beta U^{\un a}
+ \tfrac16 [ D^\alpha, \bar D^{\dt \alpha}] \S
+\partial^{\un a} \P
\\
\hyperlabel{E:d}
\d
-\tfrac 14 (D^2 + \bar D^2) \S
\\
\hyperlabel{E:dY}
\dY^i
&:=\tfrac13 D^\alpha \KK_\alpha^i
-\tfrac23 \partial^{a} \X_{a}^i
+\tfrac i{12} (D^2 - \bar D^2) \T^i
\\
\hyperlabel{E:dYY}
\dYY_{ij}&:= -\tfrac1{32} \psi_{ij}{}^{kl} \big[
D^\alpha \spinorYY_{\alpha \, kl}
+\bar D_{\dt \alpha} \bar \spinorYY_{kl}^{\dt \alpha }
\big]
-\tfrac16 \varphi_{ijk} \partial^k \S
+\tfrac14 \hitchin_{ijk\, lmn}\partial^k \F^{lmn}
\end{align}
Another familiar 4D, $N=1$ auxiliary is that of the chiral scalar field, of which we have only one such representation $\Phi_{ijk}$.
Covariantizing its auxiliary $\auxiliary f$-component, we find
\begin{align}
\hyperlabel{E:f}
\f_{ijk}
&:= D^2 \Z_{ijk}
+\tfrac i3 (\hitchin^{-1})_{ijk\, lmn}\varepsilon^{lmnpqrs}\partial_p \Z_{qrs}
\end{align}
(Here $\Z_{ijk}$ is the partial covariantization of the scalars $\F_{ijk}$ defined in \eqref{E:Z}.)
Finally, there are the additional auxiliary fields of the gravitino multiplet.
This multiplet is reviewed in detail in appendix C of [16].
In addition to the dimension-1/2 auxiliary field \eqref{E:grinEoM}, there are two dimension-1 components
\begin{align}
\hyperlabel{E:y}
\y^i_{\alpha \dt \alpha}&:= - \bar D_{\dt \alpha} \l_\alpha^i
\\
\hyperlabel{E:t}
\t_{\alpha \beta i} &:=\tfrac12 D_{(\alpha} \KK_{\beta) i}
-\tfrac i4 \big[ D_{(\beta} \bar D^{\dt \alpha}
+3 \bar D^{\dt \alpha} D_{(\beta}
\big] \X_{\alpha) \dt \alpha i}
\cr&=
-\tfrac14 \bar D^2 D_{(\alpha} \Psi_{\beta) i}
+\tfrac i4 \partial_i \big[ D_{(\beta} \bar D^{\dt \alpha}
+3 \bar D^{\dt \alpha} D_{(\beta} \big] U_{\alpha)\dt \alpha}
+ \tfrac i2 \partial_{(\beta}{}^{\dt \alpha} \gravitonXY_{\alpha)\dt \alpha
\end{align}
comprising a complex vector and a 2-form,
and the top component of $\Psi$
\begin{align}
\hyperlabel{E:r}
\r^i_\alpha&:=-\tfrac14 D^\beta \bar D^2 D_{(\alpha} \Psi_{\beta)}^i
- i \partial^i \big[
D_\alpha \G +\tfrac12 \bar D^{\dt \alpha} D^2 U_{\un a}
\big]
\end{align}
\end{subequations}
completing the set of auxiliary fields with a dimension-3/2 spinor.
The auxiliary fields defined in this section are important because they are part of a linear representation of the subalgebra generated by 4 of the 32 supersymmetries.
Many of them appear explicitly in the supersymmetry transformation \eqref{E:SupersymmetryTransformationsGravitino} of the eleven-dimensional gravitino.
To compare this rule to the known on-shell transformation \eqref{E:LinSUSY} and to prove that they do not contribute propagating degrees of freedom, we examine in section \ref{S:Dynamics} how they appear in the off-shell dynamics.
\paragraph{Useful Alternative Composite Fields}
The supersymmetry transformations of the gravitino are mixtures of the auxiliary fields and 4-form flux.
The gauge-invariant combinations that appear naturally in the differential algebra are
\begin{subequations}
\begin{align}
\hyperlabel{E:R}
\R &:= -\tfrac 16 \bar D^2 \S = \tfrac i6 \bar D^2 \P
= -\tfrac 1{12} \bar D^2 [ \bar \G + i \partial_{\un a} U^{\un a} ]
= \tfrac13 \d + \tfrac i{6\cdot 4!} \varepsilon^{abcd} \FSXXXX_{abcd}
\\
\hyperlabel{E:Ga}
\Ga^{a}_i &:= -i\bar D^2 \X^{a}_i
+\tfrac i2 \bar D \bar \sigma^a D \T_i
\cr
&= \tfrac 1{2i} \big[ \y_i^{a} - \bar \y_i^{a} \big]
+\tfrac16 \psi_{ijkl} \FSXYYY^{a jkl}
+i \tilde \FSXXXY_i^{a}
\\
\hyperlabel{E:m}
\m^{ijk}
&:= \hitchin^{ijk\, lmn} D^2 \Z_{lmn}
- \tfrac 13 \epsilon^{ijk lmnp}\partial_l \bar \Phi_{mnp}
%\cr &
= (\hitchin)^{ijk\, lmn} \f_{lmn}-\tfrac 1{12} \varepsilon^{ijklmnp} \FSYYYY_{lmnp}
\\
\hyperlabel{E:Ztorsion}
\Ztorsion_i{}^j
- i \hmap_i{}^j ( \f_{lmn})
+\tfrac i{36}\varphi_i{}^{jk}\psi_k{}^{lmn} \f_{lmn}
-8 \pi_{\bm{14}} \dYY_i{}^j
\end{align}
\end{subequations}
The first of these is the chiral scalar field strength of old-minimal supergravity containing the complex auxiliary field, the spin-1/2 part of the gravitino, and the curvature scalar [21, 22, 23].
In the 3-form formulation of the theory employed here, the imaginary part of the auxiliary field is dualized to the 4-form component $_abcd$ [31, 32].
The second is a complex linear field $D̅^(β ^α)α_i = 0$. In fact, $_i^αα = 12i D̅^α _i^α$, which explains why it is this combination that features prominently in the gravitino supersymmetry transformations.
Expanding it out explicitly, we see that it is related to the imaginary part of the gravitino auxiliary field $_i^a$. (Note, however, that the real part of $$ does appear in the transformation of $_a^βj$ \eqref{E:SupersymmetryTransformationsGravitino} cf.\ also eq.\ \ref{E:DgrinoXY}.)
The third is a modification of the auxiliary field of the scalars \eqref{E:f} by a real field strength.
It is an interesting alternative that is chiral.\footnote{
Contraction with the constant $\varepsilon_{i_1\dots i_7}$ symbol defines a chiral 4-form
\begin{align*}
\bm E_{ijkl}
&:= -\tfrac 1{12} \varepsilon_{ijklmnp} \bar \m^{mnp}
= 4 \partial_{[i} \Phi_{jkl]}
- \tfrac1{12} \varepsilon_{ijklmnp} \bar D^2 \left[ \hitchin^{mnp\,qrs} \bar \Z_{qrs}\right]
\end{align*}
} %end footnote
The last combination $_i^j$ of auxiliary fields appears in the supersymmetry transformation of $_i^βj$ (cf.\ eq.\ \ref{E:SupersymmetryTransformationsGravitino}).
Equivalently, this final invariant is the component of the dimension-1 torsion $_αj^γk =4 δ_α^γ_j^k +⋯$ that is a singlet under the 4D part of the Lorentz group (see eq.\ \ref{E:TorsionY}).
In the form presented, the decomposition $49 = 28 + 7 + 14$ is evident.
Of these the $28 = 27 +1$ and $7$ are complex, whereas the $14$ is real.
This agrees with the counting of the first-order invariants of the $G_2$ structure, with the real $49 = 27 + 1 + 7+14$ corresponding to the four torsion forms \eqref{E:TorsionForms} and the imaginary $35= 27 + 1+7$ to
the analogous derivatives of the gauge 3-form (the derivative in the $14$ is not gauge invariant).
Finally, we mention for completeness that additional invariants are possible that look new, but are actually linear combinations of the invariants defined above.
For example, it is easy to check that the dimension-3/2 spinor $D̅^2 D^α_i$ is invariant. It is somewhat less trivial to show that it satisfies
\begin{align}
\tfrac {3i}{4} \bar D^2 D_\alpha \T_i
&= \r_{\alpha i}
+ 2 (\sigma^{ab})_\alpha{}^\beta \partial_{[a}\grinoXX_{b] \beta i}
-12 (\sigma^a)_{\alpha \dt \alpha}
\partial_a \bar \grinoYX_i{}^{\dt \alpha}
-\partial_i \bar \grinoXX_a{}^{\dt \alpha}
\end{align}
We conclude that this invariant is not new, since it is a combination of
the previously-defined auxiliary field $_̊i^α$ \eqref{E:r} and curls of the gravitini \eqref{E:grinoXX} and \eqref{E:grinoYX}.
(The choice of the dimension-3/2 spinor depends on how we order the derivatives $D$ and $D̅$ in the definition of the auxiliary component.)
This conclusion was guaranteed by the component results of [16] where a Wess-Zumino gauge analysis was carried out to identify the spectrum: Any new invariant would have to correspond to a physical component field, but this would be in contradiction with the Wess-Zumino analysis since we already accounted for all of those fields.
\section{Dynamics}
\label{S:Dynamics}
The prepotential superfields of tabel \ref{T:Components} are not constrained except for reality and chirality conditions.
Therefore, we can write an action in terms of them to determine the dynamical equations.
This was done at the linearized level in [16] (and to low order in a gravitino superfield expansion in [15]).\footnote{The gravitino and Kaluza-Klein field were treated non-linearly in reference [29] by including them into the superspace geometry. In this approach $\Psi_i^\alpha$ and $\mathcal V^i$ all appear in the connections alongside $U^a$.}
Defining the variation $E_ϕ := δS/δϕ$ of that action with respect to the prepotential $ϕ$, we obtain the supercurrents
\begin{subequations}
\label{E:SuperEoM}
\begin{align}
\hyperlabel{E:EoMX}
\EoMX{}_a %E_{U^{a}}
&= - \dX_a +\tfrac1{18} \varphi^{ijk} \FSXYYY_{a\, ijk}
\\
\hyperlabel{E:EoMgrino}
\EoMgrino{}^i_\alpha &= \tfrac i4 \l_\alpha^i
\\
\hyperlabel{E:EoMY}
\EoMY{}_i
&= - 3\dY_i
+\tfrac16\varphi^{jkl} \FSYYYY_{ijkl}
\\
\hyperlabel{E:EoM}
\EoM
&= -\tfrac23 \d
+\tfrac 1{72}\varphi_{ijk}\hitchin^{ijk\, lmn} (\f_{lmn} +\bar \f_{lmn})
+\tfrac 1{48}\psi^{ijkl}\FSYYYY_{ijkl}
\cr
&= -\tfrac23 \d
-\tfrac 1{96}\varphi_{ijk}(\m^{ijk} +\bar \m^{ijk})
+\tfrac 1{96}\psi^{ijkl}\FSYYYY_{ijkl}
\\
\label{E:EoMSigma}
\vanishing{E_{\Sigma} }{}_\alpha^i &=
\tfrac i{16}\bar D^2 \l_\alpha^i
\\
\hyperlabel{E:EoMYY}
\EoMYY{}^{ij}
&= 2\dYY^{ij}
\\
\hyperlabel{E:EoMYYY}
\EoMYYY^{ijk}
\tfrac i{48} \hitchin^{ijk\, lmn} \bar \f_{lmn}
% -\tfrac i{2^6 \cdot 3^2} \varepsilon^{ijklmnp} \FSYYYY_{lmnp}
-\tfrac i{576} \varepsilon^{ijklmnp} \FSYYYY_{lmnp}
+\tfrac i6 \varphi^{ijk} \R
\cr
\tfrac i{48} \bar \m^{ijk}
+\tfrac i{12} \varphi^{ijk} \R
\end{align}
\end{subequations}
These are in bold font because they are gauge-invariant, and we have colored them red to indicate that setting them to zero defines the superspace mass-shell condition.\footnote{This overloading of notion should not be confused with the compensating component fields, which are unrelated and have played no role since the earliest part of section \ref{S:Prepotentials}.}
This is a useful distinction because, although {\em we will continue to work off-shell}, these equations are needed to prove the claim that auxiliary fields do not contribute any new degrees of freedom on-shell: The auxiliary fields differ from components of the physical 4-form field strength only by the superfields $E_∗$ that vanish on-shell.
Explicitly, we can solve these equations \eqref{E:SuperEoM} for the auxiliary fields to find
\begin{subequations}
\begin{align}
\dX_a &=\tfrac1{18} \varphi^{ijk} \FSXYYY_{a\, ijk}
- \EoMX{}_a
\\
\l_\alpha^i &= -4i \EoMgrino{}_\alpha^i
\\
3\dY_i&= \tfrac16\varphi^{jkl} \FSYYYY_{ijkl}
- \EoMY{}_i
\\
\tfrac12 \d
-\tfrac 1{96}\psi^{ijkl}\FSYYYY_{ijkl}
-\tfrac i2 \varphi_{ijk} ( \EoMYYY - \overline {\EoMYYY} )^{ijk}
\\
\dYY_{ij} &= \tfrac12 \EoMYY{}_{kl}
\\
\hitchin^{ijk\, lmn} \f_{lmn} &=
\tfrac 1{36}\left[
\varphi^{ijk}(i \varepsilon^{abcd} \FSXXXX_{abcd}
+ \psi^{lmnp} \FSYYYY_{lmnp} )
+3 \varepsilon^{ijklmnp} \FSYYYY_{lmnp}
\right]
\cr
+\tfrac{4i}3 \varphi^{ijk} \varphi_{lmn} ( \EoMYYY - \overline {\EoMYYY})^{lmn}
-\tfrac 83\varphi^{ijk} \EoM
\end{align}
\end{subequations}
The auxiliary fields not appearing here ${ _i^a, _i^αβ, _̊i^α}$ are those of the gravitino superfield.
However, it is not difficult to show that
\begin{align}
\y_{i}^a&= -2i \bar D \bar \sigma^a \EoMgrino{}_i
\cr
\t_{\alpha \beta i} &=
-\tfrac12 \varphi_i{}^{jk} \FSXXYY_{\alpha \beta jk}
- D_{(\alpha}\EoMgrino{}_{\beta) i}
\cr
\r_i^\alpha
-\tfrac 1{12} (\bar D\bar \sigma^a)^{\alpha}
\left[
\psi_i{}^{jkl} \FSXYYY_{a\, jkl}
-i \varepsilon_a{}^{bcd} \FSXXXY_{bcd\, i}
\right]
+2i \psi_i{}^{jkl} \partial_j \spinorYY_{kl}^\alpha
- i \bar D_{\dt \alpha} D^\alpha \overline \EoMgrino^{\dt \alpha}_{ i}
\end{align}
It follows that $$ and $$ contribute nothing new on-shell.
Similarly, up to equations of motion, $$̊ is a sum of three curls, each of which is separately an invariant of dimension 3/2—that is—they are components of the gravitino curl.
Finally, we note that the action of the superspace derivatives on the field strength $\FS{G}=d\physical{C}$ (i.e. its supersymmetry transformation) generate gravitino curls by (<ref>), up to supersymmetry spacetime derivatives of $\sim D\l_i $.
The latter can be rewritten in terms auxiliary fields and field strengths by (<ref>) and
\begin{align}
D_{(\alpha} \l_{\beta)}^i & = 4i \t_{\alpha \beta}{}^i + 2i \varphi^{ijk} \FSXXYY_{\alpha \beta jk}
\\
D^\alpha \l_{\alpha}^i &=
6i \dY^i
+ 4\varphi^{ijk} \dYY_{jk}
-\tfrac i6 \psi^{ijkl} \f_{jkl}
\end{align}
implying that supersymmetric derivatives of the dimension-1 field strength generate no new invariants.
This concludes the demonstration that the $N=1/8$ off-shell component spectrum reduces correctly to the physical spectrum if the component equations of motion are imposed by taking $\color{Red}\bm E_\ast| \to 0$.
§ TOWARD ELEVEN-DIMENSIONAL OFF-SHELL SUPERGEOMETRY
The analysis that leads to the results of the previous sections also implies the constraints on the supergeometry needed to solve the superspace Bianchi identities.
Such a top-down approach defines the superspace torsion and curvature through the commutator
\begin{align}
\label{E:TorsionCurvature11D}
[ \nabla_{\bm A}, \nabla_{\bm B} ] =
T_{\bm A \bm B}{}^{\bm C} \nabla_{\bm C}
+\tfrac12 R_{\bm A \bm B}{}^{\bm c\bm d} M_{\bm c \bm d}
\end{align}
where $\nabla_{\bm A} = \nabla_{\bm \alpha} , \nabla_{\bm a}$ is the covariant superspace derivative of eleven-dimensional supergravity (${}_{\bm \alpha} = {}_1,\dots,{}_{32}$ and ${}_{\bm a} = {}_0,\dots , {}_{10}$).
These are subject to the Bianchi identities that follow from the nested commutator identity
\begin{align}
\label{E:Jacobi11D}
[[ \nabla_{\bm A}, \nabla_{\bm B} ] , \nabla_{\bm C} ] ] + \textrm{permutations}= 0
\end{align}
The non-trivial part says that
\begin{align}
\label{E:BI11D}
\nabla_{[\bm C} T_{\bm A \bm B]}{}^{\bm D}
+T_{[\bm A \bm B]}{}^{\bm E} T_{|\bm E| \bm C] }{}^{\bm D}
+\tfrac12 R_{\bm A \bm B}{}^{\bm c\bm d} (\Gamma_{\bm c \bm d})_{\bm C}{}^{\bm D} = 0
\end{align}
These identities become non-trivial when covariant conditions are imposed on various components of the torsion and curvature. Some of these conditions are a matter of convention that follows from the possibility of shifting the definitions of the various connections (most importantly, the spin connection).[
This can be formalized in terms of first-order-flat $G$ structures [33, 34].
The obstruction theory for this is known as Spencer cohomology (cf. e.g. Schwachhöfer's article “Holonomy Groups and Algebras” in [35]). The Spencer cohomology for 11D supergravity was considered in [36].]
For eleven-dimensional supergravity in superspace, these are
\begin{align}
\label{E:TorsionConstraints11D}
T_{\bm \alpha \bm \beta}{}^{\bm \gamma} = 0
T_{\bm \alpha \bm \beta}{}^{\bm c} = -2i (\Gamma^{\bm c} )_{\bm \alpha \bm \beta}
T_{\bm a\bm b}{}^{\bm c} = 0
\end{align}
Others follow from the deformation of constraints defining representations of the flat space supersymmetry algebra (most notably, chiral).
There is no known analog of this type of constraint in eleven-dimensional supergravity, presumably because there are no matter multiplets with spins $ < 3/2$ in this case.
Historically, appropriate torsion constraints were often found by making an Ansatz and solving the Bianchi identities.
Poor Ansätze over-constrain the torsions leading to trivial solutions or on-shell geometries. We can circumvent this problem by reducing the structure group so that there are fewer torsion constraints and more possibilities for deformed representation constraints.
Of course, guessing the torsion constraints is complicated by the lack of symmetry.
But now we notice that we already know the solution to the linearized constraints, even though we do not know the constraints!
This is because in the case of off-shell supergeometry, the solution to the superspace Bianchi identities culminates in explicit expressions for the field strengths, torsions, and curvatures in terms of unconstrained prepotentials. We already know from the derivation of the correct spectrum [16], that these must be equivalent to the prepotentials (<ref>).
Starting with these, then, it must be possible to derive the constraints from this solution directly.
In principle, this can be done by acting repeatedly on the prepotentials with superspace derivatives in all possible ways to generate a graded space of superfields.
In this large space, we will find (among many other things) invariants and relations between them.
The invariants are the aforementioned field strengths, torsions, curvatures, and their derivatives, since these are the only invariants of a superspace by definition.
The relations between them will then be the (reduced) Bianchi identities.
In this paper, we have taken a shortcut by building the connections and covariantizing the auxiliary fields. The remaining invariants are the curls of the connections. The constraints can then be gleaned from the spectrum of invariants and the supersymmetry transformation of the connections.
Off-shell 11D, $N=1/8$ Supergeometry
Let us now specialize to the supergeometry appropriate to the eleven-dimensional superspace considered in this paper.
Since only $N=1/8$ is manifest, we retain from the superspace covariant derivatives above only $\{\nabla_\alpha, \bar \nabla_{\dt \alpha}, \nabla_a, \nabla_i\}$. Similarly, we keep only those parts of the Bianchi identity (<ref>) with the implied free indices.
However, the on-shell eleven-dimensional torsion (<ref>) is a tensor, so any non-vanishing component of it must be retained including $T_{\alpha\, \beta j}{}^k = -2i \varepsilon_{\alpha \beta} \delta_j^k $.
The implication of this is that torsion components of the form $T_{\alpha \beta j}{}^{\ast}$ can appear through the quadratic terms in (<ref>).
The remaining dimension-0 constraints of relevance from (<ref>) are
\begin{align}
T_{\alpha \beta}{}^c = 0
T_{\alpha \dt \beta}{}^c = -2i (\sigma^c)_{\alpha \dt \beta}
T_{\alpha \dt \beta}{}^k = 0
\end{align}
together with their conjugates.
This is consistent with our finding that there are no invariants with dimension $<1/2$.
Less obvious—but not difficult to verify—is that there are also no dimension-1/2 torsions with our choice of gravitino components:
In this form, the gravitino auxiliary field appears only in the dimension-1/2 component
\begin{align}
\FS G_{\alpha b jk} = -\tfrac16 (\sigma_b \bar \l^i)_\alpha \varphi_{ijk}
\end{align}
of the super-4-form field strength.
This leaves the dimension-1 conventional constraints.
Since there is no bosonic eleven-dimensional torsion, we have chosen the constraint accordingly: $T_{\bm{ab}}{}^{\bm c} =0$ defines the 11D spin connection in terms of the frame as we have chosen it in section <ref>.
We could now begin to solve the superspace Bianchi identities subject to these conditions.
This would be useful as a direct method of deriving the field strengths, perhaps even beyond their linearized approximation.
Instead, we observe that the dimension-1 torsion components can be deduced directly from the linearized gravitino transformations
\begin{gather}
\delta_\epsilon \psi_{\bm a}{}^{\bm \beta}
\sim \delta_\epsilon E_{\bm a}{}^{\bm \beta}
\sim -\epsilon D E_{\bm a}{}^{\bm \beta}
\sim -\epsilon^{\gamma} \Torsion_{\gamma \bm a}{}^{\bm \beta}
-\epsilon^{\dt \gamma} \Torsion_{\dt \gamma \bm a}{}^{\bm \beta}
\end{gather}
Collecting the relevant parts from (<ref>) (or app. <ref>), we find for the $\Torsion^{\gamma}$ part
\begin{gather}
\label{E:TorsionX}
\Torsion_{\alpha b}{}^{\gamma} =
\tfrac i2 [\delta^\gamma_\alpha \eta_{bc} + (\sigma_{bc})_\alpha {}^{\gamma}] \dX^c
\Torsion_{\alpha b}{}^{\dt \gamma} =
\tfrac 12 (\sigma_b)_{\alpha}{}^{\dt \gamma} \R
\cr
\Torsion_{\alpha j}{}^{\gamma} =
\tfrac i2 \delta_\alpha^\gamma \dY_j
+\tfrac i{12} (\sigma^{ab})_\alpha {}^\gamma \t_{abj}
\Torsion_{\alpha j}{}^{\dt \gamma} = \tfrac i{12} (\sigma_a)_\alpha{}^{\dt \gamma} \bar \Ga_j^a
\end{gather}
The first line is the same as those of textbook treatments of 4D, $N=1$ superspace [21, 22, 23].
They are essentially the definitions of the $\dX^a$ and $\R$ auxiliary fields that contain the 4D part of the Ricci tensor and scalar curvature in their $\theta$-expansion.
We see that for D $>4$, these are extended by $\dY_i$, $\t_{ab \,k}$, and $\bar \Ga_i^a$.
The analogous dimension-1 torsion $\Torsion^{\gamma k}$ with $\bm 7$-valued spinor index decomposes as
\begin{align}
\label{E:TorsionY}
\Torsion_{\alpha b}{}^{\gamma k} &=
\tfrac i{12} [5 \delta_\alpha^\gamma \eta_{bc} + 8 (\sigma_{bc})_\alpha {}^\gamma ] \bar \Ga^{ck}
+ \tfrac i8 \delta_\alpha^\gamma \Ga_b^k
-\tfrac 18 \delta_\alpha^\gamma
[ \bar D \bar \sigma_b \l^k + D\sigma_b \bar \l{}^k]
\cr
\Torsion_{\alpha b}{}^{\dt \gamma k} &=
i(\sigma_b)_\alpha{}^{\dt \gamma} \dY^k
+\tfrac i6 (4 \eta_{bc} \sigma_d - i \epsilon_{bcde} \sigma^e)_\alpha {}^{\dt \gamma} \t^{kcd} ,
\cr
\Torsion_{\alpha j}{}^{\gamma k} &=
\tfrac12 \delta_\alpha^\gamma \Ztorsion_j{}^k
+ i (\sigma^{ab})_\alpha{}^\gamma \pi_{\bm {14}}\FSXXYY_{ab\, j}{}^k
\cr
\Torsion_{\alpha j}{}^{\dt \gamma k} &=
i (\sigma^a)_\alpha{}^{\dt \gamma} [
\hmap_j{}^k(\FSXYYY_{a\,lmn})
-\tfrac 1{6} \varphi_j{}^{kl} \bar \Ga_{a l}
- \tfrac 16 \varphi_j{}^{kl} D_\alpha \bar \l^{\dt \gamma}_l
\end{align}
The new structures include the real and imaginary parts of the descendant $\y \sim \bar D\l$ of the gravitino auxiliary (which vanish on-shell), various projections of the 4-form field strength, and the superspace covariantization $\Ztorsion_i{}^j$ of the $G_2$ torsion from (<ref>).
The remaining torsion component was defined already at the end of section <ref> as the spacetime curl of the eleven-dimensional gravitino.
The dimension-2 invariants (curvature) were also defined in that section.
In summary, we have demonstrated that the only field strengths of the eleven-dimensional supergeometry with 4 off-shell supersymmetries are the dimension 1 4-form $\FS G_{\bm{abcd}}$, the dimension-3/2 torsion $\Torsion_{\bm {ab}}{}^{\bm \gamma}$, and the dimension-2 curvature $\Curvature_{\bm{ab}}{}^{\bm{cd}}$. These are all superfields with the implied component field strengths as their leading Taylor coefficient.
In addition, we find the auxiliary superfields (<ref>) with $1/2\leq$ dimension $\leq 3/2$, which are reduced dynamically to the aforementioned physical spectrum by setting the supercurrents (<ref>) to zero.
§ EPILOGUE
We have built up the linearized tensors of eleven-dimensional supergravity directly from the unconstrained prepotentials of this theory with the structure group reduced to $SO(3,1) \times G_2$.
This gives a superfield for each physical component field of eleven-dimensional supergravity.
In addition, the construction provides superfields embedding the auxiliary fields needed for a multiplet furnishing a linear representation with four off-shell supersymmetries manifest.
The calculations are quite cumbersome, but the results are easy to understand:
By reducing the structure supergroup of eleven-dimensional supergravity, it is possible to separate the torsion constraints of that supergeometry into a part that can be solved off-shell and a remainder that is responsible for putting the theory on-shell.
Ignoring the latter, and solving the former by the usual method in a top-down approach, gives a collection of so-called reduced field strength superfields.
These correspond to our collection of superspace invariants as expressed in (<ref>) and (<ref>).
These invariants still satisfy a large set of identities expressing, for example, superspace derivatives of one in terms of another.
Solving these identities results in explicit expressions for the reduced field strengths in terms of the unconstrained superfields (<ref>).
The miraculous-looking series of fortunate events required in this calculation was guaranteed to have this outcome, because we had previously shown that the set of unconstrained superfields (<ref>) correctly describes eleven-dimensional supergravity [16].
Acting repeatedly with superspace derivatives could, therefore, never generate an equation of motion or an inconsistent constraint, but was guaranteed to construct invariants corresponding to their reduced field strengths, their reduced Bianchi identities, et cetera.
What was unclear from the outset, was the expression of the torsions in terms of prepotentials, and, in particular, the torsion constraints needed to begin a top-down approach to the problem.
With the supergeometry torsion constraints so clarified, we could now “start over” with the usual superspace Bianchi identity analysis. The advantage of doing this would be a more direct derivation of the super-torsion, super-curvature, and super-4-form. These can be used to define higher-derivative invariants directly that can be used to study deformations of the supergeometry.
Alternatively, we can proceed to construct higher-derivative deformations of the action directly by making appropriate combinations of the tensors derived herein. (This is related to the approach above, but the methods used in the construction are different.)
Of most direct interest to us are the higher-derivative invariants needed to deform the two-derivative action.
For M-theory, these are the ones containing $R^4$ terms constructed from the curvature 2-form and those terms related to them by supersymmetry transformations.
In our hybrid superspace, these are divided into groups related by the manifest supersymmetries which are, in turn, related by the non-manifest ones.
We proceed by constructing $R^4$ invariants of the former that are closed as differential forms (as needed for terms like $\sim \int C \wedge R^4$) in superspace.
To check that such an approach is feasible, we have carried out the analogous calculation in a five-dimensional version of this scenario in [37].
There the invariant is of the form $\int A \wedge R^2$ where $A$ is the gravi-photon.
(Such an analysis was possible because enough of the 5D hybrid supergeometry had been worked out previously in [38].)
A comparison of the 11D and 5D geometries reveals that the latter is embedded in the former.
This implies that the results of the 5D analysis for $R^2$ can be extended to 11D, as we intend to demonstrate explicitly in future work.
§ ACKNOWLEDGEMENTS
It is a pleasure to thank
Jim Gates,
Konstantinos Koutrolikos,
and Warren Siegel
for illuminating discussions.
This work is partially supported by National Science Foundation grant PHY-1820921.
§ 4D, N=1 SUPERSPACE IDENTITIES
In this appendix, we rederive the $D$-identities in flat space for Minkowski superspace $\mathbf R^{4|4}$ in the conventions of reference [23] (which, for most practical calculations, agrees with [39]).
The bosonic subspace is $\mathbf R^4$ with Minkowski metric (components $\eta_{ab}$ with ${}_{a,b} = {}_{0,1,2,3}$) and a compatible $SL(2, \mathbf C)$ structure (components $\varepsilon_{\alpha \beta}$ and conjugate $\varepsilon_{\dt\alpha \dt \beta}$ with (anti-)Weyl spinor indices ${}_{\alpha, \beta, \dt \alpha, \dt \beta} = {}_{1,2}$).
These are related by the Pauli matrices $(\sigma_a)_{\alpha \dt \alpha}$ which may be interpreted as the components of isomorphisms $v^a \mapsto v_{\un a} :=v_{\alpha \dt \alpha}:= (\sigma_a)_{\alpha \dt \alpha} v^a$ between vectors and hermitian matrices acting on Weyl spinors.
The operation is so pervasive in superspace calculations that we use this special notation throughout the paper without qualification. Using it, the compatibility of the Minkowski and $SL(2, \mathbf C)$ structures is expressed by the Fierz identity
\begin{align}
\label{E:Fierz}
\eta_{\un a \un b} =-2 \varepsilon_{\alpha \beta} \varepsilon_{\dt \alpha \dt \beta}
(\sigma_a)_{\alpha \dt \alpha} (\sigma_b)_{\beta \dt \beta} \eta^{ab} =-2 \varepsilon_{\alpha \beta} \varepsilon_{\dt \alpha \dt \beta}
\end{align}
$N=1$ algebra Here we recall some formulae which all follow from the fundamental definition
\begin{align}
\left\{ D_\alpha , \bar D_{\dot \alpha}\right\} = -2i \partial_{\un a}
\end{align}
First we define $\triangle_{\un a} := \left[ D_\alpha , \bar D_{\dot \alpha}\right]$ and find
\begin{align}
\triangle_{\un a} = \left\{
\begin{array}{r}
2i \partial_{\un a} + 2 D_\alpha\bar D_{\dot \alpha}\\
-2i \partial_{\un a} - 2\bar D_{\dot \alpha} D_\alpha
\end{array}
\right.
\end{align}
With three $D$s we have
\begin{align}
[ \bar D^2, D_\alpha]= 4i \partial_{\un a} \bar D^{\dot \alpha} ~~~&,~~~
[ D^2, \bar D_{\dot \alpha}]= -4i \partial_{\un a} D^{\alpha}
\\
\label{E:D*DD*}
\{ \bar D^2, D_\alpha \} = - 2 \bar D_{\dot \alpha}D_{\alpha}\bar D^{\dot \alpha} ~~~&,~~~
\{ D^2, \bar D_{\dot \alpha} \} = - 2 D^{\alpha}\bar D_{\dot \alpha}D_{\alpha}
\\
\label{E:Diamond}
D^\alpha (D_\alpha \bar D_{\dt \alpha} + 2\bar D_{\dt \alpha} D_\alpha ) &= -\bar D_{\dt \alpha} D^2
\\
(D_\alpha \bar D_{\dt \alpha} + 2\bar D_{\dt \alpha} D_\alpha) \bar D^{\dt \alpha} &= - \bar D^2 D_\alpha
\\
(D_\alpha \bar D_{\dt \alpha} + 2\bar D_{\dt \alpha} D_\alpha) D^\beta \Psi_\beta &=
D^2 \bar D_{\dt \alpha} \Psi_\alpha + 2 D^\beta \bar D_{\dt \alpha} D_{(\alpha} \Psi_{\beta)}
\end{align}
and usual ones with four
\begin{align}
\left\{ D^2, \bar D^2\right\}= 2D^\gamma\bar D^2 D_\gamma +16 \Box ~~~&,~~~
\left[ D^2, \bar D^2\right] = -4i \partial^{\un a} \triangle_{\un a}
\end{align}
\begin{align}
D_{(\alpha} \bar D^2 D_{\beta)}= -2i \partial_{(\alpha}{}^{\dt \gamma} \triangle_{\beta) \dt \gamma}
\bar D_{(\dt \alpha} D^2 \bar D_{\dt \beta)}= - 2i \partial^\gamma{}_{(\dt \alpha} \triangle_{\gamma \dt \beta)}
\end{align}
There are other useful identities involving commutators. First of all, there is the symmetric part of the product
\begin{align}
\label{CommSymm}
\left\{ \triangle_{\un a}, \triangle_{\un b}\right\} &= -8 \partial_{\un a} \partial_{\un b} + 2 \varepsilon_{\alpha \beta} \varepsilon_{\dot \alpha \dot \beta} D^\gamma \bar D^2 D_\gamma
\end{align}
and the anti-symmetric part
\begin{align}
\label{CommAnti}
\left[ \triangle_{\un a}, \triangle_{\un b}\right] &=
2\varepsilon_{\dot \alpha \dot \beta} D_{(\alpha} \bar D^2 D_{\beta)}
- 2 \varepsilon_{\alpha \beta}\bar D_{(\dot \alpha} D^2 \bar D_{\dot \beta)}
\cr
&=- 4 i \varepsilon_{\dot \alpha \dot \beta}\partial_{(\alpha}{}^{\dot \gamma}\triangle_{\beta)\dot \gamma}
+ 4 i \varepsilon_{\alpha \beta}\partial^\gamma{}_{(\dot \alpha}\triangle_{ \gamma \dot \beta)}
\end{align}
This operator maps $V\mapsto \widetilde F_{ab}(V)$. The imaginary operator
\begin{align}
-4i \partial_{[{\un a}} \triangle_{{\un b}]} &= \varepsilon_{\dot \alpha \dot \beta} D_{(\alpha} \bar D^2 D_{\beta)} +\varepsilon_{\alpha \beta}\bar D_{(\dot \alpha} D^2 \bar D_{\dot \beta)}\cr
&=-2i \varepsilon_{\dot \alpha \dot \beta}\partial_{(\alpha}{}^{\dot \gamma}\triangle_{\beta)\dot \gamma}
-2i \varepsilon_{\alpha \beta}\partial^\gamma{}_{(\dot \alpha}\triangle_{ \gamma \dot \beta)}
\end{align}
maps $V\mapsto F_{ab}(V)$ (up to factors). Finally we point out that the dimension-3 operator
\begin{align}
D^\gamma \bar D^2 D_\gamma \triangle_{\un a} = -8 \partial^{\un b}\partial_{[{\un a}}\triangle_{{\un b}]}
\end{align}
maps $V$ to the component Maxwell equation. In this sense the operator $D^\gamma \bar D^2 D_\gamma$ maps $A_a \mapsto \partial^b F_{ab}(A)$.
§ LIE ALGEBRA MISCELLANEA
In this appendix, we review some $G_2$ linear algebra [40, 41, 42, 43, 44].
§.§ Linear Algebra
Set $Y = \mathbf R^7$ and let $\varphi \in \Lambda^3(Y)$ be a fixed constant 3-form.
Define the map $s: \Lambda^3(Y) \to S^2(Y)$ from 3-forms to symmetric tensors by
\begin{align}
\label{E:smap}
s(\omega)_{ij} := -\tfrac1{144}
\varepsilon^{k_1\cdots k_7} \varphi_{i k_1k_2} \varphi_{j k_3k_4} \omega_{k_5k_6k_7}
= - \tfrac1{24} \varphi_{i kl} \tilde \omega^{klmn} \varphi_{j mn}
\end{align}
Then $s(\varphi)_{ij}$ is a symmetric bilinear form.
When this form is invertible, $\varphi$ is said to be stable.
A stable 3-form on the tangent spaces of $Y$ reduces the structure group $GL(7) \to G_2$ so that $Y$ is a $G_2$-structure manifold.
In this case, $s_{ij}(\varphi)$ is definite, and we may assume it to be positive-definite (by flipping $\varphi\to -\varphi$, if necessary).
When this is true, we may take it to be [45]
\begin{align}
- \varphi= e^{123} + e^{145}
+e^{167} + e^{246}
-e^{257} - e^{347}
\end{align}
Then $s$ is normalized so that $s(\varphi)_{ij} = \delta_{ij}$.
Define the Hodge dual $\psi : = \ast \varphi$. (Recall $\ast^2 = \mathrm {id}$.)
These tensors satisfy the algebraic identities
\begin{align}
&{\psi}^{ijkl}{\psi}_{i'j'k'l} =
6\delta^{i}_{[i'} \delta^{j}_{j'}\delta^{k}_{k]}
-9 \delta^{[i}_{[i'}{\psi}_{j'k']}{}^{jk]}
\\
{\psi}^{ijkl} {\psi}_{ijk'l'} = 8 \delta_{[k'}^k \delta_{l']}^l - 2 {\psi}_{k'l'}{}^{kl}
\varphi^{ijk} \varphi_{ij'k'} = 2\delta_{[j'}^j\delta_{k']}^k - {\psi}_{j'k'}{}^{jk}
\label{E:contractions7projector}
\\
\label{E:psiphi}
&\psi^{ijk l} \varphi_{i'j' l} = 6 \delta^{[i}_{[i'} \varphi_{j']}{}^{jk]}
\\
\label{E:phiphiclutch}
&\varphi^{ijk} \varphi_{ijk'} = 6\delta_{k'}^k
{\psi}^{ijkl} {\psi}_{ijkl'} = 24 \delta_{l'}^l
\varphi_i{}^{lm} {\psi}_{jk lm} = - 4 \varphi_{ijk}
\\
\label{E:3phis}
&\epsilon^{lmnpqrs} \varphi_{imn}\varphi_{jpq}\varphi_{krs} = -48 \sqrt{g}\delta^l_{(i} g_{jk)}
\\
\varphi_{ijk}\epsilon^{jkl mnpq} = 10 \psi^{[lmnp}\delta^{q]}_i
\psi_{ijkl}\epsilon^{jkl mnpq} = 4! \varphi^{[mnp}\delta^{q]}_i
\end{align}
Under the reduction $GL(7)\to G_2$, the $\mathbf{21}$-dimensional space of 2-forms on $Y$ decomposes into $G_2$ representations as $\mathbf{21}=\mathbf{7}\oplus \mathbf{14}$.
Similarly, the $\mathbf{35}$-dimensional space of 3-forms on $Y$ decomposes as $\mathbf{35}=\mathbf{1}\oplus\mathbf{7}\oplus \mathbf{27}$.
For any $p$-form $\omega$, let $\omega_\mathbf i := \pi_{\bm {i}} \omega$ denote the projection to the $\mathbf{i}$-dimensional representation.
Explicitly, for any 2-form $\eta$ and 3-form $\omega$,
\begin{align}
\label{E:14to7}
\pi_{\bm 7} \eta_{ij} &= \left( \tfrac13\delta_i^k \delta_j^l - \tfrac16{\psi}_{ij}{}^{kl} \right) \eta_{kl}
= \tfrac16 \varphi_{ijk}\varphi^{klm} \eta_{lm}
\label{E:21to14}
\pi_{\bm {14}} \eta_{ij} &= \left( \tfrac23\delta_i^k \delta_j^l +\tfrac16 {\psi}_{ij}{}^{kl} \right) \eta_{kl}
\label{E:35to1}
\pi_{\bm {1}} \omega_{ijk} &= \tfrac1{42} \varphi_{ijk} \varphi^{i'j'k'} \omega_{i'j'k'}
\label{E:35to7}
\pi_{\bm {7}} \omega_{ijk} &= \left(
\tfrac14\delta_i^{i'}\delta_j^{j'}\delta_k^{k'}
-\tfrac38 {\psi}_{[ij}{}^{i'j'}\delta_{k]}^{k'}
-\tfrac1{24}\varphi_{ijk} \varphi^{i'j'k'}\right) \omega_{i'j'k'}
= -\tfrac1{24} \psi_{ijkl}\psi^{lmnp} \omega_{mnp}
\label{E:35to27}
\pi_{\bm {27}} \omega_{ijk} &= \left(
\tfrac34\delta_i^{i'}\delta_j^{j'}\delta_k^{k'}
+\tfrac38 {\psi}_{[ij}{}^{i'j'}\delta_{k]}^{k'}
+\tfrac1{56}\varphi_{ijk} \varphi^{i'j'k'}\right) \omega_{i'j'k'}
\end{align}
The $\bm 7$-projections of 2- and 3-forms play an important role in the gravitino analysis. We define for such projections the vectors fields[That is, for any $\eta \in \Lambda^2(Y)$ and $\omega \in \Lambda^3(Y)$, we are defining the vectors $\vec{\eta}$ and $\vec{\omega}$ on $Y$ such that
\begin{align}
\iota_{\vec{\eta}} \varphi = \pi_{\bm 7} \eta
\iota_{\vec{\omega}} {\psi} = 2 \pi_{\bm 7} \omega
\end{align}
\begin{align}
\label{E:7components2}
\eta^i:=\tfrac16 \varphi^{ijk} \pi_{\bm 7} \eta_{jk} = \tfrac16 \varphi^{ijk} \eta_{jk}
~~~&\Leftrightarrow~~~ \pi_{\bm 7} \eta_{ij} = \varphi_{ijk} \eta^k
\\
\label{E:7components3}
\omega^i := \tfrac1{12} {\psi}^{ijkl} \pi_{\bm 7} \omega_{jkl}= \tfrac1{12}{\psi}^{ijkl} \omega_{jkl}
\pi_{\bm 7} \omega_{ijk} = -\tfrac12 {\psi}_{ijkl} \omega^l
\end{align}
Note that this implies that there are conversion factors in squares
\begin{align}
\label{E:squares}
(\eta^i)^2 = \tfrac16 (\pi_{\bm 7} \eta_{ij})^2
(\omega^i)^2 = \tfrac16 (\pi_{\bm 7} \omega_{ijk})^2
\end{align}
The dual 4-form ${\psi}$ acts on 2-forms as ${\psi}_{ij}{}^{kl} \eta_{\bm 7 kl} = -4 \eta_{\bm 7 ij}$ and ${\psi}_{ij}{}^{kl} \eta_{\bm {14} kl} = 2\eta_{\bm {14} ij}$, or
\begin{align}
\label{E:2formSquare}
{\psi}^{ij kl} \eta_{ij} \eta_{kl}
= -4 \eta_{\bm 7 ij}^2 + 2 \eta_{\bm {14} ij}^2
= -24 (\eta^i)^2 + 2 \eta_{\bm {14} ij}^2
\end{align}
Hitchin metric
The co-calibration with raised indices $\psi^{ijkl} = \tfrac1{3!} \epsilon^{ijklmnp} \varphi_{mnp}$ acts as a(n indefinite) metric on the space of 2-forms on $Y$.
There is also a natural (indefinite) metric on the space of 3-forms resulting from the second variation of the Hitchin functional with respect to $\varphi$
\begin{align}
\hyperlabel{E:HitchinMetric}
\hitchin^{ijk}{}_{lmn}&:= \tfrac12 \delta^{[i}_{[l}\delta^j_m\delta^{k]}_{n]} + \tfrac1{36} \varphi^{ijk}\varphi_{lmn} +\tfrac 34 \delta^{[k}_{[l} \psi^{ij]}{}_{mn]}
\end{align}
to which we will refer as the Hitchin metric.[This definition differs from that of our previous papers by a burdensome factor of 18.]
In terms of projectors (<ref>)|(<ref>), it is given by $\hitchin = - \tfrac43 \pi_{\bm {1}} -\pi_{\bm {7}} +\pi_{\bm {27}}$.
Its partial contractions satisfy
\begin{align}
\varphi_{ijk} \hitchin^{jkl\, mnp} &= - \tfrac43\delta_i^{[l}\varphi^{mnp]}
\\
\psi_{ijkl} \hitchin^{jkl\, mnp} &= - \psi_i{}^{mnp}
\\
\varepsilon_{ijklqrs} \hitchin^{qrs\, mnp}
\tfrac12 \varepsilon_{ijkl}{}^{mnp}
+\tfrac16 \psi_{ijkl}\varphi^{mnp}
+18 \delta_{[i}^{[m}\delta_{j}^{n} \varphi_{kl]}{}^{p]}
\end{align}
Torsion of the $G_2$ structure
As we have reviewed, a stable 3-form $\varphi$ on a smooth 7-manifold $Y$ defines a Riemannian metric $g(\varphi)$ and, therefore, a unique compatible torsion-free connection $\nabla$.
The exterior derivative of the calibration $\nabla_m \varphi_{ijk}$ need not vanish; without loss of generality, it can be parameterized as
\begin{gather}
\nabla_m \varphi_{ijk} = \psi_{ijkl}T^l_m
\end{gather}
in terms of a $7\times 7$ matrix called the torsion of the $G_2$ structure [43].
Under the $G_2$ action, this torsion decomposes into $\bm {49}=\bm {1}+\bm {7}+\bm {14}+\bm {27}$.
Since the connection is torsion-free, antisymmetrizing all indices reduces the left-hand side to the exterior derivative of the calibration.
This is a generic 4-form which may be expanded as
\begin{align}
d \varphi = \tau_0 \psi + 3\tau_1 \wedge \varphi + \ast \tau_3
\end{align}
where $\tau_\mu$ is a $\mu$-form with $\mu=0,1,3$ in the $\bm {1}$-, $\bm {7}$-, and $\bm {27}$-dimensional representations, respectively.
(Evidently the $14$-dimensional part drops out.)
Similarly, the differential of the co-calibration can be expanded as
\begin{align}
d\psi = 4 \tau_1 \wedge \psi + \tau_2 \wedge \varphi
\end{align}
in terms of a 1-form and a 2-form corresponding to $\bm {7}$- and $\bm {14}$-dimensional representations.
The 1-forms in these expansions are the same because $\psi^{ijkl} \partial_{[m} \psi_{ijkl]}$ is proportional to $\varphi^{ijk}\partial_{[m}\varphi_{ijk]}$.
It follows that the $49$ components of the $G_2$ torsion $T_i^j$ may be expressed instead in terms of four $\mu$-forms with $\mu=0,1,2,3$.
Explicitly [43]
\begin{align}
\label{E:Forms2Torsion}
(\tau_0) &= \tfrac{4}7 T_k{}^k
\cr
(\tau_1)_i &= - \tfrac16 \varphi_{ij}{}^k T_k{}^j
\cr
(\tau_2)_{ij} &= -\tfrac43 T_{[i}{}^k\delta_{j]k} +\tfrac13 \psi_{ijk}{}^l T_l{}^k
\cr
(\tau_3)_{ijk} &=
-\tfrac32 (T_{[i}{}^l+T^l{}_{[i})\varphi_{jk]l}
+\tfrac37 \varphi_{ijk} T_l{}^l
\end{align}
Note that $\tau_2$ is in the $\bm {14}$ representation since it is $6\pi_{\bm {14}}$ acting on $T$ considered as a 2-form $T_{[i}{}^k\delta_{j]k}$, and $\tau_3$ is in the $\bm {27}$-dimensional one since the trace of the symmetric part $T_{(i}{}^k\delta_{j)k}$ is projected out.
\begin{align}
\label{E:Torsion2Forms}
T_i{}^j = \tfrac1{4} \delta_i^j (\tau_0)
+\varphi_i{}^{jk} (\tau_1)_k
-\tfrac1{2} (\tau_2)_i{}^j
-\tfrac1{4} \varphi^{jkl} (\tau_3)_{ikl}
\end{align}
Linearized Riemannian metric
The map (<ref>) is $s_{ij}(\omega) = -\frac16 \ast( \varphi_i \wedge \varphi_j \wedge \omega)$, but we can use the identities (<ref>) to give alternative forms for it.
For example, when $\omega$ is obtained by expanding a stable 3-form around the calibration, we obtain the symmetrical expression
\begin{align}
\label{E:smap2}
s(\omega)_{ij} = -\tfrac1{3\cdot 144}\varepsilon^{k_1\cdots k_7}
\left[
\omega_{i k_1k_2} \varphi_{j k_3k_4} \varphi_{k_5k_6k_7}
+\varphi_{i k_1k_2} \omega_{j k_3k_4} \varphi_{k_5k_6k_7}
+\varphi_{i k_1k_2} \varphi_{j k_3k_4} \omega_{k_5k_6k_7}
\right]
\end{align}
This is the same map, because the first and third term together are symmetric and evaluate to the second term (with a factor of 2).
Going the other direction, we can use (<ref>) to show that either form reduces to the more economical
\begin{align}
\label{E:smap3}
s(\omega)_{ij} = \tfrac1{6} \varphi_{(i}{}^{kl}\omega_{j)kl}
\end{align}
In this form, it is also clear from (<ref>) and (<ref>) that $s$ projects out the $\bm 7$ component of $\omega$.
\begin{align}
\label{E:smapKer}
s \circ \pi_{\bm 7}(\omega) \equiv 0
\end{align}
This is the entire kernel, as we conclude by counting dimensions.
Next, define the map $t: \Lambda^1 \otimes \Lambda^\ast_1 \to \Lambda^3$ from vector-valued 1-forms to 3-forms by
\begin{align}
\hyperlabel{E:tmap}
\tmap(M)_{ijk} := \tfrac32 M_{[i}{}^l \varphi_{jk] l}
\end{align}
for any matrix $M \in \Lambda^1 \otimes \Lambda^\ast_1 $.
This matrix space $\Lambda^1 \otimes \Lambda^\ast_1 \simeq \Lambda^2 \oplus S^2$ so its dimension is $\bm{49} = \bm{21} \oplus \bm{28}$ whereas the image is at most $\bm{35}$-dimensional.
This suggests that the $\bm{14}$-dimensional projection of the anti-symmetric part of $M$ does not survive, and it is easily checked that $(\pi_{\bm {14}} \eta)_{[i}{}^l \varphi_{jk] l} \equiv 0$ with the projectors above.
Similarly $(\pi_{\bm {7}} \eta)_{[i}{}^l \varphi_{jk] l} = \eta_{[i}{}^l \varphi_{jk] l}$ as required.
Composing with $s$, we find $s\circ t (M) = - \tfrac13 M_{(i}{}^k\delta_{j)k} - \tfrac16 \delta_{ij} M_k{}^k $, so the symmetric part of $M$ does not come back to itself even up to a coefficient.
This suggests we define a new map
\begin{align}
\hyperlabel{E:hmap}
h_{ij} := 3 s_{ij} -\tfrac13 \delta_{ij} \delta^{kl} s_{kl}
h \circ t (M) = M_{(i}{}^k\delta_{j)k}
\end{align}
Note that this map projects out the $\bm 7$ representation by (<ref>).
The significance of it is that it defines the graviton as a metric fluctuation from the linearized calibration.
§ PREPOTENTIAL TRANSFORMATIONS AND PARTIAL INVARIANTS
In this section, we collect the linearized transformations of the prepotentials (<ref>) found in [16].
They contain abelian transformations of the M-theory 3-form [20], a non-abelian gauging thereof under 7D diffeomorphisms [19], local 4D, $N=1$ superconformal transformations [21], and extensions thereof (gravitino transformations) [16].
Abelian Tensor Hierarchy
Under abelian 2-form symmetry only the prepotentials corresponding to the components of the 3-form transform. Explicitly, [20]
\begin{align}
\delta_{ath} \Phi_{ijk} &=
\\
\delta_{ath} V_{ij} &=
\tfrac1{2i} \left( \Lambda_{ij} - \bar \Lambda_{ij} \right)
-2 \partial_{[i} u_{j]}
\\
\delta_{ath} \Sigma^\alpha_i &=
-\tfrac14 \bar D^2 D^\alpha u_i + \partial_i \Upsilon^\alpha
\\
\delta_{ath} X &=
\tfrac1{2i} \left( D^\alpha \Upsilon_\alpha - \bar D_{\dt \alpha} \bar \Upsilon^{\dt \alpha}\right)
\end{align}
where $\Lambda_{ij}$ is chiral, $u_i$ is real, and $\Upsilon^\alpha$ is chiral. Together they define an eleven-dimensional 2-form.
The field strengths of the abelian tensor hierarchy are
\begin{align}
\hyperlabel{E:G}
\G &= -\tfrac14 \bar D^2 X
\\
\hyperlabel{E:H}
\H_i &= \tfrac1{2i} \left( D^\alpha \Sigma_{\alpha i} - \bar D_{\dt \alpha} \bar \Sigma^{\dt \alpha}_i\right)
-\partial_i X
\\
\hyperlabel{E:W}
\W_{ij}^\alpha &= -\tfrac14 \bar D^2 D^\alpha V_{ij}
+2\partial_{[i} \Sigma_{j]}^\alpha
\\
\hyperlabel{E:F}
\F_{ijk} &= \tfrac1{2i}\left( \Phi_{ijk} - \bar \Phi_{ijk} \right)
-3\partial_{[i} V_{jk]}
\\
\hyperlabel{E:E}
\E_{ijkl} &= ~~4\partial_{[i} \Phi_{jkl]}
\end{align}
Note that $\G$, $\W$, and $\E$ are chiral and that $\H$ and $$ are real.
It is almost manifest that these combinations are invariant under the linearized non-abelian gauge transformations \eqref{E:ATHXf}.
\paragraph{Kaluza-Klein Gauge Field}
The abelian tensor hierarchy was coupled to the non-abelian Kaluza-Klein field gauging the 7D diffeomorphisms in [19].
At the linearized level, this field is described by a real superfield $𝒱^i$ carrying a 7D vector index, and transforming as
\begin{align}
\delta_7 \mathcal V^i = \tfrac1{2i} (\tau^i -\bar \tau^i)
\bar D_{\dt \alpha} \tau^i =0
\end{align}
Its field strength
\begin{align}
\hyperlabel{E:KK}
\KK_\alpha^i &= -\tfrac14 \bar D^2 D_\alpha \mathcal V^i
\end{align}
is invariant and chiral.
\paragraph{Extended 4D Superconformal Transformations}
The 4D, $N=1$ superconformal transformations are parameterized by a spinor parameter superfield $L^α$ [21].
The prepotential $U^a$ containing the conformal part of the 4D frame suffers the pregauge transformation
\begin{align}
\delta U^{\alpha \dt \alpha} := \bar D^{\dt \alpha} L^\alpha - D^\alpha \bar L^{\dt \alpha}
\end{align}
where $U^a = U^αα := (σ̅_a)^αα U^a$ is by definition the contraction of the vector index by the Pauli matrices (this invertible operation is the same as Feynman's slash but with the diacritical mark directly on the index instead of the field).
In calculations, it is useful to note that $D̅ L$ appears instead of $L$ itself, so $L^α$ has a gauge-for-gauge ambiguity corresponding to a shift by a chiral spinor field.
As explained in [46], this parameter must also enter the transformation of the gravitino superfield:
\begin{align}
\delta \Psi_i^{\alpha} = \Xi_i^\alpha + D^\alpha \Omega_i + 2i \partial_i L^\alpha
\end{align}
Here $Ξ$ is chiral and $Ω$ is complex and completely unconstrained. These two parameters are the conformal supersymmetry parameters of the matter gravitino multiplet [21, 47] (see also appendix C of [16] for a review of this multiplet).
In [18], it was observed that the field $X$ in the tensor hierarchy carrying 4D polarizations of the 3-form, was also coupling exactly as the chiral conformal compensator of old-minimal supergravity [32]. This implies that it must transform in a specific way under the $L^α$ part of the superconformal transformations.
It was also conjectured there that all the fields of the tensor hierarchy suffer similar compensating superconformal transformations.
This is needed for the spectrum of component fields to match as summarized in table \ref{T:Components}.
This claim was demonstrated explicitly in [16] where it was shown that under superconformal transformations,
\begin{subequations}
\begin{align}
\delta_{sc} \mathcal V^i &= -\tfrac12( \Omega^i + \bar \Omega^i)
\\
\delta_{sc} X &= D^\alpha L_\alpha + \bar D_{\dt \alpha} \bar L^{\dt \alpha}
\\
\delta_{sc} \Sigma^\alpha_i &= -\Xi^\alpha_i
\\
\delta_{sc} V_{ij} &= \tfrac1{2i} \varphi_{ijk} (\Omega^k - \bar \Omega^k)
\\
\delta_{sc} \Phi_{ijk} &= -\tfrac i2 \psi_{ijkl} \bar D^2 \bar \Omega^l
\end{align}
\end{subequations}
Crucially, many of these transformations are St\"uckelberg shifts---the hallmark of compensating fields [21].
For the field strengths of the non-abelian tensor hierarchy, this gives
\begin{subequations}
\begin{align}
\delta_{sc} \KK_\alpha^i &= \tfrac18 \bar D^2 D_\alpha ( \Omega^i + \bar \Omega^i)
\\
\delta_{sc} \G &= -\tfrac14 \bar D^2 D^\alpha L_\alpha
\\
\delta_{sc} \H_{i}&= -\tfrac1{2i} \left( D^\alpha \Xi_{\alpha i} - \bar D_{\dt \alpha} \bar \Xi_i^{\dt \alpha} \right)
- \partial_i \left( D^\alpha L_{\alpha} + \bar D_{\dt \alpha} \bar L^{\dt \alpha} \right)
\\
\delta_{sc} \W^\alpha_{ij} &=
\tfrac i8 \varphi_{ijk} \bar D^2 D^\alpha (\Omega^k-\bar\Omega^k)
- 2 \partial_{[i} \Xi^\alpha_{j]}
\\
\delta_{sc} \F_{ijk} &= -\tfrac14 \psi_{ijkl} \left( D^2 \Omega^l + \bar D^2 \bar \Omega^l \right)
+\tfrac {3i}2 \varphi_{l[ij} \partial_{k]} (\Omega^l-\bar\Omega^l)
\\
\delta_{sc} \E_{ijkl} &= - 2i \psi_{m[ijk} \partial_{l]}\bar D^2 \bar \Omega^m
\end{align}
\end{subequations}
\paragraph{Partial Invariants}
These transformations can be removed by combining them with the superfields $U^a$ and $Ψ_i^α$ requiring the compensation.
We do this step-wise, first defining partially-invariant building blocks.
A convenient set is
\begin{subequations}
\begin{align}
\hyperlabel{E:X}
\X_i^{\alpha \dt \alpha} &:= \tfrac 1{2i} ( \bar D^{\dt \alpha} \Psi_i^\alpha
+ D^\alpha \bar \Psi_i^{\dt \alpha} ) - \partial_i U^{\alpha \dt \alpha}
\\
\hyperlabel{E:T}
\T_i &:= \tfrac1{2i} ( D^\alpha \Psi_{\alpha i} - \bar D_{\dt \alpha} \bar \Psi_i^{\dt \alpha} ) + \H_i
\\
\hyperlabel{E:hatW}
\hatW_{ij}^\alpha &:= \W_{ij}^\alpha +2\partial_{[i} \Psi_{j]}^\alpha
\\
\hyperlabel{E:J}
\J_\alpha^i &:=\KK^i_\alpha - \tfrac i2\varphi^{i jk} \hatW_{\alpha jk}
+ \tfrac i{12} \psi^{i jkl} \spinorYYY_{\alpha jkl}
\\
\hyperlabel{E:Z}
\Z_{ijk}
&:= \F_{ijk} -\tfrac i2 \psi_{ijkl} \T^l
- 3i \varphi_{l[ij} \partial_{k]} \mathcal V^l
\end{align}
\end{subequations}
These combinations are invariant under the transformations generated by the $L^α$ and $Ξ_i^α$ parameters:
\begin{subequations}
\begin{align}
\delta \X_{\alpha \dt \alpha}^i &= \tfrac1{2i} \left( \bar D_{\dt \alpha} D_\alpha \Omega^i
+ D_\alpha \bar D_{\dt \alpha} \bar\Omega^i
\right)
\\
\delta \T^{i}&= \tfrac1{2i} \left( D^2 \Omega^i - \bar D^2 \bar \Omega^i \right)
\\
\delta \hatW^\alpha_{ij} &=
\tfrac i8 \varphi_{ijk} \bar D^2 D^\alpha (\Omega^k-\bar\Omega^k)
+ 2 \partial_{[i} D^\alpha \Omega_{j]}
\\
\delta \J_\alpha^i &= \tfrac14 \bar D^2 D_\alpha \left( 2 \Omega^i -\bar \Omega^i\right)
\\
\delta \Z_{ijk} &= - \tfrac 12 \psi_{ijkl} D^2 \Omega^l
+ 3i \varphi_{l[ij}\partial_{k]}( \Omega^l -i \bar \tau^l)
\end{align}
\end{subequations}
The first of these is a partial covariantization of the vector component of $Ψ$ by the derivative of $U^a$. Alternatively, it is the partial covariantization of the derivative of $U^a$.\footnote{The other two combinations cannot be covariantized under the $L^\alpha$ transformation in this way, because there are no fields analogous to $U^a$ and $X$ transforming into $\bar D \Psi - D\bar \Psi$ and $DL - \bar D\bar L$.} The second is an analog of this for the scalar component of $Ψ$ or the derivative of the compensator $X$.
$W$ is just the covariantization of $$ with respect to the $Ξ$ parameter.
Next, $$ is a combination of the spin-1/2 fields in the $7$-dimensional representation of $G_2$: the KK-ino, the abelian gaugino, and the tensorino.
Finally, $$ is the complex combination of scalars appearing in the tensorino $^α_ijk$ \eqref{E:spinorYYY} covariantized by the derivative of the KK prepotential $𝒱^i$. Note that it is the only combination transforming under the $τ^i$ parameter.
Additionally, we define combinations
\begin{subequations}
\begin{align}
\hyperlabel{E:S}
\S &:=\tfrac12(\G+\bar \G) -\tfrac14 [D_\alpha \bar D_{\dt \alpha} ] U^{\un a}
\\
\hyperlabel{E:P}
\P &:=\tfrac 1{2i} (\G-\bar \G) + \partial_a U^a
\end{align}
\end{subequations}
transforming {\em only} under $L^α$
\begin{subequations}
\begin{align}
\delta \S &
= \tfrac38 ( D^\alpha \bar D^2 L_\alpha + \bar D_{\dt \alpha} D^2 \bar L^{\dt \alpha} )
= -\tfrac32 ( D^\alpha \varepsilon_\alpha + \bar D_{\dt \alpha} \bar \varepsilon^{\dt \alpha} )
\\
\label{E:NNMxF}
\delta \P &
= \tfrac i8 ( D^\alpha \bar D^2 L_\alpha - \bar D_{\dt \alpha} D^2 \bar L^{\dt \alpha} )
= \tfrac 1{2i} ( D^\alpha \varepsilon_\alpha - \bar D_{\dt \alpha} \bar \varepsilon^{\dt \alpha} )
\end{align}
\end{subequations}
These are essentially a slight rewriting of the 4D, $N=1$ chiral compensator $$ that transforms nicely (as new-minimal and virial compensators [48, 49, 50]).
\paragraph{4-form Field Strength}
Taking combinations of superspace derivatives of these partial invariants, we can construct superfields that are invariant at the linearized level under consideration (covariant at non-linear order).
For example, since the 4D, $N=1$ supersymmetry parameter $ε^α= -14 D̅^2 L^α$ is chiral, the chiral superfield $= -16 D̅^2 $, is a dimension-1 invariant.
In fact, it is the linearized chiral scalar curvature invariant of old-minimal supergravity that contains the complex auxiliary scalar as the leading component in a Taylor expansion and the four-dimensional curvature scalar at higher order [21, 22].
In the modification of old-minimal supergravity employed by eleven-dimensional supergravity, the prepotential of the chiral compensator is real. The effect of this is that one of the real components of the complex auxiliary field becomes (the 4D Hodge dual of) the 4-form curl of a gauge 3-form [31, 32].
Continuing to make other dimension-1 invariants in this way, we find (among other things) the following combinations
\begin{subequations}
\label{E:4formInvariants}
\begin{align}
\hyperlabel{E:FSXXXX}
\FSXXXX_{abcd}&=3i \epsilon_{abcd} (\R - \bar \R)
\\
\hyperlabel{E:FSXXXY}
\FSXXXY_{abc\, i} &:= \tfrac14 \epsilon_{abc\un d} \left[
(D^2 +\bar D^2) \X_{i}^{\un d}
-[D^\delta, \bar D^{\dt \delta}]\T_i
\right]
\\
\tilde \FSXXXY^{\un a}_{i} &:= \tfrac12 [D^\alpha , \bar D^{\dt \alpha}]\T_i - \tfrac12 (D^2 +\bar D^2) \X_{i}^{\un a}
\\
\hyperlabel{E:FSXXYY}
\FSXXYY_{\alpha \beta ij} &:=
- \tfrac i2 D_{(\alpha} \hatW_{\beta)ij}
+ \tfrac12 \varphi_{ijk} \partial_{(\beta}{}^{\dt \alpha} \X_{\alpha) \dt \alpha}^k
\\
\hyperlabel{E:FSXYYY}
\FSXYYY_{\un a ijk}
&:= \tfrac i2 \left[ \bar D_{\dt \alpha} \spinorYYY_{\alpha ijk}
+D_\alpha \bar \spinorYYY_{\dt \alpha ijk}\right]
-3 \varphi_{l [ij} \partial_{k]} \X_{\un a}^l
\cr
&= -\tfrac12 \left[ \bar D_{\dt \alpha} D_\alpha \Z_{ijk}
-D_\alpha \bar D_{\dt \alpha} \bar \Z_{ijk} \right]
-3 \varphi_{l [ij} \partial_{k]} \X_{\un a}^l
\cr
&=\tfrac12 [D_\alpha, \bar D_{\dt \alpha} ] \F_{ijk}
+ \tfrac12 \psi_{ijkl} \partial_{\un a} \T^l
- 3 \varphi_{l[ij} \partial_{k]} ( \X_{\un a}^l - \partial_{\un a} \mathcal V^l)
\\
\hyperlabel{E:FSYYYY}
\FSYYYY_{ijkl} &:= 2 \partial_{[i} \left[ \Phi_{jkl]} + \bar \Phi_{jkl]} + \psi_{jkl] m}\T^m \right]
\end{align}
\end{subequations}
It can be checked that the new superfields represent the dual of the seven 3-forms $_abc i = ϵ_abcd ^d_i$, the 21 2-forms $_abij = (σ_ab)_α^β_ijβ^α+$ h.c., the 35 1-forms and the 35 0-form field strengths.
Instead of doing this, we explicitly construct the gauge 3-form in section \ref{S:11DSuperfields} (cf.\ eq.\ \ref{E:gauge3form}).
Taking the bosonic curl of these components gives an alternative (equivalent) derivation of the dimension-1 4-form invariants.
There, we also define the other eleven-dimensional connections (\ref{E:11Dframe} and \ref{E:grino}) and, through those, the torsion and curvature invariants.
The off-shell spectrum is completed by auxiliary fields that arise by acting on the connections with supersymmetry transformations.
\section{Gravitino Descendants}
\label{S:GrinoDescendants}
The manifest supersymmetry transformations act in superspace by translations in the fermionic directions.
In this section, we present the result of acting with the fermionic derivatives on the gravitino superfields \eqref{E:grino}.
The process that gives rise to this result is not algorithmic: The desired result is obtained only up to unknown field-dependent gauge transformation and spin connection terms, and in a form that is not easily parsed into 4-form terms and combinations of auxiliary fields.
We present the solution to this superspace crossword puzzle as irreducible projections of $D$ and $D̅$ on each of the four gravitino polarizations: The 1-form index may lie along 4D (X) or 7D (Y), and the spinor may have a $G_2$-singlet index (X) or a $7$-dimensional one (Y).
The fermionic derivatives of the gravitino with polarization XX may be decomposed as
%%% \grinoXX descendants
\begin{subequations}
\begin{align}
D^{(\gamma} \grinoXX_{\un a}{}^{\beta)}
&=\omega_{\un a}{}^{\beta \gamma}
-\tfrac i2 \delta_\alpha^{(\beta}\varepsilon^{\gamma)\delta} \dX_{\delta\dt \alpha}
\\
D_{\beta} \grinoXX_{\un a}{}^{\beta}
&= \partial_{\un a} N + i \dX_{\un a}
\\
\bar D^{\dt \beta} \grinoXX_{\un a}{}^{\beta}
&= i\delta_\alpha^{ \beta}\delta_{\dt \alpha}^{\dt \beta} \R
\end{align}
\end{subequations}
The YX descendants are
%%% \grinoYX descendants
\begin{subequations}
\begin{align}
D^{(\gamma} \grinoYX_i{}^{\beta)} &=
\omega_i{}^{\beta \gamma}
- \tfrac i6 \t_i^{\beta \gamma}
\\
D_{\beta} \grinoYX_{i}{}^{\beta}
&=\partial_i N + i \dY_i
\\
\bar D^{\dt \beta} \grinoYX_{i}{}^{\beta }
&= -\tfrac i{12} \Ga_i^{\un b}
\end{align}
\end{subequations}
The XY gravitino is messier:
%%% \grinoXY descendants
\begin{subequations}
\begin{align}
D^{(\gamma} \grinoXY_{\un a}{}^{\beta)j}
-\tfrac{2i}3 \delta_\alpha^{(\beta}\varepsilon^{\gamma)\delta} \bar \Ga^j_{\delta \dt \alpha}
\\
\label{E:DgrinoXY}
D_{\beta} \grinoXY_{\un a}{}^{\beta j}
\partial_{\un a} N^j
+ 2\varphi^j{}_{kl} \omega_{\un a}{}^{kl}
+\tfrac i2\Ga^j_{\un a}
+\tfrac{5i}6 \bar \Ga^j_{\un a}
+\tfrac1{2} \big[
\bar D_{\dt \alpha} \l^j_{\alpha}
- D_{\alpha} \bar \l^j_{\dt \alpha }
\big]
\\
\bar D^{\dt \beta} \grinoXY_{\un a}{}^{\beta j}
\partial_{\un a} N^{\un b \,j}
+ 2 \omega_{\un a}{}^{\un b \,j}
+2i \delta_{\dt \alpha}^{\dt \beta} \t^j_{\alpha}{}^{\beta}
+\tfrac {2i}3 \delta_\alpha^\beta \bar \t^j_{\dt \alpha}{}^{\dt \beta}
+2i \delta_\alpha^\beta \delta_{\dt \alpha}^{\dt \beta} \dY^j
\end{align}
\end{subequations}
Finally, the YY part gives
%%% \grinoYY descendants
\begin{subequations}
\begin{align}
D^{(\gamma} \grinoYY_{i}{}^{\beta)j} &=
- {2i} \pi_{\bm {14}}( \FSXXYY^{\gamma\beta}{}_i{}^j)
\\
D_{\beta} \grinoYY_{i}{}^{\beta j}
&= \partial_i N^j
+ 2\varphi^j{}_{kl} \, \omega_i{}^{kl}
\\
\bar D^{\dt \beta} \grinoYY_{i}{}^{\beta j}
\partial_{i} N^{\un b j}
+2\omega_i{}^{\un b j}
- i\gravitonYY_i{}^j (\FSXYYY^{\un b}{}_{klm} )
+\tfrac 1{6} \varphi_i{}^{jk}\left[
\bar D^{\dt \beta} \l_k^\beta
+i\Ga_k^{\un b}
\right]
\end{align}
\end{subequations}
In this derivation, the field-dependent Wess-Zumino gauge transformations are
\begin{align}
N &:= \tfrac 13 \S-i\P
\cr
&:= \tfrac16 \psi^{ijkl} \left[ \F_{jkl} -\tfrac i2 \psi_{jklm}\T^m\right]
\cr
N^{\un a j} &=
-\gravitonXY^{\un a j}
-2i \X^{\un a j}
-\bar D^{\dt \alpha} \Psi^{\alpha j}
+ D^{\alpha} \bar \Psi^{\dt \alpha j}
\end{align}
We will not need this explicit form, but it is important that the normalizations of the shared parameters agree across equations
(and is an important technical aid in determining the correct decomposition of the descendants).
The spin connections are given in terms of the frame fields as
\begin{subequations}
\begin{align}
\omega_{\un c}{}^{\un a\un b}&=\partial^{[\un a} h^{\un b]}_{\un c}
-\partial_{\un c} e^{[\un a\un b]}
\\
\hyperlabel{E:spinYXX}
\spinYXX_k{}^{\alpha \beta} &=
-\tfrac14 \partial^{(\alpha}{}_{\dt \beta} \gravitonXY_k^{\beta) \dt \beta}
-\tfrac18 \partial_k [D^{(\alpha} , \bar D_{\dt \beta} ] U^{\beta)\dt \beta}
=\tfrac14 \varepsilon_{\dt \alpha \dt \beta} \left[ \partial^{[\un a} h_k^{\un b]}
-\partial_k e^{[\un a\un b]}
\right]
\\
\omega_{\un c}{}^{\un a \,j} &= \tfrac12 \partial^{\un a} \gravitonXY_{\un c}^j
-\tfrac12 \partial^j \gravitonXX_{\un c}^{\un a}
\\
\omega_i{}^{\un b j} &=\tfrac12 \partial^{\un b} \gravitonYY_i{}^{j}
-\tfrac12 \partial^j \gravitonXY_i{}^{\un b}
\\
\omega_{\un c}{}^{ij} &= \partial^{[i} \gravitonXY_{\un c}^{j]}
\\
\omega_k{}^{ij} &= \partial^{[i} \gravitonYY_k^{j]}
\end{align}
\end{subequations}
These transform as $δω_c^ab = ∂_c λ^ab$ with local Lorentz parameters
\begin{align}
\label{E:localLorentz}
\lambda_{\un a}{}^{\un b} &:=\tfrac 12\left[
\delta_{\dt \alpha}^{\dt \beta} D^{(\beta} \bar D^2 L_{\alpha)}
- \delta_\alpha^\beta \bar D^{(\dt \beta} D^2 \bar L_{\dt \alpha)}
\right]
= -2 \left[
\delta_{\dt \alpha}^{\dt \beta} D^{(\beta} \epsilon_{\alpha)}
- \delta_\alpha^\beta \bar D^{(\dt \beta} \bar \epsilon_{\dt \alpha)}
\right]
\cr
\lambda_{\un a}{}^j &:= -\tfrac12 \left[
\bar D_{\dt \alpha} D_\alpha \Omega^j
- D_\alpha \bar D_{\dt \alpha} \bar \Omega^j
\right]
\cr
\lambda_i{}^{\un b} &:=
- \delta_{ij} \varepsilon^{\beta\alpha} \varepsilon^{\dt \beta \dt \alpha} \lambda_{\un a}{}^j
\cr
\lambda_i{}^j &= \tfrac14 \varphi_i{}^{jk} ( D^2 \Omega_k + \bar D^2 \bar \Omega_k)
+\tfrac i4 \psi_i{}^{jkl} \partial_k ( \Omega_l - \bar \Omega_l )
\end{align}
In these equations $Ω^j$ should be interpreted as the quantity $Ω^j -i τ̅$.
This is gauge(-for-gauge)-equivalent to $Ω^j$ but properly takes into account the non-abelian gauge transformation.
% {\footnotesize
% \bibliography{/Users/wdlinch3/Dropbox/Rashoumon/LaTeX/BibTex/BibTex}
% \bibliographystyle{unsrt}
% }
\begin{thebibliography}{10}
[1]
W.~Siegel and M.~Ro{\v c}ek.
\newblock {On Off-shell Supermultiplets}.
\newblock {\em Phys.Lett.}, B105:275, 1981.
\newblock \href{http://inspirehep.net/record/164156?ln=en}{[{\sc in}SPIRE
[2]
A.~S. Galperin, E.~A. Ivanov, V.~I. Ogievetsky, and E.~S. Sokatchev.
\newblock {\em {Harmonic superspace}}.
\newblock Cambridge Monographs on Mathematical Physics. Cambridge University
Press, 2007.
[3]
A.~Galperin, E.~Ivanov, S.~Kalitsyn, V.~Ogievetsky, and E.~Sokatchev.
\newblock {Unconstrained N=2 Matter, Yang-Mills and Supergravity Theories in
Harmonic Superspace}.
\newblock {\em Class.Quant.Grav.}, 1:469--498, 1984.
\newblock \href{http://inspirehep.net/record/202528?ln=en}{[{\sc in}SPIRE
[4]
U.~Lindstr\"om and M.~Ro{\v c}ek.
\newblock {$N=2$ Superyang-mills Theory in Projective Superspace}.
\newblock {\em Commun.Math.Phys.}, 128:191, 1990.
\newblock \href{http://inspirehep.net/record/278888?ln=en}{[{\sc in}SPIRE
[5]
A.S. Galperin, Paul~S. Howe, and K.S. Stelle.
\newblock {The Superparticle and the Lorentz group}.
\newblock {\em Nucl. Phys. B}, 368:248--280, 1992.
[6]
Francois Delduc, Alexander Galperin, and Emery Sokatchev.
\newblock {Lorentz harmonic (super)fields and (super)particles}.
\newblock {\em Nucl. Phys. B}, 368:143--171, 1992.
[7]
Nathan Berkovits.
\newblock {Super Poincare covariant quantization of the superstring}.
\newblock {\em JHEP}, 0004:018, 2000.
\newblock \href{http://arXiv.org/abs/hep-th/0001035}{[hep-th/0001035]}.
[8]
Martin Cederwall.
\newblock {D=11 supergravity with manifest supersymmetry}.
\newblock {\em Mod. Phys. Lett.}, A25:3201--3212, 2010.
\newblock \href{http://arxiv.org/abs/1001.0112}{[arXiv:1001.0112]}.
[9]
Nathan Berkovits and Max Guillen.
\newblock {Equations of motion from Cederwall\textquoteright{}s pure spinor
superspace actions}.
\newblock {\em JHEP}, 08:033, 2018.
[10]
Warren Siegel.
\newblock {Fields}.
\newblock 1999.
\newblock \href{http://arXiv.org/abs/hep-th/9912205}{hep-th/9912205}.
[11]
Stanley Mandelstam.
\newblock {Light Cone Superspace and the Ultraviolet Finiteness of the N=4
\newblock {\em Nucl. Phys.}, B213:149--168, 1983.
\newblock \href{http://inspirehep.net/record/179486?ln=en}{[{\sc in}SPIRE
[12]
Neil Marcus, Augusto Sagnotti, and Warren Siegel.
\newblock {Ten-dimensional Supersymmetric {Yang-Mills} Theory in Terms of
Four-dimensional Superfields}.
\newblock {\em Nucl. Phys.}, B224:159, 1983.
\newblock \href{http://inspirehep.net/record/189742}{[{\sc in}SPIRE entry]}.
[13]
Nathan Berkovits.
\newblock {A Ten-dimensional superYang-Mills action with off-shell
\newblock {\em Phys. Lett.}, B318:104--106, 1993.
\newblock \href{https://arxiv.org/abs/hep-th/9308128}{[arXiv:hep-th/9308128]}.
[14]
Jonathan~M. Evans.
\newblock {Supersymmetry algebras and Lorentz invariance for d = 10
\newblock {\em Phys. Lett. B}, 334:105--112, 1994.
[15]
Katrin Becker, Melanie Becker, Daniel Butter, and William~D. Linch.
\newblock {$N=1$ supercurrents of eleven-dimensional supergravity}.
\newblock {\em JHEP}, 05:128, 2018.
\newblock \href{https://arxiv.org/abs/1803.00050}{[arXiv:1803.00050]}.
[16]
Katrin Becker, Melanie Becker, Daniel Butter, Sunny Guha, William~D. Linch, and
Daniel Robbins.
\newblock {Eleven-Dimensional Supergravity in 4D, $N=1$ Superspace}.
\newblock {\em JHEP}, 11:199, 2017.
\newblock \href{https://arxiv.org/abs/1709.07024}{[arXiv:1709.07024]}.
[17]
Katrin Becker, Melanie Becker, William~D. Linch, III, Stephen Randall, and
Daniel Robbins.
\newblock {All Chern-Simons Invariants of 4D, N = 1 Gauged Superform
\newblock {\em JHEP}, 04:103, 2017.
\newblock \href{https://arxiv.org/abs/1702.00799}{[arXiv:1702.00799]}.
[18]
Katrin Becker, Melanie Becker, Sunny Guha, William~D. Linch~III, and Daniel
\newblock {M-theory potential from the $G_{2}$ Hitchin functional in
\newblock {\em JHEP}, 12:085, 2016.
\newblock \href{https://arxiv.org/abs/1611.03098}{[arXiv:1611.03098]}.
[19]
Katrin Becker, Melanie Becker, William~D. Linch~III, and Daniel Robbins.
\newblock {Chern-Simons actions and their gaugings in 4D, $N =$ 1 superspace}.
\newblock {\em JHEP}, 06:097, 2016.
\newblock \href{https://arxiv.org/abs/1603.07362}{[arXiv:1603.07362]}.
[20]
Katrin Becker, Melanie Becker, William~D. Linch~III, and Daniel Robbins.
\newblock {Abelian tensor hierarchy in 4D, $N = 1$ superspace}.
\newblock {\em JHEP}, 03:052, 2016.
\newblock \href{http://arxiv.org/abs/1601.03066}{[arXiv:1601.03066]}.
[21]
S.~James Gates~Jr., Marcus~T. Grisaru, M.~Ro{\v c}ek, and W.~Siegel.
\newblock {\em {Superspace Or One Thousand and One Lessons in Supersymmetry}}.
\newblock Frontiers in Physics, 58. Benjamin/Cummings, 1983.
\newblock \href{http://arXiv.org/abs/hep-th/0108200}{[hep-th/0108200]}.
[22]
J.~Wess and J.~Bagger.
\newblock {\em {Supersymmetry and supergravity}}.
\newblock Princeton, USA: Univ. Pr. (1992) 259 p, 1992.
\newblock \href{http://inspirehep.net/record/350988?ln=en}{[{\sc in}SPIRE
[23]
I.L. Buchbinder and S.M. Kuzenko.
\newblock {\em {Ideas and methods of supersymmetry and supergravity: Or a walk
through superspace}}.
\newblock Bristol, UK: IOP (1998) 656 p, 1998.
\newblock \href{http://inspirehep.net/record/485478?ln=en}{[{\sc in}SPIRE
[24]
Paul~S. Howe.
\newblock {Weyl superspace}.
\newblock {\em Phys. Lett. B}, 415:149--155, 1997.
\newblock \href{https://inspirehep.net/literature/446234}{[{\sc in}SPIRE entry]}.
[25]
Martin Cederwall, Ulf Gran, Mikkel Nielsen, and Bengt~E.W. Nilsson.
\newblock {Manifestly supersymmetric M-theory}.
\newblock {\em JHEP}, 0010:041, 2000.
\newblock \href{https://inspirehep.net/literature/529792}{[{\sc in}SPIRE entry]}.
[26]
Hitoshi Nishino and S.~James {Gates, Jr.}
\newblock {Toward an off-shell 11D supergravity limit of M-theory}.
\newblock {\em Phys. Lett. B}, 388:504--511, 1996.
\newblock \href{https://inspirehep.net/literature/415718}{[{\sc in}SPIRE entry]}.
[27]
S.~James {Gates, Jr.} and Hitoshi Nishino.
\newblock {Deliberations on 11D superspace for the M-theory effective action}.
\newblock {\em Phys. Lett. B}, 508:155--167, 2001.
\newblock \href{https://inspirehep.net/literature/551915}{[{\sc in}SPIRE entry]}.
[28]
S.~James {Gates, Jr.}
\newblock {Superconformal symmetry in 11D superspace and the M-theory effective action}.
\newblock {\em Nucl. Phys. B}, 616:85--105, 2001.
\newblock \href{https://inspirehep.net/literature/558607}{[{\sc in}SPIRE entry]}.
[29]
Katrin Becker and Daniel Butter.
\newblock {4D $N=1$ Kaluza-Klein superspace}.
\newblock {\em JHEP}, 09:091, 2020.
\newblock \href{http://arXiv.org/abs/2003.01790}{[arXiv:2003.01790]}.
[30]
E.~Cremmer and B.~Julia.
\newblock {The N=8 Supergravity Theory. 1. The Lagrangian}.
\newblock {\em Phys.Lett.}, B80:48, 1978.
\newblock \href{http://inspirehep.net/record/131572?ln=en}{[{\sc in}SPIRE
[31]
S.~James Gates~Jr.
\newblock {Super $p$-Form Gauge Superfields}.
\newblock {\em Nucl.Phys.}, B184:381, 1981.
\newblock \href{http://inspirehep.net/record/9990?ln=en}{[{\sc in}SPIRE
[32]
Burt~A. Ovrut and Daniel Waldram.
\newblock {Membranes and three form supergravity}.
\newblock {\em Nucl. Phys.}, B506:236--266, 1997.
\newblock \href{https://arxiv.org/abs/hep-th/9704045}{[arXiv:hep-th/9704045]}.
[33]
\newblock {Torsion constraints in supergeometry}.
\newblock {\em Commun.Math.Phys.}, 133:563--615, 1990.
\newblock \href{http://inspirehep.net/record/305705?ln=en}{[{\sc in}SPIRE
[34]
John Lott.
\newblock {The Geometry of supergravity torsion constraints}.
\newblock 2001.
\newblock \href{http://arxiv.org/abs/math/0108125}{[arXiv:math/0108125]}.
[35]
Lorenz~J. Schwachh{\"o}fer.
\newblock {\em Global Differential Geometry}.
\newblock Springer Proceedings in Mathematics 17. Springer-Verlag Berlin
Heidelberg, 1 edition, 2012.
[36]
Jos{\'e} Figueroa-O'Farrill and Andrea Santi.
\newblock {Spencer cohomology and 11-dimensional supergravity}.
\newblock {\em Commun. Math. Phys.}, 349(2):627--660, 2017.
[37]
Katrin Becker, Melanie Becker, Daniel Butter, William~D. Linch, and Stephen
\newblock {Five-dimensional Supergravity in $N = 1/2$ Superspace}.
\newblock {\em JHEP}, 03:098, 2020.
\newblock \href{https://arxiv.org/abs/1909.09208}{[arXiv:1909.09208]}.
[38]
S.~James Gates~Jr., William D.~Linch III, and J.~Phillips.
\newblock {Field strengths of linearized 5D, $\mathcal N=1$ superfield
supergravity on a 3-brane}.
\newblock {\em JHEP}, 0502:036, 2005.
\newblock \href{http://arxiv.org/abs/hep-th/0311153}{[hep-th/0311153]}.
[39]
J.~Wess and B.~Zumino.
\newblock {Superspace Formulation of Supergravity}.
\newblock {\em Phys. Lett.}, B66:361--364, 1977.
\newblock \href{http://inspirehep.net/record/123994?ln=en}{[{\sc in}SPIRE
[40]
D.D. Joyce.
\newblock {\em Compact Manifolds with Special Holonomy}.
\newblock Oxford mathematical monographs. Oxford University Press, 2000.
[41]
Nigel~J. Hitchin.
\newblock {The Geometry of Three-Forms in Six and Seven Dimensions}.
\newblock {\em J. Diff. Geom.}, 55(3):547--576, 2000.
\newblock \href{http://arxiv.org/abs/math/0010054}{[arXiv:math/0010054]}.
[42]
Nigel~J. Hitchin.
\newblock {Stable Forms and Special Metrics}.
\newblock 2001.
\newblock \href{http://arxiv.org/abs/math/0107101}{[arXiv:math/0107101]}.
[43]
R.~L. {Bryant}.
\newblock {Some remarks on $G_2$-structures}.
\newblock {\em ArXiv Mathematics e-prints}, May 2003.
\newblock \href{https://arxiv.org/abs/math/0305124}{[arXiv:math/0305124]}.
[44]
Spiro Karigiannis.
\newblock {\em Deformations of ${G}_2$ and ${Spin}(7)$ Structures on
\newblock PhD thesis, 2003.
\newblock \href{https://arxiv.org/abs/math/0301218}{[arXiv:math/0301218]}.
[45]
Sergey Grigorian.
\newblock {Moduli spaces of G(2) manifolds}.
\newblock {\em Rev. Math. Phys.}, 22:1061--1097, 2010.
\newblock \href{https://arxiv.org/abs/0911.2185}{[arXiv:0911.2185]}.
[46]
William D.~Linch III, Markus~A. Luty, and J.~Phillips.
\newblock {Five-dimensional supergravity in $\mathcal N=1$ superspace}.
\newblock {\em Phys.Rev.}, D68:025008, 2003.
\newblock \href{http://arxiv.org/abs/hep-th/0209060}{[hep-th/0209060]}.
[47]
S.~James Gates~Jr. and V.~Alan Kostelecky.
\newblock {Supersymmetric Matter Gravitino Multiplets}.
\newblock {\em Nucl. Phys.}, B248:570--588, 1984.
\newblock \href{http://inspirehep.net/record/15327}{[{\sc in}SPIRE entry]}.
[48]
I.~L. Buchbinder, S.~James Gates~Jr., William~D. Linch~III, and J.~Phillips.
\newblock {New 4D, $N=1$ superfield theory: Model of free massive superspin-3/2
\newblock {\em Phys. Lett.}, B535:280--288, 2002.
\newblock \href{http://arxiv.org/abs/hep-th/0201096}{[arXiv:hep-th/0201096]}.
[49]
S.~James Gates~Jr., Sergei~M. Kuzenko, and J.~Phillips.
\newblock {The Off-shell (3/2, 2) supermultiplets revisited}.
\newblock {\em Phys.Lett.}, B576:97--106, 2003.
\newblock \href{https://arxiv.org/abs/hep-th/0306288}{[arXiv:hep-th/0306288]}.
[50]
\newblock {Imaginary supergravity or Virial supergravity?}
\newblock {\em Nucl. Phys.}, B892:288--305, 2015.
\newblock \href{https://arxiv.org/abs/1411.1057}{[arXiv:1411.1057]}.
\end{thebibliography}
\end{document}
|
# Integrable hierarchy for the resolved conifold
Murad Alim and Arpan Saha Fachbereich Mathematik, Universität Hamburg,
Bundesstr. 55, 20146, Hamburg
###### Abstract.
We provide a direct proof of a conjecture of Brini relating the Gromov–Witten
theory of the resolved conifold to the Ablowitz–Ladik integrable hierarchy at
the level of primaries. We use a functional representation of the
Ablowitz–Ladik hierarchy as well as a difference equation for the
Gromov–Witten potential. We express the solution of the difference equation as
an analytic function which is a specialization of a Tau function put forward
by Bridgeland in the study of wall-crossing phenomena of Donaldson–Thomas
invariants.
## 1\. Introduction
The study of integrable structures as the underlying organizing principle of
Gromov–Witten (GW) theory has been a great source of insights and interactions
between various areas of mathematics such as algebraic and symplectic
geometry, the study of integrable systems, and mathematical physics.
The KdV integrable hierarchy for GW theory of a point was put forward in the
works of Witten [Wit91] and Kontsevich [Kon92], relating the intersection
theory on the Deligne–Mumford compactification of the moduli space of curves
to $2$d topological gravity. For $\mathbb{P}^{1}$, the relation of GW theory
to the Toda integrable hierarchy was studied in [EHY95, EY94, EYY95, EHX97,
Pan00]. In [Pan00], a conjectured difference equation for the GW potential
implied a corresponding difference equation for Hurwitz numbers which was
proved in [Oko00].
A general construction of an integrable hierarchy for a given Gromov–Witten
theory was given by Dubrovin and Zhang in [DZ01], see also [Dub14] for an
overview, yet explicit examples remain sparse, see also [ADK+06] for a
relation between topological string theory and integrable hierarchies. For the
Calabi–Yau threefold geometry given by the resolved conifold, Brini
conjectured in [Bri12] that the equivariant descendent Gromov–Witten potential
of the resolved conifold with anti-diagonal $\mathbb{C}^{\times}$ action
coincides with the logarithm of a tau function of the Ablowitz–Ladik (AL)
hierarchy introduced in [AL75] under a suitable identification of the
variables. In particular, the conjecture implies that the _primary_ GW
potential of the resolved conifold (i. e. without gravitational descendants)
coincides with the restriction of a tau function of the AL hierarchy to the
small phase space.
In [Bri12], Brini provided a proof of the coincidence of the descendent
Gromov–Witten potential with the logarithm of an AL tau function to genus $1$,
and of the primary Gromov–Witten potential with the restriction of the
logarithm of the tau function to the small phase space up to genus $2$. A
proof of the conjecture in the stationary sector to all genera was
subsequently given by Takasaki in [Tak13] using a melting crystal model of
Gromov–Witten theory of the resolved conifold. This was essentially a two-step
proof: First, the descendent potential was identified with a tau function of
the $2$d Toda hierarchy. And then, it was shown that owing to a certain
factorization property which the specific tau function in question possessed,
it could be interpreted as a tau function of the relativistic Toda hierarchy,
a reduction of the $2$d Toda hierarchy which was known to be equivalent to the
Ablowitz–Ladik hierarchy.
In complementary recent developments, Bridgeland put forward in [Bri19] a
Riemann–Hilbert problem associated to wall-crossing of Donaldson–Thomas
invariants. This was applied in [Bri20] to the resolved conifold, putting
forward a Tau function as a solution to the Riemann–Hilbert problem, which in
particular provides the Gromov–Witten potential as an asymptotic expansion.
In this work, we give a proof of the identification of the primary
Gromov–Witten potential for both the stationary and non-stationary sectors
with an AL tau function to all genera. The main ingredients of our proof are
two difference equations. One of these is obtained from a certain functional
representation of the AL hierarchy due to Vekslerchik [Vek98, Vek02], while
the other was proved by one of us in [Ali20] by adapting ideas which appeared
in the study of the exact WKB method in [IKT18]. In particular, by combining
this with the results of Bridgeland in [Bri20], we are able to obtain a
closed-form expression for the tau function, thus providing a new, but perhaps
expected, link between integrable hierarchies of GW theory and Bridgeland’s DT
Riemann–Hilbert problem.
The outline of this paper is as follows: In §2, we recall notions from the
Gromov–Witten theory of the resolved conifold, introduce the difference
equation proved in [Ali20], and construct a solution of it using Bridgeland’s
Tau function [Bri20]. In §3, we recall Vekslerchik’s functional representation
of the Ablowitz–Ladik hierarchy. In §4, we show how to take the dispersionless
limit of the AL hierarchy in this setting and give a version of Dubrovin’s
proof of the correspondence between the equivariant genus zero Gromov–Witten
potential of the resolved conifold and the dispersionless AL hierarchy adapted
to our setting. Finally, in §5, we bring all the ingredients together to
extend a result of Brini.
### Acknowledgements
We would like to thank Andrea Brini and Jörg Teschner for correspondence and
comments on the draft. This work is supported through the DFG Emmy Noether
grant AL 1407/2-1.
## 2\. Gromov–Witten potential and difference equation
Gromov–Witten (GW) theory of a non-singular algebraic variety $X$ is concerned
with the study of integrals over the moduli spaces of maps from Riemann
surfaces into $X$. Let $X$ be a Calabi–Yau threefold. The GW potential of $X$
is the following formal power series:
$F(\lambda,\boldsymbol{t})=\sum_{g\geq
0}\lambda^{2g-2}F^{g}(\boldsymbol{t})=\sum_{g\geq
0}\lambda^{2g-2}\sum_{\beta\in
H_{2}(X,\mathbb{Z})}N^{g}_{\beta}\,q^{\beta}\,,$ (2.1)
where $q^{\beta}:=\exp(2\pi i\langle\beta,\boldsymbol{t}\rangle)$ is a formal
variable living in a suitable completion of the effective cone in the group
ring of the small phase space $H_{2}(X,\mathbb{Z})$, $\lambda$ is a formal
parameter corresponding to the topological string coupling, and
$N^{g}_{\beta}$ are the GW invariants.
The GW potential can be written as:
$F=F_{\beta=0}+\tilde{F}\,,$ (2.2)
where $F_{\beta=0}$ denotes the contribution from constant maps (the
stationary sector) and $\tilde{F}$ the contribution from non-constant maps
(the non-stationary sector). The constant-map contributions at genus 0 and 1
are $\boldsymbol{t}$ dependent and the higher-genus constant-map contributions
take the universal form [FP00]:
$F_{\beta=0}^{g}=\frac{(-1)^{g-1}\,\chi(X)\,B_{2g}\,B_{2g-2}}{4g(2g-2)\,(2g-2)!}\,,\quad
g\geq 2\,,$ (2.3)
where $\chi(X)$ is the topological Euler characteristic of $X$ and $B_{k}$
denotes the $k$-th Bernoulli number.
This note is concerned with an equivariant version of the GW potential of the
CY threefold given by the total space of a rank two bundle over the projective
line:
$\mathcal{O}(-1)\oplus\mathcal{O}(-1)\rightarrow\mathbb{P}^{1}\,.$ (2.4)
This corresponds to the resolution of the conifold singularity in
$\mathbb{C}^{4}$ and is known as the resolved conifold. Its small phase space
is spanned by an identity element $\mathrm{id}$ and the Kähler class $\omega$
of the base $\mathbb{P}^{1}$. The associated coordinates shall be denoted
$x:=\langle\mathrm{id},\boldsymbol{t}\rangle$ and
$t:=\langle\omega,\boldsymbol{t}\rangle$ respectively. The (non-equivariant)
GW potential for this geometry was determined in physics [GV98, GV99], and in
mathematics [FP00] with the following outcome for the non-constant maps:
$\tilde{F}^{0}=\mathrm{Li}_{3}(q)\,,\quad\tilde{F}^{g}=\frac{(-1)^{g-1}B_{2g}}{2g(2g-2)!}\,\mathrm{Li}_{3-2g}(q)\,,\quad
g\geq 1\,,$ (2.5)
where $q:=\exp(2\pi{\mathrm{i}}t)$ and the polylogarithm is defined by:
$\mathrm{Li}_{s}(z)=\sum_{n=1}^{\infty}\frac{z^{n}}{n^{s}}\,,\quad
s\in\mathbb{C}\,.$ (2.6)
There is an anti-diagonal $\mathbb{C}^{\times}$ action on
$\mathcal{O}(-1)\oplus\mathcal{O}(-1)$, given by fibrewise
$\mathbb{C}^{\times}$ action on the $\mathcal{O}(-1)$ with opposite
characters. This $\mathbb{C}^{\times}$ action is especially interesting as it
acts trivially on the canonical bundle, meaning that
$\mathcal{O}(-1)\oplus\mathcal{O}(-1)$ is then _equivariantly_ Calabi–Yau.
With respect to this action, one can define an equivariant analogue of
Gromov–Witten theory where the ordinary cohomology ring of the resolved
conifold is replaced by the equivariant cohomology. The equivariant
Gromov–Witten invariants of rank $2$ bundles over curves were computed by
Bryan and Pandharipande in [BP08]. The theory significantly simplifies in the
case of the resolved conifold with anti-diagonal action: It is then just the
contribution from constant maps which is affected. In particular, the genus
zero Gromov–Witten potential becomes:
$F_{\mathrm{ad}}^{0}=\frac{(2\pi)^{3}{\mathrm{i}}x^{2}t}{2\kappa^{2}}+\tilde{F}^{0},$
(2.7)
where the $\mathrm{ad}$ in the subscript denotes that it is equivariant with
respect to the _anti-diagonal_ action and $\kappa$ is the equivariant
parameter.
The following theorem was proved in [Ali20]:
###### Theorem 1.
[Ali20] The contribution of the non-constant maps $\tilde{F}(\lambda,t)$ to
the GW potential of the resolved conifold satisfies the following difference
equation:
$\tilde{F}(\lambda,t+\check{\lambda})-2\tilde{F}(\lambda,t)+\tilde{F}(\lambda,t-\check{\lambda})=\left(\frac{1}{2\pi}\frac{\partial}{\partial
t}\right)^{2}\,\tilde{F}^{0}(t)\,,\quad\check{\lambda}=\frac{\lambda}{2\pi}\,.$
(2.8)
We proceed by determining the solution of the difference equation. The latter
already appears as a building block of the Tau function determined by
Bridgeland in [Bri20] as a solution to a Riemann–Hilbert problem associated to
wall-crossing of DT invariants of the resolved conifold. The special functions
in [Bri20] involve the multiple sine functions which are defined using the
Barnes multiple Gamma functions [Bar04]. For a variable $z\in\mathbb{C}$ and
parameters $\omega_{1},\ldots,\omega_{r}\in\mathbb{C}^{\times}$, these are
defined by:
$\sin_{r}(z\,|\,\omega_{1},\dots,\omega_{r}):=\Gamma_{r}(z\,|\,\omega_{1},\dots,\omega_{r})\cdot\Gamma_{r}\left(\sum_{i=1}^{r}\omega_{i}-z\,|\,\omega_{1},\dots,\omega_{r}\right)^{(-1)^{r}}.$
(2.9)
For further definitions, see e. g. [Bri20, Rui00] and references therein. We
will furthermore need the generalized Bernoulli polynomials, defined by the
generating function:
$\frac{x^{r}\,e^{zx}}{\prod_{i=1}^{r}(e^{\omega_{i}x}-1)}=\sum_{n=0}^{\infty}\frac{x^{n}}{n!}\,B_{r,n}(z\,|\,\omega_{1},\,\dots,\omega_{r})\,.$
(2.10)
Consider now the function $G(z\,|\,\omega_{1},\omega_{2})$ of [Bri20, Sec.
4.2], defined by:
$G(z\,|\,\omega_{1},\omega_{2}):=\exp\left(\frac{\pi i}{6}\cdot
B_{3,3}(z+\omega_{1}\,|\,\omega_{1},\omega_{1},\omega_{2})\right)\cdot\sin_{3}(z+\omega_{1}\,|\,\omega_{1},\omega_{2},\omega_{3}),$
(2.11)
and define a function111The subscript np of the function stands for non-
perturbative. In [Bri20], it is suggested that the Tau function of the DT
Riemann–Hilbert problem gives a non-perturbative definition of the GW theory
of the resolved conifold. The non-perturbative content of this function is
discussed in a companion paper.:
$F_{\mathrm{np}}(\lambda,t)=\log G(t\,|\,\check{\lambda},1)\,.$ (2.12)
We obtain the following:
###### Proposition 2.
The function $F_{\mathrm{np}}(\lambda,t)$ is a solution of the difference
equation (2.8). Moreover, $F_{\mathrm{np}}$ has an asymptotic expansion as
$\lambda\rightarrow 0$ given by:
$F_{\mathrm{np}}(\lambda,t)\sim\sum_{g=0}^{\infty}\lambda^{2g-2}\tilde{F}^{g}(t)\,,$
(2.13)
where $\tilde{F}^{g}(t)$ are the non-constant parts of the conifold free
energies defined in (2.5). The solution is furthermore unique given the above
asymptotic expansion.
###### Proof.
By [Bri20, Prop.4.2], the function $G(t\,|\,\check{\lambda},1)$ is single
valued for $\check{\lambda}\notin\mathbb{R}_{<0}$ and satisfies the difference
equation:
$\frac{G(t+\check{\lambda}\,|\,\check{\lambda},1)}{G(t\,|\,\check{\lambda},1)}=H(t+\check{\lambda}\,|\,\check{\lambda},1)^{-1}\,,$
(2.14)
where $H$ is denoted by $F$ in [Bri20] and is given by:
$H(t\,|\,\omega_{1},\omega_{2}):=\exp\left(-\frac{\pi i}{2}\cdot
B_{2,2}(t\,|\,\omega_{1},\omega_{2})\right)\cdot\sin_{2}(z\,|\,\omega_{1},\omega_{2})\,.$
(2.15)
This in turn satisfies the difference equation [Bri20, Prop 4.1]:
$\frac{H(t+\omega_{1}\,|\,\omega_{1},\omega_{2})}{H(t\,|\,\omega_{1},\omega_{2})}=\frac{1}{1-x_{2}}\,,\quad
x_{2}=\exp\left(\frac{2\pi it}{\omega_{2}}\right).$ (2.16)
From (2.14) we obtain, by taking the logarithm:
$\begin{split}&\log G(t+\check{\lambda}\,|\,\check{\lambda},1)+\log
G(t-\check{\lambda}\,|\,\check{\lambda},1)-2\log G(t\,|\,\check{\lambda},1)\\\
&=-\left(\log H(t+\check{\lambda}\,|\,\check{\lambda},1)-\log
H(t\,|\,\check{\lambda},1)\right)=\log(1-q)\,,\quad q=\exp(2\pi
it)\,.\end{split}$ (2.17)
For the asymptotic expansion of $F_{\mathrm{np}}$, we use [Bri20, Prop 4.6],
which in particular proves the asymptotic expansion of $G$ as
$\omega_{2}\rightarrow 0$ to be:
$\log G(z\,|\,\omega_{2},\omega_{1})\sim\sum_{k\geq 0}\frac{(k-1)\cdot
B_{k}\cdot\omega_{2}^{k-2}}{k!}\cdot\left(\frac{2\pi
i}{\omega_{1}}\right)^{k-2}\textrm{Li}_{3-k}\left(\exp\left(\frac{2\pi
iz}{\omega_{1}}\right)\right).$ (2.18)
It follows that:
$F_{\mathrm{np}}(\lambda,t)\sim\frac{1}{\lambda^{2}}\,\mathrm{Li}_{3}(q)+\sum_{g=1}^{\infty}\lambda^{2g-2}\frac{(-1)^{g-1}B_{2g}}{2g(2g-2)!}\,\textrm{Li}_{3-2g}(q)\,,$
(2.19)
by noting that $B_{k}=0$ for $k>1$ odd.
Finally, we prove the uniqueness of this solution given the asymptotic
expansion (2.13). Let $F_{1}$ be a solution to the difference equation (2.8)
satisfying (2.13). Then the difference $f(\lambda,t):=F_{1}(\lambda,t)-\log
G(t\,|\,\check{\lambda},1)$ satisfies the following equation:
$f(\lambda,t+\check{\lambda})-f(\lambda,t)=f(\lambda,t)-f(\lambda,t-\check{\lambda})\,.$
(2.20)
For any fixed value of $\lambda$, the function
$f(\lambda,t+\check{\lambda})-f(\lambda,t)$ is periodic in $t$ with period
$\check{\lambda}$. Thus, it must be of the form:
$f(\lambda,t+\check{\lambda})-f(\lambda,t)=\sum_{k\in\mathbb{Z}}\lambda
d_{k}(\lambda)\exp\left(\frac{2k\pi{\mathrm{i}}t}{\check{\lambda}}\right),$
(2.21)
where $d_{k}(\lambda)$ are (possibly meromorphic) functions of $\lambda$. This
can again be rewritten as:
$\begin{split}&f(\lambda,t+\check{\lambda})-(t+\check{\lambda})\sum_{k\in\mathbb{Z}}d_{k}(\lambda)\exp\left(\frac{2k\pi{\mathrm{i}}(t+\check{\lambda})}{\check{\lambda}}\right)\\\
&=f(\lambda,t)-t\sum_{k\in\mathbb{Z}}d_{k}(\lambda)\exp\left(\frac{2k\pi{\mathrm{i}}t}{\check{\lambda}}\right).\end{split}$
(2.22)
This gives another function which is periodic in $t$ with period
$\check{\lambda}$. Thus, the most general $f(\lambda,t)$ satisfying this
equation is:
$f(\lambda,t)=\sum_{k\in\mathbb{Z}}(c_{k}(\lambda)+td_{k}(\lambda))\exp\left(\frac{2k\pi{\mathrm{i}}t}{\check{\lambda}}\right),$
(2.23)
where $c_{k}(\lambda),d_{k}(\lambda)$ are (possibly meromorphic) functions of
$\lambda$.
Since $F_{1}(\lambda,t)$ and $\log G(t\,|\,\check{\lambda},1)$ both satisfy
(2.13) as $\lambda\rightarrow 0$, the functions $c_{k}(\lambda)$ and
$d_{k}(\lambda)$ are forced to vanish. This implies $F_{1}(\lambda,t)=\log
G(t\,|\,\check{\lambda},1)$ and that the solution is indeed unique. ∎
###### Corollary 3.
The equivariant Gromov–Witten potential of the resolved conifold with anti-
diagonal action is:
$F_{\mathrm{ad}}(\lambda,t)=\frac{(2\pi)^{3}{\mathrm{i}}x^{2}t}{2\kappa^{2}\check{\lambda}^{2}}+\log
G(t\,|\,\check{\lambda},1)\,.$ (2.24)
## 3\. The Ablowitz–Ladik hierarchy
The AL hierarchy [AL75] is a set of infinitely many non-linear differential-
difference equations in two functions
$a,b\colon\mathbb{Z}\rightarrow\mathbb{C}\llbracket z,\tilde{z}\rrbracket$,
where $z$ and $\tilde{z}$ represent two countably infinite tuples of (complex)
flow parameters $z_{i}$ and $\tilde{z}_{i}$ respectively. The simplest non-
trivial flow equation in the hierarchy is the AL system:
$\begin{split}{\mathrm{i}}\dot{a}&=\phantom{+}(\Lambda(a)+\Lambda^{-1}(a))(1-ab)\,,\\\
{\mathrm{i}}\dot{b}&=-(\Lambda(b)+\Lambda^{-1}(b))(1-ab)\,.\end{split}$ (3.1)
Here the dot denotes derivative with respect to a linear combination of the
flow parameters in the tuples $z$ and $\tilde{z}$, while $\Lambda$ denotes the
shift operator whose action on any function $u$ on $\mathbb{Z}$ is given by
$(\Lambda(u))(n)=u(n+1)\,.$ (3.2)
One way of constructing the full AL hierarchy is by means of three tau
functions $\sigma,\rho,\tau\colon\mathbb{Z}\rightarrow\mathbb{C}\llbracket
z,\tilde{z}\rrbracket$ associated to a solution $(a,b)$ of the AL system.
These are _defined_ by
$a=\frac{\sigma}{\tau}\,,\quad b=\frac{\rho}{\tau}\,,\quad
1-ab=\frac{\Lambda(\tau)\Lambda^{-1}(\tau)}{\tau^{2}}\,.$ (3.3)
Note that the tau functions are not independent but satisfy
$\Lambda(\tau)\Lambda^{-1}(\tau)=\tau^{2}-\sigma\rho\,.$ (3.4)
The AL hierarchy is then equivalent to the following system of Hirota bilinear
equations (cf. Vekslerchik [Vek98, Vek02]):
$\displaystyle\tau(z,\tilde{z})\tau(z+{\mathrm{i}}[\zeta],\tilde{z})-\rho(z,\tilde{z})\sigma(z+{\mathrm{i}}[\zeta],\tilde{z})$
$\displaystyle=\phantom{\zeta}\Lambda^{-1}(\tau)(z,\tilde{z})\Lambda(\tau)(z+{\mathrm{i}}[\zeta],\tilde{z})\,,$
(3.5a)
$\displaystyle\tau(z,\tilde{z})\sigma(z+{\mathrm{i}}[\zeta],\tilde{z})-\sigma(z,\tilde{z})\tau(z+{\mathrm{i}}[\zeta],\tilde{z})$
$\displaystyle=\zeta\Lambda^{-1}(\tau)(z,\tilde{z})\Lambda(\sigma)(z+{\mathrm{i}}[\zeta],\tilde{z})\,,$
(3.5b)
$\displaystyle\rho(z,\tilde{z})\tau(z+{\mathrm{i}}[\zeta],\tilde{z})-\tau(z,\tilde{z})\rho(z+{\mathrm{i}}[\zeta],\tilde{z})$
$\displaystyle=\zeta\Lambda^{-1}(\rho)(z,\tilde{z})\Lambda(\tau)(z+{\mathrm{i}}[\zeta],\tilde{z})\,,$
(3.5c)
$\displaystyle\tau(z,\tilde{z})\tau(z,\tilde{z}+{\mathrm{i}}[\zeta])-\rho(z,\tilde{z})\sigma(z,\tilde{z}+{\mathrm{i}}[\zeta])$
$\displaystyle=\phantom{\zeta}\Lambda(\tau)(z,\tilde{z})\Lambda^{-1}(\tau)(z,\tilde{z}+{\mathrm{i}}[\zeta])\,,$
(3.5d)
$\displaystyle\tau(z,\tilde{z})\sigma(z,\tilde{z}+{\mathrm{i}}[\zeta])-\sigma(z,\tilde{z})\tau(z,\tilde{z}+{\mathrm{i}}[\zeta])$
$\displaystyle=\zeta\Lambda(\tau)(z,\tilde{z})\Lambda^{-1}(\sigma)(z,\tilde{z}+{\mathrm{i}}[\zeta])\,,$
(3.5e)
$\displaystyle\rho(z,\tilde{z})\tau(z,\tilde{z}+{\mathrm{i}}[\zeta])-\tau(z,\tilde{z})\rho(z,\tilde{z}+{\mathrm{i}}[\zeta])$
$\displaystyle=\zeta\Lambda(\rho)(z,\tilde{z})\Lambda^{-1}(\tau)(z,\tilde{z}+{\mathrm{i}}[\zeta])\,.$
(3.5f)
In the above, $[\zeta]$ denotes the tuple:
$[\zeta]:=\left(\zeta,\frac{\zeta^{2}}{2},\frac{\zeta^{3}}{3},\ldots\right).$
(3.6)
The flows of the hierarchy are then obtained by expanding the above set of
equations order by order in $\zeta$. In particular, the AL system is obtained
by taking the flow along $\partial_{z_{1}}+\partial_{\tilde{z}_{1}}$.
## 4\. Dispersionless limit
The dispersionless limit of any integrable hierarchy is a limit in which the
contribution of higher-derivative terms (which cause dispersion) are
suppressed. This is achieved by scaling the length and times (i. e. the flow
parameters) by a factor $\check{\lambda}^{-1}$ inversely proportional to a
dispersion parameter $\lambda$, and then taking the limit $\lambda\rightarrow
0$. Expanding the relevant flow equations as Taylor series in $\lambda$ about
$\lambda=0$ then gives the small-dispersion expansion of the hierarchy.
The most straightforward way to do this in the case of the AL hierarchy is to
work at the level of tau functions. That is to say, we introduce tau functions
$\sigma_{\lambda},\rho_{\lambda},\tau_{\lambda}\colon\mathbb{C}\rightarrow\mathbb{C}\llbracket
z,\tilde{z}\rrbracket$ of a continuous complex variable $x$, parametrised by
$\lambda$. (It is not a coincidence that the spatial coordinate is denoted
with the same symbol as the group ring parameter associated to the identity
class in $H_{2}(X,\mathbb{Z})$; we will later see that they can be
identified.) The dependence of the tau functions on $\lambda$ is stipulated to
take the following form:
$\begin{split}\log\sigma_{\lambda}=\frac{\varpi_{\lambda}}{\check{\lambda}^{2}}+\frac{s_{\lambda}}{\check{\lambda}}+r_{\lambda}\,,\quad\log\rho_{\lambda}=\frac{\varpi_{\lambda}}{\check{\lambda}^{2}}-\frac{s_{\lambda}}{\check{\lambda}}+r_{\lambda}\,,\quad\log\tau_{\lambda}&=\frac{\varpi_{\lambda}}{\check{\lambda}^{2}}\,.\end{split}$
(4.1)
Here, $\varpi_{\lambda},s_{\lambda},r_{\lambda}$ are analytic in $\lambda$.
So, they have well-defined limits $\varpi,s,r$ as $\lambda\rightarrow 0$.
We also introduce the operator
$\Lambda_{\lambda}:=\exp(\check{\lambda}\partial_{x})$. The small-dispersion
expansion of the AL hierarchy is then given by making the following
replacements in (3.5):
$\sigma\mapsto\sigma_{\lambda}\,,\quad\rho\mapsto\rho_{\lambda}\,,\quad\tau\mapsto\tau_{\lambda}\,,\quad\Lambda\mapsto\Lambda_{\lambda}\,,\quad[\zeta]\mapsto\check{\lambda}[\zeta]\,.$
(4.2)
In the conjectured correspondence between GW theory and integrable
hierarchies, the dispersion parameter $\lambda$ plays the role of the genus
expansion parameter. In particular, the genus zero GW theory should correspond
to the dispersionless limit of the associated hierarchy. In order to relate
the dispersionless limit of the AL hierarchy to the resolved conifold, we will
first identify a suitable set of dependent variables and describe the
dynamical equations of the hierarchy in terms of them.
###### Proposition 4.
The dispersionless limit of the AL hierarchy is given by the following system
of equations in the functions $v=\partial_{x}s$ and $u=-\log(1-e^{2r})$:
$\displaystyle D^{\zeta}_{z}v$
$\displaystyle={\mathrm{i}}\,\frac{\partial}{\partial
x}\log\left(\frac{1-\zeta e^{v}+\sqrt{(1+\zeta e^{v})^{2}-4\zeta
e^{v}e^{-u}}}{2}\right),$ (4.3a) $\displaystyle D^{\zeta}_{z}u$
$\displaystyle={\mathrm{i}}\,\frac{\partial}{\partial
x}\log\left(\frac{1+\zeta e^{v}+\sqrt{(1+\zeta e^{v})^{2}-4\zeta
e^{v}e^{-u}}}{2}\right),$ (4.3b) $\displaystyle D^{\zeta}_{\tilde{z}}v$
$\displaystyle={\mathrm{i}}\,\frac{\partial}{\partial
x}\log\left(\frac{1-\zeta e^{-v}+\sqrt{(1+\zeta e^{-v})^{2}-4\zeta
e^{-v}e^{-u}}}{2}\right),$ (4.3c) $\displaystyle D^{\zeta}_{\tilde{z}}u$
$\displaystyle=-{\mathrm{i}}\,\frac{\partial}{\partial
x}\log\left(\frac{1+\zeta e^{-v}+\sqrt{(1+\zeta e^{-v})^{2}-4\zeta
e^{-v}e^{-u}}}{2}\right),$ (4.3d)
where $D^{\zeta}_{z}$ and $D^{\zeta}_{\tilde{z}}$ denote the operators:
$D^{\zeta}_{z}=\sum_{j=1}^{\infty}\frac{\zeta^{j}}{j}\frac{\partial}{\partial
z_{j}}\,,\quad
D^{\zeta}_{\tilde{z}}=\sum_{j=1}^{\infty}\frac{\zeta^{j}}{j}\frac{\partial}{\partial\tilde{z}_{j}}\,.$
(4.4)
###### Proof.
Making the replacements (4.2) in (3.5) and dividing equations (3.5a) and
(3.5b) by the respective first terms on the left-hand side, we get:
$\displaystyle
1-\left(\frac{\rho_{\lambda}}{\tau_{\lambda}}\right)(z,\tilde{z})\left(\frac{\sigma_{\lambda}}{\tau_{\lambda}}\right)(z+{\mathrm{i}}\check{\lambda}[\zeta],\tilde{z})$
$\displaystyle\quad=\left(\frac{\Lambda^{-1}_{\lambda}(\tau_{\lambda})}{\tau_{\lambda}}\right)(z,\tilde{z})\left(\frac{\Lambda_{\lambda}(\tau_{\lambda})}{\tau_{\lambda}}\right)(z+{\mathrm{i}}\check{\lambda}[\zeta],\tilde{z})\,,$
(4.5a) $\displaystyle
1-\left(\frac{(\sigma_{\lambda}/\tau_{\lambda})(z+{\mathrm{i}}\check{\lambda}[\zeta],\tilde{z})}{(\sigma_{\lambda}/\tau_{\lambda})(z,\tilde{z})}\right)^{-1}$
$\displaystyle\quad=\zeta\left(\frac{\Lambda^{-1}_{\lambda}(\tau_{\lambda})}{\tau_{\lambda}}\right)(z,\tilde{z})\left(\frac{\Lambda_{\lambda}(\tau_{\lambda})}{\tau_{\lambda}}\right)(z+{\mathrm{i}}\check{\lambda}[\zeta],\tilde{z})\left(\frac{\Lambda_{\lambda}(\sigma_{\lambda}/\tau_{\lambda})}{\sigma_{\lambda}/\tau_{\lambda}}\right)(z+{\mathrm{i}}\check{\lambda}[\zeta],\tilde{z})\,.$
(4.5b)
Substituting (4.1) into the above then gives us:
$\displaystyle
1-\exp\left(\frac{s_{\lambda}(z+{\mathrm{i}}\check{\lambda}[\zeta],\tilde{z})-s_{\lambda}(z,\tilde{z}))}{\check{\lambda}}\right)\exp\left(r_{\lambda}(z+{\mathrm{i}}\check{\lambda}[\zeta],\tilde{z})+r_{\lambda}(z,\tilde{z})\right)$
$\displaystyle\quad=\exp\left(\frac{(\Lambda_{\lambda}^{-1}-1)(\varpi_{\lambda})(z,\tilde{z})+(\Lambda_{\lambda}-1)(\varpi_{\lambda})(z+{\mathrm{i}}\check{\lambda}[\zeta],\tilde{z})}{\check{\lambda}^{2}}\right),$
(4.6a) $\displaystyle
1-\exp\left(-\frac{s_{\lambda}(z+{\mathrm{i}}\check{\lambda}[\zeta],\tilde{z})-s_{\lambda}(z,\tilde{z}))}{\check{\lambda}}\right)\exp\left(-r_{\lambda}(z+{\mathrm{i}}\check{\lambda}[\zeta],\tilde{z})+r_{\lambda}(z,\tilde{z}))\right)$
$\displaystyle\quad=\zeta\exp\left(\frac{(\Lambda_{\lambda}^{-1}-1)(\varpi_{\lambda})(z,\tilde{z})+(\Lambda_{\lambda}-1)(\varpi_{\lambda})(z+{\mathrm{i}}\check{\lambda}[\zeta],\tilde{z})}{\check{\lambda}^{2}}\right)$
$\displaystyle\quad\quad\times\exp\left((\Lambda_{\lambda}-1)\left(\frac{s_{\lambda}(z+{\mathrm{i}}\check{\lambda}[\zeta],\tilde{z})}{\check{\lambda}}+\pi{\mathrm{i}}t_{\lambda}(z+{\mathrm{i}}\check{\lambda}[\zeta],\tilde{z})\right)\right).$
(4.6b)
For convenience, we now introduce the following notation:
$\begin{split}X(z,\tilde{z})&=\frac{s_{\lambda}(z+{\mathrm{i}}\check{\lambda}[\zeta],\tilde{z})-s_{\lambda}(z,\tilde{z}))}{\check{\lambda}}\,,\\\
Y(z,\tilde{z})&=\frac{r_{\lambda}(z+{\mathrm{i}}\check{\lambda}[\zeta],\tilde{z})-r_{\lambda}(z,\tilde{z})}{\check{\lambda}}\,,\\\
Z(z,\tilde{z})&=(\Lambda_{\lambda}-1)\left(\frac{s_{\lambda}(z+{\mathrm{i}}\check{\lambda}[\zeta],\tilde{z})}{\check{\lambda}}+r_{\lambda}(z+{\mathrm{i}}\check{\lambda}[\zeta],\tilde{z})\right).\end{split}$
(4.7)
Eliminating $\varpi_{\lambda}$ from (4.6a) and (4.6b) gives us a quadratic
equation in $e^{X(z,\tilde{z})}$:
$\zeta
e^{Z(z,\tilde{z})}e^{\check{\lambda}Y(z,\tilde{z})}e^{2r_{\lambda}(z,\tilde{z})}e^{2X(z,\tilde{z})}+\left(1-\zeta
e^{Z(z,\tilde{z})}\right)e^{X(z,\tilde{z})}-e^{-\check{\lambda}Y(z,\tilde{z})}=0\,.$
(4.8)
This has two solutions but only one such that $X(z,\tilde{z})$ has a well-
defined limit as $\zeta\rightarrow 0$:
$X(z,\tilde{z})=\log\left(\frac{-1+\zeta e^{Z(z,\tilde{z})}+\sqrt{(1-\zeta
e^{Z(z,\tilde{z})})^{2}+4\zeta
e^{Z(z,\tilde{z})}e^{2r_{\lambda}(z,\tilde{z})}}}{2\zeta
e^{Z(z,\tilde{z})}e^{\check{\lambda}Y(z,\tilde{z})}e^{2r_{\lambda}(z,\tilde{z})}}\right).$
(4.9)
In the dispersionless limit $\lambda\rightarrow 0$, this becomes:
${\mathrm{i}}D^{\zeta}_{z}s=\log\left(\frac{-1+\zeta e^{v}+\sqrt{(1-\zeta
e^{v})^{2}+4\zeta e^{v}e^{2r}}}{2\zeta e^{v}e^{2r}}\right).$ (4.10)
Differentiating with respect to $x$ on both sides and substituting
$e^{2r}=1-e^{-u}$ then gives us:
$D^{\zeta}_{z}v={\mathrm{i}}\,\frac{\partial}{\partial
x}\log\left(\frac{-1+\zeta e^{v}+\sqrt{(1+\zeta e^{v})^{2}-4\zeta
e^{v}e^{-u}}}{2\zeta e^{v}(1-e^{-u})}\right).$ (4.11)
Equation (4.3a) then follows from the identity:
$\frac{-1+\zeta e^{v}+\sqrt{(1+\zeta e^{v})^{2}-4\zeta e^{v}e^{-u}}}{2\zeta
e^{v}(1-e^{-u})}=\frac{2}{1-\zeta e^{v}+\sqrt{(1+\zeta e^{v})^{2}-4\zeta
e^{v}e^{-u}}}\,.$ (4.12)
We will now use what we have established so far to prove (4.3b). First, we
note that (4.6a) in the limit $\lambda\rightarrow 0$ becomes:
$1-\exp({\mathrm{i}}D^{\zeta}_{z}s)e^{2r}=\exp\left(\frac{\partial^{2}\varpi}{\partial
x^{2}}+{\mathrm{i}}\frac{\partial D^{\zeta}_{z}\varpi}{\partial x}\right).$
(4.13)
Setting $\zeta=0$ then gives us:
$u=-\log(1-e^{2r})=-\frac{\partial^{2}\varpi}{\partial x^{2}}\,.$ (4.14)
Substituting this back into (4.13), we obtain:
$\frac{\partial D^{\zeta}_{z}\varpi}{\partial
x}=-{\mathrm{i}}\log\left(\frac{1-\exp({\mathrm{i}}D^{\zeta}_{z}s)e^{2r}}{e^{-u}}\right).$
(4.15)
Differentiating with respect to $x$ on both sides and substituting (4.10) and
$e^{2r}=1-e^{-u}$ then gives
$D^{\zeta}_{z}u={\mathrm{i}}\,\frac{\partial}{\partial
x}\log\left(\frac{1+\zeta e^{v}-\sqrt{(1+\zeta e^{v})^{2}-4\zeta
e^{v}e^{-u}}}{2\zeta e^{v}e^{-u}}\right).$ (4.16)
Equation (4.3b) then follows from the identity:
$\frac{1+\zeta e^{v}-\sqrt{(1+\zeta e^{v})^{2}-4\zeta e^{v}e^{-u}}}{2\zeta
e^{v}e^{-u}}=\frac{2}{1+\zeta e^{v}+\sqrt{(1+\zeta e^{v})^{2}-4\zeta
e^{v}e^{-u}}}\,.$ (4.17)
A similar argument involving (3.5d) and (3.5e) leads to (4.3c) and (4.3d). ∎
###### Remark 5.
By setting $\zeta=0$ in (4.6a) and taking the logarithm of both sides, we
obtain a difference equation involving shifts in $x$:
$\log(1-\exp(2r_{\lambda}))=\frac{\Lambda_{\lambda}(\varpi_{\lambda})-2\varpi_{\lambda}+\Lambda_{\lambda}^{-1}(\varpi_{\lambda})}{\check{\lambda}^{2}}\,.$
(4.18)
###### Remark 6.
By applying $\zeta\partial_{\zeta}$ to (4.3) we get the following alternative
form for the dispersionless AL hierarchy:
$\displaystyle\Delta^{\zeta}_{z}v$
$\displaystyle=-{\mathrm{i}}\,\frac{\partial}{\partial x}\left(\frac{1+\zeta
e^{v}}{2\sqrt{(1+\zeta e^{v})^{2}-4\zeta e^{v}e^{-u}}}\right)$
$\displaystyle=-{\mathrm{i}}\,\frac{\partial^{2}}{\partial x\partial
u}\tanh^{-1}\left(\frac{1+\zeta e^{v}}{\sqrt{(1+\zeta e^{v})^{2}-4\zeta
e^{v}e^{-u}}}\right),$ (4.19a) $\displaystyle\Delta^{\zeta}_{z}u$
$\displaystyle=\phantom{+}{\mathrm{i}}\,\frac{\partial}{\partial
x}\left(\frac{1-\zeta e^{v}}{2\sqrt{(1+\zeta e^{v})^{2}-4\zeta
e^{v}e^{-u}}}\right)$
$\displaystyle=-{\mathrm{i}}\,\frac{\partial^{2}}{\partial x\partial
v}\tanh^{-1}\left(\frac{1+\zeta e^{v}}{\sqrt{(1+\zeta e^{v})^{2}-4\zeta
e^{v}e^{-u}}}\right),$ (4.19b) $\displaystyle\Delta^{\zeta}_{\tilde{z}}v$
$\displaystyle=-{\mathrm{i}}\,\frac{\partial}{\partial x}\left(\frac{1+\zeta
e^{-v}}{2\sqrt{(1+\zeta e^{-v})^{2}-4\zeta e^{-v}e^{-u}}}\right)$
$\displaystyle=-{\mathrm{i}}\frac{\partial^{2}}{\partial x\partial
u}\tanh^{-1}\left(\frac{1+\zeta e^{-v}}{\sqrt{(1+\zeta e^{-v})^{2}-4\zeta
e^{-v}e^{-u}}}\right),$ (4.19c) $\displaystyle\Delta^{\zeta}_{\tilde{z}}u$
$\displaystyle=-{\mathrm{i}}\,\frac{\partial}{\partial x}\left(\frac{1-\zeta
e^{-v}}{2\sqrt{(1+\zeta e^{-v})^{2}-4\zeta e^{-v}e^{-u}}}\right)$
$\displaystyle=-{\mathrm{i}}\,\frac{\partial^{2}}{\partial x\partial
v}\tanh^{-1}\left(\frac{1+\zeta e^{-v}}{\sqrt{(1+\zeta e^{-v})^{2}-4\zeta
e^{-v}e^{-u}}}\right),$ (4.19d)
where $\Delta^{\zeta}_{z}$ and $\Delta^{\zeta}_{\tilde{z}}$ denote the
operators:
$\Delta^{\zeta}_{z}=\sum_{j=1}^{\infty}\zeta^{j}\frac{\partial}{\partial
z_{j}}\,,\quad\Delta^{\zeta}_{\tilde{z}}=\sum_{j=1}^{\infty}\zeta^{j}\frac{\partial}{\partial\tilde{z}_{j}}\,.$
(4.20)
In particular, this is a dispersionless _Hamiltonian_ hierarchy with
generating functions of Hamiltonian densities:
$\begin{split}h(\zeta)&=-{\mathrm{i}}\tanh^{-1}\left(\frac{1+\zeta
e^{v}}{\sqrt{(1+\zeta e^{v})^{2}-4\zeta e^{v}e^{-u}}}\right),\\\
\tilde{h}(\zeta)&=-{\mathrm{i}}\tanh^{-1}\left(\frac{1+\zeta
e^{-v}}{\sqrt{(1+\zeta e^{-v})^{2}-4\zeta e^{-v}e^{-u}}}\right),\end{split}$
(4.21)
and the following Poisson brackets:
$\\{u(x),u(y)\\}=\\{v(x),v(y)\\}=0\,,\quad\\{u(x),v(y)\\}=\\{v(x),u(y)\\}=\delta^{\prime}(x-y)\,.$
(4.22)
There are a number of interesting aspects of the Hamiltonian formulation of
the dispersionless limit of the AL hierarchy. Firstly, the Hamiltonian
densities depend only on $u$ and $v$ explicitly and not on their spatial
derivatives. Secondly, the coefficients of $\delta^{\prime}(x)$ in the Poisson
brackets form a constant symmetric matrix. A Hamiltonian hierarchy that has
this structure (possibly after a suitable change of dependent variables) is
said to be of _hydrodynamic type_. Such hierarchies were first defined and
studied by Dubrovin and Novikov [DN84].
A key fact concerning Hamiltonian hierarchies of hydrodynamic type is that
there is a natural such hierarchy called the _principal hierarchy_ associated
to any (almost) Frobenius manifold, which conversely may be viewed as a
differential-geometric structure on the initial data of such a hierarchy. The
small phase space $H_{2}(X,\mathbb{Z})$ of a projective variety happens to
carry the structure of a(n almost) Frobenius manifold, so this gives a rather
general construction of a dispersionless integrable hierarchy associated to
the (equivariant) genus zero GW theory of any projective variety. It was noted
by Dubrovin in [Dub08] that the dispersionless limit of the AL hierarchy is
contained within the principal hierarchy associated to the equivariant GW
theory of the resolved conifold with anti-diagonal action (hereafter referred
to as simply the principal hierarchy of the resolved conifold). We will now
present a proof of this result in our notation to ensure that the dependent
variables $u$ and $v$ that we have introduced are indeed the variables of the
principal hierarchy identified by Dubrovin.
###### Proposition 7.
The dispersionless limit of AL hierarchy with the initial conditions
$v=2\pi{\mathrm{i}}x/\kappa$ and $u=-2\pi{\mathrm{i}}t$ is contained in the
principal hierarchy of the resolved conifold with the following identification
on the small phase space:
$(2\pi)^{2}\varpi|_{(z,\tilde{z})=(0,0)}=\left.\left(-\frac{uv^{2}}{2}+\tilde{F}^{0}\right)\right|_{(z,\tilde{z})=(0,0)}.$
(4.23)
###### Proof.
The principal hierarchy associated to a(n almost) Frobenius manifold $V$ is
defined on the formal loop space $L(S^{1},V)$, whose elements are maps from
$S^{1}=\mathbb{R}/\mathbb{Z}$ to $V$. In our case, $V$ is the small phase
space $H_{2}(X,\mathbb{Z})$, while $w^{1}:=v$ and $w^{2}:=u$ can be
interpreted as maps from $S^{1}$ to $V$. (Note that the map $w^{1}$ is
distinguished.) Meanwhile, $x$ is interpreted as a coordinate on the universal
cover $\mathbb{R}$ of $S^{1}$.
A potential $\Phi$ on $V$ encodes a metric $\eta$ and an associative product
$\star$ with identity $\partial_{w^{1}}$ on the tangent bundle $TV$:
$\begin{split}\eta_{\mu\nu}:=\eta\left(\frac{\partial}{\partial
w^{\mu}}\,,\frac{\partial}{\partial
w^{\nu}}\right)&=\frac{\partial^{3}\Phi}{\partial w^{1}\partial
w^{\mu}\partial w^{\nu}}\,,\\\ \eta\left(\frac{\partial}{\partial
w^{\kappa}}\,,\frac{\partial}{\partial w^{\mu}}\star\frac{\partial}{\partial
w^{\nu}}\right)&=\frac{\partial^{3}\Phi}{\partial w^{\kappa}\partial
w^{\mu}\partial w^{\nu}}\,.\end{split}$ (4.24)
We also have a Poisson bracket on $L(S^{1},V)$ in terms of the the inverse
metric $\eta^{\mu\nu}$:
$\\{w^{\mu}(x),w^{\nu}(y)\\}=\eta^{\mu\nu}\delta^{\prime}(x-y)-\eta^{\mu\kappa}\Gamma^{\nu}_{\kappa\gamma}\frac{\partial
w^{\gamma}}{\partial x}\delta(x-y)\,,$ (4.25)
where $\Gamma^{\nu}_{\kappa\gamma}$ are the Christoffel symbols of the metric
$\eta$ and repeated indices are summed over. Comparing this with (4.22), we
immediately obtain:
$\eta_{11}=\eta_{22}=0,\quad\eta_{12}=\eta_{21}=1\,.$ (4.26)
Thus, the most general form that the potential $\Phi$ can take is:
$\Phi=\frac{uv^{2}}{2}+f(u)\,.$ (4.27)
A consequence of the associativity of the product $\star$ is the existence of
a pencil of flat torsion-free connections $\nabla^{\xi}$ parametrised by a
parameter $\xi\in\mathbb{C}$:
$\nabla^{\xi}_{\partial/\partial w^{\mu}}\frac{\partial}{\partial
w^{\nu}}=\xi\,\frac{\partial}{\partial w^{\mu}}\star\frac{\partial}{\partial
w^{\nu}}\,.$ (4.28)
As these connections are flat, they admit local basis of parallel $1$-forms
$\alpha_{1}(\xi)$ and $\alpha_{2}(\xi)$ analytic in $\xi$. And as these
connections are torsion-free, the anti-symmetric parts of
$\nabla^{\xi}\alpha_{1}$ and $\nabla^{\xi}\alpha_{2}$ are
$\mathrm{d}\alpha_{1}$ and $\mathrm{d}\alpha_{2}$ (where $\mathrm{d}$ is the
exterior derivative on $V$). Thus, $1$-forms $\alpha_{1}(\xi)$ and
$\alpha_{2}(\xi)$ are closed, and so by Poincaré’s lemma, we can locally find
functions $g_{1}(\xi)$ and $g_{2}(\xi)$ analytic in $\xi$ such that we have:
$\alpha_{1}=\mathrm{d}g_{1}\,,\quad\alpha_{2}=\mathrm{d}g_{2}\,.$ (4.29)
More explicitly, the functions $g_{1}$ and $g_{2}$ are solutions to the
following system of equations:
$\frac{\partial^{2}g}{\partial v^{2}}=\xi\,\frac{\partial g}{\partial
v}\,,\quad\frac{\partial^{2}g}{\partial v\,\partial u}=\xi\,\frac{\partial
g}{\partial u}\,,\quad\frac{\partial^{2}g}{\partial u^{2}}=\xi
f^{\prime\prime\prime}(u)\,\frac{\partial g}{\partial v}\,.$ (4.30)
The functions $g_{1}(\xi)$ and $g_{2}(\xi)$ are generating functions of
Hamiltonian densities, which generate the principal hierarchy via the Poisson
bracket (4.25). The generating functions $h(\zeta)$ and $\tilde{h}(\zeta)$ are
related to these in a non-trivial way (see Remark 8). We can get around the
need to obtain an explicit relationship between them by eliminating $\xi$ from
the first and third equations in (4.30). This gives us the following equation
that by virtue of its linearity is valid for any generating function of
Hamiltonian densities:
$\frac{\partial^{2}g}{\partial
u^{2}}=f^{\prime\prime\prime}(u)\,\frac{\partial^{2}g}{\partial v^{2}}\,.$
(4.31)
Substituting $g=h$ and $g=\tilde{h}$ as in (4.21) into the above, we obtain:
$f^{\prime\prime\prime}(u)=\frac{1}{e^{u}-1}=\mathrm{Li}_{0}(e^{-u})\,.$
(4.32)
This is solved by
$f(u)=-\tilde{F}^{0}(t)=-\mathrm{Li}_{3}(e^{2\pi{\mathrm{i}}t})$ for
$u=-2\pi{\mathrm{i}}t$. So the dispersionless AL hierarchy is contained in the
principal hierarchy of the resolved conifold. The identification (4.23)
meanwhile follows from (4.14) and the initial condition
$v=2\pi{\mathrm{i}}x/\kappa$. ∎
The exponential $e^{\Phi}$ of the potential $\Phi$ of the (almost) Frobenius
manifold will be henceforth called the _topological tau function_ of the
principal hierarchy.
###### Remark 8.
The first two equations of (4.30) imply that any solution $g$ necessarily has
the form:
$g=A(\xi,u)e^{\xi v}+B(\xi)\,,$ (4.33)
where $A$ is a function independent of $v$ while $B$ is a function independent
of $v$ and $u$. Since we are only interested in the flows of the Hamiltonians,
we can without loss of generality set $B$ to be identically zero. Conversely,
any function $g=A(\xi,u)e^{\xi v}$ which also satisfies (4.31) is a solution
of (4.30). Now consider the Mellin transforms:
$\begin{split}\int_{0}^{\infty}h\zeta^{-1-\xi}\mathrm{d}\zeta&=-{\mathrm{i}}e^{\xi
v}\int_{0}^{\infty}\tanh^{-1}\left(\frac{1+\zeta}{\sqrt{(1+\zeta)^{2}-4\zeta
e^{-u}}}\right)\zeta^{-1-\xi}\mathrm{d}\zeta\\\
&=\phantom{+}{\mathrm{i}}e^{\xi
v}\int^{0}_{\infty}\,\,\tanh^{-1}\left(\frac{1+\zeta^{-1}}{\sqrt{(1+\zeta^{-1})^{2}-4\zeta^{-1}e^{-u}}}\right)\zeta^{1+\xi}\,\frac{\mathrm{d}\zeta}{\zeta^{2}}\\\
&=-{\mathrm{i}}e^{\xi
v}\int_{0}^{\infty}\tanh^{-1}\left(\frac{\zeta+1}{\sqrt{(\zeta+1)^{2}-4\zeta
e^{-u}}}\right)\zeta^{-1+\xi}\mathrm{d}\zeta\,,\\\
\int_{0}^{\infty}\tilde{h}\zeta^{-1+\xi}\mathrm{d}\zeta&=-{\mathrm{i}}e^{\xi
v}\int_{0}^{\infty}\tanh^{-1}\left(\frac{1+\zeta}{\sqrt{(1+\zeta)^{2}-4\zeta
e^{-u}}}\right)\zeta^{-1+\xi}\mathrm{d}\zeta\,.\end{split}$ (4.34)
This is of the form $A(\xi,u)e^{\xi v}$. Moreover, being an integral transform
of $h$ and $\tilde{h}$, it satisfies (4.31) with
$f(u)=-\mathrm{Li}_{3}(e^{-u})$ as well. Hence, it must be a solution $g$ of
(4.30). In other words, it is some constant linear combination of the
Hamiltonian density generating functions $g_{1}$ and $g_{2}$ of the principal
hierarchy of the resolved conifold.
## 5\. Identification of the Tau function
In this section, we will extend the genus zero identification of the
dispersionless AL tau function with the topological tau function of the
principal hierarchy of the resolved conifold to an all-genera small-phase-
space identification of the tau function of the small-dispersion expansion of
the AL hierarchy with Bridgeland’s Tau function. The genus $2$ truncation of
this identification was essentially the content of Theorem 1.4 in [Bri12] and
it was proved by analyzing dispersive perturbations of the principal
hierarchy. This is a computationally challenging task, however the control
over expansion in the dispersive parameter offered by the functional
representation of the AL hierarchies will allow us to sidestep these
difficulties.
Here is the gist of our approach. The solution of Prop. 2 is used to define
initial conditions for the tau function of the AL hierarchy while the other
difference equation in Remark 5 is used to show that these initial conditions
reduce to those of the principal hierarchy of the resolved conifold in the
dispersionless limit.
We now prove our main theorem, which extends Theorem 1.4 in [Bri12] to all
genera.
###### Theorem 9.
There exists a solution to the small-dispersion expansion of the AL hierarchy
satisfying the following conditions:
1. (1)
The tau function $\tau_{\lambda}$ of the solution reduces at leading order in
$\lambda$ to the topological tau function of the principal hierarchy of the
resolved conifold as $\lambda\rightarrow 0$,
2. (2)
The restriction of the tau function $\tau_{\lambda}$ of the solution to the
small phase space coincides with the exponential of the equivariant GW
potential of the resolved conifold with anti-diagonal action, namely:
$\exp(F_{\mathrm{ad}}(\lambda,t))=\exp\bigg{(}\frac{(2\pi)^{3}{\mathrm{i}}x^{2}t}{2\kappa^{2}\check{\lambda}^{2}}\bigg{)}G(t\,|\,\check{\lambda},1)\,,$
(5.1)
to all orders in $\lambda$.
###### Proof.
We impose the following initial conditions on
$v_{\lambda}:=\check{\lambda}^{-1}(\Lambda_{\lambda}-1)s_{\lambda}$ and
$\varpi_{\lambda}$:
$v_{\lambda}|_{(z,\tilde{z})=(0,0)}=\frac{2\pi{\mathrm{i}}x}{\kappa}\,,\quad(2\pi)^{2}\varpi_{\lambda}|_{(z,\tilde{z})=(0,0)}=\frac{(2\pi)^{3}{\mathrm{i}}x^{2}t}{2\kappa^{2}}+\check{\lambda}^{2}\log
G(t\,|\,\check{\lambda},1)\,.$ (5.2)
This entails condition (2). Moreover, the difference equation in Remark 5
implies the following initial condition for
$u_{\lambda}:=-\log(1-\exp(2r_{\lambda}))$:
$u_{\lambda}|_{(z,\tilde{z})=(0,0)}=-2\pi{\mathrm{i}}t\,.$ (5.3)
The initial conditions for $u_{\lambda}$ and $v_{\lambda}$ reduce to those for
$u$ and $v$ in the principal hierarchy of the resolved conifold in the limit
$\lambda\rightarrow 0$. Since the full solution to the principal hierarchy is
fixed by the initial conditions, condition (1) holds as well. ∎
## References
* [ADK+06] Mina Aganagic, Robbert Dijkgraaf, Albrecht Klemm, Marcos Mariño, and Cumrun Vafa. Topological strings and integrable hierarchies. Comm. Math. Phys., 261(2):451–516, 2006.
* [AL75] Mark J. Ablowitz and John F. Ladik. Nonlinear differential-difference equations. Journal of Mathematical Physics, 16(3):598–603, March 1975.
* [Ali20] Murad Alim. Difference equation for the Gromov–Witten potential of the resolved conifold. arXiv preprint arXiv:2011.12759, 2020.
* [Bar04] Ernest W. Barnes. On the theory of multiple gamma function. Trans. Cambridge Phil. Soc., 19:374–425, 1904.
* [BP08] Jim Bryan and Rahul Pandharipande. The Local Gromov–Witten theory of curves. J. Am. Math. Soc., 21:101–136, 2008.
* [Bri12] Andrea Brini. The local Gromov–Witten theory of $\mathbb{C}\mathbb{P}^{1}$ and integrable hierarchies. Comm. Math. Phys., 313(3):571–605, 2012.
* [Bri19] Tom Bridgeland. Riemann–Hilbert problems from Donaldson-Thomas theory. Invent. Math., 216(1):69–124, 2019.
* [Bri20] Tom Bridgeland. Riemann–Hilbert problems for the resolved conifold. J. Differential Geom., 115(3):395–435, 2020.
* [DN84] Boris A. Dubrovin and Sergei P. Novikov. On Poisson brackets of hydrodynamic type. Dokl. Akad. Nauk SSSR, pages 294–297, 1984.
* [Dub08] Boris A. Dubrovin. On universality of critical behaviour in Hamiltonian PDEs. arXiv e-prints, page arXiv:0804.3790, April 2008.
* [Dub14] Boris A. Dubrovin. Gromov–Witten invariants and integrable hierarchies of topological type. In Topology, geometry, integrable systems, and mathematical physics, volume 234 of Amer. Math. Soc. Transl. Ser. 2, pages 141–171. Amer. Math. Soc., Providence, RI, 2014.
* [DZ01] Boris A. Dubrovin and Youjin Zhang. Normal forms of hierarchies of integrable PDEs, Frobenius manifolds and Gromov–Witten invariants. 8 2001.
* [EHX97] Tohru Eguchi, Kentaro Hori, and Chuan-Sheng Xiong. Quantum cohomology and Virasoro algebra. Phys. Lett. B, 402(1-2):71–80, 1997.
* [EHY95] Tohru Eguchi, Kentaro Hori, and Sung-Kil Yang. Topological $\sigma$ models and large-$N$ matrix integral. Internat. J. Modern Phys. A, 10(29):4203–4224, 1995.
* [EY94] Tohru Eguchi and Sung-Kil Yang. The topological ${\bf C}{\rm P}^{1}$ model and the large-$N$ matrix integral. Modern Phys. Lett. A, 9(31):2893–2902, 1994.
* [EYY95] Tohru Eguchi, Yasuhiko Yamada, and Sung-Kil Yang. On the genus expansion in the topological string theory. Rev. Math. Phys., 7(3):279–309, 1995.
* [FP00] Carel F. Faber and Rahul Pandharipande. Hodge integrals and Gromov–Witten theory. Invent. Math., 139(1):173–199, 2000.
* [GV98] Rajesh Gopakumar and Cumrun Vafa. M theory and topological strings. 1. 9 1998.
* [GV99] Rajesh Gopakumar and Cumrun Vafa. On the gauge theory/geometry correspondence. Adv. Theor. Math. Phys., 3(5):1415–1443, 1999.
* [IKT18] Kohei Iwaki, Tatsuya Koike, and Yumiko Takei. Voros coefficients for the hypergeometric differential equations and Eynard–Orantin’s topological recursion – Part I: For the Weber equation. arXiv preprint arXiv:1805.10945, 2018.
* [Kon92] Maxim Kontsevich. Intersection theory on the moduli space of curves and the matrix Airy function. Comm. Math. Phys., 147(1):1–23, 1992.
* [Oko00] Andrei Okounkov. Toda equations for Hurwitz numbers. Math. Res. Lett., 7(4):447–453, 2000.
* [Pan00] Rahul Pandharipande. The Toda equations and the Gromov–Witten theory of the Riemann sphere. Lett. Math. Phys., 53(1):59–74, 2000.
* [Rui00] Simon N. M. Ruijsenaars. On Barnes’ multiple zeta and gamma functions. Adv. Math., 156(1):107–132, 2000.
* [Tak13] Kanehisa Takasaki. A modified melting crystal model and the Ablowitz–Ladik hierarchy. Journal of Physics A: Mathematical and Theoretical, 46(24):245202, May 2013.
* [Vek98] Vadym E. Vekslerchik. Functional representation of the Ablowitz–Ladik hierarchy. Journal of Physics A: Mathematical and General, 31(3):1087–1099, Jan 1998.
* [Vek02] Vadym E. Vekslerchik. Functional representation of the Ablowitz–Ladik hierarchy. II. Journal of Nonlinear Mathematical Physics, 9(2):157–180, 2002.
* [Wit91] Edward Witten. Two-dimensional gravity and intersection theory on moduli space. In Surveys in differential geometry (Cambridge, MA, 1990), pages 243–310. Lehigh Univ., Bethlehem, PA, 1991.
|
# LS-HDIB: A Large Scale Handwritten Document Image Binarization Dataset
###### Abstract
Handwritten document image binarization is challenging due to high variability
in the written content and complex background attributes such as page style,
paper quality, stains, shadow gradients, and non-uniform illumination. While
the traditional thresholding methods do not effectively generalize on such
challenging real-world scenarios, deep learning-based methods have performed
relatively well when provided with sufficient training data. However, the
existing datasets are limited in size and diversity. This work proposes LS-
HDIB - a large-scale handwritten document image binarization dataset
containing over a million document images that span numerous real-world
scenarios. Additionally, we introduce a novel technique that uses a
combination of adaptive thresholding and seamless cloning methods to create
the dataset with accurate ground truths. Through an extensive quantitative and
qualitative evaluation over eight different deep learning based models, we
demonstrate the enhancement in the performance of these models when trained on
the LS-HDIB dataset and tested on unseen images.
Index Terms— Document Image Binarization, Deep Learning, Adaptive
Thresholding.
## 1 Introduction
11footnotetext: This work is supported by SERB IMPRINT-2 grant.
Handwritten document image binarization is generally modeled as a
classification problem in which intra-image pixels are assigned to either of
the two classes: handwritten content (the foreground) or the background.
Document image binarization has been an active research area for decades owing
to its importance as an essential pre-processing step in facilitating several
document image processing tasks such as optical character recognition,
handwriting matching, document translation, document summarization, and
changing the background.
Fig. 1: A few samples of handwritten document images obtained from the
proposed LS-HDIB dataset.
Handwritten documents range from ancient documents, old legal records, and
ledgers to music scores and handwritten bills. These documents often degrade
over time and become difficult to comprehend. Common degradation scenarios
include crumpled pages, poor foreground-background contrast, stains, paper
aging, fainted characters, uneven illumination, and bleed/show-through.
Furthermore, handwritten documents posses various page styles, i.e., grids,
lines, staff annotation styles, and partially blank pages that increase the
difficulty of segmenting the foreground content. The variability in the type
and the thickness of strokes also increases the complexity. These challenges
make handwritten document image binarization extremely difficult. Traditional
methods like Otsu image segmentation [1] and adaptive thresholding [2] fail to
address the aforementioned challenges completely. Since these algorithms
utilize only the low-level features, they fail to capture the wide range of
variability inherent in the handwritten documents, limiting their ability to
distinguish the background from the foreground content. The high-level
features can differentiate text pixels from background noises handling the
degradations better. However, using them solely can cause loss of low-level
information like character edges and contours, making it insufficient to
address the binarization problem. Thus far, deep learning methods have shown
promising results in segmenting foreground and background content [3, 4] by
incorporating both low-level and high-level image features. Deep learning
models rely heavily on a large amount of data for better generalization [5].
However, document image binarization lacks such a large and diverse dataset
that covers numerous real-world scenarios. While the focus has been on
designing new robust networks, little attention has been given to scale up the
existing datasets.
To address the aforementioned challenges and to generate a large scalable
dataset, we propose a Large Scale Handwritten Document Image Binarization
dataset (LS-HDIB) that contains over a million handwritten document images. We
propose a simple and effective method for generating the LS-HDIB dataset with
accurate ground truths containing segmented handwritten content (see Fig. 1).
Interestingly, the proposed method requires no manual intervention for
generating these segmented ground truths. The primary contributions of this
work are: (i) A large scale dataset (LS-HDIB) containing over a million images
with accurate ground truths for handwritten document image binarization. (ii)
A scalable and efficient method to generate and extend the proposed dataset.
## 2 Related Work
The standard approaches for document image binarization are classified into
(i) global methods[1, 6] which use a single threshold value and (ii) local
methods [7] which use adaptive threshold values for separating the foreground
and the background content. While the global methods fail to handle complex
degradations, the local methods are computationally expensive and driven
mainly by manual parameter tuning. Several deep neural network architectures
have been designed for document image binarization [8, 9, 10]. Researchers
have used fully convolutional neural networks [11, 12], recurrent neural
networks [13], encoder-decoder frameworks [9, 12], and generative adversarial
networks [11, 14] to address document binarization. While the performance of
these learning-based frameworks depends on the robustness of the network
design, another aspect critical to their performance is the size and the
versatility of the training data. The existing datasets [15, 16, 17, 18, 19,
20, 21, 22, 23, 24] do not completely span complex degradations, page styles,
and illumination variations. Owing to a very large space spanned by the
possible handwritten content and background variations, there is a need to
develop a method to create a large and diverse dataset for handwritten
document image binarization. In this work, we propose a simple yet effective
method to create a large-sized dataset that can potentially circumvent the
aforementioned limitations.
## 3 Method
### 3.1 Dataset Generation
We propose a novel data generation technique that uses a combination of
adaptive thresholding [2] and mixed gradient seamless cloning [25], as
described in Fig. 2. We collect the images of a variety of handwritten content
over a plain background and refer to these images as full-length document
images $(I_{doc})$. We collected over $450$ full-length document images.
Around $400$ images contain handwritten content in various forms such as
alpha-numeric characters, electrical circuit diagrams, control system
schematics, chemical molecular structures, and flow charts that are not
present in the existing datasets. Further, to diversify the types, thickness,
and styles of strokes, we obtained these images from $21$ different persons.
Around $50$ document images were digitally created, using Google Translate, in
$13$ different languages, including English, Urdu, Mandarin, Portuguese,
Russian, French, Hindi, Telugu, Malayalam, Punjabi, Gujarati, Japanese, and
Korean in $21$ different font styles of various sizes and colors.
Fig. 2: Block schematic of the proposed method for generating LS-HDIB dataset.
We apply the adaptive thresholding $(\mathcal{T})$ [2] on $(I_{doc})$ to
obtain the segmented ground truth images $(I_{gt})$ such that
$I_{gt}=\mathcal{T}(I_{ref})$. Next, we generate a total of $10,944$ unique
content images $(I_{c})$ by cropping and augmenting multiple patches of size
$480\times 480$ from the full-length document images $(I_{doc})$. Figure 3(a)
shows a few sample images containing a variety of written content in different
languages and font styles, including diagrams and texts from different subject
domains. While the written content (foreground) associated with the generated
ground truth remains unchanged, the background can vary depending on different
page styles and degradation scenarios. We obtain multiple document images by
merely changing the background with essentially the same ground truth.
Moreover, this method can automatically generate ground truths saving hours of
tedious manual annotation. To generate different backgrounds, we manually
capture pages with a wide variety of page styles $(I_{p})$ and degradation
effects $(I_{d})$ that are predominant in the real-world. We then use mixed
gradient-based seamless cloning $(\mathcal{C})$ [25] to blend multiple patches
of $(I_{p})$ and $(I_{d})$ to generate $20,484$ photorealistic background
images $(I_{bg})$ such that $I_{bg}=\mathcal{C}(I_{p},I_{d})$.
Datasets | Page Styles | Degradation Effects
---|---|---
| Uniform
---
ruled lines
| Non-uniform
---
ruled lines
| Grid
---
lines
| Staff notation
---
lines
| Partially
---
blank pages
Plain page | | Shadow
---
gradients
| Oily
---
patches
| Ink
---
bleed-through
| crumpled
---
pages
| Non-uniform
---
illumination
| Noisy
---
background
| Liquid
---
stains
| Poor foreground-
---
background contrast
| Punched, stapled
---
or torn pages
DIBCO09 | x | x | x | x | ✓ | ✓ | x | x | ✓ | x | x | x | ✓ | ✓ | x
HDIBCO10 | ✓ | x | x | x | x | ✓ | x | ✓ | ✓ | x | x | ✓ | ✓ | ✓ | x
DIBCO11 | x | x | x | x | x | ✓ | x | x | ✓ | x | x | ✓ | ✓ | ✓ | x
HDIBCO12 | ✓ | x | x | x | x | ✓ | x | ✓ | ✓ | x | ✓ | ✓ | ✓ | ✓ | x
DIBCO13 | x | x | x | x | x | ✓ | x | ✓ | ✓ | ✓ | x | ✓ | ✓ | ✓ | x
HDIBCO14 | x | x | x | x | x | ✓ | x | x | ✓ | x | x | x | ✓ | ✓ | x
PHIBD12 | ✓ | x | x | x | x | ✓ | x | ✓ | ✓ | x | x | ✓ | ✓ | ✓ | ✓
HDIBCO16 | x | x | x | x | x | ✓ | x | ✓ | ✓ | x | ✓ | ✓ | ✓ | ✓ | x
DIBCO17 | x | x | x | x | x | ✓ | x | ✓ | ✓ | x | x | ✓ | ✓ | x | x
DIBCO18 | x | x | x | x | x | ✓ | x | ✓ | ✓ | x | x | ✓ | ✓ | ✓ | x
Bickley Diary | x | x | x | x | ✓ | ✓ | x | ✓ | x | x | x | ✓ | ✓ | ✓ | x
Palm Leaf Manuscript | x | x | x | x | ✓ | ✓ | ✓ | ✓ | x | x | x | x | ✓ | ✓ | ✓
LS-HDIB (Ours) | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ | x
Table 1: Comparison of page styles and degradation effects available across
different publicly available datasets and LS-HDIB dataset. Fig. 3: A few
sample images depicting different page styles available in LS-HDIB dataset.
For a better visual understanding, we show multiple example images of
different page styles and degradation effects in Fig. 3(a) and Fig. 3(b),
respectively. Finally, we combine the content images $(I_{c})$ and the
background images $(I_{bg})$ again using seamless cloning $(\mathcal{C})$ to
generate the handwritten document images $(I_{in})$ such that
$I_{in}=\mathcal{C}(I_{c},I_{bg})$. To generate the LS-HDIB dataset, we
randomly sample $100$ background images $(I_{bg})$ for each of the $10,994$
handwritten content images $(I_{c})$. In this way, we obtain a total of $1.09$
million images in the LS-HDIB dataset. It is important to note that we can
scale up the dataset size for up to $200$ times by considering all the
background images instead of just $100$. With adaptive thresholding and mixed
gradient-based seamless cloning, we have been able to generate accurate ground
truths without the requirement of any manual annotation and generate a wide
variety of degraded handwritten document images. In addition to the inherent
scalability and ease of ground truth generation, the proposed dataset
generation method is computationally less expensive compared to the other
deep-learning-based generative methods such as the one proposed in [14]. As
evident from Table 1, most of the publicly available datasets contain the
written content only over plain pages. In contrast, the LS-HDIB dataset
includes images with various page styles like ruled lines, gridlines, and
partially blank pages that are evident in our day-to-day encounters. Further,
they lack the document images with realistic degradations like crumpled pages,
non-uniform illumination, and shadow gradients that are well incorporated in
the proposed LS-HDIB dataset. These attributes enhance the diversity and the
versatility of the proposed dataset.
### 3.2 Training Details
We use eight widely used deep networks - DeepLabV3 [26], DeepLabV3+ [27],
Feature Pyramid Networks (FPN) [28], LinkNet [29], Multi-scale Attention net
(MANet) [30], Pyramid Attention Network (PAN) [31], Pyramid Scene Parsing
Network (PSPN) [32], and U-Net [33] \- to understand the effectiveness of the
proposed dataset for the handwritten document image binarization task. We have
carefully chosen these networks as they collectively span the various deep
learning based approaches [11, 12, 9, 14] that have been adopted thus far for
the binarization task.
Each of the eight networks is trained under three different training regimes.
We follow the standard train, validation, and test split of $80\%$, $10\%$,
and $10\%$, respectively, for each regime.
(i) Regime 1: The deep models are trained only on the baseline dataset
obtained by combining ten different publicly available datasets: DIBCO09 [15],
HDIBCO10 [16], DIBCO11 [17], HDIBCO12[18], DIBCO13 [19], HDIBCO14 [20],
PHIBD12 [24], HDIBCO16 [21], DIBCO17 [22], DIBCO18 [23]. However, even after
combining ten different datasets, the size of the baseline dataset is
relatively small (order of 1000 images). Therefore, we crop the full-length
document images of the baseline dataset to the size $480\times 480$ with a
stride of $240$ pixels and perform rotation ($90^{\circ},180^{\circ}$, and
$270^{\circ}$) and horizontal flip to obtain around $6000$ images, $5000$ for
training and $1000$ for testing. While the size of the baseline dataset can
further be increased by reducing the stride value, this leads to greater
overlap and high redundancy in the foreground content across different content
images.
(ii) Regime 2: Each deep model is trained only on the LS-HDIB training set.
Although the LS-HDIB dataset has over 1 million images, we use only $5000$
images for training to have a fair performance comparison.
(iii) Regime 3: We combine both the LS-HDIB and the baseline dataset, by
randomly selecting $2500$ images from each dataset to train all the deep
models on a total of $5000$ images. Regime 3 is targeted towards establishing
the efficacy of augmenting the proposed dataset to the existing ones.
The models are trained for a maximum of $20$ epochs with learning rate of
$0.01$ using the Adam optimizer with default parameters. The training is
performed on the NVIDIA RTX 2080 Ti GPU with the batch size of $8$.
Loss Function. We use binary cross-entropy loss to train the segmentation
models, as described in Equation 1.
$\mathcal{L}=-I_{gt}\mathrm{log}\left(\mathcal{F}(I_{in})\right)-(1-I_{gt})\mathrm{log}\left(1-\mathcal{F}(I_{in})\right)$
(1)
Here, $\mathcal{F}$ represents the functional form of deep model. Binary
cross-entropy loss is found to be more effective than MSE loss for
classification tasks [34].
Fig. 4: Qualitative result on (a) the LS-HDIB test set (b) Bickley Diary
dataset, and (c) Palm Leaf Manuscript dataset.
## 4 Experimental Analysis
We demonstrate the effectiveness of the proposed dataset for handwritten
document image binarization through an extensive quantitative and qualitative
analysis. We compare the performance of different deep models trained to
observe how well the LS-HDIB dataset enhances the generalization capability of
the deep models.
We use four different datasets containing challenging scenarios for testing
the deep models: the Bickley Diary dataset [35], the Palm Leaf Manuscript
dataset [36], the DIBCO test set, and the LS-HDIB test set. We use three
popular metrics to evaluate the network performance trained under different
regimes: F-measure $(\mathrm{F_{score}})$ [37], pseudo F-measure
$(\mathrm{PF_{score}})$ [37], and Peak Signal to Noise Ratio $(\mathrm{PSNR})$
[37].
As shown in Table 2, the $\mathrm{F_{score}}$, $\mathrm{PF_{score}}$, and
$\mathrm{PSNR}$ is maximum over all the deep models trained under Regime 2 for
LS-HDIB and the Bickley Diary dataset. Further, the performance under Regime 3
is better than that of Regime 1, indicating that augmenting the LS-HDIB
dataset to the baseline dataset enhances network performance. Qualitatively,
the foreground content of the image affected by liquid stains, noisy
background, and poor foreground to background contrast is well recovered under
Regime 2, as shown in Fig. 4(a) and 4(b). For the Palm Leaf dataset, the
strokes in the estimated foreground content corresponding to Regime 1 (and 3)
are relatively thicker when compared to those corresponding to Regime 2 across
all the models (see Fig. 4(c)). Since the stroke widths are consistent with
the ground truth, the $\mathrm{PSNR}$ continues to be higher for Regime 2.
However, the $\mathrm{F_{score}}$ and $\mathrm{PF_{score}}$ of some models
under Regime 2 and 3 are less than Regime 1. This is attributed to the
presence of minor discontinuities in the strokes obtained under Regime 2 when
compared to that of Regime 1 (see Fig. 4(c)). However, on average, the overall
performance of each of the eight deep models is the highest when trained under
Regime 2 across the three different test datasets, as evident from Table 2.
Further, Fig. 4(c) depicts that models under Regime 1 and 3 fail to segment
out the thread punched out through the document. However, almost every deep
model trained on the LS-HDIB dataset (Regime 2) precisely identifies the
appropriate foreground content even in the presence of such background
artifacts.
For the DIBCO test set, the performance under Regime 1 is better than Regime 2
(Table 2). This is inline with the expectation as the networks are trained on
DIBCO train set itself. Interestingly, Regime 3 offers the best performance
across different models and test sets indicating that the LS-HDIB dataset when
augmented with the standard DIBCO datasets enhances the network performance.
Metric | Dataset | DeepLabv3 | DeepLabv3++ | FPN | LinkNet | MANet | PANet | PSPNet | U-Net
---|---|---|---|---|---|---|---|---|---
| | 0.64 | 0.71 | 0.72 | 0.71 | 0.72 | 0.72 | 0.58 | 0.74
| | 0.66 | 0.77 | 0.77 | 0.82 | 0.84 | 0.74 | 0.65 | 0.83
| Bickley Diary | 0.60 | 0.75 | 0.74 | 0.80 | 0.82 | 0.69 | 0.60 | 0.77
| | 0.49 | 0.59 | 0.58 | 0.62 | 0.62 | 0.57 | 0.47 | 0.62
| | 0.48 | 0.57 | 0.59 | 0.60 | 0.62 | 0.57 | 0.48 | 0.60
| Palm Leaf | 0.48 | 0.58 | 0.57 | 0.61 | 0.62 | 0.57 | 0.44 | 0.61
| | 0.48 | 0.68 | 0.60 | 0.66 | 0.68 | 0.56 | 0.50 | 0.67
| | 0.43 | 0.57 | 0.46 | 0.58 | 0.60 | 0.41 | 0.43 | 0.58
| DIBCO | 0.52 | 0.62 | 0.61 | 0.72 | 0.71 | 0.58 | 0.46 | 0.72
| | 0.44 | 0.52 | 0.54 | 0.56 | 0.57 | 0.51 | 0.42 | 0.55
| | 0.54 | 0.68 | 0.68 | 0.86 | 0.87 | 0.65 | 0.53 | 0.87
F-score | LS-HDIB | 0.53 | 0.66 | 0.67 | 0.85 | 0.85 | 0.64 | 0.52 | 0.86
| | 0.64 | 0.76 | 0.75 | 0.80 | 0.79 | 0.71 | 0.59 | 0.78
| | 0.66 | 0.79 | 0.79 | 0.87 | 0.89 | 0.76 | 0.66 | 0.89
| Bickley Diary | 0.60 | 0.77 | 0.75 | 0.82 | 0.80 | 0.73 | 0.61 | 0.82
| | 0.50 | 0.60 | 0.57 | 0.61 | 0.62 | 0.58 | 0.48 | 0.63
| | 0.48 | 0.58 | 0.59 | 0.63 | 0.64 | 0.57 | 0.48 | 0.62
| Palm Leaf | 0.48 | 0.59 | 0.58 | 0.62 | 0.64 | 0.57 | 0.44 | 0.63
| | 0.49 | 0.68 | 0.60 | 0.66 | 0.68 | 0.56 | 0.50 | 0.67
| | 0.41 | 0.52 | 0.51 | 0.58 | 0.51 | 0.41 | 0.43 | 0.58
| DIBCO | 0.52 | 0.62 | 0.62 | 0.73 | 0.72 | 0.59 | 0.47 | 0.72
| | 0.43 | 0.51 | 0.53 | 0.56 | 0.56 | 0.51 | 0.42 | 0.54
| | 0.53 | 0.67 | 0.66 | 0.87 | 0.88 | 0.64 | 0.52 | 0.88
PF-score | LS-HDIB | 0.52 | 0.65 | 0.66 | 0.86 | 0.86 | 0.63 | 0.52 | 0.87
| | 10.07 | 12.32 | 12.11 | 13.16 | 12.83 | 11.55 | 9.61 | 12.89
| | 10.06 | 12.75 | 12.69 | 14.42 | 14.82 | 12.10 | 10.01 | 14.64
| Bickley Diary | 9.69 | 12.40 | 12.29 | 13.41 | 13.49 | 11.64 | 9.78 | 13.24
| | 9.02 | 10.53 | 10.68 | 11.04 | 11.27 | 10.41 | 9.04 | 11.11
| | 8.93 | 11.57 | 11.47 | 12.45 | 12.31 | 11.12 | 8.99 | 12.41
| Palm Leaf | 9.19 | 11.34 | 11.36 | 12.26 | 12.24 | 11.03 | 9.36 | 12.32
| | 8.26 | 9.76 | 9.63 | 10.84 | 11.13 | 9.01 | 7.90 | 11.06
| | 6.75 | 8.97 | 8.74 | 10.96 | 9.54 | 10.53 | 6.56 | 11.01
| DIBCO | 8.38 | 10.47 | 10.34 | 12.54 | 12.36 | 9.74 | 8.39 | 12.56
| | 7.58 | 9.29 | 9.83 | 9.63 | 9.86 | 9.26 | 7.49 | 9.05
| | 9.39 | 12.15 | 12.05 | 16.99 | 17.09 | 11.58 | 9.37 | 17.29
PSNR | LS-HDIB | 9.11 | 11.88 | 12.08 | 16.63 | 16.74 | 11.33 | 9.10 | 17.07
Table 2: Fscore, PFscore, and PSNR evaluated over eight deep models under
three different regimes: Regime 1, Regime 2, and Regime 3 over test datasets.
Fig. 5: Effect of varying dataset size on the model performance evaluated over
the three test datasets.
Given that nearly all the models have performed the best when trained on LS-
HDIB dataset across all the three test datasets, we further investigate the
effect of dataset size on the network performance. For LS-HDIB test dataset,
the $\mathrm{F_{score}}$, $\mathrm{PF_{score}}$, and $\mathrm{PSNR}$ are
observed to increase with the dataset size, as shown in Fig. 5. This indicates
the requirement of a large-scale dataset to span complex real-world scenarios
encountered in the handwritten document images. Fig. 5 shows that the
performance over the Bickley Diary and Palm Leaf dataset peaks at the dataset
size of $20$K and $5$K, respectively. Overall, we have established that the
deep models are more robust to various degradation effects and page styles
encountered in the real world when trained on the LS-HDIB dataset and
necessitate the need for such a dataset.
Note. Owing to the space constraints, we provide more qualitative results,
detailed dataset statistics, training and validation logs for different models
(for better selectivity), and the accompanying code on our
website111https://kaustubh-sadekar.github.io/LS-HDIB/.
## 5 Conclusion
We propose a large-scale dataset (LS-HDIB) for handwritten document image
binarization and a simple yet effective method to generate it. When trained on
the LS-HDIB dataset, different deep models can generalize better on unseen
document images with a wide variety of degradations encountered in our day-to-
day lives. This is possible due to the inherent diversity of the proposed
dataset. Further, this work highlights that the fundamental image processing
algorithms can be used as practical tools to support the existing deep-
learning-based methods in producing significantly better results. In our case,
we use adaptive thresholding and mixed gradient-based seamless cloning to
generate this large-scale dataset.
## References
* [1] N. Otsu, “A threshold selection method from gray-level histograms,” IEEE Transactions on Systems, Man, and Cybernetics, vol. 9, no. 1, pp. 62–66, 1979.
* [2] Pierre D Wellner, “Adaptive thresholding for the digitaldesk,” Xerox, EPC1993-110, pp. 1–19, 1993.
* [3] Chris Tensmeyer and T. Martinez, “Document image binarization with fully convolutional neural networks,” 2017 14th IAPR International Conference on Document Analysis and Recognition (ICDAR), vol. 01, pp. 99–104, 2017.
* [4] Quang Nhat Vo, Soo Hyung Kim, Hyung Jeong Yang, and Gueesang Lee, “Binarization of degraded document images based on hierarchical deep supervised network,” Pattern Recogn., vol. 74, no. C, pp. 568–586, Feb. 2018.
* [5] Sergey Levine, Peter Pastor, Alex Krizhevsky, Julian Ibarz, and Deirdre Quillen, “Learning hand-eye coordination for robotic grasping with deep learning and large-scale data collection,” The International Journal of Robotics Research, vol. 37, no. 4-5, pp. 421–436, 2018.
* [6] Wen-Hsiang Tsai, “Moment-preserving thresolding: A new approach,” Computer vision, graphics, and image processing, vol. 29, no. 3, pp. 377–393, 1985.
* [7] Basilios Gatos, Ioannis Pratikakis, and Stavros J Perantonis, “Adaptive degraded document image binarization,” Pattern recognition, vol. 39, no. 3, pp. 317–327, 2006.
* [8] Joan Pastor-Pellicer, S España-Boquera, Francisco Zamora-Martínez, M Zeshan Afzal, and Maria Jose Castro-Bleda, “Insights on the use of convolutional neural networks for document image binarization,” in International Work-Conference on Artificial Neural Networks. Springer, 2015, pp. 115–126.
* [9] Jorge Calvo-Zaragoza, Gabriel Vigliensoni, and Ichiro Fujinaga, “Pixel-wise binarization of musical documents with convolutional neural networks,” in 2017 Fifteenth IAPR International Conference on Machine Vision Applications (MVA). IEEE, 2017, pp. 362–365.
* [10] Muhammad Zeshan Afzal, Joan Pastor-Pellicer, Faisal Shafait, Thomas M Breuel, Andreas Dengel, and Marcus Liwicki, “Document image binarization using lstm: A sequence learning approach,” in Proceedings of the 3rd international workshop on historical document imaging and processing, 2015, pp. 79–84.
* [11] Chris Tensmeyer, Mike Brodie, Daniel Saunders, and Tony Martinez, “Generating realistic binarization data with generative adversarial networks,” in 2019 International Conference on Document Analysis and Recognition (ICDAR). IEEE, 2019, pp. 172–177.
* [12] Xujun Peng, Huaigu Cao, and Prem Natarajan, “Using convolutional encoder-decoder for document image binarization,” in 2017 14th IAPR International Conference on Document Analysis and Recognition (ICDAR). IEEE, 2017, vol. 1, pp. 708–713.
* [13] Florian Westphal, Niklas Lavesson, and Håkan Grahn, “Document image binarization using recurrent neural networks,” in 2018 13th IAPR International Workshop on Document Analysis Systems (DAS). IEEE, 2018, pp. 263–268.
* [14] A. K. Bhunia, A. K. Bhunia, A. Sain, and P. P. Roy, “Improving document binarization via adversarial noise-texture augmentation,” in 2019 IEEE International Conference on Image Processing (ICIP), 2019, pp. 2721–2725.
* [15] B. Gatos, K. Ntirogiannis, and I. Pratikakis, “Icdar 2009 document image binarization contest (dibco 2009),” in 2009 10th International Conference on Document Analysis and Recognition, 2009, pp. 1375–1382.
* [16] I. Pratikakis, B. Gatos, and K. Ntirogiannis, “H-dibco 2010 - handwritten document image binarization competition,” in 2010 12th International Conference on Frontiers in Handwriting Recognition, 2010, pp. 727–732.
* [17] I. Pratikakis, B. Gatos, and K. Ntirogiannis, “Icdar 2011 document image binarization contest (dibco 2011),” in 2011 International Conference on Document Analysis and Recognition, 2011, pp. 1506–1510.
* [18] I. Pratikakis, B. Gatos, and K. Ntirogiannis, “Icfhr 2012 competition on handwritten document image binarization (h-dibco 2012),” in 2012 International Conference on Frontiers in Handwriting Recognition, 2012, pp. 817–822.
* [19] I. Pratikakis, B. Gatos, and K. Ntirogiannis, “Icdar 2013 document image binarization contest (dibco 2013),” in 2013 12th International Conference on Document Analysis and Recognition, 2013, pp. 1471–1476.
* [20] K. Ntirogiannis, B. Gatos, and I. Pratikakis, “Icfhr2014 competition on handwritten document image binarization (h-dibco 2014),” in 2014 14th International Conference on Frontiers in Handwriting Recognition, 2014, pp. 809–813.
* [21] Ioannis Pratikakis, Konstantinos Zagoris, George Barlas, and Basilis Gatos, “Icfhr2016 handwritten document image binarization contest (h-dibco 2016),” 10 2016, pp. 619–623.
* [22] Ioannis Pratikakis, Konstantinos Zagoris, George Barlas, and Basilis Gatos, “Icdar2017 competition on document image binarization (dibco 2017),” 11 2017, pp. 1395–1403.
* [23] Ioannis Pratikakis, Konstantinos Zagori, Panagiotis Kaddas, and Basilis Gatos, “Icfhr 2018 competition on handwritten document image binarization (h-dibco 2018),” in 2018 16th International Conference on Frontiers in Handwriting Recognition (ICFHR), 2018, pp. 489–493.
* [24] H. Z. Nafchi, S. M. Ayatollahi, R. F. Moghaddam, and M. Cheriet, “An efficient ground truthing tool for binarization of historical manuscripts,” in 2013 12th International Conference on Document Analysis and Recognition, 2013, pp. 807–811.
* [25] Patrick Pérez, Michel Gangnet, and Andrew Blake, “Poisson image editing,” in ACM SIGGRAPH 2003 Papers, New York, NY, USA, 2003, SIGGRAPH ’03, p. 313–318, Association for Computing Machinery.
* [26] Liang-Chieh Chen, G. Papandreou, Florian Schroff, and Hartwig Adam, “Rethinking atrous convolution for semantic image segmentation,” ArXiv, vol. abs/1706.05587, 2017.
* [27] Liang-Chieh Chen, Yukun Zhu, George Papandreou, Florian Schroff, and Hartwig Adam, “Encoder-decoder with atrous separable convolution for semantic image segmentation,” in Computer Vision – ECCV 2018, 2018.
* [28] Selim Seferbekov, Vladimir Iglovikov, Alexander Buslaev, and Alexey Shvets, “Feature pyramid network for multi-class land segmentation,” in 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), 2018, pp. 272–2723.
* [29] Abhishek Chaurasia and Eugenio Culurciello, “Linknet: Exploiting encoder representations for efficient semantic segmentation,” 2017 IEEE Visual Communications and Image Processing (VCIP), Dec 2017.
* [30] Tongle Fan, Guanglei Wang, Yan Li, and Hongrui Wang, “Ma-net: A multi-scale attention network for liver and tumor segmentation,” IEEE Access, vol. 8, pp. 179656–179665, 2020.
* [31] Hanchao Li, Pengfei Xiong, Jie An, and Lingxue Wang, “Pyramid attention network for semantic segmentation,” 2018.
* [32] Hengshuang Zhao, Jianping Shi, Xiaojuan Qi, Xiaogang Wang, and Jiaya Jia, “Pyramid scene parsing network,” in 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2017, pp. 6230–6239.
* [33] Olaf Ronneberger, Philipp Fischer, and Thomas Brox, “U-net: Convolutional networks for biomedical image segmentation,” in MICCAI 2015. 2015, pp. 234–241, Springer International Publishing.
* [34] Pavel Golik, Patrick Doetsch, and Hermann Ney, “Cross-entropy vs. squared error training: a theoretical and experimental comparison,” 08 2013, pp. 1756–1760.
* [35] A. F. Bickley., “Bickley diary dataset,” 1926\.
* [36] Jean, Mickaël, Setiawan, Kesiman, and MarcOgier, “Cfhr 2016 competition on the analysis of handwritten text in images of balinesepalm leaf manuscripts,” 10 2016, pp. 596–601.
* [37] K. Ntirogiannis, B. Gatos, and I. Pratikakis, “Performance evaluation methodology for historical document image binarization,” IEEE Transactions on Image Processing, vol. 22, no. 2, pp. 595–609, 2013.
|
# Nilpotent varieties in symmetric spaces and twisted affine Schubert
varieties
Jiuzu Hong Department of Mathematics, University of North Carolina at Chapel
Hill, Chapel Hill, NC 27599-3250, U.S.A<EMAIL_ADDRESS>and Korkeat
Korkeathikhun Department of Mathematics, University of North Carolina at
Chapel Hill, Chapel Hill, NC 27599-3250, U.S.A<EMAIL_ADDRESS>
###### Abstract.
We relate the geometry of Schubert varieties in twisted affine Grassmannian
and the nilpotent varieties in symmetric spaces. This extends some results of
Achar-Henderson in the twisted setting. We also get some applications to the
geometry of the order 2 nilpotent varieties in certain classical symmetric
spaces.
## 1\. Introduction
Let $G$ be a reductive group over $\mathbb{C}$. Let $\mathcal{N}$ denote the
nilpotent cone of the Lie algebra $\mathfrak{g}$ of $G$. Let
$\operatorname{Gr}_{G}$ be the affine Grassmannian of $G$. Each spherical
Schubert cell $\operatorname{Gr}_{\lambda}$ is parametrized by a dominant
coweight $\lambda$. When $G=\mathrm{GL}_{n}$, Lusztig [Lu] defined an
embedding from $\mathcal{N}$ to $\operatorname{Gr}_{G}$, and showed that each
nilpotent variety in $\mathfrak{gl}_{n}$ can be openly embedded into certain
affine Schubert variety $\overline{\operatorname{Gr}}_{\lambda}$. This
embedding identifies the geometry of nilpotent varieties and certain affine
Schubert varieties in type $A$. However, there is no direct generalization for
general reductive groups.
In [AH], Achar-Henderson took a different idea for a general algebraic simple
group $G$. Let $\operatorname{Gr}_{0}^{-}$ be the opposite open Schubert cell
in $\operatorname{Gr}_{G}$. One can naturally define a map
$\pi:\operatorname{Gr}_{0}^{-}\to\mathfrak{g}$. Achar-Henderson showed that
$\pi(\operatorname{Gr}_{0}^{-}\cap\operatorname{Gr}_{\lambda})$ is contained
in $\mathcal{N}$ if and only if $\lambda$ is small in the sense of Broer [Br]
and Reeder [Re], i.e. $\lambda\nsucceq 2\gamma_{0}$, where $\gamma_{0}$ is the
highest short coroot of $G$. They also proved that $\pi:\operatorname{Gr}_{\rm
sm}\cap\operatorname{Gr}_{0}^{-}\to\pi(\operatorname{Gr}_{\rm
sm}\cap\operatorname{Gr}_{0}^{-})$ is a finite map whose fibers admits
transitive $\mathbb{Z}/2\mathbb{Z}$-actions, where $\operatorname{Gr}_{\rm
sm}$ is the union of all $\operatorname{Gr}_{\lambda}$ such that $\lambda$ is
small. Moreover, with respect to $\pi$, Achar-Henderson [AH, AH2] related the
geometric Satake correspondence and Springer correspondence.
In this paper, we consider a twisted analogue, and we will extend some results
of Achar-Henderson in [AH]. Let $\sigma$ be a diagram automorphism of order
$2$, and let $\sigma$ act on the field $\mathcal{K}=\mathbb{C}((t))$ via
$\sigma(t)=-t$ and $\sigma|_{\mathbb{C}}={\rm Id}_{\mathbb{C}}$. Then, we may
define a twisted affine Grassmannian
$\mathcal{G}r:=G(\mathcal{K})^{\sigma}/G(\mathcal{O})^{\sigma}$, where
$\mathcal{O}=\mathbb{C}[[t]]$. Each twisted Schubert cell
$\mathcal{G}r_{\bar{\lambda}}$, i.e. a $G(\mathcal{O})^{\sigma}$-orbit, is
parametrized by the image $\bar{\lambda}$ of a dominant coweight $\lambda$ in
the coinvariant lattice $X_{*}(T)_{\sigma}$ with respect to the induced action
of $\sigma$, where $X_{*}(T)$ is the coweight lattice of $G$. In fact,
$X_{*}(T)_{\sigma}$ can be regarded as the weight lattice of a reductive group
$H:=(\check{G})^{\sigma}$, where $\check{G}$ is the Langlands dual group of
$G$.
Let $\mathcal{G}r_{0}^{-}$ be the opposite open Schubert cell in
$\mathcal{G}r$. We may naturally define a map
$\pi:\mathcal{G}r_{0}^{-}\to\mathfrak{p}$, where $\mathfrak{p}$ is the
(-1)-eigenspace of $\sigma$ in $\mathfrak{g}$. Let
$\mathcal{M}_{\bar{\lambda}}$ denote the intersection
$\mathcal{G}r_{\bar{\lambda}}\cap\mathcal{G}r_{0}^{-}$, which is a nonempty
open subset of $\mathcal{G}r_{\bar{\lambda}}$. The following theorem is the
main result of this paper, and it can follow from Proposition 2.3 in Section
2.1 and Theorem 4.2 in Section 4, based on case-by-case analysis.
###### Theorem 1.1.
Assume that $G$ is of type $A_{\ell}$ or $D_{\ell+1}$. The image
$\pi(\mathcal{M}_{\lambda})$ is contained in the nilpotent cone
$\mathcal{N}_{\mathfrak{p}}$ of $\mathfrak{p}$, if and only if $\bar{\lambda}$
is a small dominant weight with respect to $H$.
Actually, in Theorem 4.2 we describe precisely
$\pi(\mathcal{M}_{\bar{\lambda}})$ as a union of nilpotent orbits in
$\mathfrak{p}$ for each small $\bar{\lambda}$. In Theorem 4.5, Theorem 4.6,
and Theorem 4.14, we also determine all small $\bar{\lambda}$ such that
$\pi(\mathcal{M}_{\bar{\lambda}})$ is a nilpotent orbit and
$\pi:\mathcal{M}_{\bar{\lambda}}\to\pi(\mathcal{M}_{\bar{\lambda}})$ is an
isomorphism. Furthermore, We describe all fibers of
$\pi:\mathcal{M}\to\pi(\mathcal{M})$ in Proposition 4.11 and Proposition 4.15,
where $\mathcal{M}$ is the union of all small twisted Schubert cells. The
fibers are closely related to anti-commuting nilpotent varieties for symmetric
spaces. When $G$ is of type $A_{2\ell-1}$ (resp. $D_{\ell+1}$), the fiber
$\pi^{-1}(0)$ is actually the minimal (resp. maximal) order 2 nilpotent
variety in $\mathfrak{sp}_{2\ell}$ (resp. $\mathfrak{so}_{2\ell+1}$). This is
a very different phenomenon with the untwisted setting in the work of Achar-
Henderson [AH], and it actually makes the twisted setting more challenging.
For general simple Lie algebra $\mathfrak{g}$ and general diagram automorphism
$\sigma$, it was proved in [HLR, Appendix C] by Haines-Lourenço-Richarz that,
when $\bar{\lambda}$ is quasi-miniscule and $\overline{\mathcal{O}}$ is the
minimal nilpotent variety in $\mathfrak{p}$, the map
$\pi:\overline{\mathcal{G}r}_{\bar{\lambda}}\cap\mathcal{G}r_{0}^{-}\to\overline{\mathcal{O}}$
is an isomorphism. In fact, we have also obtained this result independently,
cf. [Ko]. Also, under the same assumption as in Theorem 1.1, this isomorphism
is a special case of our Theorem 4.5, Theorem 4.6, and Theorem 4.14.
The geometric Satake correspondence for $\mathcal{G}r$ was proved by Zhu [Zh],
and it exactly recovers the Tannakian group $H$. On the other hand, the
Springer correspondence for symmetric spaces is more sophisticated than the
usual Lie algebra setting, see a survey on this subject [Sh]. It would be
interesting to relate these two pictures as was done in [AH, AH2]. Y. Li [Li]
defined the symmetric space analogue called $\sigma$-quiver variety in the
setting of Nakajima quiver variety, and he showed that certain $\sigma$-quiver
variety can be identified with null-cone of symmetric spaces. It is an
interesting question to investigate a connection between $\sigma$-quiver
variety and twisted affine Grassmannian in the spirit of the work of Mirković-
Vybornov [MV].
From Theorem 1.1, we can deduce some applications for the order 2 nilpotent
varieties in classical symmetric spaces. Let $\langle,\rangle$ be a symmeric
or symplectic non-degenerate bilinear form on a vector space $V$. Let
$\mathcal{A}$ be the space of self-adjoint linear maps with respect to
$\langle,\rangle$. We consider ${\rm Sp}_{2n}$-action on $\mathcal{A}$ when
$\langle,\rangle$ is symplectic and $\dim V=2n$, and ${\rm SO}_{n}$-action
when $\langle,\rangle$ is symmetric and $\dim V=n$. In Section 5, we obtain
the following results.
###### Theorem 1.2.
1. (1)
If $\langle,\rangle$ is symmetric and $\dim V$ is odd, then any order 2
nilpotent variety in $\mathcal{A}$ is normal.
2. (2)
If $\langle,\rangle$ is symplectic, then there is a bijection of order 2
nilpotent varieties in $\mathfrak{so}_{2n+1}$ and in $\mathcal{A}$, such that
they have the same cohomology of stalks of IC-sheaves.
3. (3)
If $\langle,\rangle$ is symplectic, the smooth locus of any order 2 nilpotent
variety in $\mathcal{A}$ is the open nilpotent orbit.
It is known that when $\langle,\rangle$ is symplectic, any nilpotent variety
in $\mathcal{A}$ is normal, but it is not always true when $\langle,\rangle$
is symmetric, cf. [Oh]. Using our methods, we can also prove that there is a
bijection of order 2 nilpotent varieties in $\mathfrak{sp}_{2n}$ and in the
space of symmetric $(2n+1)\times(2n+1)$ matrices, such that they have the same
cohomlogy of stalks of IC-sheaves. This was already proved earlier by Chen-
Vilonen-Xue [CVX] using different methods. Also, Part 3) of Theorem 1.2 is not
true when $\langle,\rangle$ is symmetric and $\dim V$ is odd, see more
detailed discussions in Section 5.
Acknowledgments: This project grew out from a conversation with Yiqiang Li in
March 2019. We would like to thank him for inspiring discussions. J. Hong is
partially supported by NSF grant DMS-2001365.
## 2\. Notation and Preliminaries
### 2.1. Root datum
Let $G$ be a simply-connected simple algebraic group over $\mathbb{C}$, and
let $\mathfrak{g}$ be its Lie algebra. Let $\sigma$ be a diagram automorphism
of $G$ of order $r$, preserving a maximal torus $T$ and a Borel subgroup $B$
containing $T$ in $G$. Then $G$ has a root datum
$(X_{*}(T),X^{*}(T),\langle\cdot,\cdot\rangle,\check{\alpha}_{i},\alpha_{i},i\in
I)$ with the action of $\sigma$, where
* •
$X_{*}(T)$ (resp. $X^{*}(T)$ ) is the coweight (resp. weight) lattice;
* •
$I$ is the set of vertices of the Dynkin diagram of $G$;
* •
$\alpha_{i}$ (resp. $\check{\alpha}_{i}$) is the simple root (resp. coroot)
for each $i\in I$;
* •
$\langle\cdot,\cdot\rangle:X_{*}(T)\times X^{*}(T)\to\mathbb{Z}$ is the
perfect pairing.
The automorphism $\sigma$ of this root datum satisfies
* •
$\sigma(\alpha_{i})=\alpha_{\sigma(i)}$ and
$\sigma(\check{\alpha}_{i})=\check{\alpha}_{\sigma(i)}$;
* •
$\langle\sigma(\check{\lambda}),\sigma(\mu)\rangle=\langle\check{\lambda},\mu\rangle$
for any $\check{\lambda}\in X_{*}(T)$ and $\mu\in X^{*}(T)$.
As a diagram autormophism on $G$, $\sigma$ also preserves a pinning with
respect to $B$ and $T$, i.e. there exists root subgroups $x_{i},y_{i}$
associated to $\alpha_{i},-\alpha_{i}$ for each $i\in$, such that
$\sigma(x_{i}(a))=x_{\sigma(i)}(a),\quad\sigma(y_{i}(a))=y_{\sigma(i)}(a),\quad\text{
for any }a\in\mathbb{C}.$
Let $I_{\sigma}$ be the set of $\sigma$-orbits in $I$. Denote
$X^{*}(T)^{\sigma}=\\{\lambda\in X^{*}(T)\mid\sigma\lambda=\lambda\\}$ and
$X_{*}(T)_{\sigma}=X_{*}(T)/(\mathrm{Id}-\sigma)X_{*}(T)$. For each $\imath\in
I_{\sigma}$, define $\gamma_{\imath}=\bar{\check{\alpha}}_{i}\in
X_{*}(T)_{\sigma}$ for any $i\in\imath$, and define
$\check{\gamma}_{\imath}\in X^{*}(T)^{\sigma}$ by
$\check{\gamma}_{\imath}=\begin{cases}\sum_{i\in\imath}\alpha_{i}&\quad\text{if
no pairs in }\imath\text{ is adjacent, }\\\
2\sum_{i\in\imath}\alpha_{i}&\quad\text{if }\imath=\\{i,\sigma(i)\\}\text{ and
}i\text{ and }\sigma(i)\text{ are adjacent,}\\\ \alpha_{i}&\quad\text{if
}\imath=\\{i\\}.\end{cases}$
Let $\check{G}$ denote the Langlands dual group of $G$, and we still denote
the induced diagram automorphism on $\check{G}$ by $\sigma$. Denoted by
$H=(\check{G})^{\sigma}$ the $\sigma$-fixed subgroup of $\check{G}$. Then, $H$
has the root datum
$(X^{*}(T)^{\sigma},X_{*}(T)_{\sigma},\check{\gamma_{\imath}},\gamma_{\imath},\imath\in
I_{\sigma})$, cf. [HS, Section 2.2]. For $\bar{\lambda},\bar{\mu}\in
X_{*}(T)_{\sigma}$, define the partial order $\bar{\mu}\preceq\bar{\lambda}$
if $\bar{\lambda}-\bar{\mu}$ is a sum of positive roots of $H$. Let
$X_{*}(T)_{\sigma}^{+}$ be the set of dominant weight of $H$. In fact,
$X_{*}(T)_{\sigma}^{+}$ is the image of the quotient map $X_{*}(T)^{+}\to
X_{*}(T)_{\sigma}$, where $X_{*}(T)^{+}$ is the set of dominant weights of
$G$.
### 2.2. Twisted affine Grassmannian
Let $\sigma$ be a diagram automorphism of $G$ of order $r$. Let $\mathcal{O}$
denote the set of formal power series in $t$ with coefficients in $\mathbb{C}$
and denote $\mathcal{K}$ the set of Laurent series in $t$ with coefficients in
$\mathbb{C}$. Denote the automorphism $\sigma$ of order $r$ on $\mathcal{K}$
and $\mathcal{O}$ given by $\sigma$ acts trivially on $\mathbb{C}$ and maps
$t\to\epsilon t$ where we fix the primitive $r$-root of unity $\epsilon$. We
consider the following twisted affine Grassmannian attached to $G$ and
$\sigma$,
$\mathcal{G}r_{G}=G(\mathcal{K})^{\sigma}/G(\mathcal{O})^{\sigma}.$
This space has been studied intensively in [BH, HR, PR, Ri]. The ramified
geometric Satake correspondence [Zh] asserts that there is an equivalence
between the category of spherical perverse sheaves on $\mathcal{G}r_{G}$ and
the category of representations of the algebraic group
$H=(\check{G})^{\sigma}$. If there is no confusion, we write $\mathcal{G}r$
for convenience.
Let $e_{0}$ be the based point in $\mathcal{G}r$. For any $\lambda\in
X_{*}(T)$, we attach an element $t^{\lambda}\in T(\mathcal{K})$ naturally and
define the norm $n^{\lambda}\in T(\mathcal{K})^{\sigma}$ of $t^{\lambda}$ by
(1)
$n^{\lambda}:=\prod_{i=0}^{r-1}\sigma^{i}(t^{\lambda})=\epsilon^{\sum_{i=1}^{r-1}i\sigma^{i}(\lambda)}t^{\sum\sigma^{i}\lambda}.$
Let $\bar{\lambda}$ be the image of $\lambda$ in $X_{*}(T)_{\sigma}$. Set
$e_{\bar{\lambda}}=n^{\lambda}\cdot e_{0}\in\mathcal{G}r$. Then
$e_{\bar{\lambda}}$ only depends on $\bar{\lambda}$. Following [BH, Zh],
$\mathcal{G}r$ admits the following Cartan decomposition
(2) $\mathcal{G}r=\bigsqcup_{\bar{\lambda}\in
X_{*}(T)_{\sigma}^{+}}\mathcal{G}r_{\bar{\lambda}}$
where $\mathcal{G}r_{\bar{\lambda}}=G(\mathcal{O})^{\sigma}\cdot
e_{\bar{\lambda}}$ is a Schubert cell. Let
$\overline{\mathcal{G}r}_{\bar{\lambda}}$ be the closure of
$\mathcal{G}r_{\bar{\lambda}}$. Then
$\overline{\mathcal{G}r}_{\bar{\lambda}}=\bigsqcup_{\bar{\mu}\preceq\bar{\lambda}}\mathcal{G}r_{\bar{\mu}},$
and $\dim\overline{\mathcal{G}r}_{\bar{\lambda}}=\langle
2\rho,\bar{\lambda}\rangle,$ where $\rho$ is the half sum of all positive
coroots of $H$.
By abuse of notation, we still use $\sigma$ to denote the induced automorphism
on $\mathfrak{g}$ of order $r$. Then there is a grading on $\mathfrak{g}$,
$\mathfrak{g}=\mathfrak{g}_{0}\oplus\mathfrak{g}_{1}\oplus\cdots\oplus\mathfrak{g}_{r-1}$
where $\mathfrak{g}_{i}$ is the $\epsilon^{i}$-eigenspace. Set
$\mathfrak{p}=\mathfrak{g}_{1}.$
Set $\mathcal{O}^{-}=\mathbb{C}[t^{-1}]$. Consider the evaluation map ${\rm
ev}_{\infty}:G(\mathcal{O}^{-})\to G$. Let $G(\mathcal{O}^{-})_{0}$ denote its
kernel. The map ${\rm ev}_{\infty}$ factors through
$G(\mathbb{C}[t^{-1}]/(t^{-2}))\to G$. Note that the kernel of
$G(\mathbb{C}[t^{-1}]/(t^{-2}))\to G$ is canonical identified with the vector
space $\mathfrak{g}\otimes t$ with respect to the adjoint action of $G$ and
$\sigma$. It induces a $G\rtimes\langle\sigma\rangle$-equivariant map
$G(\mathcal{O}^{-})_{0}\to\mathfrak{g}\otimes t.$
Taking $\sigma$-invariance, we get a $K$-equivariant map
(3) $G(\mathcal{O}^{-})_{0}^{\sigma}\to\mathfrak{p},$
where $K:=G^{\sigma}$. Note that $K$ is a connected simply-connected simple
algebraic group, as $G$ is simply-connected.
Set $\mathcal{G}r_{0}^{-}:=G(\mathcal{O}^{-})^{\sigma}\cdot e_{0}\simeq
G(\mathcal{O}^{-})_{0}^{\sigma}$. Then $\mathcal{G}r_{0}^{-}$ is the open
opposite Schubert cell in $\mathcal{G}r$. From (3), we have the following
$G^{\sigma}$-equivariant map
(4) $\pi:\mathcal{G}r_{0}^{-}\to\mathfrak{p}.$
###### Lemma 2.1.
$\mathcal{G}r_{\bar{\lambda}}\cap\mathcal{G}r_{0}^{-}$ is nonempty for any
$\lambda\in X_{*}(T)_{\sigma}^{+}$.
###### Proof.
It suffices to show that
$G(\mathcal{O})^{\sigma}n^{\lambda}G(\mathcal{O})^{\sigma}\cap
G(\mathcal{O}^{-})^{\sigma}$. Let $I$ be the Iwahori subgroup contained in
$G(\mathcal{O})^{\sigma}$ and let $I^{-}$ be the opposite Iwahori subgroup
contained in $G(\mathcal{O}^{-})^{\sigma}$. It suffices to show that
$In^{\lambda}I\cap I^{-}\not=\emptyset$.
Let $\mathcal{G}$ be the Kac-Moody group associated to the twisted loop group
$G(\mathcal{K})^{\sigma}$, see a construction in [BH, p.14]. There is a
projection map $p:\mathcal{G}\to G(\mathcal{K})^{\sigma}$. Let $\mathcal{I}$
(resp. $\mathcal{I}^{-}$) be the preimage of $I$ (resp. $I^{-}$) via the
projection $p$. Then we are reduced to show that
$\mathcal{I}w\mathcal{I}\cap\mathcal{I}^{-}\not=\emptyset$, where $w$ is an
element of the Weyl group $\mathcal{W}$ of $\mathcal{G}$. This is true, since
$\mathcal{I}s\mathcal{I}\cap\mathcal{I}^{-}\not=\emptyset$ for any simple
reflection $s\in\mathcal{W}$, and for any $y\in\mathcal{W}$ and simple
reflection $s$, $\mathcal{I}y\mathcal{I}s\mathcal{I}=\mathcal{I}ys\mathcal{I}$
if $\ell(ys)=\ell(y)+1$, cf. [Ku, 5.1.3 (d)]. ∎
Following [Br, Re, AH], an element $\bar{\lambda}$ of $X_{*}(T)_{\sigma}^{+}$
is called small, if $\bar{\lambda}\nsucceq 2\gamma_{0}$, where $\gamma_{0}$ is
the highest short root of $H$. The set of all small dominant weights is a
lower order ideal of $X_{*}(T)_{\sigma}^{+}$, i.e., if
$\bar{\mu}\preceq\bar{\lambda}$ and $\bar{\lambda}$ is small, then $\bar{\mu}$
is also small. Let $\mathcal{G}r_{\mathrm{sm}}$ be the union of
$\mathcal{G}r_{\bar{\lambda}}$ for small dominant weights $\bar{\lambda}$. Set
$\mathcal{M}=\mathcal{G}r_{\mathrm{sm}}\cap\mathcal{G}r_{0}^{-}.$
For each small dominant weight $\bar{\lambda}$, set
$\mathcal{M}_{\bar{\lambda}}=\mathcal{G}r_{\bar{\lambda}}\cap\mathcal{G}r_{0}^{-}.$
Let $\mathcal{N}_{\mathfrak{p}}$ denote the nilpotent cone of $\mathfrak{p}$.
We shall prove in Section 4 that $\pi(\mathcal{M})$ is contained
$\mathcal{N}_{\mathfrak{p}}$, when $G$ is of type $A_{n}$ and $D_{n}$ and
$\sigma$ is of order $2$.
Recall that $\gamma_{0}$ is the highest short root of $H$. The following lemma
is a twisted analogue of [AH, Lemma 3.3].
###### Lemma 2.2.
If $\sigma$ is a diagram automorphism of order $r$, then
$\pi(\mathcal{G}r_{2\gamma_{0}}\cap\mathcal{G}r_{0}^{-})\nsubseteq\mathcal{N}_{\mathfrak{g}_{1}}$.
###### Proof.
Let $X_{N}$ be the Dynkin diagram of $G$. Following [Ka, p.128-129], we choose
the following root of $G$,
${\theta}_{0}=\begin{cases}{\alpha}_{1}+\cdots+{\alpha}_{2\ell-2},&\quad(X_{N},r)=(A_{2\ell-1},2);\\\
{\alpha}_{1}+\cdots+{\alpha}_{2\ell},&\quad(X_{N},r)=(A_{2\ell},2);\\\
{\alpha}_{1}+\cdots+{\alpha}_{\ell},&\quad(X_{N},r)=(D_{\ell+1},2);\\\
{\alpha}_{1}+{\alpha}_{2}+{\alpha}_{3},&\quad(X_{N},r)=(D_{4},3);\\\
{\alpha}_{1}+2{\alpha}_{2}+2{\alpha}_{3}+{\alpha}_{4}+{\alpha}_{5}+{\alpha}_{6},&\quad(X_{N},r)=(E_{6},2).\end{cases}$
where the label of simple roots ${\alpha}_{i}$ follows from [Ka, TABLE Fin,
p.53]. Recall from the section 2.1 that for each $\iota\in I_{\sigma}$, we
define simple roots of $H$, $\gamma_{\imath}=\bar{\check{\alpha}}_{i}\in
X_{*}(T)_{\sigma}$.
Let $\check{\theta}_{0}$ be the coroot of ${\theta}_{0}$. Then,
$\bar{\check{\theta}}_{0}=\begin{cases}\gamma_{0}&\quad\text{if
}(X_{N},r)\not=(A_{2\ell},2)\\\ 2\gamma_{0}&\quad\text{if
}(X_{N},r)=(A_{2\ell},2)\end{cases}.$
Suppose $(X_{N},r)\not=(A_{2\ell},2)$. Note that $\theta_{0}\in X^{*}(T)$ and
$\check{\theta}_{0}:\mathbb{C}^{\times}\to T$. Each $a\in\mathbb{C}^{\times}$
can be identified with $\begin{pmatrix}a&0\\\
0&a^{-1}\end{pmatrix}\in\mathrm{SL}_{2}.$ For each $i=0,...,r-1$, define a
homomorphism $\phi_{\sigma^{i}(\theta_{0})}:\mathrm{SL}_{2}\to G$ given by
$\begin{pmatrix}1&a\\\ 0&1\end{pmatrix}\mapsto
x_{\sigma^{i}(\theta_{0})}(a),\hskip 14.22636pt\begin{pmatrix}1&0\\\
a&1\end{pmatrix}\mapsto y_{\sigma^{i}(\theta_{0})}(a),\hskip
14.22636pt\begin{pmatrix}a&0\\\
0&a^{-1}\end{pmatrix}\mapsto\sigma^{i}(\check{\theta}_{0})(a).$
Let $\mathcal{S}$ be the product of $r$ copies of $\mathrm{SL}_{2}$. Then
$\check{\theta}_{0}$ can be extended to $\phi:\mathcal{S}\to G$ given by
$\phi(g_{0},...,g_{r-1})=\prod_{i=0}^{r-1}\phi_{\sigma^{i}(\theta_{0})}(g_{i}).$
This $\phi$ can extend scalar to $\mathcal{K}$. Abusing notation, define
$\sigma:\prod_{i=1}^{r}(\operatorname{SL}_{2}(\mathbb{\mathcal{K}}))_{i}\to\prod_{i=1}^{r}(\operatorname{SL}_{2}(\mathbb{\mathcal{K}}))_{i}$
by
$\sigma(g_{1}(t),g_{2}(t),...,g_{r}(t))=(g_{r}(\epsilon t),g_{1}(\epsilon
t),...,g_{r-1}(\epsilon t)).$
There exists an isomorphism
$\varphi:\mathrm{SL}_{2}(\mathcal{K})\to(\prod_{i=1}^{r}(\operatorname{SL}_{2}(\mathbb{\mathcal{K}}))_{i})^{\sigma}=\\{(g(t),g(\epsilon
t),...,g(\epsilon^{r-1}t))\mid g(t)\in\operatorname{SL}_{2}(\mathcal{K})\\}.$
Hence
$\phi\circ\varphi:\begin{pmatrix}t&0\\\
0&t^{-1}\end{pmatrix}\mapsto\left(\begin{pmatrix}\epsilon^{i}t&0\\\
0&(\epsilon^{i}t)^{-1}\end{pmatrix}\right)_{i=0,...,r-1}\mapsto\prod_{i=0}^{r-1}(\epsilon^{i}t)^{\sigma^{i}\check{\theta}_{0}}=n^{\check{\theta}_{0}}.$
Let $\mathfrak{s}$ be the product of $r$ copies of $\mathfrak{sl}_{2}$. Define
$\sigma:\mathfrak{s}\to\mathfrak{s}$ by
$\sigma(x_{1},,...,x_{r-1},x_{r})=(\epsilon x_{r},\epsilon x_{1},...,\epsilon
x_{r-1}).$
Since $\sigma$ has order $r$, we have
$\mathfrak{s}=\oplus_{i=0}^{r-1}\mathfrak{s}_{i}$ where $\mathfrak{s}_{i}$ is
the eigenspace of eigenvalue $\epsilon^{i}$. Then
$\mathfrak{s}_{1}=\\{(x,\epsilon x,...,\epsilon^{r-1}x)\mid
x\in\mathfrak{sl}_{2}\\}\cong\mathfrak{sl}_{2}$. The derivative of $\phi$ is
$\mathrm{d}\phi:\mathfrak{s}\to\mathfrak{g}$ which induces
$\mathfrak{s}_{1}\to\mathfrak{g}_{1}$. Hence we have the map
$\Psi:\mathfrak{sl}_{2}\to\mathfrak{g}_{1}$.
Consider the matrix $g(t)\in\operatorname{SL}_{2}(\mathcal{O}^{-})$,
$g(t)=\begin{pmatrix}1+t^{-1}&t^{-2}\\\
t^{-1}&1-t^{-1}+t^{-2}\end{pmatrix}=\begin{pmatrix}0&1\\\
-1&t^{2}-t+1\end{pmatrix}\begin{pmatrix}t^{2}&0\\\
0&t^{-2}\end{pmatrix}\begin{pmatrix}1&0\\\ t^{2}+t&1\end{pmatrix}.$
Then $(\phi\circ\varphi)(g(t))\in
G(\mathcal{O})^{\sigma}n^{2\check{\theta}_{0}}G(\mathcal{O})^{\sigma}$. Since
$G$ is not type $A_{2l}$, $\bar{\check{\theta}}_{0}=\gamma_{0}$ and then
$(\phi\circ\varphi)(g(t))\cdot
e_{0}\in\mathcal{G}r_{2\gamma_{0}}\cap\mathcal{G}r_{G,0}^{-}$. We have the
commutative diagram
${{\operatorname{Gr}_{\mathrm{SL}_{2},0}^{-}}}$${{\mathcal{G}r_{\mathcal{S},0}^{-}}}$${{\mathcal{G}r_{G,0}^{-}}}$${{\mathfrak{sl}_{2}}}$${{\mathfrak{s}_{1}}}$${{\mathfrak{g}_{1}}}$$\scriptstyle{g(t)\cdot
L_{0}\mapsto\varphi(g(t))\cdot
e_{0}}$$\scriptstyle{(g_{i}(t))_{i=0}^{r-1}\cdot
e_{0}\mapsto\phi(g_{i}(t))_{i=0}^{r-1})\cdot
e_{0}}$$\scriptstyle{\pi_{\mathrm{SL}_{2}}}$$\scriptstyle{\pi}$$\scriptstyle{x\mapsto(x,\epsilon
x,...,\epsilon^{r-1}x)}$
where
$\operatorname{Gr}_{\mathrm{SL}_{2},0}^{-}:=\mathrm{SL}_{2}(\mathcal{O}^{-})_{0}\cdot
e_{0}\subset\operatorname{Gr}_{\mathrm{SL}_{2}}$, and
${\mathcal{G}r_{\mathcal{S},0}^{-}}$ is defined similarly. The commutativity
follows from
$\pi((\phi\circ\varphi)(g(t))\cdot
e_{0})=\Psi(\pi_{\operatorname{SL}_{2}}(g(t)\cdot
e_{0}))=\Psi\begin{pmatrix}1&0\\\ 1&-1\end{pmatrix},$
where the latter is not nilpotent. It follows that,
$\pi(\mathcal{G}r_{2\gamma_{0}})\not\subseteq\mathcal{N}_{\mathfrak{p}}$.
Suppose that $(X_{N},r)=(A_{2n},2)$. In this case,
$\bar{\check{\theta}}_{0}=2\gamma_{0}$ and
$\sigma(\check{\theta}_{0})={\check{\theta}_{0}}$. Then $\check{\theta}_{0}$
can be extended to $\phi:\mathrm{SL}_{2}\to G$ defined by
$\begin{pmatrix}1&a\\\ 0&1\end{pmatrix}\mapsto x_{\theta_{0}}(a),\hskip
14.22636pt\begin{pmatrix}1&0\\\ a&1\end{pmatrix}\mapsto
y_{\theta_{0}}(a),\hskip 14.22636pt\begin{pmatrix}a&0\\\
0&a^{-1}\end{pmatrix}\mapsto\check{\theta}_{0}(a).$
$\phi$ can extend the scalar to $\mathcal{K}$. Define a group homomorphism
$\sigma:\mathrm{SL}_{2}(\mathcal{K})\to\mathrm{SL}_{2}(\mathcal{K})$ by
$\begin{pmatrix}a(t)&b(t)\\\
c(t)&d(t)\end{pmatrix}\mapsto\begin{pmatrix}a(-t)&-b(-t)\\\
-c(-t)&d(-t)\end{pmatrix}$
where $a(t)\in\mathcal{K}$. Then $\phi:\mathrm{SL}_{2}(\mathcal{K})\to
G(\mathcal{K})$ is $\sigma$-equivariant. The induced homomorphism
$\sigma:\mathfrak{sl}_{2}\to\mathfrak{sl}_{2}$ is given by
$\begin{pmatrix}a&b\\\ c&-a\end{pmatrix}\mapsto\begin{pmatrix}a&-b\\\
-c&-a\end{pmatrix}.$
The derivative $\mathrm{d}\phi:\mathfrak{sl}_{2}\to\mathfrak{g}$ induces the
map $\Psi:(\mathfrak{sl}_{2})_{1}\to\mathfrak{g}_{1}$. Similarly to the above
arguement, we have the commutative diagram
${{\mathcal{G}r_{\mathrm{SL}_{2},0}^{-}}}$${{\mathcal{G}r_{\mathrm{G},0}^{-}}}$${{(\mathfrak{sl}_{2})_{1}}}$${{\mathfrak{g}_{1}}}$$\scriptstyle{\pi_{\mathrm{SL}_{2}}}$$\scriptstyle{\Psi}$$\scriptstyle{\pi}$
where ($\mathfrak{sl}_{2})_{1}$ is the eigenspace of eigenvalue $-1$ under
$\sigma$. Now consider $g(t)\in\mathrm{SL}_{2}(\mathcal{O}^{-})^{\sigma}$
$g(t)=\begin{pmatrix}1&t^{-1}\\\
t^{-1}&1+t^{-2}\end{pmatrix}=\begin{pmatrix}0&t\\\
-t^{-1}&1+t^{2}\end{pmatrix}\begin{pmatrix}t^{2}&0\\\
0&t^{-2}\end{pmatrix}\begin{pmatrix}1&0\\\ t&1\end{pmatrix}.$
Then $\phi(g(t))\in
G(\mathcal{O})^{\sigma}n^{\check{\theta}_{0}}G(\mathcal{O})^{\sigma}$ and
$\phi(g(t))\cdot e_{0}\in\mathcal{G}r_{2\gamma_{0}}\cap\mathcal{G}r_{0}^{-}$.
The result follows from
$\pi(\phi(g(t))\cdot e_{0})=\Psi(\pi_{\operatorname{SL}_{2}}(g(t)\cdot
e_{0}))=\Psi\begin{pmatrix}0&1\\\ 1&0\end{pmatrix}$
where the latter is not nilpotent. It also follows that,
$\pi(\mathcal{G}r_{2\gamma_{0}})\not\subseteq\mathcal{N}_{\mathfrak{p}}$. ∎
###### Proposition 2.3.
For $\bar{\lambda}\in X_{*}(T)_{\sigma}^{+}$, if
$\pi(\mathcal{G}r_{\bar{\lambda}}\cap\mathcal{G}r_{0}^{-})\subset\mathcal{N}_{\mathfrak{p}}$,
then $\bar{\lambda}$ is small.
###### Proof.
Since $\mathcal{G}r_{0}^{-}$ is an open subset of $\mathcal{G}r$,
$\pi(\overline{\mathcal{G}r}_{\bar{\lambda}}\cap\mathcal{G}r_{0}^{-})\subset\mathcal{N}_{\mathfrak{p}}$.
By Lemma 2.2,
$\mathcal{G}r_{2\gamma_{0}}\nsubseteq\overline{\mathcal{G}r}_{\bar{\lambda}}$
which means $\bar{\lambda}\nsucceq 2\gamma_{0}$. ∎
Define the following anti-involution
$\iota:G(\mathcal{K})\to G(\mathcal{K}),\hskip 14.22636ptg(t)\mapsto
g(-t)^{-1}.$
It can be checked that $\iota$ commutes with $\sigma$, and $\iota$ preserves
$G(\mathcal{K})^{\sigma},G(\mathcal{O})^{\sigma}$ and $K^{-}$. This induces
the map
$\iota:\mathcal{G}r_{0}^{-}\to\mathcal{G}r_{0}^{-},\hskip 14.22636ptg(t)\cdot
e_{0}\mapsto g(-t)^{-1}\cdot e_{0}.$
The following lemma will be used in Section 4.
###### Lemma 2.4.
For $\bar{\lambda}\in X_{*}(T)_{\sigma}^{+}$,
$\iota(\mathcal{M}_{\bar{\lambda}})\subset\mathcal{M}_{\bar{\lambda}}$.
###### Proof.
It suffices to prove $\iota(n^{\lambda})\in\mathcal{M}_{\bar{\lambda}}$ for
each $\lambda\in X_{*}(T)^{+}$.
$\displaystyle\iota(n^{\lambda})$
$\displaystyle=\iota(\epsilon^{\sigma\lambda+2\sigma^{2}\lambda+...+(r-1)\sigma^{r-1}\lambda}t^{\sum_{i=0}^{r-1}\sigma^{i}\lambda})$
$\displaystyle=\epsilon^{-(\sigma\lambda+2\sigma^{2}\lambda+...+(r-1)\sigma^{r-1}\lambda)}(-1)^{\sum_{i=0}^{r-1}\sigma^{i}\lambda}t^{-\sum_{i=0}^{r-1}\sigma^{i}\lambda}$
$\displaystyle=(-1)^{\sum_{i=0}^{r-1}\sigma^{i}\lambda}n^{-\lambda}.$
Since $(-1)^{\sum_{i=0}^{r-1}\sigma^{i}\lambda}$ is fixed by $\sigma$,
$\iota(n^{\lambda})\in
G(\mathcal{O})^{\sigma}n^{-\lambda}G(\mathcal{O})^{\sigma}$. Let $W$ be the
Weyl group of $G$ with respect to the maximal torus $T$ and $\omega_{0}$ the
longest element of $W$. We can choose a representative $\dot{\omega}_{0}\in G$
of $\omega_{0}$ such that $\sigma(\dot{\omega}_{0})=\dot{\omega}_{0}$, cf.
[HS, Section 2.3].
When $G$ is of type $D_{2\ell}$ with $\ell\geq 2$, $w_{0}=-1$; otherwise,
$w_{0}=-\sigma$ and $\sigma$ is of order $2$, cf. [Hu2, Ex 5, p.71]. If
$w_{0}=-1$, it is easy to see that $n^{-\lambda}=w_{0}n^{\lambda}w_{0}^{-1}$.
If $w_{0}=-\sigma$ and $\sigma$ has order 2,
$n^{-\lambda}=(-1)^{-\sigma\lambda}t^{-(\lambda+\sigma\lambda)}=(-1)^{w_{0}\lambda}t^{w_{0}(\lambda+\sigma\lambda)}=w_{0}(-1)^{\lambda}t^{\lambda+\sigma\lambda}w_{0}^{-1}=w_{0}(-1)^{\lambda+\sigma\lambda}n^{\lambda}w_{0}^{-1}.$
In any case, $\iota(n^{\lambda})\in
G(\mathcal{O})^{\sigma}n^{\lambda}G(\mathcal{O})^{\sigma}$. ∎
## 3\. Nilpotent orbits in the space of self-adjoint maps
In this section, we will review some facts on the nilpotent orbits in certain
symmetric spaces. These results are known, cf. [Se]. We provide proofs here,
as the proofs in [Se] are omitted.
Let $B=\langle\cdot,\cdot\rangle$ be a nondegenerate symmetric or skew-
symmetric bilinear form on a vector space $V=\mathbb{C}^{m}$ and $\mathcal{A}$
the set of self-adjoint linear maps under the bilinear form. In this section,
we describe the classification of nilpotent orbits in the space $\mathcal{A}$
in Theorem 3.3, Theorem 3.4 and Theorem 3.6.
The isometry group of the form $B$ is
$I_{B}=\\{g\in\mathrm{GL}(V)\mid\langle gu,gv\rangle=\langle u,v\rangle\text{
for all }u,v\in V\\},$
whose Lie alegbra is
(5) $\mathfrak{g}_{B}:=\\{X\in\mathfrak{sl}(V)\mid\langle Xu,v\rangle+\langle
u,Xv\rangle=0\text{ for all }u,v\in V\\}.$
When $B$ is symplectic, $\dim V$ is even, $I_{B}\cong\mathrm{Sp}_{2n}$ and
$\mathfrak{g}_{B}\simeq\mathfrak{sp}_{2n}$ where $m=2n$. When $B$ is
symmetric, $I_{B}\cong\mathrm{O}_{m}$ and
$\mathfrak{g}_{B}\cong\mathfrak{so}_{m}$.
The group $I_{B}$ acts on the space of self-adjoint linear maps
(6) $\mathcal{A}=\\{X\in\operatorname{End}(V)\mid\langle Xu,v\rangle=\langle
u,Xv\rangle\text{ for all }u,v\in V\\}$
by conjugation. The orbit is called nilpotent if it is the orbit of a
nilpotent element of $\mathcal{A}$.
Let $\mathfrak{g}$ be a complex semisimple Lie algebra. Suppose that
$\mathfrak{g}$ has $\mathbb{Z}_{m}$-grading
$\mathfrak{g}=\bigoplus_{i\in\mathbb{Z}_{m}}\mathfrak{g}_{i}$
so that $[\mathfrak{g}_{k},\mathfrak{g}_{\ell}]\subset\mathfrak{g}_{k+\ell}$.
We have the following graded version of Jacobson-Morozov theorem.
###### Lemma 3.1.
Let $X$ be a nonzero nilpotent element in $\mathfrak{g}_{i}$. There exists an
$\mathfrak{sl}_{2}$-triple $H,X,Y$ such that $H\in\mathfrak{g}_{0}$ and
$Y\in\mathfrak{g}_{-i}$.
###### Proof.
This Lemma follows from the usual Jacobson-Morozov Theorem, and the proof is
similar to [EK, Lemma 1.1]. ∎
For each $A\in\mathfrak{sl}(V)$, as a linear map, we denote its adjoint by
$A^{*}$ under the form $B$. We define an involution $\sigma$ on
$\mathfrak{g}=\mathfrak{sl}(V)$ by
(7) $\sigma(A)=-A^{*}.$
Then $\mathfrak{g}$ is the direct sum of eigenspaces,
$\mathfrak{g}=\mathfrak{g}_{0}\oplus\mathfrak{g}_{1}$. Thus
$\mathfrak{g}_{0}=\mathfrak{g}_{B}$ and $\mathfrak{g}_{1}=\mathcal{A}$. Fix a
nonzero nilpotent element $X\in\mathcal{A}$. By Lemma 3.1, there exists
$Y\in\mathcal{A}$ and $H\in\mathfrak{g}_{B}$, such that $X,Y,H$ is an
$\mathfrak{sl}_{2}$-triple. This induces a representation of
$\mathfrak{sl}_{2}$ on $V$ and hence we have a decomposition
(8) $V=\bigoplus_{r\geq 0}M(r)$
where $M(r)$ is a finite direct sum of irreducible representation of
$\mathfrak{sl}_{2}$ of highest weight $r$. For $r\geq 0$, let $H(r)$ be the
highest weight space in $M(r)$. Define a new bilinear form $(\cdot,\cdot)$ on
$H(r)$ by
$(u,v)_{r}=\langle u,Y^{r}v\rangle.$
###### Lemma 3.2.
For any $r\geq 0$, $(\cdot,\cdot)_{r}$ is symplectic (resp. symmetric) if $B$
is symplectic (resp. symmetric).
###### Proof.
We assume $B$ is symplectic. The proof is similar when $B$ is symmetric. It is
easy to see that $(\cdot,\cdot)_{r}$ is skew-symmetric. It remains to show
that $(\cdot,\cdot)_{r}$ is nondegerate. Let $V_{r}$ be an $r$-weight space in
$\mathbb{C}^{2n}$. For any $u\in V_{r}$, $v\in V_{s}$ with $s\neq-r$,
$(r+s)\langle u,v\rangle=\langle ru,v\rangle+\langle u,sv\rangle=\langle
Hu,v\rangle+\langle u,Hv\rangle=0$
This implies that $V_{r}$ and $V_{s}$ are
$\langle\cdot,\cdot\rangle$-orthogonal. Let
$W=\text{Span}\\{u\in V_{r}\mid u=Yv\text{ for some }v\in\mathbb{C}^{2n}\\}.$
It can be seen that $V_{r}=H(r)\oplus W$. For $u\in H(r)$ and $v\in W$, write
$v=Yv^{\prime}$,
$(u,v)_{r}=\langle u,Y^{r}v\rangle=\langle u,Y^{r+1}v^{\prime}\rangle=\langle
Y^{r+1}u,v^{\prime}\rangle=0.$
Hence $H(r)$ is $(\cdot,\cdot)_{r}$-orthogonal to $W$.
We claim that $\langle\cdot,\cdot\rangle:(Y^{r}\cdot H(r))\times
H(r)\to\mathbb{C}$ is nondegenerate. Let $u=Y^{r}u^{\prime}\in Y^{r}\cdot
H(r)$ be such that $\langle u,v\rangle=0$ for all $v\in H(r)$. For each
$w\in\mathbb{C}^{2n}$, write $w=\sum_{s}w_{s}$ where each $w_{s}$ belongs to
$V_{s}$. Since $u\in V_{-r}$, $\langle u,w_{s}\rangle=0$ for $s\neq r$. Write
$w_{r}=w_{1}+w_{2}$ where $w_{1}\in H(r)$ and $w_{2}=Yw^{\prime}_{2}\in W$. By
the assumption $\langle u,w_{1}\rangle=0$ and hence
$\langle u,w_{r}\rangle=\langle u,w_{2}\rangle=\langle
Y^{r}u^{\prime},Yw^{\prime}_{2}\rangle=\\-\langle
Y^{r+1}u^{\prime},w^{\prime}_{2}\rangle=0.$
We obtain $\langle u,w\rangle=0$ for any $w$ and hence $u=0$. This claim
implies that $(\cdot,\cdot)_{r}$ is nondegenerate. ∎
A partition of a positive integer is denoted by a tuple
$[d_{1},d_{2},...,d_{k}]$ of positive integers. We use the exponent notation
$[a_{1}^{i_{1}},...,a_{r}^{i_{r}}]$
to denote a partition where $a_{j}^{i_{j}}$ means there are $i_{j}$ copies of
$a_{j}$. For example, $[3^{2},1^{4}]=[3,3,1,1,1,1]$ is a partition of 10. Put
$r_{i}=|\\{j\mid d_{j}=i\\}|$ and $s_{i}=|\\{j\mid d_{j}\geq i\\}|$. In fact,
each partition can be illustrated by Young diagram and then $s_{i}$ is the
$i$-th part of the dual diagram. The following Theorem gives the
parametrization of nilpotent $I_{B}$-orbits in $\mathcal{A}$.
###### Theorem 3.3.
There exists one-to-one correspondences
$\\{\text{nilpotent }\mathrm{Sp}_{2n}\text{-orbits in
}\mathcal{A}\\}\leftrightarrow\left\\{\begin{array}[]{c}\text{partitions of
}2n\text{ such that}\\\ \text{every part occurs with even
multiplicity}\end{array}\right\\}.$
and
$\\{\text{nilpotent }\mathrm{O}_{m}\text{-orbits in
}\mathcal{A}\\}\leftrightarrow\\{\text{partitions of }m\\}.$
###### Proof.
The proof is similar to [CM, Lemma 5.1.17]. For the case that $B$ is
symplectic, it suffices to show that any nilpotent element in $\mathcal{A}$
gives arise the partition of $2n$ such that every part occurs with even
multiplicity. Given nilpoent $X\in\mathcal{A}$, a number of Jordan blocks of
size $r+1$ equals to the multiplicity of $M(r)$ in $\mathbb{C}^{2n}$ which is
exactly $\dim H(r)$. By Lemma 3.2, $\dim H(r)$ is even for every $r$.
If $B$ is symmetric, there are no constraints on $\dim H(r)$ which means there
are no conditions on partitions of $m$. ∎
###### Theorem 3.4.
There exists one-to-one correspondence
$\\{\text{nilpotent }\mathrm{SO}_{2n+1}\text{-orbits in
}\mathcal{A}\\}\leftrightarrow\left\\{\text{partitions of }2n+1\right\\}.$
###### Proof.
Since $\mathrm{O}_{2n+1}=\mathrm{SO}_{2n+1}\times\\{\pm I_{2n+1}\\}$, the
orbits under $\mathrm{O}_{2n+1}$ and $\mathrm{SO}_{2n+1}$ coincide. The
results immediately follows from Theorem 3.3 ∎
Consider the case that $B$ is symmetric and $m=2n$. Given nilpotent elements
$X,X^{\prime}\in\mathcal{A}$ whose partitions are the same and have at least
one odd part. Say that they are conjugated by an element
$g\in\mathrm{O}_{2n}$. If $\det g=1$, we conclude that $X,X^{\prime}$ are in
the same $\mathrm{SO}_{2n}$-orbits. Suppose that $\det g=-1$. We modify this
$g$ so that it has determinant $1$. Note that either $X$ or $X^{\prime}$ gives
the same decomposition (8). An odd part in the partition corresponds to an odd
dimensional irreducible representation $S$ of $\mathfrak{sl}_{2}$ in
$\mathbb{C}^{2n}$. We put $h=g$ except that $h(v)=-g(v)$ for $v\in S$.
Therefore, $\det h=1$, and $X$ and $X^{\prime}$ are conjugated by $h$. If
there is no odd parts, we need the following Lemma.
###### Lemma 3.5.
Let $X$ be a nilpotent element in $\mathcal{A}$ whose partition contains only
even parts, and $k\in\mathrm{O}_{2n}$ such that $k\cdot X=kXk^{-1}=X$. Then
$\det k=1$.
###### Proof.
Let $\mathrm{O}_{2n}^{X}$ be the stabilizer group of $O_{2n}$ at $X$. Then
$k\in\mathrm{O}_{2n}^{X}$. By multiplicative Jordan decomposition, cf. [Bo,
Theorem 4.4, p.83], let $k_{s}\in O_{2n}^{X}$ be the semisimple part of $k$.
Then $\det k_{s}=\det k$. Hence we may assume that $k$ is semisimple. Let
$\sigma$ be an automorphism on $\mathfrak{g}=\mathfrak{sl}(V)$ defined by (7).
Then $\sigma$ commutes with ${\rm Ad}k$ on $\mathfrak{g}$, as
$k\in\mathrm{O}_{2n}$. Then
$\mathfrak{g}=\mathfrak{g}_{0}\oplus\mathfrak{g}_{1}$, and we have the
decomposition of $k$-stabilizers
$\mathfrak{g}^{k}=\mathfrak{g}_{0}^{k}\oplus\mathfrak{g}_{1}^{k}$ where
$\mathfrak{g}_{i}^{k}=\mathfrak{g}_{i}\cap\mathfrak{g}^{k}$. Since
$\mathfrak{g}^{k}$ is reductive and $X\in\mathfrak{g}_{1}^{k}$, by Lemma 3.1,
there exists an $\mathfrak{sl}_{2}$-triple $H,X,Y$ such that
$X,Y\in\mathfrak{g}_{1}^{k}$ and $H\in\mathfrak{g}_{0}^{k}$ and hence we have
the decomposition (8). It is easy to see that $k(M(r))\subset M(r)$, and also,
$k$ stabilizes each weight space of $M(r)$.
Recall that $\langle\cdot,\cdot\rangle$ is a nondegenerate symmetric form on
$V=\mathbb{C}^{2n}$ and a form on $H(r)$ given by $(u,v)_{r}=\langle
u,Y^{r}v\rangle$ is also symmetric for any $r\geq 0$. For any $u,v\in H(r)$,
$(ku,kv)_{r}=\langle ku,Y^{r}kv\rangle=\langle ku,kY^{r}v\rangle=\langle
u,Y^{r}v\rangle=(u,v)_{r}.$
Hence $k\big{|}_{H(r)}\in\mathrm{O}(H(r))$ for any $r$. In particular
$\det\left(k\big{|}_{H(r)}\right)=\pm 1$.
Let $M(r)_{\ell}$ be an $\ell$-weight space, and $L(r)$ the lowest weight
space in $M(r)$. Observe that $X\big{|}_{M(r)_{\ell}}$ is an isomorphism from
$M(r)_{\ell}$ to $M(r)_{\ell+2}$ and the diagram
${{L(r)}}$${{M(r)_{\ell}}}$${{M(r)_{\ell+2}}}$${{H(r)}}$${{L(r)}}$${{M(r)_{\ell}}}$${{M(r)_{\ell+2}}}$${{H(r)}}$$\scriptstyle{X\big{|}_{M(r)_{\ell}}}$$\scriptstyle{k\big{|}_{M(r)_{\ell}}}$$\scriptstyle{k\big{|}_{M(r)_{\ell+2}}}$$\scriptstyle{X\big{|}_{M(r)_{\ell}}}$$\scriptstyle{k\big{|}_{L(r)}}$$\scriptstyle{k\big{|}_{H(r)}}$
commutes. Then $k\big{|}_{M(r)_{\ell}}$ has the same determinant for all
$\ell$. Since $X$ has only even parts, a number of weight spaces in $M(r)$ is
even for each $r$. Then
$\det
k=\prod_{r}\det\left(k\big{|}_{M(r)}\right)=\prod_{r}\prod_{\ell}\det\left(k\big{|}_{M(r)_{\ell}}\right)=\prod_{r}\left(\det\left(k\big{|}_{H(r)}\right)\right)^{\ell}=1$
as desired. ∎
Now suppose that $X,X^{\prime}\in\mathcal{A}$ have the same partition and
contain only even parts. If $\det g=1$, they are in the same
$\mathrm{SO}_{2n}$-orbits. Suppose that $\det g=-1$ and they are conjugated by
another element $h\in\mathrm{SO}_{2n}$. Say $g\cdot X=X^{\prime}=h\cdot X$ and
let $k=g^{-1}h$. Then $\det k=(\det g^{-1})(\det h)=-1$ but this contradicts
to Lemma 3.5. In this case, it means $X,X^{\prime}$ are conjugated by an
element in $\mathrm{O}_{2n}$ of determinant -1 only. We have the following
theorem:
###### Theorem 3.6.
Nilpotent $\mathrm{SO}_{2n}$-orbits in $\mathcal{A}$ are parametrized by
partitions of $2n$ except that the partitions with only even parts correspond
to two orbits.
For each nilpotent element $X$ having the partition $[d_{1},d_{2},...,d_{k}]$,
we denote the $I_{B}$-orbits of $X$ by
$\mathcal{O}_{X}$,$\mathcal{O}_{[d_{1},d_{2},...,d_{k}]}$, or simply
$[d_{1},d_{2},...,d_{k}]$. We are now ready to compute the dimension of
nilpotent $I_{B}$-orbits.
###### Theorem 3.7.
Let $X$ be a nilpotent element in $\mathcal{A}$. Then the dimension of
$I_{B}$-orbit of $X$ is
$\dim\mathcal{O}_{X}=\frac{1}{2}\left(m^{2}-\sum_{i}s_{i}^{2}\right).$
###### Proof.
Suppose that $B$ is symplectic on $\mathbb{C}^{m}$, $m=2n$. Recall that we
have the decomposition (8). For each $Z\in\mathfrak{g}_{B}^{X}$, we consider
how $Z$ sends $M(d)$. Suppose that $Z$ sends $M(d)$ to $M(e)$ for $d\neq e$.
Since $Z$ and $X$ commute, by theory of representations of
$\mathfrak{sl}_{2}$, $Z$ is uniquely determined by a linear map $L(d)\to M(e)$
where $L(d)$ is the lowest weight space in $M(d)$. In this case, $Z$ sends
$L(e)$ to $M(d)$. Thus we can assume $d<e$. For $v\in L(d)$, $X^{d+1}v=0$ and
then $ZX^{d+1}v=X^{d+1}Zv=0$ is the constraint of $Zv=0$. Note that
$r_{d+1}=\dim L(d)$. Therefore the set of all linear maps from $L(d)$ to
$M(e)$ forms a vector space of dimension $(d+1)r_{d+1}r_{e+1}$.
Now consider the case $Z$ sends $M(d)$ to itself. We consider where $Z$ sends
$L(d)$ to. Suppose that $Z$ sends $L(d)$ to $H(d)$. We define a new bilinear
form $(\cdot,\cdot)_{d}$ on $L(d)$ given by $(u,v)_{d}=\langle u,Zv\rangle$.
It can be checked $(\cdot,\cdot)_{d}$ is symmetric and completely determine
the map $Z$. The set of all such $(\cdot,\cdot)_{d}$ forms a vector space of
dimension $\frac{1}{2}r_{d+1}(r_{d+1}+1)$. If $Z$ sends $L(d)$ to
$(d-2)$-weight space in $M(d)$, we define the new form by $(u,v)_{d-2}=\langle
u,XZv\rangle$. Again, this form is symmetric and completely determine the map
$Z$. Continue this process up to the case $Z$ sends $L(d)$ to itself. We
obtain
$\displaystyle\dim\mathfrak{k}^{X}$ $\displaystyle=\sum_{d\geq
0}\left[(d+1)\left(\sum_{e>d}r_{d+1}r_{e+1}\right)+\dfrac{d+1}{2}r_{d+1}(r_{d+1}+1)\right]$
$\displaystyle=\left[r_{1}(r_{2}+r_{3}+\cdots)+\dfrac{1}{2}r_{1}(r_{1}+1)\right]+\left[2r_{2}(r_{3}+r_{4}+\cdots)+\dfrac{2}{2}r_{2}(r_{2}+1)\right]$
$\displaystyle\hskip
56.9055pt+\left[3r_{3}(r_{4}+r_{5}+\cdots)+\dfrac{3}{2}r_{3}(r_{3}+1)\right]+\cdots$
$\displaystyle=\left[\dfrac{1}{2}r_{1}(r_{1}+2r_{2}+2r_{3}+...)+\dfrac{1}{2}r_{1}\right]+\left[\dfrac{2}{2}r_{2}(r_{2}+2r_{3}+2r_{4}+...)+\dfrac{2}{2}r_{2}\right]$
$\displaystyle\hskip
56.9055pt+\left[\dfrac{3}{2}r_{3}(r_{3}+2r_{4}+2r_{5}+...)+\dfrac{3}{2}r_{3}\right]+\cdots$
$\displaystyle=\left[\dfrac{1}{2}(s_{1}-s_{2})(s_{1}+s_{2})+\dfrac{1}{2}r_{1}\right]+\left[\dfrac{2}{2}(s_{2}-s_{3})(s_{2}+s_{3})+\dfrac{2}{2}r_{2}\right]$
$\displaystyle\hskip
56.9055pt+\left[\dfrac{3}{2}(s_{3}-s_{4})(s_{3}+s_{4})+\dfrac{3}{2}r_{3}\right]+\cdots$
$\displaystyle=\dfrac{1}{2}\sum_{i}s_{i}^{2}+\dfrac{1}{2}(r_{1}+2r_{2}+3r_{3}+\cdots)+\cdots$
$\displaystyle=\dfrac{1}{2}\sum_{i}s_{i}^{2}+\dfrac{1}{2}\sum_{i}s_{i}.$
$\displaystyle=n+\dfrac{1}{2}\sum_{i}s_{i}^{2}$
and hence
$\dim\mathcal{O}_{X}=\dim\mathfrak{g}_{B}-\dim\mathfrak{g}_{B}^{X}=(2n^{2}+n)-\left(n+\dfrac{1}{2}\sum_{i}s_{i}^{2}\right)=\dfrac{m^{2}}{2}-\frac{1}{2}\sum_{i}s_{i}^{2}.$
If $B$ is symmetric, the arguement is similar except that the form $(u,v)_{d}$
is symplectic and hence the vector space consisting of such forms
$(\cdot,\cdot)_{d}$ has dimension $\frac{1}{2}r_{d+1}(r_{d+1}-1)$. ∎
###### Remark 3.8.
The dimension of $I_{B}$-orbits can also be obtained from [Se, 3.1.c, 3.2.b],
where the formulae are not uniform and the proofs are also omitted.
The closure relation on the set of nilpotent orbits in $\mathcal{A}$ is given
by
$\mathcal{O}_{Y}\preceq\mathcal{O}_{X}\text{ if
}\mathcal{O}_{Y}\subset\overline{\mathcal{O}}_{X}$
for nilpotent elements $X,Y\in\mathcal{A}$. Given two partitions
$\bar{d}=[d_{1},...,d_{N}],\bar{f}=[f_{1},...,f_{N}]$ of $N$ (put some
$d_{i},f_{i}=0$ if needed). We say that $\bar{d}$ dominates $\bar{f}$, denoted
by $\bar{d}\succeq\bar{f}$ if
$d_{1}\geq f_{1}$
$d_{1}+d_{2}\geq f_{1}+f_{2}$
$\vdots$
$d_{1}+...+d_{N}\geq f_{1}+...+f_{N}.$
###### Theorem 3.9.
Let $X,Y$ be nilpotent elements in $\mathcal{A}$ having partition
$\bar{d},\bar{f}$, respectively. Then
$\mathcal{O}_{\bar{d}}\succeq\mathcal{O}_{\bar{f}}$ if and only if $\bar{d}$
dominates $\bar{f}$.
###### Proof.
See [Oh, Theorem 1]. ∎
For example, all nilpotent $\mathrm{Sp}_{10}$-orbits in
$\mathcal{A}\subset\mathfrak{sl}_{10}$ are
$\mathcal{O}_{[5^{2}]}\succeq\mathcal{O}_{[4^{2},1^{2}]}\succeq\mathcal{O}_{[3^{2},2^{2}]}\succeq\mathcal{O}_{[3^{2},1^{4}]}\succeq\mathcal{O}_{[2^{4},1^{2}]}\succeq\mathcal{O}_{[2^{2},1^{6}]}\succeq\mathcal{O}_{[1^{10}]}.$
The dimensions are 40, 36, 32, 28, 24, 16, 0, respectively.
## 4\. The connection between Schubert cells and nilpotent $K$-orbits
The goal of this section is to show that for any small $\bar{\lambda}$,
$\mathcal{M}_{\bar{\lambda}}$ is sent to the nilpotent cone
$\mathcal{N}_{\mathfrak{p}}$ by the map $\pi$, and show that how each
$\mathcal{M}_{\bar{\lambda}}$ is sent to nilpotent $K$-orbits in the
$\mathcal{N}_{\mathfrak{p}}$. Theorem 4.2 describes the image
$\pi(\mathcal{M}_{\bar{\lambda}})$ where the proofs are provided by case-by-
case consideration in this section.
Let $X_{N}$ be the type of Dynkin diagram of $G$ and $\sigma$ the diagram
automorphism on $G$ of order $r$, denoted by the pair $(X_{N},r)$. We consider
the cases $(X_{N},r)=(A_{2\ell},2),(A_{2\ell-1},2)$, and $(D_{\ell+1}2)$. Then
$H=(\check{G})^{\sigma}$ is either of type $B_{\ell}$ or $C_{\ell}$. We make
the following labelling for simple roots $\gamma_{i}$:
(9)
$\begin{cases}\gamma_{1}=\gamma_{\\{1,2\ell-1\\}},\dots,\gamma_{\ell-1}=\gamma_{\\{\ell-1,n+1\\}},\gamma_{n}=\gamma_{\\{\ell\\}}&\quad(X_{N},r)=(A_{2\ell-1},2);\\\
\gamma_{1}=\gamma_{\\{1,2\ell\\}},\dots,\gamma_{\ell-1}=\gamma_{\\{\ell-1,\ell+2\\}},\gamma_{\ell}=\gamma_{\\{\ell,\ell+1\\}}&\quad(X_{N},r)=(A_{2\ell},2);\\\
\gamma_{1}=\gamma_{\\{1\\}},\dots,\gamma_{\ell-1}=\gamma_{\\{\ell-1\\}},\gamma_{\ell}=\gamma_{\\{\ell,\ell+1\\}}&\quad(X_{N},r)=(D_{\ell+1},2).\\\
\end{cases}$
This labelling of vertices of type $B_{\ell}$ and $C_{\ell}$ agrees with the
labelling in [Ka, TABLE Fin, p.53]. Then the highest short root $\gamma_{0}$
of $H$ can be described in the following table.
$(X_{N},r)$ | $G$ | $H$ | Simple roots of $H$ | Highest short root $\gamma_{0}$ of $H$
---|---|---|---|---
$(A_{2\ell},2)$ | $\mathrm{SL}_{2\ell+1}$ | $\mathrm{PSO}_{2\ell+1}$ | $\gamma_{i}=\bar{\check{\alpha}}_{i}=\bar{\check{\alpha}}_{2\ell-i+1}$, $i=1,...,\ell$ | $\gamma_{1}+\gamma_{2}+...+\gamma_{\ell}.$
$(A_{2\ell-1},2)$ | $\mathrm{SL}_{2\ell}$ | $\mathrm{PSp}_{2\ell}$ | $\gamma_{i}=\bar{\check{\alpha}}_{i}=\bar{\check{\alpha}}_{2\ell-i}$, $i=1,...,\ell$ | $\gamma_{1}+2\gamma_{2}+...+2\gamma_{\ell-1}+\gamma_{\ell}.$
$(D_{\ell+1},2)$ | $\mathrm{Spin}_{2\ell+2}$ | $\mathrm{PSO}_{2\ell+1}$ | $\gamma_{i}=\bar{\check{\alpha}}_{i},i=1,...,\ell-1$ $\gamma_{\ell}=\bar{\check{\alpha}}_{\ell}=\bar{\check{\alpha}}_{\ell+1}$ | $\gamma_{1}+\gamma_{2}+...+\gamma_{\ell}.$
From Section 2.1, we can identify the weight lattice of $H$ with
$X_{*}(T)_{\sigma}$, and the set of dominant weights of $H$ with
$X_{*}(T)^{+}_{\sigma}$. Then, from the construction of root system of
classical Lie algebras given in [Hu2, §12], we can make the following
identifications:
$X_{*}(T)_{\sigma}\cong\begin{cases}\mathbb{Z}^{\ell}&\text{if }H=B_{\ell};\\\
\\{(a_{1},...,a_{\ell})\in\mathbb{Z}^{\ell}\mid a_{1}+\cdots+a_{\ell}\in
2\mathbb{Z}\\}&\text{if }H=C_{\ell}\end{cases}$
and
$X_{*}(T)_{\sigma}^{+}\cong\\{(a_{1},...,a_{\ell})\in X_{*}(T)_{\sigma}\mid
a_{1}\geq\cdots\geq a_{\ell}\geq 0\\}$
for any cases of $H$. This identification preserves the relation on
$X_{*}(T)_{\sigma}^{+}$ and the dominance relation on
$\\{(a_{1},...,a_{\ell})\in X_{*}(T)_{\sigma}\mid a_{1}\geq\cdots\geq
a_{\ell}\geq 0\\}$.
In the Table 1, we can further make the following identifications for simple
roots and fundamental weights of $H$. Those fundamental weights follows from
[Hu2, Table 1., p.69].
$H$ | Simple roots | $\gamma_{0}$ | Fundamental weights
---|---|---|---
$B_{\ell}$ | $\gamma_{1}=(1,-1,0,0,...,0)$ $\gamma_{2}=(0,1,-1,0,...,0)$ $\vdots$ $\gamma_{\ell-1}=(0,0,0,...,1,-1)$ $\gamma_{\ell}=(0,0,0,...,0,1)$ | $(1,0,0,...,0,0)$ | $\omega_{j}=(1^{j}0^{\ell-j})$, $j=0,...,\ell-1$ $\omega_{\ell}=(\dfrac{1}{2},\dfrac{1}{2},...,\dfrac{1}{2})$
$C_{\ell}$ | $\gamma_{1}=(1,-1,0,0,...,0)$ $\gamma_{2}=(0,1,-1,0,...,0)$ $\vdots$ $\gamma_{\ell-1}=(0,0,0,...,1,-1)$ $\gamma_{\ell}=(0,0,0,...,0,2).$ | $(1,1,0,...,0,0)$ | $\omega_{j}=(1^{j}0^{\ell-j}),j=0,...,\ell$
Table 1. Simple roots, highest short root, and fundamental weights of $H$ in
term of tuples
The following lemma is well-known, cf.[AH]. We give a self-contained proof
here.
###### Lemma 4.1.
All small dominant weights of $H$ are
1. (1)
$\omega_{j}=(1^{j}0^{\ell-j})$, $j=0,...,\ell-1$,
$2\omega_{\ell}=(1,1,...,1)$, if $H$ has the type $B_{\ell}$.
2. (2)
$\omega_{1}+\omega_{2j+1}=(21^{2j}0^{\ell-2j-1})$,
$j=0,...,\lfloor\frac{\ell-1}{2}\rfloor$
$\omega_{2j}=(1^{2j}0^{\ell-2j})$, $j=0,...,\lfloor\frac{\ell}{2}\rfloor$, if
$H$ has the type $C_{\ell}$.
###### Proof.
Suppose that $H$ has the type $B_{\ell}$. The highest short root is
$\gamma_{0}=(1,0,...,0)$. By definition, a dominant weight
$(a_{1},...,a_{\ell})\in X_{*}(T)_{\sigma}^{+}$ is small if and only if
$(a_{1},...,a_{\ell})\nsucceq(2,0,...,0)$ which is equivalent to $a_{1}\leq
1$. This proves the first part.
Now assume that $H$ has the type $C_{\ell}$. The highest short root is
$\gamma_{0}=(1,1,...,0)$. Let $(a_{1},...,a_{\ell})\in X_{*}(T)_{\sigma}^{+}$
be a small dominant weight. Then $(a_{1},...,a_{\ell})\nsucceq(2,2,...,0)$ and
so $a_{1}\leq 2$. If $a_{1}=1$, then
$(a_{1},...,a_{\ell})=(1^{2j}0^{\ell-2j})$. If $a_{1}=2$, then $a_{2}<2$ and
hence $(a_{1},...,a_{\ell})=(21^{2j}0^{\ell-2j-1})$. ∎
Let $\bar{\mu}$ be the maximal element among all small dominant weights of
$H$, then
$\mathcal{G}r_{\mathrm{sm}}=\coprod_{{\bar{\lambda}}\preceq\bar{\mu},\\\
\bar{\lambda}\text{
small}}\mathcal{G}r_{\bar{\lambda}}=\overline{\mathcal{G}r}_{\bar{\mu}}.$
Since $\mathcal{G}r_{\mathrm{sm}}$ is irreducible and $\mathcal{M}$ is an open
subset of $\mathcal{G}r_{sm}$, $\mathcal{M}$ is irreducible.
The following theorem is the main result of this section.
###### Theorem 4.2.
If $\bar{\lambda}$ is small, then $\pi(\mathcal{M}_{\bar{\lambda}})$ is
contained in $\mathcal{N}_{\mathfrak{p}}$. Moreover, the image
$\pi(\mathcal{M}_{\bar{\lambda}})$ can be described as the union of nilpotent
orbits as the following table:
$(X_{N},r)$ | Small dominant weight $\overline{\lambda}$ of $H$ | Orbits in $\pi(\mathcal{M}_{\bar{\lambda}})$
---|---|---
$(A_{2\ell},2)$ | $(1^{j}0^{\ell-j}),j=0,1,\dots,\ell$ | $[2^{j}1^{2\ell-2j+1}]$
$(A_{2\ell-1},2)$ | $(1^{2j}0^{\ell-2j}),j=0,1,\dots,\lfloor\frac{\ell}{2}\rfloor$ | $[2^{2j}1^{2\ell-4j}]$
$(20^{\ell-1})$ | $0,[2^{2}1^{2\ell-4}]$
$(21^{2}0^{\ell-3})$ | $[2^{2}1^{2\ell-4}],[2^{4}1^{2\ell-8}],[3^{2}1^{2\ell-6}]$
$(21^{2j}0^{\ell-2j-1}),j=2,\dots,\lfloor\frac{\ell-3}{2}\rfloor$ | $[2^{2j}1^{2\ell-4j}],[2^{2j+2}1^{2\ell-4j-4}],$ $[3^{2}2^{2j-2}1^{2\ell-4j-2}],[3^{2}2^{2j-4}1^{2\ell-4j+2}]$
$(21^{2\lfloor\frac{\ell-1}{2}\rfloor}0^{\ell-2\lfloor\frac{\ell-1}{2}\rfloor-1})$ | $[2^{\ell-2}1^{4}],[2^{\ell}],$ $[3^{2}2^{\ell-4}1^{2}],[3^{2}2^{\ell-6}1^{4}]$ , if $\ell$ is even
| $[2^{\ell-1}1^{2}]$, $[3^{2}2^{\ell-3}],[3^{2}2^{\ell-5}1^{4}]$ , if $\ell$
is odd
$(D_{\ell+1},2)$ | $(1^{j}0^{\ell-j}),j=0,1,\dots,\ell$ | $0$, if $j=0$
$0,[31^{2\ell-1}]$, if $j$ is even, $j\geq 2$
| | $[31^{2\ell-1}]$, if $j$ is odd
where the nilpotent orbit $[a_{1}^{i_{1}},...,a_{r}^{i_{r}}]$ in the above
table means empty set if the associated partition is invalid for some small
$\ell$.
This theorem follows from Theorem 4.5, Theorem 4.6, Theorem 4.10, and Theorem
4.14, which will be proved separately case by case.
The partial orders of small dominant weights of $H$ are shown in the below
picture, where the partial order is compatible with the height.
${{2\omega_{\ell}}}$${{\omega_{1}+\omega_{2\lfloor\frac{\ell-1}{2}\rfloor+1}}}$${{\omega_{\ell-1}}}$${{\omega_{1}+\omega_{2\lfloor\frac{\ell-1}{2}\rfloor-1}}}$${{\omega_{\ell-2}}}$${{\omega_{1}+\omega_{2\lfloor\frac{\ell-1}{2}\rfloor-3}}}$${{\omega_{2\lfloor\frac{\ell}{2}\rfloor}}}$${{\omega_{2\lfloor\frac{\ell}{2}\rfloor-2}}}$${{\omega_{1}+\omega_{3}}}$${{\omega_{2}}}$${{2\omega_{1}}}$${{\omega_{4}}}$${{\omega_{1}}}$${{\omega_{2}}}$${{0}}$${{0}}$${\text{Type
}B_{\ell}}$${\text{Type }C_{\ell}}$
We first recall a crucial lemma from [AH, Lemma 4.3].
###### Lemma 4.3.
Let $g=\sum_{i=N}^{\infty}x_{i}t^{i}\in\mathrm{SL}_{n}(\mathcal{K})$, where
$x_{N}\neq 0$. Let ${\lambda}=(a_{1},a_{2},...,a_{n})$ be a tuple of integers
such that $a_{1}\geq a_{2}\cdots\geq a_{n}$ and $\sum a_{i}=0$, and
$g(t)\in\mathrm{SL}_{n}(\mathcal{O})t^{{\lambda}}\mathrm{SL}_{n}(\mathcal{O})$.
Then
1. (1)
$N=a_{n}$.
2. (2)
The rank of $x_{N}$ equals to the number of $j$ such that $a_{j}=a_{n}$.
3. (3)
For any $s\geq 1$,
$\operatorname{rk}\begin{pmatrix}x_{N}&x_{N+1}&\cdots&x_{N+s-2}&x_{N+s-1}\\\
0&x_{N}&\cdots&x_{N+s-3}&x_{N+s-2}\\\ \vdots&\vdots&\ddots&\vdots&\vdots\\\
0&0&\cdots&x_{N}&x_{N+1}\\\
0&0&\cdots&0&x_{N}\end{pmatrix}=\sum_{j=1}^{n}\max\\{s-(a_{j}-a_{n}),0\\}.$
We have the following lemma for the twisted version.
###### Lemma 4.4.
Let $g(t)=\sum_{i=N}^{\infty}x_{i}t^{i}\in G(\mathcal{K})^{\sigma}$ where
$x_{i}\in\operatorname{Mat}_{m\times m}$, $x_{N}\neq 0$. Let
$\bar{\lambda}=(a_{1},...,a_{\ell})\in X_{*}(T)_{\sigma}^{+}$ be such that
$g(t)\in G(\mathcal{O})^{\sigma}n^{\lambda}G(\mathcal{O})^{\sigma}$. Then
1. (1)
$N=\begin{cases}-a_{1}&\text{if }(X_{N},r)=(A_{2\ell},2),(A_{2\ell-1},2);\\\
-2a_{1}&\text{if }(X_{N},r)=(D_{\ell+1},2).\end{cases}$
2. (2)
The rank of $x_{N}$ is equal to the number of $j$ such that $a_{j}=a_{1}$.
###### Proof.
We can write
$\bar{\lambda}=(a_{1},...,a_{\ell})=\sum_{i=1}^{\ell}\left(\sum_{j=1}^{i}a_{j}\right)\gamma_{i}$
where $\gamma_{i}$ are simple roots of $H$ as labelled by (9). We choose a
representative $\lambda\in X_{*}(T)$ of $\bar{\lambda}$ by
$\lambda=\begin{cases}\sum_{i=1}^{\ell}\left(\sum_{j=1}^{i}a_{j}\right)\check{\alpha}_{i}&\text{if
}(X_{N},r)=(A_{2\ell},2),(D_{\ell+1},2);\\\
\sum_{i=1}^{\ell-1}\left(\sum_{j=1}^{i}a_{j}\right)\check{\alpha}_{i}+\dfrac{1}{2}\left(\sum_{j=1}^{\ell}a_{j}\right)\check{\alpha}_{\ell}&\text{if
}(X_{N},r)=(A_{2\ell-1},2)\\\ \end{cases}$
so that
$\lambda+\sigma\lambda=\begin{cases}\sum_{i=1}^{\ell}\left(\sum_{j=1}^{i}a_{j}\right)\check{\alpha}_{i}+\sum_{i=\ell+1}^{2\ell}\left(\sum_{j=1}^{2\ell-i+1}a_{j}\right)\check{\alpha}_{i}&\text{if
}(X_{N},r)=(A_{2\ell},2);\\\
\sum_{i=1}^{\ell}\left(\sum_{j=1}^{i}a_{j}\right)\check{\alpha}_{i}+\sum_{i=\ell+1}^{2\ell-1}\left(\sum_{j=1}^{2\ell-i}a_{j}\right)\check{\alpha}_{i}&\text{if
}(X_{N},r)=(A_{2\ell-1},2);\\\
\sum_{i=1}^{\ell-1}\left(\sum_{j=1}^{i}2a_{j}\right)\check{\alpha}_{i}+\left(\sum_{j=1}^{\ell}a_{j}\right)(\check{\alpha}_{\ell}+\check{\alpha}_{\ell+1})&\text{if
}(X_{N},r)=(D_{\ell+1},2).\\\ \end{cases}$
The simple coroots of ${G}$ are identified with tuples of integers through the
construction of root system given from [Hu2, §12].
Let $\rho:G\to{\rm GL}(V)$ be the standard representation of $G$. We will
determine the double ${\rm SL(V_{\mathcal{O}})}$-coset in ${\rm
SL}(V_{\mathcal{K}})$ that $\rho(g(t))$ belongs to.
If ${G}$ is of the type $A_{m}$, then $\check{\alpha}_{i}$, $i=1,...,m$, are
identified with the following $(m+1)$-tuples
$\check{\alpha}_{1}=(1,-1,0,0,...,0)$
$\check{\alpha}_{2}=(0,1,-1,0,...,0)$
$\vdots$
$\check{\alpha}_{m}=(0,0,0,...,1,-1)$
and hence, as the coweight of $\mathrm{SL}_{m}$, $\lambda+\sigma\lambda$
corresponds to the following tuples
$\lambda+\sigma\lambda=\begin{cases}(a_{1},a_{2},...,a_{\ell},0,-a_{\ell},...,-a_{2},-a_{1})&\text{if
}(X_{N},r)=(A_{2\ell},2);\\\
(a_{1},a_{2},...,a_{\ell},-a_{\ell},...,-a_{2},-a_{1})&\text{if
}(X_{N},r)=(A_{2\ell-1},2).\end{cases}$
Assume that ${G}$ has the type $D_{\ell+1}$. Then $\check{\alpha}_{i}$,
$i=1,...,\ell+1$, are identified with following $(\ell+1)$-tuples
$\check{\alpha}_{1}=(1,-1,0,0,...,0)$
$\check{\alpha}_{2}=(0,1,-1,0,...,0)$
$\vdots$
$\check{\alpha}_{\ell}=(0,0,0,...,1,-1)$
$\check{\alpha}_{\ell+1}=(0,0,0,...,1,1).$
Then $\lambda+\sigma\lambda=(2a_{1},2a_{2},...,2a_{\ell},0)$ as the coweight
of $G=\mathrm{Spin}_{2\ell+2}$. Choose an appropriate maximal torus and a
positive root system in $G$. Composing with $\rho:G\to{\rm SL}_{2\ell+2}$, as
the coweight of $\mathrm{SL}_{2\ell+2}$, $\lambda+\sigma\lambda$ corresponds
to the following tuple
$(2a_{1},2a_{2},...,2a_{\ell},0,0,-2a_{\ell},...,-2a_{2},-2a_{1}).$
We write $g(t)=A(t)n^{\lambda}B(t)$, where $n^{\lambda}$ is a norm of
$t^{\lambda}$ defined by (1) and $A(t),B(t)\in G(\mathcal{O})^{\sigma}$. Hence
$\rho(g(t))\in\mathrm{SL}_{m}(\mathcal{O})\rho(t^{\lambda+\sigma\lambda})\mathrm{SL}_{m}(\mathcal{O})$.
By above descriptions of $\rho(t^{\lambda+\sigma\lambda})$, this lemma follows
from Lemma 4.3. ∎
### 4.1. Case $(X_{N},r)=(A_{2\ell},2)$
Let $\langle\cdot,\cdot\rangle$ be a nondegenerate symmetric bilinear form on
$V=\mathbb{C}^{2\ell+1}$ whose matrix is
$J=\begin{pmatrix}&&&&&&1\\\ &&&&&-1&\\\ &&&&1&&\\\ &&&-1&&&\\\
&&\reflectbox{$\ddots$}&&&&\\\ &-1&&&&&\\\ 1&&&&&&\end{pmatrix}.$
The diagram automorphism $\sigma$ on $\mathfrak{g}$ given by (7) becomes
$\sigma(A)=-JA^{T}J^{-1}$ and the diagram automorphism $\sigma$ on $G$ is
given by
(10) $\sigma(A)=JA^{-T}J^{-1}.$
This $\sigma$ gives the decomposition
$\mathfrak{g}=\mathfrak{k}\oplus\mathfrak{p}$ to 1 and -1 eigenspaces
$\mathfrak{k}$ and $\mathfrak{p}$, respectively. Let
$K:=(\mathrm{SL}_{2\ell+1})^{\sigma}=\mathrm{SO}_{2\ell+1}$. The
classification of nilpotent $K$-orbits in $\mathfrak{p}$ and their dimensions
follow from the Theorem 3.4 and 3.7.
Set
(11)
$\mathcal{N}_{\mathfrak{p},2}=\\{x\in\mathcal{N}_{\mathfrak{p}}\,|\,x^{2}=0\\}.$
###### Theorem 4.5.
$\pi$ maps $\mathcal{M}$ isomorphically onto $\mathcal{N}_{\mathfrak{p},2}$.
Moreover, $\pi$ maps $\mathcal{M}_{(1^{j}0^{\ell-j})}$ isomorphically to
$[2^{j}1^{2\ell-2j+1}]$.
###### Proof.
Let $g(t)\cdot e_{0}\in\mathcal{M}$. Then $g(t)\cdot
e_{0}\in\mathcal{M}_{(1^{j}0^{\ell-j})}$ for some $j$. By Lemma 4.4,
$g(t)=I+xt^{-1}$ for some $x\in\operatorname{Mat}_{2\ell+1\times 2\ell+1}$. By
Lemma 2.4, $\iota(g(t)\cdot e_{0})\in\mathcal{M}_{(1^{j}0^{\ell-j})}$. Hence
$\iota(g(t))=(I-xt^{-1})^{-1}=I+zt^{-1}$ for some
$z\in\operatorname{Mat}_{2\ell+1\times 2\ell+1}$, and so $x^{2}=0$.
Conversely, let $x\in\mathcal{N}_{\mathfrak{p}}$ be such that $x^{2}=0$. Then
$I+xt^{-1}\in G(\mathcal{K})^{\sigma}$. By (2), $(I+xt^{-1})\cdot
e_{0}\in\mathcal{G}r_{\bar{\lambda}}$ for some
$\bar{\lambda}=(a_{1},...,a_{\ell})\in X_{*}(T)_{\sigma}^{+}$. By lemma 4.4,
$\bar{\lambda}=(1^{j}0^{\ell-j})$. We have proved that
$\mathcal{M}=\\{(I+xt^{-1})\cdot e_{0}\mid x\in\mathfrak{p},x^{2}=0\\}$.
Let $(I+xt^{-1})\cdot e_{0}\in\mathcal{M}_{(1^{j}0^{\ell-i})}$. Then $x^{2}=0$
and $x$ has the Jordan blocks of size at most 2. By Lemma 4.4,
$\operatorname{rk}x=j$ and then $x$ has the partition $[2^{j}1^{2\ell-2j+1}]$.
It is obvious that $\pi$ is injective. To prove surjectivity, let
$x\in\mathcal{N}_{\mathfrak{p},2}$ having the partition
$[2^{j}1^{2\ell-2j+1}]$. Then, $(I+xt^{-1})\cdot e_{0}\in\mathcal{M}$ and
hence $(I+xt^{-1})\cdot e_{0}\in\mathcal{M}_{(1^{k}0^{\ell-k})}$ for some $k$.
In fact, $k=j$ since $\pi((I+xt^{-1})\cdot e_{0})=x\in[2^{k}1^{2\ell-2k+1}].$
∎
### 4.2. Case $(X_{N},r)=(A_{2\ell-1},2)$
Let $\langle\cdot,\cdot\rangle$ be a symplectic bilinear form on
$V=\mathbb{C}^{2\ell}$ whose matrix is
$J=\begin{pmatrix}&&&&&&1\\\ &&&&&-1&\\\ &&&&1&&\\\ &&&-1&&&\\\
&&\reflectbox{$\ddots$}&&&&\\\ &1&&&&&\\\ -1&&&&&&\end{pmatrix}.$
The diagram automorphism $\sigma$ on $\mathfrak{g}$ given by (7) becomes
$\sigma(A)=-JA^{T}J^{-1}$ and the action on $G$ is given by
(12) $\sigma(A)=JA^{-T}J^{-1}.$
This $\sigma$ gives the decomposition
$\mathfrak{g}=\mathfrak{k}\oplus\mathfrak{p}$ to 1 and -1 eigenspaces
$\mathfrak{k}$ and $\mathfrak{p}$, respectively. Let
$K:=(\mathrm{SL}_{2\ell})^{\sigma}=\mathrm{Sp}_{2\ell}$. The classification of
nilpotent $K$-orbits in $\mathfrak{p}$ and their dimensions follow from the
Theorem 3.3 and 3.7.
Define constructible sets
$\mathcal{M}^{\prime}:=\bigcup_{j=0}^{\lfloor\frac{\ell}{2}\rfloor}\mathcal{M}_{(1^{2j}0^{\ell-2j})},\hskip
14.22636pt\mathcal{M}^{\prime\prime}:=\bigcup_{j=0}^{\lfloor\frac{\ell-1}{2}\rfloor}\mathcal{M}_{(21^{2j}0^{\ell-2j-1})}.$
By Lemma 4.4, the element of $\mathcal{M}^{\prime}$ is of the form
$(I+xt^{-1})\cdot e_{0}$ and the element of $\mathcal{M}^{\prime\prime}$ is of
the form $(I+xt^{-1}+yt^{-2})\cdot e_{0}$. Let $\mathcal{N}_{\mathfrak{p},2}$
be the order 2 nilpotent cone defined as in (11).
###### Theorem 4.6.
$\pi$ maps $\mathcal{M^{\prime}}$ isomorphically onto
$\mathcal{N}_{\mathfrak{p},2}$. Moreover, $\pi$ maps
$\mathcal{M}_{(1^{2j}0^{\ell-2j})}$ isomorphically to $[2^{2j}1^{2\ell-4j}]$.
###### Proof.
The proof is the same as the proof of Theorem 4.5, where in this case we use
Lemma 4.4 for $(A_{2\ell-1},2)$. ∎
Before we describe elements of $M^{\prime\prime}$, we need the following
lemma.
###### Lemma 4.7.
Let $g(t)\cdot e_{0}=(I+xt^{-1}+yt^{-2})\cdot
e_{0}\in\mathcal{M}^{\prime\prime}$. Then
1. (1)
$\iota(g(t))\neq g(t)$
2. (2)
If $g(t)\cdot e_{0}\in\mathcal{M}_{(21^{2j}0^{\ell-2j-1})}$, then
$\operatorname{rk}\begin{pmatrix}y&x\\\ 0&y\end{pmatrix}=2j+2.$
###### Proof.
Note that $\iota(g(t))=g(t)$ if and only if
$(I+xt^{-1}+yt^{-2})(I-xt^{-1}+yt^{-2})=I$
which is equivalent to $y=\dfrac{1}{2}x^{2}$ and $x^{4}=0$. Suppose that
$\iota(g(t))=g(t)$. Observe that
$\operatorname{rk}x^{3}\leq\operatorname{rk}x^{2}=\operatorname{rk}y=1$. If
$\operatorname{rk}x^{3}=1$, then
$\operatorname{rk}x^{4}=\operatorname{rk}x^{3}=1$ which is impossible. Hence
$x^{3}=0$. Since $\operatorname{rk}x^{2}=1$, $x\in\mathfrak{p}$ is nilpotent
having the partition $[32^{j}1^{2l-2j-3}]$ but this contradicts to Theorem
3.3. This proves the first part.
Assume that $g(t)\cdot e_{0}\in\mathcal{M}_{(21^{2j}0^{\ell-2j-1})}$. We write
$g(t)=At^{(2,1^{2j},0^{2\ell-4j-2},(-1)^{2j},-2)}B$
where $A=\sum_{i=0}A_{i}t^{i},B=\sum_{i=0}B_{i}t^{i}\in
G(\mathcal{O})^{\sigma}$. In particular,
$g(t)\in\mathrm{SL}_{2\ell}(\mathcal{O})t^{\check{\lambda}}\mathrm{SL}_{2\ell}(\mathcal{O})$
where $\check{\lambda}=(2,1^{2j},0^{2\ell-4j-2},(-1)^{2j},-2)$. By Lemma 4.3,
$\operatorname{rk}\begin{pmatrix}y&x\\\
0&y\end{pmatrix}=\sum_{j=1}^{2\ell}\max\\{-a_{j},0\\}=2j+2$
as desired. ∎
Let $g(t)=I+xt^{-1}+yt^{-2}$. By Lemma 2.4,
$\iota(g(t))=I+x^{\prime}t^{-1}+y^{\prime}t^{-2}$ for some matrices
$x^{\prime},y^{\prime}$. Hence
$(I-xt^{-1}+yt^{-2})(I+x^{\prime}t^{-1}+y^{\prime}t^{-2})=I=(I+x^{\prime}t^{-1}+y^{\prime}t^{-2})((I-xt^{-1}+yt^{-2})$
which implies
(13) $x^{\prime}=x,\hskip 7.11317ptx^{2}=y+y^{\prime},\hskip
7.11317ptxy=y^{\prime}x,\hskip 7.11317ptxy^{\prime}=yx,\hskip
7.11317ptyy^{\prime}=y^{\prime}y=0.$
By Lemma 4.4 and Lemma 4.7, $\operatorname{rk}y=\operatorname{rk}y^{\prime}=1$
and $y^{\prime}\neq y$. Since $\sigma(g(t))=g(t)$, $y^{\prime}=Jy^{T}J^{-1}$
which means that $y$ and $y^{\prime}$ are adjoint each other. We set
$\mathcal{M}_{(21^{2j}0^{\ell-2j-1})}^{\mathrm{I}}:=\\{(I+xt^{-1}+yt^{-2})\cdot
e_{0}\in\mathcal{M}_{(21^{2j}0^{\ell-2j-1})}\mid L=L^{\prime}\\},$
$\mathcal{M}_{(21^{2j}0^{\ell-2j-1})}^{\mathrm{II}}:=\\{(I+xt^{-1}+yt^{-2})\cdot
e_{0}\in\mathcal{M}_{(21^{2j}0^{\ell-2j-1})}\mid L\neq L^{\prime}\\}.$
The following lemma will be used in the proofs of Lemma 4.9 and Theorem 4.14.
###### Lemma 4.8.
Let $(\cdot,\cdot)$ be a nondegenerate bilinear form on a vector space $V$
over a field $\mathbb{C}$ and let $T$ a linear map on $V$. Denote the adjoint
of $T$ by $T^{*}$. Assume that $\operatorname{Im}T=\operatorname{Im}T^{*}$ and
$\operatorname{rk}T=1$. Then $T$ is self-adjoint or skew-adjoint.
###### Proof.
It is easy to see that $\ker T=(\operatorname{Im}T^{*})^{\perp}$ and $\ker
T^{*}=(\operatorname{Im}T)^{\perp}$. Say that $\operatorname{Im}T=\mathbb{C}v$
and $\operatorname{Im}T^{*}=\mathbb{C}v^{\prime}$ for some $v,v^{\prime}\in
V$. Then $Tw=v$ for some $w\in V$. Since
$\operatorname{Im}T=\operatorname{Im}T^{*}$, we have $T^{*}w=\lambda v$ for
some $\lambda\in\mathbb{C}$. Let $u\in V$. Then $Tu=kv$ for some
$k\in\mathbb{C}$. Since $T(u-kw)=0$, $u-kw\in\ker
T=(\operatorname{Im}T^{*})^{\perp}=(\operatorname{Im}T)^{\perp}=\ker T^{*}$.
Hence $T^{*}u=T^{*}(kw)=k\lambda v=\lambda Tu$. Since $u$ is arbitrary,
$T^{*}=\lambda T$. Consider $T+T^{*}=(1+\lambda)T$. Then
$(1+\lambda)T=T+T^{*}=(T+T^{*})^{*}=(1+\lambda)T^{*}.$
Therefore $1+\lambda=0$ or $T=T^{*}$ which means that $T$ is skew-adjoint or
self-adjoint. ∎
###### Lemma 4.9.
If $(I+xt^{-1}+yt^{-2})\cdot
e_{0}\in\mathcal{M}_{(21^{2j}0^{\ell-2j-1})}^{\mathrm{I}}$, then
$y^{\prime}=-y$.
###### Proof.
We know that $y\neq y^{\prime}$ are adjoint each other, they have the same
images, and $\operatorname{rk}y=1$. By Lemma 4.8, $y$ is skew-adjoint, i.e.,
$y^{\prime}=-y$. ∎
###### Theorem 4.10.
1. (1)
If $\ell$ is even, then
$\pi(\mathcal{M}_{(21^{2j}0^{\ell-2j-1})}^{\mathrm{I}})=[2^{2j}1^{2\ell-4j}]\cup[2^{2j+2}1^{2\ell-4j-4}]$
for $j=0,1,...,\frac{\ell-2}{2}$. If $\ell$ is odd, then
$\pi(\mathcal{M}_{(21^{2j}0^{\ell-2j-1})}^{\mathrm{I}})=\begin{cases}[2^{2j}1^{2\ell-4j}]\cup[2^{2j+2}1^{2\ell-4j-4}]&\hskip
1.42271pt\text{if }\ell\geq 3,\,0\leq j\leq\frac{\ell-3}{2};\\\
[2^{\ell-1}1^{2}]&\hskip 1.42271pt\text{if }j=\frac{\ell-1}{2}.\end{cases}$
2. (2)
When $\ell\geq 3$, for $j=1,...,\lfloor\frac{\ell-1}{2}\rfloor$, we have
$\pi(\mathcal{M}_{(21^{2j}0^{\ell-2j-1})}^{\mathrm{II}})=\begin{cases}[3^{2}1^{2\ell-6}]&\hskip
1.42271pt\text{if }j=1;\\\
[3^{2}2^{2j-2}1^{2\ell-4j-2}]\cup[3^{2}2^{2j-4}1^{2\ell-4j+2}]&\hskip
1.42271pt\text{if }\ell\geq 4,\,2\leq
j\leq\lfloor\frac{\ell-1}{2}\rfloor.\end{cases}$
Moreover, for any $\ell\geq 1$, $\mathcal{M}_{(20^{\ell-1})}^{\mathrm{II}}$ is
empty.
###### Proof.
Let $g(t)\cdot e_{0}=(I+xt^{-1}+yt^{-2})\cdot
e_{0}\in\mathcal{M}_{(21^{2j}0^{\ell-2j-1})}^{\mathrm{I}}$. By Lemma 4.7,
$\operatorname{rk}x\leq\operatorname{rk}\begin{pmatrix}y&x\\\
0&y\end{pmatrix}=2j+2\leq
2\operatorname{rk}y+\operatorname{rk}x=2+\operatorname{rk}x.$
Since $x^{2}=y+y^{\prime}=0$, $x$ is nilpotent whose partition is
$[2^{2k}1^{2\ell-4k}]$ so that $\operatorname{rk}x=2k$. Hence $k=j$ or $j+1$.
If $\ell$ is odd and $j=\frac{\ell-1}{2}$, then $k=j$.
Let $E_{ij}\in\operatorname{Mat}_{2\ell\times 2\ell}$ be the matrix which has
1 at the entry $i,j$ and 0 elsewhere. For each
$j=0,1,\dots,\lfloor\frac{\ell-1}{2}\rfloor$, let
$x_{j}=\operatorname{diag}(0,J_{2},..,J_{2},0_{2\ell-4j-2},-J_{2},..,-J_{2},0)$
where there are $j$ blocks of $J_{2}=\begin{pmatrix}0&1\\\ 0&0\end{pmatrix}$
and $j$ blocks of $-J_{2}$, and $0_{2\ell-4j-2}$ is the square zero matrix of
size $2\ell-4j-2$. Then $x_{j}\in\mathfrak{p}$ is nilpotent and has the
partition $[2^{2j}1^{2\ell-4j}]$. It is easy to check that
$g(t)=I+x_{j}t^{-1}+E_{1,2\ell}t^{-2}\in G(\mathcal{K})^{\sigma}$ and
$\iota(g(t))=I+x_{j}t^{-1}-E_{1,2\ell}t^{-2}$. By (2), $g(t)\cdot
e_{0}\in\mathcal{G}r_{\bar{\lambda}}$ for some
$\bar{\lambda}=(a_{1},...,a_{\ell})\in X_{*}(T)_{\sigma}^{+}$ with $a_{1}\geq
a_{2}...\geq a_{\ell}\geq 0$ and $\sum a_{i}$ is even. By Lemma 4.4, since
$\operatorname{rk}E_{1,2\ell}=1$, $\bar{\lambda}=(21^{2k}0^{\ell-2k-1})$ for
some $k$. By Lemma 4.7,
$2k+2=\operatorname{rk}\begin{pmatrix}E_{1,2\ell}&x_{j}\\\
0&E_{1,2\ell}\end{pmatrix}=2j+2.$
Then $(I+x_{j}t^{-1}+E_{1,2\ell}t^{-2})\cdot
e_{0}\in\mathcal{M}_{(21^{2j}0^{\ell-2j-1})}^{\mathrm{I}}$. For
$j=0,1,\dots,\lfloor\frac{\ell-2}{2}\rfloor$, let
$x^{\prime}_{j}=x_{j}+E_{1,2j+2}-E_{2\ell-2j-1,2\ell}.$
Then $x^{\prime}_{j}\in\mathfrak{p}$ is nilpotent and has the partition
$[2^{2j+2}1^{2\ell-4j-4}]$. Similarly, one can show that
$(I+x^{\prime}_{j}t^{-1}+E_{1,2\ell}t^{-2})\cdot
e_{0}\in\mathcal{M}_{(21^{2j}0^{\ell-2j-1})}^{\mathrm{I}}$. Since $\pi$ is
$K$-invariant, this proves the first part.
Now, we prove the second part. Let $g(t)\cdot e_{0}=(I+xt^{-1}+yt^{-2})\cdot
e_{0}\in\mathcal{M}_{(21^{2j}0^{\ell-2j-1})}^{\mathrm{II}}$. Set
$U=L+L^{\prime}$. Since $y\neq y^{\prime}$, $\dim U=2$ and
$U=\operatorname{Im}x^{2}$. Assume that
$L=\mathbb{C}v,L^{\prime}=\mathbb{C}v^{\prime}$. By (13), we have
$xy=y^{\prime}x$ and $xy^{\prime}=yx$. Hence $xv=bv^{\prime}$ and
$xv^{\prime}=av$ for some $a,b\in\mathbb{C},v,v^{\prime}\in\mathbb{C}^{2l}$.
Then
$x\big{|}_{U}=\begin{pmatrix}0&a\\\ b&0\end{pmatrix},\hskip
14.22636ptx^{2}\big{|}_{U}=\begin{pmatrix}ab&0\\\ 0&ab\end{pmatrix}$
Suppose that $ab\neq 0$. To show that
$\langle\cdot,\cdot\rangle\big{|}_{U\times U}$ is nondegenerate, let
$u\in\mathbb{C}^{2\ell}$ be such that $\langle x^{2}u,x^{2}v\rangle=0$ for all
$v\in\mathbb{C}^{2\ell}$. Since $x^{2}$ is self-adjoint, $\langle
x^{4}u,v\rangle=0$ for all $v$. Therefore, $x^{4}u=0$ and so $x^{2}u\in\ker
x^{2}\big{|}_{U}=0$. Since $y,y^{\prime}$ are adjoint each other and
$yy^{\prime}=0=y^{\prime}y$, $L^{\prime}\subset\ker y=(L^{\prime})^{\perp}$
and $L\subset\ker y^{\prime}=L^{\perp}$. This implies $\langle
v,v\rangle=\langle v^{\prime},v^{\prime}\rangle=0$ and $\langle
v,v^{\prime}\rangle\neq 0$. Observe that
$ab\langle v,v^{\prime}\rangle=\langle v,x^{2}v^{\prime}\rangle=\langle
xv,xv^{\prime}\rangle=\langle bv^{\prime},av\rangle=-ab\langle
v,v^{\prime}\rangle$
which implies $ab=0$, a contradiction. This proves $x^{4}=0$. Since
$x\in\mathfrak{p}$ and $\operatorname{rk}x^{2}=2$, by Theorem 3.3, $x$ is
nilpotent having the partition $[3^{2}2^{2k}1^{2\ell-4k-6}]$. By Lemma 4.7,
$\operatorname{rk}x\leq\operatorname{rk}\begin{pmatrix}y&x\\\
0&y\end{pmatrix}=2j+2\leq
2\operatorname{rk}y+\operatorname{rk}x=2+\operatorname{rk}x.$
Since $\operatorname{rk}x=2k+4$, $k=j-1$ or $j-2$. Here we see that $j\neq 0$
and hence $\mathcal{M}_{(20^{\ell-1})}^{\mathrm{II}}$ is empty. When $j=1$, we
see that $k=0$. Let
$J_{3}=\begin{pmatrix}0&1&0\\\ 0&0&1\\\ 0&0&0\end{pmatrix},\hskip
5.69046ptJ_{2}=\begin{pmatrix}0&1\\\ 0&0\end{pmatrix}.$
For each $j=1,...,\lfloor\frac{\ell-1}{2}\rfloor$, let
$x_{j-1}=\operatorname{diag}(J_{3},J_{2},...,J_{2},0_{2\ell-4j-2},-J_{2},...,-J_{2},-J_{3})$
where there are $j-1$ blocks of $J_{2}$, and $j-1$ blocks of $-J_{2}$. Then
$x_{j-1}\in\mathfrak{p}$ is nilpotent having the partition
$[3^{2}2^{2j-2}1^{2\ell-4j-2}]$. Note that
$g(t):=1+x_{j-1}t^{-1}+E_{13}t^{-2}\in G(\mathcal{K})^{\sigma}$ and
$\iota(g(t))=1+x_{j-1}t^{-1}+E_{2\ell-2,2\ell}t^{-2}$. By (2), $g(t)\cdot
e_{0}\in\mathcal{G}r_{\bar{\lambda}}$ for some
$\bar{\lambda}=(a_{1},...,a_{\ell})\in X_{*}(T)_{\sigma}^{+}$ with $a_{1}\geq
a_{2}...\geq a_{\ell}\geq 0$ and $\sum a_{i}$ is even. By Lemma 4.4, since
$\operatorname{rk}E_{13}=1$, $\bar{\lambda}=(21^{2k}0^{\ell-2k-1})$ for some
$k$. By Lemma 4.7,
$2k+2=\operatorname{rk}\begin{pmatrix}E_{13}&x_{j-1}\\\
0&E_{13}\end{pmatrix}=2j+2.$
Then $(I+x_{j}t^{-1}+E_{13}t^{-2})\cdot
e_{0}\in\mathcal{M}_{(21^{2j}0^{\ell-2j-1})}^{\mathrm{II}}$. For
$j=2,...,\lfloor\frac{\ell-1}{2}\rfloor$, let
$x^{\prime}_{j-2}=\operatorname{diag}(0,J_{3},J_{2},...,J_{2},0_{2\ell-4j},-J_{2},...,-J_{2},-J_{3},0)$
where there are $j-2$ blocks of $J_{2}$, and $j-2$ blocks of $-J_{2}$. Then
$x^{\prime}_{j-2}\in\mathfrak{p}$ is nilpotent having the partition
$[3^{2}2^{2j-4}1^{2\ell-4j+2}]$. One can check that
$h(t):=1+x^{\prime}_{j-2}t^{-1}+(E_{24}+E_{1,2\ell})t^{-2}\in
G(\mathcal{K})^{\sigma}$ and
$\iota(h(t))=1+x^{\prime}_{j-2}t^{-1}+(E_{2\ell-3,2\ell-1}-E_{1,2\ell})t^{-2}$.
Similarly, $h(t)\cdot e_{0}\in\mathcal{G}r_{\bar{\lambda}}$ where
$\bar{\lambda}=(21^{2k}0^{\ell-2k-1})$ for some $k$. By Lemma 4.7,
$2k+2=\operatorname{rk}\begin{pmatrix}E_{24}+E_{1,2\ell}&x^{\prime}_{j-2}\\\
0&E_{24}+E_{1,2\ell}\end{pmatrix}=2j+2.$
Then $(I+x^{\prime}_{j-2}t^{-1}+(E_{24}+E_{1,2\ell})t^{-2})\cdot
e_{0}\in\mathcal{M}_{(21^{2j}0^{\ell-2j-1})}^{\mathrm{II}}$. ∎
In the following proposition, we describe the fibers of
$\pi:\mathcal{M}\to\pi(\mathcal{M})$.
###### Proposition 4.11.
For any $x\in\pi(\mathcal{M})$,
$\pi^{-1}(x)\cong\\{z\in\mathfrak{sp}_{2\ell}\mid
xz+zx=0,z^{2}=0,\operatorname{rk}(z+\frac{1}{2}x^{2})\leq 1\\}.$
In particular, $\pi^{-1}(0)$ is isomorphic to the closure of nilpotent orbit
$[21^{2\ell-2}]$ in $\mathfrak{sp}_{2\ell}$ and $\dim\pi^{-1}(0)=2\ell+1$.
###### Proof.
Note that $(1+xt^{-1}+yt^{-2})\cdot e_{0}\in\mathcal{M}$ if and only if
$\det(1+xt^{-1}+yt^{-2})=1,\operatorname{rk}y\leq 1$ and
(14) $x^{T}J-Jx=0,\quad-x^{T}Jx+y^{T}J+Jy=0,\hskip
14.22636ptx^{T}Jy-y^{T}Jx=0,\hskip 14.22636pty^{T}Jy=0.$
Set $z=y-\frac{1}{2}x^{2}$, (14) is equivalent to
(15) $x\in\mathfrak{p},\quad z\in\mathfrak{k},\quad xz+zx=0,\quad z^{2}=0.$
If $xz+zx=0$, then $\det(1+xt^{-1}+(z+\frac{1}{2}x^{2})t^{-2})=1$. Since
$\pi^{-1}(0)=\\{(1+yt^{-2})\cdot e_{0}\mid
y\in\mathfrak{sp}_{2\ell},y^{2}=0,\operatorname{rk}y\leq 1\\}$, the dimension,
cf. [CM, Corollary 6.1.4], is
$\dim\pi^{-1}(0)=\dim[21^{2\ell-2}]=(2\ell^{2}+\ell)-\frac{1}{2}((2\ell-1)^{2}+1^{2})-\frac{1}{2}(2\ell-2)=2\ell+1.$
∎
${{\mathcal{M}_{(21^{4})}^{\mathrm{I}}}}$${{\mathcal{M}_{(21^{4})}^{\mathrm{II}}}}$${{[3^{2}2^{2}]}}$${{\mathcal{M}_{(21^{2}0^{2})}^{\mathrm{I}}}}$${{[3^{2}1^{4}]}}$${{\mathcal{M}_{(21^{2}0^{2})}^{\mathrm{II}}}}$${{\mathcal{M}_{(1^{4}0)}}}$${{[2^{4}1^{2}]}}$${{\mathcal{M}_{(20^{4})}^{\mathrm{I}}}}$${{\mathcal{M}_{(20^{4})}^{\mathrm{II}}=\emptyset}}$${{[2^{2}1^{6}]}}$${{\mathcal{M}_{(1^{2}0^{3})}}}$${{0}}$${{0}}$$\scriptstyle{\sim}$$\scriptstyle{\sim}$
Figure 1. $\mathcal{M}_{\bar{\lambda}}$ for small dominant weight
$\bar{\lambda}$ and their image under the map $\pi$ in type
$(X_{N},r)=(A_{9},2)$
In [AH, Theorem 1.2], they proved that there are finitely many $G$-orbits in
$\mathrm{Gr}_{\lambda}\cap\mathrm{Gr}_{0}^{-}$ for small dominant coweight
$\lambda$. In the case of $(A_{2\ell},2)$, it is easy to see that $K$ acts
transitively on $\mathcal{M}_{(1^{j}0^{\ell-j})}$ and hence there are finitely
many $K$-orbits in $\mathcal{M}$. For the case $(A_{2\ell-1},2)$, it is not
obvious to determine if there are finitely many $K$-orbits in
$\mathcal{M}_{(21^{j}0^{\ell-j-1})}$. If
$g(t)=1+xt^{-1}+(z+\frac{1}{2}x^{2})t^{-2}\in\mathcal{M}_{(21^{j}0^{\ell-j-1})}$,
then $g(t)$ satisfies (15). If the action of $K$ on the following anti-
commuting nilpotent variety
$\\{(x,z)\in\mathcal{N}_{\mathfrak{p}}\times\mathcal{N}_{\mathfrak{k}}\mid
xz+zx=0\\}$
by diagonal cojugation has finitely many orbits, then there are finitely many
$K$-orbits in $\mathcal{M}_{(21^{j}0^{\ell-j-1})}$.
###### Example 4.12.
Consider the case $(X_{N},r)=(A_{9},2)$ . In this case, $G=\mathrm{SL}_{10}$
and $\mathfrak{g}=\mathfrak{sl}_{10}=\mathfrak{k}\oplus\mathfrak{p}$. The
diagram as shown in Figure 1 describes the image of
$\mathcal{M}_{\bar{\lambda}}$ for each small dominant weight $\bar{\lambda}$.
For instance, $\mathcal{M}_{(21^{2}0^{2})}$ consists of two parts,
$\mathcal{M}_{(21^{2}0^{2})}^{\mathrm{I}}$ and
$\mathcal{M}_{(21^{2}0^{2})}^{\mathrm{II}}$. By Theorem 4.10,
$\pi(\mathcal{M}_{(21^{2}0^{2})}^{\mathrm{I}})$ is precisely the union of two
nilpotent orbits $[2^{4}1^{2}]$ and $[2^{2}1^{6}]$ in $\mathfrak{p}$ while
$\pi(\mathcal{M}_{(21^{2}0^{2})}^{\mathrm{II}})$ is the single nilpotent orbit
$[3^{2}1^{4}]$ in $\mathfrak{p}$. Since $(21^{2}0^{2})\succeq(1^{4}0)$,
$\mathcal{M}_{(21^{2}0^{2})}\succeq\mathcal{M}_{(1^{4}0)}$. Similarly,
$\mathcal{M}_{(21^{2}0^{2})}\succeq\mathcal{M}_{(20^{4})}$. According to the
table in Theorem 4.2, the image of certain $\mathcal{M}_{\bar{\lambda}}$ is a
union of 4 nilpotent orbits. It does not happen in this case since $\ell=5$ is
not large enough. In the case of $(X_{N},r)=(A_{13},2)$,
$\pi(\mathcal{M}_{(21^{4}0^{2})})$ is a union of nilpotent orbits
$[2^{4}1^{6}],[2^{6}1^{2}],[3^{2}2^{2}1^{4}],[3^{2}1^{8}]$ in
$\mathfrak{p}\subset\mathfrak{sl}_{14}$.
### 4.3. Case $(X_{N},r)=(D_{\ell+1},2)$
In this case, it is more convenient to work with $G=\mathrm{SO}_{2\ell+2}$ and
$\sigma$ is a diagram automorphism on $G$. It is known that
$G^{\sigma}\simeq\mathrm{SO}_{2\ell+1}\times\\{\pm I\\}$. Let
$G(\mathcal{O})^{\sigma,\circ}$ denote the identity component of the group
$G(\mathcal{O})^{\sigma}$. Then, the action of ${\rm
Spin}_{2\ell+2}(\mathcal{O})^{\sigma}$ on the twisted affine Grassmanian
$\mathcal{G}r$ of ${{\rm Spin}_{2n+2}}$ factors through
$G(\mathcal{O})^{\sigma,\circ}$. Let $G(\mathcal{O}^{-})_{0}^{\sigma}$ be the
kernel of the evaluation map $G(\mathcal{O}^{-})^{\sigma}\to G^{\sigma}$. The
action of ${\rm Spin}_{2\ell+2}(\mathcal{O})_{0}^{-}$ on $\mathcal{G}r$
factors through $G(\mathcal{O}^{-})_{0}^{\sigma}$. Hence, the opposite open
Schubert cell $\mathcal{G}r_{0}^{-}$ is a
$G(\mathcal{O}^{-})_{0}^{\sigma}$-orbit. In fact, $\mathcal{G}r$ is naturally
the neutral component of the twisted affine Grassmannian associated to
$(G,\sigma)$, whose definition is a bit more sophisticated.
We can realize the group $G$ as $\\{g\in{\rm SL}_{2\ell+2}|gJg^{T}=J\\}$, and
the Lie algebra of $G$ as
$\mathfrak{g}=\mathfrak{so}_{2\ell+2}(J)=\\{x\in\mathfrak{gl}_{2n}\mid
Jx+x^{T}J=0\\}$ where
$J=\begin{pmatrix}&&&&&1\\\ &&&&1&\\\ &&&1&&\\\ &&\reflectbox{$\ddots$}&&&&\\\
&1&&&&\\\ 1&&&&&\end{pmatrix}.$
The diagram automorphism $\sigma$ of order 2 on $\mathfrak{g}$ can be given by
$\sigma(x)=wxw$ where
$w=\operatorname{diag}\left(I_{\ell},\begin{pmatrix}0&1\\\
1&0\end{pmatrix},I_{\ell}\right).$
The diagram automorphism $\sigma$ on $G$ is also defined the same. We also
have the decomposition $\mathfrak{g}=\mathfrak{k}\oplus\mathfrak{p}$. Let $K$
be the identity component of $G^{\sigma}$. $K$ has Lie algebra $\mathfrak{k}$
and acts on $\mathfrak{p}$ by conjugation. It can be checked that $J=A^{T}A$
where
$A=\begin{pmatrix}\frac{1}{\sqrt{2}}&0&\cdots&\cdots&0&\frac{1}{\sqrt{2}}\\\
0&\ddots&\cdots&\cdots&\reflectbox{$\ddots$}&0\\\
\vdots&\vdots&\frac{1}{\sqrt{2}}&\frac{1}{\sqrt{2}}&\vdots&\vdots\\\
\vdots&\vdots&\frac{i}{\sqrt{2}}&-\frac{i}{\sqrt{2}}&\vdots&\vdots\\\
0&\reflectbox{$\ddots$}&\cdots&\cdots&\ddots&0\\\
\frac{i}{\sqrt{2}}&0&\cdots&\cdots&0&-\frac{i}{\sqrt{2}}\end{pmatrix}.$
Another realization of $\mathfrak{so}_{2\ell+2}$ is
$\mathfrak{so}_{2\ell+2}(I)=\\{x\in\mathfrak{gl}_{2\ell}\mid x+x^{T}=0\\}$.
There exists an isomorphism from $\mathfrak{so}_{2\ell+2}(J)$ to
$\mathfrak{so}_{2\ell+2}(I)$ given by $x\mapsto AxA^{-1}$. Under
$\mathfrak{so}_{2\ell+2}(I)$, the diagram automorphism $\sigma_{0}$ is defined
by $\sigma_{0}(x)=w_{0}xw_{0}$ where
$w_{0}=(PA)w(PA)^{-1}=\operatorname{diag}(-1,1,1,...,1)$ and $P$ is some
matrix of change of basis.
###### Proposition 4.13.
1. (1)
If $x$ is a nonzero nilpotent element in $\mathfrak{p}$, then $x$ has the
partition $[31^{2\ell-1}]$.
2. (2)
There are exactly 2 nilpotent $K$-orbits in $\mathfrak{p}$: $\\{0\\}$ and
$\mathcal{N}_{\mathfrak{p}}\setminus\\{0\\}.$
###### Proof.
Since $w_{0}xw_{0}=-x$, $x$ has the form
$x=\begin{pmatrix}0&-u^{t}\\\ u&0\end{pmatrix}$
where $u\in\mathbb{C}^{2l+1}$ is a nonzero column vector. Then
$x^{2}=\begin{pmatrix}-u^{t}u&0\\\ 0&-uu^{t}\end{pmatrix}.$
If $x^{2}=0$, then $uu^{t}=0$ which implies $u=0$, a contradiction. Since
$\operatorname{rk}x=2$ and $x^{2}\neq 0$, $x$ has the partition $[31^{2n-1}]$.
This proves the first part.
The element of $K$ has the form
$k=\begin{pmatrix}1&0\\\ 0&g\end{pmatrix}.$
where $g\in\mathrm{SO}_{2\ell+1}$ and $k$ acts on $x\in\mathfrak{p}$ by
$k\cdot x=kxk^{-1}=\begin{pmatrix}1&-(gu)^{t}\\\ gu&0\end{pmatrix}.$
Hence the action of $K$ on $\mathfrak{p}$ is the same as the action of
$\mathrm{SO}_{2\ell+1}$ on $\mathbb{C}^{2\ell+1}$. Note that for every $k\geq
3$, $x^{k}$ has the scalar $u^{T}u$ on every nonzero entry. Since $x$ is
nilpotent, $u^{T}u=0$. The result immediately follows since the action of
$\mathrm{SO}_{2\ell+1}$ on $\\{z\in\mathbb{C}^{2\ell+1}\mid z^{T}z=0\\}$ has
two orbits. ∎
###### Theorem 4.14.
For $j=0,1,...,\ell$, we have
$\pi(\mathcal{M}_{(1^{j}0^{\ell-j})})=\begin{cases}\mathcal{N}_{\mathfrak{p}}\setminus\\{0\\}&\quad\text{
if }j\text{ is odd};\\\ \\{0\\}&\quad\text{ if }j=0;\\\
\mathcal{N}_{\mathfrak{p}}&\quad\text{ if }j\text{ is even and }j\geq
2.\end{cases}.$
Moreover,
$\mathcal{M}_{(10^{\ell-1})}=\\{(1+xt^{-1}+\frac{1}{2}x^{2}t^{-2})\cdot
e_{0}\mid x\in\mathcal{N}_{\mathfrak{p}}\setminus\\{0\\}\\}$. Consequently,
$\pi$ maps $\mathcal{M}_{(10^{\ell-1})}$ isomorphically onto
$\mathcal{N}_{\mathfrak{p}}\setminus\\{0\\}$.
###### Proof.
By Lemma 4.4, let $g(t)\cdot e_{0}=(1+xt^{-1}+yt^{-2})\cdot
e_{0}\in\mathcal{M}_{(1^{j}0^{\ell-j})}.$ We work under
$\mathfrak{so}_{2\ell+2}(I)$ and $\mathrm{SO}_{2\ell+2}(I)$. Since $g(t)$ is
fixed by $\sigma_{0}$, $w_{0}xw_{0}=-x$ and $w_{0}yw_{0}=y$. Similarly to the
proof of Proposition 4.13, $x,y$ are in the form
(16) $x=\begin{pmatrix}0&-u^{T}\\\ u&0\end{pmatrix},\hskip
14.22636pty=\begin{pmatrix}y_{0}&0\\\ 0&D\end{pmatrix}$
where $u\in\mathbb{C}^{2\ell+1}$ is a column vector, $y_{0}\in\mathbb{C}$, and
$D\in\operatorname{Mat}_{(\ell-1)\times(\ell-1)}$. Since $g^{T}g=I$, the
following equations hold:
(17) $x^{T}+x=0,\hskip 14.22636ptx^{T}x+y^{T}+y=0,\hskip
14.22636ptx^{T}y+y^{T}x=0,\hskip 14.22636pty^{T}y=0.$
Then $y_{0}=0$ and $u^{T}u=0$ which implies $x^{3}=0$. Hence
$x\in\mathcal{N}_{\mathfrak{p}}$.
Suppose that $j$ is odd and $x=0$. We have $y+y^{T}=0$ and $y^{T}y=0$. Then
$y_{0}=0$ and $D$ is a nilpotent element of $\mathfrak{so}_{2\ell+1}$ with
$D^{2}=0$. Since $\operatorname{rk}D=\operatorname{rk}y=j$, $D$ has the
partition $[2^{j}1^{2\ell-2j+2}]$. This contradicts to the classification of
nilpotent orbits of type B, [CM, Theorem 5.1.2]. Hence $x\neq 0$. Now,
consider the matrix $x_{0}\in\mathcal{N}_{\mathfrak{p}}\setminus\\{0\\}$
defined by
$x_{0}=\begin{pmatrix}0&\cdots&0&-1&-i\\\ \vdots&&&&\\\ 0&&&&\\\ 1&&&&\\\
i&&&&\end{pmatrix}.$
Let $N$ be a nilpotent element in $\mathfrak{so}_{2\ell-1}(I)$ having the
partition $[2^{j-1}1^{2\ell-j+1}].$ Such a matrix $N$ exists in view of [CM,
Theorem 5.1.2]. Then the matrices $x_{0}$ and
$y_{0}:=\operatorname{diag}(0,...,0,N,0,0)+\dfrac{1}{2}x_{0}^{2}$
satisfy the relations in (17). Hence $g(t):=1+x_{0}t^{-1}+y_{0}t^{-2}\in
G(\mathcal{O})^{\sigma}$. By (2), $g(t)\cdot
e_{0}\in\mathcal{G}r_{\bar{\lambda}}$ for some
$\bar{\lambda}=(a_{1},a_{2},...,a_{\ell})\in X_{*}(T)_{\sigma}^{+}$ with
$a_{1}\geq...\geq a_{\ell}$. Since $\operatorname{rk}y_{0}=j$, by Lemma 4.4,
$\bar{\lambda}=(1^{j}0^{\ell-j})$ and hence $g(t)\cdot
e_{0}\in\mathcal{M}_{(1^{j}0^{\ell-j})}$. Since $\pi$ is $K$-equivariant, the
first part is done.
Suppose that $j$ is even. For each
$j=0,2,4,...,2\lfloor\frac{\ell}{2}\rfloor$, consider the $\ell\times\ell$
matrix
$\begin{pmatrix}&&1&&&&&\\\ &\reflectbox{$\ddots$}&&&&&&\\\ 1&&&&&&&\\\
&&&&&&&\\\ &&&&&&&\\\ &&&&&&&-1\\\ &&&&&&\reflectbox{$\ddots$}&\\\
&&&&&-1&&\end{pmatrix}$
where there are $\frac{j}{2}$ copies of each 1 and -1. Denote $z_{j}$ the
square zero matrix of size $2\ell+2$ whose $\ell\times\ell$ submatrix on the
right top is replaced by the above matrix. Now we work under
$\mathfrak{so}_{2\ell+2}(J)$ and $\mathrm{SO}_{2\ell+2}(J)$. Since
$wz_{j}w=z_{j}$ and $\operatorname{rk}z_{j}=j$, we have $(1+z_{j}t^{-2})\cdot
e_{0}\in\mathcal{M}_{(1^{j}0^{\ell-j})}$ and then $\pi((1+z_{j}t^{-2})\cdot
e_{0})=0$. Let $x_{0}$ be the square zero matrix of size $2\ell+2$ whose
$4\times 4$ submatrix at the center is replaced by
$\begin{pmatrix}0&1&-1&0\\\ 0&0&0&1\\\ 0&0&0&-1\\\ 0&0&0&0\end{pmatrix}.$
Then $x_{0}\in\mathcal{N}_{\mathfrak{p}}\setminus\\{0\\}$, Set
$y_{0}=\frac{1}{2}x_{0}^{2}+z_{j}$. Then $\operatorname{rk}y_{0}=j$. It can be
checked that $x_{0},y_{0}$ satisfy $wx_{0}w=-x_{0},wy_{0}w=y_{0}$, and
(18) $\displaystyle x_{0}^{T}J+Jx_{0}=0,\hskip
14.22636ptx_{0}^{T}Jx_{0}+y_{0}^{T}J+Jy_{0}=0,$ $\displaystyle
x_{0}^{T}Jy_{0}+y_{0}^{T}Jx_{0}=0,\hskip 14.22636pty_{0}^{T}Jy_{0}=0.$
Hence $h(t):=1+x_{0}t^{-1}+y_{0}t^{-2}\in G(\mathcal{O})^{\sigma}$. Similarly,
one can show that $h(t)\cdot e_{0}\in\mathcal{M}_{(1^{j}0^{\ell-j})}$. This
proves the second part.
To prove the last part, let $x$ be a nonzero nilpotent element in
$\mathfrak{p}$. Since $x$ has the partition $[31^{2\ell}]$,
$\operatorname{rk}x^{2}=1$. It is easy to check that
$(1+xt^{-1}+\frac{1}{2}x^{2}t^{-2})\cdot e_{0}\in\mathcal{M}_{(10^{\ell-1})}$.
Conversely, let $g(t)\cdot e_{0}=(1+xt^{-1}+yt^{-2})\cdot
e_{0}\in\mathcal{M}_{(10^{\ell-1})}$. Let
$\iota(g(t))=1+xt^{-1}+y^{\prime}t^{-2}$. Since
$g(t)=g(t)^{-T}=(\iota(g(-t)))^{T}$, $y=(y^{\prime})^{T}$. Then $y$ and
$y^{\prime}$ are adjoint each other under the symmetric form whose matrix is
$I$. Note that $\operatorname{rk}y=\operatorname{rk}y^{\prime}=1$. If
$\operatorname{Im}y\neq\operatorname{Im}y^{\prime}$, then
$\operatorname{rk}x^{2}=\operatorname{rk}y+\operatorname{rk}y^{\prime}=2$, a
contradiction. Hence $\operatorname{Im}y=\operatorname{Im}y^{\prime}$. By
Lemma 4.8, $y^{\prime}=y$ or $y^{\prime}=-y$. By (13), $x^{2}=y+y^{\prime}$
and hence $y^{\prime}=y$. By (17), $x^{T}+x=0$ and $x^{T}x+y^{T}+y=0$. Then
$y+y^{\prime}=x^{2}=y+y^{T}$, so $y=y^{\prime}=y^{T}$. Therefore,
$g(t)=1+xt^{-1}+yt^{-2}=1+xt^{-1}+\frac{1}{2}x^{2}t^{-2}.$ ∎
###### Proposition 4.15.
For $x\in\mathcal{N}_{\mathfrak{p}}$, write $x$ as in (16). Then
$\pi^{-1}(x)\cong\\{D\in\mathfrak{so}_{2\ell+1}\mid Du=0,D^{2}=0\\}.$
In particular, $\pi^{-1}(0)$ is isomorphic to the maximal order 2 nilpotent
variety in $\mathfrak{so}_{2\ell+1}$, and
$\dim\pi^{-1}(0)=\begin{cases}\ell^{2}&\quad\text{if }\ell\text{ is even;}\\\
\ell^{2}-1&\quad\text{if }\ell\text{ is odd.}\\\ \end{cases}$
###### Proof.
Under the realization $\mathfrak{so}_{2\ell+2}(I)$ and
$\mathrm{SO}_{2\ell+2}(I)$, and the diagram automorphism $\sigma_{0}$, we have
that $1+xt^{-1}+yt^{-2}\in\mathcal{M}$ if and only if $w_{0}yw_{0}=y$ and the
conditions (17) hold. Set $z=y-\frac{1}{2}x^{2}$, these conditions are
equivalent to
(19) $z=\begin{pmatrix}0&\\\ &D\end{pmatrix},\quad
D\in\mathfrak{so}_{2\ell+1},\quad D^{2}=0,\quad Du=0.$
Hence $\pi^{-1}(0)\cong\\{D\in\mathfrak{so}_{2\ell+1}\mid D^{2}=0\\}$ which is
$\overline{\mathcal{O}}_{[2^{k}1^{2\ell-2k+1}]}$ in $\mathfrak{so}_{2\ell+1}$
where $k$ is the maximal even integer. By the dimension formula, cf. [CM,
Corollary 6.1.4],
$\dim\pi^{-1}(0)=\begin{cases}\dim\mathcal{O}_{[2^{\ell}1]}=\ell^{2}&\quad\text{if
}\ell\text{ is even;}\\\
\dim\mathcal{O}_{[2^{\ell-1}1^{3}]}=\ell^{2}-1&\quad\text{if }\ell\text{ is
odd}\\\ \end{cases}$
as desired. ∎
Similar to the case $(A_{2\ell-1},2)$, it is not obvious to see if there are
finitely many $K$-orbits in $\mathcal{M}_{(1^{j}0^{\ell-j})}$. If
$g(t)=1+xt^{-1}+(z+\frac{1}{2}x^{2})t^{-2}\in\mathcal{M}_{(1^{j}0^{\ell-j})}$,
then $g(t)$ satisfies (19). If the action of $K$ on the following anti-
commuting nilpotent variety
$\\{(x,z)\in\mathfrak{so}_{2\ell+2}(I)\times\mathfrak{so}_{2\ell+2}(I)\mid
xz+zx=0,x,z\text{ nilpotent}\\}$
by diagonal cojugation has finitely many orbits, then there are finitely many
$K$-orbits in $\mathcal{M}_{(1^{j}0^{\ell-j})}$.
## 5\. Applications
In this section, we describe some applications to the geometry of order 2
nilpotent varieties in the certain classical symmetric spaces.
Let $\langle,\rangle$ be a symmetric or symplectic non-degenerate bilinear
form on a vector space $V$. Recall that $\mathcal{A}$ is the space of all
self-adjoint linear maps with respect to $\langle,\rangle$. Set
$\mathcal{N}_{\mathcal{A},2}$ denote the space of all nilpotent operators $x$
in $\mathcal{A}$ such that $x^{2}=0$. If $\langle,\rangle$ is symmetric and
$\dim V=2n+1$, then ${\rm SO}_{2n+1}$-orbits in $\mathcal{N}_{\mathcal{A},2}$
are classified by the partitions $[2^{j}1^{2n+1-2j}]$ with $0\leq j\leq n$; if
$\langle,\rangle$ is symplectic and $\dim V=2n$, then ${\rm Sp}_{2n}$-orbits
in $\mathcal{N}_{\mathcal{A},2}$ are classified by the partitions
$[2^{2j}1^{2n-2j}]$ with $0\leq j\leq\lfloor\frac{n}{2}\rfloor$.
###### Theorem 5.1.
Assume that $\langle,\rangle$ is symplectic or symmetric and $\dim V$ is odd.
Then any order 2 nilpotent variety in $\mathcal{A}$ is normal.
###### Proof.
By Theorem 4.5 and Theorem 4.6, for any order 2 nilpotent variety
$\overline{\mathcal{O}}$ in $\mathcal{A}$, $\overline{\mathcal{O}}$ is
isomorphic to
$\overline{\mathcal{M}}_{\bar{\lambda}}:=\overline{\mathcal{G}r}_{\bar{\lambda}}\cap\mathcal{G}r_{0}^{-}$
for a small dominant weight $\bar{\lambda}$ of $H$. Note that
$\overline{\mathcal{M}}_{\bar{\lambda}}$ is an open subset of the twisted
Schubert variety $\overline{\mathcal{G}r}_{\bar{\lambda}}$ and
$\overline{\mathcal{G}r}_{\bar{\lambda}}$ is a normal variety. It follows that
$\overline{\mathcal{O}}$ is also normal. ∎
In fact, when $\langle,\rangle$ is symplectic, any nilpotent variety in
$\mathcal{A}$ is normal, see [Oh]. In loc.cit., Ohta also showed that not all
nilpotent varieties are $\mathcal{N}_{\mathfrak{p}}$ is normal, when
$\langle,\rangle$ is symmetric. When $\langle,\rangle$ is symmetric and $\dim
V$ is odd, this theorem seems to be new.
###### Remark 5.2.
Theorem 5.1 is true for any field $\mathrm{k}$ of characteristic $p>2$, as one
can easily see that the classification theorem in Section 3 still holds for
order 2 nilpotent orbits, and the arguments in Theorem 4.5, Theorem 4.6
applies as well. The same remark applies to the following Theorem 5.3 and
Theorem 5.4
For any variety $X$, let ${\rm IC}_{X}$ denote the intersection cohomology
sheaf on $X$. The perverse sheaf ${\rm IC}_{X}$ captures the singularity of
the variety $X$. For any $x\in X$, we denote by $\mathscr{H}^{k}_{x}({\rm
IC}_{X})$ the k-th cohomology of the stalk of ${\rm IC}_{X}$ at $x$.
###### Theorem 5.3.
1. (1)
When $\langle,\rangle$ is symmetric and $\dim V=2n+1$, for any $0\leq j\leq
n$, let $\mathcal{O}_{j}$ denote the nilpotent orbit in $\mathcal{A}$
associated to the partition $[2^{j}1^{2n+1-2j}]$ and let
$\mathcal{O}^{\prime}_{j}$ denote the nilpotent orbit in $\mathfrak{sp}_{2n}$
associated to the partition $[2^{j}1^{2n-2j}]$, we have
$\dim\mathcal{O}_{j}=\dim\mathcal{O}^{\prime}_{j}=j(2n+1-j).$
Moreover, for any $x\in\mathcal{O}_{[2^{i}1^{2n+1-2i}]}$ and
$x^{\prime}\in\mathcal{O}^{\prime}_{[2^{i}1^{2n-2i}]}$, and for any
$k\in\mathbb{Z}$,
$\dim\mathscr{H}_{x}^{k}({\rm
IC}_{\overline{\mathcal{O}}_{j}})=\dim\mathscr{H}_{x}^{k}({\rm
IC}_{\overline{\mathcal{O}^{\prime}}_{j}}).$
2. (2)
When $\langle,\rangle$ is symplectic and $\dim V=2n$, for any $0\leq
j\leq\lfloor\frac{n}{2}\rfloor$, let $\mathcal{O}_{2j}$ denote the nilpotent
orbit in $\mathcal{A}$ associated to the partition $[2^{2j}1^{2n-4j}]$ and let
$\mathcal{O}^{\prime}_{2j}$ denote the nilpotent orbit in
$\mathfrak{so}_{2n+1}$ associated to the partition $[2^{2j}1^{2n+1-4j}]$, we
have
$\dim\mathcal{O}_{2j}=\dim\mathcal{O}^{\prime}_{2j}=4j(n-j),$
Moreover, for any integer $0\leq i\leq j$, $x\in\mathcal{O}_{2i}$,
$x^{\prime}\in\mathcal{O}^{\prime}_{2i}$, and for any $k\in\mathbb{Z}$, we
have
$\dim\mathscr{H}_{x}^{k}({\rm
IC}_{\overline{\mathcal{O}}_{2j}})=\dim\mathscr{H}_{x}^{k}({\rm
IC}_{\overline{\mathcal{O}^{\prime}}_{2j}}).$
###### Proof.
We first prove part 1). By Theorem 3.7 and [CM, Corollary 6.1.4], it is easy
to verify $\dim\mathcal{O}_{j}=\dim\mathcal{O}^{\prime}_{j}=j(2n+1-j)$. By
Theorem 4.5, $\overline{\mathcal{O}}_{j}$ can be embedded into an open subset
in the twisted affine Schubert variety $\overline{\mathcal{G}r}_{\omega_{j}}$
associated to $(\mathrm{SL}_{2n+1},\sigma)$. On the other hand, in view of
[AH], $\overline{\mathcal{O}^{\prime}}_{j}$ can be embedded into the untwisted
affine Schubert variety
$\overline{\operatorname{Gr}}^{\omega_{j}}_{\mathrm{Sp}_{2n}}$ in the affine
Grassmannian $\overline{\operatorname{Gr}}_{\mathrm{Sp}_{2n}}$ of
$\mathrm{Sp}_{2n}$. Set
$\mathcal{F}={\rm
IC}_{\overline{\mathcal{O}}_{j}}[-\dim\overline{\mathcal{O}}_{j}],\quad\text{
and }\mathcal{F}^{\prime}={\rm
IC}_{\overline{\mathcal{O}^{\prime}}_{j}}[-\dim\overline{\mathcal{O}^{\prime}}_{j}].$
By purity vanishing property of intersection cohomology sheaf of Schubert
varieties (cf. [KL]),
$\mathscr{H}_{x}^{k}(\mathcal{F})=\mathscr{H}_{x^{\prime}}^{k}(\mathcal{F}^{\prime})=0$
when $k$ is odd. Equivalently,
$\mathscr{H}_{x}^{k}({\rm
IC}_{\overline{\mathcal{O}}_{j}})=\mathscr{H}_{x}^{k}({\rm
IC}_{\overline{\mathcal{O}^{\prime}}_{j}})=0$
for any odd integer $k$, as
$\dim\overline{\mathcal{O}}_{j}=\dim\overline{\mathcal{O}^{\prime}}_{j}$ is
even.
Note that the affine Grassmannian $\operatorname{Gr}_{\mathrm{Sp}_{2n}}$ and
the twisted affine Grassmannian $\mathcal{G}r_{{\rm SL}_{2n+1}}$ have the same
underlying affine Weyl group. Applying the results in [KL], the polynomials
$\sum\dim\mathscr{H}_{x}^{2k}(\mathcal{F})q^{k}$ and
$\sum\mathscr{H}_{x}^{2k}(\mathcal{F}^{\prime})q^{k}$ are both equal to the
same Kazhdan-Lusztig polynomial $P_{\omega_{i},\omega_{j}}(q)$ for the affine
Weyl group of $\mathfrak{so}_{2n+1}$. It follows that
$\dim\mathscr{H}_{x}^{k}({\rm
IC}_{\overline{\mathcal{O}}_{j}})=\dim\mathscr{H}_{x}^{k}({\rm
IC}_{\overline{\mathcal{O}^{\prime}}_{j}})$
for all even integer $k$. Alternatively, one can see these two polynomials are
equal, as they both coincide with the jump polynomial of the Brylinsky-Kostant
filtration on the irreducible representation $V_{\omega_{j}}$ of $H$, see
[Bry, Zh].
For the second part of the theorem, the proof is almost the same, except that
by Theorem 4.6, $\overline{\mathcal{O}}_{2j}$ can be openly embedded into the
twisted affine Schubert variety $\overline{\mathcal{G}r}_{\omega_{2j}}$
associated to $(\mathrm{SL}_{2n},\sigma)$, and
$\overline{\mathcal{O}}^{\prime}_{2j}$ can be openly embedded into the affine
Schubert variety
$\overline{\operatorname{Gr}}^{\omega_{2j}}_{\mathrm{Spin}_{2n+1}}$. ∎
Part 1) of this theorem was due to Chen-Xue-Vilonen [CVX] by different
methods. This theorem shows that there is a natural bijection between order 2
nilpotent varieties in $\mathcal{A}$ and order 2 nilpotent varieties in its
dual classical Lie algebras, such that they share similar geometry and
singularities. However, under such bijection the associated two nilpotent
varieties are not necessarily isomorphic. For example, when $\langle,\rangle$
is symmetric and $\dim V=2n+1$, $\overline{\mathcal{O}}_{[21^{2n-1}]}$ is the
minimal nilpotent variety in $\mathcal{A}$. In fact
$\overline{\mathcal{O}}_{[21^{2n-1}]}$ is smooth, since the quasi-miniscule
twisted affine Schubert variety $\overline{\mathcal{G}r}_{\omega_{1}}$ is
smooth, cf. [HR, Section 5.1]. On the other hand, the minimal nilpotent
variety $\overline{\mathcal{O}^{\prime}}_{[21^{2n-2}]}$ in
$\mathfrak{sp}_{2n}$ is not smooth, as the smooth locus of the quasi-miniscule
affine Schubert variety $\overline{\operatorname{Gr}}_{\omega_{1}}$ is the
open cell $\operatorname{Gr}_{\omega_{1}}$.
We now describe another application.
###### Theorem 5.4.
If $\langle,\rangle$ is symplectic, then the smooth locus of any order 2
nilpotent variety in $\mathcal{A}$ is the open nilpotent orbit.
###### Proof.
Let $\overline{\mathcal{O}}$ be any order 2 nilpotent variety in
$\mathcal{A}$. By Theorem 4.6, $\overline{\mathcal{O}}$ can be openly embedded
into a twisted Schubert variety $\overline{\mathcal{G}r}_{\bar{\lambda}}$ with
$\bar{\lambda}$ small, in the twisted affine Grassmannian
$\mathcal{G}r_{\mathrm{SL}_{2n}}$. Then this theorem follows from [BH, Theorem
1.2]. ∎
## References
* [AH] P. Achar and A. Henderson, Geometric Satake, Springer correspondence and small representations. Selecta Math. (N.S.) 19 (2013), no. 4, 949-986.
* [AH2] P. Achar and A. Henderson and S. Riche, Geometric Satake, Springer correspondence, and small representations II . Represent. Theory 19 (2015), 94-166.
* [BH] J. Hong and M. Besson, Smooth locus of twisted affine Schubert varieties and twisted affine Demazure modules, arXiv:2010.11357.
* [Bo] A. Borel, Linear algebraic groups. Second edition. Graduate Texts in Mathematics, 126. Springer-Verlag, New York, 1991.
* [Br] A. Broer, The sum of generalized exponents and Chevalley’s restriction theorem for modules of covariants, Indag. Math. (N.S.) 6 (1995), no. 4, 385-396.
* [Bry] R. Brylinski, Limits of weight spaces, Lusztig’s q-analogs, and fiberings of adjoint orbits. J. Amer. Math. Soc. 2 (1989), no. 3, 517-533.
* [CM] D. H. Collingwood and W. M. McGovern. Nilpotent Orbits in Semisimple Lie Algebras, Van Nostrand Reinhold, New York, 1993.
* [CVX] T. Chen, K. Vilonen and T. Xue. Springer correspondence, hyperelliptic curves, and cohomology of Fano varieties. Mathematical Research Letters, Vol. 27, No. 5 (2020), pp. 1281-1323. arXiv:1510.05986.
* [EK] A.G. Elashvili and V.G. Kac Classification of Good Gradings of Simple Lie Algebras, preprint arXiv:math-ph/0312030
* [HLR] T. Haines, J. Lourenço and T. Richarz, On the normality of Schubert varieties: remaining cases in positive characteristic, arXiv:1806.11001.
* [HR] T.J.Haines and T.Richarz, Smoothness of Schubert varieties in twisted affine Grassmannians. Duke Math. J. 169, no. 17, 3233-3260
* [HS] J. Hong and L. Shen. Tensor invariants, saturation problems and diagram automorphisms, Advances in Mathematics, Volume 285, Pages 629-657, 2015.
* [Hu1] J. E. Humphreys. Linear Algebraic Groups, volume 21 of Graduate Texts in Mathematics. Springer, New York, 1975.
* [Hu2] J. E. Humphreys. Introduction to Lie Algebras and Representation Theory, volume 9 of Graduate Texts in Mathematics. Springer, New York, 1972.
* [Ko] K. Korkeathikhun, Nullcones of Symmetric Spaces and Twisted Affine Grassmannian. Oral exam at UNC, March 4, 2020.
* [Ka] V. Kac. Infinite Dimensional Lie algebras, 3rd edition. Cambridge University Press, Cambridge, 1990.
* [Ku] S. Kumar, Kac-Moody Groups, their Flag Varieties and Representation Theory . Progress in Mathematics, 204. Birkhäuser Boston, Inc., Boston, MA, 2002.
* [KL] D. Kazhdan and G. Lusztig, Schubert varieties and Poincaré duality. Geometry of the Laplace operator (Proc. Sympos. Pure Math., Univ. Hawaii, Honolulu, Hawaii, 1979), pp. 185-203, Proc. Sympos. Pure Math., XXXVI, Amer. Math. Soc., Providence, R.I., 1980.
* [Li] Y. Li, Quiver varieties and symmetric pairs. Representation Theory, 23 (2019), 1-56.
* [Lu] G. Lusztig. Green Polynomials and Singularities, Advances in Mathematics 42:169-178, 1981.
* [MV] I. Mirković and M. Vybornov, On quiver varieties and affine Grassmannians of type A. C. R. Math. Acad. Sci. Paris 336 (2003), no. 3, 207-212.
* [PR] G.Pappas and M.Rapoport. Twisted loop groups and their flag varieties, Adv.Math.219,118-198 (2008).
* [Oh] T. Ohta, The singularities of the closures of nilpotent orbits in certain symmetric pairs, Tohoku Math. J. (2) 38 (1986), no. 3, 441-468.
* [Re] M. Reeder, Zero weight spaces and the Springer correspondence, Indag. Math. (N.S.) 9 (1998), no. 3, 431-441.
* [Ri] T. Richarz, Schubert varieties in twisted affine flag varieties and local models. J. Algebra 375 (2013), 121-147.
* [Se] J. Sekiguchi, The nilpotent subvariety of the vector space associated to a symmetric pair, Publ. Res. Inst. Math. Sci. 20 (1984), no. 1, 155-212.
* [Sh] T. Shoji, Springer correspondence for symmetric spaces. arXiv:1909.06744.
* [Zh] X. Zhu, The geometric Satake correspondence for ramified groups, Ann.Sci.Eć. Norm. Supér. (4) 48 (2015), no. 2, 409-451.
|
# The AL-Gaussian Distribution as the Descriptive Model for the Internal
Proactive Inhibition in the Standard Stop Signal Task
Mohsen Soltanifara, Michael Escobarb, Annie Dupuisc, Andre Chevrierd and
Russell Schachare CONTACT Mohsen Soltanifar. Email:
<EMAIL_ADDRESS>a,b,c Biostatistics Division, Dalla Lana
School of Public Health, University of Toronto, Toronto, M5T 3M7, ON, Canada ;
a,c,d,e Department of Psychiatry, The Hospital for Sick Children, 555
University Avenue, Toronto, M5G 1X8, ON, Canada
###### Abstract
Measurements of response inhibition components of reactive inhibition and
proactive inhibition within the stop-signal paradigm have been of particular
interest to researchers since the 1980s. While frequentist nonparametric and
Bayesian parametric methods have been proposed to precisely estimate the
entire distribution of reactive inhibition, quantified by stop signal reaction
times(SSRT), there is no method yet in the stop-signal task literature to
precisely estimate the entire distribution of proactive inhibition. We
introduce an Asymmetric Laplace Gaussian (ALG) model to describe the
distribution of proactive inhibition. The proposed method is based on two
assumptions of independent trial type(go/stop) reaction times and Ex-Gaussian
(ExG) models. Results indicated that the four parametric, ALG model uniquely
describes the proactive inhibition distribution and its key shape features;
and, its hazard function is monotonically increasing, as are its three
parametric ExG components. In conclusion, both response inhibition components
can be uniquely modeled via variations of the four parametric ALG model
described with their associated similar distributional features.
###### keywords:
Proactive Inhibition, Reaction Times, Ex-Gaussian, Asymmetric Laplace
Gaussian, Bayesian Parametric Approach, Hazard function.
## 1 Introduction
Response inhibition refers to one’s ability to stop responses or impulses that
just became inappropriate or unwanted within continually changing
environments,[1]. This process’s importance lies in one’s being in continually
changing conditions, which require new, updated courses of action,[2]. Some
instances of response inhibition in daily life include braking while driving a
vehicle into an intersection in reaction to a sudden traffic change, changing
direction during a tennis game, and resisting an extra piece of pizza at a
birthday party. Two paradigms have been proposed to study the lab setting’s
response inhibition, [3]: the stop-signal task and the Go/No-go task. In the
standard stop-signal task, as used in this study, the task consists of a two-
choice, response time task called the “go task” and the “stop task.” The go
task is the primary task in which the participants are asked to correctly
press a right or left button, in response to stimulus presentation, an “X” or
“O” on the computer screen. The stop task is the occasional, secondary task in
which (with a probability of stop signal $p_{ss}$ ) the participants are
presented with a stop signal alarm after a temporal delay; Participants are
instructed to withhold their responses the ongoing go task. Successful
response inhibition occurs when participants successfully withhold their
response to the “X” or “O” on the screen in the stop task (Figure 1).
Figure 1: The standard stop signal task with two inhibition components:
proactive inhibition, reactive inhibition [4]
Response inhibition has two distinctive temporal-dynamic components: reactive
inhibition and proactive inhibition. Both components have been utilized within
the standard stop-signal task, or its varieties, to discriminate different
clinical groups, [5, 6]. We refer to reactive inhibition as the outright
inhibition triggered by an external cause, while proactive inhibition is
restraint of actions in preparation for stopping by external conditions, [7].
Each of these inhibition components have been quantified in distinctive
methods as constant point estimate or distribution in the stop-signal task
(SST) literature[1, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22,
23].
Reactive inhibition has been quantified as Stop Signal Reaction Times (SSRT)
in the SST literature from both point estimation and distributional
perspectives. Primary point estimations of the reactive inhibition up to now
include the Crude SSRT, the Logan 1994 SSRT, [1], the Weighted SSRT, the
Mixture SSRT, [8], and the time series-based SSRT,[9]. On the other hand,
primary distributional estimations of the reactive inhibition include the
Colonius’s nonparametric method,[10], the Bayesian Parametric Approach (BPA),
[11, 12] — with two subtype methods: the Individual BPA (IBPA) and the
Hierarchical BPA(HBPA) — and the mixture method, [13]. In the mixture method,
the entire SST data is partitioned to type A SST data, trials following a go
trial, and type B SST data, trials following a stop trial; The trial type
weights are defined as $W_{A}^{c}=$proportion of type A stops within all
stops, $W_{B}^{c}=1-W_{A}^{c}.$ Now, for $W_{A}\sim
Bernoulli(W_{A}^{c}),W_{B}\sim Bernoulli(W_{B}^{c}),$ and fitted cluster type
$SSRT_{A},SSRT_{B}$ distributions by the former two methods, the mixture index
for the reactive inhibition is defined as:
$\displaystyle SSRT_{Mixture}=^{d}W_{A}\times SSRT_{A}+W_{B}\times SSRT_{B}.$
(1)
Note that for the case of parametric mixture SSRT, the cluster type components
$SSRT_{A},$ and $SSRT_{B}$ may take a variety of proposed reaction time (RT)
models, such as Ex-Gaussian (ExG), Ex-Wald, Wald, Gamma, Weibull, and
lognormal, [14, 15, 16]. However, given the ExG model’s practical advantages
to others, it is widely considered the parametric model for the reactive
inhibition,[11, 12].
Proactive inhibition has been quantified in the SST literature merely as point
estimation from different perspectives. The first such estimation is defined
through a dynamic Bayesian model,[17]. Here, using the mixture assumption for
the predictive distribution for the probability of the kth trial being a stop
task $``p(Stop),”$ the reactive inhibition is defined as the following
positive Pearson correlation called the ”Sequential Effect(SE)”:
$\displaystyle SE$ $\displaystyle=$ $\displaystyle Corr(p(Stop),GORT).$ (2)
The second estimation is based on the variation of differences of reaction
times in a go trial (GORTs) in the associated arms of the modified standard
stop signal task paradigm,[18, 19, 20, 21, 22, 23]. Generally, for two given
probabilities of stop signals in two arms of the modified SST where $0\ll
p_{ss}^{(1)}<p_{ss}^{(2)}\ll 1,$ the arm type proactive inhibition
$\Delta_{AT}GORT$ is defined by as:
$\displaystyle\Delta_{AT}GORT$ $\displaystyle=$ $\displaystyle
mean(GORT_{(p_{ss}^{(2)})})-mean(GORT_{(p_{ss}^{(1)})}).$ (3)
The last type of the point estimation of proactive inhibition is based on the
differences of GORTs in the trial type clusters of the standard SST paradigm
[8, 9](Figure.2.). Here, for the fixed stop signal probability
(e.g.$p_{ss}=0.25$) and the type A GORTA and type B GORTB, the trial type
proactive inhibition $\Delta_{TT}GORT$ is defined as:
$\displaystyle\Delta_{TT}GORT$ $\displaystyle=$ $\displaystyle
mean(GORT_{B})-mean(GORT_{A}).$ (4)
Figure 2: Trial type point estimation of proactive inhibition in the standard
stop signal task
However, little information is available in the SST literature on the entire
distribution of proactive inhibition and its key features. This subject is
significant as measures of central tendencies, such as mean or median, are
insufficient — and even unnecessary — to compare mostly skewed response
inhibition distributions,[24]. Besides, masking prominent features of the
proactive inhibitions by using central tendency measures may result in
incorrect conclusions about their make-up. As an example, two different
clinical groups may have the same mean of proactive inhibition, but the shape
of their distributions may differ: one may be more positively skewed, or more
leptokurtic, or possess a higher domain of variance. The methods mentioned
earlier of the estimation of proactive inhibition do not allow for precise
estimation and description of the appropriate models for the entire set of
proactive inhibition distributions.
This study proposes a four parametric, Asymmetric Laplace Gaussian (ALG) model
for the entire proactive inhibition, given the assumption of independent trial
type (go/stop) GORTs within the standard stop-signal task. The study outline
is as follows. First, as in [8, 9] the overall SST data for each participant
is partitioned to type A SST data and type B SST data. Second, using the
Individual Bayesian Parametric Approach (IBPA) method [11, 12], the fitted ExG
GORT mean posterior parameters are calculated for the cluster type SST data.
Then, the distribution of proactive response inhibition is introduced as the
difference of two independent fitted GORT ExG models, and it is shown that it
has a four-parametric ALG distribution. Third, the descriptive statistics,
shape statistics, and vital distributional properties of the ALG model, such
as component decomposition, shape and tail behavior, and monotonic hazard
function, are discussed. Finally, an empirical example is presented to
manifest the above-presented results. Table 1 presents the summary of
estimations methods presented in the literature, including this study given
the type of inhibition component.
Table 1: Summary of Estimation Methods of Inhibition Components
[b] Estimation Inhibition Component Reactive Proactive Constant Index SSRT
$\Delta GORT$,$SE$ Examples $SSRT_{Crude},SSRT_{Mixture},SSRT_{Weighted}$
$\Delta_{AT}GORT,\Delta_{TT}GORT$ $SSRT_{Logan1994},SSRT_{SS.Logan1994}$
Distribution Index SSRT $\Delta GORT$ Examples ExG,LN,Wald ALG Ex-Wald, Gamma
## 2 Materials & Methods
### 2.1 Data
The study data have been previously described [25, 8, 9, 13]. Data were
collected at the Ontario Science Center in Toronto, Canada, in 2009-2010.
Included were 16,099 participants aged 6 to 17. The participants’ parents
provided the required ethical consent for the SST experiment. Each participant
completed the SST task, including four blocks of 24 trials with a total of 96
trials, including random 25% stop trials (24 stops) and 75% go trials(72
goes). The SST tracking algorithm was designed so that at the end of trials,
each participant achieves a 50% probability of successful inhibition.
### 2.2 Participants
The study participants are the same as those described in [9, 13]. Included
here is a unique subsample of 44 participants with a mean age of 12.1 years,
with 96 SST trials for each, and an almost balanced number of trial type stop
trials (10-14 type B stop trials vs. 14-10 type A stop trials, respectively).
This almost balanced number of trial type stop trials yields 30-42 type B go
trials vs. 42-30 type A stop trials, respectively.
### 2.3 The SST Clusters
The study Stop Signal Task clusters have been described before in [8, 9, 13].
Each participant’s SST data was partitioned to type A SST data, where a go
trial preceded all trials, and type B SST data, all trials preceded by a stop
trial. Hence, each participant has three types of SST data clusters: Type-A
SST cluster (i.e., 56 trials), Type B SST cluster(i.e., 40 trials), and Type-S
SST cluster (all 96 trials). Then, using IBPA, the parameters of the
corresponding Ex-Gaussian (ExG) GORT’s parameters (i.e.
$\theta_{A}=(\mu_{A},\sigma_{A},\tau_{A}),\theta_{B}=(\mu_{B},\sigma_{B},\tau_{B}),\theta_{S}=(\mu_{S},\sigma_{S},\tau_{S})$
were computed as described in the upcoming results section 3.2.
### 2.4 Preliminaries on Component Distributions
The first two definitions provide the critical features of the components of
our upcoming calculations, namely Ex-Gaussian distribution (ExG), [26], and
the Asymmetric Laplace distribution (AL), [27].
###### Definition 2.1.
A random variable has an Ex-Gaussian (ExG) distribution with parameters
$(\mu,\sigma,\tau)$ whenever it is considered as the sum of an independent
normal random variable with parameters $(\mu,\sigma^{2})$ and an exponential
random variable with parameter $\tau$:
$\displaystyle ExG(\mu,\sigma,\tau)$ $\displaystyle=^{d}$ $\displaystyle
N(\mu,\sigma^{2})\oplus Exp(\tau).$ (5)
The density, the moment generating function, the nth cumulant $(n\geq 1),$ the
variance, the skewness and the kurtosis of the ExG distribution are given by:
$\displaystyle PDF$ $\displaystyle
f_{ExG}(t|\mu,\sigma,\tau)=\frac{1}{\tau}exp(\frac{\mu-t}{\tau}+\frac{\sigma^{2}}{2\tau^{2}})*\Phi(\frac{\mu-t}{\sigma}-\frac{\sigma}{\tau}):\sigma,\tau>0,t\in\mathbb{R},$
$\displaystyle MGF$ $\displaystyle
m_{ExG}(t)=(1-t\tau)^{-1}exp(\mu.t+\frac{\sigma^{2}}{2}t^{2}):t<\tau^{-1},$
$\displaystyle n^{th}Cumulant$
$\displaystyle\kappa_{n}^{ExG}=(n-1)!\tau^{n}+1_{n=1}(n)\mu+1_{n=2}(n)\sigma^{2}:1\leq
n,$ $\displaystyle Mean$ $\displaystyle E(ExG)=\mu+\tau,$ $\displaystyle
Variance$ $\displaystyle Var(ExG)=\sigma^{2}+\tau^{2},$ $\displaystyle
Skewness$ $\displaystyle\gamma_{ExG}=2(1+\sigma^{2}\tau^{-2})^{-3/2},$
$\displaystyle Kurtosis$
$\displaystyle\kappa_{ExG}=3\frac{(1+2\sigma^{-2}\tau^{2}+3\sigma^{-4}\tau^{4})}{(1+\sigma^{-2}\tau^{2})^{2}}.$
(6)
###### Definition 2.2.
A random variable has an Asymmetric Laplace (AL) distribution with parameters
$(\alpha_{1},\alpha_{2})$ whenever it is considered as the difference of two
independent exponential random variables with parameters $\alpha_{2}$, and
$\alpha_{1}$, respectively:
$\displaystyle AL(\alpha_{1},\alpha_{2})$ $\displaystyle=^{d}$ $\displaystyle
Exp(\alpha_{2})\ominus Exp(\alpha_{1}).$ (7)
The density, the moment generating function, the nth cumulant $(n\geq 1),$ the
variance, the skewness and the kurtosis of the AL distribution are given by:
$\displaystyle PDF$ $\displaystyle
f_{AL}(t|\alpha_{1},\alpha_{2})=\frac{exp(\frac{t}{\alpha_{1}})1_{(-\infty,0)}(t)+exp(\frac{-t}{\alpha_{2}})1_{[0,\infty)}(t)}{\alpha_{1}+\alpha_{2}}\
t\in\mathbb{R},$ $\displaystyle MGF$ $\displaystyle
m_{AL}(t)=(1+(\alpha_{1}-\alpha_{2})t-\alpha_{1}\alpha_{2}.t^{2})^{-1}\
-\alpha_{1}^{-1}<t<\alpha_{2}^{-1},$ $\displaystyle n^{th}Cumulant$
$\displaystyle\kappa_{n}^{AL}=(n-1)!((-\alpha_{1})^{n}+(\alpha_{2})^{n})\
1\leq n,$ $\displaystyle Mean$ $\displaystyle E(AL)=-(\alpha_{1}-\alpha_{2}),$
$\displaystyle Variance$ $\displaystyle
Var(AL)=\alpha_{1}^{2}+\alpha_{2}^{2},$ $\displaystyle Skewness$
$\displaystyle\gamma_{AL}=-2(\alpha_{1}^{3}-\alpha_{2}^{3})\times(\alpha_{1}^{2}+\alpha_{2}^{2})^{-3/2},$
$\displaystyle Kurtosis$
$\displaystyle\kappa_{AL}=3(3\alpha_{1}^{4}+2\alpha_{1}^{2}\alpha_{2}^{2}+3\alpha_{2}^{4})\times(\alpha_{1}^{2}+\alpha_{2}^{2})^{-2}.$
(8)
The convolution of two independent, AL random variables and Gaussian random
variables, called ALG or Normal-Laplace (NL) random variables, has been of
special attention in the literature [28, 29].
###### Definition 2.3.
A random variable has Asymmetric Laplace-Gaussian (ALG) distribution with
parameters $(\alpha_{1},\alpha_{2},\mu,\sigma)$ whenever it is considered as
the sum of two independent, Asymmetric Laplace random variables with
parameters $(\alpha_{1},\alpha_{2})$, and a Normal random variable with
parameters $(\mu,\sigma^{2})$, respectively:
$\displaystyle ALG(\alpha_{1},\alpha_{2},\mu,\sigma)$ $\displaystyle=^{d}$
$\displaystyle AL(\alpha_{1},\alpha_{2})\oplus N(\mu,\sigma^{2}).$ (9)
Note that since $AL(0^{+},\alpha_{2})=^{d}Exp(\alpha_{2}),$ it follows that
$ALG(0^{+},\alpha_{2},\mu,\sigma)=^{d}ExG(\mu,\sigma,\alpha_{2})$.
Consequently, the ExG model can be considered a special degenerate ALG model.
Next, the following key Theorems equip us to propose the ALG distribution as
the model for the proactive inhibition and compute the key descriptive and
shape statistics of the ALG distribution in terms of its Laplacian and
Gaussian components,[30].
###### Theorem 2.4.
Let $X,Y$ be two independent, real-valued random variables with finite, moment
generating function $m_{X},m_{Y},$ and cumulant functions $\kappa,\kappa,$
respectively. Then, for some $s_{0}>0:$
$\displaystyle m_{X+Y}(t)=m_{X}(t)m_{Y}(t):\ \ (-s_{0}<t<s_{0}),$ (10)
$\displaystyle\kappa_{X+Y}(t)=\kappa_{X}(t)+\kappa(t):\ \ \ (-s_{0}<t<s_{0}).$
(11)
###### Theorem 2.5.
Let $X,Y$ be two real-valued random variables with finite moment generating
functions $m_{X},m_{Y},$ respectively. Assume for some $s_{0}>0:\
m_{X}(t)=m_{Y}(t)\ (-s_{0}<t<s_{0}).$ Then, $X,Y$ have the same distribution.
Finally, the last two theorems enable us to describe the behavior of the
hazard function of the ALG model for the proactive inhibition, [31].
###### Theorem 2.6.
Let $X$ be a real-valued random variable with differentiable PDF $f_{X}$ and
CDF $F_{X}$ such that $f_{X}(t)\rightarrow 0,F_{X}(t)\rightarrow 1\ as\
t\rightarrow\infty,$ and $-ln(f_{X}(t))$ is convex(concave). Then, the hazard
function $h_{X}$ is increasing(decreasing).
###### Theorem 2.7.
Let $X,Y$ be two independent, real-valued random variables with (strictly)
increasing hazard functions $h_{X},h_{Y},$ respectively. Then, the hazard
function of their sum, $h_{X+Y}$ is (strictly) increasing as well.
### 2.5 Proactive Inhibition Index
Proactive inhibition was operationalized based on the standard stop signal
task’s internal perspective,[9]. Here, for a given fixed stop signal
probability (e.g., 0.25), type A GORT of $GORT_{A}$ (GORT for a trial after a
go trial), and type B GORT of $GORT_{B}$ (GORT for a trial after a stop
trial), the internal proactive inhibition is defined as:
$\displaystyle\Delta GORT$ $\displaystyle=^{d}$ $\displaystyle
GORT_{B}-GORT_{A}.$ (12)
Note that there are two mathematical perspectives for the proactive
inhibition: First, a model with two ExG components; Second, a model with
Asymmetric Laplace (AL) and Gaussian components. Henceforward, it is
understood within the given context which perspective is being discussed.
### 2.6 Statistical Analysis
The statistical ALG model to describe the proactive inhibition distribution
was presented using moment generating functions, [29]. Next, the ALG model’s
descriptive and shape statistics were computed in terms of parameters of the
cluster type ExG components,[30]. Finally, its hazard function behavior was
theoretically inferred using its components’ parameters [31].
The ExG components of the presented statistical model were estimated using the
IBPA method,[11, 12]. As in [13], each participant had three IBPA associated
ExG parametric estimations
$\theta_{A}=(\mu_{A},\sigma_{A},\tau_{A}),\theta_{B}=(\mu_{B},\sigma_{B},\tau_{B}),$
and $\theta_{S}=(\mu_{S},\sigma_{S},\tau_{S}),$ associated to type-A cluster
SST data, type-B cluster SST data, and the entire SST data, respectively.
These parameters were estimated as the posterior means of the associated
following IBPA procedure with three chains, 5,000 burn-ins within 20,000
simulations in Bayesian Ex-Gaussian Estimation of Stop-Signal RT distributions
(BEESTS) 2.0 software,[12]:
Data | Individual Priors
---|---
$GORT\sim ExG(\mu_{go},\sigma_{go},\tau_{go})$ |
$SRRT\sim ExG(\mu_{go},\sigma_{go},\tau_{go},\mu_{stop},\sigma_{stop},\tau_{stop},SSD)I^{+}_{[1,1000]}$ | $\mu_{go},\sigma_{go},\tau_{go}\sim U[10,2000]$
$SSRT\sim ExG(\mu_{go},\sigma_{go},\tau_{go},\mu_{stop},\sigma_{stop},\tau_{stop},SSD)I^{+}_{[1,1000]}$ | $\mu_{go},\sigma_{go},\tau_{go}\sim U[10,2000]$
Two sets of comparisons were conducted using paired t-tests (DescTools, $R$
software version 4.0.0, [32]): First, a primary comparison between the
cluster-type fitted parameter of the ExG distribution, the descriptive
statistics, and the shape statistics; second, secondary comparisons between
the ALG model descriptive and shape statistics and its associated cluster-type
ExG components.
## 3 Results
The results are divided into two subsections. In subsection 3.1, we explore
the mathematical analysis of the proposed model for the proactive inhibition
in the standard stop-signal task. This model includes a four parametric ALG
for the proactive inhibition and its prominent distributional properties. In
subsection 3.2, we present an empirical example of the case and discuss its
various distributional features.
### 3.1 Mathematical Analysis
#### 3.1.1 The Proactive Inhibition Distribution and its Parameters
First of all, we propose a mathematical model for the proactive inhibition
provided by the ALG:
###### Theorem 3.1.
(The Main Result). The four parametric
$ALG(\tau_{A},\tau_{B},\mu_{B}-\mu_{A},(\sigma_{B}^{2}+\sigma_{A}^{2})^{1/2})$
presents a model for the Internal Proactive Inhibition Index $\Delta GORT$
with trial type-related parameters
$(\mu_{A},\sigma_{A},\tau_{A},\mu_{B},\sigma_{B},\tau_{B})$ in the Standard
Stop Signal Task.
As a corollary of Theorem 3.1, the probability density function $(f_{\Delta
GORT})$ and the cumulative density function $(F_{\Delta GORT})$ of the ALG
distribution for the Internal Proactive Inhibition Index are given by [28,
29],:
$\displaystyle f_{\Delta GORT}(t)$ $\displaystyle=$
$\displaystyle\frac{1}{\tau_{A}+\tau_{B}}$ (13) $\displaystyle\times[$
$\displaystyle
e^{(\frac{\sqrt{\sigma_{B}^{2}+\sigma_{A}^{2}}}{2\tau_{B}}(\frac{\sqrt{\sigma_{B}^{2}+\sigma_{A}^{2}}}{\tau_{B}}-2\frac{t-(\mu_{B}-\mu_{A})}{\sqrt{\sigma_{B}^{2}+\sigma_{A}^{2}}}))}\times(1-\Phi(\frac{\sqrt{\sigma_{B}^{2}+\sigma_{A}^{2}}}{\tau_{B}}-\frac{t-(\mu_{B}-\mu_{A})}{\sqrt{\sigma_{B}^{2}+\sigma_{A}^{2}}}))$
$\displaystyle+$ $\displaystyle
e^{(\frac{\sqrt{\sigma_{B}^{2}+\sigma_{A}^{2}}}{2\tau_{A}}(\frac{\sqrt{\sigma_{B}^{2}+\sigma_{A}^{2}}}{\tau_{A}}+2\frac{t-(\mu_{B}-\mu_{A})}{\sqrt{\sigma_{B}^{2}+\sigma_{A}^{2}}}))}\times(1-\Phi(\frac{\sqrt{\sigma_{B}^{2}+\sigma_{A}^{2}}}{\tau_{A}}+\frac{t-(\mu_{B}-\mu_{A})}{\sqrt{\sigma_{B}^{2}+\sigma_{A}^{2}}}))]$
$\displaystyle\hskip 284.52756ptt\in\mathbb{R}$ and $\displaystyle F_{\Delta
GORT}(t)$ $\displaystyle=$
$\displaystyle\frac{1}{\tau_{A}^{-1}+\tau_{B}^{-1}}\times[(\tau_{A}^{-1}+\tau_{B}^{-1})\Phi(\frac{t-(\mu_{B}-\mu_{A})}{\sqrt{\sigma_{B}^{2}+\sigma_{A}^{2}}})$
(14) $\displaystyle-$
$\displaystyle\tau_{A}^{-1}e^{(\frac{\sqrt{\sigma_{B}^{2}+\sigma_{A}^{2}}}{2\tau_{B}}(\frac{\sqrt{\sigma_{B}^{2}+\sigma_{A}^{2}}}{\tau_{B}}-2\frac{t-(\mu_{B}-\mu_{A})}{\sqrt{\sigma_{B}^{2}+\sigma_{A}^{2}}}))}\times(1-\Phi(\frac{\sqrt{\sigma_{B}^{2}+\sigma_{A}^{2}}}{\tau_{B}}-\frac{t-(\mu_{B}-\mu_{A})}{\sqrt{\sigma_{B}^{2}+\sigma_{A}^{2}}}))$
$\displaystyle+$
$\displaystyle\tau_{B}^{-1}e^{(\frac{\sqrt{\sigma_{B}^{2}+\sigma_{A}^{2}}}{2\tau_{A}}(\frac{\sqrt{\sigma_{B}^{2}+\sigma_{A}^{2}}}{\tau_{A}}+2\frac{t-(\mu_{B}-\mu_{A})}{\sqrt{\sigma_{B}^{2}+\sigma_{A}^{2}}}))}\times(1-\Phi(\frac{\sqrt{\sigma_{B}^{2}+\sigma_{A}^{2}}}{\tau_{A}}+\frac{t-(\mu_{B}-\mu_{A})}{\sqrt{\sigma_{B}^{2}+\sigma_{A}^{2}}}))]$
$\displaystyle\hskip 284.52756ptt\in\mathbb{R},$
respectively. Here $\Phi$ denotes the standard normal cumulative distribution
function.
Next, given trial type parameters, we estimate the descriptive and shape
statistics for the proposed ALG model of the proactive inhibition:
###### Theorem 3.2.
The descriptive statistics and the shape statistics of the Proactive
Inhibition ALG distribution with trial type-related parameters
$(\mu_{A},\sigma_{A},\tau_{A},\mu_{B},\sigma_{B},\tau_{B})$ in the standard
stop signal task are given by:
$\displaystyle n^{th}Cumulant$
$\displaystyle\kappa_{n}^{ALG}=(n-1)!((-\tau_{A})^{n}+\tau_{B}^{n})$
$\displaystyle\hskip
28.45274pt+1_{(n=1)}(n)(\mu_{B}-\mu_{A})+1_{(n=2)}(n)(\sigma_{B}^{2}+\sigma_{A}^{2}):\
\ 1\leq n$ $\displaystyle Mean$ $\displaystyle
E(ALG)=\tau_{B}-\tau_{A}+\mu_{B}-\mu_{A},$ $\displaystyle Variance$
$\displaystyle
Var(ALG)=\tau_{A}^{2}+\tau_{B}^{2}+\sigma_{A}^{2}+\sigma_{B}^{2},$
$\displaystyle Skewness$
$\displaystyle\gamma_{ALG}=\frac{2(\tau_{B}^{3}-\tau_{A}^{3})}{(\tau_{A}^{2}+\tau_{B}^{2}+\sigma_{A}^{2}+\sigma_{B}^{2})^{3/2}},$
$\displaystyle Kurtosis$
$\displaystyle\kappa_{ALG}=\frac{6(\tau_{B}^{4}+\tau_{A}^{4})}{(\tau_{A}^{2}+\tau_{B}^{2}+\sigma_{A}^{2}+\sigma_{B}^{2})^{2}}.$
(15)
#### 3.1.2 The Proactive Inhibition’s Key ALG Distributional Properties
In this section, we present key distributional properties for the ALG model
for proactive inhibition including: (i) component decompositions in terms of
trial type GORT; (ii) shape and tail behavior; and, (iii) behavior of the
hazard function.
###### Theorem 3.3.
(Component Decomposition). An ALG model for proactive inhibition emerges from
uncountable pairs of trial type related GORT $(GORT_{A},GORT_{B})$
distributions.
We remind the reader that Theorem3.3 presents a process to simulate a
plausible four parametric ALG distribution for the proactive inhibition.
###### Theorem 3.4.
(Shape and Tail Behavior). An ALG model for proactive inhibition has unimodal,
generally asymmetric, infinite, differentiable density with extreme large
values proportionate to the $Exp(1/\tau_{B})$ distribution.
We remind the reader that contrary to the ALG model’s mean for proactive
inhibition, there are no closed-form formulas for the mode and the median,
respectively. Similar to the ExG model for reactive inhibition with increasing
hazard function, we have:
###### Theorem 3.5.
(Hazard Function’s Behavior). An ALG model for proactive inhibition has
increasing hazard function.
### 3.2 The Empirical Example
#### 3.2.1 The ALG Model for Proactive Inhibition
This section presents an example of the empirical data for the theoretical
results inferred in the previous section on the ALG model for proactive
inhibition and its descriptive, shape, and hazard function’s key features.
These results are based on the cluster type IBPA estimation of mean posterior
ExG parameters $\theta=(\mu,\sigma,\tau)$ presented in Table 5(Appendix B).
The ExG model for each of the two components of proactive inhibition has the
following key features presented by Table 2. First, while type B $\mu,$, and
$\sigma$ parameters are significantly larger than their type A counterparts,
there is no difference for the $\tau$ parameter. In addition, the sample
average proactive inhibition is 92.1 ms (95% CI = (69.4,114.9)). Second, both
ExG components are positively skewed and leptokurtic. Finally, there is no
significant difference between their trial-type skewness and their trial-type
kurtosis, respectively. Figure3a presents the trial type ExG modeled
components of the ALG model.
Table 2: Descriptive and paired t-test [mean ($95\%CI$)] results for
parameters, descriptive and shape statistics of fitted Ex-Gaussian
distribution to cluster type GORT and AL-Gaussian distribution to $\Delta
GORT,(n=44).$
[b] ExG model ALG model Cluster Comparison Cluster Type A Type B Type B vs.
Type A Type S $\alpha_{1}$ - - - 104.2 - - - (90.4,117.9) $\alpha_{2}$ - - -
142.4 - - - (125.9,158.8) Parameter $\mu$ 478.8 532.8 53.9*** 53.9
(448.0,509.7) (498.6,566.9) (30.9,76.9) (30.9,76.9) $\sigma$ 109.9 133.1 23.2
179.2 (90.5,129.3) (108.4,157.8) (-0.1,46.4) (151.4,206.9) $\tau$ 104.2 142.4
38.2*** - (90.4,117.9) (125.9,158.8) (19.6,56.8) - Mean 583.0 675.1 92.1***
92.1 (553.0,612.9) (633.8,716.4) (69.4,114.9) (69.4,114.9) Statistics St.D
160.6 202.4 41.8*** 260.4 (143.5,177.8) (177.9,226.9) (25.9,57.6)
(232.3,288.6) Skewness 0.787 0.918 0.131 0.186 (0.602,0.973) (0.751,1.085)
(-0.113,0.375) (0.076,0.296) Kurtosis 4.966 5.300 0.334 1.153 (4.414,5.518)
(4.790,5.808) (-0.397,1.064) (0.923,1.384)
* •
Notes: ∗p-value$<0.05$;∗∗p-value$<0.005$;∗∗∗p-value$<0.0005$.
The ALG model for proactive inhibition has the following features, given
cluster type parameter estimations. First, as a primary corollary of Theorem
3.2, the model is positively skewed whenever $\tau_{B}>\tau_{A}.$ The
negatively skewed and symmetric cases hold whenever the strict greater
inequality $>$ is replaced with $<$ and $=,$ respectively. According to
appendix Table 5 data, all three cases exist (case 10: positive skew; case 16:
symmetric; case 11: negative skew). Figure 3b presents all the mentioned three
cases. Overall, the model is positively skewed given the results in Table 2.
Second, as the second corollary of Theorem 3.2, the model is leptokurtic
whenever
$(2(\tau_{A}^{4}+\tau_{B}^{4}))^{1/2}-(\tau_{A}^{2}+\tau_{B}^{2})>\sigma_{A}^{2}+\sigma_{B}^{2}.$
The platykurtic and mesokurtic cases hold whenever the strict greater
inequality $>$ is replaced with $<$ and $=,$ respectively. In particular, for
the case, $\tau_{A}\approx\tau_{B},$ the model is always platykurtic. Overall,
the model is platykurtic given results in Table 2. Third, while the ALG
model’s standard deviation is larger than its two ExG’s components, its
skewness and kurtosis are significantly smaller. Finally, the ALG model has a
strictly increasing hazard function for various skewness cases, as mentioned
in Theorem3.5 and presented in Figure 3c.
Figure 3: The ALG density and its trial type ExG component densities; (b) The
ALG density for the positively skewed, symmetric and negatively skewed cases;
(c) The ALG hazard function for the positively skewed, symmetric and
negatively skewed cases.
#### 3.2.2 Proactive Inhibition ALG Model versus Reactive Inhibition ExG
Model
This section compares the ALG distribution of the proactive inhibition and the
ExG distribution of the reactive inhibition in descriptive and shapes
statistics. Table 3 presents the corresponding statistics for both models. As
it is seen, the proactive ALG inhibition distribution has a significantly
lower mean, lower skewness, lower kurtosis, and higher standard deviation than
reactive inhibition ExG distribution. Also, while the proactive inhibition ALG
distribution is platykurtic, the corresponding reactive ExG distribution is
leptokurtic. However, the two distributions are both positively skewed.
Overall, the two distributions are significantly distinctive.
Table 3: Comparison of proactive inhibition ALG model versus reactive
inhibition ExG model in terms of descriptive and shape statistics $(n=44).$
[b] Inhibition Reactive Proactive Proactive vs. Reactive Index $SSRT$ $\Delta
GORT$ $\Delta GORT\ vs.\ SSRT$ Model ExG ALG ALG vs. ExG Mean 196.8 92.1
-104.6*** (173.5,220.1) (69.4,114.9) (-140.6,-68.7) Statistics St.D 157.8
260.4 102.6*** (139.4,176.2) (232.3,288.6) (71.8,133.6) Skewness 0.578 0.186
-0.401*** (0.500,0.674) (0.076,0.296) (-0.540,-0.261) Kurtosis 4.231 1.153
-3.077*** (3.998,4.465) (0.923,1.384) (-3.381,-2.775)
* •
Notes: ∗p-value$<0.05$;∗∗p-value$<0.005$;∗∗∗p-value$<0.0005$.
## 4 Discussion
### 4.1 Present Work
This paper presents a four parametric model for the entire proactive
inhibition distribution, the ALG model. This model is based on the two
independent ExG components fitted to the trial type GORTs. Considering ExG
models as a degenerate ALG model, this work indicates that a four parametric
ALG model can model both response inhibition components.
The proposed ALG model for proactive inhibition has several important aspects.
First, the model is based on the independent assumption of GORTA and GORTB.
Such speculation is warranted by the independent assignment of go trials and
stop trials in the entire stop-signal task trials. Hence, via conditioning,
their following associated type A go trial and type B go trials are
independent, and so are their associated GORTs (i.e., GORTA and GORTB).
Second, contrary to the point estimations of proactive inhibition in the form
of mean [9], or correlation [17], it presents the entire distribution of the
proactive inhibition. Third, the ALG model for proactive inhibition is
entirely distinctive from the ExG model for reactive inhibition in terms of
the mean, standard deviation, and kurtosis. This result provides more evidence
on the distinction between proactive inhibition and reactive inhibition as a
whole distribution. Fourth, similar to the ExG model for reactive inhibition,
the ALG model for proactive inhibition is skewed to the right and has a
monotonically increasing hazard function. Finally, the ALG model for proactive
inhibition with its associated ExG modeled components is unique because other
non-ExG RT models for its components (e.g., Gamma, Weibull, Lognormal, Wald,
and Ex-Wald) do not yield to any known closed-form distribution for proactive
inhibition. This limitation is easily verifiable by repeating the proof of
Theorem 3.1 based on the uniqueness of moment generating functions for other
non-ExG RT models for the trial type GORTs.
The proposed ALG model for the proactive inhibition is estimated (or
simulated) in two different methods. First, one may use the Bayesian or
frequentist-based methods to fit the ExG parametric models to its trial-type
components and then estimate the four parametric ALG model using Theorem 3.1,
as done in Section 3.2. Second, one may subtract the trial type GORTs and fit
the ALG model directly to the differenced GORT data using Maximum Likelihood
(ML) or Expectation-Maximization (EM) algorithms [33].
The proposed ALG model uniquely distinguishes the reactive inhibition and
proactive inhibition distribution in terms of vital distributional features.
Table 4 presents an overall comparison for proactive inhibition and reactive
inhibition in terms of the ALG model(with considering ExG as its particular
case):
Table 4: Comparison of proactive inhibition and reactive inhibition in terms
of ALG model properties.
[b] Inhibition Index $\\#$ Parameters $\\#$ Estimations Mean StD
Skewness($+$) Kurtosis Hazard Proactive $\Delta GORT$ 4 2 lower higher lower
platykurtic increasing Reactive $SSRT$ 3 1 higher lower higher leptokurtic
increasing
There are some limitations in the proposed ALG model for proactive inhibition.
First, since GORTA data and GORTB data are unmatched, there is no way to
calculate their correlation quantitatively. Hence, from a quantitative
perspective, checking the validity of the assumption of independent GORTA and
GORTB is difficult. Second, by its definition, proactive inhibition takes only
non-negative values while the presented ALG model takes negative values.
Third, similar to the ExG model for reactive inhibition, the ALG model for
proactive inhibition has a monotonically increasing hazard function preventing
it from being the best fitting model for the cases of proactive inhibition
with peaked hazards. Finally, given the calculations’ structure of the ALG
model parameters, based on those of ExG components, its parameters’ cognitive
interpretations are highly dependent on its ExG components inheriting their
constraints.
### 4.2 Future Work
Future research should replicate the proposed approach in modeling the
proactive inhibition distribution in this study in other different directions.
This further work may include the following perspectives: First, one model
should consider peaked hazard functions for the ALG model components to
address RT data with such features. Second, there is a need to interpret the
proposed ALG distribution parameters in terms of inhibition mechanisms in the
brain and vise versa. Third, there is a lack of investigation comparing the
proactive inhibition distribution in terms of the usual stochastic order, the
descriptive and shape statistics across a spectrum of clinical groups such as
ADHD, OCD, schizophrenia, and drug users. Finally, similar investigations on
comparing the proactive inhibition distribution and its above key statistics
are plausible in terms of the participants’ age.
### 4.3 Conclusion
In conclusion, the ALG model provides a practical description of the proactive
inhibition distribution that takes full advantage of its ExG components fitted
for the trial type GORTs. It also offers a straightforward, computational
analog of the proactive inhibition, comparable to the ExG model for reactive
inhibition. Given the advantages of estimating the entire distribution of
proactive inhibition over former point estimations, the researchers recommend
considering the ALG model as the latest optimal choice to describe the
distribution of proactive inhibition.
Author Contributions: The authors contributed to the study in the following
manner: Conceptualization, M.S.; methodology, M.S.; formal analysis, M.S.;
validation, M.E, M.S., A.D and R.S.; investigation, M.E., M.S., A.D and R.S.;
data curation, R.S.; writing—original draft preparation, M.S.; writing—review
and editing, M.S.,M.E., A.D., A.C., and R.S.; All authors have read and agreed
to the submitted version of the manuscript.
## Disclosure Statement
Mohsen Soltanifar, Michael Escobar, Annie Dupuis and Andre Chevrier have no
financial interests to disclose. Russell Schachar has equity in ehave and has
been the on the Scientific Advisory Board(SAB) in Lilly and Highland
Therapautic Inc, Toronto, Canada.
## Funding
This work has no external funding.
## ORCID
Mohsen Soltanifar: https://orcid.org/0000-0002-5989-0082
Michael Escobar: https://orcid.org/0000-0001-9055-4709
Annie Dupuis: https://orcid.org/0000-0002-8704-078X
Andre Chevrier: https://orcid.org/0000-0002-4298-9529
Russell Schachar: https://orcid.org/0000-0002-2015-4395
## Abbreviations
$\begin{array}[]{ll}\text{ADHD}&\text{Attention Deficit Hyperactivity
Disorder}\\\ \text{AL}&\text{Asymmetric Laplace distribution}\\\
\text{ALG}&\text{Asymmetric Laplace Gaussian distribution}\\\
\text{BEESTS}&\text{Bayesian Ex-Gaussian Estimation of Stop Signal RT
distributions}\\\ \text{BPA}&\text{Bayesian Parametric Approach}\\\
\text{CDF}&\text{Cumulative Density Function}\\\ \text{EM}&\text{Expectation
Maximization}\\\ \text{ExG}&\text{Ex-Gaussian distribution}\\\
\text{GORT}&\text{Reaction Time in a go trial}\\\ \text{GORTA}&\text{Reaction
Time in a type A go trial}\\\ \text{GORTB}&\text{Reaction Time in a type B go
trial}\\\ \text{HBPA}&\text{Hierarchical Bayesian Parametric Approach}\\\
\text{IBPA}&\text{Individual Bayesian Parametric Approach}\\\
\text{MGF}&\text{Moment Generating Function}\\\ \text{ML}&\text{Maximum
Likelihood}\\\ \text{NL}&\text{Normal-Laplace distribution}\\\
\text{OCD}&\text{Obsessive Compulsive Disorder}\\\
\text{PDF}&\text{Prpobability Density Function}\\\ \text{SE}&\text{Sequential
Effect}\\\ \text{SSD}&\text{Stop Signal Delay}\\\ \text{SRRT}&\text{Reaction
Times in a failed stop trial}\\\ \text{SSRT}&\text{Stop Signal Reaction Times
in a stop trial}\\\ \text{SSRTA}&\text{Stop Signal Reaction Times in a type A
stop trial}\\\ \text{SSRTB}&\text{Stop Signal Reaction Times in a type B stop
trial}\\\ \text{SST}&\text{Stop Signal Task }\\\ \text{$=^{d}$}&\text{Equality
in distribution}\\\ \text{$\oplus$}&\text{Sum of independent random
variables}\\\ \text{$\ominus$}&\text{Difference of independent random
variables}\end{array}$
## References
* [1] Matzke, D.; Verbrugen, F; $\&$ Logan G.D. (2018). The Stop Signal Paradigm. In Steven Handbook of Experimental Psychology and Cognitive Neuroscience (4th ed). Vol 5. Methodology. John Wiley $\&$ Sons. Inc.
* [2] Aron, A. R. (2011). From reactive to proactive and selective control: Developing a richer model for stopping inappropriate responses. Biological Psychiatry, 69(12). e55-e68.
* [3] Johnston S, Dinoske A, Smith J, Bang R.J, et al. (2007). The Development of Stop Signal and Go/No-go Response Inhibition in Children Aged 7-12 ‘years: Performance and Event Related Potential Indices . International Journal of Psychology, 63: 25-38.
* [4] Chevrier, A., Schachar, R.J. (2020). BOLD differences normally attributed to inhibitory control predict symptoms, not task-directed inhibitory control in ADHD. Journal of Neurodevelop Disorders 12(8) . https://doi.org/10.1186/s11689-020-09311-8
* [5] Zandbelt, B.B. ; Van Buuren M.; Kahn, R.S; $\&$ Vink, M. (2011). Reduced Proactive Inhibition in Schizophrenia is related to Corticotriatal dysfunction and poor working memory. Biological Psychiatry. 70(12), 1151-1158.
* [6] Van Rooji, S.J.H; Rademaker, A.R.; Kennis, M; Vink, M; Kahn, R.S.; $\&$ Geuze, E. (2014). Impaired right inferior frontal gyrus-response to contextual cues in male veterans with PTSD during response inhibition. Journal of Psychiatry and Neuroscience. 39(5), 330-338.
* [7] Vink, M; Zandbelt, B.B.; Gladwin, T; Hillegers, M; Hoogendam, JM; Van der Wildenberg, WPM; et al. (2014). Frontostriatal Activity and Connectivity Increase During Proactive Inhibition Across Adolescence and Early Adulthood. Human Brain Mapping. 35, 4415-4427.
* [8] Soltanifar, M.; Dupuis, A; Schachar, R.;$\&$ Escobar, M. (2019).A frequentist mixture modelling of stop signal reaction times. Biostatistics and Epidemiology, 3, 90-108.
* [9] Soltanifar, M.; Knight, K.; Dupuis, A.; Schachar, R $\&$ Escobar, M. (2020). A Time Series-Based Point Estimation of Stop Signal Reaction Times: More Evidence on the Role of Reactive Inhibition-Proactive Inhibition Interplay on the SSRT Estimations. Brain Science , 10, 598.
* [10] Colonius, H.(1990). A. Note on the Stop Signal Paradigm, or How to Observe the Unobservable. Psychological Review, 97, 309-312.
* [11] Matzke D, Dolan C.V, Logan G.D, Brown S.D, and et al. (2013). Bayesian Parametric Estimation of Stop Signal Reaction Time Distributions. Journal of Experimental Psychology: General, 142(4): 1047-1073.
* [12] Matzke D, Love J, Wiecki T.V, Brown S.D, and et al.(2013). Release the BEESTS: Bayesian Estimation of Ex-Gaussian Stop Signal Reaction Time Distributions. Frontiers in Psychology, 4: Article 918.
* [13] Soltanifar, M.; Escobar, M.; Dupuis, A.$\&$ Schachar, R. (2021). A Bayesian Mixture Modelling of Stop Signal Reaction Time Distributions. arXiv: 2010. 12705.v2[stat.ME].
* [14] Schwarz W. (2001). The Ex-Wald Distribution as a Descriptive Model of Reaction Time Data. Behavioural Research Methods, Instruments and Computers, 33(4). 457-469.
* [15] Prado Martine M.d., $\&$ Fermin Dr. (2008). A Theory of Reaction Times Distributions, (unpublished). Retrieved on December 15, 2020 (http:cggprints.org/6310).
* [16] Rouder J.N. (2005). Are Un-shifted Distributional Models Appropriate for Response Times ? Psychometrica, 70, 377-381.
* [17] Ide, J.S.; Shenoy, P. ; Yu, A.J.;$\&$ Li, C.S. (2013). Bayesian Prediction and Evaluation in the Anterior Cingulate Cortex. The Journal of Neuroscience, 33, 2039-2047.
* [18] Bartholly, S; Rennalls, S.J; Jacques, C; Darby, H; Campbell, IC; Schmidt U; $\&$ ODaly, OG. (2017). Proactive and reactive inhibitory control in eating disorders . Psychiatry Research, 255, 432-440.
* [19] Castro-Meneses, LJ; Johnston, BW;$\&$ Sowman, DF. (2018). The effects of impulsivity and proactive inhibition on reactive inhibition and go process: Insights from vocal and manual stop signal tasks. Frontiers in Human Neuroscience, 9, 529.
* [20] Brevers, D; Voubuisson, E.; Dejonghe, F; Dutrieux, J; Detiea, M; Cheron, G; Verbanck, D; $\&$ Foucart, J. (2018). Proactive and Reactive Motor Inhibition in Top Athletes versus Nonathletes. Perceptual Motor Skills, 125, 289-312.
* [21] Zandbelt, B.B; Bloemendaal, M; Neggers, S.F.W; Kahn, R.S;$\&$ Vink, M. (2013). Expectations and Variations: Declining the Neural Network of Proactive Inhibitory Control. Human Brain Mapping, 34, 2015-2024.
* [22] Dicaprio, V; Modugno, N; Mancini, C; Olivia, E;$\&$ Mirabella, G. (2020). Early Stage Parkinson’s Patients Show Selective Impairment in Reactive but not Proactive Inhibition. Moving Disorders, 35, 409-418.
* [23] Ramautar, JR; Kok, A;$\&$ Rielderinkhof, KR. (2004). Effects of Stop Signal Probability in the Stop Signal Paradigm: The N2/P3 Complex further validated. Brain Cognition , 56, 234-252.
* [24] Rousselet, GA;$\&$ Wilcox, RR. (2020). Reaction Times and Other Skewed Distributions: Problems with the Mean and Median. Meta Psychology, 4, MP 2019.1630.
* [25] Crosbie, J.; Arnold, P.; Peterson, A.D.; Swanson, J., Dupuis, A.; Li, X.; Shan, J.; Goodale, T.; Tam, C.; Strug, L.J.; et al. (2013). Response Inhibition and ADHD Traits: Correlates and heritability in a Community Sample. Journal of Abnormal Child Psychology, 41, 497-597.
* [26] Heathcote A. (1996). RTSYS: A DOS Application for the Analysis of Reaction Times Data. Behavior Research Methods, Instruments and Computers, 28(3), 427-445.
* [27] Kotz, S; Kozubowski, T.J;$\&$ Podgorski, K. (2001). The Laplace Distribution and Generalizations: A revisit with applications to communications, economics, engineering and finance. Birkhauser, Boston, USA. pp. 133-144.
* [28] Reed, W. (2006). The Normal-Laplace distribution and its relatives. In: Balakrishnan, N., Castillo, E., Sarabia Alegria, J.-M. (eds.) Advances in Distribution Theory, Order Statistics, and Inference, pp. 61–74. Birkhäuser Boston, New York.
* [29] Amini Z;$\&$ Rabbani, H. (2017). Letter to the editor: Correction to “The Normal-Laplace distribution and its relatives”, Communications in Statistics - Theory $\&$ Methods,46(4), 2076-2078.
* [30] Evans, M.J $\&$ Rosenthal, J.S. (2010). Probability and Statistics: The Science of Uncertainty (2ed). W. H. Freeman and Company: New York, USA. p. 167.
* [31] Luce, R.D. (1986). Response Times: Their Role in Inferring Elementary Mental Organizations (No.8). Oxford University Press: New York, USA. 16,24
* [32] Signorell A; Aho K; Alfons A; Anderegg N; Aragon T; Arachchige C; et al. (2020). DescTools: Tools for Descriptive Statistics. R package version 0.99.38, https://cran.r-project.org/package=DescTools.
* [33] Reed, W.J. $\&$ Jorgensen, M. (2004). The Double Pareto-lognormal distribution-A New Parametric Model for Size Distributions. Communications in Statistics-Theory $\&$ Methods, 33: 1733-1753.
Appendices
## Appendix A Proofs
This appendix presents proofs for the new results presented in Section 3.1.
### A.1 Proof of Theorem 3.1.
Proof. Method(1): Let $\Delta GORT=GORT_{B}-GORT_{A}$ where
$GORT_{B},GORT_{A}$ are independent with ExG distribution with associated
parameters
$\theta_{B}=(\mu_{B},\sigma_{B},\tau_{B}),\theta_{A}=(\mu_{A},\sigma_{A},\tau_{A}),$
respectively. Then, by Definition 2.1, Definition2.2 and Theorem2.4 it follows
that:
$\displaystyle m_{\Delta GORT}(t)$ $\displaystyle=$ $\displaystyle
m_{GORT_{B}}(t)\times m_{GORT_{A}}(-t)$ $\displaystyle=$ $\displaystyle
m_{ExG(\theta_{B})}(t)\times m_{ExG(\theta_{A})}(-t)$ $\displaystyle=$
$\displaystyle((1-t\tau_{B})^{-1}exp(\mu_{B}.t+\frac{\sigma_{B}^{2}}{2}t^{2}))\times((1+t\tau_{A})^{-1}exp(-\mu_{A}.t+\frac{\sigma_{A}^{2}}{2}t^{2}))$
$\displaystyle=$
$\displaystyle((1-t\tau_{B})(1+t\tau_{A}))^{-1}exp((\mu_{B}-\mu_{A}).t+\frac{\sigma_{B}^{2}+\sigma_{A}^{2}}{2}t^{2})$
$\displaystyle=$
$\displaystyle(1+(\tau_{A}-\tau_{B})t-\tau_{A}\tau_{B}.t^{2})^{-1}exp((\mu_{B}-\mu_{A}).t+\frac{\sigma_{B}^{2}+\sigma_{A}^{2}}{2}t^{2})$
$\displaystyle=$ $\displaystyle m_{AL(\tau_{A},\tau_{B})}(t)\times
m_{N(\mu_{B}-\mu_{A},\sigma_{B}^{2}+\sigma_{A}^{2})}(t):\ \
-\tau_{A}^{-1}<t<\tau_{B}^{-1}.$
Accordingly, by Theorem2.5 it follows that:
$\Delta GORT=^{d}AL(\tau_{A},\tau_{B})\oplus
N(\mu_{B}-\mu_{A},\sigma_{B}^{2}+\sigma_{A}^{2})=^{d}ALG(\tau_{A},\tau_{B},\mu_{B}-\mu_{A},(\sigma_{B}^{2}+\sigma_{A}^{2})^{1/2}).$
Method(2): Using Definition2.1, Definition2.2, Definition2.3 and equation(12)
it follows that:
$\displaystyle\Delta GORT$ $\displaystyle=^{d}$ $\displaystyle GORT_{B}\ominus
GORT_{A}$ $\displaystyle=^{d}$ $\displaystyle
ExG(\mu_{B},\sigma_{B},\tau_{B})\ominus ExG(\mu_{A},\sigma_{A},\tau_{A})$
$\displaystyle=^{d}$ $\displaystyle(N(\mu_{B},\sigma_{B}^{2})\oplus
Exp(\tau_{B}))\ominus(N(\mu_{A},\sigma_{A}^{2})\oplus Exp(\tau_{A}))$
$\displaystyle=^{d}$ $\displaystyle(N(\mu_{B},\sigma_{B}^{2})\ominus
N(\mu_{A},\sigma_{A}^{2}))\oplus(Exp(\tau_{B})\ominus Exp(\tau_{A})$
$\displaystyle=^{d}$ $\displaystyle
N(\mu_{B}-\mu_{A},\sigma_{B}^{2}+\sigma_{A}^{2})\oplus AL(\tau_{A},\tau_{B})$
$\displaystyle=^{d}$ $\displaystyle AL(\tau_{A},\tau_{B})\oplus
N(\mu_{B}-\mu_{A},\sigma_{B}^{2}+\sigma_{A}^{2})$ $\displaystyle=^{d}$
$\displaystyle
ALG(\tau_{A},\tau_{B},\mu_{B}-\mu_{A},(\sigma_{B}^{2}+\sigma_{A}^{2})^{1/2}).$
$\Box$
### A.2 Proof of Theorem 3.2.
Proof. By Definition2.2 and Theorem2.4 for the nth cumulant of the ALG
distribution with four parameters $(\alpha_{1},\alpha_{2},\mu,\sigma)$ we
have:
$\displaystyle\kappa_{n}^{ALG(\alpha_{1},\alpha_{2},\mu,\sigma)}$
$\displaystyle=$
$\displaystyle\kappa_{n}^{AL(\alpha_{1},\alpha_{2})}+\kappa_{n}^{N(\mu,\sigma^{2})}$
$\displaystyle=$
$\displaystyle((n-1)!((-\alpha_{1})^{n}+\alpha_{2}^{n}))+(1_{(n=1)}(n)\mu+1_{(n=2)}(n)\sigma^{2}):\
\ 1\leq n.$
Consequently, the descriptive and the shape statistics for the ALG
distribution with four parameters $(\alpha_{1},\alpha_{2},\mu,\sigma)$, it
follows that:
$\displaystyle Mean$ $\displaystyle E(ALG)=\alpha_{2}-\alpha_{1}+\mu,$
$\displaystyle Variance$ $\displaystyle
Var(ALG)=\alpha_{1}^{2}+\alpha_{2}^{2}+\sigma^{2},$ $\displaystyle Skewness$
$\displaystyle\gamma_{ALG}=\frac{2(\alpha_{2}^{3}-\alpha_{1}^{3})}{(\alpha_{1}^{2}+\alpha_{2}^{2}+\sigma^{2})^{3/2}},$
$\displaystyle Kurtosis$
$\displaystyle\kappa_{ALG}=\frac{6(\alpha_{1}^{4}+\alpha_{2}^{4})}{(\alpha_{1}^{2}+\alpha_{2}^{2}+\sigma^{2})^{2}}.$
Finally, the assertion follows for
$\alpha_{1}=\tau_{A},\alpha_{2}=\tau_{B},\mu=\mu_{B}-\mu_{A},$ and
$\sigma^{2}=\sigma_{B}^{2}+\sigma_{A}^{2}.$
$\Box$
### A.3 Proof of Theorem 3.3.
Proof. Given a four parametric $ALG(\alpha_{1},\alpha_{2},\mu,\sigma)$ for the
proactive inhibition. Then, it can be written in the form:
$ALG(\alpha_{1},\alpha_{2},\mu,\sigma)=\mu\oplus\sigma
N(0,1)\oplus\alpha_{2}Exp_{2}(1)\ominus\alpha_{1}Exp_{1}(1)$
where the random variables $N(0,1),Exp_{i}(1)\sim Exp(1)(i=1,2)$ are mutually
independent. Accordingly, there are uncountably many solutions
$(\mu_{A},\sigma_{A},\tau_{A},\mu_{B},\sigma_{B},\tau_{B})$ for the following
equations:
$\tau_{A}=\alpha_{1},\tau_{B}=\alpha_{2},\mu_{B}-\mu_{A}=\mu,\sigma_{B}^{2}+\sigma_{A}^{2}=\sigma^{2}.$
$\Box$
### A.4 Proof of Theorem 3.4.
Proof. These are straightforward from the probability density function and
[28].
$\Box$
### A.5 Proof of Theorem 3.5.
Proof. Let, $\Delta
GORT=^{d}ALG(\alpha_{1},\alpha_{2},\mu,\sigma)=^{d}AL(\alpha_{1},\alpha_{2})\oplus
N(\mu,\sigma^{2})$ where
$(\alpha_{1},\alpha_{2},\mu,\sigma^{2})=(\tau_{A},\tau_{B},\mu_{B}-\mu_{A},\sigma_{A}^{2}+\sigma_{B}^{2}).$
For, $X=^{d}AL(\alpha_{1},\alpha_{2}),Y=^{d}N(\mu,\sigma^{2}),$ and two
applications of Theorem2.6 for the following convex functions show that both
components of the ALG distribution have increasing hazard functions:
$\displaystyle-ln(f_{X}(t))$ $\displaystyle=$ $\displaystyle
ln(\alpha_{1}+\alpha_{2})\times(\frac{-t}{\alpha_{1}})1_{(-\infty,0]}+\frac{t}{\alpha_{2}}1_{[0,\infty)}(t))\
\ -\infty<t<\infty,$ $\displaystyle-ln(f_{Y}(t))$ $\displaystyle=$
$\displaystyle ln(\sqrt{2\pi}\sigma)+\frac{(t-\mu)^{2}}{2\sigma^{2}}.\ \
-\infty<t<\infty$
Accordingly, by an application of the Theorem2.7, the plausible result
follows.
$\Box$
## Appendix B ExG models parameters
This appendix presents IBPA parameter estimations for cluster type ExG model.
Table 5: Mean posterior Ex-Gaussian parameters estimations across trial types
by IBPA (n = 44).
[b] $\mu-$ parameter $\sigma-$ parameter $\tau-$ parameter $\\#$ $\mu_{S}$
$\mu_{A}$ $\mu_{B}$ $\sigma_{S}$ $\sigma_{A}$ $\sigma_{B}$ $\tau_{S}$
$\tau_{A}$ $\tau_{B}$ 1 357 350 372 32 35 14 86 96 68 2 637 599 732 175 170
132 47 48 68 3 469 484 411 60 57 42 76 73 111 4 597 567 631 163 165 96 66 69
149 5 640 618 608 156 133 58 47 62 121 6 452 431 469 108 106 64 66 65 135 7
689 668 664 136 130 146 47 60 115 8 665 609 660 145 91 237 51 103 120 9 543
484 640 166 151 118 120 147 157 10 470 468 483 56 59 52 98 87 156 11 414 399
597 46 37 118 177 168 80 12 557 534 597 132 128 146 53 58 123 13 550 538 564
137 133 98 55 38 190 14 319 318 365 307 295 370 170 137 264 15 421 416 437 61
56 90 138 142 149 16 358 342 389 61 57 61 48 56 57 17 594 599 561 130 130 133
78 62 196 18 467 397 747 229 190 299 127 131 159 19 426 426 424 67 75 50 102
103 110 20 423 449 504 62 74 129 122 65 169 21 521 519 487 144 157 96 91 97
125 22 397 346 463 87 58 110 94 132 101 23 540 525 588 80 78 79 94 88 128 24
592 571 529 176 136 304 46 69 180 25 577 459 602 165 70 244 69 181 124 26 562
555 694 79 75 160 172 154 148 27 446 436 541 71 60 166 240 236 233 28 486 476
629 82 64 196 172 155 151 29 414 363 391 133 66 213 62 115 111 30 486 484 541
87 86 146 141 127 181 31 546 502 656 137 118 157 90 100 137 32 436 421 462 107
109 90 72 81 88 33 452 454 458 38 46 40 156 156 165 34 404 422 408 105 109 42
95 72 92 35 470 549 595 230 200 298 207 171 136 36 429 400 448 116 139 95 158
163 245 37 521 497 507 89 130 68 112 125 222 38 284 271 321 40 37 53 100 108
91 39 424 432 416 57 55 87 70 52 131 40 419 418 476 52 53 196 145 148 105 41
533 537 517 105 151 93 72 35 159 42 388 445 497 53 145 206 145 57 116 43 506
467 539 97 82 100 66 96 78 44 842 824 822 175 341 165 34 95 320
* •
Notes: $\mu_{S},\sigma_{S},\tau_{S}:$ ExG GORT parameters for single cluster
SST data; $\mu_{A},\sigma_{A},\tau_{A}:$ ExG GORT parameters for type-A
cluster SST data; $\mu_{B},\sigma_{B},\tau_{B}:$ ExG GORT parameters for
type-B cluster SST data; IBPA: $\\#$Chains = 3; Simulations = 20,000; Burn-in
= 5,000 (for all parameters).
|
# Hybrid Neural Pareto Front (HNPF): A Two-Stage Neural-Filter Approach for
Pareto Front Extraction
Gurpreet Singh † The University of Texas at Austin Soumyajit Gupta †
Department of Computer Science Matthew Lease School of Information Clint
Dawson Oden Institute for Computational Engineering and Sciences
###### Abstract
Pareto solutions represent optimal frontiers for jointly optimizing multiple
competing objective functions over the feasible set of solutions satisfying
imposed constraints. Extracting a Pareto front is computationally challenging
today with limited scalability and solution accuracy. Popular generic
scalarization approaches do not always converge to a global optimum and can
only return one solution point per run. Consequently, multiple runs of a
scalarization problem are required to guarantee a Pareto front, where all
instances must converge to their respective global optima. We propose a
robust, low cost hybrid Pareto neural-filter (HNPF) optimization approach that
is accurate and scales (compute space and time) with data dimensions, and the
number of functions and constraints. A first-stage neural network first
efficiently extracts a weak Pareto front, using Fritz-John conditions as the
discriminator, with no assumptions of convexity on the objectives or
constraints. A second-stage, low-cost Pareto filter then extracts the strong
Pareto optimal subset from the weak front. Fritz-John conditions provide
strong theoretical bounds on approximation error between the true and the
network extracted weak Pareto front. Numerical experiments demonstrates the
accuracy and efficiency of our approach.
†††contributed equally to this work.
## 1 Introduction
Multi-Objective Optimization (MOO) problems arise frequently across diverse
fields such as engineering [22], finance [32], and supply chain management
[33]. Such problems share the common requirement to satisfy multiple competing
objectives under a set of constraints imposed by physical or economic limits.
A Pareto optimal solution [28] for an MOO problem is defined as the solution
point away from which no single objective can be improved without diminishing
at least one other objective. A Pareto front is then defined as the set of all
such optimal points that satisfy this definition. Since all solutions reflect
optimal tradeoff points between competing objective functions, choosing
between solutions depends on the user’s preferred tradeoff of objectives.
Computing a Pareto solution to an MOO problem requires optimizing competing
(often non-convex) objective functions under constraints. This optimization
problem is quite challenging: the solution set to an MOO can seldom be
formulated as a closed-form expression, and solving this for practical
problems is compute intensive and often not feasible. Consequently, most
research seeking practical solutions has focused on developing efficient
approximations of the Pareto optimal solution set [8, 15, 14, 29].
The ability to accurately and efficiently compute a Pareto optimal solution
set, with theoretical guarantees and interpretability, would have tremendous
practical value. For example, imagine an online news search for ‘influential
CEOs’ in which the user seeks results that are not only relevant, but also
recent and reliable. In addition to optimizing for these three objective
functions (information relevance, recency, and reliability), imagine the user
further wishes to impose a demographic parity constraint on search results to
ensure racial or gender parity (since implicit data biases might otherwise
yield search result coverage skewed toward white males). The Pareto optimal
solution set would not only satisfy this parity constraint, but further enable
the user to vary the composition of optimal search results based on their
tradeoff preference between getting more recent, breaking news (which may be
less reliable) vs. getting more reliable news (which may be less recent).
Our review of prior work reveals significant limitations: accuracy [5],
compute time [30, 9, 26], and scalability, not to mention limited
interpretability and verfiability. Recent methods for algorithmic fairness
invoking Pareto optimality [1, 20, 23, 34, 35, 36] often suffer from
inconsistent Pareto definitions and impractical assumptions of convex
objective functions and constraints. Furthermore, a notable absence of
benchmarks against known analytical forms makes it difficult to assess
reported results, verify optimality, and A/B test alternative methods. This is
in contrast to studies on computational methods [8, 14, 29, 15] in which such
comparative benchmarking and verification is well established. Although
accurate and verifiable, existing computational methods tend to generate
Pareto points with low density (i.e. providing a coarser representation of the
underlying Pareto front) and poor scalability, with compute times ranging from
hours to days as data dimensionality increases.
In this work, we propose a novel two-stage architecture called Hybrid Neural
Pareto Front (HPNF) for inducing Pareto optimal solution sets. Stage 1
consists of an interpretable and robust neural network that extracts a weak
Pareto solution manifold as the output, given a dataset as input. Following
this, Stage 2 provides a low-cost Pareto filter. For the network loss
function, we use a discriminator based on Fritz-John conditions [19] that
accounts for multiple objectives and constraints. An approximate weak Pareto
manifold is extracted as a weighted output of the softmax function from the
last layer of the network. The softmax activation classifies weak Pareto vs.
non-Pareto data points.
Method | | Generates only
---
Pareto points
| Generates
---
Even Spread
| Ease
---
of Use
| Efficient
---
and Scalable
Fair Pareto [34] | ✗ | ✗ | ✗ | ✗
mCHIM [14] | ✓ | ✗ | ✗ | ✗
PK [29] | ✓ | ✗ | ✗ | ✗
NBI [8] | ✓ | ✓ | ✓ | ✗
Our HNPF | ✓ | ✓ | ✓ | ✓
Table 1: HNPF vs. existing state-of-the-art methods.
Our network architecture has few trainable parameters, making it robust to
outliers and over-fitting. Furthermore, we empirically show computational
efficiency vs. current state-of-the-art methods [8, 15, 29]. Our approach
produces only Pareto points (no false positives) with an even spread and
higher density than possible with existing approaches. Furthermore, our
approach is scalable with both increasing dimensions of the input data, and
the number of functions and constraints. Table 1 summarizes key properties of
our HNPF approach vs. existing methods (see Section 2).
Contributions. Our key contributions are as follows:
1. 1.
A manifold solution strategy for weak Pareto front identification based on
Fritz-John conditions as the discriminator.
2. 2.
A robust neural network for approximating the weak solution manifold for both
convex and non-convex scenarios.
3. 3.
Design of a computationally efficient Pareto filter to extract the strong
Pareto set, compared to existing Pareto filters.
4. 4.
Compared to other neural Pareto approaches our method extracts only Pareto
optimal points with an even spread.
5. 5.
HNPF is computationally scalable as the dimension of variable space, or
functions and constraints increases.
6. 6.
The final layer of the neural net is fully interpretable in terms of
extracting the efficient set of input data as a manifold.
7. 7.
The approximate weak Pareto is bounded below by $0\leq\epsilon\leq 1$ w.r.t.
the true manifold upon convergence.
8. 8.
We will share our source code and benchmark datasets for reproducibility upon
acceptance.
## 2 Related Work
As noted earlier, since all Pareto optimal solutions reflect optimal tradeoff
points between competing objective functions, choosing between solutions
depends on the user’s preferred tradeoff of objectives. Prior work can be
organized around four directions for managing user preferences: 1) No
preference [37]: user preference criteria are not explicitly specified; 2) a
priori [12]: preference criteria are explicitly specified before computation;
3) a posteriori [8]: preference criteria are explicitly specified after
computation; and 4) Interactive methods [25]: preference criteria are
continuously consulted to isolate one of the optimal solutions.
### 2.1 Generic and Enhanced Scalarization
One common approach is to convert an MOO problem into a Single Objective
Optimization (SOO) problem via scalarization. However, generic scalarization
methods [1, 20, 23, 34, 35] suffer from various limitations. Firstly, these
approaches can only extract one solution point at a time given that the
minimization problem converges to the global optimum. However for practical
applications, with non-convex objectives and constraints, ensuring global
optimality is non-trivial. Secondly, multiple runs with different trade-off
parameters must be performed in order to extract the weak Pareto solution set,
resulting in substantial computational overhead [35]. Finally, the Pareto
solution set can still form a non-convex manifold even when the objectives are
convex [14] due to the presence of non-convex constraints (see Case III in
Section 5). These challenges prove to be major obstacles in the deployment of
scalarization approaches as a practical tool for Pareto set extraction.
Generic scalarization should not be confused with enhanced scalarization
approaches [8, 14, 29], whose strength lies in the specific localization of
the objective space that allows treatment of non-convex functions and
constraints. Although accurate and complete, enhanced scalarization approaches
suffer from low computational scalability and low density of Pareto points on
the solution manifold. For example, the $30$ dimensional benchmark in Section
5 Case V shows enhanced scalarization methods (mCHIM and PK) generating a
Pareto set in approximately 18 hours.
Enhanced scalarization methods fall under category (3) of a posteriori
methods. One such enhanced approach to solve an MOO involves constructing a
local linear or epsilon scalarization based SOO. These methods include Normal
Boundary Intersection (NBI) [8], Normal Constraint (NC) [24], Successive
Pareto Optimization [27], modified Convex Hull of Individual Minimum (mCHIM)
[14] and Pirouz-Khorram (PK) [29]. NBI [8] produces an evenly distributed set
of Pareto points given an evenly distributed set of weights. Furthermore, NBI
produces Pareto points in the non-convex parts of the Pareto curve while being
independent of the relative scales of the objective functions. It uses the
concept of Convex Hull of Individual Minima (CHIM) to break down the
boundary/hull into evenly spaced segments and then trace the weak Pareto
points.
As an improvement over the NBI method, mCHIM uses a quasi-normal procedure to
update the aforementioned CHIM set iteratively, to obtain a strong Pareto set.
PK [29], on the other hand, uses a local $\epsilon$-scalarization based
strategy that searches for the Pareto front using controllable step-lengths in
a restricted search region, thereby accounting for non-convexity. Gobbi et al.
[15] proposed a framework using Fritz-John conditions [19] to obtain
analytical solutions for convex functions and constraints with high point
density. Note that, all of these aforementioned enhanced methods are
guaranteed to converge to the Pareto front under their respective assumptions
on the function property each method can handle.
### 2.2 Bayesian and Genetic Approaches
Methods that are a priori (2) require a prior distribution or initial seed
parameters to be specified beforehand. Examples include Bayesian [17, 4, 16]
and Evolutionary [30, 9, 26] methods. Khan et al. [17]’s Bayesian method
showed convergence to the Pareto front, but only under a linear setting, which
is the strictest form of convexity. In recent Bayesian methods [4, 16], not
only was convexity assumed, but even in actual convex cases significant error
was still incurred. Deb et al. [9] introduced the Non-dominated Sorting
Genetic Algorithm II (NSGA-II) algorithm that involves recombination, mutation
and selection of a population representing the set of solutions points
considered to be Pareto, each having one or more assigned objective values.
The population is maintained to consist of diverse solutions, resulting in a
set of non-dominated individuals that are expected to be near (not on) the
real Pareto front. Other variants include NSGA-I [30] and NSGA-III [26].
However, convergence and reproducibility are not guaranteed with Genetic
Algorithms, and significant hyper-parameter tuning is required.
### 2.3 Approaches in the Fairness Literature
Pareto optimality is being increasingly pursued in classification and fairness
research (e.g. see survey [5]). However, we are not aware of any work in this
area providing verifiable solutions for benchmarking scenarios wherein the
ground truth is known.
Several works [1, 23] seek to balance classification accuracy vs. a no
unnecessary harm notion of fairness relying upon convexity assumptions without
justification. The Weighted Sum Method (WSM) [7], commonly used in fairness
literature, is a linear scalarization approach to convert an MOO into an SOO
using a convex combination of objective functions and constraints. However,
this is viable only when the functions and constraints are also convex [14].
Valdivia et al. [34] present a group-Fairness based trade-off model for
decision tree based classifiers using the aforementioned genetic algorithm
NSGA-II, which has the same convergence and reproducibility issues mentioned
earlier. In addition, their reported results violate fundamental definitions
of Pareto optimality. Wei and Niethammer [35] provide the first neural
architecture for Pareto front computation for Fairness vs. Accuracy on
classification datasets. They rely upon a Chebyshev scalarization, which
assumes that objective functions must sum up to a constant, but do not justify
this assumption. In the Fair-Recommendation literature, Xiao et al. [36] seek
to balance social welfare and group fairness for movie recommendations. They
also propose a linear scalarization-based formulation which arrives at the
true front for convex functions only. Lin et al. [20] claim Pareto optimality
using KKT conditions, which is guaranteed to converge only if the functions
and constraints are convex under linear scalarization.
### 2.4 Our Inspiration
We draw inspiration from three seminal works: (1) Das and Dennis [8] proposed
to break the functional domain boundary into uniform and evenly spaced
segments (see CHIM in [8]) to identify weak Pareto points with guarantees.
Motivated by this, we first identify the weak Pareto front using a robust
neural network (Stage 1). (2) Messac et al. [24] proposed the first Pareto
filter to obtain the set of strong Pareto points from the aforementioned weak
Pareto set. The filter uses an all-pair comparison criterion to reject
dominated points from the weak Pareto set. This filter motivates our low-cost
Pareto filter (Stage 2) design, which avoids the expensive all-pair comparison
using a plane search strategy. (3) Gobbi et al. [15] presented the matrix form
of the Fritz-John conditions satisfying the existence of Pareto points.
Although their approach is only valid for convex cases, we extend the Fritz-
John matrix form as a discriminator to identify Pareto point even for non-
convex cases.
## 3 Pareto Optimality
A general multi-objective optimization problem can formulated as:
$\displaystyle\underset{}{min}\quad F(x)$
$\displaystyle=(f_{1}(x),f_{2}(x),\ldots,f_{k}(x))$ (1)
$\displaystyle\text{s.t.}\quad x\in S$
$\displaystyle=\\{x\in\mathbb{R}^{n}|G(x)=(g_{1}(x),g_{2}(x),\ldots,g_{m}(x)\leq
0\\}$
in $n$ variables $(x_{1},\ldots,x_{n})$, $k$ objective functions
$(f_{1},\ldots,f_{k})$, and $m$ constraint functions $(g_{1},\ldots,g_{m})$.
Here, $S$ is the feasible set i.e. the set of input values $x$ that satisfy
the constraints $G(x)$. For a multi-objective optimization problem there is
typically no single global solution, and it is often necessary to determine a
set of points that all fit a predetermined definition for an optimum.
### 3.1 Definitions
We borrow and adapt existing definitions of Pareto optimality [28] from Marler
and Arora [22].
Strongly Pareto Optimal: The Pareto optimal solution $x^{*}$ for Eq. (1)
satisfies the conditions:
$\displaystyle\nexists x_{j}:f_{p}(x_{j})\leq
f_{p}(x^{*}),\quad\textrm{for}\quad p=1,2,\ldots,k$ $\displaystyle\exists
l:f_{l}(x_{j})<f_{l}(x^{*})$ (2)
This states that for $x^{*}\in X$ to be strongly Pareto optimal, there does
not exist another $x_{j}\in X$, s.t. $f_{p}(x_{j})\leq f_{p}(x^{*})$ for all
functions $f_{p},\forall p\in[1,k]$ and $f_{l}(x_{j})<f_{l}(x^{*})$ for at
least one function $f_{l}$.
Weakly Pareto Optimal: A weak Pareto point $\tilde{x}^{*}$ satisfies:
$\displaystyle\nexists
x_{j}:f_{p}(x_{j})<f_{p}(\tilde{x}^{*}),\quad\textrm{for}\quad p=1,2,\ldots,k$
(3)
A point is weakly Pareto optimal if no other point exists that improves all of
the objectives simultaneously. This is different from a strongly Pareto
optimal point, s.t. no point exists, that improves at least one objective
without detriment to other objectives.
Efficient and Inefficient Points: A Pareto efficient point is defined on the
domain $X$ as a point $x^{*}$ iff no other point $x$ exists s.t. $F(x)\leq
F(x^{*})$ with at least one $f_{p}(x)<f_{p}(x^{*})$. The point $x^{*}$ is
considered inefficient otherwise.
Dominated and Non-Dominated Points: Dominated points are defined w.r.t. the
objective function in the criterion space. The objective function vector
$F(x^{*})$ is non-dominated iff no other vector $F(x)$ exists s.t. $F(x)\leq
F(x^{*})$ with at least one $f_{p}(x)<f_{p}(x^{*})$. The vector $F(x^{*})$ is
considered dominated otherwise.
(a) Convex form
(b) Non-Convex form
Figure 1: Pareto optimal set under different objectives. Note that the red
line corresponds to the min-min strong Pareto optimal front for both problems
and the dashed line corresponds to the weak Pareto Front for the entire
Feasible Set.
Fig. 1 shows two separate MOOs for two competing functions. The feasible set
of points are shown as gray shaded area with the Pareto boundary shown as
dashed lines. For both problems, a joint minimization problem is considered,
resulting in a Pareto optimal set (red curves) facing the origin. A different
Pareto boundary can be obtained if a joint-maximization or a mixed min-max
problem is considered. As shown in Fig. 1 (b), a strong Pareto optimal
solution from a non-convex weak Pareto front can result in a discontinuous
manifold. In the following sections, although we rely on dominated and non-
dominated points for visualization purposes, the approximation errors and our
algorithm rely primarily on efficient and inefficient points for computational
purposes.
### 3.2 Fritz John Conditions
Let the objective and constraint function in Eq. (1) be differentiable once at
a decision vector $\tilde{x}^{*}\in\mathcal{S}$. The Fritz-John necessary
conditions for $\tilde{x}^{*}$ to be weak Pareto optimal is that vectors must
exists for $0\leq\lambda\in\mathbb{R}^{k}$, $0\leq\mu\in\mathbb{R}^{m}$ and
$(\lambda,\mu)\neq(0,0)$ (not identically zero) s.t. the following holds:
$\displaystyle\sum_{i=1}^{k}\lambda_{i}\nabla
f_{i}(\tilde{x}^{*})+\sum_{j=1}^{m}\mu_{j}\nabla g_{j}(\tilde{x}^{*})=0$ (4)
$\displaystyle\mu_{j}g_{j}(\tilde{x}^{*})=0,\forall j=1,\ldots,m$
Gobbi et al. [15] presented an $L$ matrix form, comprising the gradients of
the functions and constraints as follows:
$\displaystyle L=\begin{bmatrix}\nabla F&\nabla G\\\
\mathbf{0}&G\end{bmatrix}\quad[(n+m)\times(k+m)]$ (5) $\displaystyle\nabla
F_{n\times k}=[\nabla f_{1},\ldots,\nabla f_{k}]$ $\displaystyle\nabla
G_{n\times m}=[\nabla g_{1},\ldots,\nabla g_{m}]$ $\displaystyle G_{m\times
m}=diag(g_{1},\ldots,g_{m})$
The matrix equivalent of Fritz John Conditions for $x^{*}$ to be Pareto
optimal, is to show the existence of $\lambda\in\mathbb{R}^{k+m}$ in Eq. (4)
such that:
$\displaystyle L\cdot\delta=0\quad\text{s.t.}\quad
L=L(\tilde{x}^{*}),\mathbf{\delta}\geq 0,\mathbf{\delta}\neq 0$ (6)
The non-trivial solution ($\mathbf{\delta}$ is not identically zero) for Eq.
(6) is:
$\displaystyle det(L^{T}L)=0$ (7)
The weak Pareto front is characterized by the set of points such that matrix
$L$ is low rank. This ensures that the points identified are either inside the
feasible set or at boundaries dictated by the constraints. For e.g. if
$\mu_{1}=0$ for any $\lambda_{i}$, then $\sum_{i}\lambda_{i}f_{i}=0$ must be
satisfied for the corresponding internal point $x^{*}$ to be Pareto. Similarly
if $\mu_{1}\neq 0,\mu_{j\neq 1}=0$ in the aforementioned case, then $g_{1}=0$
holds true for the corresponding boundary point $x^{*}$ to be Pareto. Note
that all Pareto points satisfy $\nabla f_{i}=0$ for at least one $i$ whether
they lie inside the feasible set or on the boundaries. This is to say that all
points $x^{*}$ are local optimizers for at least one $f_{i}$. The rank of the
matrix $L^{T}L$ determines the dimension of the Pareto manifold. Furthermore,
the necessary condition written as $det(L^{T}L)$ is independent of the
preference parameters $\lambda_{i},\,\mu_{j}$. Eq. (7) now serves as a
condensed discriminator to identify a weak Pareto front. In what follows, we
use this matrix form of Fritz-John conditions to approximate the Pareto front
using a robust, low-weight, neural network.
## 4 HNPF Framework
In this section we lay out the details of the proposed computationally
efficient hybrid two-stage architecture for Pareto set detection.
### 4.1 Stage 1: Neural Net for Weak Pareto Front
The proposed neural network consists of feed-forward layers with tanh
activation. There are three layers of dense connections with eight neurons
each, to smoothly approximate the optimal solution manifold $M(X)$ as
$\tilde{M}(X)$, shown in Fig. 2. The last layer of the network has two neurons
with softmax activation for binary classification of Pareto vs. non-Pareto
points. In other words, this layer approximates the separation manifold that
distinguishes inefficient points from the weak Pareto points, in the feasible
set $S$. Note that our network loss is representation driven, since the Fritz
John discriminator (Eq. (7)) explicitly classifies points as being weak Pareto
or not. The network accepts or rejects the input data points $X$ based on the
Fritz-John discriminator described by the objective functions and constraints.
The Fritz-John necessary conditions for weak Pareto optimality, as pointed out
earlier, require that the $D=det(L(X)^{T}L(X))=0$. Therefore, $1-D$ and $D$
naturally provide us with binary labels for the softmax activated output
layer. A binary cross entropy loss ensures that the distribution of the
extracted manifold $\tilde{M}(X)$ matches the distribution of the weak Pareto
front satisfying the Fritz-John conditions. The network architecture is
purposely kept low weight for weak Pareto manifold extraction to provide
robustness against outliers.
Figure 2: Proposed Neural network for finding the weak Pareto points. The last
layer performs Fritz John criteria enabled binary cross entropy classification
of data points, using softmax activation, ensuring weak Pareto optimality.
Error bound. For a user-prescribed relaxation margin $0\leq\epsilon\leq 1$,
the approximation error between the network extracted manifold
$\tilde{M}(\tilde{X})$ and the true solution $M(X^{*})$ is bounded below by
$\|\tilde{M}(\tilde{X})-M(X^{*})\|_{2}\leq\epsilon$, upon convergence. See
Appendix C for proof.
### 4.2 Stage 2: Pareto filter for Strong Pareto Set
A Pareto filter is an algorithm that, given a set of weak Pareto points in
objective space, retains a subset of non-dominated points. This corresponds to
the strong Pareto set s.t. none of the points are dominated. In other words,
the filter eliminates all dominated points from the given set. A state of the
art Pareto filter, defined in Messac et al. [24], is used as a post-processing
step to extract a strong Pareto set. However, note that this Pareto filter
requires an all-pairs comparison, an $O(n!)$ calculation, and thus becomes
computationally expensive as the point set grows. This necessitates that the
number of weak Pareto points be small when using this filter. However, since
Stage 1 of our approach generates weak Pareto points with high density, the
filter proves to be quite expensive.
We present an efficient Pareto filter algorithm for finding a strong Pareto
set which is computationally scalable to arbitrary dimensions. The algorithm
is based on a plane search strategy, inspired by Kd-Trees [3], well known for
efficient data partitioning and storage. The compute complexity of our
approach is determined by the number of competing functions while being
linearly proportional to the number of points. The inputs to the algorithm
(Alg. 1) are the number of functions $k$, their minimum and maximum bounds
$f(min),f(max)$, discretization level $h$ of the function space and the weak
Pareto points $P$. The output is the strong Pareto set $\\{x^{*}\\}$.
Following Alg. 1, we can estimate its time complexity. There are three nested
for loops which carry the load. In the worst case that the points in the weak
Pareto set are all strong Pareto, then the cardinality of the set $P$ remains
unchanged. Let us also define $z=(f_{i}(max)-f_{i}(min))/h$ to denote the
number of chunks into which the function space is divided. Thus, the worst
case complexity of the proposed Pareto filter is $\mathbf{O(kzn)}$. For
scenarios, where the strong Pareto set is a subset of the original weak set
$P$, the complexity reduces in the factor guided by $n$.
Algorithm 1 Pareto filter
1:Data $P=\\{\tilde{x}^{*}\\}\in\mathbb{R}^{n}$ weak Pareto points
2:Input $f(min),f(max):$ bounds of each function $f_{i},\forall i\in k$
3:Input $k:$ number of functions, $h:$ discretization level
4:for $i\in k$ do $\triangleright$ Loop over all functions
5: $level=f_{i}(min)$
6: for $j\in(f_{i}(max)-f_{i}(min))/h)$ do $\triangleright$ Loop over all
levels
7: $temp=\varnothing$
8: for $p\in P$ do
9: if $level\leq f_{i}(p)<level+h$ then
10: $temp=temp\cup p$
11: if $card(temp)>1$ then $\triangleright$ Cardinality of set
12: $x_{p}=min\,f_{q}(x),x\in temp,q=i+1$ $\triangleright$ Efficient point
13: $P=P\backslash(temp\backslash x_{p})$ $\triangleright$ Remove inefficient
points
14: $level=level+h$
15:Output: Strong Pareto set $x^{*}=P$
(a) Weak Pareto Set
(b) First pass
(c) Second pass
Figure 3: Illustration of the proposed Pareto Filter on a two-function
scenario for visualization purposes
It is easy to visualize the working of the proposed Pareto filter in Fig. 3.
We start with the weak Pareto set $P$ (Fig. 3 (a)) for a non-convex form. The
first pass over $f_{1}$ removes a set (segment DEF) of dominated Pareto points
(Fig. 3 (b)). The leftover points in $P$ are then filtered again based on
$f_{2}$ (Fig. 3 (c)), where the dominated points (segment ABC) as per $f_{2}$
are removed. Points surviving the filtering process belong to the strong
Pareto set.
## 5 Results
In this section, we present five numerical experiments for benchmarking and
analysis (see Appendix A for two additional cases). These experiments address
standard analytical forms, with increasing complexity and scale in the number
of functions $(k)$, constraints $(m)$ and dimension of variables $(n)$. While
cases may appear synthetic, they arise from practical physical domains in
various engineering fields. We compare our results vs. those from two current
state-of-the-art methods: mCHIM [14] and PK [29].
Sampling. Since we are only provided with objective functions and constraints,
we must sample data points from the variable domain in order to generate
candidates to test for optimality. Firstly, if there are any direct
constraints on variable values, we consider this feasible domain for sampling,
as in the benchmark cases. Secondly, lacking any prior knowledge of where the
Pareto front may reside, we sample values random uniform distribution in the
feasible variable domain. Objective functions evaluated at these points
generate a quantized, topographic map of the function domain that is then used
to identify optimal points. For each benchmark test case below, we generate
11K points from a random uniform distribution in the feasible variable domain,
to serve as training data. The training-validation split is 90-10%. Once the
manifold is learned by the network, we feed in 90k points within the
permissible domain to plot the Pareto set for visualization.
### 5.1 Experimental Setup
Experiments use an Nvidia 2060 RTX Super 8GB GPU, Intel Core i7-9700F 3.0GHz
8-core CPU and 16GB DDR4 memory. We use the Keras [6] library on a Tensorflow
2.0 backend with Python 3.7 to train the networks in this paper. For
optimization, we use AdaMax [18] with parameters (lr=0.001) and $1000$ steps
per epoch.
While neural approaches often pre-initialize the network with layer-wise
training [2], a strength of HPNF is that all network weights can be simply
drawn from a uniform random distribution. Since the data domain is discrete,
an exact zero might not be achievable. We therefore use a slightly relaxed
criterion of $\epsilon=0.001$ as the classification margin. Any point above
this value will be classified as weak Pareto. For all results, the extracted
Pareto set (shaded red) overlaps the true Pareto set with an $\epsilon$
spread.
Due to stochastic variation, neural network studies often report variance
across several runs. However, the only approximation errors with our method
lie in the extracted manifold over runs. Error bounds are given in Section 4.1
and Appendix C. Since the manifold remains constant across runs, the loss
itself is the approximation error with a minimum achievable value of 0 at
machine precision. With respect to this 0 over multiple runs, the loss
function is the deviation from the true manifold. We thus do not report mean-
variance across runs. Section 5.9 shows loss profiles.
### 5.2 Case I: n=2, k=2, m=2
This problem was originally proposed in [11]. Jointly minimize
$\displaystyle
f_{1}(x_{1},x_{2})=1-exp(-[(x_{1}-1/\sqrt{(}2))^{2}+(x_{2}-1/\sqrt{(}2))^{2}])$
$\displaystyle
f_{2}(x_{1},x_{2})=1-exp(-[(x_{1}+1/\sqrt{(}2))^{2}+(x_{2}+1/\sqrt{(}2))^{2}])$
$\displaystyle\text{s.t.}\quad g_{1},g_{2}:-1/\sqrt{2}\leq x_{1},x_{2}\leq
1/\sqrt{2}$
The Pareto set can be computed by applying linear scalarization (WSM) for this
problem since all the functions and constraints are convex. Gobbi et al. [15],
NBI, mCHIM and PK are able to extract the Pareto solution set. Fig. 4 shows
the solution from our proposed network with high point density, where we can
visually verify (Fig. 4(b)) that the network approximated the Pareto manifold,
validating that the Pareto points indeed closely satisfy $x_{1}=x_{2}$.
(a) Function Domain
(b) Variable Domain
Figure 4: Pareto Front for Case I. Please refer to color plots for proper
visualization for all following figures.
### 5.3 Case II: n=2, k=2, m=2
This problem was proposed in [14]. Jointly minimize
$\displaystyle f_{1}(x_{1},x_{2})=x_{1}$ $\displaystyle
f_{2}(x_{1},x_{2})=1+x_{2}^{2}-x_{1}-0.1sin3\pi x_{1}$
$\displaystyle\text{s.t.}\quad g_{1},g_{2}:0\leq x_{1}\leq 1,-2\leq x_{2}\leq
2$
Gobbi et al. [15] does not consider this scenario due to non-convexity of
$f_{2}$. As shown in [14], WSM can only identify a subset of the Pareto
optimal points. NBI, mCHIM and PK methods are able to identify points in this
case with equal density. Fig. 5 shows the results from our model with high
point density. It also satisfies closely the true Pareto manifold given by
$0\leq x_{1}\leq 1,x_{2}=0$ in Fig. 5(b).
(a) Function Domain
(b) Variable Domain
(c) WSM
(d) NBI
(e) mCHIM
Figure 5: Pareto Front for Case II. Note the even spread of point HNPF
produces. NBI produce an even spread while mCHIM cannot. WSM fails for non-
convexity in function.
### 5.4 Case III: n=2, k=2, m=4
This problem was proposed in [31]. Jointly minimize
$\displaystyle f_{1}(x_{1},x_{2})=x_{1}$ $\displaystyle
f_{2}(x_{1},x_{2})=x_{2}$ $\displaystyle\text{s.t.}\quad
g_{1}(x_{1},x_{2})=(x_{1}-0.5)^{2}+(x_{2}-0.5)^{2}\leq 0.5$ $\displaystyle
g_{2}(x_{1},x_{2})=x_{1}^{2}+x_{2}^{2}-1-0.1\cos(16\arctan(\frac{x_{1}}{x_{2}}))\geq
0$ $\displaystyle g_{3},g_{4}:0\leq x_{1},x_{2}\leq\pi$
This form is convex in $f_{1},f_{2}$ but the non-convex constraints in
$g_{1},g_{2}$ forces the Pareto front to be non-convex. While NBI (without
Pareto filter) fails in this scenario, both mCHIM and PK extracts the non-
dominated Pareto points with limited density $(\sim 40)$. HNPF extracts point
with higher density (Fig. 6). Since the front is strongly affected by the
constraints, Fig. 6(a) shows a sinusoidal weak Pareto front. To arrive at the
non-dominated Pareto set, we then post-process this result using the efficient
Pareto filter proposed in sub-section 4.2. The updated discontinuous set of
non-dominated Pareto points can be seen in Fig. 6(b) following the visual
explanation in Fig. 3.
(a) Dominated
(b) Non-Dominated
(c) mCHIM
Figure 6: Strong Pareto Front for Case III. All dominated points are removed
from set after applying the Pareto filter.
### 5.5 Case IV: n=3, k=3, m=4
This problem was proposed in [14]. Jointly minimize
$\displaystyle f_{1}(x_{1},x_{2},x_{3})=x_{1}$ $\displaystyle
f_{2}(x_{1},x_{2},x_{3})=x_{2}$ $\displaystyle f_{3}(x_{1},x_{2},x_{3})=x_{3}$
$\displaystyle\text{s.t.}\quad
g_{1}(x_{1},x_{2},x_{3})=(x_{1}-1)^{2}+(x_{2}-1)^{2}+(x_{3}-1)^{2}\leq 1.0$
$\displaystyle g_{2},g_{3},g_{4}:x_{1},x_{2},x_{3}\geq 0$
This form is convex in $f_{1},f_{2},f_{3}$ but the non-convex constraint in
$g_{1}$ forces the Pareto front to be non-convex. The results using our
method, as shown in Fig. 7, are in good agreement with mCHIM and PK methods
with a higher point density.
(a) Function Domain
(b) Variable Domain
Figure 7: Pareto Front for Case IV
### 5.6 Case V: n=30,k=2,m=30
This problem was proposed in [38]. Jointly minimize
$\displaystyle f_{1}(x)=x_{1}+\frac{2}{|J_{1}|}\sum_{j\in J_{1}}y_{j}^{2}$
$\displaystyle f_{2}(x)=1-\sqrt{x_{1}}+\frac{2}{|J_{2}|}\sum_{j\in
J_{2}}y_{j}^{2}$ $\displaystyle\text{s.t.}\quad g_{1},\ldots,g_{30}:0\leq
x_{1}\leq 1,-1\leq x_{j}\leq 1,j=2,\ldots,m$ $\displaystyle
J_{1}=\\{j|j\,\textrm{is odd},2\leq j\leq m\\},J_{2}=\\{j|j\,\textrm{is
even},2\leq j\leq m\\}$ $\displaystyle
y_{j}=\left\\{\begin{matrix}x_{j}-[0.3x_{1}^{2}\cos(24\pi
x_{1}+\frac{4j\pi}{m})+0.6x_{1}]cos(6\pi x_{1}+\frac{j\pi}{m})\quad j\in
J_{1}\\\ x_{j}-[0.3x_{1}^{2}\cos(24\pi
x_{1}+\frac{4j\pi}{m})+0.6x_{1}]cos(6\pi x_{1}+\frac{j\pi}{m})\quad j\in
J_{2}\end{matrix}\right.$
This form is non-convex in both $f_{1},f_{2}$. The dimension of the design
variable space is $m=30$. The corresponding Pareto front is non-convex. The
results using our method, as shown in Fig. 8, are in good agreement with mCHIM
and PK methods
(a) Function Domain
(b) Variable Domain
(c) mCHIM
Figure 8: Pareto Front for Case V. Note the density difference between HNPF
and mCHIM in the variable space.
### 5.7 Summary of Results
Linear Scalarization, as in WSM [7], is a well-known approach for Pareto set
detection in Fairness and Classification literature. This approach fails for
all but one (Case I) of the cases presented above, since either the functions
or constraints or both are non-convex. This raises serious concerns regarding
validation in Fairness literature: whether the points in the solution set are
Pareto optimal or not. If not, then all such works are generating points which
are non-Pareto (weak, strong or otherwise) in any sense of the definitions
posed in Section 3. Case III highlights the fallacies of such convexity
assumptions, where in spite of functions being convex themselves, the
analytical front is non-convex due to the interaction of the functions and
non-convex constraints.
| HNPF | mCHIM
---|---|---
Case | Density | Points | Evals | Density | Points | Evals
Case II | 1.83 | 1648 | 90K | 4.39e-2 | 33 | 75,152
Case III | 1.37 | 1241 | 90K | 1.01e-2 | 33 | 328,375
Case IV | 6.57 | 5915 | 90K | 5.86e-3 | 43 | 733,752
Case V | 0.20 | 184 | 90K | 1.38e-6 | 33 | 2,379,459,895
Table 2: Pareto optimal point density % (ratio of #extracted optimal points to
#function evaluations). HNPF finds many more optimal points with many fewer
function evaluations. Case I has not been considered by mCHIM, hence left out.
NBI [8] works for cases where the detected weak Pareto front consists of non-
dominated points. Therefore, NBI generates correct solution sets in Cases I,
II, IV, V with equal density of points on the Pareto front. In essence,
applying the Pareto filter on the NBI generated solution set would resolve the
discontinuous cases too.
Gobbi [15], a Genetic Algorithm solution strategy, works for Case I. Their
algorithm is developed only for cases where all the functions and constraints
are convex.
NBI, mCHIM,PK and HNPF produce only Pareto points, which is not guaranteed by
WSM [7]. Additionally, HNPF generates Pareto points uniformly with high
density, while state-of-the-art mCHIM [14] and PK [29], although accurate,
limit themselves to low point density ($\sim 40$) with large computational
overhead as the variable dimension scales. Table 2 shows a comparison between
HNPF and mCHIM. See Appendix D for a similar comparison against PK.
### 5.8 Runtime Comparison
Using numerical experiments, we previously verified that mCHIM, PK and our
method always arrives at the correct results for all the considered cases. We
now perform a compute time analysis against mCHIM and PK, to demonstrate
improved performance using our proposed approach. The trajectories in Fig. 9
show the compute times for the high dimensional Case VII. Note that for mCHIM
and PK, the timings are reported for dimensions n=30 and n=4, respectively.
For our method, the runtimes are reported for Case VII with the variable space
dimension ranging from $[2-30]$.
Figure 9: Runtime of HNPF vs. mCHIM and PK, as the variable dimension
increases. All methods have a linear increase in runtime with dimension, but
HNPF scales much better.
The reported runtime with two dimensional variable space might give the false
notion that mCHIM and PK are more efficient than HNPF. However, as the
variable dimension increases, both mCHIM and PK become far more expensive, as
shown in Fig. 9. These methods also produce a low density of Pareto points
$(\sim 40)$, while HNPF yields high density $(\sim 1k)$. Since both mCHIM and
PK are based on enhanced scalarization, solving the resulting problem to
extract Pareto points suffers from scaling issues.
### 5.9 Loss Profile
We now briefly discuss the training process for the cases shown above. On an
average, the network takes around $10/20/30$ epochs for the
simple/moderate/hard cases as visualized in Fig. 10. Since the last layer of
the network is classifying points as being weak Pareto or not, the runtime is
dictated by the complexity of the curve in the design variable space. The more
non-linear the solution manifold, the more training time is required to
approximate it.
Case II and IV both converge within 10 epochs although they lie in 2D and 3D
space, respectively. Per Fig. 5 and 7 (b), the design variable space is convex
and so the solution manifold is less complicated. Although in 2D variable
space, Case III takes 20 epochs owing to the sinusoidal solution manifold.
Case V converges in 30 epochs, the design space is 30 dimensional, hence the
compute complexity increases due to the construction of a larger $L$ matrix.
The validation loss curve lies below the training loss (but strictly at
scale), suggesting that our low-weight network did not over/underfit.
(a) Case II
(b) Case III
(c) Case IV
(d) Case V
Figure 10: Training (blue) and validation (orange) curves for 4 cases. An
increase in the dimensions of the design variable space results in increased
costs for constructing the L matrix. Consequently, the network takes more
epochs to converge.
## 6 Modeler Interpretability
Motivated by Lipton [21]’s definitions of model interpretability and trust, we
adopt the persona of a modeler in assessing the interpretability of our model.
In all of the problems above, the approximate manifold $\tilde{M}$ is
described by the user specified loss function. If a domain specific analytical
solution ($M(X^{*})=0$) is known, then the approximate network generated
solution set ($\tilde{M}(\tilde{X})=0$) can be verified by comparing
$\tilde{X}$ and $X^{*}$. Additionally, a domain-specific modeler can also
compare the approximate manifold $\tilde{M}$ (see Fig. 11 for case III), at
the last layer of the network, against the true manifold $M$ known from the
analytical form. If the modeler is able to verify that the network classifies
the correct (truly Pareto optimal) data points in the variable space as being
Pareto optimal (high probability value), the trust in the network’s working is
established.
Figure 11: Classification boundary of weak Pareto points for Case III. The
final layer assigns a probability score to each point in the variable space as
being Pareto optimal or not. A modeler needs to examine the produced
points/manifold against the known analytical form in the variable domain to
verify the correctness of the network’s working.
## 7 An Application: Fair Search
Imagine a new policy for predictive policing is under consideration, with
various public arguments being published for and against adoption. By
reviewing this body of arguments, one might arrive at an informed and balanced
understanding of the issue and public debate surrounding it. This, in turn,
could guide citizens or lawmakers in voting, or help a journalist to provide
balanced reporting.
Assume a search engine is used to find information that is both relevant and
balanced. Specifically, assume we wish to maximize two objective functions
computed on the set of retrieved search results: 1) relevance and 2) diversity
(i.e., balanced inclusion of search results that are for and against the
proposed policy). Note that our diversity target is specified here as a soft
objective function to maximize rather than a hard constraint to rigidly
enforce. Let us make a further independence assumption between relevance and
diversity: knowing that a retrieved document is relevant does not provide any
indication as to whether or not that document presents an argument for or
against the proposed policy.
Gao and Shah [13] present such a fair search problem as follows (albeit with a
different motivating back story). Let $S$ denote the search result set of
cardinality $|S|=s$. Assume each document $d_{i}$ has binary relevance
$r_{i}\in\\{0,1\\}$ and group assignment $g_{i}\in\\{0,1\\}$ (i.e., for or
against the policy, in our scenario), and that $r_{i}$ and $g_{i}$ are
independent, as above. The optimization goal is dual maximization of the
average relevance $f_{r}=\overline{r_{i}}=\frac{\sum r_{i}}{n}$ and the
entropy $f_{g}=H(\frac{\sum g_{i}}{n})$ of the search result set, with entropy
used as the measure of diversity (aka parity, balance, or group fairness).
### 7.1 Insights from Pareto Framing
Gao and Shah [13] pose several questions, including “1. What are the possible
relevance and fairness scores of a solution set $S$ (the solution space)?” and
“2. What is the trade-of between fairness and relevance?” They define the
solution space as the set of all possible subsets $S\subset D$ in document
collection $D$ having cardinality $|S|=s$. They then proceed to investigate
these questions via simulation: generating different subsets of search results
and inspecting the empirical distributions of scores. In contrast, we suggest
a conceptual Pareto framing provides more direct and informed answers.
While we have emphasized generality of Pareto optimality under competing
objectives with constraints, Pareto optimality is of course also applicable to
simpler optimization problems, such as posed here. Firstly, since there are no
constraints on the solution space, the feasible set spans the full range of
$f_{r}\in[0,1]$ and $f_{g}\in[0,1]$. Secondly, since relevance and entropy are
independent, they do not compete: maximizing one does not preclude maximizing
the other, and each can be considered separately in turn. Relevance is
maximized when all search results are relevant ($f_{r}=1$), while entropy is
maximized when search results are evenly split across the two groups
($f_{g}=1$). This yields weak Pareto fronts at $f_{r}=1$ and $f_{g}=1$, with
the only strong Pareto solution at the intersection of both fronts
($f_{r}=f_{g}=1$), when search results are all relevant and are evenly split
across the two groups being represented.
With regard to the probability of observing any given ($f_{r},f_{g})$ score
for a given search result set $S$ (i.e., the chance of achieving the optimum
$f_{r}=f_{g}=1$ or any other feasible point), since relevance and entropy are
independent, their joint distribution is defined simply by the product of
probability distribution functions (PDFs) for relevance $P(r_{i})=p_{r}$ and
group membership $P(g_{i})=p_{g}$. In practice, we must induce these PDFs from
data, but this is standard estimation and not unique to this particular
problem setting. Moreover, this permits analytical analysis without
simulation.
Finally, whereas Gao and Shah [13]’s optimization problem is relatively easy
(no constraints or competing objectives), one could easily introduce further
complications. For example, imagine public opinion is highly skewed, such that
nearly all relevant information supports one side of the argument. In this
scenario, in order to get more balanced coverage of the minority position, we
would need to include more non-relevant search results, forcing the user to
sift through a larger result set to find relevant and balanced information. As
a second example, given controversy surrounding predictive policing policies,
one might like to constrain search results to enforce racial parity across
authors of retrieved documents in the results set. In either case, our Pareto
optimality framing would allow principled reasoning about the resulting
solution spaces.
### 7.2 Another Test Case for HNPF
As in Section 5, evaluating our HNPF approach on solutions with known
analytical forms allows us to verify its accuracy. As discussed above, we know
from first principles that the fair search problem considered here has weak
Pareto fronts at $f_{r}=1$ and $f_{g}=1$, with the only strong Pareto solution
at the intersection of both fronts ($f_{r}=f_{g}=1$). By running HNPF on this
problem, we can verify it induces these expected fronts as another check on
its correctness.
Gao and Shah [13] simulate $r_{i}$ and $g_{i}$ values based on observed
statistics of the YOW RSS Feed dataset [39]. They consider $n=48$ data points
drawn from this dataset, with empirical $\hat{p_{r}}=0.56$ and and
$\hat{p_{g}}=0.5$. Sampling from binomial distributions $R\sim
B_{r}(n=48,p_{r}=0.56)$ and $G\sim B_{g}(n=48,p_{r}=0.50)$, in expectation we
anticipate $R=(\sum^{n=48}r_{i})\approx 27$ relevant documents and an even
split of $G=(\sum^{n=48}g_{i})=24$ across groups. Note that whereas average
relevance $f_{r}$ is maximized when all documents are relevant, the entropy
$f_{g}$ is maximized when documents are evenly split by group. The probability
distributions above will thus naturally tend to yield near optimal entropy
$f_{g}\approx 1$ (with $p_{g}=0.5$) but only mediocre average relevance
$f_{g}\approx\overline{r_{i}}=0.56$. With $p_{r}=0.56$, the probability of
maximum relevance $P(f_{r}=1)=P(R=48)\approx 10^{-60}$.
(a) Function Domain
(b) Variable Domain
Figure 12: Weak Pareto Front for RSS dataset
Figure 12 shows the weak Pareto front for (a) the function domain
$(y=f_{r},x=f_{g})$ and (b) the variable domain $(x=R,y=G)$. As expected, we
see the sample distribution lays roughly symmetrically about the expectation
in both $R$ and $G$ in the variable domain, yielding near optimal entropy
$f_{g}$ and mediocre average relevance $f_{r}$ in the function domain. Given
this sample, the network correctly identifies the Pareto front for entropy at
$f_{g}=1$, corresponding to $G=24$ in the variable domain. Note that this is
still a weak front, where all points to the left are dominated by those to the
right, hence the need for the Pareto filter to identify the non-dominated set.
However, the network cannot identify the Pareto front for relevance at
$f_{r}=1$ due to sample sparsity relative to the sampling distribution. As
noted above, with $p_{r}=0.56$, $P(f_{r}=1)\approx 10^{-60}$, suggesting we
would need $10^{60}$ samples to observe $f_{r}=1$.
Our results above follow Gao and Shah [13] in sampling from the variable
domain according to ($p_{r},p_{g}$). This makes sense if we want to explore
the solution space via simulation, as they do. However, if our goal is
actually to identify Pareto fronts (e.g., to measure how far a given solution
is from optimality), we can instead probe the solution space far more
efficiently by uniformly sampling the variable domain ($r,g\in[0,1]$).
Appendix E presents these results.
## 8 Conclusion
A hybrid, two-stage, neural-Pareto filter based optimization framework is
presented for extracting the Pareto optimal solution set for multi-objective,
constrained optimization problems. The proposed method is computationally
efficient and scales well with increasing dimensionality of design variable
space, objective functions, and constraints. Results on verifiable benchmark
problems show that our Pareto solution set accuracy compares well with
existing state of the art NBI, mCHIM, and PK methods. The proposed neural
architecture is fully interpretable with a Fritz-John conditions inspired
discriminator for weak Pareto manifold classification. We also show that the
approximation error between the true and extracted Pareto manifold can be
easily verified for analytical solutions.
## References
* Balashankar et al. [2019] Ananth Balashankar, Alyssa Lees, Chris Welty, and Lakshminarayanan Subramanian. What is fair? exploring pareto-efficiency for fairness constrained classifiers. _arXiv preprint arXiv:1910.14120_ , 2019.
* Bengio et al. [2007] Yoshua Bengio, Pascal Lamblin, Dan Popovici, Hugo Larochelle, et al. Greedy layer-wise training of deep networks. _Advances in neural information processing systems_ , 19:153, 2007.
* Bentley [1990] Jon Louis Bentley. K-d trees for semidynamic point sets. In _Proceedings of the sixth annual symposium on Computational geometry_ , pages 187–197, 1990.
* Calandra et al. [2014] Roberto Calandra, Jan Peters, and MP Deisenrothy. Pareto front modeling for sensitivity analysis in multi-objective bayesian optimization. In _NIPS Workshop on Bayesian Optimization_ , volume 5, 2014.
* Caton and Haas [2020] Simon Caton and Christian Haas. Fairness in machine learning: A survey. _arXiv preprint arXiv:2010.04053_ , 2020.
* Chollet [2015] François Chollet. keras. https://github.com/fchollet/keras, 2015.
* Cohon [2004] Jared L Cohon. _Multiobjective programming and planning_ , volume 140. 2004\.
* Das and Dennis [1998] Indraneel Das and John E Dennis. Normal-boundary intersection: A new method for generating the pareto surface in nonlinear multicriteria optimization problems. _SIAM journal on optimization_ , 8(3):631–657, 1998.
* Deb et al. [2002] Kalyanmoy Deb, Amrit Pratap, Sameer Agarwal, and TAMT Meyarivan. A fast and elitist multiobjective genetic algorithm: Nsga-ii. _IEEE transactions on evolutionary computation_ , 6(2):182–197, 2002.
* Dutta and Kaya [2011] Joydeep Dutta and C Yalçın Kaya. A new scalarization and numerical method for constructing the weak pareto front of multi-objective optimization problems. _Optimization_ , 60(8-9):1091–1104, 2011.
* Fonseca and Fleming [1998] Carlos M Fonseca and Peter J Fleming. Multiobjective optimization and multiple constraint handling with evolutionary algorithms. ii. application example. _IEEE Transactions on systems, man, and cybernetics-Part A: Systems and humans_ , 28(1):38–47, 1998.
* Gal [1980] Tomas Gal. Multiple objective decision making-methods and applications: A state-of-the art survey: Ching-lai hwang and abu syed md. masud springer, berlin, 1979, xii+ 351 pages, dfl. 46.90. soft cover, 1980.
* Gao and Shah [2019] Ruoyuan Gao and Chirag Shah. How fair can we go: Detecting the boundaries of fairness optimization in information retrieval. In _Proceedings of the 2019 ACM SIGIR International Conference on Theory of Information Retrieval_ , 2019.
* Ghane-Kanafi and Khorram [2015] A Ghane-Kanafi and E Khorram. A new scalarization method for finding the efficient frontier in non-convex multi-objective problems. _Applied Mathematical Modelling_ , 39(23-24):7483–7498, 2015.
* Gobbi et al. [2015] Massimiliano Gobbi, F Levi, Gianpiero Mastinu, and Giorgio Previati. On the analytical derivation of the pareto-optimal set with applications to structural design. _Structural and Multidisciplinary Optimization_ , 51(3):645–657, 2015.
* Hernández-Lobato et al. [2016] Daniel Hernández-Lobato, Jose Hernandez-Lobato, Amar Shah, and Ryan Adams. Predictive entropy search for multi-objective bayesian optimization. In _International Conference on Machine Learning_ , pages 1492–1501, 2016.
* Khan et al. [2002] Nazan Khan, David E Goldberg, and Martin Pelikan. Multi-objective bayesian optimization algorithm. In _Proceedings of the 4th Annual Conference on Genetic and Evolutionary Computation_ , pages 684–684. Citeseer, 2002.
* Kingma and Ba [2014] Diederik P Kingma and Jimmy Ba. Adam: A method for stochastic optimization. _arXiv preprint arXiv:1412.6980_ , 2014.
* Levi and Gobbi [2006] Francesco Levi and Massimiliano Gobbi. An application of analytical multi-objective optimization to truss structures. In _11th AIAA/ISSMO multidisciplinary analysis and optimization conference_ , page 6975, 2006.
* Lin et al. [2019] Xiao Lin, Hongjie Chen, Changhua Pei, Fei Sun, Xuanji Xiao, Hanxiao Sun, Yongfeng Zhang, Wenwu Ou, and Peng Jiang. A pareto-efficient algorithm for multiple objective optimization in e-commerce recommendation. In _Proceedings of the 13th ACM Conference on Recommender Systems_ , pages 20–28, 2019.
* Lipton [2018] Zachary C Lipton. The mythos of model interpretability. _Queue_ , 16(3), 2018.
* Marler and Arora [2004] R Timothy Marler and Jasbir S Arora. Survey of multi-objective optimization methods for engineering. _Structural and multidisciplinary optimization_ , 26(6):369–395, 2004.
* Martinez et al. [2020] Natalia Martinez, Martin Bertran, and Guillermo Sapiro. Minimax pareto fairness: A multi objective perspective. In _Proceedings of the 37th International Conference on Machine Learning_ , 2020.
* Messac et al. [2003] Achille Messac, Amir Ismail-Yahaya, and Christopher A Mattson. The normalized normal constraint method for generating the pareto frontier. _Structural and multidisciplinary optimization_ , 25(2):86–98, 2003.
* Miettinen [2012] Kaisa Miettinen. _Nonlinear multiobjective optimization_ , volume 12. Springer Science & Business Media, 2012.
* Miriam et al. [2020] A Jemshia Miriam, R Saminathan, and S Chakaravarthi. Non-dominated sorting genetic algorithm (nsga-iii) for effective resource allocation in cloud. _Evolutionary Intelligence_ , pages 1–7, 2020.
* Mueller-Gritschneder et al. [2009] Daniel Mueller-Gritschneder, Helmut Graeb, and Ulf Schlichtmann. A successive approach to compute the bounded pareto front of practical multiobjective optimization problems. _SIAM Journal on Optimization_ , 20(2):915–934, 2009.
* Pareto [1906] Vilfredo Pareto. Manuale di economica politica, societa editrice libraria. milan. _English translation as Manual of Political Economy, Kelley, New York_ , 1906.
* Pirouz and Khorram [2016] Behzad Pirouz and Esmaile Khorram. A computational approach based on the $\varepsilon$-constraint method in multi-objective optimization problems. _Advances and Applications in Statistics_ , 49(6):453, 2016.
* Srinivas and Deb [1994] Nidamarthi Srinivas and Kalyanmoy Deb. Muiltiobjective optimization using nondominated sorting in genetic algorithms. _Evolutionary computation_ , 2(3):221–248, 1994\.
* Tanaka et al. [1995] Masahiro Tanaka, Hikaru Watanabe, Yasuyuki Furukawa, and Tetsuzo Tanino. Ga-based decision support system for multicriteria optimization. In _1995 IEEE International Conference on Systems, Man and Cybernetics. Intelligent Systems for the 21st Century_ , volume 2, pages 1556–1561. IEEE, 1995.
* Tapia and Coello [2007] Ma Guadalupe Castillo Tapia and Carlos A Coello Coello. Applications of multi-objective evolutionary algorithms in economics and finance: A survey. In _2007 IEEE Congress on Evolutionary Computation_ , pages 532–539. IEEE, 2007.
* Trisna et al. [2016] Trisna Trisna, Marimin Marimin, Yandra Arkeman, and T Sunarti. Multi-objective optimization for supply chain management problem: A literature review. _Decision Science Letters_ , 5(2):283–316, 2016\.
* Valdivia et al. [2020] Ana Valdivia, Javier Sánchez-Monedero, and Jorge Casillas. How fair can we go in machine learning? assessing the boundaries of fairness in decision trees. _arXiv preprint arXiv:2006.12399_ , 2020.
* Wei and Niethammer [2020] Susan Wei and Marc Niethammer. The fairness-accuracy pareto front. _arXiv preprint arXiv:2008.10797_ , 2020.
* Xiao et al. [2017] Lin Xiao, Zhang Min, Zhang Yongfeng, Gu Zhaoquan, Liu Yiqun, and Ma Shaoping. Fairness-aware group recommendation with pareto-efficiency. In _Proceedings of the Eleventh ACM Conference on Recommender Systems_ , pages 107–115, 2017.
* Zeleny [1973] Milan Zeleny. Compromise programming. _Multiple criteria decision making_ , 1973.
* Zhang et al. [2008] Qingfu Zhang, Aimin Zhou, Shizheng Zhao, Ponnuthurai Nagaratnam Suganthan, Wudong Liu, and Santosh Tiwari. Multiobjective optimization test instances for the cec 2009 special session and competition. 2008\.
* Zhang [2005] Yi Zhang. Bayesian graphical models for adaptive information filtering, 2005.
## Appendix A Additional Cases
### A.1 Case VI: n=2, k=2, m=5
This problem was proposed in [10]. Jointly minimize
$\displaystyle f_{1}(x_{1},x_{2})=x_{1}$ $\displaystyle
f_{2}(x_{1},x_{2})=x_{2}$ $\displaystyle\text{s.t.}\quad
g_{1}(x_{1},x_{2})=(x_{1}-0.5)^{2}+(x_{2}-0.5)^{2}\leq 0.5$ $\displaystyle
g_{2}(x_{1},x_{2})=x_{1}^{2}+x_{2}^{2}-1-0.1\cos(16\arctan(\frac{x_{1}}{x_{2}}))\geq
0$ $\displaystyle g_{3}(x_{1},x_{2})=max(|x_{1}-0.6|,|x_{2}-0.7|)-0.2\geq 0$
$\displaystyle g_{4},g_{5}:0\leq x_{1},x_{2}\leq\pi$
This form is an extension of Case III, with an additional max boundary
constraint $g_{3}$. Both mCHIM and PK computes the true Pareto front with
limited density $n=40$ points. Fig. 13 (a) shows the weak Pareto front with
dominated points. As before, after post-processing with the proposed Pareto
filter we arrive at the Pareto set with non-dominated points shown in Fig. 13
(b).
(a) Dominated
(b) Non-Dominated
(c) PK
(d) NBI
Figure 13: Strong Pareto Front for Case VI. Note that all the dominated points
are removed from the set after application of Pareto filter. Also note the low
density of points in PK and the even spread of points in NBI. Also note that
PK is not able to remove some of the dominated points on the horizontal and
vertical lines.
### A.2 Case VII: n=30, k=2, m=30
This problem was proposed in [29], albeit with a discrepancy222Although the
normalization term proposed in $f(x)$ is $m-1$, it does not generate the curve
reported in [29]. We were able to replicate the shown curve, in our
experiments, by choosing a normalizing constant of $10000$.. Jointly minimize
$\displaystyle f_{1}(x)=x_{1}$ $\displaystyle
f_{2}(x)=f(x)\left(1-\left(\frac{f_{1}(x)}{f(x)}\right)^{0.5}-\frac{f_{1}(x)}{f(x)}sin(10\pi
x_{1})\right)$ $\displaystyle\text{s.t.}\quad
f(x)=1+\frac{9}{m-1}\sum_{i=2}^{n}x_{i}^{2}$ $\displaystyle
g_{1},\ldots,g_{30}:x_{1}\in[0,1],x_{i}\in[-1,1],\forall i=2,\ldots,m$
This form is convex in $f_{1}$ and non-convex in $f_{2}$. The dimension of the
design variable space is $m=30$. The corresponding Pareto front is non-convex.
The results using our method, as shown in Fig. 14, are in good agreement with
mCHIM and PK methods. Even in this high-dimensional setting, we obtain a weak
Pareto front with high point density as shown in Fig. 14 (a). The distribution
of the objective space in this setting is such that the entire space is the
front itself. Hence, we cannot see the cyan points in Fig. 14 (a). As before,
after post-processing with the proposed Pareto filter we arrive at the Pareto
set with non-dominated points shown in Fig. 14 (b), where the dominated points
are now visible.
(a) Dominated
(b) Non Dominated
(c) PK
(d) NBI
Figure 14: Pareto Front for Case VII. Note that all the dominated points are
removed from the set after application of Pareto filter. Also note the low
density of points in PK and the even spread of points in NBI.
## Appendix B Working of Pareto filter
The algorithm starts with the set of all weak Pareto points $p\in P$, which
will be refined through the iterative process. The loop (line 4) iterates over
all the functions $f_{i}$. It checks for set of dominating and non-dominating
points for all discretization levels (line 6) and appends them to a temporary
list (line 10). If multiple points do exist (line 11) for a given level
(cardinality $>1$), then there certainly are dominated points. The non-
dominated point is one which has the lowest function value for the next
function $f_{q},q=i+1$. This (line 12) states that a point which seems non-
dominated for a given function $f_{i}$ might be dominated for other functions
$f_{r},r\in k,r\neq i$, but will be taken care of when iterating through
function $f_{r}$. Once the non-dominated point has been found, it implies that
all the other points are in fact dominated, hence should not be considered for
further evaluation. They are rejected (line 13) from the active set
$P=P\backslash(temp\backslash x_{p})$. The output is the set of strong Pareto
points which were all non-dominated for every function $f_{i}$, and are
essentially the points in the weak Pareto set $P$ that survived the filtering
process (line 13) for every $f_{i}$.
## Appendix C Error Bound
For a user-prescribed relaxation margin $0\leq\epsilon\leq 1$, the
approximation error between the network extracted manifold
$\tilde{M}(\tilde{X})$ and the true solution $M(X^{*})$ is bounded below by
$\|\tilde{M}(\tilde{X})-M(X^{*})\|_{2}\leq\epsilon$. Assuming the $L$ matrix
from Eq. (6) is square, we have $det(L)=0$. This form holds for some of the
problems chosen in our numerical experiments (Cases I, II, III), where the
number of functions and constraints are equal. From Leibnitz formula for
determinants:
$\displaystyle det(L)=det\Big{(}\begin{bmatrix}\nabla F&\nabla G\\\
\mathbf{0}&G\end{bmatrix}\Big{)}=det(\nabla F)det(G)=0$
Further assume that,
$\displaystyle|det(L(\tilde{x}))|\leq\epsilon,\,\tilde{x}\neq x^{*}$ (8)
where $\epsilon>0$, and $x^{*}$ and $\tilde{x}$ are the optimal points and the
network generated approximate solution points, respectively. The Fritz John
necessary conditions in Eq. (4) for weak Pareto optimality is:
$\displaystyle det(L(x^{*}))=0$ (9)
Combining the assumption in Eq. (8) and Eq. (9), we have
$\displaystyle|det(L(\tilde{x}))-det(L(x^{*}))|\leq\epsilon$ (10)
The solution manifold $M(X)=det(L(X))=0$ is weakly Pareto optimal. We assume a
low precision manifold $M(\tilde{x})$ such that:
$\displaystyle\|M(\tilde{X})-M(X^{*})\|_{2}\leq\epsilon$ (11)
When the network converges, Eq. (11) will hold for the network approximated
$\tilde{M}(x)$. Here, $\tilde{x}\in\tilde{X}=\\{x|\tilde{M}(x)=0\\}$ and
$X^{*}$ is the set of true optimal points such that $M(x^{*})=0,\forall x\in
X^{*}$. Since we explicitly specify $\epsilon$ in our loss description, we
know that the network generated solution is $\epsilon$ close to $M(\tilde{x})$
if:
$\displaystyle\|M(\tilde{X})-\tilde{M}(\tilde{X})\|_{2}\leq C\epsilon,\quad
0\leq C\leq 1$ (12)
The form in Eq. (12) implies that if we are able to find such a $C$, then we
implicitly satisfy Eq. (11). Hence,
$\displaystyle\|\tilde{M}(\tilde{X})-M(X^{*})\|_{2}\leq\epsilon$ (13)
## Appendix D Density Comparison
| HNPF | PK
---|---|---
Case | Density | Points | Evals | Density | Points | Evals
Case V | 1.34 | 1206 | 90K | 2.21e-4 | 100 | 45,126,324
Case VI | 0.20 | 180 | 90K | 1.32e-2 | 151 | 1,139,781
Case VII | 6.57 | 5915 | 90K | 9.29e-5 | 101 | 108,685,605
Table 3: Pareto optimal point density % (ratio of #extracted optimal points to
#function evaluations). HNPF finds many more optimal points with many fewer
function evaluations.
## Appendix E Verification Case for Fairness
We now demonstrate the nature of the extracted Pareto front under a uniform
distribution for the problem setting in Section 7.2. While the scenario
presented in Gao and Shah [13] considered points being drawn from a Binomial
distribution, it is highly improbable to reach a relevance value of 1. We
therefore consider a unbiased uniform distribution, where we sample integers
uniformly between $[0,48]\times[0,48]$ for both relevance ($R$) and entropy
($G$).
(a) Function Domain
(b) Variable Domain
Figure 15: Weak Pareto Front for RSS dataset under a uniform setting. Note
that now both min-min and the max-max solutions are achievable, as these
points are visible to the network.
In Fig. 15 we show the extracted weak Pareto front from the HNPF neural
network. Under the uniform data distribution, the maximum value of achievable
entropy corresponds to $G=24$. Similarly, the minimum achievable entropy value
corresponds to both ($G=0,48$) which indicates that all documents belong to
only one source. The neural network extracts the entire max-max weak Pareto
front which when post-processed using the Pareto filter, results in the
expected $(1,1)$ solution in the function domain. Note that the Stage-1 of
HNPF is capable of extracting the entire weak Pareto front for any MOO
problem. The middle red line in the variable domain corresponds to the top red
line in the function domain. The two remaining red lines (top and bottom) in
the variable domain collapses to the bottom in the function domain showing the
weak Pareto front for a min-min MOO problem. Applying the Stage-2 Pareto
filter now gives $(0,0)$ as the optimal min-min solution.
|
Data Analysis and Management (Computer Science and Engineering)
June 2020 May, 2020
Victor LempitskyAssociate Professor, PhD Yury MalkovPhD
TODOTODO, Skoltech
# CNN with large memory layers
Rasul Karimov
This work is centred around the recently proposed product key memory structure
[46], implemented for a number of computer vision applications. The memory
structure can be regarded as a simple computation primitive suitable to be
augmented to nearly all neural network architectures. The memory block allows
implementing sparse access to memory with square root complexity scaling with
respect to the memory capacity. The latter scaling is possible due to the
incorporation of Cartesian product space decomposition of the key space for
the nearest neighbour search. We have tested the memory layer on the
classification, image reconstruction and relocalization problems and found
that for some of those, the memory layers can provide significant
speed/accuracy improvement with the high utilization of the key-value
elements, while others require more careful fine-tuning and suffer from dying
keys. To tackle the later problem we have introduced a simple technique of
memory re-initialization which helps us to eliminate unused key-value pairs
from the memory and engage them in training again. We have conducted various
experiments and got improvements in speed and accuracy for classification and
PoseNet relocalization models.
We showed that the re-initialization has a huge impact on a toy example of
randomly labeled data and observed some gains in performance on the image
classification task. We have also demonstrated the generalization property
perseverance of the large memory layers on the relocalization problem, while
observing the spatial correlations between the images and the selected memory
cells.
### Acknowledgments
We wish to express our deep sense of gratitude and profound thanks to Karim
Iskakov and all the engineers in Samsung AI Moscow who contributed to the
project in one way or another. We are hugely indebted to Samsung Research
Center for the provided resources that gave us the chance to implement the
models and conduct the required experiments.
###### Contents
1. 1 Introduction
2. 2 Related work
3. 3 Methods
1. 3.1 Product key memory
1. 3.1.1 Design overview
2. 3.2 Re-initialization trick
1. 3.2.1 Overview
2. 3.2.2 Problem of dying keys
3. 3.2.3 Key re-sampling
4. 3.2.4 Re-sorting and key-value reinititalization
5. 3.2.5 Re-initialization complexity
3. 3.3 Classification pipeline
4. 3.4 Regression pipeline
5. 3.5 Image reconstruction pipeline
4. 4 Experiments
1. 4.1 Experiments on random labels
1. 4.1.1 Results on CIFAR-10
2. 4.1.2 Experiments on ImageNet
2. 4.2 Memory in PoseNet
1. 4.2.1 Analysis of the memory layer
3. 4.3 Image reconstruction experiments
4. 4.4 Other experiments
5. 5 Conclusion
###### List of Figures
1. 3.1 Overview of the product key-value memory design. Feature generator is the baseline image model which outputs the latent vector that is projected by the the query generator, normalized and divided into the two sub queries $q_{1}$/$q_{2}$ to work with product key values. As the result we have k selected keys and the corresponding values from $\mathcal{V}$.
2. 3.2 The illustration of the key re-initalization procedure. (a) We have 5 keys with three of which does not pass the threshold of being identified as the ”utilized keys”. We are removing those keys in the step (b). And in the step (c) we are initializing new keys sampled from existing ones and perturbed with an additive Gaussian noise ($\epsilon\sim\mathcal{N}(0,\sigma_{d}^{2})$).
3. 3.3 Bottleneck [30] augmented with a memory block $\mathcal{M}$.
4. 3.4 The overview of the modified squeeze-and-excitation block augmented with the memory layer. $F_{GB}$ is the function of global pooling which reduces the dimension of 3 dimensional feature map to the signle dimension, and $F_{add}$ is channel-wise addition of a resulting vector from memory to the original feature tensor.
5. 4.1 Results for the random data on the various models. The left graph is the plot for top-1 validation accuracy results, top-5 is pictured on the right. We can see that the model with no memory is not able to fit the data. Setting 8 heads for multi-head attention memory model, on the other hand, helps the model to easily fit the data. Re-initialization helps to get nearly full convergence with on only 1 head, i.e. multi-head mode disabled. Also, $h=1$ in the graph is the notation for the single head model.
6. 4.2 Relation between $\sigma_{d}$ hyper parameter of the Gaussian noise square variance parameter in re-initialization procedure and the values of train accuracy and loss for the random data. Setting higher $\sigma_{d}$ helps significantly, giving lower loss.
7. 4.3 Comparison of the top-1 scores, memory utilization and the inference speed for the model with no memory augmentation and three memory augmented models with the number of heads in the set of {1, 4, 8}. There are 20 runs for the each experiment with the different initial seeds. As it can be observed, there is no evident increase in the accuracy while the performance of the models with the heads is much worse compared to the baseline models. Inference values are calculated on GTX-1080Ti cards with fp32 mode, the results are approximate.
8. 4.4 We picture the key access distribution for each head. As we see there is a little perturbation in the distribution when we apply the random region crop in the input image.
9. 4.5 Change in values of the standard deviation of keys, memory usage, gradient of memory values, and key gradients during a training phase.
10. 4.6 The camera path and the four sample images captures in the given coordinates. We see the correlation between the traveled distance and the indices selected.
11. 4.7 Reconstruction results. Top row images in each section are the reconstruction results with the memory augmented autoencoder with mem_idx=2 and heads=8, the middle are the output of the baseline autoencoder while the final row are the input images. We see that there are some little details that are captured using the memory layer.
###### List of Tables
1. 4.1 Results for the modified SE blocks. WL is the table notation for the wide linear layer that replaces the memory layer, $d_{w}$ defines respectably the row and column of the projections matrices in MLP (with the row vector in the linear operator). Overall we see better results on Resnet-20 with the memory layer and with re-initialization trick we have superior memory utilization rate. Cosine similarity helps us to nearly reach the accuracy values of ResNet-100. Inference values are calculated on GTX-1080Ti cards with fp32 mode, the results are approximate.
2. 4.2 The results on the Resnet-50 and Imagenet. We have tested a number of the hyper parameters to find the best train strategy for the memory models. For now we dont see the clear picture on optimization issues.
3. 4.3 Results on the King’s College dataset, $\mathrm{trans}$ and $\mathrm{rot}$ are the translation and rotation errors respectively. We see a huge decrease in the loss on the train set for the networks with memory augmentation but also the increase in the inference time. Inference values calculated on the GTX-1080Ti with the batch size of 1.
4. 4.4 Reconstruction results after 1600 iterations. The memory utilization numbers are approximate since the amount of experiments conducted was low. The inference speed is calculated on GTX-1080Ti with batch size of 1.
## Chapter 1 Introduction
With the huge development of deep learning, neural networks have made
significant progress in various tasks such as image classification [44],
speech recognition [26], and machine translation [74]. As it was shown in [72,
54], with sufficiently large data, increasing the capacity of neural networks
could lead to far superior prediction accuracy. Therefore, scaling of both
training and model size has been a central problem of deep learning field in
recent years. For a typical neural model where the single input data sample
depends on all the parameters of the network, therefore increasing both
dataset and model sizes leads to a nearly quadratic surge in the training
costs [69].
Previous works [15, 12] proposed several ideas of increasing the model size
with no proportional increase in a computational complexity. Most of the
proposed models rely on the sparse gating mechanism which determines whenever
a particular node of the graph should be calculated during the forward pass of
the network. This is the type of a branching problem with a discrete decision
which is being solved with REINFORCE algorithm [75]. It was applied in [8] and
gave good results on MNIST dataset and CIFAR-10 [43] with a reasonable speed
up.
Instead of REINFORCE estimator, one can also apply the ideas from [49, 34] by
relaxing the discrete skipping decisions with reparametrization technique
adoption. However, these approaches usually find sub-optimal solutions due to
the approximation error introduced by the reparametrization trick [82, 1].
Other approaches rely on learning binary masks as the sparse $l_{0}$
regularization term for the final objective. Works like [53] employ a
rectified sigmoid proposed in [48] to learn binary decision choices. Authors
apply regularization during post-processing to quantize the weights of the
network, but the idea could be used in the training phase too. Recently, the
paper [58] on differentiable tabular data with neural networks has leveraged
the entmax transformation [55] to learn ”hard” decisions for binary branching
problem in decision trees.
Though solving the issue of scalability, models still fall short in giving
promising results due to the following challenges:
* •
GPUs are optimized to work faster with arithmetic tasks rather than branching.
* •
Batching reduces the batch sizes for conditionally activated chunks, therefore
complicating parallelization.
* •
REINFORCE estimator has a large variance, making it hard to get a strong bias
during train. There are some variance reduction tricks [28, 51] that try to
solve the issue but most of them skew the bias-variance trade-off with no
tuning on hyperparameters applied.
* •
Nearly all the methods suffer from neuron dying problem - if at some moment of
the training a gate is not open for any input sample, this means it is highly
unlikely it will be open at any further moment since the gate receives only
zero gradient.
A recent work [46] on the over-parametrized language models, on the other
hand, rely on the key-value memory structures to scale the set of parameters
in the neural network. Authors rely on the product key space decomposition for
nearest neighbours to scale the networks with a little or no change in the
performance and the memory consumption. These results encouraged us to
research these methods in the computer vision applications.
As an extension of [46], we augment the product key layer with the key-value
re-initialization mechanism which allows to solve the dying neuron problem.
The mechanism is based on re-initialization of dead or underutilized keys-
values pairs using the information from more successful key-values pairs.
## Chapter 2 Related work
Memory layers in neural models Memory augmented neural networks (MANNs)
augment neural networks with external memory block which allows for explicit
access with differentiable write and read operations. The memory is usually
implemented as a key-value differentiable dictionary [59], array-structured
memory such as Turing machine (NTM) [27], or recently proposed product-key
memory layers [46].
Key-value memory architectures were analyzed extensively in deep learning
literature. Mostly the models are based on the architectural designs described
in [73, 78] and used mainly in natural language processing field (NLP) such as
document reading and question answering tasks. The key-value paired structure
can be seen as a generalization of how the huge context for question answering
is stored in the memory. That makes key-value memory a natural choice for
these tasks. And with the recent advancements in attention models (AM) [5,
80], it is becoming the predominant concept in natural language processing
literature with the usage case in nearly every NLP task.
The key-value design structure is a sparse version of attention models, and as
previously described in [14, 7] the self-attention mechanism could be
competitive in replacing convlolutions as a computation primitive for object
detection and image classification. We hope to leverage and improve upon those
ideas for computer vision applications.
There were also some works in extending the key-value structure, [25] using
unbound cache model to provide better modelling of infrequent words that occur
in a recent context, including the out-of-vocabulary words. This is the
evidence of interpretability of learned key-value structure which provides the
linear interpolation between the values of the table. Other works [39] focus
on the interpretability of memory blocks by linearly interpolating baseline
language models (LM) with k-nearest neighbours (k-NN) models and assessing the
generalization and memorization capability of the final models. Authors report
increased perplexity in all of the experiments.
Other approaches have successfully applied memory layers to store image
features, more specifically for image captioning [13], image generation tasks
[41], and video summarization [47].
Some neural network designs include non-differentiable memory layers with
unsupervised updates. These models mostly rely on the architectural ideas of
[89] and rely on contrastive loss to update the memory block. Authors of [89]
have demonstrated the the efficiency of their memory block in few-shot and
one-shot learning tasks, while [84] has shown the advantage of using the
memory in style transfer [20] tasks with limited data. While in supervised
approaches of memory usage where we are learning the mapping function between
two spaces, in the unsupervised approach memory block is used for storing
latent vectors with the ability to interpolate between them. This is the
important property of memory blocks that is implicitly used in most of the
models.
Some works incorporate memory-like structure to store a bank of weights
accessing them sparsely and using k-nearest neighbours to retrieve a set of
indices that allows to encode the feature vector into the discrete code. There
are some promising results in auto-regressive models [79] giving high-fidelity
image reconstruction results in [63]. Authors argue that the discrete
representation is a more natural fit for complex reasoning, planning and
predictive learning.
Moreover, memory layers were successfully incorporated in graph knowledge
representation problems, with promising results in graph classification tasks
[40].
Learning compact representations, ANN. Since the task of exact top-k neighbour
search is expensive in practice for high dimensional embeddings, i.e. linear
scale to the search set size, practitioners in the field usually resort to
more scalable approximate nearest neighbours methods (ANN). Popular methods
include Locality sensitive hashing (LSH) [22, 2] that relies on random space
partitioning, graph methods like Navigable small world graphs (NSW) [50] and
Hierarchical NSW (HNSW) [6] based on constructing the search algorithm by
finding the clusters of data. Another important subset of methods utilize
quantization [21, 18, 4] to reduce the memory consumption and speed up the
distance calculations. Many of those methods exploit the idea of product
decomposition, e.g. assumption that the space can be decomposed into a
Cartesian product of low dimensional spaces.
Product decomposition in neural models Most of this thesis is inspired by the
work of [46] which are showing the efficiency of using product key memory
layer in language modelling tasks. Here product key is a structure that allows
more efficient scaling by scarifying some expressiveness of the model. Authors
find that the language model augmented with memory with only 12 layers can
outperform in accuracy a baseline transformer model with 24 layers while
giving two times less inference time. The [46] didn’t however address the
problem of dying keys other than by adding noise to the query (via batchnorm)
and was focused solely on NLP applications.
Product quantization in general has also been used in many computer vision
applications, starting from scalable supervised [56] to semi-supervised [35]
image retrieval tasks. There are some promising results [83] in face video
generation with memory augmented Generative adversarial networks (GAN) [23].
Classification networks Huge chunk of work [76, 30, 45, 70] is done in
designing the neural networks for the image classification problems. In our
experiments we mainly focus on the ResNet-like [30] networks. Some recent work
[77] demonstrated the SOTA results in ImageNet-2012 dataset [44] with the help
of the reinforcement learning to tune the models and the whole architecture.
Most of the existing neural networks for image classification rely on the
convolutional block but there were some recent works suggesting the self-
attention mechanism with promising results [7, 62].
## Chapter 3 Methods
### 3.1 Product key memory
#### 3.1.1 Design overview
The overall pipeline of the differentiable product memory layer is similar to
most of the key-value data structures that are augmented into neural network
models [27, 78, 73]. More specifically, product memory design in our work is
heavily inspired by previously proposed architecture in [46]. Here we build
models upon this design to solve classification, regression, and
reconstruction computer vision tasks.
Higher view of the architecture is illustrated in Figure 3.1. The central idea
is to augment baseline convolutional neural networks with sparse memory
access. The input of the memory layer is the latent feature vector that
describes the given input image. Depending on where we place the memory layer,
the query can represent features like brightness gradients or colours with
more complex patterns in later layers [87]. Therefore, the choice of memory
access placement is important. Given the input query, memory block finds the
distance scores by comparing it with all of the keys in the table and
selecting the values associated with top-k distance scores. The scores are
then used to produce the output $m(x)$ via a weighted sum over the values
associated with the selected keys:
$m(x)=\sum_{i\in\mathcal{I}}w_{i}v_{i}$
where $\mathcal{I}$ is the set of top-k indices by distance scores, $w_{i}$
are the scores, and $v_{i}$ are the values in the memory layer.
Query generation. The memory block consists of the query generation network
which is a learnable projection function $q:x\rightarrow
q(x)\in\mathbb{R}^{d_{q}}$ mapping the d-dimensional input vector x into the
$d_{q}$-dimensional query vector. Typical dimension sizes of the query vectors
in our experiments are from 256 up to 2048.
Also, since the keys are initialized in a fixed range, we follow [46] adding
BatchNorm [33] layer on the top of the query network. This allows a better
overlap between the distribution of keys and queries. And as in [46] we
observe higher utilization of the memory with BatchNorm enabled.
Key assignment We have a resulting query $q(x)$ that should be assigned to the
closest keys with the selected metric for the distance. Let
$\mathcal{K}=\\{k_{1},\ldots,k_{|\mathcal{K}|}\\}$ is the set of all keys, the
set is composed of $|\mathcal{K}|$ $d_{q}$-dimensional vectors that are
uniformly initialized in the $\mathbb{R}^{d_{q}}$ space. We can define the
differentiable procedure to find a weighted sum over the value vectors (the
memories associated with top-k keys). The sum is weighted by the distance
scores between the subset of the top-k keys and the query value $q(x)$. Top-k
procedure finds the most closest keys to the given query, i.e. maximization of
the chosen similarity measure $d(\cdot,\cdot)$. The overall algorithm is:
$\displaystyle\mathcal{I}=\mathcal{T}_{k}\left(d(q(x),k_{i})\right)$
$\displaystyle
w=\operatorname{softmax}\left(\left\\{d(q(x),k_{i})\right\\}_{i\in\mathcal{I}}\right)$
$\displaystyle m(x)=\sum_{i\in\mathcal{I}}w_{i}v_{i}$
where $\mathcal{T}_{k}$ denotes the top-k operation which finds k largest
values based on the similarity measure $d(\cdot,\cdot)$. $\mathcal{I}$ denotes
the index set of the most similar keys to the query $q(x)$ and $w$ represents
the normalized scores associated with the selected indices. The resulting
value $m(x)$ is the sum of the selected values weighted by the normalized
scores. As we see due to the summation over the normalized values in the
$\operatorname{softmax}$ operation, the gradients can be calculated. Note that
it is not possible to find the gradient for top-1 function, since there is no
summation.
Figure 3.1: Overview of the product key-value memory design. Feature generator
is the baseline image model which outputs the latent vector that is projected
by the the query generator, normalized and divided into the two sub queries
$q_{1}$/$q_{2}$ to work with product key values. As the result we have k
selected keys and the corresponding values from $\mathcal{V}$.
Product keys We see that the bottleneck of the given procedure is the
calculation of the $\mathcal{T}_{k}$ the operation which has linear complexity
over the size of the key set $\mathcal{K}$, so it is hard to scale this
procedure for large memory sizes. The remainder of operations are done for the
reduced set of selected indices, e.g. the summation over top-k normalized
weight values.
To solve the performance issue, authors of [46] propose to represents the key
set in the form of two independent sets of half dimension size $d_{q}/2$
vector sets $\mathcal{K}_{1}$ and $\mathcal{K}_{2}$ which constructs the
Cartesian product set of resulting values with size
$|\mathcal{K}|=|\mathcal{K}_{1}|\times|\mathcal{K}_{2}|$. The query vector
should also splitted into two sub-queries $q_{1}$ and $q_{2}$ to work in each
of the key sets. We then find the closest keys in both sets as:
$\mathcal{I}_{\mathcal{K}_{1}}=\mathcal{T}_{k}\left(\left\\{d(q_{1}(x),k_{i}^{1})\right\\}_{i\in\\{1\ldots|\mathcal{K}_{1}|\\}}\right),\quad\mathcal{I}_{\mathcal{K}_{2}}=\mathcal{T}_{k}\left(\left\\{d(q_{2}(x),k_{j}^{(2)})\right\\}_{j\in\left\\{1\ldots\left|\mathcal{K}_{2}\right|\right\\}}\right)$
Then the two subsets of the keys associated with the index sets
$\mathcal{I}_{\mathcal{K}_{1}}$ and $\mathcal{I}_{\mathcal{K}_{2}}$ are
multiplied together to form a new Cartesian product set. We are applying the
top-k operation on the newly created set and find the final subset of the
top-k keys.
Choice of distance Authors in [46] experiment with the inner product as the
single similarity measure for the provided experiments. We observe that using
cosine similarity not only provides us with better numbers in some experiments
but also gives us control over the selection process of the keys. Since the
dot product is proportional to the vector norm, the key vectors with the
largest vector lengths will be selected in most of the cases, while low norm
vectors may be completely ignored. This means that the distance measure
captures the most popular candidates, the latter can skew the similarity
metric. We balance the skew by introducing the hyperparameter $\alpha$ and
raising the length to an exponent $\alpha<1$ to calculate the distance as:
$d_{cos}(q,k,\alpha)=|q|^{\alpha}|k|^{\alpha}\cos(\theta)=|q|^{\alpha}|k|^{\alpha}\frac{q^{T}k}{|q|\cdot|k|}$
(3.1)
Multi-head mode To make the model more expressive we are using the multi-head
mechanism [80] which splits queries and keys into multiple chunks of vectors
to be calculated independently. The similar calculations are conducted on each
head and the results are concatenated. Due to the reduced dimension of each
head, the overall computational complexity of the layer is similar to the
single-head attention.
Complexity Naive top-K key selection requires $\mathcal{K}$ comparisons of
$d_{q}$ sized vectors, which gives us $\mathcal{O}\left(|\mathcal{K}|\times
d_{\mathrm{q}}\right)$ operations overall. When using the product space
$\mathcal{K}=K_{1}\times K_{2}$, we have two separate sets of keys for
subspaces with significantly reduced carnality
$|K_{1}|=|K_{2}|=\sqrt{|\mathcal{K}|}$. The overall complexity for the first
step then is: $\mathcal{O}\left(|K_{1}|\times d_{q}/2+|K_{2}|\times
d_{q}/2\right)=\mathcal{O}\left(|K_{1}|d_{q}\right)=\mathcal{O}\left(\sqrt{\mathcal{|K|}}d_{q}\right)$.
The second step is performed on the reduced subset of $k\times k$ elements so
it will require $\mathcal{O}\left(k^{2}\times d_{1}\right)$ operations.
Therefore overall complexity is:
$\mathcal{O}\left(\left(\sqrt{|\mathcal{K}|}+k^{2}\right)\times
d_{\mathrm{q}}\right)$
### 3.2 Re-initialization trick
#### 3.2.1 Overview
While conducting our initial experiments on random data, we have observed that
a toy neural network augmented with memory block struggles to fit the data
with no multi-head mode enabled even though the model should have had enough
capacity to fit the whole dataset. By conducting some ablation study and
literature review [3] we have concluded that the problem is due to the correct
initialization of the memory layer. Additionally, authors in [81] suggest that
most of the heads in the attention mechanism can be pruned without serious
effect on the performance. To tackle the initialization issues we are
introducing the re-initialization trick that dynamically initializes unused
keys during the training phase. We are describing the whole procedure below.
Figure 3.2: The illustration of the key re-initalization procedure. (a) We
have 5 keys with three of which does not pass the threshold of being
identified as the ”utilized keys”. We are removing those keys in the step (b).
And in the step (c) we are initializing new keys sampled from existing ones
and perturbed with an additive Gaussian noise
($\epsilon\sim\mathcal{N}(0,\sigma_{d}^{2})$).
#### 3.2.2 Problem of dying keys
Let’s assume that we are working with the dataset of size $|\mathcal{D}|$
which is equal to the number of values in the memory $|\mathcal{M}|$. We could
assume that augmenting the neural network with the memory layer could lead to
the full convergence, i.e. perfect accuracy, because of the one-to-one mapping
between the input and the memory elements. We hovewer did not observe this in
our experiments with random data (description is in the experiments section),
and classification tasks. Instead, we discover continuously reduced
cardinality of the selected key set at each iteration of the optimization,
reaching some fixed value $|\mathcal{K}^{\prime}|$:
$\displaystyle|\mathcal{K}^{\prime}|=\alpha|\mathcal{K}|,\,\alpha\ll 1$ (3.2)
$\displaystyle\mathcal{K}^{\prime}=\left\\{k_{i}\in\mathcal{K}|c_{i}>0,c_{i}\in\mathcal{C}\right\\}$
(3.3)
where $K^{\prime}$ is the set of selected keys during the inference,
$\mathcal{K}$ is the set of all keys, and $c_{i}$ is the utilization of the
key $k_{i}\in\mathcal{K}$ summed for the whole dataset, i.e. number of times
the key $k_{i}$ was selected. In the experiments we are not able to get full
utilization of the selected keys and observed low final accuracy. We call this
a problem of dying keys, when the optimizer is unable to pass gradients
through certain key-value pair in the memory layer, leading us to the dead
keys, useless for inference but still having a computational burden.
#### 3.2.3 Key re-sampling
To solve the problem we implement a simple trick of key re-initialization,
which is being executed during the training phase at certain points. We
observe that during training, the key utilization converges to some specific
number, as it is given in Equation (3.2). We assume that the main reason for
this is dying keys problem discussed in the previous section. For this reason,
we are running the pipeline of key re-initialization when the utilization
plateau is reached.
Here we describe the algorithm for a single product space key subset but the
algorithm is applicable for both of key subsets. Let $\mathcal{K}$ define the
set of all the keys in memory and $\mathcal{K^{\prime}}$ is the subset of
utilized keys where $|\mathcal{K^{\prime}}|\ll|\mathcal{K}|$. We also
introduce the hyperparameter $k_{a}$ which will control how many keys should
be re-sampled at the each call of our key initialization procedure. Then we
have:
$\displaystyle\mathcal{I}_{a}=\left\\{i|i\sim
U\\{0,|\mathcal{K}^{\prime}|\\}\right\\}$
$\displaystyle\mathcal{K}_{a}=\left\\{k_{i}\in\mathcal{K}^{\prime}|i\in\mathcal{I}_{a}\right\\}+\epsilon,\epsilon\sim\mathcal{N}(0,\sigma_{n}^{2})$
$\displaystyle\mathcal{K}=\mathcal{K}_{a}\cup\mathcal{K}^{\prime}$
where $\mathcal{I}_{a}$ is the set of indices sampled uniformly from the used
keys $\mathcal{K}^{\prime}$, $\mathcal{K}_{a}$ is the sampled set of utilized
keys perturbed with guassian additive noise. We have an additional hyper
parameter $\sigma_{n}$ that controls the noise variance of the re-initialized
keys, i.e. the magnitude of difference between the original keys and re-
initialized ones. Then the existing set of utilized keys are expanded by
$\mathcal{K}_{a}$. The sampling mechanism we discussed above is very basic,
but sampling more from the regions of high density/low density could
potentially bring us more gain both in prediction accuracy and the compactness
of the final representation. This, however, requires the re-initialization
algorithm to be able to sample key points in the regions with higher density.
Something like rejection sampling algorithms, i.e. Metroplolis-Hastings
algorithm [66] could save us here, by defining the multimodal normal
distribution and the utilization of the key values as the mean parameter. But
because of the difficulty of tuning the rejection sampling algorithm, we plan
to test those algorithm in the future and resorting to simple re-sorting
discussed in the following section.
#### 3.2.4 Re-sorting and key-value reinititalization
To give more priority on the regions of high density during sampling, we are
sorting the keys in the set by the utilization coefficients
$c_{i}\in\mathcal{C}$, and adapting a naive thresholding logic to eliminate
the least utilized keys by removing those with the values less than the hyper
parameter $d_{k}$ Then the index set calculated is:
$\displaystyle\mathcal{K}_{d_{k}}=\left\\{k_{i}\geq
d_{k}|k_{i}\in\mathcal{K}\right\\}$
$\displaystyle\mathcal{I}_{a}=\left\\{i\in\\{0..|K|\\}|k_{i}\in\mathcal{K}_{d_{k}}\right\\}$
After we resample the keys by eliminating the least utilized, we to initialize
new values $\mathcal{U}_{a}$ that will be mapped to the elements of the
Cartesian product set of the new keys $\mathcal{K}_{a}$. Because of the set
product, adding single key to the subset will add
$|\mathcal{K}_{a}||\mathcal{K}|$ new values into the memory. For each key from
the first product set, we are initializing new values associated with the
resulting keys concatenated with the given key from the first set and all the
existing keys in the second set. The same applies to the values associated
with the second product set. The overall algorithm for the re-sampling step is
demonstrated in Algo 1.
Algorithm 1 Re-initialization algorithm
1:existing set of keys $\mathcal{K}^{\\{0,1\\}}$, noise variance $\sigma_{n}$,
utilization set $\mathcal{C}^{\\{0,1\\}}$, threshold parameter $d_{k}$, and
the weight values $\mathcal{U}$
2:Reinitialized set of keys $\mathcal{K}$, Reinitialized values $\mathcal{U}$
3:function keysort($\mathcal{K},d_{k}$)
4: $\mathcal{K}_{d_{k}}=\left\\{k_{i}\geq d_{k}|k_{i}\in\mathcal{K}\right\\}$
5: return $\mathcal{K}_{d_{k}}$
6:end function
7:for $j\in\\{0,1\\}$ do
8: $\mathcal{K}_{d_{k}}^{j}=\operatorname{keysort}(\mathcal{K}^{j},d_{k})$
9:
$\mathcal{U}=\mathcal{U}\setminus\left\\{\mathcal{U}_{i}|i\in\left(\mathcal{K}^{j}\setminus
K_{d_{k}}^{j}\right)\times\mathcal{K}^{(-j)}\right\\}$
10: $\mathcal{I}_{a}^{j}\sim
U\left\\{i\in\\{0..|\mathcal{K}^{j}|\\}|k_{i}\geq\mathcal{K}^{j}_{d_{k}}\right\\}$
sample $a$ indices from the discrete distribution
11:
$\mathcal{K}_{a}^{j}=\left\\{k_{i}\in\mathcal{K}^{j}|i\in\mathcal{I}_{a}^{j}\right\\}+\epsilon,\epsilon\sim\mathcal{N}(0,\sigma_{n}^{2})$
12: $\mathcal{K}^{j}=\mathcal{K}^{j}_{a}\cup\mathcal{K}^{j}_{d_{k}}$
13: $\mathcal{U}^{\prime}=\left\\{u_{i}^{\prime}\sim
U|i\in[0..|\mathcal{K}_{a}^{j}||\mathcal{K}^{-j}|]\right\\}$
14: $\mathcal{U}^{j}=\mathcal{U}^{j}\cup\mathcal{U}_{a}^{j}$ We need to
associate the indices of newly created values with keys
15:end for
16:$\textbf{return}\,\mathcal{K}^{\\{0,1\\}},\mathcal{U}$
#### 3.2.5 Re-initialization complexity
Taking the constant complexity of random number generation we can assume that
the index sampling from discrete distribution is also constant. Then the
complexity of key generation for both subsets is
$\mathcal{O}(d_{q}\times|\mathcal{K}_{a}|)$. The complexity for value re-
sampling is $\mathcal{O}(|\mathcal{K}_{a}|\times|\mathcal{K}|\times d_{1})$
which in result give us the complexity of the whole procedure as:
$\mathcal{O}(d_{q}\times|\mathcal{K}_{a}|)+\mathcal{O}(|\mathcal{K}_{a}|\times|\mathcal{K}|\times
d_{v})$
where $d_{v}$ is the dimension of the memory value vectors.
### 3.3 Classification pipeline
Figure 3.3: Bottleneck [30] augmented with a memory block $\mathcal{M}$.
We are augmenting various types of the classification neural networks with the
memory layer defined in the sections above. ResNet [30] is the baseline
architecture for most of the experiments. The first idea is to augment the
Bottleneck block with the memory layer. The memory is inserted after the
($3\times 3$) kernel size convolution layer. We could also add the memory
access before the middle convolution layer but we didn’t find any differences
between the two methods so we just stick with the first design. We are keeping
the baseline high-level architecture the same by only replacing the single
Bottleneck block with the augmented version. Replacing a single layer should
be enough to observe the effect of the memory, while having only a single
layer with relatively low spatial size allows less carrying about the
efficiency of the layer implementation.
Inspired by [32] we are also adding the memory access in squeeze-and-
excitation (SE) like manner. SE is a computation block that can be built upon
any transformation in the network. It consists of two parts. First, Squeeze
which, given the feature map, captures the global spatial information and
squeeze it into the channel descriptor. Authors use global average pooling
with that goal. Second, Excitation, which employs the simple gating mechanism
upon the projection of squeezed vector with a sigmoid activation. The
projection is the bottleneck with two fully connected layers around the non-
linearity. The bottleneck layer successfully limits the complexity of the SE
block by introducing the dimensionality-reduction with ratio
$r=d_{in}/d_{out}$, where $d_{in}$ is the dimension size of input vector and
$d_{out}$ is the dimension size of the vector after the first projection in
the bottleneck.
We setup nearly the same design but with three main differences, first, we are
replacing only one block instead of every/several blocks in [32] (fewer SE
blocks give worse final score). This reduces the number of parameters to be
stored in the memory and the overall FLOPS required in the inference. Second,
channel-wise feature response is fed to the memory instead of the MLP with two
fully-connected (FC) layers around the non-linearity. This design helps us to
tackle the issues of large spatial shapes of the query input and therefore
softens the overall performance drop. Finally, instead of re-scaling the
values of the feature map with the gating output, we are simply adding the
embedding pixel-wise, i.e. replacing multiplication by addition operation and
adding the embedding to each pixel of the feature map. The overall model of
memory augmented squeeze-and-excitation block is illustrated in Figure 3.4.
Another option is to simply add the memory block as an additional layer
between the existing ones. This way we still have the issues with large
spatial shapes, especially for the earlier layers. We are testing this design
type with the ImageNet dataset [44].
Figure 3.4: The overview of the modified squeeze-and-excitation block
augmented with the memory layer. $F_{GB}$ is the function of global pooling
which reduces the dimension of 3 dimensional feature map to the signle
dimension, and $F_{add}$ is channel-wise addition of a resulting vector from
memory to the original feature tensor.
### 3.4 Regression pipeline
To test the capability of the memory layer to work on regression problems, we
are also experimenting with the camera relocalization problem [38]. The task
is to find the camera extrinsics, i.e. the camera’s location in the real world
and its direction, from the single image. Inferring the camera extrinsics is a
crucial task for mobile robotics, navigation, augmented reality.
Authors of the PoseNet neural network [38] construct the simple neural model
which consists of the backbone network and two additional layers that map the
feature representation of the input into the pose and the direction values.
First, it is the regression feature generation network as a basis for the
architecture. For that purpose GoogleNet [76] is used, we are replacing it
with ResNet [30] to conduct our experiments. The output of the backbone is fed
to the fully connected layer, which is then regressing values for direction
and orientation separately. Authors of the paper suggest to parametrize the
direction with quaternions because of the overall simplicity compared to the
rotational matrice, i.e. advantage in size: 4 scalars vs 9 in rotation matrix
and speed since quaternion multiplication is much faster compared to a matrix-
vector product used for rotation matrices. Additionally, since the rotation
matrices $n\times n$ are the members of $SO(n)$ [37], they have the
orthonormality property that should be preserved during the optimization,
which is generally the hard problem.
Since quaternion, q, is identical to -q, this leads us to the issue of non-
injectivity of the rotation value. To solve it authors normalize a quaternion
rotation to a unit length:
$\mathcal{L}_{q}(I)=\left\|\mathbf{q}-\frac{\hat{\mathbf{q}}}{\|\hat{\mathbf{q}}\|}\right\|$
For position loss, $L_{2}$ Euclidean norm is used. Introducing scaling factor
$\beta$, we can balance the overall loss, by keeping expected values of two
losses approximately equal. We are not trying to tune the scaling factor in
our experiments since it is not the main direction of this research, but we
still experiment with a large grid of hyperparameters including various values
for the scaling factor. The overall objective loss function is:
$\operatorname{loss}(I)=\|\hat{\mathbf{x}}-\mathbf{x}\|_{2}+\beta\left\|\hat{\mathbf{q}}-\frac{\mathbf{q}}{\|\mathbf{q}\|}\right\|_{2}$
We are experimenting with memory block by replacing the fully connected layer
before the final regressor of feature size 2048. Since the data size (King’s
College [38]) on which the experiments are conducted is relatively small, we
are constraining ourselves with setting the memory size to 1k/10k values. We
also regularize the memory layer by augmenting weights with Dropout
(multiplicative binomial noise) but find far worse results.
### 3.5 Image reconstruction pipeline
To test the memory layers further we are working with an image reconstruction
problem on the Imagenet-2012 [44] dataset. Image reconstruction is the type of
dimensionality reduction problem to learn the projection function that could
inject the given image into the latent representation in the compact manifold
(data encoding) and then generate the image from the given latent. Autoencoder
is a neural approach that helps us to tackle the problem in an unsupervised
fashion. In the basic design of the autoencoders, we have two networks:
encoder which maps the image into the small latent code and a decoder which
generates the image from the code.
We are experimenting with several autoencoder designs but stick to: DCGAN [60]
generator as the decoder network and the encoder as the custom 2D neural
network consisting of five ResNet blocks. The image latent is the 1024
dimensional vector. The architectural choice of the augmentation is described
in the section about the classification pipeline. We are using the basic
method of augmentation by inserting an additional memory layer in the decoder
network.111More details on the architecture will be given in the released
code.
We observe that the location of the memory layer is important on how the
memory is utilized on the train/validation sets and the final reconstruction
results.
## Chapter 4 Experiments
### 4.1 Experiments on random labels
The heart of our methodology to test the memory layers and re-initialization
technique is a well-known variant of randomized test in non-parametric
statistics [19, 88]. Drawing motivation from [88] we have conducted several
experiments to test the ability of our memory layer to fit the randomly
labelled set of data points. For this reason, we have prepared a simple data
set with $N$ sample points. We are regulating the number of samples to much
the memory size $M$. This is because our goal was having the one-to-one
correspondence between the input data and the memory values, i.e. ideally
overfitted memory layer. Sample points are the vectors uniformly generated in
$\mathbb{R}^{8}$ space, i.e. points in 8 dimensional cube. There are a total
of $m$ classes that are uniformly chosen for each data point. We have
experimented with the data set of 100k data points with 10 classes,
consequently setting $|\mathcal{M}|$ to 100k also.
Architecturally we have limited ourselves with the simplest model design with
two linear projections before and after the memory layer. It is the basic
architecture we could think of with no convolutional neural networks involved.
Moreover, we observe that using convolutional layers allows us to fit the
model to the dataset ideally. There is some research on the connection between
the multi-head self-attention and convolutional layers [14], so we have tried
to avoid the ambiguity and focused on the fully connected layers as the
projections in our network.
Also to compare our key-value structure with classic dense layers, we have
replaced memory access with very wide linear layers and point-wise non-
linearity, i.e. ReLU, sigmoid. As it is described in [11], wide layer networks
are a powerful memorizers, though in [86] authors are able to get great
memorization for small ReLU networks also, with some initialization technique
for SGD [64]. So it was interesting to see how the key-value structure
memorization capability can be compared with the wide dense layers. We have
used two fully connected layers with the ReLU in the middle. The weight matrix
of the layers are set to project the 512-dimensional vector to the
$\mathbb{R}^{15k}$ space and after applying the nonlinearity, acting as the
discriminant function in the feature space divided by hyperplane decision
surfaces, we are projecting the vector back to the space $\mathbb{R}^{512}$.
This network of two projections and the nonlinearity in the middle is the
approximation of our memory layer. This is because the k-nn function also acts
as the discriminator function, more on this in [24] (Chapter 12).
We have trained our models with an Adam optimizer [42], with an initial
learning rate of $10^{-3}$, and $\beta_{1}=0.9$, $\beta_{2}=0.98$. The models
were implemented in Pytorch [57]. For the memory values we have chosen the
SparseEmbedding structure which calculates and stores only the sparse
gradients in the backward. We have chosen the SparseAdam (implemented in
Pytorch) to update the sparse gradients in the memory layer. Because of the
sparse updates in the memory, we have multiplied the learning rate for the
sparse parameters by the factor of 10. For key parameter update, we have used
the same optimizer as for the dense parameters. Due to the usage of re-
initialization trick and Adam optimizer which stores the values of past
gradients and square gradients, these values should also be dynamically
updated. The results for the models with memory blocks and wide dense layers
compared in Figure 4.1.
Figure 4.1: Results for the random data on the various models. The left graph
is the plot for top-1 validation accuracy results, top-5 is pictured on the
right. We can see that the model with no memory is not able to fit the data.
Setting 8 heads for multi-head attention memory model, on the other hand,
helps the model to easily fit the data. Re-initialization helps to get nearly
full convergence with on only 1 head, i.e. multi-head mode disabled. Also,
$h=1$ in the graph is the notation for the single head model.
In our experiments, we varied the hyper-parameters of the memory model, such
as memory size, number of heads, k parameter in top-k operator, etc. We
provide the results only for {$k=10$, $N=M$} hyperparameter set with different
values for the number of heads $h$ and the re-initialization trick
enabled/disabled since other combinations contain no interest in these
experiments.
We observe that setting the number of heads to 8 gives us perfect fit to the
data, i.e. full top-1 validation accuracy. As it is shown in Figure 4.1,
replacing the memory layer with wide dense layer doesn’t help us with the
accuracy. Lowering the number of heads, we see the declining accuracy in the
validation. We speculate that this is caused by the poor initialization due to
which the pair of the same keys could be selected for the two very close query
vectors. Using uniform initialization to maximize the key coverage at the
initial state of the model didn’t help us to resolve the issue as we have
observed that the utilization of keys converged to some small subset.
To overcome the problem, we have experimented with re-initialization trick
that was introduced in the chapter above. As it is seen from Figure 4.1, re-
initialization helps us to get nearly ideal validation accuracy even with a
single head. We are setting $\epsilon_{d}=10^{-6}$ to get the results above.
We haven’t experimented much with the special scheduler for the re-
initialization trick, but early experiments showed that the frequency with
which the re-initialization procedure is called and the number of added keys
for each call can have the significant influence on the final accuracy we get.
More experiments are required in this direction.
We have conducted some additional experiments to see how the variance
$\sigma_{d}$ parameter of the additive noise added to the re-initialized keys
and the memory values affect the final accuracy. We are giving the results in
in Figure 4.2
Figure 4.2: Relation between $\sigma_{d}$ hyper parameter of the Gaussian
noise square variance parameter in re-initialization procedure and the values
of train accuracy and loss for the random data. Setting higher $\sigma_{d}$
helps significantly, giving lower loss.
#### 4.1.1 Results on CIFAR-10
We have implemented several architectural ideas to test the performance of
memory augmented models on CIFAR-10 [43]. The first idea was to augment the
bottleneck blocks [30] with the memory layer and replace the single bottleneck
in the network with the modified block. We have also experimented with
replacing multiple bottlenecks blocks but didn’t find anything reasonable to
stick with it in the experiments because of the overall increased inference
time we have observed.
The logic behind the bottleneck augmentation is given in the chapter above.
Here we describe the architectural choices we’ve made to incorporate the
augmented bottleneck in the most optimal way possible taking into
consideration the inference time and the final validation accuracy. The real
hurdle during the experiments was the speed issues of the inference. It didn’t
allow us to set up experiments with more broader set of models because of the
time limitations and the general difficulty of running large grids rapidly for
slower models. We were able to partly mitigate the issues by using a lower
spatial size of the query input. Taking all this into the consideration, we
have chosen the last layer to be augmented with the memory layer as it gave us
the smallest spatial size possible in the ResNet-type network. We have
abandoned the experiments with larger spatial sizes in classification
experiments for CIFAR-10 since the balance between the performance and the
accuracy wasn’t reasonable. But we still have conducted experiments with
larger spatial sizes with the autoregressive models, the results are available
in the sections below.
We have chosen the ResNet-50 [30] to be the baseline network for the
experimental models. The baseline consists of two projections and 16
Bottleneck blocks in the middle. We have added the memory layer in the 14th
Bottleneck block and have illustrate the results in Figure 4.3. The training
loop design described in [30] have been implemented with the SGD [65]
optimizer, learning rate of $10^{-1}$ weight decay of $0.0001$ and momentum
equal to $0.9$. Since the model contained the sparse parameters, we weren’t
able to use the standard implementation of the SGD optimizer in PyTorch [57].
For that, we have implemented the SparseSGD optimizer with disabled weight
decay. As for the momentum, to our knowledge, there is no mathematical ground
of using it to accelerate sparse gradients, but we have still set it to 0.9 in
all of our experiments. More information on the sparse SGD can be found here
[17].
We have adopted the weight initialization as in [29] and the batch
normalization (BN) [33]. The augmentation is the same as in [30] with 4-pixel
padding on each side of the image and 32$\times$ 32 crop randomly sampled from
the padded image and random horizontal flip. All of the experiments have been
conducted with the memory size of 100k, top-k operator parameter k of 30 and
no dropout on the retrieved indices applied. Setting the memory size to 100k
we have two sets of product keys of size 100,
$|\mathcal{K}_{1}|=|\mathcal{K}_{2}|=100$.
Figure 4.3: Comparison of the top-1 scores, memory utilization and the
inference speed for the model with no memory augmentation and three memory
augmented models with the number of heads in the set of {1, 4, 8}. There are
20 runs for the each experiment with the different initial seeds. As it can be
observed, there is no evident increase in the accuracy while the performance
of the models with the heads is much worse compared to the baseline models.
Inference values are calculated on GTX-1080Ti cards with fp32 mode, the
results are approximate.
We have calculated the distributions of top-1 scores for 20 runs with
different seed values. As it is seen in the Figure 4.3 we weren’t able to gain
any improvements in the accuracy scores with the memory augmentation, while
the performance of the memory model, i.e. iterations per second in the train,
decreases significantly and continues to decrease with the higher number of
heads. We have also calculated the distributions of the memory utilization and
observed that for larger heads we see the increase in the overall utilization.
These findings mirror the results in [46].
Evaluation metrics for memory layer. As the simple evaluation metric of how
well the memory is being utilized during the training phase, we have
calculated the memory usage score which represents the fraction of accessed
values $\\#\\{\mathcal{C}_{i}\neq 0\\}$, where $\mathcal{C}_{i}\in\mathcal{C}$
is the number of the times the key $\mathcal{K}_{i}$ is accessed summed for
the whole validation set. Authors in [46] also use Kullback-Leibler (KL) [68]
divergence between the distribution of summed values of the softmax results
for the whole validation dataset with a uniform distribution. We have
implemented the KL divergence metric in our experiments and found it giving
more accurate numbers with the small changes of the real memory utilization.
But in the given results here we have constrained our experiments to the first
evaluation metric because of its simplicity and the numerical
interpretability.
So as we can see in Figure 4.3 the utilization of the memory is increasing
with a larger number of heads. These findings were consistent during all the
experiments with the classification networks.
As the results failed on the BottleNeck blocks, we have changed the focus to
other architectures. Since we had the problem with the performance due to the
large spatial size, we have decided to limit ourselves with the image of
spatial size $1\times 1$ as the input query for the memory layer. Therefore in
the next experiments, we have leverage the architectural design of Squeeze-
and-Excitation [32] with some changes that were described in the chapter
above.
For the experiments with the modified SE blocks, we have chosen the Resnet-20
as the baseline network. We have kept the training pipeline the same but
modified the scheduler replacing it with the ReduceLRonPlateau111For eg.
Pytorch implementation of ReduceLRonPlateau with the reduction factor of
$0.1$. All the experiments with the memory layer enabled have been run with a
memory size of 100k, top-k k parameter of 30 and no dropout on the selected
indices. We have listed the most interesting results in the Table 4.1
| accuracy, top1 | inference, ms | FLOPs | utilization, %
---|---|---|---|---
ResNet-20 | 92.73(92.46$\pm$0.15) | 6.7ms | $\sim$40.92M | -
SE-ResNet-20 | 93.31(93.16$\pm$0.13) | 7.4ms | $\sim$41.49M | -
ResNet-110 | 93.63(93.41$\pm$0.18) | 25ms | $\sim$254.98M | -
ResNet-20+WL, $d_{w}=15k$ | 92.23 | 7.2ms | $\sim$43M | -
ResNet-20+Memory, scalar, h=8 | 92.42 | 19.5ms | $\sim$41.05M | 1-2%
ResNet-20+Memory, cosine, $\alpha=0.9$, h=8 | 92.45 | 19.5ms | $\sim$41.05M | 1-2%
ResNet-20+Memory/RI, scalar, h=8 | 93.16 | 19.5ms | $\sim$41.05M | 60-75%
ResNet-20+Memory/RI, cosine, $\alpha=0.9$, h=4 | 93.02 | 12ms | $\sim$41.05M | 40-50%
ResNet-20+Memory/RI, cosine, $\alpha=0.9$, h=8 | 93.51(93.34$\pm$0.14) | 19.5ms | $\sim$41.05M | 60-75%
Table 4.1: Results for the modified SE blocks. WL is the table notation for
the wide linear layer that replaces the memory layer, $d_{w}$ defines
respectably the row and column of the projections matrices in MLP (with the
row vector in the linear operator). Overall we see better results on Resnet-20
with the memory layer and with re-initialization trick we have superior memory
utilization rate. Cosine similarity helps us to nearly reach the accuracy
values of ResNet-100. Inference values are calculated on GTX-1080Ti cards with
fp32 mode, the results are approximate.
As we can see the re-initialization trick helps us with the utilization of the
memory which in turn gives us better top-1 accuracy overall. We have also
compared the memory block with the very wide MLP that consists of two large
projections matrices and the pointwise nonlinearity in the middle. We are
setting the row/column of two matrices to $d_{k}=15\text{k}$, meaning that we
have two linear operators $W_{1}\in\mathbb{R}^{d_{k}\times d_{in}}$ and
$W_{2}\in\mathbb{R}^{d_{in}\times d_{k}}$ that map the input vector
$v\in\mathbb{R}^{d_{in}}$ to the $d_{k}$ dimensional vector, applies ReLU
pointwise and project back to the vector $v^{\prime}\in\mathbb{R}^{d_{in}}$.
We can see in Table 4.1 that adding the large MLP doesn’t affect the
performance at the level compared to the memory layers. It is because the GPUs
can easily parallelize the matrix multiplication while stumbling with the
operations that require random access to the main memory[36] . We see this as
the fundamental problem of the approach with the sparse memories.
We haven’t conducted experiments with augmenting the ResNet-110 network with a
memory layer because the goal of these experiments was to understand how the
memory layer can help us with the very small networks to bit results of large
ones. And since the inference speed of the small models was inferior compared
to ResNet and SE-ResNet blocks we have changed our focus to different
applications. But more experiments should be conducted to determine whenever
ResNet-100+M results compares to the results of SE-ResNet-100 both in the
final prediction scores and the performance.
Analysis of memory layer. To find how good the introduced memory can
generalize to the given images and overall get the better picture on how the
properties of a convolutional neural network, e.g shift-invariance, are
maintained with the memory augmentation, we have conducted more experiments in
which we have randomly cropped the small region (4$\times$4) of a sample image
from the validation and compared the accessed keys for the cropped and the
original image. We see in Figure (4.4) that the small perturbation of the
input data has insignificant affect on how the keys selected. Therefore we
could assume that the generalization properties of the memory networks are
maintained that could be crucial in other applications, e.g. pose regression
on smaller datasets for which we have conducted additional experiments.
Figure 4.4: We picture the key access distribution for each head. As we see
there is a little perturbation in the distribution when we apply the random
region crop in the input image.
#### 4.1.2 Experiments on ImageNet
We have conducted further experiments on the ImageNet-2012 dataset [67],
assuming that the large size of the train set of the ImageNet would be more
natural fit for our memory layer. The only issue was time limitations we had
and the hard task of tuning the optimizer for the memory layer. Since it takes
90 epochs for the ImageNet to finish training with Resnet-50 and on NVIDIA M40
GPU it takes 14 days [85], the experiments with the size of 224$\times$224
weren’t reasonable. And since we have decided to increase the spatial size of
the query input in the memory, the inference performance of the models
plummeted. Therefore we have decided to resize the sizes of the images in
train and validation to 64$\times$64 and run the pipeline. We kept the same
augmentation pipeline as in [30].
First steps were to run the ResNet-50 augmented with the memory layer and
compare it with the baseline results. The augmentation logic we have chosen
for the ImageNet experiments were simpler. We have inserted the additional
layer before the 44th layer of the network, where the image has the 7$\times$7
spatial size, this meant that the queries consisted of 49 feature vectors that
are batched together to be fed to the memory layer. The memory size of the
experiments was set to 256k, we have looked at the top-30 indices during the
memory access and the batch size was set to 256. As the distance metric, we
have chosen the cosine similarity with $\alpha=1$. We haven’t used dropout on
the retrieved indices. SGD [65] was chosen as the optimizer with the initial
learning rate of $10^{-1}$, weight decay for dense parameters of 0.0001 and
momentum of 0.9. We haven’t set the weight decay for memory parameters because
of the inferior results, more experiments should be conducted to find the
reason for this.
The results are given in 4.2.
| accuracy, top1 | utilization | FLOPs | inference, ms
---|---|---|---|---
ResNet-50, 64$\times$64 | 61.45 (61.24$\pm$0.12) | - | $\sim$338M | 11ms
ResNet-50+M, 64$\times$64, skip=True,heads=8 | 61.61 | $\sim$98% | $\sim$382M | 23ms
ResNet-50+M, 64$\times$64, skip=True,heads=8, $\text{lr}_{key}$=$10^{2}$ | 61.82 (61.59$\pm$0.17) | $\sim$96% | $\sim$382M | 23ms
ResNet-50+M, 64$\times$64, skip=True,heads=8, $\text{lr}_{key}$=$10^{3}$ | 60.85 | $\sim$86% | $\sim$382M | 23ms
ResNet-50+M, 64$\times$64, skip=False,heads=8 | 54.12 | $\sim$15% | $\sim$382M | 23ms
ResNet-50, 32$\times$32, skip=True | 42.81 | - | $\sim$86.1M | 6ms
ResNet-50+M, 32$\times$32, skip=True,heads=8 | 43.49 | $\sim$98% | $\sim$100M | 12ms
ResNet-50+M, 32$\times$32, skip=False,heads=8 | 35.68 | $\sim$18% | $\sim$100M | 12ms
Table 4.2: The results on the Resnet-50 and Imagenet. We have tested a number
of the hyper parameters to find the best train strategy for the memory models.
For now we dont see the clear picture on optimization issues.
We can see from the table that there is a small increase in the validation
accuracy for the models augmented with the memory layer but the large drop in
the performance (inference in the table). This is not a reasonable way of
incorporating the memories with the classification models and that is why we
have tried to analyze how the values in the memory were used in the inference
and how did they change during the training phase. We hoped to find a way of
increasing the accuracy of memory augmented models by tweaking the training
pipeline. For that we have logged the gradients of the keys, memory values,
memory utilization and standard deviation of the keys during the training
phase.
We have observed that for the activated residual connection on the memory
layer, skip=True in Figure 4.5, gradients were overall higher both for memory
and key values. The utilization of the skip=True was way higher reaching
almost 100%, while the skip=False run plummeted to nearly 20%. What is most
interesting is that the standard deviations of the keys in skip=True were not
even during all training iterations. Our first assumption of the reason for
this phenomena was the low learning keys for the key parameters. Further
experiments are needed to tune the learning rate parameters. As a first step,
we have conducted more experiments with super-convergence [71] to find the top
value learning rate for key parameters in a single cycle train. We have
observed that the super-convergence leaarning rate schedule reaches $10^{6}$
before the overall loss starts to increase.
Figure 4.5: Change in values of the standard deviation of keys, memory usage,
gradient of memory values, and key gradients during a training phase.
We are not aware of all the underlining issues that do not allow us to get the
learning rate in a reasonable range. Maybe setting the learning rate value to
$10^{6}$ is logical too, but for now, we don’t know that yet. Also, we require
the augmented models to get a way better final accuracy results taking into
the consideration the performance issues of the memory blocks and the amount
of the additional parameters introduced into the network, i.e.
$|\mathcal{M}|\cdot(d_{q}+d_{v})$ where $d_{v}$ is the dimension of values in
the memory. Because of this we stop our analysis here and acknowledge the need
for more experiments.
### 4.2 Memory in PoseNet
We have conducted some experiments on PoseNet [38] for 6-DOF camera
relocalization. As it was mentioned in the previous chapter, authors of the
paper have used the GoogleNet [76] the backbone for feature extraction. We
have replaced it in a favour of ResNet-34 and run several models tuning the
hyperparameter set, especially the scaling factor $\beta$. The Adam was the
optimizer choice in these experiments. We have set the initial learning rate
of $10^{-3}$ which decreases every 80 epoch. The weight decay for the dense
parameters, i.e. all the parameters except the key and value vectors in the
memory layer, was set to $2\times 10^{-4}$. In all the results listed, we set
weight decay of memory layer to zero, Setting higher values for the weight
decay was the plan of our initial experiments also, as we have hoped to
provide some regularization for the values in the memory, but even the
smallest weight decay failed to give us any reasonable results. We acknowledge
that the additional work should be done here to find the reason behind this
issue. We have set the memory size to 1k/10k and compared the results. Overall
we have trained the models to the 250th epoch and have observed the plateau in
the train loss.
We have initiated the experiments on the King’s College outdoor relocalization
[38] which is the dataset with the images from Cambridge landmark. There are
overall 1220 images in the train set and nearly 350 in the validation set. The
small train set size has discouraged us to apply larger memory sizes
$|\mathcal{M}|$. Since the validation set is relatively large, we have assumed
that the validation accuracy could give us an overview of how good the memory
layers generalize to the dataset. For the augmentation part, we have resized
the images to 256$\times$256 and applied a random crop of 224$\times$224, the
same set of transformations have been done in the validation. We have set the
batch size of the train set to 75 and run the experiments, the results are
listed in Table 4.3.
| validation | train loss | inference | FLOPs | utiliazation
---|---|---|---|---|---
$L_{r}(q_{1},q_{2})=\left\|\mathbf{q}_{1}-\mathbf{q}_{2}\right\|,q^{0}_{i}\in\mathbb{R}^{+}$ | | | | |
PN (PoseNet), $\beta=1$ | 3.02m, 3.17∘ | 37.5 | 5.6ms | $\sim$3B | -
PN, $\beta=0.1$ | 1.57m, 5.73∘ | 6.57 | 5.6ms | $\sim$3B | -
PN+LM, $\beta=1$ | 3.24m, 3.43∘ | 34.5 | 5.8ms | $\sim$(3B+10M) | -
PN+M, heads=1, $\beta=1$, $|\mathcal{M}|=1k$ | 3.47m, 3.49∘ | 10.81 | 6ms | $\sim$(3B+16k) | 20-30%
PN+M, heads=16, $\beta=1$, $|\mathcal{M}|=1k$, dp=0.3 | 4.43m, 7.23∘ | 14.27 | 13ms | $\sim$(3B+16k) | 95-100%
PN+M, heads=16, $\beta=1$, $|\mathcal{M}|=1k$ | 2.19m, 2.99∘ | 7.94 | 13ms | $\sim$(3B+16k) | 95-100%
PN+M, heads=16, $\beta=0.1$, $|\mathcal{M}|=1k$ | 1.47m, 5.02∘ | 1.76 | 13ms | $\sim$(3B+16k) | 95-100%
PN+M, heads=16, $\beta=0.1$, $|\mathcal{M}|=10k$ | 1.44m, 4.98∘ | 1.68 | 15ms | $\sim$(3B+50k) | 50-55%
$L_{r}(q_{1},q_{2})=\min\left\\{\left\|\mathbf{q}_{1}-\mathbf{q}_{2}\right\|,\left\|\mathbf{q}_{1}+\mathbf{q}_{2}\right\|\right\\}$ | | | | |
PN (PoseNet), $\beta=1$ | 2.86m, 3.42∘ | 33.1 | 5.6ms | $\sim$3B | -
PN, $\beta=0.1$ | 1.59m, 5.36∘ | 6.45 | 5.6ms | $\sim$3B | -
PN+M, heads=16, $\beta=1$, $|\mathcal{M}|=1k$ | 1.87m, 2.53∘ | 7.41 | 13ms | $\sim$(3B+16k) | 80-95%
PN+M, heads=16, $\beta=0.1$, $|\mathcal{M}|=1k$ | 1.42m, 4.38∘ | 1.63 | 13ms | $\sim$(3B+16k) | 80-95%
Table 4.3: Results on the King’s College dataset, $\mathrm{trans}$ and
$\mathrm{rot}$ are the translation and rotation errors respectively. We see a
huge decrease in the loss on the train set for the networks with memory
augmentation but also the increase in the inference time. Inference values
calculated on the GTX-1080Ti with the batch size of 1.
We have compared the memory networks with the wide MLP layer that is defined
as LM in the table. As in the classification experiments, the MLP layer
consists of the two projection matrices and the nonlinearity between. First
projections matrix maps the input vector $v_{in}\in\mathbb{R}^{d_{in}}$ to the
$\mathbb{R}^{2k}$ then applying ReLU on the result we project the vector back
to $\mathbb{R}^{d_{in}}$. We are using the residual on the MLP. As it can be
seen from the table replacing the memory with MLP increases the train and the
validation results both for rotation and positional losses. For now, we don’t
understand why the replacement of the memory layer with the MLP can’t compete
in the final score with the memory block augmentation.
Overall we have seen the huge decrease in the train loss for memory models,
while the validation loss, though decreased both for rotation and position
loss, didn’t give us as a steep decrease in the value as we have expected. We
assume that the more elaborate regularization technique could be applied here.
But for now we have applied the naive dropout regularization on retrieved keys
which didn’t give us any promising results (dp=0.3 for dropout rate in the
Table 4.3).
Though getting better numbers overall we are seeing the huge inference time
increase for all the models augmented with the memory.
#### 4.2.1 Analysis of the memory layer
Figure 4.6: The camera path and the four sample images captures in the given
coordinates. We see the correlation between the traveled distance and the
indices selected.
To get a better picture of how the memory layer is utilized in regressing the
position coordinates, we have plotted the distribution of the accessed keys
for each image. We have scattered 200 first images in the validation set by
their x,y coordinates. We have set the colour for each point ranging it from 0
to $|\mathcal{M}|$. To calculate the colour for a particular image we have
gathered the key indices that were used in the forward operation, averaged and
rounded to the nearest integer. The results are given in Figure 4.6.
We see that there is a correlation between the distance passed by the camera
and the colours of the point, as they get darker with more distance, i.e. use
lower key indices. We could assume that the memory can capture the spatial
differences between the images and interpret them in the right manner.
### 4.3 Image reconstruction experiments
We have conducted several experiments to test the performance of the memory
layers in the reconstruction tasks. For that, we have constructed a naive
encoder-decoder neural network with the memory augmentation in the decoder.
The overall overview of the architecture is described above.
We have experiment with the various types of memory placement: right after the
latent vector (mem_idx=0), after the first layer in the decoder (mem_idx=1)
and so on. We have used the Adam [42] optimizer with the initial learning rate
of $10^{-3}$ and $\beta_{1}=0.9$, $\beta_{2}=0.98$. ImageNet samples were
resized to 64 $\times$ 64 before training the model. We have chosen $L_{2}$
norm as the objective. Also, we used the memory size of 100k, k parameter
top-k procedure of 30 and disabled dropout. The results are given in the Table
4.4.
| train loss | validation loss | inference, ms | utilization, %
---|---|---|---|---
Baseline | 0.092 | 0.0076 | 4.3ms | -
Baseline+M, heads=1, mem_idx=0 | 0.0773 | 0.0056 | 4.3ms | 0-10
Baseline+M, heads=4, mem_idx=0 | 0.075 | 0.0045 | 5ms | 10-20
Baseline+M, heads=8, mem_idx=0 | 0.0734 | 0.0044 | 6.3ms | 20-40
Baseline+M, heads=1, mem_idx=1 | 0.0721 | 0.0042 | 12.1ms | 40-50
Baseline+M, heads=4, mem_idx=1 | 0.0693 | 0.0037 | 15.4ms | 50-60
Baseline+M, heads=8, mem_idx=1 | 0.0675 | 0.0035 | 17.7ms | 60-65
Baseline+M, heads=1, mem_idx=2 | 0.0628 | 0.0026 | 21.4ms | 85-90
Baseline+M, heads=4, mem_idx=2 | 0.0544 | 0.0019 | 26.7ms | 90-100
Table 4.4: Reconstruction results after 1600 iterations. The memory
utilization numbers are approximate since the amount of experiments conducted
was low. The inference speed is calculated on GTX-1080Ti with batch size of 1.
We see the steady decline in the train and validation losses with increasing
the number of heads and the index of the layer of the decoder where the memory
is being inserted, mem_idx. Utilization numbers increase which again supports
the experiments we have conducted before. As the inference time giving us the
degraded performance with mem_idx=2. This didn’t allow us to conduct more
experiments with large spatial shapes of the input images. We include some
reconstruction examples from the validation set in Figure 4.7.
The overall pipeline and the more details on the final architecture will be
given in the released code. For now, it is important to get an understating of
the overall reconstruction improvements with the memory augmentation and if it
is reasonable to be used with the performance issues in mind.
Figure 4.7: Reconstruction results. Top row images in each section are the
reconstruction results with the memory augmented autoencoder with mem_idx=2
and heads=8, the middle are the output of the baseline autoencoder while the
final row are the input images. We see that there are some little details that
are captured using the memory layer.
### 4.4 Other experiments
We have also applied other experiments with memory usage in distillation [31],
implicit shape modelling [10] and NERF (Representing Scenes as Neural Radiance
Fields for View Synthesis) [52]. Overall, for now we can conclude that the
large batch sizes of these models’ training pipelines, i.e. point coordinate
samples for implicit modeling and the sampled rays for view synthesis in NERF,
is the hurdle which won’t allow the memory to be used in the most efficient
way, because of the difficulty of the random access parallelization with
modern GPUs. Though we see some potential in knowledge distillation from very
large models and more work should be conducted in this direction.
## Chapter 5 Conclusion
This work analyzes the usage of product key-value memory in computer vision
applications with a deep dive into the problems like image classification,
regression (in the context of the camera relocalization, and the image
reconstruction). We have found that for some of the problems the product key
memory layer was able to provide significant speed/accuracy improvements with
the high utilization of the key-value elements, while others require more
careful fine-tuning and an efficient regularization strategy. We also find
that the ”dying keys” affect the image classification problems. To help us
tackle with it, we introduce a simple technique of memory re-initialization
which helps us to eliminate ”unused” key-value pairs from the memory and
cleverly re-initialize new keys that, with high probability, will be used in
next iterations of training.
We show that the re-initialization has a huge impact on a toy example of
randomly labelled data and observe some gains in performance on the image
classification tasks.
In addition to the promising results in the experiments with the camera
relocalization, we have also shown that the choice of the set of memory
accessed indices in the inference depends on the spatial correlations of the
input images. This signals us about the perseverance of the generalization
property of the memory layer with no additional regularization required.
Still, validation results didn’t meet our expectations and at this point, we
could only assume that more work is required in defining more elaborate
regularization strategies.
We hope that the re-initialization training strategy could be applied in large
language models with exaggerated number of heads [61, 16, 9]. As we have
analyzed, smaller number of heads gave a significant boost in performance.
## References
* [1] Davide Abati, Jakub Tomczak, Tijmen Blankevoort, Simone Calderara, Rita Cucchiara, and Babak Ehteshami Bejnordi. Conditional channel gated networks for task-aware continual learning, 2020\.
* [2] Alexandr Andoni, Piotr Indyk, Thijs Laarhoven, Ilya Razenshteyn, and Ludwig Schmidt. Practical and optimal lsh for angular distance. In C. Cortes, N. D. Lawrence, D. D. Lee, M. Sugiyama, and R. Garnett, editors, Advances in Neural Information Processing Systems 28, pages 1225–1233. Curran Associates, Inc., 2015.
* [3] Joris Baan, Maartje ter Hoeve, Marlies van der Wees, Anne Schuth, and Maarten de Rijke. Understanding multi-head attention in abstractive summarization. arXiv preprint arXiv:1911.03898, 2019.
* [4] Artem Babenko and Victor Lempitsky. The inverted multi-index. IEEE transactions on pattern analysis and machine intelligence, 37(6):1247–1260, 2014.
* [5] Dzmitry Bahdanau, Kyunghyun Cho, and Yoshua Bengio. Neural machine translation by jointly learning to align and translate. arXiv preprint arXiv:1409.0473, 2014.
* [6] Dmitry Baranchuk, Artem Babenko, and Yury Malkov. Revisiting the inverted indices for billion-scale approximate nearest neighbors. In The European Conference on Computer Vision (ECCV), September 2018\.
* [7] Irwan Bello, Barret Zoph, Ashish Vaswani, Jonathon Shlens, and Quoc V. Le. Attention augmented convolutional networks. In The IEEE International Conference on Computer Vision (ICCV), October 2019.
* [8] Emmanuel Bengio, Pierre-Luc Bacon, Joelle Pineau, and Doina Precup. Conditional computation in neural networks for faster models. arXiv preprint arXiv:1511.06297, 2015.
* [9] Tom B. Brown, Benjamin Mann, Nick Ryder, Melanie Subbiah, Jared Kaplan, Prafulla Dhariwal, Arvind Neelakantan, Pranav Shyam, Girish Sastry, Amanda Askell, Sandhini Agarwal, Ariel Herbert-Voss, Gretchen Krueger, Tom Henighan, Rewon Child, Aditya Ramesh, Daniel M. Ziegler, Jeffrey Wu, Clemens Winter, Christopher Hesse, Mark Chen, Eric Sigler, Mateusz Litwin, Scott Gray, Benjamin Chess, Jack Clark, Christopher Berner, Sam McCandlish, Alec Radford, Ilya Sutskever, and Dario Amodei. Language models are few-shot learners, 2020.
* [10] Zhiqin Chen and Hao Zhang. Learning implicit fields for generative shape modeling. 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Jun 2019.
* [11] Heng-Tze Cheng, Mustafa Ispir, Rohan Anil, Zakaria Haque, Lichan Hong, Vihan Jain, Xiaobing Liu, Hemal Shah, Levent Koc, Jeremiah Harmsen, and et al. Wide & deep learning for recommender systems. Proceedings of the 1st Workshop on Deep Learning for Recommender Systems - DLRS 2016, 2016.
* [12] Kyunghyun Cho and Yoshua Bengio. Exponentially increasing the capacity-to-computation ratio for conditional computation in deep learning. arXiv preprint arXiv:1406.7362, 2014.
* [13] Cesc Chunseong Park, Byeongchang Kim, and Gunhee Kim. Attend to you: Personalized image captioning with context sequence memory networks. In The IEEE Conference on Computer Vision and Pattern Recognition (CVPR), July 2017.
* [14] Jean-Baptiste Cordonnier, Andreas Loukas, and Martin Jaggi. On the relationship between self-attention and convolutional layers. In International Conference on Learning Representations, 2020.
* [15] Andrew Davis and Itamar Arel. Low-rank approximations for conditional feedforward computation in deep neural networks. 2013\.
* [16] Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805, 2018.
* [17] Xiaohan Ding, Xiangxin Zhou, Yuchen Guo, Jungong Han, Ji Liu, et al. Global sparse momentum sgd for pruning very deep neural networks. In Advances in Neural Information Processing Systems, pages 6379–6391, 2019.
* [18] M. Douze, H. Jegou, and C. Schmid. Product quantization for nearest neighbor search. IEEE Transactions on Pattern Analysis & Machine Intelligence, 33(01):117–128, jan 2011.
* [19] Eugene Edgington and Patrick Onghena. Randomization Tests. Statistics: A Series of Textbooks and Monographs. Addison-Wesley, Reading, Massachusetts, 2007.
* [20] Leon A Gatys, Alexander S Ecker, and Matthias Bethge. A neural algorithm of artistic style. arXiv preprint arXiv:1508.06576, 2015.
* [21] Tiezheng Ge, Kaiming He, Qifa Ke, and Jian Sun. Optimized product quantization for approximate nearest neighbor search. In The IEEE Conference on Computer Vision and Pattern Recognition (CVPR), June 2013.
* [22] Aristides Gionis, Piotr Indyk, and Rajeev Motwani. Similarity search in high dimensions via hashing. VLDB, 1999.
* [23] Ian Goodfellow, Jean Pouget-Abadie, Mehdi Mirza, Bing Xu, David Warde-Farley, Sherjil Ozair, Aaron Courville, and Yoshua Bengio. Generative adversarial nets. In Z. Ghahramani, M. Welling, C. Cortes, N. D. Lawrence, and K. Q. Weinberger, editors, Advances in Neural Information Processing Systems 27, pages 2672–2680. Curran Associates, Inc., 2014.
* [24] GFosta H. Granlund and Hans Knutsson. Signal Processing for Computer Vision. Kluwer Academic Publishers, USA, 1995.
* [25] Edouard Grave, Moustapha M Cisse, and Armand Joulin. Unbounded cache model for online language modeling with open vocabulary. In I. Guyon, U. V. Luxburg, S. Bengio, H. Wallach, R. Fergus, S. Vishwanathan, and R. Garnett, editors, Advances in Neural Information Processing Systems 30, pages 6042–6052. Curran Associates, Inc., 2017.
* [26] A. Graves, A. Mohamed, and G. Hinton. Speech recognition with deep recurrent neural networks. In 2013 IEEE International Conference on Acoustics, Speech and Signal Processing, pages 6645–6649, 2013.
* [27] Alex Graves, Greg Wayne, and Ivo Danihelka. Neural turing machines. arXiv preprint arXiv:1410.5401, 2014.
* [28] Evan Greensmith, Peter L. Bartlett, and Jonathan Baxter. Variance reduction techniques for gradient estimates in reinforcement learning. In T. G. Dietterich, S. Becker, and Z. Ghahramani, editors, Advances in Neural Information Processing Systems 14, pages 1507–1514. MIT Press, 2002.
* [29] Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Delving deep into rectifiers: Surpassing human-level performance on imagenet classification. In Proceedings of the IEEE international conference on computer vision, pages 1026–1034, 2015.
* [30] Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 770–778, 2016.
* [31] Geoffrey Hinton, Oriol Vinyals, and Jeff Dean. Distilling the knowledge in a neural network. arXiv preprint arXiv:1503.02531, 2015.
* [32] Jie Hu, Li Shen, and Gang Sun. Squeeze-and-excitation networks. In IEEE Conference on Computer Vision and Pattern Recognition, 2018\.
* [33] Sergey Ioffe and Christian Szegedy. Batch normalization: Accelerating deep network training by reducing internal covariate shift. In Francis Bach and David Blei, editors, Proceedings of the 32nd International Conference on Machine Learning, volume 37 of Proceedings of Machine Learning Research, pages 448–456, Lille, France, 07–09 Jul 2015\. PMLR.
* [34] Eric Jang, Shixiang Gu, and Ben Poole. Categorical reparameterization with gumbel-softmax. 2017\.
* [35] Young Kyun Jang and Nam Ik Cho. Generalized product quantization network for semi-supervised hashing. arXiv preprint arXiv:2002.11281, 2020.
* [36] Alexander V Kashkovsky, Anton A Shershnev, and Pavel V Vashchenkov. Aspects of gpu perfomance in algorithms with random memory access. In AIP Conference Proceedings, volume 1893, page 030047. AIP Publishing LLC, 2017.
* [37] Alex Kendall and Roberto Cipolla. Geometric loss functions for camera pose regression with deep learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 5974–5983, 2017.
* [38] Alex Kendall, Matthew Grimes, and Roberto Cipolla. Posenet: A convolutional network for real-time 6-dof camera relocalization. In Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV), ICCV ’15, page 2938–2946, USA, 2015. IEEE Computer Society.
* [39] Urvashi Khandelwal, Omer Levy, Dan Jurafsky, Luke Zettlemoyer, and Mike Lewis. Generalization through memorization: Nearest neighbor language models. In International Conference on Learning Representations, 2020.
* [40] Amir Hosein Khasahmadi, Kaveh Hassani, Parsa Moradi, Leo Lee, and Quaid Morris. Memory-based graph networks. In International Conference on Learning Representations, 2020.
* [41] Youngjin Kim, Minjung Kim, and Gunhee Kim. Memorization precedes generation: Learning unsupervised GANs with memory networks. In International Conference on Learning Representations, 2018.
* [42] Diederick P Kingma and Jimmy Ba. Adam: A method for stochastic optimization. In International Conference on Learning Representations (ICLR), 2015\.
* [43] Alex Krizhevsky. Learning multiple layers of features from tiny images. Technical report, 2009.
* [44] Alex Krizhevsky, Ilya Sutskever, and Geoffrey E Hinton. Imagenet classification with deep convolutional neural networks. In F. Pereira, C. J. C. Burges, L. Bottou, and K. Q. Weinberger, editors, Advances in Neural Information Processing Systems 25, pages 1097–1105. Curran Associates, Inc., 2012.
* [45] Alex Krizhevsky, Ilya Sutskever, and Geoffrey E Hinton. Imagenet classification with deep convolutional neural networks. In F. Pereira, C. J. C. Burges, L. Bottou, and K. Q. Weinberger, editors, Advances in Neural Information Processing Systems 25, pages 1097–1105. Curran Associates, Inc., 2012.
* [46] Guillaume Lample, Alexandre Sablayrolles, Marc' Aurelio Ranzato, Ludovic Denoyer, and Herve Jegou. Large memory layers with product keys. In H. Wallach, H. Larochelle, A. Beygelzimer, F. d'Alché-Buc, E. Fox, and R. Garnett, editors, Advances in Neural Information Processing Systems 32, pages 8548–8559. Curran Associates, Inc., 2019.
* [47] Sangho Lee, Jinyoung Sung, Youngjae Yu, and Gunhee Kim. A memory network approach for story-based temporal summarization of 360° videos. In The IEEE Conference on Computer Vision and Pattern Recognition (CVPR), June 2018.
* [48] Christos Louizos, Max Welling, and Diederik P. Kingma. Learning sparse neural networks through $l_{0}$ regularization. In International Conference on Learning Representations, 2018.
* [49] Chris J. Maddison, Andriy Mnih, and Yee Whye Teh. The concrete distribution: A continuous relaxation of discrete random variables. 2016\. cite arxiv:1611.00712.
* [50] Y. A. Malkov and D. A. Yashunin. Efficient and robust approximate nearest neighbor search using hierarchical navigable small world graphs. IEEE Transactions on Pattern Analysis and Machine Intelligence, 42(4):824–836, 2020.
* [51] Hongzi Mao, Shaileshh Bojja Venkatakrishnan, Malte Schwarzkopf, and Mohammad Alizadeh. Variance reduction for reinforcement learning in input-driven environments. In International Conference on Learning Representations, 2019.
* [52] Ben Mildenhall, Pratul P Srinivasan, Matthew Tancik, Jonathan T Barron, Ravi Ramamoorthi, and Ren Ng. Nerf: Representing scenes as neural radiance fields for view synthesis. arXiv preprint arXiv:2003.08934, 2020.
* [53] Markus Nagel, Rana Ali Amjad, Mart van Baalen, Christos Louizos, and Tijmen Blankevoort. Up or down? adaptive rounding for post-training quantization, 2020.
* [54] Behnam Neyshabur, Zhiyuan Li, Srinadh Bhojanapalli, Yann LeCun, and Nathan Srebro. Towards understanding the role of over-parametrization in generalization of neural networks. arXiv preprint arXiv:1805.12076, 2018.
* [55] Vlad Niculae, André F. T. Martins, Mathieu Blondel, and Claire Cardie. Sparsemap: Differentiable sparse structured inference, 2018.
* [56] Qingqun Ning, Jianke Zhu, Zhiyuan Zhong, Steven CH Hoi, and Chun Chen. Scalable image retrieval by sparse product quantization. IEEE Transactions on Multimedia, 19(3):586–597, 2016.
* [57] Adam Paszke, Sam Gross, Francisco Massa, Adam Lerer, James Bradbury, Gregory Chanan, Trevor Killeen, Zeming Lin, Natalia Gimelshein, Luca Antiga, Alban Desmaison, Andreas Kopf, Edward Yang, Zachary DeVito, Martin Raison, Alykhan Tejani, Sasank Chilamkurthy, Benoit Steiner, Lu Fang, Junjie Bai, and Soumith Chintala. Pytorch: An imperative style, high-performance deep learning library. In H. Wallach, H. Larochelle, A. Beygelzimer, F. d'Alché-Buc, E. Fox, and R. Garnett, editors, Advances in Neural Information Processing Systems 32, pages 8024–8035. Curran Associates, Inc., 2019.
* [58] Sergei Popov, Stanislav Morozov, and Artem Babenko. Neural oblivious decision ensembles for deep learning on tabular data. In International Conference on Learning Representations, 2020.
* [59] Alexander Pritzel, Benigno Uria, Sriram Srinivasan, Adrià Puigdomènech Badia, Oriol Vinyals, Demis Hassabis, Daan Wierstra, and Charles Blundell. Neural episodic control. In Doina Precup and Yee Whye Teh, editors, Proceedings of the 34th International Conference on Machine Learning, volume 70 of Proceedings of Machine Learning Research, pages 2827–2836, International Convention Centre, Sydney, Australia, 06–11 Aug 2017. PMLR.
* [60] Alec Radford, Luke Metz, and Soumith Chintala. Unsupervised representation learning with deep convolutional generative adversarial networks. arXiv preprint arXiv:1511.06434, 2015.
* [61] Alec Radford, Jeff Wu, Rewon Child, David Luan, Dario Amodei, and Ilya Sutskever. Language models are unsupervised multitask learners. 2019\.
* [62] Prajit Ramachandran, Niki Parmar, Ashish Vaswani, Irwan Bello, Anselm Levskaya, and Jonathon Shlens. Stand-alone self-attention in vision models. arXiv preprint arXiv:1906.05909, 2019.
* [63] Ali Razavi, Aaron van den Oord, and Oriol Vinyals. Generating diverse high-fidelity images with vq-vae-2. In H. Wallach, H. Larochelle, A. Beygelzimer, F. d'Alché-Buc, E. Fox, and R. Garnett, editors, Advances in Neural Information Processing Systems 32, pages 14866–14876. Curran Associates, Inc., 2019.
* [64] Herbert E. Robbins. A stochastic approximation method. Annals of Mathematical Statistics, 22:400–407, 2007.
* [65] Herbert E. Robbins. A stochastic approximation method. Annals of Mathematical Statistics, 22:400–407, 2007.
* [66] Christian P. Robert. The metropolis-hastings algorithm. arXiv preprint arXiv:1504.01896, 2015.
* [67] Olga Russakovsky, Jia Deng, Hao Su, Jonathan Krause, Sanjeev Satheesh, Sean Ma, Zhiheng Huang, Andrej Karpathy, Aditya Khosla, Michael Bernstein, et al. Imagenet large scale visual recognition challenge. International journal of computer vision, 115(3):211–252, 2015\.
* [68] Igal Sason and Sergio Verdu. $f$ -divergence inequalities. IEEE Transactions on Information Theory, 62(11):5973–6006, Nov 2016.
* [69] Noam Shazeer, Azalia Mirhoseini, Krzysztof Maziarz, Andy Davis, Quoc Le, Geoffrey Hinton, and Jeff Dean. Outrageously large neural networks: The sparsely-gated mixture-of-experts layer. 2017\.
* [70] Karen Simonyan and Andrew Zisserman. Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556, 2014.
* [71] Leslie N. Smith and Nicholay Topin. Super-convergence: very fast training of neural networks using large learning rates. Artificial Intelligence and Machine Learning for Multi-Domain Operations Applications, May 2019.
* [72] S Spigler, M Geiger, S d’Ascoli, L Sagun, G Biroli, and M Wyart. A jamming transition from under-to over-parametrization affects generalization in deep learning. Journal of Physics A: Mathematical and Theoretical, 52(47):474001, 2019.
* [73] Sainbayar Sukhbaatar, arthur szlam, Jason Weston, and Rob Fergus. End-to-end memory networks. In C. Cortes, N. D. Lawrence, D. D. Lee, M. Sugiyama, and R. Garnett, editors, Advances in Neural Information Processing Systems 28, pages 2440–2448. Curran Associates, Inc., 2015.
* [74] Ilya Sutskever, Oriol Vinyals, and Quoc V Le. Sequence to sequence learning with neural networks. In Z. Ghahramani, M. Welling, C. Cortes, N. D. Lawrence, and K. Q. Weinberger, editors, Advances in Neural Information Processing Systems 27, pages 3104–3112. Curran Associates, Inc., 2014.
* [75] Richard S Sutton, David A. McAllester, Satinder P. Singh, and Yishay Mansour. Policy gradient methods for reinforcement learning with function approximation. In S. A. Solla, T. K. Leen, and K. Müller, editors, Advances in Neural Information Processing Systems 12, pages 1057–1063. MIT Press, 2000\.
* [76] Christian Szegedy, Wei Liu, Yangqing Jia, Pierre Sermanet, Scott Reed, Dragomir Anguelov, Dumitru Erhan, Vincent Vanhoucke, and Andrew Rabinovich. Going deeper with convolutions. In Computer Vision and Pattern Recognition (CVPR), 2015.
* [77] Mingxing Tan and Quoc Le. EfficientNet: Rethinking model scaling for convolutional neural networks. In Kamalika Chaudhuri and Ruslan Salakhutdinov, editors, Proceedings of the 36th International Conference on Machine Learning, volume 97 of Proceedings of Machine Learning Research, pages 6105–6114, Long Beach, California, USA, 09–15 Jun 2019. PMLR.
* [78] Chen-Tse Tsai, Wen-tau Yih, Chris J.C. Burges, and Scott Wen-tau Yih. Web-based question answering: Revisiting askmsr. Technical Report MSR-TR-2015-20, April 2015.
* [79] Aaron van den Oord, Oriol Vinyals, and koray kavukcuoglu. Neural discrete representation learning. In I. Guyon, U. V. Luxburg, S. Bengio, H. Wallach, R. Fergus, S. Vishwanathan, and R. Garnett, editors, Advances in Neural Information Processing Systems 30, pages 6306–6315. Curran Associates, Inc., 2017.
* [80] Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, Ł ukasz Kaiser, and Illia Polosukhin. Attention is all you need. In I. Guyon, U. V. Luxburg, S. Bengio, H. Wallach, R. Fergus, S. Vishwanathan, and R. Garnett, editors, Advances in Neural Information Processing Systems 30, pages 5998–6008. Curran Associates, Inc., 2017.
* [81] Elena Voita, David Talbot, Fedor Moiseev, Rico Sennrich, and Ivan Titov. Analyzing multi-head self-attention: Specialized heads do the heavy lifting, the rest can be pruned. Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, 2019.
* [82] Xin Wang, Fisher Yu, Zi-Yi Dou, Trevor Darrell, and Joseph E. Gonzalez. Skipnet: Learning dynamic routing in convolutional networks. Lecture Notes in Computer Science, page 420–436, 2018.
* [83] Ran Yi, Zipeng Ye, Juyong Zhang, Hujun Bao, and Yong-Jin Liu. Audio-driven talking face video generation with learning-based personalized head pose. arXiv preprint arXiv:2002.10137, 2020.
* [84] Seungjoo Yoo, Hyojin Bahng, Sunghyo Chung, Junsoo Lee, Jaehyuk Chang, and Jaegul Choo. Coloring with limited data: Few-shot colorization via memory augmented networks. In The IEEE Conference on Computer Vision and Pattern Recognition (CVPR), June 2019.
* [85] Yang You, Zhao Zhang, James Demmel, Kurt Keutzer, and Cho-Jui Hsieh. Imagenet training in 24 minutes. arXiv preprint arXiv:1709.05011, 2017.
* [86] Chulhee Yun, Suvrit Sra, and Ali Jadbabaie. Small relu networks are powerful memorizers: a tight analysis of memorization capacity. In H. Wallach, H. Larochelle, A. Beygelzimer, F. d'Alché-Buc, E. Fox, and R. Garnett, editors, Advances in Neural Information Processing Systems 32, pages 15558–15569. Curran Associates, Inc., 2019.
* [87] Matthew D. Zeiler and Rob Fergus. Visualizing and understanding convolutional networks. Lecture Notes in Computer Science, page 818–833, 2014.
* [88] Chiyuan Zhang, Samy Bengio, Moritz Hardt, Benjamin Recht, and Oriol Vinyals. Understanding deep learning requires rethinking generalization. arXiv preprint arXiv:1611.03530, 2016.
* [89] Łukasz Kaiser, Ofir Nachum, Aurko Roy, and Samy Bengio. Learning to remember rare events. arXiv preprint arXiv:1703.03129, 2017.
|
# Hadamard Extensions and the Identification of Mixtures of Product
Distributions
Spencer L. Gordon Engineering and Applied Science, California Institute of
Technology<EMAIL_ADDRESS>Leonard J. Schulman Engineering and Applied
Science, California Institute of Technology<EMAIL_ADDRESS>Research
supported in part by NSF grant CCF-1909972.
###### Abstract
The Hadamard Extension of a matrix is the matrix consisting of all Hadamard
products of subsets of its rows. This construction arises in the context of
identifying a mixture of product distributions on binary random variables:
full column rank of such extensions is a necessary ingredient of
identification algorithms. We provide several results concerning when a
Hadamard Extension has full column rank.
## 1 Introduction
The Hadamard product for row vectors $u=(u_{1},\ldots,u_{k})$,
$v=(v_{1},\ldots,v_{k})$ is the mapping
$\odot:\mathbb{R}^{k}\times\mathbb{R}^{k}\to\mathbb{R}^{k}$ given by
$\displaystyle u\odot v$ $\displaystyle:=(u_{1}v_{1},\ldots,u_{k}v_{k})$
The identity for this product is the all-ones vector $\mathbbm{1}$. We
associate with vector $v$ the linear operator
$v_{\odot}=\operatorname{diag}(v)$, a $k\times k$ diagonal matrix, so that
$u\cdot v_{\odot}=v\odot u.$
Throughout this paper ${\mathbf{m}}$ is a real matrix with row set
$[n]:=\\{1,\ldots,n\\}$ and column set $[k]$; write ${\mathbf{m}}_{i}$ for a
row and ${\mathbf{m}}^{j}$ for a column.
As a matter of notation, for a matrix $Q$ and nonempty sets $R$ of rows and
$C$ of columns, let $Q|_{R}^{C}$ be the restriction of $Q$ to those columns
and rows (with either index omitted if all rows or columns are retained).
###### Definition 1.
The Hadamard Extension of ${\mathbf{m}}$, written $\mathbb{H}({\mathbf{m}})$,
is the $2^{n}\times k$ matrix with rows ${\mathbf{m}}_{S}$ for all
$S\subseteq[n]$, where, for $S=\\{i_{1},\ldots,i_{\ell}\\}$,
${\mathbf{m}}_{S}={\mathbf{m}}_{i_{1}}\odot\cdots\odot{\mathbf{m}}_{i_{\ell}}$;
equivalently ${\mathbf{m}}_{S}^{j}=\prod_{i\in S}{\mathbf{m}}_{i}^{j}$. (In
particular ${\mathbf{m}}_{\emptyset}=\mathbbm{1}$.)
This construction has arisen recently in learning theory [3, 8] where it is
essential to source identification for a mixture of product distributions on
binary random variables. We explain the connection further in Section 5.
Motivated by this application, we are interested in the following two
questions:
(1) If $\mathbb{H}({\mathbf{m}})$ has full column rank, must there exist a
subset $R$ of the rows, of bounded size, such that
$\mathbb{H}({\mathbf{m}}|_{R})$ has full column rank?
(2) In each row of ${\mathbf{m}}$, assign distinct colors to the distinct real
values. Is there a condition on the coloring that ensures
$\mathbb{H}({\mathbf{m}})$ has full column rank?
In answer to the first question we show in Section 2:
###### Theorem 2.
If $\mathbb{H}({\mathbf{m}})$ has full column rank then there is a set $R$ of
no more than $k-1$ of the rows of ${\mathbf{m}}$, such that
$\mathbb{H}({\mathbf{m}}|_{R})$ has full column rank.
Considering the more combinatorial second question, observe that if
${\mathbf{m}}$ possesses two identical columns then the same is true of
$\mathbb{H}({\mathbf{m}})$, and so it cannot be full rank. Extending this
further, suppose there are three columns $C$ in which only one row $r$ has
more than one color. Then
$\operatorname{Rowspace}\mathbb{H}({\mathbf{m}}|^{C})$ is spanned by
$\mathbbm{1}|^{C}$ and $r|^{C}$, so again $\mathbb{H}({\mathbf{m}})$ cannot be
full rank. Motivated by these necessary conditions, set:
###### Definition 3.
For a matrix $Q$ let $\text{NAE}(Q)$ be the set of nonconstant rows of $Q$
(NAE=“not all equal”); let ${\varepsilon}(Q|^{C})=|\text{NAE}(Q|^{C})|-|C|$;
and let
$\overline{{\varepsilon}}(Q)=\min_{C\neq\emptyset}{\varepsilon}(Q|^{C})$. If
$\overline{{\varepsilon}}(Q)\geq-1$ we say $Q$ satisfies the NAE condition.
In answer to the second question we have the following:
###### Theorem 4.
If ${\mathbf{m}}$ satisfies the NAE condition then
1. (a)
There is a restriction of ${\mathbf{m}}$ to some $k-1$ rows $R$ such that
$\overline{{\varepsilon}}({\mathbf{m}}|_{R})=-1$.
2. (b)
$\mathbb{H}({\mathbf{m}})$ is full column rank.
(As a consequence also $\mathbb{H}({\mathbf{m}}|_{R})$ is full column rank.)
Apparently the only well-known example of the NAE condition is when
${\mathbf{m}}$ contains $k-1$ rows which are identical and whose entries are
all distinct. Then the vectors
${\mathbf{m}}_{\emptyset},{\mathbf{m}}_{\\{1\\}},{\mathbf{m}}_{\\{1,2\\}},\ldots,{\mathbf{m}}_{\\{1,\ldots,k-1\\}}$
form a nonsingular Vandermonde matrix. This example shows that the bound of
$k-1$ in (a) is best possible.
For another example in which the NAE condition ensures that
$\operatorname{rank}\mathbb{H}({\mathbf{m}})=k$, take the $(k-1)$-row matrix
with ${\mathbf{m}}_{i}^{j}=1$ for $i\leq j$ and ${\mathbf{m}}_{i}^{j}=1/2$ for
$i>j$. Here the NAE condition is only minimally satisfied, in that for every
$\ell\leq k$ there are $\ell$ columns $C$ s.t.
${\varepsilon}({\mathbf{m}}|^{C})=-1$.
For $k>3$ the NAE condition is no longer necessary for
$\mathbb{H}({\mathbf{m}})$ to have full column rank. E.g., for $k=2^{\ell}$,
the $\ell\times k$ “Hamming matrix” ${\mathbf{m}}_{i}^{j}=(-1)^{j_{i}}$ where
$j$ is an $\ell$-bit string $j=(j_{1},\ldots,j_{\ell})$, forms
$\mathbb{H}({\mathbf{m}})=$ the Fourier transform for the group
$(\mathbb{Z}/2)^{\ell}$ (often called a Hadamard matrix), which is invertible.
Furthermore, almost all (in the sense of Lebesgue measure) $\lceil\lg
k\rceil\times k$ matrices ${\mathbf{m}}$ form a full-rank
$\mathbb{H}({\mathbf{m}})$. (This is because $\det\mathbb{H}({\mathbf{m}})$ is
a polynomial in the entries of ${\mathbf{m}}$, and the previous example shows
the polynomial is nonzero.) Despite this observation, the Vandermonde case, in
which $k-1$ rows are required, is very typical, as it is what arises in
$\mathbb{H}({\mathbf{m}})$ for a mixture model of observables $X_{i}$ that are
iid conditional on a hidden variable.
## 2 Some Theory for Hadamard Products, and a Proof of Theorem 2
For $v\in\mathbb{R}^{k}$ and $U$ a subspace, extend the definition $v_{\odot}$
to
$v_{\odot}(U)=\\{u\cdot v_{\odot}:u\in U\\}$
and introduce the notation
$v_{\bar{\odot}}(U)=\operatorname{span}\\{U\cup v_{\odot}(U)\\}.$
We want to understand which subspaces $U$ are invariant under
$v_{\bar{\odot}}$. Let $v$ have distinct values
$\lambda_{1}>\ldots>\lambda_{\ell}$ for $\ell\leq k$. Let the polynomials
$p_{v,i}$ ($i=1,\ldots,\ell$) of degree $\ell-1$ be the Lagrange interpolation
polynomials for these values, so $p_{v,i}(\lambda_{j})=\delta_{ij}$ (Kronecker
delta). Let $B(v)$ denote the partition of $[k]$ into blocks
$B(v)_{(i)}=\\{j:v_{j}=\lambda_{i}\\}$. Let $V_{(i)}$ be the space spanned by
the elementary basis vectors in $B(v)_{(i)}$, and $P_{(i)}$ the projection
onto $V_{(i)}$ w.r.t. standard inner product. We have the matrix equation
$p_{v,i}(v_{\odot})=P_{(i)}.$
The collection of all linear combinations of the matrices $P_{(i)}$ is a
commutative algebra, the _$B(v)$ projection algebra_, which we denote
$A_{B(v)}$. The identity of the algebra is $I=\sum P_{(i)}$.
###### Definition 5.
A subspace of $\mathbb{R}^{k}$ respects $B(v)$ if it is spanned by vectors
each of which lies in some $V_{(i)}$.
For $U$ respecting $B(v)$ write $U=\operatorname{span}(\bigcup U_{(i)})$ for
$U_{(i)}\subseteq V_{(i)}$. Let $D_{(i)}=(U_{(i)})^{\perp}\cap V_{(i)}$. Then
$(U_{(i)})^{\perp}=D_{(i)}\oplus\bigoplus_{j\neq i}V_{(j)}$.
###### Lemma 6.
Subspace $U^{\perp}$ respects $B(v)$ if $U$ does.
###### Proof.
In general, $(\operatorname{span}(W\cup W^{\prime}))^{\perp}=W^{\perp}\cap
W^{\prime\perp}$. So $U^{\perp}=\bigcap(U_{(i)})^{\perp}=\bigoplus D_{(i)}$. ∎
###### Lemma 7.
Subspace $U$ respects $B(v)$ iff $U=\bigoplus(P_{(i)}U)$.
###### Proof.
($\Leftarrow$): Because this gives an explicit representation of $U$ as a
direct sum of subspaces each restricted to some $V_{(i)}$. ($\Rightarrow$): By
definition $U$ is spanned by some collection of subspaces
$V^{\prime}_{(i)}\subseteq V_{(i)}$; since these subspaces are necessarily
orthogonal, $U=\bigoplus V^{\prime}_{(i)}$. Moreover, since $P_{(i)}$
annihilates $V_{(j)},j\neq i$, and is the identity on $V_{(i)}$, it follows
that each $V^{\prime}_{(i)}=P_{(i)}U$. ∎
###### Theorem 8.
Subspace $U$ is invariant under $v_{\bar{\odot}}$ iff $U$ respects $B(v)$.
###### Proof.
($\Leftarrow$): It suffices to show $U^{\perp}$ is invariant under
$v_{\bar{\odot}}$. By the previous lemma, it is equivalent to suppose that
$U^{\perp}$ respects $B(v)$. So let $d\in U^{\perp}$ and write $d=\sum
d_{i},d_{i}\in D_{(i)}$. Then $v\odot d_{i}=\lambda_{i}d_{i}\in D_{(i)}$. So
$v\odot d=\sum v\odot d_{i}\in\bigoplus D_{(i)}=U^{\perp}$.
($\Rightarrow$): If $U=v_{\bar{\odot}}(U)$ then these also equal
$v_{\bar{\odot}}(v_{\bar{\odot}}(U))$, etc., so $U$ is an invariant space of
$A_{B(v)}$, meaning, $aU\subseteq U$ for any $a\in A_{B(v)}$. In particular
for $a=P_{(i)}$. So $U\supseteq\bigoplus(P_{(i)}U)$. On the other hand, since
$\sum P_{(i)}=I$, $U=(\sum P_{(i)})U\subseteq\bigoplus(P_{(i)}U)$. So
$U=\bigoplus(P_{(i)}U)$. Now apply Lemma 7. ∎
The symbol $\subset$ is reserved for strict inclusion.
###### Lemma 9.
If $S,T\subseteq[n]$ and
$\operatorname{Rowspace}\mathbb{H}({\mathbf{m}}|_{S})\subset\operatorname{Rowspace}\mathbb{H}({\mathbf{m}}|_{S\cup
T})$, then there is a row $t\in T$ such that
$\operatorname{Rowspace}\mathbb{H}({\mathbf{m}}|_{S})\subset\operatorname{Rowspace}\mathbb{H}({\mathbf{m}}|_{S\cup\\{t\\}})$.
###### Proof.
Without loss of generality $S,T$ are disjoint. Let $T^{\prime}\subseteq T$ be
a smallest set s.t. $\exists S^{\prime}\subseteq S$ s.t.
${\mathbf{m}}_{S^{\prime}}\odot{\mathbf{m}}_{T^{\prime}}\notin\operatorname{Rowspace}\mathbb{H}({\mathbf{m}}_{S})$.
Select any $t\in T^{\prime}$ and write
${\mathbf{m}}_{S^{\prime}}\odot{\mathbf{m}}_{T^{\prime}}={\mathbf{m}}_{S^{\prime}}\odot{\mathbf{m}}_{T^{\prime}-\\{t\\}}\odot{\mathbf{m}}_{t}$.
By minimality of $T^{\prime}$,
${\mathbf{m}}_{S^{\prime}}\odot{\mathbf{m}}_{T^{\prime}-\\{t\\}}\in\operatorname{Rowspace}\mathbb{H}({\mathbf{m}}_{S})$.
But then
${\mathbf{m}}_{S^{\prime}}\odot{\mathbf{m}}_{T^{\prime}}\in\operatorname{Rowspace}\mathbb{H}({\mathbf{m}}_{S\cup\\{t\\}})$,
so
$\operatorname{Rowspace}\mathbb{H}({\mathbf{m}}|_{S})\subset\operatorname{Rowspace}\mathbb{H}({\mathbf{m}}|_{S\cup\\{t\\}})$.
∎
Theorem 2 is now a consequence of Lemma 9. ∎
It follows from Theorem 2 that we can check whether
$\operatorname{rank}\mathbb{H}({\mathbf{m}})=k$ in time $O(n)^{k}$ by
computing $\operatorname{rank}\mathbb{H}({\mathbf{m}}|_{S})$ for each
$S\in\binom{[n]}{k-1}$.
## 3 Combinatorics of the NAE Condition: Proof of Theorem 4 (a)
Recall we are to show: 4 (a): If
$\overline{{\varepsilon}}({\mathbf{m}})\geq-1$ then ${\mathbf{m}}$ has a
restriction to some $k-1$ rows on which $\overline{{\varepsilon}}=-1$.
###### Proof.
We induct on $k$. The (vacuous) base-case is $k=1$. For $k>1$, we induct on
$n$, with base-case $n=k-1$.
Supposing the Theorem fails for $k$, $k>1$, let ${\mathbf{m}}$ be a $k$-column
counterexample with least $n$. Necessarily every row is in
$\text{NAE}({\mathbf{m}})$, and $n>k-1\geq 1$. We will show ${\mathbf{m}}$ has
a restriction ${\mathbf{m}}^{\prime}$ to $n-1$ rows, for which
$\overline{{\varepsilon}}({\mathbf{m}}^{\prime})\geq-1$; this will imply a
contradiction because, by minimality of ${\mathbf{m}}$,
${\mathbf{m}}^{\prime}$ has a restriction to $k-1$ rows on which
$\overline{{\varepsilon}}=-1$.
If $\overline{{\varepsilon}}({\mathbf{m}})\geq 0$ then we can remove any
single row of ${\mathbf{m}}$ and still satisfy
$\overline{{\varepsilon}}\geq-1$.
Otherwise, $\overline{{\varepsilon}}({\mathbf{m}})=-1$, so there is a nonempty
$S$ such that $|\text{NAE}({\mathbf{m}}|^{S})|=|S|-1$; choose a largest such
$S$. It cannot be that $S=[k]$ (as then $n=k-1$). Arrange the rows
$\text{NAE}({\mathbf{m}}|^{S})$ as the bottom $|S|-1$ rows of the matrix. As
discussed earlier, for the NAE condition one may regard the distinct real
values in each row of ${\mathbf{m}}$ simply as distinct colors; relabel the
colors in each row above $\text{NAE}({\mathbf{m}}|^{S})$ so the color above
$S$ is called “white.” (There need be no consistency among the real numbers
called white in different rows.) See Figure 1.
Figure 1: Argument for Theorem 4 (a). Upper-left region is white. Entries
$(t,f(t))$ are not white.
Due to the maximality of $|S|$, there is no white rectangle on $\ell$ columns
and $n-|S|-\ell+1$ rows inside
${\mathbf{m}}|_{[n]-\text{NAE}({\mathbf{m}}|^{S})}^{[k]-S}$ for any $\ell\geq
1$. That is to say, if we form a bipartite graph on right vertices
corresponding to the columns $[k]-S$, and left vertices corresponding to the
rows $[n]-\text{NAE}({\mathbf{m}}|^{S})$, with non-white cells being edges,
then any subset of the right vertices of size $\ell\geq 1$ has at least
$\ell+1$ neighbors within the left vertices.
By the induction on $k$ (since $S\neq\emptyset$), for the set of columns
$[k]-S$ there is a set $R^{\prime\prime}$ of $k-|S|-1$ rows such that
$\overline{{\varepsilon}}({\mathbf{m}}|_{R^{\prime\prime}}^{[k]-S})=-1$.
Together with the rows of $\text{NAE}({\mathbf{m}}|^{S})$ this amounts to at
most $k-2$ rows, so since $n\geq k$, we can find two rows outside this union;
delete either one of them, leaving a matrix ${\mathbf{m}}^{\prime}$ with $n-1$
rows. This matrix has the rows $\text{NAE}({\mathbf{m}}|^{S})$ at the bottom,
and $n-|S|$ remaining rows which we call $R^{\prime}$. The lemma will follow
by showing that $\overline{{\varepsilon}}({\mathbf{m}}^{\prime})\geq-1$.
In ${\mathbf{m}}^{\prime}$, the induced bipartite graph on right vertices
$[k]-S$ and left vertices $R^{\prime}$ has the property that any right subset
of size $\ell\geq 1$ has a neighborhood of size at least $\ell$ in
$R^{\prime}$. Applying Hall’s Marriage Theorem, there is an injective
$f:[k]-S\to R^{\prime}$ employing only edges of the graph.
Now consider any set of columns $T$, $T=T_{1}\cup
T_{2},T_{1}\subseteq[k]-S,T_{2}\subseteq S$. We need to show that
${\varepsilon}({\mathbf{m}}|^{T})\geq-1$. Let
$R_{1}=\text{NAE}({\mathbf{m}}|^{T_{1}})\cap R^{\prime\prime}$,
$R_{2}=\text{NAE}({\mathbf{m}}|^{T_{2}})\subseteq\text{NAE}({\mathbf{m}}|^{S})$,
and note that $|R_{1}|\geq|T_{1}|-1$, $|R_{2}|\geq|T_{2}|-1$. If
$T_{2}=\emptyset$ we simply use $R_{1}$. Likewise if $T_{1}=\emptyset$, we use
$R_{2}$.
If both $T_{1}$ and $T_{2}$ are nonempty,
$\text{NAE}({\mathbf{m}}|^{T_{2}})\subseteq\text{NAE}({\mathbf{m}}|^{S})$, and
$|\text{NAE}({\mathbf{m}}|^{T_{2}})|\geq|T_{2}|-1$. Now use the matching $f$.
The set of rows $f(T_{1})$ lies in $R^{\prime}$ and is therefore disjoint from
$\text{NAE}({\mathbf{m}}|^{T_{2}})$. Moreover since $T_{2}\neq\emptyset$,
every entry $(t,j)$ for $t\in T_{2},j\in R^{\prime}$ is white. On the other
hand due to the construction of $f$, for every $t\in T_{1}$ the entry
$(t,f(t))$ is non-white. Therefore every row in $f(T_{1})$ is in
$\text{NAE}({\mathbf{m}}|^{T_{1}\cup T_{2}})$. So
$|\text{NAE}({\mathbf{m}}|^{T_{1}\cup T_{2}})|\geq|T_{2}|-1+|T_{1}|$. Thus
$\overline{{\varepsilon}}({\mathbf{m}}^{\prime})\geq-1$. ∎
## 4 From NAE to Rank: Proof of Theorem 4 (b)
Recall we are to show: 4 (b): $\mathbb{H}({\mathbf{m}})$ has full column rank
if $\overline{{\varepsilon}}({\mathbf{m}})\geq-1$.
###### Proof.
The case $k=1$ is trivial. Now suppose $k\geq 2$ and that Theorem 4 (b) holds
for all $k^{\prime}<k$. Any constant rows of ${\mathbf{m}}$ affect neither the
hypothesis nor the conclusion, so remove them, leaving ${\mathbf{m}}$ with at
least $k-1$ rows. Now pick any set, $C$, of $k-1$ columns of ${\mathbf{m}}$.
By Theorem 4 (a) there are some $k-2$ rows of ${\mathbf{m}}$, call them
$R^{\prime}$, on which
$\overline{{\varepsilon}}({\mathbf{m}}|_{R^{\prime}}^{C})=-1$. Let $v$ be a
row of ${\mathbf{m}}$ outside $R^{\prime}$. Call the rows of ${\mathbf{m}}$
apart from $v$, $R^{\prime\prime}$. Since $R^{\prime\prime}$ contains
$R^{\prime}$, by induction
$\dim\operatorname{Rowspace}\mathbb{H}({\mathbf{m}}|_{R^{\prime\prime}}^{C})=k-1$.
Therefore
$U:=\operatorname{Rowspace}\mathbb{H}({\mathbf{m}}|_{R^{\prime\prime}})\subseteq\mathbb{R}^{k}$
is of dimension at least $k-1$. We claim now that $\dim U=k$.
Suppose to the contrary that $\dim U=k-1$. If $v_{\odot}(U)\subseteq U$ then
as proven earlier in Theorem 8, $U$ respects $B(v)$. Since $v$ is nonconstant,
$B(v)$ is a partition of $[k]$ into $\ell\geq 2$ nonempty blocks $B(v)_{(i)}$,
and $U=\bigoplus_{i=1}^{\ell}U_{(i)}$ with $U_{(i)}=P_{(i)}U_{(i)}$. So there
is some $i_{0}$ for which $U_{(i_{0})}\subset V_{(i_{0})}$; specifically,
$U_{(i)}=V_{(i)}$ for all $i\neq i_{0}$, and $\dim U_{(i_{0})}=\dim
V_{(i_{0})}-1$. Since $|B(v)_{(i_{0})}|<k$, we know by induction that the rows
of $\mathbb{H}({\mathbf{m}})$ span $V_{(i_{0})}$. Thus in fact
$U=\mathbb{R}^{k}$. (Further detail for the last step: let
$w\in\mathbb{R}^{k}$. Since the rows of $\mathbb{H}({\mathbf{m}})$ span
$V_{(i_{0})}$, there is a
$w^{\prime}\in\operatorname{Rowspace}\mathbb{H}({\mathbf{m}})$ s.t.
$P_{(i_{0})}w^{\prime}=P_{(i_{0})}w$. Moreover since $U_{(i)}=V_{(i)}$ for all
$i\neq i_{0}$, there is a $w^{\prime\prime}\in U$ s.t.
$w^{\prime\prime}=(I-P_{(i_{0})})(w-w^{\prime})$. Then
$w^{\prime}+w^{\prime\prime}\in\operatorname{Rowspace}\mathbb{H}({\mathbf{m}})$,
and $w^{\prime}+w^{\prime\prime}=w$.) ∎
## 5 Motivation
Consider observable random variables $X_{1},\ldots,X_{n}$ that are
statistically independent conditional on $H$, a hidden random variable $H$
supported on $\\{1,\ldots,k\\}$. (See causal diagram.)
$\textstyle{H}$$\textstyle{X_{1}}$$\textstyle{X_{2}}$$\textstyle{\cdots}$$\textstyle{X_{i}}$$\textstyle{\cdots}$$\textstyle{X_{n}}$
The most fundamental case is that the $X_{i}$ are binary. Then we denote
${\mathbf{m}}_{i}^{j}=\Pr(X_{i}=1|H=j)$. The model parameters are
${\mathbf{m}}$ along with a probability distribution (the mixture
distribution) $\pi=(\pi_{1},\ldots,\pi_{k})$ on $H$.
Finite mixture models were pioneered in the late 1800s in [13, 14]. The
problem of learning such distributions has drawn a great deal of attention.
For surveys see, e.g., [5, 17, 11, 12]. For some algorithmic papers on
discrete $X_{i}$, see [9, 4, 7, 2, 6, 1, 15, 10, 3, 8]. The source
identification problem is that of computing $({\mathbf{m}},\pi)$ from the
joint statistics of the $X_{i}$. Put another way, the problem is to invert the
multilinear moment map
$\displaystyle\mu:({\mathbf{m}},\pi)$ $\displaystyle\to\mathbb{R}^{2^{[n]}}$
$\displaystyle\mu({\mathbf{m}},\pi)_{S}$
$\displaystyle=\Pr(X_{S}=1)\quad\text{ where
}S\subseteq[n],\;X_{S}=\prod\nolimits_{i\in S}X_{i}$
$\displaystyle={\mathbf{m}}_{S}\cdot\pi^{\top}$
The last line shows the significance of $\mathbb{H}({\mathbf{m}})$ to mixture
model identification, since ${\mathbf{m}}_{S}^{j}=\Pr(X_{S}=1|H=j)$.
### Connection to $\operatorname{rank}\mathbb{H}({\mathbf{m}})$.
In general $\mu$ is not injective (even allowing for permutation among the
values of $H$). For instance it is clearly not injective if ${\mathbf{m}}$ has
two identical columns (unless $\pi$ places no weight on those). More
generally, and assuming all $\pi_{j}>0$, it cannot be injective unless
$\mathbb{H}({\mathbf{m}})$ has full column rank.
One sufficient condition for injectivity, due to [16], is that there be $2k-1$
“separated” observables $X_{i}$; $X_{i}$ is separated if all
${\mathbf{m}}_{i}^{j}$ are distinct, or in our terminology, if no color recurs
in ${\mathbf{m}}_{i}$. (Further [8], one can lower bound the distance between
$\mu({\mathbf{m}},\pi)$ and any $\mu({\mathbf{m}}^{\prime},\pi^{\prime})$ in
terms of $\min_{i,j}|{\mathbf{m}}_{i}^{j}-{\mathbf{m}}_{i}^{j^{\prime}}|$ and
the distance between $({\mathbf{m}},\pi)$ and
$({\mathbf{m}}^{\prime},\pi^{\prime})$.)
A weaker sufficient condition for injectivity of $\mu$, due to [8], is that
for every $i\in[n]$ there exist two disjoint sets $A,B\subseteq[n]-\\{i\\}$
such that $\mathbb{H}({\mathbf{m}}|_{A})$ and $\mathbb{H}({\mathbf{m}}|_{B})$
have full column rank. (It is not known whether two disjoint such $A,B$ are
strictly necessary, but the implied $n\leq 2k-1$ is in general best possible
[15].)
## References
* [1] A. Anandkumar, D. Hsu, and S. M. Kakade. A method of moments for mixture models and hidden Markov models. In Proc. 25th Ann. Conf. on Computational Learning Theory, pages 33.1–33.34, 2012.
* [2] K. Chaudhuri and S. Rao. Learning mixtures of product distributions using correlations and independence. In Proc. 21st Ann. Conf. on Computational Learning Theory, pages 9–20, 2008.
* [3] S. Chen and A. Moitra. Beyond the low-degree algorithm: mixtures of subcubes and their applications. In Proc. 51st Ann. ACM Symp. on Theory of Computing, pages 869–880, 2019.
* [4] M. Cryan, L. Goldberg, and P. Goldberg. Evolutionary trees can be learned in polynomial time in the two state general Markov model. SIAM J. Comput., 31(2):375–397, 2001. Prev. FOCS ’98.
* [5] B. S. Everitt and D. J. Hand. Mixtures of discrete distributions, pages 89–105. Springer Netherlands, Dordrecht, 1981.
* [6] J. Feldman, R. O’Donnell, and R. A. Servedio. Learning mixtures of product distributions over discrete domains. SIAM J. Comput., 37(5):1536–1564, 2008.
* [7] Y. Freund and Y. Mansour. Estimating a mixture of two product distributions. In Proc. 12th Ann. Conf. on Computational Learning Theory, pages 183–192, July 1999.
* [8] S. L. Gordon, B. Mazaheri, Y. Rabani, and L. J. Schulman. Source identification for mixtures of product distributions. arXiv:2012.14540, 2020.
* [9] M. Kearns, Y. Mansour, D. Ron, R. Rubinfeld, R. Schapire, and L. Sellie. On the learnability of discrete distributions. In Proc. 26th Ann. ACM Symp. on Theory of Computing, pages 273–282, 1994.
* [10] J. Li, Y. Rabani, L. J. Schulman, and C. Swamy. Learning arbitrary statistical mixtures of discrete distributions. In Proc. 47th Ann. ACM Symp. on Theory of Computing, pages 743–752, 2015.
* [11] B. G. Lindsay. Mixture models: theory, geometry and applications. In NSF-CBMS regional conference series in probability and statistics, pages i–163. JSTOR, 1995.
* [12] G. J. McLachlan, S. X. Lee, and S. I. Rathnayake. Finite mixture models. Annual Review of Statistics and Its Application, 6(1):355–378, 2019\.
* [13] S. Newcomb. A generalized theory of the combination of observations so as to obtain the best result. American Journal of Mathematics, 8(4):343–366, 1886.
* [14] K. Pearson. Contributions to the mathematical theory of evolution III. Philosophical Transactions of the Royal Society of London (A.), 185:71–110, 1894.
* [15] Y. Rabani, L. J. Schulman, and C. Swamy. Learning mixtures of arbitrary distributions over large discrete domains. In Proc. 5th Conf. on Innovations in Theoretical Computer Science, pages 207–224, 2014.
* [16] B. Tahmasebi, S. A. Motahari, and M. A. Maddah-Ali. On the identifiability of finite mixtures of finite product measures. IEEE International Symposium on Information Theory (ISIT) 2018 and arXiv:1807.05444v1, 2018.
* [17] D. M. Titterington, A. F. M. Smith, and U. E. Makov. Statistical Analysis of Finite Mixture Distributions. John Wiley and Sons, Inc., 1985.
|
11institutetext: imec-IDLab-UGent, Department of Electronics and Information
Systems, Ghent University, Technologiepark-Zwijnaarde 126, 9052 Zwijnaarde,
Belgium 22institutetext: IPEM-UGent, Department of Art History, Musicology and
Theatre Studies, Ghent University, Miriam Makebaplein 1, 9000 Gent, Belgium
33institutetext: imec-mict-UGent, Department of Communication Sciences, Ghent
University, Miriam Makebaplein 1, 9000 Gent, Belgium
# Art and Science Interaction Lab
A highly flexible and modular interaction science research facility
Niels Van Kets§ 11 0001-5495-2240 Bart Moens§ 22 0002-1281-9406 Klaas Bombeke§
33 0003-2056-1246 Wouter Durnez§ 33 0001-8045-8801 Pieter-Jan Maes 22
0002-9237-3298 Glenn Van Wallendael§ 11 0001-9530-3466 Lieven De Marez 33
0001-7716-4079 Marc Leman 22 0002-9780-2194 Peter Lambert§ 11 0001-5313-4158
###### Abstract
The Art and Science Interaction Lab (“ASIL”) is a unique, highly flexible and
modular “interaction science” research facility to effectively bring, analyse
and test experiences and interactions in mixed virtual/augmented contexts as
well as to conduct research on next-gen immersive technologies. It brings
together the expertise and creativity of engineers, performers, designers and
scientists creating solutions and experiences shaping the lives of people. The
lab is equipped with state-of-the-art visual, auditory and user-tracking
equipment, fully synchronized and connected to a central backend. This
synchronization allows for highly accurate multi-sensor measurements and
analysis.
###### Keywords:
Immersive Experience Virtual Reality User Testing
## 1 Introduction
The Art and Science Interaction Lab (ASIL) team supports innovation in
different key domains. Within these domains, the team focuses on interaction
research in virtualized environments, unraveling complex user interactions and
experiences in order to design and create novel applications and interfaces.
The application domains span from smart home appliances, health, safety, smart
public places to more artistic and creative applications. Furthermore, the lab
infrastructure is used for fundamental research on virtual reality
technologies (e.g. auralization, virtual acoustics, 6 degrees of freedom VR,
multi-person VR…) and is dark fiber connected to three concert halls in Ghent.
The team is an interdisciplinary consortium combining the expertise of three
Ghent University research groups (IDLab, IPEM and mict) and has been co-funded
under the medium-scale research infrastructure program governed by the
Research Foundation Flanders (FWO). The combination of humanities,
engineering, psychology and social sciences makes the ASIL a one of its kind
research facility. Moreover, the research tracks in the ASIL target both
industry and academia to deliver a unique interdisciplinary approach in
measuring, analyzing and creating our next-generation appliances, interfaces
and experiences.
## 2 Technical infrastructure
The Art and Science Interaction lab is located in ”De Krook” building (Fig. 1)
in the city center of Ghent. Next to the city of Ghent library, ”De Krook”
also houses the three Ghent University research groups involved in the Art and
Science Interaction Lab (IDLab, IPEM and mict), as well as imec, an R&D hub
for nano- and digital technologies. Both IDLab and mict are affiliated imec
research groups.
The Lab is located in the sub-level floors of the building and has a volume of
10m x 10m and spans a height of two full floors ($\approx$ 6m).
Figure 1: De Krook
In view of the audiovisual applications envisioned in this room, the walls and
ceiling all have been acoustically treated, delivering a RT60 Reverberation
Time of 0.5s. The RT60 Reverberation Time describes how long it takes for
sound to decay by 60dB in a room with a diffuse soundfield. By insulating the
room, a reduction of approximately 2s was met.
Furthermore, in order to deliver a highly flexible and modular research
infrastructure, the lab has been equipped with a state of the art trussing
system (Fig. 2, including 5 motorized trusses of 7m. This trussing system
allows for a myriad of experimental setups.
(a)
(b)
Figure 2: Trussing system
In order to limit the length and amount of cables to be used in setting up new
experiments, a large amount of patch boxes have been installed on both the
fixed and moving trusses. All patch boxes deliver power, data, UTP, coaxial,
DMX and XLR connections to several locations in the lab (Fig. 3a). Each of
these patch-points are directly linked to the patch rack (Fig. 3b), from where
connections towards the machine room or other patch points can be made.
(a)
(b)
Figure 3: Patching infrastructure
## 3 Audiovisual equipment
The Art and Science Interaction Lab is equipped with 80 individual speakers
connected to a fully IP-based audio distribution system. The system is capable
of delivering highly realistic spatial audio projection by making use of a
dedicated audio wavefield processor. This audio system allows for accurate
recreation and simulation of room acoustics.
In terms of visual modalities, the lab is equipped with 2 fully untethered HTC
Vive Pro Eye, 2 tethered HTC Vive Pro and 2 Microsoft Hololens version 2
devices, allowing free roaming spanning the full 10m x 10m area and allowing
for multi person AR/VR. A 7m x 4m acoustically transparent projection screen
in combination with a 12.000 lumen 4K projector delivers compelling and high-
end immersive visualizations. Both the audio and visual systems are connected
to a powerful state-of-the-art processing backend.
### 3.1 Audio equipment
An immersive speaker system and a state-of-the-art sound processing backend
allows compelling auditory experiences. Due to the acoustic insulation, sound
reflections and deformations are vastly reduced in order to deliver the most
accurate auditory stimuli to the listeners.
#### 3.1.1 Equipment list
* •
80 calibrated speakers (8 subs, 2 speaker rings and an overhead speaker set)
* •
Barco IOSONO [2] core wavefield synthesis system
* •
Fully IP-based Dante [1] audio infrastructure
* •
10 amplifiers including built-in DSP for each individual speaker channel
* •
Highly flexible XLR and IP-audio patch possibilities
* •
Fully synced with 48kHz clock
* •
Software/frameworks: Ableton Live, Max/MSP, Ambisonics…
#### 3.1.2 Applications:
Figure 4: Audio Applications - Musical Performance
The audio installation in the Art and Science Interaction Lab allows for
accurate object-based projection of 3D audio in space by using wavefield
synthesis. This enables the (re-)creation of immersive auditory environments,
bringing a myriad of application and research opportunities such as the
recreation of multi-track musical performances and the use of interactive
sound objects that can dynamically move in space.
Furthermore, such a system allows for the (re-)creation of real-life or
simulated acoustic environments. This allows one to experience the acoustic
properties of a certain location (e.g. concert hall, church, outdoors…) or the
simulation of acoustic properties of future buildings or expositions.
### 3.2 Video and Mixed AR/VR equipment
The visual installation caters towards multi-user applications, with the focus
on usability and freedom of movement. State-of-the-art virtual and augmented
reality headsets are readily available in the lab. Furthermore, a high-
resolution projection system is capable of delivering compelling grouped or
single experiences.
#### 3.2.1 Equipment
* •
2x HTC Vive Pro Eye untethered VR headsets with built-in eye tracking
* •
2x HTC Vive Pro tethered VR headsets
* •
2x Microsoft Hololens v2 AR glasses
* •
7m x 4m acoustically transparent projection screen
* •
12.000 lumens 4K projector with active stereo 3D and Extended Dynamic Range
* •
high-end rendering with latest NVIDIA GPU compute capabilities
#### 3.2.2 Applications
Figure 5: Visual applications
The visual systems integrated in the Art and Science Interaction Lab allow for
the delivery of immersive and interactive virtual and augmented reality
experiences while maintaining the highest degree of free movement possible.
The state-of-the-art processing backend allows for real-time rendering and
visualization of complex 6 degrees of freedom experiences (6DoF).
Secondly, the systems allow for multi-user VR and AR applications, allowing
multiple users to interact naturally in a shared virtual environment.
Integrated with the aforementioned positional audio and the motion capture
system that will be described in the following section, these interactions can
be of very high realism.
Lastly, the high resolution, high dynamic range projection allows for
immersive context creation without the use of a virtual reality headset, as
well as delivering compelling visual experiences to larger groups.
### 3.3 Motion capture
A dense Qualisys motion capture setup allows full body tracking of multiple
users (up to 5) with an accuracy of $<$ 1mm3, on a 81m2 floor area and a 5m
configurable volume height. This allows detailed tracking of (multi) user
movement. Furthermore, real-time integration with both the audio and
visualization solutions allow for interactive audiovisual experiences based on
human movement.
#### 3.3.1 Equipment
* •
Qualisys Oqus 7+ infrared cameras [3] (14 fixed+ 4 mobile units)
* •
Qualisys Miqus Video cameras [4] (4 fixed + 1 mobile unit)
* •
Synced with the Art and Science Interaction Lab 120Hz clock signal
* •
High-end real-time processing backend allowing for real-time skeleton tracking
and streaming
* •
Tracking compatible with VR headsets
#### 3.3.2 Applications
Figure 6: Motion Capture - Qualisys Oqus 7+ IR camera
The motion capture installation in the Art and Science Interaction Lab allows
for highly accurate measurements of human behavior (motion) in interactive
scenarios. The system allows for fine-granular motion tracking (e.g. finger
tracking while playing an instrument, face expression tracking…) but also
robust full body tracking of multiple users in the same area.
The system is capable of mapping marker data onto skeletons in order to
animate avatars in VR and in real-time. This allows for accurate real-time
representations of users in a virtual (e.g. VR) space.
Next to tracking users, the motion capture system allows for accurate 6
degrees-of-freedom tracking of objects. This by applying reflective markers to
real-life objects.
## 4 Sensor equipment
The Art and Science Interaction Lab is the go-to facility for interaction and
user-experience research. The infrastructure provides a wide variety of
synchronized sensors capable of measuring different aspects of an experience
and is backed by a strong team of UX researchers.
The interdisciplinary team of the Art and Science Interaction Lab can deliver
a detailed unraveling of user experiences and interactions, even those where
users are not aware of.
#### 4.0.1 Equipment:
* •
2x clinical grade untethered EEG headsets (64 channels with active
electrodes), which can be used in combination with our wireless VR headsets.
* •
Eye trackers, both built into VR headsets and standalone.
* •
Skin conductance sensors
* •
Heart rate sensors
* •
EMG sensors
* •
Synced with the Art and Science Interaction Lab 120Hz clock signal
#### 4.0.2 Applications:
Figure 7: Sensor Applications - Physiological Analysis
The available sensors in the Art and Science Interaction Lab and their
integration and synchronization allow the measurement of a highly accurate
physiological footprint of a user experience. By combining different sensors
such as EEG, EMG, skin conductance… and the aforementioned motion capture
system, physiological processes within human interactions and experiences can
be unraveled on a granular scale. This analysis can be used to discover points
of pain in a certain user-experience.
Furthermore, in a more artistic setting, real-time sensor data can be used in
order to create novel audiovisual experiences based on real-time biofeedback.
## 5 Interconnectivity
The Art and Science Interaction Lab serves as a central hub towards different
concert halls in the city center of Ghent. A 10Gbps dark fiber towards ’De
Vooruit’, ’De Minard’ and the future ’Wintercircus’ allows for real-time
shared experiences between the Art and Science Interaction lab and three
important cultural venues in Ghent.
Next to the cultural venues, the Art and Science Interaction Lab is also
directly connected towards URGENT, the Ghent University Student radio, which
also resides in De Krook.
Lastly, the lab has a direct dark fiber link to an off-site back-up facility,
delivering a secure and redundant data storage.
#### 5.0.1 Applications:
The direct link with the concert venues and URGENT radio allow for real-time
streaming of multi-track audio between these facilities in order to enable the
(re-)creation of a cultural experience in the Art and Science Interaction Lab
as well as creating novel experiences based on real-time audio feeds.
Furthermore, real-time motion capture allows for novel experiences where
motion in the Art and Science Interaction Lab can be used in real-time in the
cultural venues (e.g. the creation of virtual avatars or experiences).
## References
* [1] Audinate: Audinate dante (2021), https://www.audinate.com
* [2] Barco: Barco iosono (2019), https://www.barco.com/nl/product/iosono-core/specsheet
* [3] Systems, Q.M.C.: Qualisys 5+, 6+ and 7+ (2018), https://cdn-content.qualisys.com/2018/11/PI_Camera-5-6-and-7.pdf
* [4] Systems, Q.M.C.: Miqus video (2020), https://cdn-content.qualisys.com/2020/04/PI_Miqus_Video.pdf
|
# Robust and Efficient Single-Pixel Image Classification with Nonlinear Optics
Santosh Kumar Department of Physics, Stevens Institute of Technology,
Hoboken, NJ, 07030, USA Center for Quantum Science and Engineering, Stevens
Institute of Technology, Hoboken, NJ, 07030, USA These authors contributed
equally Ting Bu Department of Physics, Stevens Institute of Technology,
Hoboken, NJ, 07030, USA Center for Quantum Science and Engineering, Stevens
Institute of Technology, Hoboken, NJ, 07030, USA These authors contributed
equally He Zhang Department of Physics, Stevens Institute of Technology,
Hoboken, NJ, 07030, USA Center for Quantum Science and Engineering, Stevens
Institute of Technology, Hoboken, NJ, 07030, USA Irwin Huang Department of
Physics, Stevens Institute of Technology, Hoboken, NJ, 07030, USA Yuping
Huang Department of Physics, Stevens Institute of Technology, Hoboken, NJ,
07030, USA Center for Quantum Science and Engineering, Stevens Institute of
Technology, Hoboken, NJ, 07030, USA Corresponding author<EMAIL_ADDRESS>
###### Abstract
We present a hybrid image classifier by mode-selective image upconversion,
single pixel photodetection, and deep learning, aiming at fast processing a
large number of pixels. It utilizes partial Fourier transform to extract the
signature features of images in both the original and Fourier domains, thereby
significantly increasing the classification accuracy and robustness. Tested on
the MNIST handwritten digit images, it boosts the accuracy from 81.25% to
99.23%, and achieves an 83% accuracy for highly contaminated images whose
signal-to-noise ratio is only -17 dB. Our approach could prove useful for fast
lidar data processing, high resolution image recognition, occluded target
identification, atmosphere monitoring, and so on.
††journal: ol
Machine learning techniques based on deep neural networks (DNNs) have scored
considerable success in image classification [1, 2], speech recognition [3],
image generation [4], image reconstruction [5], and so on. They use a feed-
forward multi-layer neural architecture to achieve high performance for
complex operations [6, 7]. Despite notable success for their digital
implementations, it remains a challenge to adopt them directly for embedded
systems due to the high memory-demand, limited scalability, and excessive
energy budget [8]. On the other hand, optical systems naturally offer extreme
parallelism and multiplexing yet with little energy consumption and fast data-
processing [9, 10, 11]. Thus far, optical Fourier transformation, diffraction,
interference, and filtering have been exploited for optical pattern
recognition [12, 13] and restoration [14], phase retrieval [15, 16], and
optical information processing [17]. Recently, optically assisted image
processing has become a vivid pursuit in the field of artificial intelligence
[18], computer vision [19], robotics [20, 21], medical science [22], and so
on.
While most of these studies are based on linear optics, nonlinear optics is
poised to enable even richer and more complex operations and lift our
processing capability to yet another level. To this end, optical neural
network with nonlinear activation functions was implemented for object
identification and classification of ordered and disordered phases of Ising
models [23]. Spontaneous parametric down-conversion was utilized for quantum-
correlated pattern recognition with spatially structured photons [24].
Artificial neural networks based on nonlinear optics have shown advantages in
reconstructing the amplitude and phase profiles of ultrashort pulses [25, 26]
and neuromorphic computing [27, 28]. A super-Ising emulator was demonstrated
with unprecedented four-body interaction using spatial light modulation and
second-harmonic generation [29].
Recently, we have proposed and experimentally demonstrated a nonlinear-optics
approach to pattern recognition with single-pixel imaging and a deep neural
network [30]. It employs mode-selective image up-conversion to project a raw
image with up to mega pixels onto a set of coherent spatial modes, whereby its
signature features are extracted optically in a nonlinear manner. Our
experimental results promise distinct applications in online classification of
large-size images, fast lidar data analyses [31], complex pattern recognition,
and so on.
In this work, we continue to study the aforementioned hybrid machine learning
technique, focusing on further improving the efficiency and effectiveness in
the feature extraction. In our previous system [30], the mode-selective
conversion is applied to the fully Fourier-transformed images, which samples
the signature features in the Fourier domain, but overlooks some important
identifying features in the original image domain. Here, we study a new
approach where the upconversion is applied to the images undergoing only
partial Fourier transform. While challenging to calculate partial Fourier
transform numerically [32, 33], it is easily realized in the current system by
displacing the nonlinear crystal from the focal point of the Fourier lens for
the images. This simple adjustment allows both the original and Fourier
features of the images to be sampled at the same time, leading to high
classification accuracy using only a limited number of pump modes for the
upconversion. Tested on the MNIST (Modified National Institute of Standards
and Technology) handwritten digit images, a 40 mm crystal displacement
significantly increases the accuracy from 81.25% to 99.23%. The same technique
enables robust pattern recognition when the images are mixed with strong
noise, where a high accuracy $\sim 83\%$ is achieved even when the signal-to-
noise (SNR) is as low as -17 dB. All of our experimental results agree well
with the simulated results, confirming this effect. Finally, we numerically
compare two choices of the pump mode bases for the upconversion that can
affect the feature extraction, showing better performance with the Laguerre
Gaussian (LG) mode bases than the Hermite-Gaussian (HG) modes, likely due to a
lower cross-talk for the former [34]. Our results highlight a viable approach
to image recognition through feature extraction in both the original and
Fourier domains by mode-selective image upconversion, for high accuracy, high
efficiency, and exceptional robustness.
Figure 1: Experimental setup for the feature extraction via frequency up-
conversion. The PPLN crystal is moved away from the focal point to the focal
lens by a distance $\Delta$, as shown in the lower right inset. SLM: Spatial
Light Modulator, BS: Beamsplitter, WDM: Wavelength-division multiplexing,
PPLN: Magnesium-doped Periodic Poled Lithium Niobate crystal; PM: power meter.
The experimental setup for the present nonlinear optical pattern-recognition
scheme is outlined in Fig. 1. The phase patterns of the Gaussian beams in
signal and pump arms are modulated by two separate spatial light modulators
(SLM). In the signal arm, SLM1 is used to upload phase patterns of input
images while the SLM2 is applied to map phase patterns of spatial LG (or HG)
modes on pump beam. The phase value is wrapped in the interval between 0 and
$2\pi$ to express on the SLMs [35, 36]. The phase modulated two beams are then
merged at a beam splitter (BS). A Fourier lens ($f$=200 mm) is used to focus
the coupled beam in a temperature-stabilized periodic poled lithium niobate
(PPLN) crystal for sum-frequency (SF) generation process. The power readings
of the SF generation are extracted features from images, which are inputs to a
deep neural network for digits recognition. Further details of our setup can
be found in our previous work [30]. The difference between the current setup
and the previous one is the position of the PPLN. As illustrated in the lower
right inset in Fig. 1, $\Delta$ is the of the displacement of PPLN crystal
towards the Fourier lens.
In simulation, for each pair of the signal beam $E_{s}$ and pump beam $E_{p}$,
their corresponding SF field $E_{f}$ is solved under the slowly-varying-
envelope approximation as:
$\displaystyle
2ik_{s}\partial_{z}E_{s}+(\partial_{x}^{2}+\partial_{y}^{2})E_{s}=-2\frac{\omega_{s}^{2}}{c^{2}}\chi^{(2)}E_{p}^{*}E_{f}e^{i\triangle
kz},$ (1) $\displaystyle
2ik_{p}\partial_{z}E_{p}+(\partial_{x}^{2}+\partial_{y}^{2})E_{p}=-2\frac{\omega_{p}^{2}}{c^{2}}\chi^{(2)}E_{s}^{*}E_{f}e^{i\triangle
kz},$ (2) $\displaystyle
2ik_{f}\partial_{z}E_{f}+(\partial_{x}^{2}+\partial_{y}^{2})E_{f}=-2\frac{\omega_{f}^{2}}{c^{2}}\chi^{(2)}E_{p}E_{s}e^{-i\triangle
kz},$ (3)
where $\omega_{s}$, $\omega_{p}$, and $\omega_{f}$ are the beam waists of
signal, pump and SF light, respectively. $k_{s}={n_{s}\omega_{s}}/{c}$ ,
$k_{p}={n_{p}\omega_{p}}/{c}$ and $k_{f}={n_{f}\omega_{f}}/{c}$ are their wave
numbers. $\Delta k=k_{s}+k_{p}-k_{f}-2\pi/\Lambda$ gives their phase mismatch,
$\Lambda=19.36$ $\mu$m is the poling period of the PPLN. $z$ is the distance
in propagation direction with $z=0$ being the focal point. $\chi^{(2)}$ is the
second-order nonlinear susceptibility. We can use the standard split-step
Fourier and adaptive step size methods [37] to numerically solve Eqs.(1)-(3)
with an initial location $z=-\frac{L}{2}-\Delta$, where $L=1$ $cm$ is the
length of the PPLN.
Figure 2: Normalized confusion matrices for testing the performance of the
hand-written digits with three different crystal positions. The top panels
(a-c) are the simulated results using 40 LG pump modes with $\Delta=0$, 27 mm
and 40 mm, respectively. The bottom panels (d-f) are the corresponding
experimental results.
In this paper, one of the most popular MNIST handwritten digit database is
selected as our benchmark, which is a 10-class database including handwritten
images representing digits from "0" to "9" [38]. A subset of the original
database consisting of the first 200 handwritten images of each digit is
applied as our database. These images in our database are shuffled and
separated into training (1600 images) and testing (400 images) sets. The
resolution of these 28$\times$28 images are increased to 400$\times$400 pixels
to match with the input signal beam size and SLM pitch. The SF power readings
of each handwritten image are normalized between 0 and 1, after collecting all
results for our training and testing data sets. This normalization is
different from our previous work in [30], where the upconversion takes place
in the Fourier plane and data are normalized within each LG mode. While the
original and Fourier planes offer equivalent descriptions [39], the amount of
information that can be extracted using a small number of pump modes is quite
different. In [30], the upconversion was performed in the Fourier domain,
which resulted in similar SF power distributions for the first few pump modes.
The signature features are contained in the higher order modes where the SF
power is relatively low. To distinguish those digits therefore requires
“amplifying” the SF power difference in the higher order modes, which was
realized through SF power normalization by mode. While high performance is
permissible, that method renders the system prone to noise and fluctuations.
Also, after adding new training or testing data, the whole data set needs to
be normalized again and new training process is warranted for the best
performance.
Figure 3: (a) An example of random noise mask with $\textrm{SNR}=-10$ dB. The
zoomed windows are representing the noise clusters with cluster size S equal
to the length of 10 SLM pixels (i.e. one noise pixel size, $\textrm{S}=10$ SLM
pixels or 104 $\mu$m). (b) Experimental and simulated accuracy vs SNR. Blue
and green dotted curves are the experimental and simulated accuracy for
$\Delta$ = 27 mm, respectively, and red and orange dashed curves are for
$\Delta$ = 40 mm.
In contrast, when the upconversion is applied to partially Fourier transformed
images, the signature features of the images are extracted in both the
original and Fourier domains. They are readily manifested in SF power for the
first few LG modes of high readings, resulting in distinct power distribution
over the 40 LG modes. This eliminates the need to amplify the SF power for
higher-order modes, and allows normalization for each image. It makes the
system more robust and efficient, while avoiding the need of repeated
normalization within each mode over the entire data set when new data are
added.
The structure of the deep neural network is same as the one in our previous
work in [30], which is a combination of one convolutional layer and five fully
connected layers. In the training process, the adaptive moment estimation
(ADAM) gradient descent algorithm is selected to minimize the loss between
predictions and their ground truth. The activation function softmax is applied
to normalize the output values between 0 and 1 so that they can represent the
probabilities of each class. Then the predicted class of each input image will
be chosen as the one with the largest probability. Finally, the classification
accuracy can be calculated as the fraction of correctly classified images.
To optimally extract the spatial and Fourier features at the same time, we
displace the PPLN crystal closer to the Fourier lens with its center away from
the focal point by $\Delta$, as shown in Fig.1. While this leads to efficient
and robust classification, a large $\Delta$ de-focuses the light and
suppresses the SF generation. In light of this trade-off, we choose
$\Delta=0$, 27 mm and 40 mm in this study. A larger $\Delta$ may be
accommodated by using a higher pump power to compensate for the SF efficiency
loss. Also, here we follow [30] to use a set of 40 LG modes, with $l\in[-2,2]$
and $p\in[0,7]$. Figure 2 lists the normalized confusion matrices using the 40
LG pump modes for features extraction. Figure 2(a-c) are the simulated results
and (d-f) are the experimental results with $\Delta=0,$ 27 mm and 40 mm,
respectively. When the crystal is centered at the Fourier plane with $\Delta$
= 0, the setup is the same with that in [30]. In this case, the simulated
accuracy is only 77.25% and experimental accuracy is 81.25%. This discrepancy
is likely attributed to the measurement errors in the upconverted SF power
readings for high-order pump projection modes. When the crystal is displaced
by $\Delta=27$ mm, both the simulated and experimental accuracy reached above
97% as shown in Fig. 2(b) and (e), respectively. This shows the advantage of
feature extraction in both spatial and Fourier domains. Figure 2(c) and (f)
show that the accuracy can be further increased over 99% with $\Delta=40$ mm.
To test the robustness of our technique against noise, we next mix the images
with uniformly distributed random noise. The noise level is quantified by the
signal-to-noise ratio (SNR), defined as:
$\textrm{SNR}=10\leavevmode\nobreak\
\textrm{log}_{10}(\sigma_{s}^{2}/\sigma_{n}^{2}),$ (4)
where $\sigma_{s}^{2}$ and $\sigma_{n}^{2}$ are the variance of per-pixel
phase values in the digit images and the added noise masks, respectively.
Figure 3(a) shows an example of one uniformly distributed noise mask with SNR
= -10 dB and cluster size $\textrm{S}=104$ $\mu$m, where the cluster size is
the number of pixels in one noise cluster (which is 10 SLM pixels in this
case). The white noise masks are added to the original handwritten digit
images and are uploaded on SLM1 as the inputs. The results in Fig. 3(b) show
that, with a larger SNR, the accuracy is higher and the difference is smaller
between the simulated and experimental results. Similar to those in Fig. 2 for
clear images, for the noise-contaminated images, both results reveal better
performance with $\Delta$ = 40 mm than $\Delta$ = 27 mm. The advantage reaches
12.18% when the SNR is -15 dB. This again validates our approach for efficient
and robust information extraction.
Figure 4: The accuracy for different sets of LG modes with different azimuthal
($l$) and radial ($p$) indexes, with $\textrm{SNR}=-15$ dB and $\textrm{S}=52$
$\mu$m.
In addition to SNR, the cluster size S is also an import factor to the
classification accuracy. When the noise feature size is comparable with the
image’s characteristic features, it is difficult to distinguish them in both
original and Fourier domains. The simulated and experimental results with
SNR=-15 dB and $\Delta$ = 40 mm in Fig. 3 (b) are 59% and 72.68%,
respectively. However, when we reduce the cluster size by half
($\textrm{S}=52$ $\mu$m), the accuracy increases to 91.25% and 93.4% in
simulation and experiment, respectively.
Thus far, 40 LG modes {LG${}^{p}_{l}$} are chosen for the pump, with
$l\in[-2,2]$ and $p\in[0,7]$, to achieve the accuracy for clear images while
aiming at high efficiency. In practice, one may further optimize the mode
bases for a particular application. To illustrate this opportunity, Fig. 4
plots the simulated accuracy using different sets of LG mode basis with
$\textrm{SNR}=-15$ dB and $\textrm{S}=52$ $\mu$m. It shows that with more
modes, the accuracy can be significantly higher when more than 40 LG modes are
used. With 176 LG modes ($l\in[-5,5]$ and $p\in[0,15]$), the accuracy improved
to 96% from 91%. On the other hand, when the number of LG modes is smaller
than the 40 modes, the accuracy could drop quickly from 87% with $l\in[-2,2]$
and $p\in[0,4]$ to 48% with $l=0$ and $p\in[0,4]$).
Figure 5: Simulated normalized confusion matrices using 49 HG pump modes with
(a) $\Delta$ = 27 mm and (b) $\Delta$ = 40 mm.
Lastly, we compare the choices of LG and HG pump modes to drive the
upconversion for image classification. HG modes are another popular class of
orthogonal mode basis which, unlike LG modes of rotational symmetric
intensity, exhibit rectangular symmetry in intensity. Figure 5 shows the
simulated results with the 49 lowest order HG modes, {HGmn} with $m$ and
$n\in[0,7]$. The classification accuracy of the same data base is 88.5% with
$\Delta=27$ mm and 96.75% with $\Delta=40$ $mm$, which are all inferior to the
simulated results with 40 LG modes in Fig. 2(b) and (c). Therefore, LG modes
are more efficient and suitable for the feature extractions of the handwritten
digits than HG modes. Of course, in practice the optimum mode choice shall
depend on the spatial characteristics of the target images.
In summary, we have demonstrated a new approach to efficient and robust image
classification using a hybrid system integrating nonlinear optics and a deep
neural network. It combines mode selective frequency up-conversion and partial
Fourier transformation to sample the signature features in the original and
Fourier-transformed images through a single nonlinear optical stage. The
partial Fourier transform is conveniently implemented by moving the up-
conversion crystal away from the focal point. It significantly improves the
classification accuracy for both clear and noise-contaminated images. Our
results tap on an inherent advantage of coherent optics for complex machine
learning tasks: the ease and flexibility of linear algebra operations for high
volume data. The robustness of this approach against strong noise invites its
future applications in challenging tasks such as identifying occluded targets,
wide-field surveillance, and remote sensing under low visibility.
Disclosures. The authors declare no conflicts of interest.
## References
* [1] O. A. Rosso, R. Ospina, and A. C. Frery, PloS one 11, e0166868 (2016).
* [2] A. T. Sahlol, D. Yousri, A. A. Ewees, M. A. A. Al-qaness, R. Damasevicius, and M. A. Elaziz, Scientific Reports 10, 15364 (2020).
* [3] L. Deng, G. Hinton, and B. Kingsbury, “New types of deep neural network learning for speech recognition and related applications: An overview,” in _2013 IEEE international conference on acoustics, speech and signal processing,_ (IEEE, 2013), pp. 8599–8603.
* [4] A. Krizhevsky, I. Sutskever, and G. E. Hinton, Communications of the ACM 60, 84 (2017).
* [5] G. Wang, J. C. Ye, and B. De Man, Nature Machine Intelligence 2, 737 (2020).
* [6] J. Schmidhuber, Neural Networks 61, 85 (2015).
* [7] I. Goodfellow, Y. Bengio, and A. Courville, _Deep Learning_ (MIT Press, 2016). http://www.deeplearningbook.org.
* [8] A. A. Moreno, J. Olivito, J. Resano, and H. Mecha, IEEE Transactions on Very Large Scale Integration (VLSI) Systems 28, 1993 (2020).
* [9] I. Chakraborty, G. Saha, and K. Roy, Phys. Rev. Applied 11, 014063 (2019).
* [10] R. Hamerly, L. Bernstein, A. Sludds, M. Soljačić, and D. Englund, Phys. Rev. X 9, 021032 (2019).
* [11] J. Feldmann, N. Youngblood, M. Karpov, H. Gehring, X. Li, M. Stappers, M. Le Gallo, X. Fu, A. Lukashchuk, A. S. Raja, J. Liu, C. D. Wright, A. Sebastian, T. J. Kippenberg, W. H. P. Pernice, and H. Bhaskaran, Nature 589, 52 (2021).
* [12] S. Jiao, J. Feng, Y. Gao, T. Lei, Z. Xie, and X. Yuan, Opt. Lett. 44, 5186 (2019).
* [13] Z. Zhang, X. Li, S. Zheng, M. Yao, G. Zheng, and J. Zhong, Opt. Express 28, 13269 (2020).
* [14] D. Brunner, A. Marandi, W. Bogaerts, and A. Ozcan, Nanophotonics 9, 4053 (2020).
* [15] A. Goy, K. Arthur, S. Li, and G. Barbastathis, Phys. Rev. Lett. 121, 243902 (2018).
* [16] J. Qian, S. Feng, Y. Li, T. Tao, J. Han, Q. Chen, and C. Zuo, Opt. Lett. 45, 1842 (2020).
* [17] G. R. Steinbrecher, J. P. Olson, D. Englund, and J. Carolan, npj Quantum Information 5, 1 (2019).
* [18] G. Wetzstein, A. Ozcan, S. Gigan, S. Fan, D. Englund, M. Soljačić, C. Denz, D. A. B. Miller, and D. Psaltis, Nature 588, 39 (2020).
* [19] P. Latorre-Carmona, V. J. Traver, J. S. Sánchez, and E. Tajahuerce, Image and Vision Computing 86, 28 (2019).
* [20] Y. J. jae Shin, W. Shin, T. Taniguchi, K. Watanabe, P. Kim, and S.-H. Bae, 2D Materials (2020).
* [21] P. A. Xu, A. K. Mishra, H. Bai, C. A. Aubin, L. Zullo, and R. F. Shepherd, Science Robotics 4, eaaw6304 (2019).
* [22] D. P. Sullivan, C. F. Winsnes, L. Åkesson, M. Hjelmare, M. Wiking, R. Schutten, L. Campbell, H. Leifsson, S. Rhodes, A. Nordgren, K. Smith, B. Revaz, B. Finnbogason, A. Szantner, and E. Lundberg, Nature Biotechnology 36, 820 (2018).
* [23] Y. Zuo, B. Li, Y. Zhao, Y. Jiang, Y.-C. Chen, P. Chen, G.-B. Jo, J. Liu, and S. Du, Optica 6, 1132 (2019).
* [24] X. Qiu, D. Zhang, W. Zhang, and L. Chen, Phys. Rev. Lett. 122, 123901 (2019).
* [25] R. Ziv, A. Dikopoltsev, T. Zahavy, I. Rubinstein, P. Sidorenko, O. Cohen, and M. Segev, Optics Express 28, 7528 (2020).
* [26] T. Zahavy, A. Dikopoltsev, D. Moss, G. I. Haham, O. Cohen, S. Mannor, and M. Segev, Optica 5, 666 (2018).
* [27] G. Marcucci, D. Pierangeli, and C. Conti, Phys. Rev. Lett. 125, 093901 (2020).
* [28] D. Ballarini, A. Gianfrate, R. Panico, A. Opala, S. Ghosh, L. Dominici, V. Ardizzone, M. De Giorgi, G. Lerario, G. Gigli _et al._ , Nano Letters 20, 3506 (2020).
* [29] S. Kumar, H. Zhang, and Y.-P. Huang, Communications Physics 3, 108 (2020).
* [30] T. Bu, S. Kumar, H. Zhang, I. Huang, and Y.-P. Huang, Optics Letters 45, 6771 (2020).
* [31] P. Rehain, Y. M. Sua, S. Zhu, I. Dickson, B. Muthuswamy, J. Ramanathan, A. Shahverdi, and Y.-P. Huang, Nature Communications 11, 921 (2020).
* [32] J. C. Bowman and Z. Ghoggali, Journal of Scientific Computing 76, 1578 (2018).
* [33] Y. chan Park, J.-G. Jang, and U. Kang, “Fast partial fourier transform,” (2021).
* [34] S. Restuccia, D. Giovannini, G. Gibson, and M. Padgett, Opt. Express 24, 27127 (2016).
* [35] S. Kumar, H. Zhang, S. Maruca, and Y.-P. Huang, Opt. Lett. 44, 98 (2019).
* [36] H. Zhang, S. Kumar, and Y.-P. Huang, Scientific Reports 9, 17481 (2019).
* [37] G. Agrawal, _Nonlinear Fiber Optics_ (Elsevier Science, 2007).
* [38] Y. LeCun and C. Cortes, “Mnist handwritten digit database,” (2010).
* [39] J. W. Goodman, _Introduction to Fourier optics_ (McGraw-hill, 2008).
ref
|
IMPARTIAL GAMES WITH ENTAILING MOVES
Urban<EMAIL_ADDRESS>
School of Computing, National University of Singpore, Singapore
Richard J<EMAIL_ADDRESS>
Department of Mathematics and Statistics, Dalhousie University, Canada
Carlos P. Santos333Partially supported by UID/MAT/04721/2019 strategic
project<EMAIL_ADDRESS>
Center for Functional Analysis, Linear Structures and Applications,
University of Lisbon & ISEL--IPL
###### Abstract
Combinatorial Game Theory has also been called ‘additive game theory’,
whenever the analysis involves sums of independent game components. Such
disjunctive sums invoke comparison between games, which allows abstract values
to be assigned to them. However, there are rulesets with entailing moves that
break the alternating play axiom and/or restrict the other player’s options
within the disjunctive sum components. These situations are exemplified in the
literature by a ruleset such as nimstring, a normal play variation of the
classical children’s game dots & boxes, and top entails, an elegant ruleset
introduced in the classical work Winning Ways, by Berlekamp Conway and Guy.
Such rulesets fall outside the scope of the established normal play theory.
Here, we axiomatize normal play via two new terminating games,
${\boldsymbol{\infty}}$ (Left wins) and $\overline{\infty}$ (Right wins), and
a more general theory is achieved. We define affine impartial, which extends
classical impartial games, and we analyze their algebra by extending the
established Sprague-Grundy theory, with an accompanying minimum excluded rule.
Solutions of nimstring and top entails are given to illustrate the theory.
## 1 Introduction
Combinatorial Game Theory (CGT), as described in [1, 3, 5, 8], considers
disjunctive sums of normal play games. In order to evaluate the outcome of a
sum of such games, it suffices to analyze the components individually, and
then add the individual values.
However, some classical impartial rulesets, such as nimstring and top entails
fall slightly outside the usual CGT axioms. In nimstring, certain moves
require a player to play again, or carry-on, which is a violation of the
alternating play axiom. And in top entails, certain moves enforce the next
player to play in the same component, which violates the standard definition
of a disjunctive sum. Thus, values of individual components is no longer a
relevant measure, given the standard CGT axioms. The type of moves mentioned
in this paragraph will be gathered under the term entailing moves.444Entailing
means “involve something as a necessary or inevitable part or consequence”.
The purpose of this paper is to extend impartial normal play games
sufficiently to include games with entailing moves. While accomplishing this,
we expand the classical Sprague-Grundy theory to fit this extension.
We will rebuild the normal play axioms by using so-called terminating games,
or infinities, ${\boldsymbol{\infty}}$ and $\overline{\infty}$. Here we focus
on the impartial setting, and the general comprehensive theory for partizan
games will appear in [6].555There are partizan rulesets, in the literature and
in recreational play, with similar entailing and terminating moves. Probably
the most prominent ones are the game of chess, and a version of go called
atari go. These theories are called affine impartial and affine normal play
respectively.
Although we consider only impartial games in this paper, we will keep the
players distinguished as Left and Right. In particular, Left wins if either
player plays to ${\boldsymbol{\infty}}$, in any component, and Right wins in
case of play to $\overline{\infty}$. Note that the normal play zero is
restored by defining
$0=\left\\{\overline{\infty}\\!\mid\\!{\boldsymbol{\infty}}\right\\}$, a first
player losing position.
It is well-known that, in classical Combinatorial Game Theory, the impartial
values are nimbers. We will prove that there is exactly one more value modulo
affine impartial, a game $\leftmoon$ , called the _moon_. This value was
anticipated in the classical work Winning Ways, by Berlekamp, Conway and Guy.
In [3], volume 2, page 398, one can read “A loony move is one that loses for a
player, no matter what other components are.”. Before developing the theory,
we illustrate how the infinities are used in the motivating rulesets,
nimstring ([2, 3]) and top entails ([3], volume 2).
Let us first briefly mention the organization of the paper. To facilitate the
development of the new impartial theory, Section 2 considers the basic
properties of unrestricted affine normal play, aiming for a game comparison
result, Theorem 14.The affine impartial theory is developed in Section 3. The
main result is Theorem 34 which shows that values in this extension are the
nimbers plus one more value. Theorem 39 gives an algorithm to find the value
of a given position, and notably, if there are no infinities in the options,
then the nimbers are obtained by the usual mex-rule. We finish off with two
case studies. In Section 4, we compute the value of an interesting nimstring
position, anticipated in Section 1.1. In Section 5, we compute the values for
top entail heaps of sizes $1$ through $12$, and Theorem 41 provides
theoretical justification for computing top entails values.
### 1.1 The ruleset nimstring
In nimstring, a player draws a line between two horizontally or vertically
adjacent points, in a finite grid, not already joined by a line. If a player
completes a $1\times 1$ square, then they must draw another line, and if they
cannot do so, they lose.
Figure 1 shows an example position, where no square can be completed in the
next move. Later, through the new theory, we will see that the position $H$
equals $*2$ modulo affine impartial.
Figure 1: A nimstring position, $H$.
In Figure 2, we show a position, $G$, with two options, one of which is an
entailing move. Namely, if the the top bar is drawn, the next player
continues, but if the middle bar is drawn, then the current player has to
carry-on.
Figure 2: A nimstring position, $G$, with its two options, a ‘double-box’ and
an entailing carry-on position.
When we develop the theory, we will see that the position $G$, to the left in
Figure 2, is the abstract game
$\displaystyle\left\\{\left\\{{\boldsymbol{\infty}}\\!\mid\\!0\right\\}\\!,0\\!\mid\\!\left\\{0\\!\mid\\!\overline{\infty}\right\\}\\!,0\right\\}.$
(1)
The option 0 is obtained by drawing the top bar. The intuition for this is as
follows: the player who moves inside the ‘double box’ loses, if $G$ is played
in isolation, because they have to play again.
If a player draws the middle bar in $G$, then they have to carry-on, and this
is represented by the abstract option
$\left\\{{\boldsymbol{\infty}}\\!\mid\\!0\right\\}$, if Left moved. There is
an infinite urgency in this game: Right has to play here, or lose. And so, the
effect is the desired: Left plays again, and alternating play is restored.
Hence disjunctive sum play is also restored, within the affine impartial
convention. Moreover, the Right option in this threat should be 0, because
Left loses by playing this option if $G$ is played alone. If the sum is $G+H$,
with $H$ as in Figure 1, then the next player wins, by playing this entailing
middle bar in $G$.
### 1.2 The ruleset top entails
Top entails is played on heaps of tokens. A player may either remove the top
token from exactly one heap, or split a heap into two non-empty heaps. If the
top token is removed from a heap, then the next move (in alternating play)
must be played on the same heap.
A heap with one token, say $H$, is a first player win, in any situation.
Namely, a move in $H$ forces the opponent to play in the same heap, where no
move remains. Note that the abstract game
$H=\left\\{{\boldsymbol{\infty}}\\!\mid\\!\overline{\infty}\right\\}$ settles
this behaviour. The player who moves first in $H$ wins independently of
existence of other components. The point we wish to make here is that this
abstract representation settles the problem of independecny of a heap of size
one with other disjunctive sum components.
Consider $G$, in Figure 3, a pile of size 3. There are two options, as
depicted in Figures 4 and 5.
Figure 3: A pile $G$ of top entails, of size 3.
The option in Figure 4 splits $G$ into two piles and the next player’s options
are unrestricted. By the terminating effect of playing in a heap of size one,
this composite game should be equal to the game
$H=\left\\{{\boldsymbol{\infty}}\\!\mid\\!\overline{\infty}\right\\}$.
Figure 4: The game $G$ is split into two components.
The option in Figure 5 is an entailing move, and the next player must continue
in this component, even if other moves are available. Therefore, the game form
of the entailing option in Figure 5 is
$\left\\{{\boldsymbol{\infty}}\\!\mid\\!\bf 1+\bf 1,\bf
1_{\text{entail}}\right\\},$
if Left just moved, and where 1 denotes a heap of size one. The terminating
threat forces Right to play here, instead of possibly using other options.
Figure 5: An entailing option.
Intuitively, either way of responding reduces to a game of the form
$H=\left\\{{\boldsymbol{\infty}}\\!\mid\\!\overline{\infty}\right\\}$. In
conclusion, the heap of size three should be equal to the game $H$, and
disjunctive sum play has been restored. All this intuition will be rigorously
justified in the coming complete theory for affine impartial play.
It turns out that affine impartial games require only a small extension to the
Sprague-Grundy theory. Namely, the game in (1), obtained from the nimstring
position in Figure 2, equals the game
$H=\left\\{\overline{\infty}\\!\mid\\!{\boldsymbol{\infty}}\right\\}$ in the
previous paragraph, modulo affine impartial, and later we will devote the
value ‘ $\leftmoon$ ’ to the equivalence class of such games.
## 2 Affine literal forms and order
This section aims at Theorem 13, a comparison theorem for affine normal play
that suffices for the purpose of this paper. We begin by defining the
fundamental concepts for affine normal play, and we wait with the restriction
to affine impartial until the next section.
In classical Combinatorial Game Theory, the normal play forms, $\mathbb{Np}$,
are recursively constructed from the empty set. The form
$\\{\varnothing\,|\,\varnothing\\}=0$ is the only form of day zero and the
only form without options. The forms $\\{0\,|\,\varnothing\\}=1$ ,
$\\{\varnothing\,|\,0\\}=-1$, $\\{0\,|\,0\\}=*$, are born on day 1, and so on.
The forms of affine normal play, denoted $\mathbb{Np}^{\infty}$, are
recursively constructed from the games $\infty$ (infinity) and
$\overline{\infty}$ (minus infinity) [6]. The forms $\infty$ and
$\overline{\infty}$ are the only forms without options. The forms
$\left\\{\overline{\infty}\\!\mid\\!{\boldsymbol{\infty}}\right\\}=0,\left\\{{\boldsymbol{\infty}}\\!\mid\\!\overline{\infty}\right\\}=\pm{\boldsymbol{\infty}},\left\\{{\boldsymbol{\infty}}\\!\mid\\!{\boldsymbol{\infty}}\right\\}$
and $\left\\{\overline{\infty}\\!\mid\\!\overline{\infty}\right\\}$ are born
on day zero. And so on.
The order of $\mathbb{Np}^{\infty}$ is defined in the standard way. Consider
the four perfect play outcome classes $\mathscr{L}$ (Left wins), $\mathscr{N}$
(Next player wins), $\mathscr{P}$ (Previous player wins), and $\mathscr{R}$
(Right wins). From Left’s perspective, the first outcome is the best (she
wins, regardless of playing first or second) and the fourth is the worst (she
loses, regardless of whether playing first or second). On the other hand,
regarding $\mathscr{N}$ and $\mathscr{P}$, the victory depends on playing
first or second, so these outcomes are not comparable. These considerations
explain the partial order in an ‘outcome diamond’:
We write $G\in\mathscr{L}$, or equivalently $o(G)=\mathscr{L}$, if the outcome
of $G\in\mathbb{Np}^{\infty}$ is Left wins, and so on. The evaluation of games
in $\mathbb{Np}^{\infty}$ is based on the following axiomatic list:
###### Axiom 1 (Absorbing Nature of Infinities).
The infinities satisfy
1. 1.
$\infty\in\mathscr{L}$;
2. 2.
$\overline{\infty}\in\mathscr{R}$;
3. 3.
For all $X\in\mathbb{Np}^{\infty}\setminus\\{\overline{\infty}\\}$,
$\infty+X=\infty$;
4. 4.
For all $X\in\mathbb{Np}^{\infty}\setminus\\{{\boldsymbol{\infty}}\\}$,
$\overline{\infty}+X=\overline{\infty}$;
5. 5.
‘$\infty+\overline{\infty}$’ is not defined.
Addition of games is defined as usual, apart from items 3 and 4. The fifth
item is natural in terms of perfect play, since if ${\boldsymbol{\infty}}$
appears, then $\overline{\infty}$ cannot appear and vice versa.
The definitions of equality and partial order of games are based on the
outcome diamond.
###### Definition 2 (Order and Equality of Games).
Let $G,H\in\mathbb{Np}^{\infty}$. Then, $G\succcurlyeq H$ if, for every form
$X\in\mathbb{Np}^{\infty}\setminus\\{\infty,\overline{\infty}\\}$,
$o(G+X)\geqslant o(H+X)$. Moreover $G=H$ if $G\succcurlyeq H$ and
$H\succcurlyeq G$.
Note that the exclusion of the infinities does not diminish the generality of
the definition, but is necessary due to Axiom 5. As usual, we have the
following observations. If $G=H$ then replacing $H$ by $G$ or $G$ by $H$ do
not hurt the players under any circumstances. Similarly, if $G\succcurlyeq H$
then replacing $H$ by $G$ does not hurt Left, and replacing $G$ by $H$ does
not hurt Right.
###### Theorem 3.
Let $G\in\mathbb{Np}^{\infty}$. Then $\infty\succcurlyeq G$ and
$G\succcurlyeq\overline{\infty}$.
###### Proof.
If $X\in\mathbb{Np}^{\infty}\setminus\\{\infty,\overline{\infty}\\}$ then, by
Axiom 3, $\infty+X=\infty$. Hence, by Axiom 1,
$o(\infty+X)=o(\infty)=\mathscr{L}$. Therefore, for every
$X\in\mathbb{Np}^{\infty}\setminus\\{\infty,\overline{\infty}\\}$, we have
$o(\infty+X)\geqslant o(G+X)$, and so $\infty\succcurlyeq G$. Proving that
$G\succcurlyeq\overline{\infty}$ is analogous. ∎
The concept of a check is fundamental to $\mathbb{Np}^{\infty}$. Indeed, this
is an alternative, and perhaps more explicit, at least for those Chess playing
readers, term for an entailing move, as seen in the Introduction.
###### Definition 4 (Check Games).
Consider $G\in\mathbb{Np}^{\infty}$. If $\infty\in{G^{\mathcal{L}}}$
($\overline{\infty}\in{G^{\mathcal{R}}}$) then $G$ is a _Left-check_ (_Right-
check_). If $G$ is a Left-check or a Right-check then $G$ is a _check_. Denote
by $G^{\mathrel{\mathop{L}\limits^{\vbox
to0.0pt{\kern-1.5pt\hbox{$\scriptstyle\rightarrow$}\vss}}}}$
($G^{\mathrel{\mathop{R}\limits^{\vbox
to0.0pt{\kern-1.5pt\hbox{$\scriptstyle\leftarrow$}\vss}}}}$) a Left (Right)
option of $G$ that is a Left-check (Right-check).
Of course, all checks are asymmetric, apart from the ‘trivial check’,
$\left\\{{\boldsymbol{\infty}}\\!\mid\\!\overline{\infty}\right\\}$. A player
would not use this check, because the opponent ‘check mates’ by defending.
###### Definition 5 (Quiet Games).
Let $G\in\mathbb{Np}^{\infty}$. If $G\neq{\boldsymbol{\infty}}$
($G\neq\overline{\infty}$) and $G$ is not a Left-check (Right-check) then $G$
is _Left-quiet_ (_Right-quiet_). If $G$ is Left-quiet and Right-quiet then $G$
is _quiet_.
###### Definition 6 (Conway Forms and Games).
A game $G\in\mathbb{Np}^{\infty}$ is a _Conway form_ if
$G\not\in\\{\infty,\overline{\infty}\\}$, and $G$ has no checks as followers.
Let $\mathbb{Np}^{\rm C}\subseteq\mathbb{Np}^{\infty}$ denote the substructure
of Conway forms. A game $G\in\mathbb{Np}^{\infty}$ is a _Conway game_ if it
equals a Conway form.
###### Example 7.
The game
$G=\left\\{\left\\{\overline{\infty}\\!\mid\\!\infty\right\\}\\!\mid\\!\left\\{\overline{\infty}\\!\mid\\!\infty\right\\}\right\\}=\left\\{0\\!\mid\\!0\right\\}=*$
is a Conway form (no checks as followers). The game
$G^{\prime}=\left\\{\left\\{\infty\\!\mid\\!*\right\\}\\!\mid\\!\left\\{*\\!\mid\\!\overline{\infty}\right\\}\right\\}$
is not a Conway form because there are checks as followers. However, later, we
will see that $G^{\prime}=G$. Therefore, $G^{\prime}$ is a Conway game.
In general, when we say form, we mean the literal form, and when we say game,
we usually mean (any member in) the full equivalence class of games. When we
write $G\in\mathbb{Np}^{\infty}$, we usually refer to the literal form, but
the context may decide.
Some classical theorems are still available in $\mathbb{Np}^{\infty}$.
###### Theorem 8 (Fundamental Theorem of Affine Normal Play).
If $G\in\mathbb{Np}^{\infty}$ then $G\succcurlyeq 0$ if and only if
$G\in\mathscr{L}\cup\mathscr{P}$.
###### Proof.
Assume that $G\succcurlyeq 0$. We have $0\in\mathscr{P}$, and so, by order of
outcomes, $G\in\mathscr{L}\cup\mathscr{P}$.
Suppose now that $G\in\mathscr{L}\cup\mathscr{P}$. If $G=\infty$, by Theorem
3, $G\succcurlyeq 0$; hence, assume $G\neq\infty$. Let
$X\in\mathbb{Np}^{\infty}\setminus\\{\infty,\overline{\infty}\\}$.
If, playing first, Left wins $X$ with the option $X^{L}$, then she also wins
$G+X$ with the option $G+X^{L}$. Essentially, she mimics the strategy used
when $X$ is played alone, answering locally when Right plays in $G$. Due to
the assumption $G\in\mathscr{L}\cup\mathscr{P}$, this is a winning strategy
for Left in $G+X$.
If Left, playing second, wins $X$. Then, on $G+X$, she can respond to each of
Right’s moves locally, with a winning move on the same component, because
$G\in\mathscr{L}\cup\mathscr{P}$. Thus Left can win $G+X$ playing second.
Therefore, $o(G+X)\geqslant o(X)$ and so, $G\succcurlyeq 0$. ∎
###### Corollary 9 (Order-Outcome Bijection).
If $G\in\mathbb{Np}^{\infty}$ then
* •
$G\succ 0$ if and only if $G\in\mathscr{L}$;
* •
$G=0$ if and only if $G\in\mathscr{P}$;
* •
$G\operatorname{\|}0$ if and only if $G\in\mathscr{N}$;
* •
$G\prec 0$ if and only if $G\in\mathscr{R}$.
###### Proof.
The statement of Theorem 8 can equivalently be “$G\preccurlyeq 0$ if and only
if $G\in\mathscr{R}\cup\mathscr{P}$”, so we can use that fact too.
Suppose that $G\succ 0$. By Theorem 8, $G\in\mathscr{L}\cup\mathscr{P}$. But,
we cannot have $G\in\mathscr{P}$, for otherwise
$G\in\mathscr{R}\cup\mathscr{P}$ and $G\preccurlyeq 0$. Therefore,
$G\in\mathscr{L}$. Conversely, suppose that $G\in\mathscr{L}$. By Theorem 8,
we have $G\succcurlyeq 0$. But, we cannot have $G=0$, for otherwise
$G\preccurlyeq 0$, and $G\in\mathscr{R}\cup\mathscr{P}$. Hence, $G\succ 0$.
Thus, the first equivalence holds.
The proof of the fourth equivalence is analogous.
For the second equivalence, if $G=0$, then $G\succcurlyeq
0\;\wedge\;G\preccurlyeq 0$. So,
$G\in(\mathscr{L}\cup\mathscr{P})\cap(\mathscr{R}\cup\mathscr{P})=\mathscr{P}$.
The third equivalence is a consequence of eliminating all other possibilities.
∎
It is known that $\mathbb{Np}$ is a group. By Corollary 9 we may deduce that
$\mathbb{Np}^{\infty}$ is only a monoid. Namely, if $G=\\{\infty\,|\,0\\}$
then, for any
$X\in\mathbb{Np}^{\infty}\setminus\\{\infty,\overline{\infty}\\}$,
$G+X\in\mathscr{L}\cup\mathscr{N}$ (playing first, Left wins). Hence, for all
$X$, $G+X\neq 0$ and $G$ is non-invertible. Thus, in general, the comparison
of $G$ with $H$ cannot be done by playing the game ‘$G-H$’, because,
sometimes, ‘$-H$’ does not exist.
However, the following theorem shows that not everything is lost.
The _conjugate_ of a given game switches roles of the players.
###### Definition 10 (Conjugate).
The conjugate of $G\in{\mathbb{Np}^{\infty}}$ is
$\mathrel{\mathop{G}\limits^{\vbox
to0.0pt{\kern-1.5pt\hbox{$\scriptstyle\curvearrowleftright$}\vss}}}\,=\begin{cases}\overline{\infty},&\mbox{if
$G=\infty$}\\\ \infty,&\mbox{if $G=\overline{\infty}$}\\\
\left\\{\mathrel{\mathop{{G^{\mathcal{R}}}}\limits^{\vbox
to0.0pt{\kern-1.5pt\hbox{$\scriptstyle\curvearrowleftright$}\vss}}}\,\\!\mid\\!\;\mathrel{\mathop{{G^{\mathcal{L}}}}\limits^{\vbox
to0.0pt{\kern-1.5pt\hbox{$\scriptstyle\curvearrowleftright$}\vss}}}\right\\},&\mbox{otherwise},\end{cases}$
where $\mathrel{\mathop{{G^{\mathcal{L}}}}\limits^{\vbox
to0.0pt{\kern-1.5pt\hbox{$\scriptstyle\curvearrowleftright$}\vss}}}$ denotes
the set of literal forms $\mathrel{\mathop{G^{L}}\limits^{\vbox
to0.0pt{\kern-1.5pt\hbox{$\scriptstyle\curvearrowleftright$}\vss}}}$, for
$G^{L}\in{G^{\mathcal{L}}}$, and similarly for ${G^{\mathcal{R}}}$.
###### Theorem 11.
If $G\in\mathbb{Np}^{\infty}$ is a Conway game, then $G$ is invertible and
$-G=\,\mathrel{\mathop{G}\limits^{\vbox
to0.0pt{\kern-1.5pt\hbox{$\scriptstyle\curvearrowleftright$}\vss}}}$.
###### Proof.
Suppose first that $G$ is a Conway form. If $G=0$ then
$\mathrel{\mathop{G}\limits^{\vbox
to0.0pt{\kern-1.5pt\hbox{$\scriptstyle\curvearrowleftright$}\vss}}}=0$, and
thetheorem holds. Otherwise, let us verify that
$G+\mathrel{\mathop{G}\limits^{\vbox
to0.0pt{\kern-1.5pt\hbox{$\scriptstyle\curvearrowleftright$}\vss}}}$ is a
$\mathscr{P}$-position. If Left, playing first, chooses
$G^{L}+\mathrel{\mathop{G}\limits^{\vbox
to0.0pt{\kern-1.5pt\hbox{$\scriptstyle\curvearrowleftright$}\vss}}}$, because
this game is not $\infty$ ($G$ is not a check), Right can answer with
$G^{L}+\mathrel{\mathop{G^{L}}\limits^{\vbox
to0.0pt{\kern-1.5pt\hbox{$\scriptstyle\curvearrowleftright$}\vss}}}$ and, by
induction, because $G^{L}$ is a Conway form with no checks as followers, that
option is equal to zero. Because of that, by Corollary 9, that option is a
$\mathscr{P}$-position, and Right wins. Analogous arguments work for the other
options of the first player, and so, $G+\mathrel{\mathop{G}\limits^{\vbox
to0.0pt{\kern-1.5pt\hbox{$\scriptstyle\curvearrowleftright$}\vss}}}$ is a
$\mathscr{P}$-position. Again, by Corollary 9,
$G+\mathrel{\mathop{G}\limits^{\vbox
to0.0pt{\kern-1.5pt\hbox{$\scriptstyle\curvearrowleftright$}\vss}}}=0$.
Suppose now that $G$ is not a Conway form. Because it is a Conway game, by
definition, it is equal to some $G^{\prime}\in\mathbb{Np}^{\rm C}$. The first
paragraph proved that $G^{\prime}+\mathrel{\mathop{G^{\prime}}\limits^{\vbox
to0.0pt{\kern-1.5pt\hbox{$\scriptstyle\curvearrowleftright$}\vss}}}=0$. Also,
by symmetry, $\mathrel{\mathop{G}\limits^{\vbox
to0.0pt{\kern-1.5pt\hbox{$\scriptstyle\curvearrowleftright$}\vss}}}$ is equal
to $\mathrel{\mathop{G^{\prime}}\limits^{\vbox
to0.0pt{\kern-1.5pt\hbox{$\scriptstyle\curvearrowleftright$}\vss}}}$.
Therefore, $G^{\prime}+\mathrel{\mathop{G^{\prime}}\limits^{\vbox
to0.0pt{\kern-1.5pt\hbox{$\scriptstyle\curvearrowleftright$}\vss}}}=0$ implies
$G+\mathrel{\mathop{G}\limits^{\vbox
to0.0pt{\kern-1.5pt\hbox{$\scriptstyle\curvearrowleftright$}\vss}}}=0$. ∎
###### Lemma 12.
Let $G,H\in\mathbb{Np}^{\infty}$, and let $J$ be an invertible form of
$\mathbb{Np}^{\infty}$. Then
$G\succcurlyeq H\Leftrightarrow G+J\succcurlyeq H+J.$
###### Proof.
($\Rightarrow$) Consider any
$X\in\mathbb{Np}^{\infty}\setminus\\{\infty,\overline{\infty}\\}$ and let
$X^{\prime}=J+X$. Since $J$ is invertible, $J$ is neither $\infty$ nor
$\overline{\infty}$, and so, $X^{\prime}$ is neither $\infty$ nor
$\overline{\infty}$. Definition of order implies $o(G+X^{\prime})\geqslant
o(H+X^{\prime})$, that is $o(G+J+X))\geqslant o(H+J+X)$. Thus, the
arbitrariness of
$X\in\mathbb{Np}^{\infty}\setminus\\{\infty,\overline{\infty}\\}$ implies
$G+J\succcurlyeq H+J$.
($\Leftarrow$) Consider any
$X\in\mathbb{Np}^{\infty}\setminus\\{\infty,\overline{\infty}\\}$ and let
$X^{\prime}=-J+X$ ($J$ is invertible, i.e. $-J$ exists and $J-J=0$). Since
$-J$ is invertible, $-J$ is neither $\infty$ nor $\overline{\infty}$, and so,
$X^{\prime}$ is neither $\infty$ nor $\overline{\infty}$. By definition of
order, $o(G+J+X^{\prime})\geqslant o((H+J+X^{\prime})$, that is
$o(G+J-J+X)\geqslant o(H+J-J+X)$. Hence, $o(G+X)\geqslant o(H+X)$, and so,
given the arbitrariness of
$X\in\mathbb{Np}^{\infty}\setminus\\{\infty,\overline{\infty}\\}$,
$G\succcurlyeq H$. ∎
###### Theorem 13.
Let $G$ be any form of $\mathbb{Np}^{\infty}$ and suppose that $H$ is an
invertible form of $\mathbb{Np}^{\infty}$. Then,
$G\succcurlyeq H\Leftrightarrow G-H\in\mathscr{L}\cup\mathscr{P}\text{ and
}G=H\Leftrightarrow G-H\in\mathscr{P}.$
###### Proof.
By Lemma 12, $G\succcurlyeq H\Leftrightarrow G-H\succcurlyeq H-H$. Therefore,
we have $G\succcurlyeq H\Leftrightarrow G-H\succcurlyeq 0$. By Theorem 8, this
is the same as $G\succcurlyeq H\Leftrightarrow
G-H\in\mathscr{L}\cup\mathscr{P}$.
Finally, $G=H\Leftrightarrow G-H\in\mathscr{P}$, by $G\succcurlyeq H\wedge
H\succcurlyeq G$. ∎
###### Theorem 14.
Let $G$ be any form of $\mathbb{Np}^{\infty}$ and let
$H\in\mathbb{Np}^{\infty}$ be a Conway game. Then
* •
$G\succcurlyeq H$ if and only if $G\,+\mathrel{\mathop{H}\limits^{\vbox
to0.0pt{\kern-1.5pt\hbox{$\scriptstyle\curvearrowleftright$}\vss}}}\,\in\mathscr{L}\cup\mathscr{P}$
* •
$G=H$ if and only if $G\,+\mathrel{\mathop{H}\limits^{\vbox
to0.0pt{\kern-1.5pt\hbox{$\scriptstyle\curvearrowleftright$}\vss}}}\,\in\mathscr{P}$.
###### Proof.
These are direct consequences of Theorems 11 and 13. ∎
In a follow up paper [6], where we study the full game space
$\mathbb{Np}^{\infty}$, we provide a solution of the general case of
$G\succcurlyeq H$.
## 3 Affine impartial theory
In order to propose an extension of the Sprague-Grundy theory, we first define
the concept of an affine impartial game.666In terms of ruleset: here ‘affine
impartial’ is an abbreviation of affine normal play impartial, in the sense
that if the player to move cannot complete their move they lose. Of course,
rulesets like nimstring should be impartial.
###### Definition 15 (Symmetric Game).
Consider a form $G\in{\mathbb{Np}^{\infty}}$. Then $G$ is _symmetric_ if
$G\not\in\\{\infty,\overline{\infty}\\}$ and
${G^{\mathcal{R}}}=\;\mathrel{\mathop{{G^{\mathcal{L}}}}\limits^{\vbox
to0.0pt{\kern-1.5pt\hbox{$\scriptstyle\curvearrowleftright$}\vss}}}$.
###### Definition 16 (Affine Impartial).
A form $G\in\mathbb{Np}^{\infty}$ is _affine impartial_ if it is symmetric and
all quiet followers of $G$ are symmetric. The subset of affine impartial games
is $\mathbb{Im}^{\boldsymbol{\infty}}\subset{\mathbb{Np}^{\infty}}$.
Of course, a non-quiet game either has no option, or is a check, and so
(unless a trivial check) is by definition asymmetric. But this is the only
exception of symmetry in the world of affine impartial impartial games. It is
easy to check that $\mathbb{Im}^{\boldsymbol{\infty}}$ satisfies the standard
closure properties of combinatorial games, i.e. closure of taking options,
addition, and conjugates.
The following result must hold for any class of games that claims to be
“impartial”.
###### Theorem 17 (Affine Impartial Outcomes).
If $G$ is a symmetric form, then $G\in\mathscr{N}\cup\mathscr{P}$.
###### Proof.
This proof uses a strategy-stealing argument. Suppose that $G\in\mathscr{L}$.
Then Left wins $G$ playing first with some option $G^{L}$. Hence, by symmetry,
Right wins $G$ playing first with $\mathrel{\mathop{G^{L}}\limits^{\vbox
to0.0pt{\kern-1.5pt\hbox{$\scriptstyle\curvearrowleftright$}\vss}}}$. That
contradicts $G\in\mathscr{L}$. A similar argument holds against
$G\in\mathscr{R}$. ∎
We want to restrict our analysis to $\mathbb{Im}^{\boldsymbol{\infty}}$.
Therefore, we define equality modulo $\mathbb{Im}^{\boldsymbol{\infty}}$.
###### Definition 18 (Impartial Equality).
Consider forms $G,H\in\mathbb{Im}^{\boldsymbol{\infty}}$. Then,
$G=_{\mathbb{Im}^{\boldsymbol{\infty}}}H$ if, for every form
$X\in\mathbb{Im}^{\boldsymbol{\infty}}$, $o(G+X)=o(H+X)$.
###### Observation 19.
Of course, $G=H$ in $\mathbb{Np}^{\infty}$ implies
$G=_{\mathbb{Im}^{\boldsymbol{\infty}}}H$. The opposite direction is not true.
We can have $G=_{\mathbb{Im}^{\boldsymbol{\infty}}}H$ and $G\neq H$ in
$\mathbb{Np}^{\infty}$, if the there is no distinguishing game in
$\mathbb{Im}^{\boldsymbol{\infty}}$. A simple example is
$G=\left\\{\left\\{{\boldsymbol{\infty}}\\!\mid\\!0\right\\}\\!,0\\!\mid\\!\left\\{0\\!\mid\\!\overline{\infty}\right\\}\\!,0\right\\}$
and
$H=\left\\{\left\\{{\boldsymbol{\infty}}\\!\mid\\!*\right\\},*\\!\mid\\!\left\\{*\\!\mid\\!\overline{\infty}\right\\},*\right\\}$.
As we will see, these games are indistinguishable modulo
$\mathbb{Im}^{\boldsymbol{\infty}}$. However, the game
$X=\left\\{0\\!\mid\\!-1\right\\}$ distinguishes them in
${\mathbb{Np}^{\infty}}$; playing first, Left wins $G+X$, but loses $H+X$.
It is easy to verify if a form in $\mathbb{Im}^{\boldsymbol{\infty}}$ equals a
nimber.
###### Theorem 20 (Nimbers).
Let $G\in\mathbb{Im}^{\boldsymbol{\infty}}$. Then,
$G=_{\mathbb{Im}^{\boldsymbol{\infty}}}*n$ if and only if
$G+*n\in\mathscr{P}$.
###### Proof.
Suppose that $G+*n\in\mathscr{P}$. By Theorem 14, $G=*n$ modulo
$\mathbb{Np}^{\infty}$, and so $G=_{\mathbb{Im}^{\boldsymbol{\infty}}}*n$.
Suppose now that $G=_{\mathbb{Im}^{\boldsymbol{\infty}}}*n$. By Theorem 17,
$G+*n\in\mathscr{N}\cup\mathscr{P}$, since impartiality is closed under
addition. If $G+*n\in\mathscr{N}$, then $G+*n\in\mathscr{N}$ and
$*n+*n\in\mathscr{P}$, contradicting
$G=_{\mathbb{Im}^{\boldsymbol{\infty}}}*n$. Hence, $G+*n\in\mathscr{P}$. ∎
###### Notation 21.
Let $\mathbb{nim}\subseteq\mathbb{Im}^{\boldsymbol{\infty}}$ denote the subset
of affine impartial games that equal nimbers.
It is well-known that, in classical Combinatorial Game Theory, the impartial
values are nimbers. We will prove that there is exactly one more value modulo
$\mathbb{Im}^{\boldsymbol{\infty}}$, a game $\leftmoon$ , called _moon_. In
[3], volume 2, page 398, one can read “A loony move is one that loses for a
player, no matter what other components are.”. The following general
definition is motivated by that idea.
###### Definition 22 (Loony Game).
A game $G\in{\mathbb{Np}^{\infty}}$ is _loony_ if, for all quiet
$X\in{\mathbb{Np}^{\infty}}\cap(\mathscr{N}\cup\mathscr{P})$,
$G+X\in\mathscr{N}$.
Thus, in our interpretation, a ‘loony move’ exposes a loony game.
There are no loony games in $\mathbb{Np}$. Suppose that
$G\in{\mathbb{Np}}\cap(\mathscr{P}\cup\mathscr{L}\cup\mathscr{R})$ is a loony
game. Of course, $G+0\in\mathscr{P}\cup\mathscr{L}\cup\mathscr{R}$ and that is
a contradiction. Suppose that $G\in{\mathbb{Np}}\cap\mathscr{N}$ is a loony
game. In that case, if $n$ is large enough,
$G+\left\\{n\\!\mid\\!0\right\\}\in\mathscr{L}$, and that is a contradiction,
since $\left\\{n\\!\mid\\!0\right\\}\in\mathscr{N}$ is quiet.
There are loony games in $\mathbb{Np}^{\infty}$ . The obvious one is
$\pm\infty=\left\\{{\boldsymbol{\infty}}\\!\mid\\!\overline{\infty}\right\\}$,
but we can also have impartial quiet loony moves. Consider
$G=\left\\{\left\\{{\boldsymbol{\infty}}\\!\mid\\!0\right\\}\\!,0\\!\mid\\!\left\\{0\\!\mid\\!\overline{\infty}\right\\}\\!,0\right\\}$
and a quiet $X\in\mathbb{Np}^{\infty}$ such that
$X\in\mathscr{P}\cup\mathscr{N}$. If $X\in\mathscr{P}$, the first player wins
moving to $X$. If $X\in\mathscr{N}$, the first player wins moving to
$\\{\infty\,|\,0\\}+X$ (Left) or to $\\{0\,|\,\overline{\infty}\\}+X$ (Right).
###### Notation 23.
The moon is the game form
$\scalebox{1.1}{$\leftmoon$}=\left\\{{\boldsymbol{\infty}}\\!\mid\\!\overline{\infty}\right\\}$.
When a player moves to $\scalebox{1.1}{$\leftmoon$}+X$, for any
$X\in\mathscr{N}\cup\mathscr{P}$, he “goes to the moon” and loses.
###### Theorem 24 (Loony Uniqueness).
All loony games are equal modulo $\mathbb{Im}^{\boldsymbol{\infty}}$.
###### Proof.
Consider $G$ and $G^{\prime}$, two loony games. We know that all quiet
$X\in\mathbb{Im}^{\boldsymbol{\infty}}$ belong to
$\mathscr{N}\cup\mathscr{P}$. By definition of a loony game, we have
$G+X\in\mathscr{N}$ and $G^{\prime}+X\in\mathscr{N}$. On the other hand, if
$X\in\mathbb{Im}^{\boldsymbol{\infty}}$ is not quiet then $\infty\in
X^{\mathcal{L}}$ and $\overline{\infty}\in X^{\mathcal{R}}$, and hence
$G+X\in\mathscr{N}$ and $G^{\prime}+X\in\mathscr{N}$. In all cases,
$o(G+X)=o(G^{\prime}+X)=\mathscr{N}$ and the theorem holds. ∎
###### Observation 25.
Two loony games may be different modulo ${\mathbb{Np}^{\infty}}$, but equal
modulo $\mathbb{Im}^{\boldsymbol{\infty}}$. The games
$\left\\{\left\\{{\boldsymbol{\infty}}\\!\mid\\!0\right\\}\\!,0\\!\mid\\!\left\\{0\\!\mid\\!\overline{\infty}\right\\}\\!,0\right\\}$
and
$\left\\{\left\\{{\boldsymbol{\infty}}\\!\mid\\!2\right\\},0\\!\mid\\!\left\\{-2\\!\mid\\!\overline{\infty}\right\\},0\right\\}$
are loony. These games are different modulo $\mathbb{Np}^{\infty}$. Left,
playing first, loses
$\left\\{\left\\{{\boldsymbol{\infty}}\\!\mid\\!0\right\\}\\!,0\\!\mid\\!\left\\{0\\!\mid\\!\overline{\infty}\right\\}\\!,0\right\\}-1$
and wins
$\left\\{\left\\{{\boldsymbol{\infty}}\\!\mid\\!2\right\\},0\\!\mid\\!\left\\{-2\\!\mid\\!\overline{\infty}\right\\},0\right\\}-1$.
However, as will follow by theory developed here, one cannot distinguish these
two games modulo $\mathbb{Im}^{\boldsymbol{\infty}}$.
In order to prove an affine impartial minimum excluded rule, we separate the
options into two classes.
###### Definition 26 (Immediate Nimbers).
Let $G\in\mathbb{Im}^{\boldsymbol{\infty}}$. The set of $G$-_immediate
nimbers_ , denoted $S_{G}$ is the set
$S_{G}={G^{\mathcal{L}}}\cap\mathbb{nim}$.
Not that, by symmetry, $S_{G}={G^{\mathcal{R}}}\cap\mathbb{nim}$, and note
that $S_{\scalebox{0.9}{$\leftmoon$}}=\varnothing$.
###### Definition 27 (Protected Nimbers).
Conisder a game form $G\in\mathbb{Im}^{\boldsymbol{\infty}}$. The set of
$G$-_protected nimbers_ $P_{G}$ is
1. 1.
$P_{G}=\mathbb{nim}$, if $\infty\in{G^{\mathcal{L}}}$;
2. 2.
$P_{G}=\\{*n:G^{\mathrel{\mathop{L}\limits^{\vbox
to0.0pt{\kern-1.5pt\hbox{$\scriptstyle\rightarrow$}\vss}}}}+*n\in\mathscr{L},G^{\mathrel{\mathop{L}\limits^{\vbox
to0.0pt{\kern-1.5pt\hbox{$\scriptstyle\rightarrow$}\vss}}}}\in{G^{\mathcal{L}}}\\}$,
otherwise.
The second item says: if $\infty\not\in{G^{\mathcal{L}}}$ then $*n\in P_{G}$
if there is a check $G^{\mathrel{\mathop{L}\limits^{\vbox
to0.0pt{\kern-1.5pt\hbox{$\scriptstyle\rightarrow$}\vss}}}}=\\{\infty\,|\,G^{L\mathcal{R}}\\}\in{G^{\mathcal{L}}}$
such that Right, playing first, loses $G^{\mathrel{\mathop{L}\limits^{\vbox
to0.0pt{\kern-1.5pt\hbox{$\scriptstyle\rightarrow$}\vss}}}}+*n$. That is,
playing first, Left is protected against those nimbers in a disjunctive sum.
Similar to Definition 26, we could have defined $P_{G}$ with respect to Right
options, to obtain the same set.
Note that $P_{\scalebox{0.9}{$\leftmoon$}}=\mathbb{nim}$. This statement holds
for the literal form $\scalebox{1.1}{$\leftmoon$}=\pm\infty$. However, one can
show that by using instead the form
$\scalebox{1.1}{$\leftmoon$}=\left\\{\left\\{{\boldsymbol{\infty}}\\!\mid\\!0\right\\}\\!,0\\!\mid\\!\left\\{0\\!\mid\\!\overline{\infty}\right\\}\\!,0\right\\}$,
as in (1), then
$P_{\scalebox{0.9}{$\leftmoon$}}=\mathbb{nim}\setminus\\{0\\}$. The output of
“protected” is sensitive to which form we choose.
When the underlying game form is understood, we simply refer to the immediate
and protected nimbers, respectively.
###### Example 28.
Let $G\in\mathbb{Im}^{\boldsymbol{\infty}}$ be such that the Left options are
$0$, $*2$, and $\\{\infty\,|\,\\{*\,|\,\overline{\infty}\\},0\\}$. Of course,
$S_{G}=\\{0,*2\\}$. On the other hand, playing first, Left can use the check
to win $G+*$. Because of that, $P_{G}=\\{*\\}$. An important observation is
that, although Left is protected against the nimber $*$, Left cannot force a
Left move to $*$ in $G$. But if Right moves to 0, Left wins $G+*$ anyway
Sometimes, Right can manoeuvre Left’s eventual play to a nimber, or worse, via
a sequence of ‘check upon check’.
###### Definition 29 (Manoeuvrable Form).
A quiet form $G\in\mathbb{Im}^{\boldsymbol{\infty}}$ is _manoeuvrable_ if
after each Left move that is neither a nimber nor $\overline{\infty}$, Right
can force, with checks, a Left move to a nimber or a move by either player to
$\overline{\infty}$.
###### Example 30.
The form
$G=\\{*2,\\{\infty\,|\,\\{0,*4\,|\,\overline{\infty}\\},0\\}\,|\,*2,\\{0,\\{\infty\,|\,0,*4\\}|\,\overline{\infty}\\}\\}$
is manoeuvrable. If Left avoids the immediate nimber $*2$, by checking, then
Right can still force Left to move to one of the nimbers $0$ or $*4$.
###### Lemma 31.
If $G\in\mathbb{Im}^{\boldsymbol{\infty}}$ is manoeuvrable, then $P_{G}$ is
finite.
###### Proof.
After a Left first move in $G$, if needed, Right can force with checks a Left
move to a nimber or a move by either player to $\overline{\infty}$. Let $C$ be
the set of nimbers that can arise through this forcing strategy by Right. Then
$C$ is finite, because we study short games. Let $*n$ be a nimber such that,
for all $*m\in C$, we have $n>m$. In $G+*n$, after a first check, say, to
$G^{L}+*n$, Right forces with checks a move by either player to
$\overline{\infty}$ or a Left move to $*m+*n$ ($n>m$). In the second case,
after the sequence, Right wins with a TweedleDee-TweedleDum move. Thus, Left
can protect against at most a finite number of nimbers. That explains why
$P_{G}$ is finite in case of manoeuvrable games. ∎
Let $\mathrm{mex}(X)$ denote the smallest nonnegative integer not in $X$. Let
$\mathcal{G}$ denote the set of Sprague-Grundy values of a set of nimbers,
i.e. if $S=\\{*n_{i}\\}$, then $\mathcal{G}(S)=\\{n_{i}\\}$.
###### Lemma 32.
If $G\in\mathbb{Im}^{\boldsymbol{\infty}}$ is manoeuvrable then $G$ equals the
nimber $*n$, where $n=\mathrm{mex}(\mathcal{G}(S_{G}\cup P_{G}))$.
###### Proof.
By Lemma 31, we know that $S_{G}\cup P_{G}$ is finite. Let
$n=\mathrm{mex}(\mathcal{G}(S_{G}\cup P_{G}))$. Let us argue that the game
$G+*n\in\mathscr{P}$. If the first player moves in $G$ to a nimber $*m\in
S_{G}$, because $n$ is excluded from $\mathcal{G}(S_{G})$, he loses. If the
first player moves in $G$ to a quiet not nimber $G^{\prime}$, because
$G^{\prime}$ is not a nimber, $G^{\prime}+*n\in\mathscr{N}$ (Theorem 20), and
the first player also loses. If the first player moves in $G$, giving a check,
because $n$ is excluded from $\mathcal{G}(P_{G})$, he also loses. Finally, if
the first player moves to $G+*n^{\prime}$ ($n^{\prime}<n$), because $n$ is the
minimum excluded from $\mathcal{G}(S_{G}\cup P_{G})$, he loses because the
opponent has a direct TweedleDee-TweedleDum move or wins with a check. Hence,
by Theorem 20, $G=*n$. ∎
###### Lemma 33.
If $G,H\in\mathbb{Im}^{\boldsymbol{\infty}}$ are not nimbers, then
$G+H\in\mathscr{N}$.
###### Proof.
Consider $G,H\in\mathbb{Im}^{\boldsymbol{\infty}}\setminus\mathbb{nim}$. For a
contradiction, assume that the sum of the birthdays, $b=b(G)+b(H)$, is the
smallest possible such that $G+H\in\mathscr{P}$. Note that, by the assumptions
on $G$ and $H$, $b>0$. Without loss of generality, we will analyze the move
from $G+H$ to $G+H^{L}$. First, we prove two claims that concern local play in
$G$ and $H$ respectively.
Claim 1. Playing second in $G$, Left can avoid Left moves to nimbers and moves
by either player to $\overline{\infty}$ until the first Right-quiet move.
Proof of Claim 1. Suppose that Right, playing first in $G$, could force a Left
move to a nimber or a move by either player to $\overline{\infty}$. If so, in
$G+H$, by giving checks in $G$, Right could force some
$G^{\mathrel{\mathop{R}\limits^{\vbox
to0.0pt{\kern-1.5pt\hbox{$\scriptstyle\leftarrow$}\vss}}}L\cdots\mathrel{\mathop{R}\limits^{\vbox
to0.0pt{\kern-1.5pt\hbox{$\scriptstyle\leftarrow$}\vss}}}L}+H=*n+H$ (Right’s
turn) or a move by either player to $\overline{\infty}+H$. Of course, the
second situation would be a victory for Right. Regarding the first case, at
that moment, the position would be $*n+H$. And, because $H$ is not a nimber,
by Theorem 20, we would have $*n+H\in\mathscr{N}$, which is a winning move for
Right. In either case, Right, as first player, would win. That would
contradict $G+H\in\mathscr{P}$.
Claim 2. There is an $H^{L}$ such that Left can avoid Left moves to nimbers
and moves by either player to $\overline{\infty}$, until the first Right-quiet
move.
Proof of Claim 2. This is exactly the same as saying that $H$ is non-
manoeuvrable. If it was manoeuvrable, by Lemma 32, it would be a nimber, and
we would have a contradiction again.
Let us return to the move from $G+H$ to $G+H^{L}$. Because
$G+H\in\mathscr{P}$, Right has a winning move from $G+H^{L}$. But, by Claims 1
and 2, Left can play such that, at any stage before a Right-quiet move, Right
is moving on $g+h$, where $g$ is a follower of $G$ and $h$ is a follower of
$H^{L}$, such that neither $g$ nor $h$ is a nimber.
Either way, by assumption, there is a winning quiet Right-move $g^{R}+h$ or
$g+h^{R}$. Since these are impartial games, we must have
$g^{R}+h\in\mathscr{P}$ or $g+h^{R}\in\mathscr{P}$. But, because $h$ and $g$
are not nimbers, it follows by Theorem 20 that $g^{R}$ and $h^{R}$ are not
nimbers.
Therefore, we have $g^{R}+h\in\mathscr{P}$ or $g+h^{R}\in\mathscr{P}$ with
both components not nimbers. But this contradicts the smallest birthday
assumption. The result follows. ∎
###### Theorem 34 (Affine Impartial Values).
Every affine impartial form equals a nimber or the game
$\scalebox{1.1}{$\leftmoon$}\pmod{\mathbb{Im}^{\boldsymbol{\infty}}}$.
###### Proof.
Let $G\in\mathbb{Im}^{\boldsymbol{\infty}}$. If there is some $*n$ such that
$G+*n\in\mathscr{P}$, then $G=_{\mathbb{Im}^{\boldsymbol{\infty}}}*n$, by
Theorem 20.
Suppose next that $G+*n\in\mathscr{N}$, for all $n$, so that $G$ does not
equal a nimber modulo $\mathbb{Im}^{\boldsymbol{\infty}}$. By Lemma 33, for
all $X\in\mathbb{Im}^{\boldsymbol{\infty}}\setminus\mathbb{nim}$, we also have
$G+X\in\mathscr{N}$. Hence, for all $X\in\mathbb{Im}^{\boldsymbol{\infty}}$,
we have $G+X\in\mathscr{N}$, and therefore $G$ is a loony game. Because, by
Theorem 24, all loony games are equal modulo
$\mathbb{Im}^{\boldsymbol{\infty}}$, and $\leftmoon$ is a loony game, we have
$G=_{\mathbb{Im}^{\boldsymbol{\infty}}}\scalebox{1.1}{$\leftmoon$}$. ∎
###### Observation 35.
A form can be loony modulo $\mathbb{Im}^{\boldsymbol{\infty}}$ and not loony
modulo ${\mathbb{Np}^{\infty}}$. An example is the form
$G=\left\\{*,\left\\{{\boldsymbol{\infty}}\\!\mid\\!*\right\\}\\!\mid\\!*,\left\\{*\\!\mid\\!\overline{\infty}\right\\}\right\\}$.
This game is not loony modulo ${\mathbb{Np}^{\infty}}$ because, if
$X=\left\\{0\\!\mid\\!-1\right\\}\in\mathscr{N}$, playing first, Left loses
$G+X$. However,
$G=_{\mathbb{Im}^{\boldsymbol{\infty}}}\scalebox{1.1}{$\leftmoon$}$. This
follows, by Theorem 34, since $G$ does not equal any nimber; if Right starts
$G+*n$, he wins, by an appropriate parity consideration.
###### Theorem 36.
The game $\leftmoon$ is absorbing modulo $\mathbb{Im}^{\boldsymbol{\infty}}$,
that is,
$\scalebox{1.1}{$\leftmoon$}+Y=_{\mathbb{Im}^{\boldsymbol{\infty}}}\scalebox{1.1}{$\leftmoon$}$,
for all $Y\in\mathbb{Im}^{\boldsymbol{\infty}}$.
###### Proof.
Since $\scalebox{1.1}{$\leftmoon$}=\pm\infty$, regardless of what
$X\in\mathbb{Im}^{\boldsymbol{\infty}}$ is, the first player wins both
$\scalebox{1.1}{$\leftmoon$}+Y+X$ and $\scalebox{1.1}{$\leftmoon$}+X$.
Therefore, by definition of equality of games,
$\scalebox{1.1}{$\leftmoon$}+Y=_{\mathbb{Im}^{\boldsymbol{\infty}}}\scalebox{1.1}{$\leftmoon$}$.
∎
###### Corollary 37.
The game $\leftmoon$ is an idempotent modulo
$\mathbb{Im}^{\boldsymbol{\infty}}$, that is,
$\scalebox{1.1}{$\leftmoon$}+\scalebox{1.1}{$\leftmoon$}=_{\mathbb{Im}^{\boldsymbol{\infty}}}\scalebox{1.1}{$\leftmoon$}$.
###### Proof.
This is a trivial consequence of Theorem 36. ∎
###### Definition 38.
The Sprague-Grundy value of the moon is $\mathcal{G}($ $\leftmoon$ $)=\infty$.
The following theorem explains how the Sparague-Grundy value of
$G\in\mathbb{Im}^{\boldsymbol{\infty}}$ is determined by the set $S_{G}\cup
P_{G}$.
###### Theorem 39 (Affine Impartial Minimum Excluded Rule).
Let $G\in\mathbb{Im}^{\boldsymbol{\infty}}$. We have the following
possibilities:
* •
If $S_{G}\cup P_{G}=\mathbb{nim}$, then $G=\scalebox{1.1}{$\leftmoon$}$ and
$\mathrm{mex}(\mathcal{G}(S_{G}\cup P_{G}))=\infty$;
* •
If $S_{G}\cup P_{G}\neq\mathbb{nim}$, then
$G=*\left(\mathrm{mex}(\mathcal{G}(S_{G}\cup P_{G}))\right)$.
###### Proof.
If $S_{G}\cup P_{G}=\mathbb{nim}$, we have $G+*n\in\mathscr{N}$ for all $n$.
Because of that, $G$ is not a nimber and, by Theorem 34,
$G=\scalebox{1.1}{$\leftmoon$}$. If $S_{G}\cup P_{G}\neq\mathbb{nim}$, we use
the same argument of the proof of Lemma 32. ∎
###### Corollary 40.
If all the options of a game $G\in\mathbb{Im}^{\boldsymbol{\infty}}$ are quiet
then $G$ is a nimber.
###### Proof.
If all the options of a game $G\in\mathbb{Im}^{\boldsymbol{\infty}}$ are
quiet, then $P_{G}=\varnothing$. Therefore, $S_{G}\cup
P_{G}=S_{G}\neq\mathbb{nim}$ and, by Theorem 39,
$G=*\left(\mathrm{mex}(\mathcal{G}(S_{G}))\right)$. ∎
## 4 Case study: nimstring
In the introduction, we promised to show that the following component equals
$*2$.
Study the positions:
All (a), (b), (c), (d), and (e) are $\mathcal{P}$-positions. The game value of
(f) is
$\scalebox{1.1}{$\leftmoon$}=\left\\{\left\\{{\boldsymbol{\infty}}\\!\mid\\!0\right\\}\\!,0\\!\mid\\!\left\\{0\\!\mid\\!\overline{\infty}\right\\}\\!,0\right\\}$.
Other positions that equal $\leftmoon$ are the following.
In the position (l), the central horizontal move is the option (d), that is
equal to $0$. The other options are (f) and (g), that are equal to
$\leftmoon$ . Therefore, the literal form is
$l=\left\\{0,\scalebox{1.1}{$\leftmoon$},\scalebox{1.1}{$\leftmoon$}\\!\mid\\!0,\scalebox{1.1}{$\leftmoon$},\scalebox{1.1}{$\leftmoon$}\right\\}$
with $S_{l}=\\{0\\}$, and $P_{l}=\varnothing$. Applying the affine impartial
minimum excluded rule, we conclude that the position is $*$.
The position (m) is also equal to $*$, i.e. $0+*$.
Now, we are ready for (n), a more complex situation. The literal form is
$n=\\{h,i,k,\\{\infty\,|\,l\\}\,|\,h,i,k,\\{l\,|\,\overline{\infty}\\}\\},$
that is,
$n=\\{\leftmoon,\leftmoon,\leftmoon,\\{\infty\,|\,*\\}\,|\,\leftmoon,\leftmoon,\leftmoon,\\{*\,|\,\overline{\infty}\\}\\}.$
Hence, $S_{n}=\varnothing$, and $P_{n}=\mathbb{nim}\setminus\\{*\\}$. Applying
the affine impartial minimum excluded rule, we conclude that the position is
$*$.
Going back to the original question, we have the following.
The form is
$\\{\textbf{{\color[rgb]{0.00,1.00,0.00}n}},\textbf{{\color[rgb]{1.00,0.00,0.00}m}},\textbf{{\color[rgb]{0.00,0.07,1.00}c}}\,|\,\textbf{{\color[rgb]{0.00,1.00,0.00}n}},\textbf{{\color[rgb]{1.00,0.00,0.00}m}},\textbf{{\color[rgb]{0.00,0.07,1.00}c}}\\}$,
that is, $\\{*,*,0\,|\,*,*,0\\}=*2$. Here n represents a play of the top or
bottom bar, m represents a play of some middle bar, and c represents play of
the left line.
## 5 Case study: top entails
We denote by $\boldsymbol{n}$ the literal form of a stack of size $n$. The
literal form of the Left removal of the top coin from a stack of size $n$ is
$\\{\infty\,|\,(\boldsymbol{n-1})^{\mathcal{R}}\\}$ (and the symmetric from
Right’s point of view). With that in mind, let us compute the first few
values. First, we do it the tedious way, and then later after Theorem 41, we
propose the slick recursive way for a few more values, in a table format.
Of course, $\boldsymbol{0}=\\{\overline{\infty}\,|\,\infty\\}$. The first
player loses. Moreover,
* $\boldsymbol{1}=\\{\\{\infty\,|\,\boldsymbol{0}^{\mathcal{R}}\\}\,|\,\\{\boldsymbol{0}^{\mathcal{L}}\,|\,\overline{\infty}\\}\\}=\\{\\{\infty\,|\,\infty\\}\,|\,\\{\overline{\infty}\,|\,\overline{\infty}\\}\\}$. Therefore, $S_{\boldsymbol{1}}=\varnothing$ and $P_{\boldsymbol{1}}=\mathbb{nim}$. Using the affine impartial minimum excluded rule, $\boldsymbol{1}=\scalebox{1.1}{$\leftmoon$}$. In the next step, for ease, we will use the form $\scalebox{1.1}{$\leftmoon$}=\pm\infty$.
* $\boldsymbol{2}=\\{\boldsymbol{1}+\boldsymbol{1},\\{\infty\,|\,\boldsymbol{1}^{\mathcal{R}}\\}\,|\,\boldsymbol{1}+\boldsymbol{1},\\{\boldsymbol{1}^{\mathcal{L}}\,|\,\overline{\infty}\\}\\}=\\{\scalebox{1.1}{$\leftmoon$},\\{\infty\,|\,\overline{\infty}\\}\,|\,\scalebox{1.1}{$\leftmoon$},\\{\infty\,|\,\overline{\infty}\\}\\}$. Therefore, $S_{\boldsymbol{2}}=\varnothing$ and $P_{\boldsymbol{2}}=\varnothing$. Using the affine impartial minimum excluded rule, $\boldsymbol{2}=0$.
* $\boldsymbol{3}=\\{\boldsymbol{1}+\boldsymbol{2},\\{\infty\,|\,\boldsymbol{2}^{\mathcal{R}}\\}\,|\,\boldsymbol{1}+\boldsymbol{2},\\{\boldsymbol{2}^{\mathcal{L}}\,|\,\overline{\infty}\\}\\}$. This game is equal to $\\{\scalebox{1.1}{$\leftmoon$},\\{\infty\,|\,\scalebox{1.1}{$\leftmoon$},\\{\infty\,|\,\overline{\infty}\\}\\}\,|\,\scalebox{1.1}{$\leftmoon$},\\{\scalebox{1.1}{$\leftmoon$},\\{\infty\,|\,\overline{\infty}\\}\,|\,\overline{\infty}\\}\\}$. Therefore, $S_{\boldsymbol{3}}=\varnothing$ and $P_{\boldsymbol{3}}=\mathbb{nim}$. Using the affine impartial minimum excluded rule, $\boldsymbol{3}=\scalebox{1.1}{$\leftmoon$}$. In the next step, for ease, we will use the form $\scalebox{1.1}{$\leftmoon$}=\pm\infty$.
* $\boldsymbol{4}=\\{\boldsymbol{1}+\boldsymbol{3},\boldsymbol{2}+\boldsymbol{2},\\{\infty\,|\,\boldsymbol{3}^{\mathcal{R}}\\}\,|\,\boldsymbol{1}+\boldsymbol{3},\boldsymbol{2}+\boldsymbol{2},\\{\boldsymbol{3}^{\mathcal{L}}\,|\,\overline{\infty}\\}\\}$. This game is equal to$\\{\scalebox{1.1}{$\leftmoon$},0,\\{\infty\,|\,\overline{\infty}\\}\,|\,\scalebox{1.1}{$\leftmoon$},0,\\{\infty\,|\,\overline{\infty}\\}\\}$. Therefore, $S_{\boldsymbol{4}}=\\{0\\}$ and $P_{\boldsymbol{4}}=\varnothing$. Using the affine impartial minimum excluded rule, $\boldsymbol{4}=*$.
* $\boldsymbol{5}=\\{\boldsymbol{1}+\boldsymbol{4},\boldsymbol{2}+\boldsymbol{3},\\{\infty\,|\,\boldsymbol{4}^{\mathcal{R}}\\}\,|\,\boldsymbol{1}+\boldsymbol{4},\boldsymbol{2}+\boldsymbol{3},\\{\boldsymbol{4}^{\mathcal{L}}\,|\,\overline{\infty}\\}\\}$. This game is equal to $\\{\scalebox{1.1}{$\leftmoon$},\scalebox{1.1}{$\leftmoon$},\\{\infty\,|\,0\\}\,|\,\scalebox{1.1}{$\leftmoon$},\scalebox{1.1}{$\leftmoon$},\\{0\,|\,\overline{\infty}\\}\\}$. Therefore, $S_{\boldsymbol{5}}=\varnothing$ and $P_{\boldsymbol{5}}=\mathbb{nim}\setminus\\{0\\}$. Using the affine impartial minimum excluded rule, $\boldsymbol{5}=0$.
* $\boldsymbol{6}=\\{\boldsymbol{1}+\boldsymbol{5},\boldsymbol{2}+\boldsymbol{4},\boldsymbol{3}+\boldsymbol{3},\\{\infty\,|\,\boldsymbol{5}^{\mathcal{R}}\\}\,|\,\boldsymbol{1}+\boldsymbol{5},\boldsymbol{2}+\boldsymbol{4},\boldsymbol{3}+\boldsymbol{3},\\{\boldsymbol{5}^{\mathcal{L}}\,|\,\overline{\infty}\\}\\}$. This game is equal to $\\{\scalebox{1.1}{$\leftmoon$},*,\scalebox{1.1}{$\leftmoon$},\\{\infty\,|\,\\{0\,|\,\overline{\infty}\\}\\}\,|\,\scalebox{1.1}{$\leftmoon$},*,\scalebox{1.1}{$\leftmoon$},\\{\\{\infty\,|\,0\\}\,|\,\overline{\infty}\\}\\}$. Therefore, $S_{\boldsymbol{6}}=\\{*\\}$ and $P_{\boldsymbol{6}}=\\{0\\}$. Using the affine impartial minimum excluded rule, $\boldsymbol{6}=*2$.
* $\boldsymbol{7}=\\{\boldsymbol{1}+\boldsymbol{6},\boldsymbol{2}+\boldsymbol{5},\boldsymbol{3}+\boldsymbol{4},\\{\infty\,|\,\boldsymbol{6}^{\mathcal{R}}\\}\,|\,\boldsymbol{1}+\boldsymbol{6},\boldsymbol{2}+\boldsymbol{5},\boldsymbol{3}+\boldsymbol{4},\\{\boldsymbol{6}^{\mathcal{L}}\,|\,\overline{\infty}\\}\\}$. This game is equal to $\\{\scalebox{1.1}{$\leftmoon$},0,\scalebox{1.1}{$\leftmoon$},\\{\infty\,|\,*,\\{\\{\infty\,|\,0\\}\,|\,\overline{\infty}\\}\\}\,|\,\scalebox{1.1}{$\leftmoon$},0,\scalebox{1.1}{$\leftmoon$},\\{*,\\{\infty\,|\,\\{0\,|\,\overline{\infty}\\}\\}\,|\,\overline{\infty}\\}\\}$. So, $S_{\boldsymbol{7}}=\\{0\\}$, $P_{\boldsymbol{6}}=\mathbb{nim}\setminus\\{0,*\\}$, and with the affine impartial minimum excluded rule, $\boldsymbol{7}=*$.
Consider a stack of size $n$. We claim that an entailing move by Left does not
protect her against an element in $S_{\boldsymbol{n-1}}$. To see this, let
$*m\in S_{\boldsymbol{n-1}}$. Moving in $\boldsymbol{n}+*m$, if Left chooses
$\\{\infty\,|\,(\boldsymbol{n-1})\mathcal{{}^{R}}\\}+*m$, Right answers
$*m+*m$ and wins. On the other hand, we observe that an entailing move by Left
does not protect her against the elements of $P_{\boldsymbol{n-1}}$. To see
this, let $*m$ be an element of $P_{\boldsymbol{n-1}}$. Moving in
$\boldsymbol{n}+*m$, if Left chooses
$\\{\infty\,|\,(\boldsymbol{n-1})\mathcal{{}^{R}}\\}+*m$, because in
$\boldsymbol{n-1}$, Right is protected against $*m$, he has an entailing
winning move in the first component. Therefore, we have the general recursion
$P_{\boldsymbol{n}}=\mathbb{nim}\setminus(S_{\boldsymbol{n-1}}\cup
P_{\boldsymbol{n-1}}).$
The set $S_{\boldsymbol{n}}$ is composed of the values of the positions of the
form $\boldsymbol{\ell}+\boldsymbol{m}$, $\ell+m=n$, $\ell,m>0$, and
disregarding any sum where $\leftmoon$ appears. Hence, the recurrence of top
entails is as follows.
###### Theorem 41.
The sets $P_{0}=S_{0}=\varnothing$, and for all $n>0$
$P_{\boldsymbol{n}}=\mathbb{nim}\setminus(S_{\boldsymbol{n-1}}\cup
P_{\boldsymbol{n-1}}),$
$S_{\boldsymbol{n}}=\\{\mathcal{G}(\boldsymbol{\ell}+\boldsymbol{m}),\boldsymbol{\ell},\boldsymbol{m}\neq\scalebox{1.1}{$\leftmoon$}\\}$.
###### Proof.
This is explained in the above paragraph. ∎
Now, we can fill a table in an easy way.
$n$ | $S_{\boldsymbol{n}}$ | $P_{\boldsymbol{n}}$ | $S_{\boldsymbol{n}}\cup P_{\boldsymbol{n}}$ | $\mathcal{G}$-value (mex rule)
---|---|---|---|---
$0$ | $\varnothing$ | $\varnothing$ | $\varnothing$ | $0$
$1$ | $\varnothing$ | $\mathbb{nim}$ | $\mathbb{nim}$ | $\infty$
$2$ | $\varnothing$ | $\varnothing$ | $\varnothing$ | $0$
$3$ | $\varnothing$ | $\mathbb{nim}$ | $\mathbb{nim}$ | $\infty$
$4$ | $\\{0\\}$ | $\varnothing$ | $\\{0\\}$ | $1$
$5$ | $\varnothing$ | $\mathbb{nim}\backslash\\{0\\}$ | $\mathbb{nim}\backslash\\{0\\}$ | $0$
$6$ | $\\{*\\}$ | $\\{0\\}$ | $\\{0,*\\}$ | $2$
$7$ | $\\{0\\}$ | $\mathbb{nim}\backslash\\{0,*\\}$ | $\mathbb{nim}\backslash\\{*\\}$ | $1$
$8$ | $\\{0,*2\\}$ | $\\{*\\}$ | $\\{0,*,*2\\}$ | $3$
$9$ | $\\{*\\}$ | $\mathbb{nim}\backslash\\{0,*,*2\\}$ | $\mathbb{nim}\backslash\\{0,*2\\}$ | $0$
$10$ | $\\{0,*3\\}$ | $\\{0,*2\\}$ | $\\{0,*2,*3\\}$ | $1$
$11$ | $\\{0,*2\\}$ | $\mathbb{nim}\backslash\\{0,*2,*3\\}$ | $\mathbb{nim}\backslash\\{*3\\}$ | $3$
$12$ | $\\{0,*,*2\\}$ | $\\{*3\\}$ | $\\{0,*,*2,*3\\}$ | $4$
With the recursion, we know that $\boldsymbol{n}=\scalebox{1.1}{$\leftmoon$}$
if and only if $S_{\boldsymbol{n-1}}\cup P_{\boldsymbol{n-1}}\subseteq
S_{\boldsymbol{n}}$. That happens for $n=2403$, $n=2505$, and $n=33243$, as
mentioned in [9]. One of three possibilities must happen: a) A finite number
of finite nimbers; b) A finite number of loony values; c) An infinite number
of finite nimbers and an infinite number of loony values. However, it is an
open problem to know what case happens.
At the first Combinatorial Games Workshop at MSRI, John Conway proposed that
an effort should be made to devise some game with entailing moves that is non-
trivial, but (unlike top entails) susceptible to a complete analysis. All
attempts which have been tried turn out to be not very interesting. As a
sequel to this work, we are finalizing a paper [7] with a proposal of a
ruleset to meet Conway’s suggestion.
## References
* [1] M. Albert, R. J. Nowakowski, D. Wolfe. Lessons in Play: An Introduction to Combinatorial Game Theory, A. K. Peters, 2007.
* [2] E. R. Berlekamp. The Dots and Boxes Game: Sophisticated Child’s Play, A. K. Peter’s Ltd., 2000.
* [3] E. R. Berlekamp, J. H. Conway, R. K. Guy. Winning Ways, Academic Press, London, 1982.
* [4] E. R. Berlekamp, D. Wolfe. _Mathematical Go: Chilling Gets the Last Point_ , A. K. Peters, Ltd., 1994.
* [5] J. H. Conway. On Numbers and Games, Academic Press, 1976.
* [6] U. Larsson, R. J. Nowakowski, C. P. Santos. Combinatorial games with checks and terminating moves, preprint.
* [7] U. Larsson, R. J. Nowakowski, C. P. Santos. Electric cables, preprint.
* [8] A. N. Siegel. Combinatorial Game Theory, American Math. Soc., 2013.
* [9] J. West. New Values for Top Entails, Games of No Chance, 29, MSRI Publications, 345–350, 1996.
|
# Better sampling in explanation methods can prevent dieselgate-like deception
Domen Vreš, Marko Robnik-Šikonja
University of Ljubljana, Faculty of Computer and Information Science
Večna pot 113, 1000 Ljubljana, Slovenia
<EMAIL_ADDRESS><EMAIL_ADDRESS>
###### Abstract
Machine learning models are used in many sensitive areas where, besides
predictive accuracy, their comprehensibility is also important.
Interpretability of prediction models is necessary to determine their biases
and causes of errors and is a prerequisite for users’ confidence. For complex
state-of-the-art black-box models, post-hoc model-independent explanation
techniques are an established solution. Popular and effective techniques, such
as IME, LIME, and SHAP, use perturbation of instance features to explain
individual predictions. Recently, Slack et al. (2020) put their robustness
into question by showing that their outcomes can be manipulated due to poor
perturbation sampling employed. This weakness would allow dieselgate type
cheating of owners of sensitive models who could deceive inspection and hide
potentially unethical or illegal biases existing in their predictive models.
This could undermine public trust in machine learning models and give rise to
legal restrictions on their use.
We show that better sampling in these explanation methods prevents malicious
manipulations. The proposed sampling uses data generators that learn the
training set distribution and generate new perturbation instances much more
similar to the training set. We show that the improved sampling increases the
LIME and SHAP’s robustness, while the previously untested method IME is
already the most robust of all.
## 1 Introduction
Machine learning models are used in many areas where besides predictive
performance, their comprehensibility is also important, e.g., in healthcare,
legal domain, banking, insurance, consultancy, etc. Users in those areas often
do not trust a machine learning model if they do not understand why it made a
given decision. Some models, such as decision trees, linear regression, and
naïve Bayes, are intrinsically easier to understand due to the simple
representation used. However, complex models, mostly used in practice due to
better accuracy, are incomprehensible and behave like black boxes, e.g.,
neural networks, support vector machines, random forests, and boosting. For
these models, the area of explainable artificial intelligence (XAI) has
developed post-hoc explanation methods that are model-independent and
determine the importance of each feature for the predicted outcome. Frequently
used methods of this type are IME (Štrumbelj & Kononenko, 2013), LIME (Ribeiro
et al., 2016), and SHAP (Lundberg & Lee, 2017).
To determine the features’ importance, these methods use perturbation
sampling. Slack et al. (2020) recently noticed that the data distribution
obtained in this way is significantly different from the original distribution
of the training data as we illustrate in Figure 1a. They showed that this can
be a serious weakness of these methods. The possibility to manipulate the
post-hoc explanation methods is a critical problem for the ML community, as
the reliability and robustness of explanation methods are essential for their
use and public acceptance. These methods are used to interpret otherwise
black-box models, help in debugging models, and reveal models’ biases, thereby
establishing trust in their behavior. Non-robust explanation methods that can
be manipulated can lead to catastrophic consequences, as explanations do not
detect racist, sexist, or otherwise biased models if the model owner wants to
hide these biases. This would enable dieselgate-like cheating where owners of
sensitive prediction models could hide the socially, morally, or legally
unacceptable biases present in their models. As the schema of the attack on
explanation methods on Figure 1b shows, owners of prediction models could
detect when their models are examined and return unbiased predictions in this
case and biased predictions in normal use. This could have serious
consequences in areas where predictive models’ reliability and fairness are
essential, e.g., in healthcare or banking. Such weaknesses can undermine
users’ trust in machine learning models in general and slow down technological
progress.
a)
b)
Figure 1: a) PCA based visualization of a part of the COMPAS dataset. The blue
points show the original instances, and the red points represent instances
generated with the perturbation sampling used in the LIME method. The
distributions are notably different. b) The idea of the attack on explanation
methods based on the difference of distributions. The attacker’s adversarial
model contains both the biased and unbiased model. The decision function that
is part of the cheating model decides if the instance is outside the original
distribution (i.e. used only for explanation) or an actual instance. If the
case of an actual instance, the result of the adversarial model is equal to
the result of the biased model, otherwise it is equal to the result of the
unbiased model.
In this work, we propose to change the main perturbation-based explanation
methods and make them more resistant to manipulation attempts. In our
solution, the problematic perturbation-based sampling is replaced with more
advanced sampling, which uses modern data generators that better capture the
distribution of the training dataset. We test three generators, the RBF
network based generator (Robnik-Šikonja, 2016), random forest-based generator,
available in R library semiArtificial (Robnik-Šikonja, 2019), as well as the
generator using variational autoencoders (Miok et al., 2019). We show that the
modified gLIME and gSHAP methods are much more robust than their original
versions. For the IME method, which previously was not analyzed, we show that
it is already quite robust. We release the modified explanation methods under
the open-source license111https://github.com/domenVres/Robust-LIME-SHAP-and-
IME.
In this work, we use the term robustness of the explanation method as a notion
of resilience against adversarial attacks, i.e. as the ability of an
explanation method to recognize the biased classifier in an adversary
environment. This type of robustness could be more formally defined as the
number of instances where the adversarial model’s bias was correctly
recognized. We focus on the robustness concerning the attacks described in
Slack et al. (2020). There are other notions of robustness in explanation
methods; e.g., (Alvarez-Melis & Jaakkola, 2018) define the robustness of the
explanations in the sense that similar inputs should give rise to similar
explanations.
The remainder of the paper is organized as follows. In Section 2, we present
the necessary background and related work on explanation methods, attacks on
them, and data generators. In Section 3, we propose a defense against the
described weaknesses of explanation methods, and in Section 4, we empirically
evaluate the proposed solution. In Section 5, we draw conclusions and present
ideas for further work.
## 2 Background and related work
In this section, we first briefly describe the background on post-hoc
explanation methods and attacks on them, followed by data generators and
related works on the robustness of explanation methods.
### 2.1 Post-hoc explanation methods
The current state-of-the-art perturbation-based explanation methods, IME,
LIME, and SHAP, explain predictions for individual instances. To form an
explanation of a given instance, they measure the difference in prediction
between the original instance and its neighboring instances, obtained with
perturbation sampling. Using the generated instances, the LIME method builds a
local interpretable model, e.g., a linear model. The SHAP and IME methods
determine the impact of the features as Shapley values from the coalitional
game theory (Shapley, 1988). In this way, they assure that the produced
explanations obey the four Shapley fairness axioms (Štrumbelj & Kononenko,
2013). Due to the exponential time complexity of Shapley value calculation,
both methods try to approximate them. The three methods are explained in
detail in the above references, and a formal overview is presented in Appendix
A, while below, we present a brief description. In our exposition of the
explanation methods, we denote with $f$ the predictive model and with $x$ the
instance we are explaining.
Explanations of instances with the LIME method is obtained with an
interpretable model $g$. The model $g$ has to be both locally accurate (so
that it can obtain correct feature contributions) and simple (so that it is
interpretable). Therefore in LIME, $g$ is a linear regression model trained on
the instances sampled in the neighborhood of the explained instance $x$, which
is first normalized.The perturbed explanation instances are generated by
adding Gaussian noise to each feature of $x$ independently. Using linear
models to determine the contributions of features, LIME is fast but may
produce poor explanations for instances close to complex decision boundaries.
SHAP estimates Shapley values using the approach of LIME (i.e. local linear
models computed with the weighted least squares algorithm) but using a
different neighborhood and loss function for local accuracy, which ensures
that Shapley values are obtained. The sampling of explanation instances in
SHAP relies on the distribution set $D$, obtained from the training set, by
computing the centroids of k-means clusters.
IME computes Shapley values using a pure sampling approach (no proxy with an
interpretable model). Its sampling replaces part of the explained instance
with existing values of selected features and then averages differences in the
prediction of explained instance and samples. This makes the method slower but
also potentially more robust as the method does not assume the shape of the
dependency.
### 2.2 Attacks on explanations
The idea of the attack on explanations presented in Slack et al. (2020) is
that the attacker creates an adversarial (biased) model whose behavior it
wants to hide from explanation methods, e.g., a racist model that does not
grant credits to blacks. As Figure 1b illustrates, a part of the attacker’s
adversarial model is an unbiased model, which, e.g., does not take disputed
attributes such as race into account. The adversarial model tries to
manipulate the explanation method by behaving like an unbiased model on
perturbed instances used for the explanation – in this case, it uses the
unbiased model’s output. On instances from the original distributions that are
a part of normal model use, the adversarial model uses the biased model’s
output. To distinguish between the original and perturbation-sampled data, the
adversarial model contains a decision model that selects the right type of
model (biased or unbiased) based on the predicted type of instances.
The reason that the decision function can distinguish between the original
data and sampled instances are the shortcomings of explanation methods that
use the perturbation sampling (see Figure 1a). For that graph, we created new
instances based on the COMPAS dataset (see Section 4) with the Gaussian noise
as in the LIME explanation method. We mapped the original and sampled
instances to 2D space using the PCA transformation. We can see that the
sampled data is distributed differently than the original data. As the
difference is revealed with PCA, other machine learning algorithms can also
effectively distinguish between sampled and original data.
The attack can be formally defined in the following way. The attacker has a
biased model $b$ and a sample $S$ from the distribution $X_{dist}$. The
attacker has no additional information about the distribution $X_{dist}$. The
attacker’s goal is to produce an adversarial model $e$ that behaves like $b$
on the data from $X_{dist}$ but does not reveal $b$’s bias to the explanation
methods. We assume that the attacker also has an unbiased model $\psi$ that
hides the $f$’s bias. The attacker creates a decision model $d$, that should
output $1$ if the input instance $x$ is from $X_{dist}$ and $0$ otherwise. The
model $d$ is trained on $S$ and generated perturbation samples. The attacker
creates the adversarial model $e$ with its output defined by the following
equation:
$e(x)=\begin{cases}b(x),&d(x)=1\\\ \psi(x),&d(x)=0\end{cases}\vspace{-5mm}$
(1)
### 2.3 Data generators
We used three different data generators based on different algorithms,
modeling the distribution of the training set: variational autoencoder with
Monte Carlo dropout (Miok et al., 2019), RBF network (Robnik-Šikonja, 2016),
and random forest ensemble (Robnik-Šikonja, 2019). In the remainder of the
paper, we refer to the listed generators consecutively as MCD-VAE, rbfDataGen,
and TreeEnsemble.
Autoencoder (AE) consists of two neural networks called encoder and decoder.
It aims to compress the input instances by passing them through the encoder
and then reconstructing them to the original values with the decoder. Once the
AE is trained, it can be used to generate new instances. Variational
autoencoder (Doersch, 2016) is a special type of autoencoder, where the
vectors $z$ in the latent dimension (output of the encoder and input of the
decoder) are normally distributed. Encoder is therefore approximating the
posterior distribution $p(z|x)$, where we assume
$p(z|x)\sim\mathcal{N}(\mu_{x},\Sigma_{x}).$ The generator proposed by Miok et
al. (2019) uses the Monte Carlo dropout (Gal & Ghahramani, 2016) on the
trained decoder. The idea of this generator is to propagate the instance $x$
through the encoder to obtain its latent encoding $z$. This can be propagated
many times through the decoder, obtaining every time a different result due to
the Monte Carlo dropout but preserving similarity to the original instance
$x$.
The RBF network (Moody & Darken, 1989) uses Gaussian kernels as hidden layer
units in a neural network. Once the network’s parameters are learned, the
rbfDataGen generator (Robnik-Šikonja, 2016) can sample from the normal
distributions, defined with obtained Gaussian kernels, to generate new
instances.
The TreeEnsemble generator (Robnik-Šikonja, 2019) builds a set of random trees
(forest) that describe the data. When generating new instances, the generator
traverses from the root to the leaves of a randomly chosen tree, setting
values of features in the decision nodes on the way. When reaching a leaf, it
assumes that it has captured the dependencies between features. Therefore, the
remaining features can be generated independently according to the observed
empirical distribution in this leaf. For each generated instance, all
attributes can be generated in one leaf, or another tree can be randomly
selected where unassigned feature values are filled in. By selecting different
trees, different features are set in the interior nodes and leaves.
### 2.4 Related work on robustness of explanations
The adversarial attacks on perturbation based explanation methods were
proposed by Slack et al. (2020), who show that LIME and SHAP are vulnerable
due to the perturbation based sampling used. We propose the solution to the
exposed weakness in SHAP and IME based on better sampling using data
generators adapted to the training set distribution.
In general, the robustness of explanation methods has been so far poorly
researched. There are claims that post-hoc explanation methods shall not be
blindly trusted, as they can mislead users (deliberately or not) and disguise
gender and racial discrimination (Lipton, 2016). Selbst & Barocas (2018) and
Kroll et al. (2017) showed that even if a model is completely transparent, it
is hard to detect and prevent bias due to the existence of correlated
variables.
Specifically, for deep neural networks and images, there exist adversarial
attacks on saliency map based interpretation of predictions, which can hide
the model’s bias (Dombrowski et al., 2019; Heo et al., 2019; Ghorbani et al.,
2019). Dimanov et al. (2020) showed that a bias of a neural network could be
hidden from post-hoc explanation methods by training a modified classifier
that has similar performance to the original one, but the importance of the
chosen feature is significantly lower.
The kNN-based explanation method, proposed by Chakraborty et al. (2020), tries
to overcomes the inadequate perturbation based sampling used in explanation
methods by finding similar instances to the explained one in the training set
instead of generating new samples. This solution is inadequate for realistic
problems as the problem space is not dense enough to get reliable
explanations. Our defense of current post-hoc methods is based on superior
sampling, which has not yet been tried. Saito et al. (2020) use the neural CT-
GAN model to generate more realistic samples for LIME and prevent the attacks
described in Slack et al. (2020). We are not aware of any other defenses
against the adversarial attacks on post-hoc explanations.
## 3 Robustness through better sampling
We propose the defense against the adversarial attacks on explanation methods
that replaces the problematic perturbation sampling with a better one, thereby
making the explanation methods more robust. We want to generate the
explanation data in such a way that the attacker cannot determine whether an
instance is sampled or obtained from the original data. With an improved
sampling, the adversarial model shown in Figure 1b shall not determine whether
the instance $x$ was generated by the explanation method, or it is the
original instance the model has to label. With a better data generator, the
adversarial model cannot adjust its output properly and the deception,
described in Section 2.2, becomes ineffective.
The reason for the described weakness of LIME and SHAP is inadequate sampling
used in these methods. Recall that LIME samples new instances by adding
Gaussian noise to the normalized feature values. SHAP samples new instances
from clusters obtained in advance with the k-means algorithm from the training
data.
Instead of using the Gaussian noise with LIME, we generate explanation samples
for each instance with one of the three better data generators, MCD-VAE,
rbfDataGen, or TreeEnsemble (see Section 2.3). We call the improved
explanation methods gLIME and gSHAP (g stands for generator-based). Using
better generators in the explanation methods, the decision function in the
adversarial model will less likely determine which predicted instances are
original and which are generated for the explanation purposes.
Concerning gLIME, we generate data in the vicinity of the given instance using
MCD-VAE, as the LIME method builds a local model. Using the TreeEnsemble and
rbfDataGen generators, we do not generate data in the neighborhood of the
given instance but leave it to the proximity measure of the LIME method to
give higher weights to instances closer to the explained one.
In SHAP, the perturbation sampling replaces the values of hidden features in
the explained instance with the values from the distribution set $D$. The
problem with this approach is that it ignores the dependencies between
features. For example, in a simple dataset with two features, house size, and
house price, let us assume that we hide the house price, but not the house
size. These two features are not independent because the price of a house
increases with its size. Suppose we are explaining an instance that represents
a large house. Disregarding the dependency, using the sampled set $D$, SHAP
creates several instances with a low price, as such instances appeared in the
training set. In this way, the sampled set contains many instances with a
small price assigned to a large house, from which the attacker can determine
that these instances were created in perturbation sampling and serve only for
the explanation.
In the proposed gSHAP, using the MCD-VAE and TreeEnsemble generators, the
distribution set $D$ is generated in the vicinity of the explained instance.
In the sampled instances, some feature values of the explained instance are
replaced with the generated values, but the well-informed generators consider
dependencies between features detected in the original distribution. This will
make the distribution of the generated instances very similar to the original
distribution. In our example, the proposed approach generates new instances
around the instance representing a large house, and most of these houses will
be large. As the trained generators capture the original dataset’s
dependencies, these instances will also have higher prices. This will make it
difficult for the attacker to recognize the generated instances used in
explanations. The advantage of generating the distribution set close to the
observed instance is demonstrated in Appendix B.
The rbfDataGen generator does not support the generation of instances around a
given instance. Therefore, we generate the sampled set based on the whole
training set and not for each instance separately (we have this option also
for TreeEnsemble). This is worse than generating the distribution set around
the explained instance but still better than generating it using the k-means
sampling in SHAP. There are at least three advantages. First, the generated
distribution set $D$ can be larger. The size of the k-means distribution set
cannot be arbitrarily large because it is limited by the number of clusters in
the training set. Second, the centroids of the clusters obtained by the
k-means algorithm do not necessarily provide a good summary of the training
set and may not be valid instances from training set distribution. They also
do not capture well the density of instances (e.g., most of the data from the
training set could be grouped in one cluster). Third, using the proposed
generators, SHAP becomes easier to use compared to the k-means generator,
where users have to determine the number of clusters, while the data
generators we tested can be used in the default mode without parameters,
## 4 Evaluation
To evaluate the proposed improvements in the explanation methods, we first
present the used datasets in Section 4.1, followed by the experiments. In
Section 4.2, we test the robustness of gLIME, gSHAP, and gIME against the
adversarial models. To be more realistic, we equipped the adversarial models
with the same improved data generators we used in the explanation methods. It
is reasonable to assume that attackers could also use better data generators
when preparing their decision function, making their attacks much stronger. As
the evaluation metric for the success of deception, we use the proportion of
instances where the adversarial model deceived the explanation methods so that
they did not detect sensitive attributes as being important in the prediction
models. In Section 4.3, we test if enhanced generators produce different
explanations than the original ones. As the attacker might decide to employ
deception only when it is really certain that the predicted instance is used
inside the explanation method, we test different thresholds of the decision
function $d$ from Equation 1 (currently set to $0.5$). We report on this
analysis in Section 4.4.
### 4.1 Sensitive datasets prone to deception
Following (Slack et al., 2020), we conducted our experiments on three data
sets from domains where a biased classifier could pose a critical problem,
such as granting a credit, predicting crime recidivism, and predicting the
violent crime rate. The basic information on the data sets is presented in
Table 1. The statistics were collected after removing the missing values from
the data sets and before we encoded categorical features as one-hot-encoded
vectors.
Data set | # inst. | # features | # categorical | sensitive | unrelated 1 | unrelated 2
---|---|---|---|---|---|---
COMPAS | $6172$ | $7$ | $4$ | race | random1 | random2
German | $1000$ | $25$ | $15$ | gender | pctOfIncome | /
CC | $1994$ | $100$ | $0$ | racePctWhite | random1 | random2
Table 1: Basic information and sensitive features in the the used data sets.
The target variable is not included in the number of features and is binary
for all data sets. The pctOfIncome full name is loanRateAsPercentOfIncome.
COMPAS (Correctional Offender Management Profiling for Alternative Sanctions)
is a risk assessment used by the courts in some US states to determine the
crime recurrence risk of a defendant. The dataset (Angwin et al., 2016))
includes criminal history, time in prison, demographic data (age, gender,
race), and COMPAS risk assessment of the individual. The dataset contains data
of 6,172 defendants from Broward Couty, Florida. The sensitive attribute in
this dataset (the one on which the adversarial model will be biased) is race.
African Americans, whom biased model associates with a high risk of
recidivism, represent $51.4\%$ of instances from the data set. This set’s
target variable is the COMPAS score, which is divided into two classes: a high
and low risk. The majority class is the high risk, which represents $81.4\%$
of the instances.
The German Credit dataset (German for the rest of the paper) from the UCI
repository (Dua & Graff, 2019) includes financial (bank account information,
loan history, loan application information, etc.) and demographic data
(gender, age, marital status, etc.) for 1,000 loan applicants. A sensitive
attribute in this data set is gender. Men, whom the biased model associates
with a low-risk, represent $69\%$ of instances. The target variable is the
loan risk assessment, divided into two classes: a good and a bad customer. The
majority class is a good customer, which represents $70\%$ of instances.
Communities and Crime (CC) data set (Redmond & Baveja, 2002) contains data
about the rate of violent crime in US communities in $1994$. Each instance
represents one community. The features are numerical and represent the
percentage of the community’s population with a certain property or the
average of population in the community. Features include socio-economic (e.g.,
education, house size, etc.) and demographic (race, age) data. The sensitive
attribute is the percentage of the white race. The biased model links
instances where whites’ percentage is above average to a low rate of violent
crime. The target variable is the rate of violent crime divided into two
classes: high and low. Both classes are equally represented in the data set.
### 4.2 Robustness of explanation methods
To evaluate the robustness of explanation methods with added generators (i.e.
gLIME, gSHAP, and gIME), we split the data into training and evaluation set in
the ratio $90\%:10\%$. We used the same training set for the training of
adversarial models and explanation methods. We encoded categorical features as
one-hot-encoded vectors.
We simulated the adversarial attack for every combination of generators used
in explanation methods and adversarial models (except for method IME, where we
did not use rbfDataGen, which cannot generate instances in the neighborhood of
a given instance). When testing SHAP, we used two variants of the TreeEnsemble
generator for explanation: generating new instances around the explained
instance and generating new instances according to the whole training set
distribution. In the LIME testing, we used only the whole training set variant
of TreeEnsemble inside the explanation. In the IME testing, we only used the
variant of Tree Ensemble that fills in the hidden values (called TEnsFillIn
variant). For training the adversarial models, we used the whole training set
based variant of TreeEnsemble, except for IME, where we used the TEnsFillIn
variant. These choices reflect the capabilities of different explanation
method and attempt to make both defense and attack realistic (as strong as
possible). More details on training the decision model $d$ inside the
adversarial model can be found in Appendix D.
In all cases, the biased model $b$ (see Section 2.2) was a simple function
depending only on the value of the sensitive feature. The unbiased model
$\psi$ depended only on the values of unrelated features. The sensitive and
unrelated features are shown on the right-hand side of Table 1. Features
random1 and random2 were uniformly randomly generated from the $\\{0,1\\}$
set. On COMPAS and CC, we simulated two attacks, one with a single unrelated
feature (the result of $\psi$ depends only on the value of the unrelated
feature 1), and another one with two unrelated features.
For every instance in the evaluation set, we recorded the most important
feature according to the used explanation method (i.e. the feature with the
largest absolute value of the contribution as determined by the explanation
method). The results are shown as a heatmap in Figure 2. The green color means
that the explanation method was deceived in less than $30\%$ of cases (i.e.
the sensitive feature was recognized as the most important one in more than 70
% of the cases as the scale on the heatmap suggests), and the red means that
it was deceived in more than $70\%$ of cases (the sensitive feature was
recognized as the most important one in less than 30 % of the cases as the
scale on the heatmap suggests). We consider deception successful if the
sensitive feature was not recognized as the most important by the explanation
method (the sensitive features are the only relevant features in biased models
$b$).
The gLIME method is more robust with the addition of rbfDataGen and
TreeEnsemble than LIME and less robust with the addition of MCD-VAE. This
suggests that parameters for MCD-VAE were not well-chosen, and this pattern
repeats for other explanation methods. Both TreeEnsemble and rbfDataGen make
gLIME considerably more robust on COMPAS and German datasets, but not on the
CC dataset. We believe the reason for that is that features in CC are strongly
interdependent, and many represent the same attribute as a fraction of value,
e.g., we have the share of the white population, the share of the Asian
population, and so on. This interdependence dictates that all fractions have
to add up to $1$. The data generators are unlikely to capture such strong
conditional dependencies, but the adversarial model’s decision function is
likely to approximate it.
The gSHAP method is most robust when using the TreeEnsemble generator, but it
shows less robust behavior than the gLIME method. The reason for that could be
in the feature value replacement strategy used by the SHAP and gSHAP methods,
which change only some of the feature’s values with the help of the
distribution set. This can lead to out-of-distribution instances if the
instances in the distribution set are not close enough to the explained
instance. With gLIME, we generate complete instances that are more likely to
be in the distribution of the training set.
IME is quite robust even with the perturbation sampling, as we expected. This
suggests that IME is the most robust of all three tested explanation methods
and shall be considered the chosen method in sensitive situations. The gIME
results when using the TreeEnsemble generator (TEnsFillIn variant) are
comparable to the original IME variant results. This suggests that sampling
from a smaller data set, which represents the neighborhood of the explained
instance, does not decrease the method’s robustness.
Figure 2: The robustness results for gLIME (top), gSHAP (middle), and gIME
(bottom). The graphs show the proportion of evaluation set instances, where
the sensitive feature was recognized as the most important by the used
explanation method. Rows represent the generators used for explanations. The
column labels consist of the name of the data set on which the experiment was
performed and the name of the generator used for training of the adversarial
model. Compas2 and CC2 denote an attack with two independent features.
Perturbation represents the original sampling used in LIME, SHAP, and IME,
TEnsFillIn represents the TreeEnsemble variant where new instances are
generated around the given one, and TreeEns represents the generation from the
whole distribution.
### 4.3 Comparing explanations of original and modified methods
We checked if improved data generators affect explanations in a non-adversary
environment. We compared the explanations of original and modified methods for
four different classification models on three different datasets. As a measure
of differences, we use the mean of the absolute difference between the
original and enhanced explanations over the whole evaluation set. Details of
these experiments are reported in Appendix F, below we shortly summarize the
results, contained in Table 4.
The differences between original LIME and gLIME explanations are considerable
(see the top part of Table 4). This is not surprising since LIME fits local
linear models in its local explanations, which can strongly differ even for
small perturbations of the model’s input. The differences in SHAP and IME
explanations are small (the average MAD is almost negligible). We can conclude
that explanations of gSHAP and gIME are not significantly different from SHAP
and IME in the non-adversary environment.
### 4.4 Behavior of adversarial classifier with different prediction
thresholds
The attacker might be wary of being discovered by the regulator and decide to
employ deception only when it is really certain that the predicted instance is
used inside the explanation method. We express different levels of attacker’s
conservatism by changing the decision function threshold $d$ from Equation 1,
where it is currently set to $0.5$. We tested the robustness of modified
explanation methods on the COMPAS dataset with adversarial models that use
different certainty thresholds for deploying the biased classifier (simulating
different levels of aggressiveness). For results and a more detailed
description of the experiment, see Appendix G. Below, we shortly summarize the
results.
Even with different thresholds, gIME is still the most robust of the three
explanation methods, and treeEnsemble still gives the best results as the data
generator. The sensitive feature is recognized as the most important more
often when the threshold is lower, which is expected as in this case the
adversarial model behaves more aggressively. While the percentage of the
instances on which the racist behavior of the adversarial model is recognized
drops with higher values of the thresholds, it remains high enough to warn the
regulator about the problematic behavior of the prediction model (especially
in the case of gSHAP and gIME using treeEnsemble as data generator). We can
conclude that the enhanced explanation methods remain robust enough, even with
more conservative adversarial models.
## 5 Conclusions
We presented the defense against adversarial attacks on explanation methods.
The attacks exploit the shortcomings of perturbation sampling in post-hoc
explanation methods. This sampling used in these methods produces instances
too different from the original distribution of the training set. This allows
unethical owners of biased prediction models to detect which instances are
meant for explanation and label them in an unbiased way. We replaced the
perturbation sampling with data generators that better capture the
distribution of a given data set. This prevents the detection of instances
used in explanation and disarms attackers. We have shown that the modified
gLIME and gSHAP explanation methods, which use better data generators, are
more robust than the original variants, while IME is already quite robust. The
difference in explanation values between original and enhanced gSHAP and gIME
is negligible, while for gLIME, it is considerable. Our preliminary results in
Appendix C show that using the TreeEnsemble generator, the gIME method
converges faster and requires from 30-50% fewer samples.
The effectiveness of the proposed defense depends on the choice of the data
generator and its parameters. While the TreeEnsemble generator turned out the
most effective in our evaluation, in practice, several variants might need to
be tested to get a robust explanation method. Inspecting authorities shall be
aware of the need for good data generators and make access to training data of
sensitive prediction models a legal requirement. Luckily, even a few non-
deceived instances would be enough to raise the alarm about unethical models.
This work opens a range of possibilities for further research. The proposed
defense and attacks shall be tested on other data sets with a different number
and types of features, with missing data, and on different types of problems
such as text, images, and graphs. The work on useful generators shall be
extended to find time-efficient generators with easy to set parameters and the
ability to generate new instances in the vicinity of a given one. Generative
adversarial networks (Goodfellow et al., 2014) may be a promising line of
research. The TreeEnsemble generator, currently written in pure R, shall be
rewritten in a more efficient programming language. The MCD-VAE generator
shall be studied to allow automatic selection of reasonable parameters for a
given dataset. In SHAP sampling, we could first fix the values of the features
we want to keep, and the values of the others would be generated using the
TreeEnsemble generator.
#### Acknowledgments
The work was partially supported by the Slovenian Research Agency (ARRS) core
research programme P6-0411. This paper is supported by European Union’s
Horizon 2020 research and innovation programme under grant agreement No
825153, project EMBEDDIA (Cross-Lingual Embeddings for Less-Represented
Languages in European News Media).
## References
* Alvarez-Melis & Jaakkola (2018) David Alvarez-Melis and Tommi S Jaakkola. On the robustness of interpretability methods. _arXiv preprint arXiv:1806.08049_ , 2018.
* Angwin et al. (2016) Julia Angwin, Jeff Larson, Surya Mattu, and Lauren Kirchner. Machine bias. _ProPublica, May 23_ , 2016.
* Chakraborty et al. (2020) Joymallya Chakraborty, Kewen Peng, and Tim Menzies. Making fair ML software using trustworthy explanation. _arXiv preprint arXiv:2007.02893_ , 2020.
* Dimanov et al. (2020) Botty Dimanov, Umang Bhatt, Mateja Jamnik, and Adrian Weller. You shouldn’t trust me: Learning models which conceal unfairness from multiple explanation methods. In _SafeAI@ AAAI_ , pp. 63–73, 2020.
* Doersch (2016) Carl Doersch. Tutorial on variational autoencoders. _ArXiv_ , abs/1606.05908, 2016.
* Dombrowski et al. (2019) Ann-Kathrin Dombrowski, Maximillian Alber, Christopher Anders, Marcel Ackermann, Klaus-Robert Müller, and Pan Kessel. Explanations can be manipulated and geometry is to blame. In _Advances in Neural Information Processing Systems_ , pp. 13589–13600, 2019.
* Dua & Graff (2019) Dheeru Dua and Casey Graff. UCI machine learning repository. http://archive.ics.uci.edu/ml, 2019. [Accessed: 9. 8. 2020].
* Gal & Ghahramani (2016) Yarin Gal and Zoubin Ghahramani. Dropout as a Bayesian Approximation: Representing Model Uncertainty in Deep Learning. In _Proceedings of The 33rd International Conference on Machine Learning_ , volume 48, pp. 1050–1059, 2016.
* Ghorbani et al. (2019) Amirata Ghorbani, Abubakar Abid, and James Zou. Interpretation of neural networks is fragile. _Proceedings of the AAAI Conference on Artificial Intelligence_ , 33:3681–3688, 2019.
* Goodfellow et al. (2014) Ian Goodfellow, Jean Pouget-Abadie, Mehdi Mirza, Bing Xu, David Warde-Farley, Sherjil Ozair, Aaron Courville, and Yoshua Bengio. Generative adversarial nets. In _Advances in neural information processing systems_ , pp. 2672–2680, 2014.
* Heo et al. (2019) Juyeon Heo, Sunghwan Joo, and Taesup Moon. Fooling neural network interpretations via adversarial model manipulation. In _Advances in Neural Information Processing Systems_ , pp. 2925–2936, 2019.
* Kroll et al. (2017) Joshua A Kroll, Joanna Huey, Solon Barocas, Edward W Felten, Joel R Reidenberg, David G Robinson, and Harlan Yu. Accountable algorithms. _University of Pennsylvania Law Review_ , 165(3):633–705, 2017.
* Lipton (2016) Zachary Chase Lipton. The mythos of model interpretability. _CoRR_ , abs/1606.03490, 2016.
* Lundberg & Lee (2017) Scott M Lundberg and Su-In Lee. A unified approach to interpreting model predictions. In _Advances in Neural Information Processing Systems 30_ , pp. 4765–4774. Curran Associates, Inc., 2017.
* Miok et al. (2019) Kristian Miok, Deng Nguyen-Doan, Daniela Zaharie, and Marko Robnik-Šikonja. Generating data using Monte Carlo dropout. In _2019 IEEE 15th International Conference on Intelligent Computer Communication and Processing (ICCP)_ , pp. 509–515, 2019.
* Moody & Darken (1989) J. Moody and C. J. Darken. Fast learning in networks of locally-tuned processing units. _Neural Computation_ , 1:281–294, 1989.
* Pedregosa et al. (2011) F. Pedregosa, G. Varoquaux, A. Gramfort, V. Michel, B. Thirion, O. Grisel, M. Blondel, P. Prettenhofer, R. Weiss, V. Dubourg, J. Vanderplas, A. Passos, D. Cournapeau, M. Brucher, M. Perrot, and E. Duchesnay. Scikit-learn: Machine learning in Python. _Journal of Machine Learning Research_ , 12:2825–2830, 2011\.
* Redmond & Baveja (2002) Michael Redmond and Alok Baveja. A data-driven software tool for enabling cooperative information sharing among police departments. _European Journal of Operational Research_ , 141:660–678, 2002.
* Ribeiro et al. (2016) Marco Tulio Ribeiro, Sameer Singh, and Carlos Guestrin. ”Why should I trust you?”: Explaining the predictions of any classifier. In _Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining_ , pp. 1135–1144, 2016.
* Robnik-Šikonja (2016) Marko Robnik-Šikonja. Data generators for learning systems based on RBF networks. _IEEE Transactions on Neural Networks and Learning Systems_ , 27(5):926–938, 2016.
* Robnik-Šikonja & Kononenko (2008) Marko Robnik-Šikonja and Igor Kononenko. Explaining classifications for individual instances. _IEEE Transactions on Knowledge and Data Engineering_ , 20:589–600, 2008.
* Robnik-Šikonja (2019) Marko Robnik-Šikonja. _semiArtificial: Generator of Semi-Artificial Data_ , 2019. URL https://cran.r-project.org/package=semiArtificial. R package version 2.3.1.
* Rousseeuw (1987) Peter J. Rousseeuw. Silhouettes: A graphical aid to the interpretation and validation of cluster analysis. _Journal of Computational and Applied Mathematics_ , 20:53 – 65, 1987.
* Saito et al. (2020) Sean Saito, Eugene Chua, Nicholas Capel, and Rocco Hu. Improving lime robustness with smarter locality sampling, 2020.
* Selbst & Barocas (2018) Andrew D Selbst and Solon Barocas. The intuitive appeal of explainable machines. _Fordham Law Review_ , 87:1085, 2018.
* Shapley (1988) Lloyd S. Shapley. A value for n-person games. In Alvin E. Roth (ed.), _The Shapley Value: Essays in Honor of Lloyd S. Shapley_ , pp. 31–40. Cambridge University Press, 1988.
* Slack et al. (2020) Dylan Slack, Sophie Hilgard, Emily Jia, Sameer Singh, and Himabindu Lakkaraju. Fooling LIME and SHAP: Adversarial attacks on post-hoc explanation methods. In _AAAI/ACM Conference on AI, Ethics, and Society (AIES)_ , 2020\.
* Štrumbelj & Kononenko (2010) Erik Štrumbelj and Igor Kononenko. An efficient explanation of individual classifications using game theory. _Journal of Machine Learning Research_ , 11:1–18, 2010\.
* Štrumbelj & Kononenko (2013) Erik Štrumbelj and Igor Kononenko. Explaining prediction models and individual predictions with feature contributions. _Knowledge and Information Systems_ , 41:647–665, 2013\.
## Appendix A Details on post-hoc explanation methods
For the sake of completeness, we present further details on the explanation
methods LIME (Ribeiro et al., 2016), SHAP (Lundberg & Lee, 2017), and IME
(Štrumbelj & Kononenko, 2013). Their complete description can be found in the
above-stated references. In our exposition of the explanation methods, we
denote with $f$ the predictive model, with $x$ the instance we are explaining,
and with $n$ the number of features describing $x$.
### A.1 LIME
Explanations of instances with the LIME method is obtained with an
interpretable model $g$. The model $g$ has to be both locally accurate (so
that it can obtain correct feature contributions) and simple (so that it is
interpretable).
The explanation of the instance $x$ for the predictive model $f$, obtained
with the LIME method, is defined with the following equation:
$\xi(x)=\operatorname*{arg\,min}_{g\in
G}(\mathcal{L}(f,g,\pi_{x})+\Omega(g)),$ (2)
where $\Omega(g)$ denotes the measure of complexity for interpretable model,
$\pi_{x}(z)$ denotes the proximity measure between $x$ and generated instances
$z$, $\mathcal{L}(f,g,\pi_{x})$ denotes the measure of local fidelity of
interpretable model $g$ to the prediction model $f$, and $G$ denotes the set
of interpretable models. We use the linear version of the LIME method, where
$G$ represents the set of linear models. With $x^{\prime}$ we denote the
normalized presentation of instance $x$, i.e. the numerical attributes have
their mean set to 0 and variance to 1, and the categorical attributes contain
value 1 if they have the same value as the exoplained instance and 0
otherwise. The proximity function $\pi_{x}$ is defined on the normalized
instances (hence we use the notation $\pi_{x^{\prime}}$), and uses the
exponential kernel:
$\pi_{x^{\prime}}(z^{\prime})=e^{\frac{d(x^{\prime},z^{\prime})}{\sigma^{2}}}$,
where $d(x^{\prime},z^{\prime})$ denotes a distance measure between
$x^{\prime}$ and $z^{\prime}$. The local fidelity measure $\mathcal{L}$ from
Equation 2 is defined as:
$\mathcal{L}(f,g,\pi_{x^{\prime}})=\sum_{z^{\prime}\in\mathcal{Z}}\pi_{x^{\prime}}(z^{\prime})(f(z)-g(z^{\prime}))^{2},$
(3)
where $\mathcal{Z}$ denotes the set of samples. In LIME, each generated sample
is obtained by adding Gaussian noise to each feature of $x^{\prime}$
independently.
Using linear models as the set of interpretable models, LIME is relatively
fast but may produce poor explanations for instances close to complex decision
boundaries.
### A.2 SHAP
We refer to SHAP as the method called Kernel SHAP by (Lundberg & Lee, 2017).
SHAP essentially estimates Shapley values using LIME’s approach, which means
that the calculation of explanations is fast due to the use of local linear
models computed with the weighted least squares algorithm. The explanation of
the instance $x$ with the SHAP method are feature contributions $\phi_{i}$,
$i=1,2,...,n$ that are coefficients of the linear model
$g(x^{\prime})=\phi_{0}+\sum_{i=1}^{n}\phi_{i}\cdot x^{\prime}_{i}$, obtained
with LIME, where $x^{\prime}_{i}\in\\{0,1\\}$ for $i\in\\{1,2,...,n\\}$.
As LIME does not compute Shapley values, this property is assured with the
proper selection of functions $\Omega(g),\pi_{x^{\prime}}(z^{\prime})$, and
$\mathcal{L}(f,g,\pi_{x^{\prime}})$. Let us first define the function
$h_{x}(z^{\prime})$ that maps from the $\\{0,1\\}^{n}$ to the original feature
space. The function $h_{x}$ is defined implicitly with the equation
$f(h_{x}(z^{\prime}))=E[f(z)|z_{i}=x_{i}\ \forall
i\in\\{j;z^{\prime}_{j}=1\\}]$. This is the expected value of the model $f$
when we do not know the values of features at indices where $z^{\prime}$
equals $0$ (these values are hidden). Functions
$\Omega(g),\pi_{x^{\prime}}(z^{\prime})$ and
$\mathcal{L}(f,g,\pi_{x^{\prime}})$ that enforce the computation of Shapley
values are:
$\Omega(g)=0,$
$\pi_{x^{\prime}}(z^{\prime})=\dfrac{n}{\binom{n}{|z^{\prime}|}\cdot|z^{\prime}|\cdot(n-|z^{\prime}|)},$
$\mathcal{L}(f,g,\pi_{x^{\prime}})=\sum_{z^{\prime}\in\mathcal{Z}}(f(h_{x}(z^{\prime}))-g(z^{\prime}))^{2}\cdot\pi_{x^{\prime}}(z^{\prime}),$
where $|z^{\prime}|$ denotes the number of nonzero features of $z^{\prime}$
and $\mathcal{Z}\subseteq 2^{\\{0,1\\}^{n}}$.
The main purpose of the sampling in this method is to determine the value
$f(h_{x}(z^{\prime}))$ because in general predictive models cannot work with
hidden values. To determine $f(h_{x}(z^{\prime}))$, SHAP uses the distribution
set $D$ which we obtain from the training set. For $D$, SHAP takes the
centroids of clusters obtained by the k-means clustering algorithm on the
training set. The number of clusters is set by the user. Value of
$f(h_{x}(z^{\prime}))$ is determined by the following sampling:
$f(h_{x}(z^{\prime}))=E[f(z)|z_{i}=x_{i}\ \forall
i\in\\{j;z^{\prime}_{j}=1\\}]=\dfrac{1}{|D|}\sum_{d\in
D}f(x_{[x_{i}=d_{i},z^{\prime}_{i}=0]}),$ (4)
where $x_{[x_{i}=d_{i},z^{\prime}_{i}=0]}$ denotes instance $x$ with features
that are $0$ in $z^{\prime}$ being set to the feature values from $d$.
### A.3 IME
The explanation of $x$ with the method IME are feature contributions
$\phi_{i}$, $i=1,2,...,n$. Štrumbelj & Kononenko (2013) shoved that Shapley
values are the only solution that takes into account contributions of
interactions for every subset of features in a fair way. The $i$-th feature
contribution can be calculated with the following expresiion:
$\phi_{i}(x)=\dfrac{1}{n}\sum_{\pi\in
S_{n}}\sum_{w\in\mathcal{X}}p(w)\cdot(f(w_{[w_{i}=x_{i},i\in
Pre^{i}(\pi)\cup\\{i\\}]})-f(w_{[w_{i}=x_{i},i\in Pre^{i}(\pi)]})),$ (5)
where $S_{n}$ denotes a group of permutations of $n$ elements, $\mathcal{X}$
denotes the training set instances, $Pre^{i}(\pi)$ represents the set of
indices that precedes $i$ in the permutation $\pi$, i.e.
$Pre^{i}(\pi)=\\{j;\pi(j)<\pi(i)\\}$). Let $p(w)$ denote the probability of
the instance $w$ in $\mathcal{X}$ and let $w_{[formula]}$ denote the instance
$w$ with some of its features values changed according to the formula. To
calculate $\phi_{i}(x)$, we have to go through $|\mathcal{X}|\cdot n!$
iterations, which can be slow. Therefore, the method IME uses the following
sampling. The sampling population of $i$-th feature is
$V_{\pi,w}=(f(w_{[w_{i}=x_{i},i\in
Pre^{i}(\pi)\cup\\{i\\}]})-f(w_{[w_{i}=x_{i},i\in Pre^{i}(\pi)]}))$ for every
combination of the permutation $\pi$ and instance $w$. IME draws $m_{i}$
samples $V_{1},...,V_{m_{i}}$ at random with repetition. The estimate of
$\phi_{i}(x)$ is defined with the equation:
$\hat{\phi_{i}}=\dfrac{1}{m_{i}}\sum_{i=1}^{m_{i}}V_{i}.$ (6)
Contrary to SHAP, IME does not use approximation with linear models, which
compute all features’ contributions at once but has to compute the Shapley
values by averaging over a large enough sample for each feature separately.
This makes the method slower but also potentially more robust as the method
does not assume the shape of the dependency in the space of normalized
features.
## Appendix B Demonstration of better sampling in SHAP
The graphs in Figure 3 show the PCA based 2D space of the evaluation part of
the COMPAS dataset (see Section 4 for the dataset description). The left-hand
side shows the SHAP-generated sampled instances using the k-means algorithm
(14 clusters determined by the silhouette score (Rousseeuw (1987)). The sample
produced with the MCD-VAE generator in gSHAP is shown on the right-hand side.
This sample is much more similar to the original distribution compared to the
SHAP sampling.
|
---|---
Figure 3: Visual comparison of original and sampled distributions for the
COMPASS dataset. The SHAP k-means based generator (left) produces instances
less similar to the original data, compared to the MCD-VAE generator (right).
## Appendix C Improved IME convergence rate with the TreeEnsemble generator
Preliminary, we tested how better generators affect the convergence rate of
the IME explanation method. The preliminary results in Table 2 show a
significant reduction in the number of needed samples and a slight increase in
the error compared to the original perturbation sampling. Note that the error
measure deployed is biased in favor of the perturbation sampling, which was
used to determine the gold standard. This was determined with the sampling
population’s variance, as described in Štrumbelj & Kononenko (2010).
COMPAS dataset
| Error | # samples |
---|---|---|---
Classifier | Perturb. | TEnsFillIn | Perturb. | TEnsFillIn | Reduction % | CA %
Naive Bayes | 0.0076 | 0.0217 | 18571 | 11278 | 39 | 83
Linear SVM | 0.0033 | 0.0080 | 8244 | 4423 | 46 | 84
Random forest | 0.0049 | 0.0221 | 45372 | 26960 | 41 | 80
Neural network | 0.0057 | 0.0130 | 16157 | 8841 | 45 | 84
German dataset
| Error | # samples |
---|---|---|---
Classifier | Perturb. | TEnsFillIn | Perturb. | TEnsFillIn | Reduction % | CA %
Naive Bayes | 0.0076 | 0.0201 | 56052 | 39257 | 30 | 77
Linear SVM | 0.0005 | 0.0013 | 3877 | 2157 | 44 | 69
Random forest | 0.0046 | 0.0141 | 92478 | 66639 | 28 | 74
Neural network | 0 | 0 | 0 | 26 | / | 69
CC dataset
| Error | # samples |
---|---|---|---
Classifier | Perturb. | TEnsFillIn | Perturb. | TEnsFillIn | Reduction % | CA %
Naive Bayes | 0.0028 | 0.0046 | 32910 | 20117 | 39 | 70
Linear SVM | 0.0009 | 0.0048 | 73324 | 39098 | 47 | 62
Random forest | 0.0032 | 0.0045 | 109958 | 58852 | 46 | 79
Neural network | 0.0012 | 0.0061 | 144183 | 70020 | 51 | 72
Table 2: Comparison of original perturbation sampling and the TreeEnsemble
generator with data fill-in inside the IME method. The results show average
scores for the evaluation set. The column Perturb. presents the perturbation
based sampling and TEnsFillIn presents the sampling using TreeEnsemble
generator with missing parts of instances filled in. The Reduction column
shows the reduction in the number of samples using the TEnsFillIn method
compared to perturbations. The CA stands for the classification accuracy of
the explained classifier on the evaluation set.
## Appendix D Training discriminator function of the attacker
Details of training attacker’s decision models $d$ (see Figure 1b) is
described in Algorithms 1, 2, and 3. We used a slightly different algorithm
for each of the three explanation methods, LIME, SHAP, and IME, as each method
uses a different sampling. Algorithms first create out-of-distribution
instances by method-specific sampling. The training sets for decision models
are created by labeling the created instances with $0$; the instances from
sample $S$ (to which the attacker has access) from distribution $X_{dist}$ are
labeled with $1$. Finally, the machine learning model dModel is trained on
this training set and returned as $d$. In our experiments, we used random
forest classifier as dModel and the training part of each evaluation dataset
as $S$.
Input: $S=\\{(x_{i})_{i=1}^{m}\\}$: training set, nSamples: number of
generated instances for each instance $x_{i}\in D$, $gen$: data generator,
dModel: machine learning algorithm
Output: Classifier $d$ that outputs $1$ if its input $x$ is from $X_{dist}$
and $0$ otherwise
$X\leftarrow\emptyset$ // Training set for dModel
$gen.fit(S)$ // Train the data generator on $S$
for _$i=1$ to $m$_ do
$X\leftarrow X\cup(x_{i},1)$ // Add an instance from distribution
$G\leftarrow gen.newdata(nSamples,x_{i})$ // Generate nSamples new samples
around $x_{i}$
for _$j=1$ to nSamples_ do // Add nSamples out of distribution instances
$X\leftarrow X\cup(G[j],0)$ // Add $j$-th instance from set $G$ to $X$
end for
end for
$d\leftarrow dModel.fit(X)$ // Fit model dModel to set $X$ and save it in $d$
return $d$
Algorithm 1 Training of the decision model $d$, used by the attacker to
distinguish between instances from distribution $X_{dist}$ and samples
produced by explanation methods LIME or gLIME.
Input: $S=\\{(x_{i})_{i=1}^{m}\\}$: training set, nSamples: number of
generated instances for each instance $x_{i}\in D$, $k$: size of the generated
distribution set, $gen$: data generator, dModel: machine learning algorithm
Output: Classifier $d$ that outputs $1$ if its input $x$ is from $X_{dist}$
and $0$ otherwise
$X\leftarrow\emptyset$ // Training set for dModel
$gen.fit(S)$ // Train the data generator on $S$
if _$gen==KMeans$ or $gen==rbfDataGen$ or $gen==treeEnsemble$_ then
$D\leftarrow gen.newdata(k)$ // Generate the distribution set with KMeans,
rbfDataGen or treeEnsemble
end if
for _$i=1$ to nSamples_ do // Add nSamples out of distribution instances
$x\leftarrow$ random instance from $S$
if _$gen==MCD-VAE$ or $gen==treeEnsembleFill$ _ then
$w\leftarrow gen.newdata(1,x)$ // Generate an instance $w$ in the vicinity of
$x$
end if
else // KMeans, treeEnsemble or rbfDataGen
$w\leftarrow$ take a random instance from $D$
end if
$M\leftarrow$ choose a random subset of $\\{1,2,...,len(x)\\}$ // Choose
random features
$x[M]\leftarrow w[M]$ // Replace the values of chosen features in $x$ with
values from $w$ as in SHAP method
$X\leftarrow X\cup(x,0)$ // Add out of distribution instance
end for
for _$i=1$ to $m$_ do // Add instances from distribution
$X\leftarrow X\cup(x_{i},1)$
end for
$d\leftarrow dModel.fit(X)$ // Fit model dModel to set $X$ and save it in $d$
return $d$
Algorithm 2 Training of the decision model $d$, used by attacker to
distinguish between instances from distribution $X_{dist}$ and samples
produced by explanation methods SHAP or gSHAP.
Input: $S=\\{(x_{i})_{i=1}^{m}\\}$: training set, nSamples: number of
generated instances for each instance $x_{i}\in D$, $gen$: data generator,
dModel: machine learning algorithm
Output: Classifier $d$ that outputs $1$ if its input $x$ is from $X_{dist}$
and $0$ otherwise
$X\leftarrow\emptyset$ // Training set for dModel
$gen.fit(S)$ // Train the data generator on set $S$
for _$i=1$ to $m$_ do
$X\leftarrow X\cup(x_{i},1)$ // Add an instance from distribution
for _$j=1$ to nSamples_ do // Add nSamples out of distribution instances
$w\leftarrow gen.newdata(1,x_{i})$ // Generate an instance $w$ in the vicinity
of $x_{i}$
$b_{1}\leftarrow w$ // First out of distribution instance
$b_{2}\leftarrow w$ // Second out of distribution instance
$\pi\leftarrow$ choose a random permutation from $S_{len(x_{i})}$ // i.e. a
random permutation of $x_{i}$’s features
$idx\leftarrow$ choose a random number from $\\{1,2,...,len(x_{i})\\}$
$M_{1}\leftarrow\\{k\in\\{1,2,...,len(x_{i})\\},\pi(k)<\pi(idx)\\}$ //
Features that precede $idx$ in permutation $\pi$
$M_{2}\leftarrow M_{1}\cup\\{idx\\}$
$b_{1}\leftarrow x_{i}[M_{1}]$ // Vector $b_{1}$ as in IME method (Štrumbelj &
Kononenko (2013))
$b_{2}\leftarrow x_{i}[M_{2}]$ // Vector $b_{2}$ as in IME method (Štrumbelj &
Kononenko (2013))
$X\leftarrow X\cup\\{(b_{1},0),(b_{2},0)\\}$ // Add out of distribution
instances
end for
end for
$d\leftarrow dModel.fit(X)$ // Fit model dModel to set $X$ and save it in $d$
return $d$
Algorithm 3 Training of the decision model $d$, used by attacker to
distinguish between instances from distribution $X_{dist}$ and samples
produced by explanation methods IME or gIME.
## Appendix E Heatmaps as tables
We present the information contained in Figure 2 in a more detailed tabular
form in Table 3.
Table 3: The robustness results for gLIME (top table), gSHAP (middle table),
and gIME (bottom table). The tables show the proportion of evaluation set
instances, where the sensitive feature was recognized as the most important by
the used explanation method. Columns represent the generators used for
explanations. The row labels consist of the name of the dataset on which the
experiment was performed and the name of the generator used for training of
the adversarial model. Compas2 and CC2 denote attacks with two independent
features. Perturbation represents the original sampling used in LIME, SHAP,
and IME, TEnsFillIn represents a variant of the TreeEnsemble generator where
new instances are generated around the given one, and TreeEns represents the
generation from the whole distribution.
## Appendix F Comparing explanations of original and modified methods
We check if improved data generators affect explanations in non-adversary
environment. We split the dataset into training and evaluation set in the
ratio $90\%:10\%$, and trained four classifiers from Python scikit-learn
(Pedregosa et al. (2011)) library: Gaussian naive Bayes, linear SVC (SVM for
classification), random forest, and neural network. We explained the
predictions of each classifier on the evaluation set with every combination of
explanation methods and generators used in the adversarial attack experiments.
For instances in the evaluation set, we measured the mean absolute difference
(MAD) of modified explanation methods, defined with the following equation:
$\text{MAD}_{gen}(x)=\frac{1}{n}\sum_{i=1}^{n}|\phi_{i}^{gen}(x)-\phi_{i}(x)|,$
(7)
where $\phi_{i}^{gen}(x)$ and $\phi_{i}(x)$ represent the explanations of
$i$-th feature returned by the modified and original explanation method,
respectively (recall that $n$ denotes the number of features in the data set).
We experimented on three datasets. The COMPAS dataset is described in Section
4.1. In addition to that, we used synthetic dataset condInd from Robnik-
Šikonja & Kononenko (2008), and Ionosphere dataset from UCI repository (Dua &
Graff (2019)). Both datasets represent a binary classification problem. Apart
from the target variable, condInd consists of $8$ binary features, while
Ionosphere consists of $34$ numerical attributes. The condInd datasets
contains $2000$ instances and Ionosphere contains $351$ instances.
The results are shown in Table 4. The differences between original LIME and
gLIME explanations are considerable (see the top table). This is not
surprising since LIME fits local linear models in its local explanations,
which can strongly differ even for small perturbations of the model’s input.
SHAP and IME explanations are very similar (the average MAD is almost
negligible). We can conclude that explanations of gSHAP and gIME are not
significantly different from SHAP and IME in the non-adversary environment.
Table 4: The differences in explanations between LIME and gLIME (top table),
SHAP and gSHAP (middle table), and IME and gIME (bottom table). The tables
show MAD and the variance of MAD on evaluation set for every combination of
the explanation method and predictive model. Columns represent MAD and the
variance of MAD on used datasets. The row labels consist of the used
classifier and generator.
## Appendix G Adversarial classifiers with different prediction thresholds
In a realistic scenario, the attacker might be wary of being discovered by the
regulator and decide to employ deception only when it is really certain that
the predicted instance is used inside the explanation method. In these
circumstances, the number of times when the sensitive variable is used inside
the model will also depend on how aggressively the biased classifier is used.
This reflection leads to another metric for the defense’s success, which
measures the proportion of instances when the sensitive variable is identified
when it is being used, i.e. at different levels of conservatism.
We express different levels of attacker’s conservatism by changing the
threshold of the decision function $d$. The output of the adversarial model
$e_{t}$, which uses threshold $t$ is now defined with the equation:
$e_{t}(x)=\begin{cases}b(x),&d_{prob}(x)\geq t\\\
\psi(x),&d_{prob}(x)<t\end{cases},$ (8)
where $d_{prob}(x)$ denotes the probability that $x$ is from distribution
$X_{dist}$, returned by decision model $d$. Threshold $t$ therefore represents
the required certainty of the attacker that given instance is from $X_{dist}$
before it deploys the biased classifier. Larger values of $t$ will result in
biased classifier being deployed fewer times but with larger certainty that it
has to be used. If we use $t=0.5$ in Equation (8), we get the adversarial
model that behaves identically to the one, described with Equation (1) in
Section 2.2.
In this experiment, we used only the COMPAS dataset, which we split into the
training and evaluation set in ratio $90\%:10\%$. We performed the same
experiment as in section 4.2, but with different thresholds for the
adversarial models. We used five different thresholds: $0.3,0.4,0.5,0.6$ and
$0.7$.
We first checked how different thresholds affect the deployment of the biased
model. For each generator used in the adversarial model’s training and the
certainty threshold used by the adversarial model, we measured the proportion
of instances from the evaluation set, where the biased classifier was
deployed. The results are shown in Table 5. As expected, the percentage of
instances where the biased classifier is deployed decreases with increasing
$t$. The biased classifier is deployed less frequently when using treeEnsemble
and rbfDataGen in its training, which is also expected, as these two
generators give the best results inside the explanation methods. This means
that for these generators, it is more difficult for an attacker to distinguish
between the actual instances and generated instances. The biased classifier is
deployed the most often in the case of the LIME adversarial model and least
often in the SHAP adversarial model.
Next, we tested the proportion of instances where the sensitive feature (in
case of COMPAS dataset this is race) was recognized as the most important
feature by the explanation method on two different subsets of instances:
instances from the evaluation set on which the biased classifier was deployed
and the whole evaluation set. We measured these numbers for all combinations
of thresholds and explanation methods. The results are shown in Tables 6
(gLIME), 7 (gSHAP) and 8 (gIME) for gLIME, gSHAP, and gIME, respectively. We
can observe the same pattern as in section 4.2: gIME is still the most robust,
and the treeEnsemble generator still gives the best results. The sensitive
attribute is recognized as the most important feature more often when the
decision threshold values are lower, which is expected as in these cases, the
adversarial model behaves more aggressively. Whilst the percentage of the
instances on which the racist behavior of the adversarial model is recognized
drops with higher thresholds, it remains high enough to warn the regulator
about the problematic behavior of the prediction model (especially in the case
of gSHAP and gIME using treeEnsemble as data generator). From that, we can
conclude that modified explanation methods remain robust enough, even with
more conservative adversaries.
Table 5: Proportions of instances in % of the evaluation set on which the
biased classifier was deployed for adversarial LIME (top table), SHAP (middle
table) and IME model (bottom table). Columns represent different threshold
used for deploying the biased classifier. The rows represent the generator
used in training of the adversarial attack. The labels ”One unrelated” and
”Two unrelated” represent the attacks with one or two unrelated features.
Table 6: Percentages of instances where the sensitive feature (race) was
recognized as the most important feature with gLIME for adversarial attacks
with one (top table) or two unrelated features (bottom table). The columns
labeled Biased pred. represent the results on instances on which the biased
classifier was deployed, while the columns labeled All represent the results
on the whole evaluation sets. The numbers above represent the used threshold.
The row labels are in the form $<$explainer$>$_$<$adversarial$>$ where
$<$explainer$>$ denotes the generator used in the explanation method and
$<$adversarial$>$ denotes the generator used in the training of the
adversarial model.
Table 7: Percentage of instances where the sensitive attribute (race) was
recognized as the most important feature with gSHAP for adversarial attacks
with one (top table) or two unrelated features (bottom table). The columns
labeled Biased pred. represent the results on instances on which the biased
classifier was deployed, while columns labeled All represent the results on
the whole evaluation sets. The numbers above represent the used threshold. The
row labels are in the form $<$explainer$>$_$<$adversarial$>$ where
$<$explainer$>$ denotes the generator used in the explanation method and
$<$adversarial$>$ denotes the generator used in the training of the
adversarial model.
Table 8: Percentages of instances where the sensitive attribute (race) was
recognized as the most important feature with gIME for adversarial attacks
with one (top table) or two unrelated features (bottom table). The columns
labeled Biased pred. represent the results on instances on which the biased
classifier was deployed, while the columns labeled All represent the results
on the whole evaluation sets. The numbers above represent the used threshold.
The row labels are in the form $<$explainer$>$_$<$adversarial$>$ where
$<$explainer$>$ denotes the generator used in the explanation method and
$<$adversarial$>$ denotes the generator used in the training of the
adversarial model.
|
# Unified Feature Extraction Framework based on Contrastive Learning
Hongjie Zhang, Wenwen Qiang, Jinxin Zhang, and Ling Jing This work was
supported by Next Generation Precision Aquaculture: R&D on intelligent
measurement, control technology (No. 2017YFE0122100), the National Natural
Science Foundation of China (Nos. 62076244, 12071024).H. Zhang is associated
with the College of Information and Electrical Engineering, China Agricultural
University, Beijing 100083, PR China; Beijing Engineering and Technology
Research Center for Internet of Things in Agriculture, Beijing 100083, PR
China. W. Qiang is associated with the University of Chinese Academy of
Sciences, Beijing, PR China, in addition to the Science & Technology on
Integrated Information System Laboratory, Institute of Software Chinese
Academy of Sciences, Beijing, PR China.J. Zhang is associated with the College
of Information and Electrical Engineering, China Agricultural University,
Beijing 100083, PR China; National Innovation Center for Digital Fishery,
China Agricultural University, Beijing 100083, PR China. L. Jing is associated
with the College of Science, China Agricultural University, Beijing 100083, PR
China; Key Laboratory of Agricultural Information Acquisition Technology,
Ministry of Agriculture, Beijing 100083, PR China. (Corresponding author: Ling
Jing, email<EMAIL_ADDRESS>
###### Abstract
Feature extraction is an efficient approach for alleviating the issue of
dimensionality in high-dimensional data. As a popular self-supervised learning
method, contrastive learning has recently garnered considerable attention. In
this study, we proposed a unified framework based on a new perspective of
contrastive learning (CL) that is suitable for both unsupervised and
supervised feature extraction. The proposed framework first constructed two CL
graph for uniquely defining the positive and negative pairs. Subsequently, the
projection matrix was determined by minimizing the contrastive loss function.
In addition, the proposed framework considered both similar and dissimilar
samples to unify unsupervised and supervised feature extraction. Moreover, we
propose the three specific methods: unsupervised contrastive learning method,
supervised contrastive learning method 1 ,and supervised contrastive learning
method 2. Finally, the numerical experiments on five real datasets
demonstrated the superior performance of the proposed framework in comparison
to the existing methods.
###### Index Terms:
feature extraction, dimension reduction, self-supervised learning, contrastive
learning
## I Introduction
Currently, high-dimensional data is widely used in pattern recognition and
data mining[1, 2, 3], which wastes considerable time and cost apart from
causing the problem known as “curse of dimensionality”[4]. Thus, feature
extraction of data poses immense significance to solve this issue[5, 6, 7, 8].
Recently, self-supervised learning[9, 10, 11] has become a popular topic in
the field of deep learning. Self-supervised learning is a method of
unsupervised learning that uses data information to supervise itself by
constructing positive and negative pairs. Self-supervised learning strives to
learn more discriminative features to effectively bridge unsupervised and
supervised learning. Consequently, contrastive learning[12, 13, 14, 15] has
attracted extensive scholarly attention as the primary method of self-
supervised learning. Tian et al. proposed contrastive multiview coding (CMC)
to process multiview data[16]. First, CMC constructs the same samples in any
two views as positive pairs and dissimilar samples as negative pairs, and
subsequently optimizes a neural network framework by minimizing the
contrastive loss function to maximize the similarity of the projected positive
pairs. In addition, Chen et al. proposed a simple framework for contrastive
learning (SimCLR) [17] to process the problem. First, it performs data
enhancement to obtain distinct representations of the same sample to consider
them as positive pairs, and thereafter, considers the representations of any
two distinct samples as negative pairs. Finally, SimCLR optimizes the network
framework by minimizing the contrastive loss, similar to CMC. However, for the
problem of single view feature extraction, the existing models based on
contrastive learning still have many defects. For example, data enhancement
will increase the running time of the algorithm, and defining all different
samples as negative pairs will lead to the separation of samples of the same
class.
Inspired by our prior research, we proposed a unified feature extraction
framework based on contrastive learning (CL-UFEF) that is suitable for both
unsupervised and supervised feature extraction for single-view data. The
proposed framework, CL-UFEF, constructed two contrastive graphs (CLG) to
establish a new approach for defining positive and negative pairs, instead of
conducting data enhancement. Moreover, the contrastive loss function considers
both similar and dissimilar samples based on the CLG to unify the aspects of
unsupervised and supervised feature extraction. Furthermore, the effectiveness
of the proposed framework was verified by three specific models, including the
unsupervised contrastive learning method (u-CL), supervised contrastive
learning method 1 (s-CL1), and supervised contrastive learning method 2
(s-CL2), respectively.
The main contributions of this study are as follows:
* •
A unified feature extraction framework for single-view data based on
contrastive learning (CL-UFEF) is proposed from a new perspective that is
suitable for both unsupervised and supervised cases.
* •
CL-UFEF proposes a novel approach to define positive and negative pairs in
contrastive learning.
* •
Based on CL-UFEF, we propose three specific models, including u-CL, s-CL1, and
s-CL2, and the experiments on five real datasets proved the advantages of the
proposed framework.
The remainder of this article is organized as follows. The traditional feature
extraction methods are briefly introduced in Section II. Subsequently, the
development of the unified feature extraction framework (CL-UFEF) and the
u-CL, s-CL1, and s-CL2 models are discussed in Section III. In addition, the
extensive experiments conducted on several real-world datasets are presented
in Section IV. Finally, the conclusions of the current study are detailed in
Section V.
## II Related Work
Recently, several feature extraction methods[18, 19, 20, 21, 22] have been
proposed, which can be categorized into unsupervised[23], semi-supervised[24],
and supervised[25] methods. In this study, we focused on unsupervised and
supervised feature extraction. In the unsupervised feature extraction models,
classical principal-component analysis (PCA)[18] provides a wide range of
applicability and effectiveness, as it seeks the maximum variance of samples
in the subspace, which is more conducive to subsequent classification,
clustering, or other tasks. Nonetheless, PCA is a linear feature extraction
method that is unsuitable for handling nonlinear data. With the development of
manifold learning, numerous nonlinear feature extraction methods such as
isometric feature mapping (ISOMAP)[26], Laplacian eigenmap (LE)[27], and local
linear embedding (LLE)[28] have been proposed to solve this problem. However,
these nonlinear feature extraction methods cannot be applied to new sample
points because they directly obtain the low-dimensional representation of
samples without using a projection matrix. Therefore, several nonlinear
methods based on manifold assumptions have been modified into linearized
versions. In particular, locality preserving projections (LPP)[29],
neighborhood preserving embedding (NPE)[30], and isometric projection are
considered as linearized LE, LLE, and ISOMAP, respectively. Although all the
feature extraction methods preserve the manifold structures in subspace, they
have various requirements for manifold learning. For instance, LPP obtains the
neighbor graph of the original data in advance, and then the samples in the
subspace maintain the same neighbor relationship. In contrast, NPE expects
that the linear reconstruction relationship of the original space can be
maintained between the samples and neighborhood samples after feature
extraction. However, the above graph-based methods only considered the local
structures of the data and neglected the global structures. Thus, sparsity
preserving projections (SPP)[31], collaborative representation-based
projections (CRP)[32], and low-rank preserving embedding (LRPE)[33] have been
proposed to effectively solve this problem. SPP constructs a $l1$-graph with
adaptive neighbors by utilizing the sparsity technique, whereas CRP aims to
build a $l2$-graph by calculating the linear reconstruction coefficients of
each sample based on the remaining samples; LRPE constructs a nuclear norm
graph with adaptive neighbors using low-rank representation. However, the
unsupervised graph-based methods closely projects the original similar samples
in the subspace and neglects the dissimilar samples.
Supervised feature extraction methods obtain more discriminant information
using sample labels. For instance, linear discriminant analysis (LDA)[34]
seeks an embedding transformation such that the within-class scatter is
minimized and the between-class scatter is maximized. However, LDA, similar to
PCA, is a linear feature extraction method as well. Consequently, it might
deliver poor performance if the samples in a class form several separate
clusters (i.e., multi-mode). Thus, researchers have proposed using local
Fisher discriminant analysis (LFDA)[35] and marginal Fisher analysis (MFA)[36]
based on GE and utilizing the approach of manifold learning in unsupervised
feature extraction models. LFDA combines the ideas of LDA and LPP to locally
construct the levels of within- and between-class scatters, which allows the
LFDA to simultaneously achieve maximum preservation of the within- and
between-class local structures at the same time. MFA is distinct from the LFDA
as it considers the local structure within classes and constructs the local
structure relationship between classes by accounting for the samples on the
edges of various classes. However, MFA is limited by the problem of class
isolation, i.e., not all samples of heterogeneous edges have local neighbor
relationships. In this context, researchers have proposed multiple marginal
Fisher analysis (MMFA)[37], which selects the nearest neighbor samples on all
heterogeneous edges to construct the local relationship between classes.
Sparsity preserving discriminant projections (SPDP)[38] is proposed based on
SPP to maintain the sparse reconstruction coefficients of samples in the
subspace. Note that these graph-based methods initially learn an affine graph
using various measure metrics, and subsequently evaluate the projection based
on graphs.
Inspired by traditional graph-based methods, we proposed a unified feature
extraction framework based on contrastive learning (CL-UFEF). Concretelly,
compared with the previous models based on contrastive learning, CL-UFEF does
not need data enhancement, and it constructs positive and negative pairs based
on two contrastive learning graphs (CLG), which will make the similar samples
in the subspace more clustered. Compared with the traditional graph-based
models, CL-UFEF is suitable for both unsupervised and supervised feature
extraction, and it considers similar and dissimilar samples in unsupervised
and spervised learning based on the CLG.
## III Methodology
Figure 1: Process of determining projection matrix $P$ using CL-UFEF. TABLE I: Notations and definitions. $X$ | Training sample set
---|---
$Y$ | Set of training samples in a low-dimensional space
$n$ | Number of training samples
$D$ | Dimensionality of the samples in the original space
$d$ | Dimensionality of embedding features
$c_{i}$ | Labels of samples $x_{i}$
$C$ | Number of classes
$P$ | Projection matrix
$CLG$ | Contrastive learning graph
$G^{pos}$ | Positive graph
$G^{neg}$ | Negative graph
$S^{pos}$ | Similarity matrix for positive pairs
$S^{neg}$ | Dissimilarity matrix of negative pairs
$\sigma$ | Positive parameter
$k$ | Number of neighbors
$NK(x_{j})$ | The $k$ nearest neighbors of $x_{j}$
$NK^{+}_{(x_{j})}$ | The $k$ nearest neighbors of $x_{j}$ in the $c_{j}$ th class
$\nabla L(P)$ | Gradient of $L(P)$ with respect to $P$
$T$ | Number of iterations
In this section, a unified feature extraction framework is proposed based on
contrastive learning (CL-UFEF) for both unsupervised and supervised feature
extraction, including three specific cases: u-CL, s-CL1, and s-CL2. The
projection matrix evaluation process using CL-UFEF is intuitively illustrated
in Figure 1.
This study considered the problem of unsupervised and supervised feature
extraction. Let us mathematically formulate the feature extraction problem as
follows.
Feature extraction problem: Given a training sample set
$X=[x_{1},x_{2},...,x_{n}]\in{R^{D\times n}}$, where $n$ and $D$ are the
number of samples and features, respectively. In the supervised case, the
class label $c_{i}$ of $x_{i}$ was provided, where $c_{i}\in\\{1,2,...,C\\}$
and $C$ represent the number of classes. The purpose of feature extraction is
to find a projection matrix $P\in{R^{D\times d}}$ to derive the low-
dimensional embedding $Y=[y_{1},y_{2},...,y_{n}]\in{R^{d\times n}}$ for $X$
calculated by $Y=P^{T}X$, where $d\ll D$.
For convenience, the symbols used in this study are summarized in Table I.
### III-A Unified Feature Extraction Framework based on Contrastive Learning
(CL-UFEF)
### Construct CLG
The contrastive learning was applied by constructing the two contrastive
learning graphs (CLG): a positive graph $G^{pos}=\\{X,{S^{pos}}\\}$ and a
negative graph $G^{neg}=\\{X,{S^{neg}}\\}$, which define the positive and
negative pairs, including the positive matrix
$S^{pos}=(S^{pos}_{i,j})_{n\times n}$ and the negative matrix
$S^{neg}=(S^{neg}_{i,j})_{n\times n}$ to measure the similarity of the
positive pairs and the dissimilarity of the negative pairs. Morever, it should
be note that the greater number of negative pairs strengthens the influence of
the model based on contrastive learning[39].
In particular, compared with the previous methods based on ciontrastive
learning, $G^{pos}$ and $G^{neg}$ define similar samples as positive pairs and
dissimilar samples as negative pairs. Compared with the traditional graph-
based methods, $G^{pos}$ and $G^{neg}$ are suitable for both supervised and
unsupervised learning, and dissimilar samples are considered in unsupervised
and spervised learning based on the $G^{neg}$.
### Model of CL-UFEF
Contrastive learning is realized using the positive matrix $S^{pos}$ and
negative matrix $S^{neg}$. We hope that the projections of the positive pair
$x_{i}$ and $x_{j}$ with larger $S^{pos}_{i,j}$ should have greater
similarity, and the projections of the negative pair $x_{i}$ and $x_{j}$ with
larger $S^{neg}_{i,j}$ will have greater dissimilarity. Thus, the optimization
problem of the CL-UFEF is proposed as follows:
$\min_{P}L(P)=\sum_{i=1}^{n}-log\frac{\sum_{j=1}^{n}S^{pos}_{i,j}exp(SIM(P^{T}x_{i},P^{T}x_{j}))}{\sum_{j=1}^{n}S^{who}_{i,j}exp(SIM(P^{T}x_{i},P^{T}x_{j}))}$
(1)
where,
$SIM(P^{T}x_{i},P^{T},x_{j})=\frac{x_{i}^{T}PP^{T}x_{j}}{\|P^{T}x_{i}\|\|P^{T}x_{j}\|\sigma}$
(2)
$\sigma$ is a positive parameter, and
$S^{who}_{i,j}=S^{pos}_{i,j}+S^{neg}_{i,j}$ is the whole similarity matrix.
Subsequently, based on the specific construction methods of $G^{pos}$ and
$G^{neg}$, three special cases, including u-CL, s-CL1, and s-CL2 are proposed
to verify the effectiveness of the proposed framework. Morever, some other
construction methods of $G^{pos}$ and $G^{neg}$ are presented in APPENDIX A.
### III-B Unspervised contrastive learning method (u-CL)
We use the k-nearest neighbor method to construct CLG, indicating that the
local $k$ nearest neighbors were positive pairs whereas the non-nearest
neighbors were negative pairs. In particular, the positive matrix $S^{pos}$
and negative matrix $S^{neg}$ can be defined as follows:
$S^{pos}_{i,j}=\left\\{\begin{aligned}
&exp(\frac{-||x_{i}-x_{j}||_{2}^{2}}{t}),&{\rm if}\,x_{i}\in NK(x_{j})\\\
&&{\rm or}\,x_{j}\in NK(x_{i});\\\ &0,&{\rm otherwise},\end{aligned}\right.$
(3)
$S^{neg}_{i,j}=\left\\{\begin{aligned} &1,&{\rm if}\,x_{i}\notin
NK(x_{j})\,{\rm and}\,x_{j}\notin NK(x_{i});\\\ &0,&{\rm
otherwise}.\end{aligned}\right.$ (4)
where $t$ is the thermal parameter used for adjusting the value range of the
weight matrix $S^{pos}$, $NK(x_{j})$ represents the $k$ nearest neighbors of
$x_{j}$ , and the parameter $k$ is a tuning parameter. The optimization
problem of u-CL can be obtained by introducing (3) and (4) into CL-UFEF.
### III-C Spervised contrastive learning method 1 (s-CL1)
We use class information samples to construct CLG, indicating that the within-
class samples are positive pairs and between-class samples are negative pairs.
In particular, the positive matrix $S^{pos}$ and negative matrix $S^{neg}$ can
be defined as
$S^{pos}_{i,j}=\left\\{\begin{aligned} &1,&{\rm if}\,c_{i}=c_{j};\\\ &0,&{\rm
if}\,c_{i}\neq c_{j},\end{aligned}\right.$ (5)
$S^{neg}_{i,j}=\left\\{\begin{aligned} &1,&{\rm if}\,c_{i}\neq c_{j};\\\
&0,&{\rm if}\,c_{i}=c_{j}.\end{aligned}\right.$ (6)
Moreover, the optimization problem of s-CL1 can be obtained by introducing (5)
and (6) into the CL-UFEF.
### III-D Spervised contrastive learning method 2 (s-CL2)
We use class information and k nearest neighbors of samples to construct CLG,
indicating that local $k$ nearest neighbors in the same class were positive
pairs and the samples of other relationships were negative pairs. In
particular, the positive matrix $S^{pos}$ and negative matrix $S^{neg}$ can be
defined as
$S^{pos}_{i,j}=\left\\{\begin{aligned}
&exp(\frac{-||x_{i}-x_{j}||_{2}^{2}}{t}),&{\rm if}\,x_{i}\in NK^{+}(x_{j})\\\
&&{\rm or}\,x_{j}\in NK^{+}(x_{i});\\\ &0,&{\rm
otherwise},\end{aligned}\right.$ (7)
$S^{neg}_{i,j}=\left\\{\begin{aligned} &1,&{\rm if}\,x_{i}\notin
NK^{+}(x_{j})\,{\rm and}\,x_{j}\notin NK^{+}(x_{i});\\\ &0,&{\rm
otherwise}.\end{aligned}\right.$ (8)
where $NK^{+}_{(x_{j})}$ represents the $k$ nearest neighbors of $x_{j}$ in
the same class. The optimization problem of s-CL2 can be obtained by
introducing (7) and (8) into CL-UFEF.
### III-E Optimization algorithm
Algorithm 1 CL-UFEF
Input:
Data matrix: $X\in R^{D\times n}$,
$\alpha,\beta_{1},\beta_{2},\epsilon,P_{0}$.
$m_{0}=0$(Initialize $1^{st}$ moment vector)
$v_{0}=0$(Initialize $2^{nd}$ moment vector)
$t=0$ (Initialize time step)
Output: Projection matrix $P$
while $P_{t}$ not converged do
$t=t+1$$g_{t}=\bigtriangledown L(P_{t-1})$ is calculated using (9) (Obtain
gradients with respect to the stochastic objective at time step
$t$)$m_{t}=\beta_{1}\cdot m_{t-1}+(1-\beta_{1})\cdot g_{t}$ (Update biased
first-moment estimate)$v_{t}=\beta_{2}\cdot v_{t-1}+(1-\beta_{2})\cdot
g_{t}^{2}$ (Update biased second raw-moment
estimate)$\hat{m}_{t}=m_{t}/(1-\beta_{1}^{t})$ (Compute bias-corrected first-
moment estimate)$\hat{v}_{t}=v_{t}/(1-\beta_{2}^{t})$ (Compute bias-corrected
second raw-moment
estimate)$P_{t}=P_{t-1}-\alpha\cdot\hat{m}_{t}/(\sqrt{\hat{v}_{t}}+\epsilon)$
end while
return result
In this section, the optimization algorithm of CL-UFEF is presented (Algorithm
1), which uses the Adam optimizer[40] to solve the optimization problem for
CL-UFEF. Adam is an advancement on the random gradient descent method and can
rapidly yield accurate results. This method calculates the adaptive learning
rate of various parameters based on the budget of the first and second moments
of the gradient. The parameters $\alpha$, $\beta_{1}$, $\beta_{2}$, and
$\epsilon$ represent the learning rate, the exponential decay rate of the
first- and second-order moment estimation, and the parameter to prevent
division by zero in the implementation, respectively. In addition, the
gradient of the loss function $L(P)$ with respect to the projection matrix $P$
was obtained from (9).
The convergence condition of Algorithm 1 was set as
$L(P_{t})-L(P_{t-1})<0.001$, where $L(P_{t})$ and $L(P_{t-1})$ are the
function values obtained after the $t$th and $t-1$th gradient descent,
respectively. Therefore, the computational complexity of Algorithm 1 is
primarily executed in the first step, and the complexity of the derivative of
the objective function is $O(Tn^{3})$, where $T$ denotes the number of
iterations.
$\displaystyle\nabla L(P)=$ (9)
$\displaystyle\sum_{i=1}^{n}\\{-\frac{\sum_{j=1}^{n}S^{who}_{i,j}exp(SIM(P^{T}x_{i},P^{T}x_{j}))}{\sum_{j=1}^{n}S^{pos}_{i,j}exp(SIM(P^{T}x_{i},P^{T}x_{j})}\cdot$
$\displaystyle[\sum_{j=1}^{n}S^{pos}_{i,j}exp(P^{T}x_{i},P^{T}x_{j})\nabla
SIM(P^{T}x_{i},P^{T}x_{j})\cdot$
$\displaystyle\sum_{j=1}^{n}S^{who}_{i,j}exp(P^{T}x_{i},P^{T}x_{j})$
$\displaystyle-\sum_{j=1}^{n}S^{who}_{i,j}exp(P^{T}x_{i},P^{T}x_{j})\nabla
SIM(P^{T}x_{i},P^{T}x_{j})\cdot$
$\displaystyle\sum_{j=1}^{n}S^{pos}_{i,j}exp(P^{T}x_{i},P^{T}x_{j})]/$
$\displaystyle[\sum_{j=1}^{n}S^{who}_{i,j}exp(P^{T}x_{i},P^{T}x_{j})]^{2}\\},$
where
$\displaystyle\nabla SIM(P^{T}x_{i},P^{T}x_{j})=$ (10)
$\displaystyle\\{(x_{i}{x_{j}}^{T}+x_{j}{x_{i}}^{T})P\cdot\|P^{T}x_{i}\|\|P^{T}x_{j}\|\sigma$
$\displaystyle-[({x_{i}}^{T}PP^{T}x_{i})^{-\frac{1}{2}}\cdot
x_{i}{x_{i}}^{T}P\cdot\|P^{T}x_{j}\|\sigma$
$\displaystyle+({x_{j}}^{T}PP^{T}x_{j})^{-\frac{1}{2}}\cdot
x_{j}{x_{j}}^{T}P\cdot\|P^{T}x_{i}\|\sigma]\cdot{x_{i}}^{T}PP^{T}x_{j}\\}/$
$\displaystyle{(\|P^{T}x_{i}\|\|P^{T}x_{j}\|\sigma)}^{2}$
The main computational complexity of each cycle in Algorithm 1 is the
derivation of the loss function in the first step, which is
$O(n^{2}(D^{2}d+Dd+D^{2}))$. Assuming that the Algorithm 1 performs a total of
$T$ cycles when converging, the main computational complexity is
$O(n^{2}T(D^{2}d+Dd+D^{2}))$.
## IV Experimental results
Figure 2: Image samples used in experiments: (a) Multiple features data set.
(b) Yale dataset. (c) COIL20 dataset. (d) MNIST data set. (e) USPS dataset.
### IV-A Data Descriptions and Experimental Setups
In this study, the numerical experiments utilized five real datasets to
demonstrate the performance advantages of the proposed CL-UFEF.
Multiple Features dataset: This dataset was acquired from the UCI machine
learning repository, comprising six distinct feature sets extracted from
handwritten numbers from 0 to 9 with 200 patterns per class (i.e., 2000
patterns in ten classes). All the datasets were digitized in binary images.
The six feature sets included fou (Fourier coefficients, 76 features), fac
(profile correlations, 216 features), kar (Karhunen–Loeve coefficients, 64
features), pix (pixel averages in $2\times 3$ windows, 240 features), zer
(Zernike moments, 47 features), and mor (morphological features, six
features). Example images are presented in Figure 2 (a).
Yale dataset: The dataset was created by Yale University Computer Vision and
Control Center, containing data of $15$ individuals, wherein each person has
11 frontal images ($64\times 64$ pixels in size) captured under various
lighting conditions. The images were edited to $50\times 40$ pixels with 256
Gy levels per pixel; examples are presented in Figure 2 (b).
COIL20 dataset: The COIL20 dataset was created by Columbia University in 1996,
containing 1440 images of 20 objects, wherein each object includes 72 images
($64\times 64$ pixels in size). In addition, each image was edited to
$32\times 32$ pixels, and each pixel had 256 Gy levels; examples are presented
in Figure 2 (c).
MNIST dataset: This dataset contains 70,000 samples of $0--9$ digital images
with a size of $28\times 28$. We randomly selected 2000 images as experimental
data, uniformly rescaled all the images to a size of $16\times 16$, and used a
feature vector of 256-level grayscale pixel values to represent each image.
Examples are presented in Figure 2 (d).
USPS dataset: This dataset contains 9298 samples of $0--9$ handwritten digital
images, and the size of each image was adjusted to $16\times 16$. A total of
1800 images were randomly selected as experimental data, and each image was
represented by a feature vector of 256-level grayscale pixel values. Examples
are depicted in Figure 2 (e).
In data processing, in order to shorten the running time, we first used PCA to
reduce the dimensions of the Yale, COIL20, MNIST, and USPS datasets, and the
details are listed in Table II. Thereafter, we separately standardized these
five datasets to improve the convergence speed of the model. Finally, we
compared u-CL, s-CL1, and s-CL2 with the traditional graph-based methods of
LPP, FLPP, LDA, LFDA, and FDLPP and the method SimCLR beased on contrastive
learning to verify the advantages of our proposed feature extraction
framework. Note that the $k$-nearest neighbor classifier ($k$ = 1) was used in
the experiment. Moreover, six samples of each class were randomly selected
from the Yale, COIL20, and each feature set of multiple feature datasets for
training, and the remaining data were used for testing. Furthermore, nine
samples of each class were randomly selected from MNIST and USPS datasets for
training, and the remaining data were used for testing. All the processes were
repeated five times, and the final evaluation criteria constituted the average
recognition accuracy and average recall rate of five repeated experiments. The
calculation method of recognition accuracy and recall rate are shown in (11)
and (12). The experiments were implemented using MATLAB R2018a on a computer
with an Intel Core i5-9400 2.90 GHz CPU and Windows 10 operating system.
${\rm recognition\;accuracy}=\frac{T_{1}+...+T_{C}}{n}$ (11)
${\rm recall\;rate}=(\frac{T_{1}}{n_{1}}+...+\frac{T_{C}}{n_{C}})/C$ (12)
where $T_{c},c=1,...,C$ is the count of true samples in $c$th class,
$n_{c},i=1,...,C$ is the count of forecasting samples in $c$th class.
TABLE II: Description of data sets. Datasets | No. of Instances | No. of Features | No. of Classes | No. of Features after PCA
---|---|---|---|---
fac | 2000 | 216 | 10 | -
fou | 2000 | 76 | 10 | -
kar | 2000 | 64 | 10 | -
pix | 2000 | 240 | 10 | -
zer | 2000 | 47 | 10 | -
mor | 2000 | 6 | 10 | -
Yale | 165 | 2000 | 15 | 100
COIL20 | 1440 | 1024 | 20 | 200
MNIST | 2000 | 256 | 10 | 100
USPS | 1800 | 256 | 10 | 100
### IV-B Parameters Setting
The performance of various feature extraction methods was evaluated by setting
certain parameters in advance. First, the more appropriate default parameters
for testing machine learning problems in Adam optimizer comprised
$\alpha=0.001$, $\beta_{1}=0.9$, $\beta_{2}=0.999$, and $\epsilon=10^{−8}$.
For all comparative algorithms, the search range of $k$ was set to
$\\{2,4,6,8,10\\}$, whereas the range of $\sigma$ for CL-LPP, CL-LDA, and CL-
LFDA was set as $\\{0.01,0.1,1,10,100,1000\\}$. In addition, the thermal
parameter was calculated by $t=\|x_{i}-x^{(7)}_{i}\|\|x_{j}-x^{(7)}_{j}\|$,
where $x^{(7)}_{i}$ and $x^{(7)}_{j}$ are the 7th nearest neighbors of $x_{i}$
and $x_{j}$, respectively.
.
## V Conclusion
In this study, we proposed a unified feature extraction framework based on
contrastive learning (CL-UFEF) that is suitable for both unsupervised and
supervised learning. The proposed method defined the positive and negative
pairs from a new perspective and subsequently applied them to the field of
feature extraction. Concretelly, compared with the previous models based on
contrastive learning, CL-UFEF does not need data enhancement, and it
constructs positive and negative pairs based on two contrastive learning
graphs (CLG), which will make the similar samples in the subspace more
clustered. Compared with the traditional graph-based models, CL-UFEF is
suitable for both unsupervised and supervised feature extraction, and it
considers similar and dissimilar samples in unsupervised and supervised
learning based on the CLG.
## References
* [1] W. Zhao, R. Chellappa, P. Phillips, and A. Rosenfeld, “Face recognition: A literature survey,” _ACM COMPUTING SURVEYS_ , vol. 35, no. 4, pp. 399–459, 2003.
* [2] Q. Gao, S. Xu, F. Chen, C. Ding, X. Gao, and Y. Li, “R-1-2-DPCA and Face Recognition,” _IEEE TRANSACTIONS ON CYBERNETICS_ , vol. 49, no. 4, pp. 1212–1223, 2019.
* [3] H. Cai, V. W. Zheng, and K. C.-C. Chang, “A Comprehensive Survey of Graph Embedding: Problems, Techniques, and Applications,” _IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING_ , vol. 30, no. 9, pp. 1616–1637, 2018.
* [4] A. Jain, R. Duin, and J. Mao, “Statistical pattern recognition: A review,” _IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE_ , vol. 22, no. 1, pp. 4–37, 2000.
* [5] Z. Zhang, M. Zhao, and T. W. S. Chow, “Binary- and multi-class group sparse canonical correlation analysis for feature extraction and classification,” _IEEE Trans. on Knowl. and Data Eng._ , vol. 25, no. 10, pp. 2192–2205, 2013\.
* [6] Z. Lai, D. Mo, W. K. Wong, Y. Xu, D. Miao, and D. Zhang, “Robust Discriminant Regression for Feature Extraction,” _IEEE TRANSACTIONS ON CYBERNETICS_ , vol. 48, no. 8, pp. 2472–2484, 2018.
* [7] J. Wen, X. Fang, J. Cui, L. Fei, K. Yan, Y. Chen, and Y. Xu, “Robust Sparse Linear Discriminant Analysis,” _IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY_ , vol. 29, no. 2, pp. 390–403, 2019.
* [8] H. Hu, D.-Z. Feng, and Q.-Y. Chen, “A novel dimensionality reduction method: Similarity order preserving discriminant analysis,” _SIGNAL PROCESSING_ , vol. 182, 2021.
* [9] T. T. Nguyen, M. T. Pham, T. T. Nguyen, T. T. Huynh, V. V. Tong, Q. V. H. Nguyen, and T. T. Quan, “Structural representation learning for network alignment with self-supervised anchor links,” _EXPERT SYSTEMS WITH APPLICATIONS_ , vol. 165, 2021.
* [10] H. Kahng and S. B. Kim, “Self-Supervised Representation Learning for Wafer Bin Map Defect Pattern Classification,” _IEEE TRANSACTIONS ON SEMICONDUCTOR MANUFACTURING_ , vol. 34, no. 1, pp. 74–86, 2021.
* [11] J. Zbontar, L. Jing, I. Misra, Y. LeCun, and S. Deny, “Barlow twins: Self-supervised learning via redundancy reduction,” _arXiv Preprint_ , 2021\.
* [12] A. van den Oord, Y. Li, and O. Vinyals, “Representation learning with contrastive predictive coding,” _arXiv Preprint_ , 2019.
* [13] T. Wang and P. Isola, “Understanding contrastive representation learning through alignment and uniformity on the hypersphere,” _arXiv Preprint_ , 2020.
* [14] M. Caron, I. Misra, J. Mairal, P. Goyal, P. Bojanowski, and A. Joulin, “Unsupervised learning of visual features by contrasting cluster assignments,” _arXiv Preprint_ , 2021.
* [15] J. Li, P. Zhou, C. Xiong, and S. C. H. Hoi, “Prototypical contrastive learning of unsupervised representations,” _arXiv Preprint_ , 2021.
* [16] Y. Tian, D. Krishnan, and P. Isola, “Contrastive multiview coding,” _arXiv Preprint_ , 2020.
* [17] T. Chen, S. Kornblith, M. Norouzi, and G. Hinton, “A simple framework for contrastive learning of visual representations,” _arXiv Preprint_ , 2020\.
* [18] J. Oh and N. Kwak, “Generalized mean for robust principal component analysis,” _PATTERN RECOGNITION_ , vol. 54, pp. 116–127, 2016.
* [19] A. Martinez and A. Kak, “PCA versus LDA,” _IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE_ , vol. 23, no. 2, pp. 228–233, 2001.
* [20] F. Dornaika and A. Khoder, “Linear embedding by joint Robust Discriminant Analysis and Inter-class Sparsity,” _NEURAL NETWORKS_ , vol. 127, pp. 141–159, 2020.
* [21] D. Mo, Z. Lai, and W. Wong, “Locally Joint Sparse Marginal Embedding for Feature Extraction,” _IEEE TRANSACTIONS ON MULTIMEDIA_ , vol. 21, no. 12, pp. 3038–3052, 2019.
* [22] Wai, Keung, Wong, Linlin, Shen, Qinghua, Hu, Yudong, Chen, and Zhihui, “Low-rank linear embedding for image recognition,” _IEEE Transactions on Multimedia_ , vol. 20, no. 12, pp. 3212–3222, 2018.
* [23] Y. Yi, J. Wang, W. Zhou, Y. Fang, J. Kong, and Y. Lu, “Joint graph optimization and projection learning for dimensionality reduction,” _PATTERN RECOGNITION_ , vol. 92, pp. 258–273, 2019.
* [24] P. Chen, L. Jiao, F. Liu, J. Zhao, Z. Zhao, and S. Liu, “Semi-supervised double sparse graphs based discriminant analysis for dimensionality reduction,” _PATTERN RECOGNITION_ , vol. 61, no. SI, pp. 361–378, 2017.
* [25] G. Belous, A. Busch, and Y. Gao, “Dual subspace discriminative projection learning,” _PATTERN RECOGNITION_ , vol. 111, no. 107581, 2021.
* [26] J. Tenenbaum, V. de Silva, and J. Langford, “A global geometric framework for nonlinear dimensionality reduction,” _SCIENCE_ , vol. 290, no. 5500, pp. 2319–2323, 2000.
* [27] M. Belkin and P. Niyogi, “Laplacian eigenmaps for dimensionality reduction and data representation,” _NEURAL COMPUTATION_ , vol. 15, no. 6, pp. 1373–1396, 2003.
* [28] S. Roweis and L. Saul, “Nonlinear dimensionality reduction by locally linear embedding,” _Science_ , vol. 290, pp. 2323–2326, 2000.
* [29] X. He, “Locality preserving projections,” _Advances in Neural Information Processing Systems_ , vol. 16, no. 1, pp. 186–197, 2003.
* [30] X. He, C. Deng, S. Yan, and H. J. Zhang, “Neighborhood preserving embedding,” in _Tenth IEEE International Conference on Computer Vision_ , 2005.
* [31] L. Qiao, S. Chen, and X. Tan, “Sparsity preserving projections with applications to face recognition,” _PATTERN RECOGNITION_ , vol. 43, no. 1, pp. 331–341, 2010.
* [32] Wang, Zhenyu, Yang, Wankou, Sun, and Changyin, “A collaborative representation based projections method for feature extraction,” _Pattern Recognition_ , vol. 48, no. 1, pp. 20–27, 2015.
* [33] Y. Zhang, M. Xiang, and B. Yang, “Low-rank preserving embedding,” _PATTERN RECOGNITION_ , vol. 70, pp. 112–125, 2017.
* [34] P. Belhumeur, J. Hespanha, and D. Kriegman, “Eigenfaces vs. Fisherfaces: Recognition using class specific linear projection,” _IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE_ , vol. 19, no. 7, pp. 711–720, JUL 1997.
* [35] M. Sugiyama, “Dimensionality reduction of multimodal labeled data by local fisher discriminant analysis,” _JOURNAL OF MACHINE LEARNING RESEARCH_ , vol. 8, pp. 1027–1061, 2007.
* [36] S. Yan, D. Xu, B. Zhang, H.-J. Zhang, Q. Yang, and S. Lin, “Graph embedding and extensions: A general framework for dimensionality reduction,” _IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE_ , vol. 29, no. 1, pp. 40–51, 2007.
* [37] Z. Huang, H. Zhu, J. T. Zhou, and X. Peng, “Multiple Marginal Fisher Analysis,” _IEEE TRANSACTIONS ON INDUSTRIAL ELECTRONICS_ , vol. 66, no. 12, pp. 9798–9807, 2019.
* [38] Y. Ren, Z. Wang, Y. Chen, and W. Zhao, “Sparsity Preserving Discriminant Projections with Applications to Face Recognition,” _MATHEMATICAL PROBLEMS IN ENGINEERING_ , vol. 2016, pp. 1–12, 2016.
* [39] K. He, H. Fan, Y. Wu, S. Xie, and R. Girshick, “Momentum contrast for unsupervised visual representation learning,” in _2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)_ , 2020, pp. 9726–9735.
* [40] D. P. Kingma and J. Ba, “Adam: A method for stochastic optimization,” _arXiv Preprint_ , 2017.
|
# Ten Simple Rules for Attending Your First Conference
Elizabeth Leininger Division of Natural Sciences, New College of Florida, USA
Kelly Shaw Computer Science Department, Williams College, USA Niema Moshiri
Computer Science & Engineering Department, University of California San Diego,
USA Kelly Neiles Department of Chemistry & Biochemistry, St. Mary’s College
of Maryland, USA Getiria Onsongo Mathematics, Statistics & Computer Science
Department, Macalester College, USA Anna Ritz Corresponding Author:
<EMAIL_ADDRESS>Biology Department, Reed College, USA
## Introduction
Conferences are a mainstay of most scientific disciplines, where scientists of
all career stages come together to share cutting-edge ideas and approaches. If
you do research, chances are you will attend one or more of these meetings in
your career. Conferences are a microcosm of their discipline, and while
conferences offer different perspectives in different disciplines, they all
offer experiences that range from a casual chat waiting in line for coffee to
watching someone present their groundbreaking, hot-off-the-press research.
Here, we share tips for trainees and their mentors. Our recommendations are
based on our experiences of attending conferences and mentoring students to
improve their conference experiences. As you head to your first scientific
conference, these rules will help you navigate the conference environment and
make the most of your experience.
### In-person Conferences:
Scientific meetings have historically been in person, where all attendees
travel to one location to share and learn about recent work in the field. In-
person conferences provide extensive opportunities to meet and interact with
other researchers. Our rules include considerations for all parts of attending
an in-person meeting, including deciding whether to attend and present
research, navigating travel and the conference schedule, and debriefing after
the conference.
### Virtual Conferences:
During the COVID-19 pandemic, many conferences have been and continue to be
offered remotely [1, 2, 3]. Even before 2020, some conferences had already
moved to a virtual format in response to climate change [4] and/or travel bans
[5], noting the environmental, financial, and geopolitical challenges posed by
in-person meetings. Virtual conferences are more inclusive than their in-
person counterparts in many ways: more people can attend them from all over
the world, virtual meetings are less costly and easier to organize, and
attendees can participate from the comfort of their homes [3, 6, 7, 8]. Given
that virtual conferences are becoming more commonplace [9, 10, 3, 7, 8], we
include considerations for each rule when attending one.
### For Mentors:
Academic mentorship takes many forms, including advising student research,
teaching students in courses, and serving as a department chair or organizer
for student activities. In many fields, graduate students mentored by a
principal investigator (PI) are expected to disseminate their research at
scientific meetings. Conference attendance is a valuable experience for
undergraduate students, whether they are presenting research [11] or just
beginning to learn about the field [12, 13]. As mentors, you might be working
with undergraduates, graduate students, or other staff who may be new to
conferences. We offer guidance on ways for you to help mentees navigate the
subtleties and assumptions of your field.
We have also developed a web portal (SI Web Portal) which contains far more
information about these rules, tables of professional societies and
conferences in different disciplines, and other resources that may come in
handy for first-time conference attendees and their mentors. We encourage any
reader to use, adapt, and contribute to these materials.
## Rule #1: Select a conference that aligns with your goals
Why go to a conference, anyway? When you are deciding whether to attend a
conference, consider the experience and how it will help you develop
intellectually and professionally. There may be more than one answer!
Conferences are a “meeting of minds” – a place for researchers to gather,
present their research, give feedback on others’ research, engage in
professional development, and network with one another.
Conferences differ greatly in size and scope. While some conferences provide a
large overview of an entire academic discipline with tens of thousands of
attendees, other conferences focus on specific subdisciplines in an informal
and personalized setting. Smaller conferences are excellent venues for
networking, sharing your data informally, etc. Some conferences are national
or international in scope, while others have a regional focus. All of these
variables impact factors such as the conference location, scope of presented
research, the cost of attendance, and more.
If you are planning to attend a conference, you might also plan to present
research. You and the others who have worked on the research would make this
decision before the conference, often many months prior, and you typically
submit materials to be considered for presenting. Consult with your mentor and
colleagues in your research group about the benefits of and opportunities for
presenting your work. In some fields, such as biology or chemistry, you might
submit a short abstract for a poster presentation or a talk. In other fields,
like computer science, you might submit a full paper that is peer-reviewed and
published upon acceptance. Conventions vary by discipline; make sure to
discuss norms in your field with your research mentor and all other authors of
the work, who need to approve any submission.
In addition to contributed talks and posters, conferences provide formal and
informal opportunities for professional development and networking (See Rule
#7). Review the professional development opportunities at the conference you
are attending, and participate in opportunities that sound interesting and
align with your goals.
### Virtual Conferences:
Nearly every annual meeting in 2020 and 2021 has shifted to a virtual format
due to the COVID-19 pandemic. Many traditionally in-person conferences aim to
retain multiple goals of their conferences – a venue for presenting the latest
work, networking, and professional development – in a virtual format.
Alongside these legacy meetings, the virtual format has also spurred new
grassroots virtual conferences [7]. For example, conferences such as
Neuromatch and Black in Neuro are newer additions to Neuroscience conference
offerings that originated as virtual experiences. Some virtual conferences
offer more focused research areas than the larger meetings, since they are
generally easier to organize and manage. They also have fewer barriers for
attendance (allowing researchers from all over the world to attend), they are
less wasteful (in terms of money and environmental impact), and they are
overall more equitable than in-person conferences [2, 3]. Consider all of
these aspects when deciding on a conference to attend.
### For Mentors:
Empower students to attend conferences. Suggest conferences for your students
that match their present goals/professional development needs and explain how
they could benefit from attending these meetings. Be aware of socioeconomic
privilege and implicit bias when recommending opportunities to trainees (e.g.,
recommend equally and equitably, help them find funding, and connect them to
your professional network at conferences).
## Rule #2: Find others to foot the bill
Organizing and running conferences cost money, and attendees usually pay to
attend these events. While registration costs are often low for virtual
conferences, in-person conferences can get expensive quickly due to additional
costs associated with travel, room and food expenses, poster printing fees,
childcare, appropriate clothing, etc. Therefore, it’s important to have a
funding strategy to attend an in-person conference. Funding sources are often
available to offset some (or even all) of your costs, and knowing where they
exist early in the process will help you take advantage of these
opportunities.
First, some conferences offer discounted registration or travel grants for
students, attendees from countries with traditionally fewer economic
resources, etc. Some travel grants may include additional support for trainees
with disabilities or trainees with small children. Other societies are
establishing separate child care grants for trainees with families. Often
these funding opportunities are made possible by professional societies,
grants from organizations such as the National Science Foundation, or by
sponsorship funds provided by industrial partners. An in-person conference
might have volunteer opportunities where you are compensated with a waived
registration fee or housing for working at the conference. Second, many
academic institutions have funds intended to support student travel. Principal
investigators may have a budget to support their students’ attendance at a
conference. Research scholarships and/or travel awards may be available at the
departmental or college level. Sometimes these opportunities are not widely
advertised so it is always good to ask around. If you do not have a research
mentor, asking the department chair may be appropriate to help you navigate
your search. Another place to look is student clubs and organizations, which
may have funding to support students presenting research in a field relevant
to the group’s goals. Third, national societies, often the same societies that
are organizing the conference, may provide support for student travel at the
society, division, and regional chapter level. Some of these opportunities are
dependent on student demographics such as being first-generation, a student
from a historically underrepresented population, or a woman in the field.
Information about these awards is generally not in a central location, so
visit each level’s website to search for potential sources (the web portal
includes a list of societies in STEM, many of which provide travel funds).
An important note: many of these awards and funding opportunities for in-
person conferences will often require that you front the money and then get
reimbursed after the conference. This means you would need to put things like
airfare, hotel reservations, and the like on your own credit card and then
submit receipts for reimbursement. If this presents a hardship for you, reach
out to your department for help. This is not the first time, nor will it be
the last, that your department has encountered this problem so you should not
hesitate to ask for assistance in navigating this process. Keep track of all
itemized receipts for the conference, and ask the funding provider for details
about reimbursement.
### Virtual Conferences:
Virtual conferences generally cost less than their in-person counterparts [3],
and many virtual conferences have reduced or waived the student fees
associated with conference registration. If paying for a virtual conference
registration is a hardship, there is no harm in asking the conference
organizer if there is a way to attend.
### For Mentors:
Funding support deadlines for conferences are often so far in advance that
students, especially undergraduate students, may miss these opportunities.
Help students identify and meet these deadlines. If your institution has
funding sources for student travel but those sources are only available to
those students who ask (or whose mentor’s ask), it might be time to suggest a
conversation about equity of those opportunities. This policy may
disproportionately disenfranchise students from historically underrepresented
populations who are not as well versed in the hidden curriculum of academia
and thus do not know to ask for these opportunities [14].
## Rule #3: Know your logistics
It is important to be aware of the Who, What, When, Where logistics (see Rule
#1 for Why) to ensure a smooth conference-going experience. Such logistics
will vary depending on if you choose to attend an in-person conference or a
virtual conference, or even depending on the format of virtual conference you
choose to attend.
Who do you know will be at the conference? Housing and transportation costs
for conferences can be expensive, complicated, and/or both. Consider traveling
together with people you know and sharing housing costs by rooming with
labmates, acquaintances, etc. Some conferences and field-specific societies
offer roommate-matching services for those seeking conference roommates. You
may also be traveling with family or dependents; check to see if the meeting
has programming or resources for companions, childcare and lactation
facilities, etc. as needed.
What will you need to bring to the conference? It is usually preferable to
pack light, and pack for the situations and weather you will encounter (see
Rule #4). Double-check baggage policies for your mode of transportation, as
policies vary drastically by carrier and fare. If you are presenting, make
sure that your presentation materials travel safely with you. If you are
giving a talk, make a few different backups of talk slides and media (store on
a thumbdrive or in the cloud) in case your computer is lost, stolen, or
broken. Posters can be transported in a poster tube or even printed on fabric
to fold in your luggage. The poster can also be shipped to the venue or
printed on-site if it is challenging to travel with it.
When do you need to make key conference travel decisions? Chronologically,
these include: writing and submitting a research abstract or paper, having
your abstract or paper accepted, registering for the conference, making travel
and housing arrangements, and actually traveling to the conference and
presenting your work. Because abstract and paper submission deadlines can fall
several months ahead of the meeting date, it is important to be aware of
abstract deadlines for conferences in your field. When scheduling travel, make
sure that you arrive before, and leave after, your scheduled presentation
date/time. If your travel has multiple legs of transportation, give yourself
enough time to make each connection. International travel may require visas,
and the lead time and expense varies greatly depending on the origin and
destination countries. Consider visas and passports in both your timeline and
your budget, and talk with your mentor about whether an international
conference is feasible. Conferences with an international scope will usually
communicate information about visas and lead time.
Where is the conference, and where will you stay while attending? Depending on
the scope of the conference (see Rule #1), housing may be easily obtainable at
the conference venue, or a mad scramble between all 30,000 attendees to find a
room near the convention center. Conferences will typically contract
guaranteed rates at nearby hotels; heed deadlines and sign up early if you
possibly can. Be aware of ADA accessibility when booking if you require rooms
with accessible features. Acquaint yourself with the transportation options to
and within the conference host city. Again, conferences will usually send
information and tips regarding transportation, housing, and the like.
### Virtual Conferences:
The Who, What, When, Where logistics of virtual conferences are also extremely
helpful to determine before the meeting.
Identify Who you know at the meeting, and make plans to meet with them
virtually during a break. Virtual conference platforms allow you to browse the
list of presenters and/or conference attendees (see Rule #7). Know What you
are expected to do in different sessions; if you plan to attend a virtual
panel, for example, think about some questions you might want to ask the
panelists.
Familiarize yourself with When the conference activities are, and if any
action is required for accessing the conference materials. The format of
virtual conferences often differs from the time-limited experience of in-
person conferences. For example, they may run longer to allow for a different
pace of engagement, and they may have components that are synchronous (e.g.
talks scheduled live for specific times) or asynchronous (pre-recorded talks
for on-demand viewing). Some virtual conferences are offered completely
asynchronously, making them more inclusive for attendees with care
responsibilities or in time zones across the world [8]. Make sure to note
conference dates and time zones of scheduled synchronous sessions as
appropriate. If presenting at a virtual conference, pay very close attention
to announcements regarding presentation logistics; you may need to upload your
presentation before the formal start of the conference and check in at various
times to field questions.
Finally, be organized by knowing Where the conference activities are
happening, often through a web interface or a collection of Zoom links. If you
are unsure about the digital platform or need extra assistance, contact the
organizer who will be able to help you get help before the conference. If the
virtual conference requires high internet bandwidth, make sure your setup at
home allows for participation.
### For Mentors:
If you are also attending the meeting, make sure that your students are at the
same point logistically as you (register for the conference together, book
flights together, etc.) If you are funding the students, putting as much of
the costs on your business card up-front reduces their financial burden. If
you are not attending the meeting, reach out to your colleagues who will be
there to help arrange housing and other connections for your students.
## Rule #4: Prepare for the environment
A good rule of thumb is to dress as if you might meet your future employer at
a conference (which you very well might do!). At the same time, wear something
you are comfortable in. Different meetings and disciplines may have a variety
of dress code standards. Attendees who are presenting may wear more formal
clothes, but many people will dress as they do when they are at their home
institutions.
When you are packing for an in-person conference, check the weather before you
go, since you may be outdoors (e.g., to reach the conference from the airport,
to walk to find dinner, or participate in planned conference outings). Between
moving around the conference venue, seeking out meals, and milling about
poster sessions, you will spend a lot of time being active. Wear shoes and
clothing that allow you to comfortably do all these things. If you have any
questions about ADA accessibility at the conference, contact the conference
venue (e.g., if it is a conference hall attached to a hotel) or contact the
conference organizer (the conference website will contain this information).
Some conferences have an app (See Rule #6) which may have a Q&A Forum.
Finally, conference days can be long - bring a water bottle to stay hydrated,
a light snack to sustain you between meals, and a comfortable bag to keep
belongings.
### Virtual Conferences:
You will likely be in the comfort of your own home during the conference.
Realize it is easy to sit and stare at the screen for hours on end. Be sure to
take breaks, stay hydrated, stretch, and focus your eyes away from your screen
every so often. It is also a good idea to test your video, audio and
background before the conference. Find a room with good lighting and make sure
your background is not distracting. Some video conferencing platforms have
video settings that will enhance your image if the lighting is low - try them
out. Conference days can be extremely long in virtual platforms – sometimes
longer than typical in-person schedules – so be sure to pace yourself (see
Rule #6).
### For Mentors:
Talk to your students about the level of formality at the meeting. If you know
that an in-person conference will have outdoor events, mention that. If you
are comfortable with sharing, tell your students the types of things you pack
for conferences.
## Rule #5: Learn how to take in the science
Science communication happens in a variety of ways at a conference. There are
almost always formal and informal mechanisms for learning about and presenting
science. Even within formal mechanisms, there is a range of how attendees
communicate their research. Some speakers are invited to give keynote
presentations, which often provide a perspective of the field or cover a broad
range of projects. Other speakers are selected to give technical
presentations, based on submitted abstracts or full papers. Both keynotes and
technical presentations provide time for asking questions at the end, and
attendees may line up in front of a microphone or raise their hand depending
on the number of people. Finally, poster presentations offer a more
interactive way to talk with researchers, where attendees walk up and chat
with authors standing in front of their posters.
But wait, there’s (often) more! Many conferences offer tutorials, workshops,
or special sessions that are held either right before or right after the main
conference, usually at the same location. These topics may be even more
narrowly focused than those at the conference, or involve emerging concepts in
the discipline. There also may be forums run by graduate/undergraduate student
societies, designed for students. Note, though, some of these may require
registration beyond the main conference.
So, how do you actually “take in the science?” First, know that it is often
impossible to see everything - the conference may have multiple conflicting
sessions, the days are quite long, and you need to take time to rest and
recharge. As you make a conference plan to prioritize what you want to see
(see Rule #6), let the conference program be your guide. Find a way that works
for you to take notes on what you learn throughout the conference. We, the
authors, have developed note-taking strategies that involve carrying around a
notebook, using Google Docs, writing in the conference program, taking notes
on our smartphones, making sketchnotes, even sending ourselves emails. In
addition to the science, take notes about whom you meet - you never know if
you will meet them again or if you might want to connect with them after the
conference. Find something that works for you. Some conferences may have rules
about taking digital photos or videos of talks and posters, so be sure to
check beforehand if you want to do either.
At in-person conferences, there are a plethora of informal ways to interact
with other researchers and learn about their science. Researchers appear to
have a common need for caffeine, and coffee breaks are a staple of every
conference. If you are not a coffee drinker, there are often other warm
beverages and snacks on hand. Conferences may also offer breakfast or lunch,
and each one is in an invitation to meet someone new (see Rule #7). Many
conferences have a “big-ticket” event - an evening banquet, an outing to a
tourist attraction, a social hosted by a sponsor, etc. If the conference has a
big-ticket event, it is not one to miss! There may also be a career fair,
special meetings for first-timers, or meet-ups for special interest or
affinity groups. Take advantage of these to find other attendees with similar
interests, backgrounds, and experiences (see Rule #8).
### Virtual Conferences:
Virtual conferences have the enormous benefit in that the talks are almost
always recorded (even when they are presented synchronously to all attendees).
Further, sometimes conferences will rebroadcast talks 12 hours later for
attendees in different time zones. The material from conferences often remains
on the virtual platform after the conference concludes, allowing attendees to
watch and rewatch talks at their leisure. In addition to the formal
presentations, virtual conferences have developed strategies for the
interactive conference portions (e.g., poster presentations, breaks, and
social events). Some digital platforms offer mechanisms to ask questions and
continue discussions after a talk has concluded. Cold Spring Harbor meetings,
for example, have a Slack workspace with channels for each session to continue
Q&A and networking. The chat functionality also allows researchers to talk to
poster presenters throughout the entire meeting and send short messages about
appreciating people’s work in a way that cannot happen at an in-person
conference. Take advantage of the interactive parts of the conference that can
help support your learning and professional development.
### For Mentors:
Share the events you plan to attend with your students, and give advice about
what talks or events will be most useful for them. Encourage your students to
ask questions, either in the Q&A after a talk or less formally at the poster
session. If you have students or colleagues who are presenting, encourage your
students to attend those talks.
## Rule #6: Make a conference strategy
The idea of a conference - days of uninterrupted learning about fascinating
ideas and exchanging insights with other folks excited about the topics you
are passionate about - might sound like a dream come true! The reality is that
conferences can be exhausting if you do not have a plan for selectively
attending activities that will provide you the most benefit and for practicing
self-care. This is especially important for neurodivergent attendees who may
get overwhelmed by the amount of information conveyed during a conference.
It’s good to fashion a draft plan several weeks before the conference in which
you prioritize the events you want to attend. Having a written (or app-built)
plan with scheduled breaks gives you a solid framework that you can tweak on
the fly as new opportunities appear. If you intend to participate in pre- or
post-conference events, it is important to factor in extra self-care to
compensate for the increasing duration.
Before making a conference strategy, gather resources to help you make a plan.
Conferences usually have a website with lots of information, and often release
a schedule overview as the keynote and technical talks and posters are
finalized. The conference may make use of an app, where attendees can find the
schedule of events and connect with others, and/or you may receive a
conference program. These are all great resources to have on hand as you make
your strategy, and it is useful to skim the conference program before the
meeting begins.
How do you prioritize what to attend? First, it is good to attend keynote and
panel sessions as they provide perspective into the wider concerns of your
field and often are forward looking to emerging challenges. Second, definitely
attend technical presentations related to your specific area of focus in order
to know what research is being done and become part of that community of
researchers. Reading papers or watching videos in advance and thinking of what
questions you might like to ask about the work is a great way to prepare so
that you can contribute to the discussion in a positive way. Third, the poster
sessions are often short, so make sure you know which posters you want to
visit while the presenter is there. Fourth, if the conference offers any
first-time or new attendee events, plan on attending those as you will make
some connections with other attendees that will make the conference more
enjoyable and less lonely. Finally, attending the networking events (see Rule
#7) helps you get to know your colleagues as individuals on a personal level
(not all discussions are about the research) and also exchanging your research
ideas.
Working self-care into your plan is essential - but how? First, if you have a
daily ritual such as exercising or going for a walk, sustain that ritual
during conference days and decide when you will fit this activity in. If you
are at an in-person conference, find out if the hotel has the equipment you
need for exercising; many downtown gyms or community centers sell day passes.
Second, recognize that your brain is going to need breaks between talks. If
the conference is not in your native language, your brain may be working
overtime to process the science. Determine time slots when the presentations
are not of particular interest to you and plan on taking a walk, getting a
coffee, or doing some other activity that will help you recharge. You will see
lots of experienced conference attendees disappear for stretches of time for
exactly this purpose. For many of us, traveling to in-person conferences is a
way to experience new parts of the country or the world. It is fine to take a
short break to experience local destinations such as a museum, park, cultural
attraction, etc.
### Virtual Conferences:
When attending virtually, it may be hard to devote entire days to attending
the conference. Prioritize attending the synchronous events such as the
keynotes and panel discussions, especially if they are not recorded for later
viewing. Take advantage of videos posted for areas you are interested in and
watch those ahead of time so you can choose to attend the sessions where
you’re interested in the question and answer portion.
### For Mentors:
Encourage your students to come up with a written plan and discuss it with
them before the conference. Suggest specific presentations that would be good
for them to attend and explain to them the importance of strategic downtime
and self-care.
## Rule #7: Make new friends but keep the old; be ready to communicate
Conferences offer a great opportunity to exchange ideas, network, and
potentially form collaborations with other researchers. Networking
opportunities can be organized or spontaneous. Organized networking events may
include events such as socials, affinity group meetings, or mentorship
opportunities that pair newcomers with established researchers in the field.
Many researchers also make efforts to meet with current or prospective
collaborators at meetings. In addition to organized events, lots of networking
happens spontaneously – waiting in line for coffee or tea, at a poster
presentation, etc. Take advantage of the opportunity to meet lots of people
who are interested in sharing their science and forming new professional
connections.
Pre-plan a few people you would like to meet at the conference. A good way to
find people is to look at the conference program, which contains presenter
names and sometimes includes a full list of attendees (See Rule #6). Start
with the research area you are interested in. Typically, the first author
presents the work and will most likely be at the conference. Read their work
before the conferences and prepare questions to start a conversation. If they
have a research website, they will likely have a picture. Find out what they
look like in advance. In-person conference attendees usually wear name badges
which helps if you’re trying to meet someone whom you’ve never met in person.
If they are giving a talk, try to attend; following up with the speaker after
their talk is a good way to strike up a conversation with them. If you did not
get a chance to talk with everyone you wanted to, you can always follow up
with them over email (see Rule #10). You can use the conference to make it not
feel like a cold email - tell them you attended their talk, saw their poster,
etc., and ask any questions. Researchers delight in being asked informed and
insightful questions about their work.
In-person conferences have many opportunities for informal networking, but it
still may feel awkward to strike up a conversation. If you are coming from a
lab or institution with colleagues also attending the conference, agree to
introduce one another to whomever you are speaking with. Also, most
conferences have events just before the official program starts such as
breakfast or socializing events in the evening. These are a great way to meet
people in a more relaxed environment. If you are feeling overwhelmed or out of
place, which is normal (see Rule #8), it is okay to spend time with people you
already know.
### Virtual Conferences:
Introducing yourself and having a one-on-one conversation can also be
challenging in virtual environments. Networking still happens at virtual
conferences, but the interactions are necessarily more intentional than
striking up a conversation during a coffee break. Many virtual conference
platforms offer ways to interact, and conferences organize social events to
foster this networking. If the online conference platform has social events or
research discussion groups, join in. Leave your camera on, if comfortable, and
participate in the chat, if appropriate. Some virtual conference platforms
have messaging capabilities, which you can use to have conversations with
individuals or small groups.
### For Mentors:
If you will also be attending an in-person conference, be intentional about
introducing your students to your professional network. Invite them to join a
group meal you plan on attending with people you already know and introduce
them to colleagues. If you are talking to someone working in an area your
student is interested in, and you see your student close by, be sure to
introduce them. Pointing students towards posters that are relevant and
interesting to them is also a great way to help attendees begin conversations
with researchers.
## Rule #8: Prepare to (safely) get out of your comfort zone
For most of us, meeting new people or joining a new community can be nerve-
wracking or intimidating, even when you have a lot in common with people in
the community. Things can be even more challenging if you come from a
historically underrepresented community and do not see yourself represented at
the conference. Academia, and by extension its conferences, is a traditional
and elite institution whose diversity, or lack thereof, can make it a less
than welcoming place at times. Imbalance in representation is gradually being
acknowledged by assessing the demographics of invited speakers and society
awardees [2, 15, 16, 17]. If you are feeling out of place, know you are not
alone, and that it can take some time to feel comfortable. Try to meet new
people and make new friends. You will likely see the same people if you attend
the conference again. And, the more conferences you attend, the easier it
gets. Some conferences host affinity group events, which is a good opportunity
to meet and network with other attendees in a safe and welcoming space. Our
web portal hosts a crowd-sourced list of affinity groups.
Come up with a plan for how to talk about your interests and research. Have a
90 second elevator pitch prepared for your initial introduction and be
prepared to switch into a slightly extended version of that pitch if the other
person expresses interest. Because it is important to learn about the other
person (people love good listeners), make sure to ask them about their
research, including asking follow-up questions after their initial answers to
show your interest. If the exchange goes well, create a note of their names
and affiliations for use later. Also, if you see them later in the conference,
acknowledge their existence with a smile or head nod.
### Virtual Conferences:
An advantage of virtual conferences is you will be in your own space. Take
advantage of this opportunity in a virtual setting to make yourself
comfortable (See Rule #4). It is okay to turn off your video if you need to
step away for a few minutes.
### For Mentors:
It can be helpful for mentees to know they are not alone in feeling out of
place. Normalize the experience of feeling nervous and out of place. Let them
know conferences can be awkward and that is okay. If you are mentoring
students from underrepresented groups, being culturally competent will help
you better support them. Cultural competency is the ability to work
effectively with people from different cultural backgrounds. In examining the
role of cultural competence in a biology classroom, Tanner and Allen highlight
the importance of cultural competence in creating an inclusive and welcoming
environment for students from underrepresented groups [18]. A good place to
start is finding out if your institution offers cultural competency training.
If it does not, the Association of American Colleges & Universities offers
cultural competency resources [19].
Finally, mentor your students – particularly marginalized students who may
find academia and conferences less-than-welcoming – through the conference
experience and their overall professional development with “compassion,
advocacy, and support” [20]. Be aware that imposter syndrome stems from
systematic bias, and calling out imposter syndrome sometimes can have the
opposite effect for mentees [21]. Become involved in helping meetings achieve
gender and racial balance in their selection of awardees and keynote speaker
invitations [15]. See our web portal for more resources on cultural competency
and improving representation at conferences.
## Rule #9: Take charge of your social interactions
Conferences bring together many people from all over the world, and navigating
a complex professional-yet-social environment can be challenging. All members
of a scientific community have a responsibility to help make a welcoming
environment and should in turn feel welcomed. As a conference attendee, you
are a member of the scientific community. At the same time, we acknowledge
that power imbalances may be prevalent at the conference, often reflective of
career stage [22]. Many societies have established codes of conduct to which
attendees must adhere [23], and the conference should include a contact if you
witness or are subjected to troubling behavior. Above all, you should never
feel unsafe or pressured to participate by anyone.
Before you travel to in-person conferences, acquaint yourself with the code of
conduct and have a contact person (e.g., your mentor, someone from your
institution, or a friend) that you can reach out to. You may need to prepare
yourself to adjust to the cultural norms of the conference location, if it is
notably different from your academic environment. While there are social
opportunities at in-person conferences, remember that these are professional
events.
### Virtual Conferences:
Social situations are different in a virtual platform. There may be fewer
awkward conversations, but online interaction can pose its own problems. For
example, you may see unwelcome visitors hack into presentations or observe
uncourteous behavior from other attendees. You may also find some features
(like the chat function or unmuted attendees) distracting during a
presentation. When participating in virtual conferences, follow any
instructions session chairs give about how to interact during presentations.
As a general rule, make sure you follow online etiquette (such as muting
yourself on Zoom unless speaking). Attendees should observe the same codes of
conduct in a virtual setting that they would observe in an in-person
conference.
### For Mentors:
It is important that students have your contact information (or the
information of someone trusted who will be at the meeting). Stay up to date on
your organization’s code of conduct. If your society has not yet established a
code of conduct (check on our web portal), encourage the leaders to do so and
support their efforts [23].
## Rule #10: Tie up the loose ends after the conference
After attending the conference you will likely want to come home and collapse
from all of the excitement, but wait…you have a few more things to do before
you are done. These loose ends pertain to both in-person and virtual
conferences. First and foremost, update any notes you have about the
conference itself before you forget. Conferences can be a fast-paced blur, so
make sure to record any feedback you get on your work at the conference, so
that you have it when you are ready to make improvements. Updating your resume
and/or curriculum vitae (CV) is another important step you should complete as
quickly after the conference as possible. In addition to any posters or
presentations you gave, you should also add any awards you received including
travel funding awards. Depending on your career stage, you might also include
other events or support you received - check with a mentor about what is
appropriate.
Next you should follow up with people you met at the conference. This helps to
solidify the relationships you began at the conference. Email people you are
interested in speaking with again and ask for an opportunity to meet. If you
spoke with companies or potential job seekers, follow up with an email
containing your resume and statement of interest. When you attend a conference
you meet so many people that it is hard to remember everyone. By reaching out
with a simple note or a LinkedIn invitation, you will be helping people to
remember you which can lead to future collaborations and/or job opportunities!
If you received a travel award (for an in-person conference, for example) that
requires reimbursement you will need to carefully follow the instructions
and/or rules of your funding agency. Reimbursements can take quite a while to
process so the faster you get it done after the conference, the better. Pay
special attention to rules regarding the submission of receipts (itemized
receipts are required in some cases) and deadlines for submission (some
agencies require reimbursement documents be submitted within a month of the
conference ending). Make a note to yourself to follow up on reimbursement; if
you do not hear back about your reimbursement within two weeks, go ahead and
reach out to inquire about your paperwork (you can also ask your mentor if it
has been longer than you were expecting).
Finally, you should take time to write thank you notes or emails to anyone who
supported your travel to the conference whether that be financially, in
conference preparation, or in your research. This not only helps you to
further strengthen relationships with those people, but also helps future
students receive these awards by leaving a good impression with the awarders.
### For Mentors:
Communicate your institution’s rules and requirements about reimbursements to
your trainees, and tell your trainees who to contact for reimbursement
paperwork (e.g. a department administrator). We suggest holding a paperwork
session after the conference so that all travelers can come together and fill
out paperwork, print and/or make copies of necessary materials, and ask
questions of each other and you. If multiple travelers attended the same
conference, then mentors can also submit all of this paperwork together to
increase the chances of its timely completion. We sometimes also bring a giant
box of thank you notes so that students can easily grab one to write a note of
thanks to those who helped them along the way. If students run into
bureaucratic difficulties with paperwork, help them get the issue resolved.
## Conclusion
Scientific conferences are an amazing intellectual and professional
opportunity, though attending one for the first time may be overwhelming. We
hope that first-time conference attendees will feel more empowered and
prepared to get the most out of their experience, be it virtual or in-person,
with knowledge of these “unwritten rules.” Happy travels!
## Supporting information
### SI Web Portal
The web portal includes additional details on tips for first time conference
attendees and their mentors, tables of societies and scientific conferences,
and affinity group organizations.
https://sites.google.com/macalester.edu/simplerules/home.
## Acknowledgments
We thank Erik Zornik and Andrew Bray for their early contributions to the
conference advice template. This work is supported by the National Science
Foundation (DBI-1750981 to AR).
## References
* [1] Olena A. COVID-19 Ushers in the Future of Conferences. The Scientist. 2020.
* [2] Sarabipour S. Research Culture: Virtual conferences raise standards for accessibility and interactions. Elife. 2020;9:e62668.
* [3] Sarabipour S, Khan A, Seah YFS, Mwakilili AD, Mumoki FN, Sáez PJ, et al. Changing scientific meetings for the better. Nature Human Behaviour. 2021; p. 1–5.
* [4] Ligozat AL, Névéol A, Daly B, Frenoux E. Ten simple rules to make your research more sustainable. PLoS Computational Biology. 2020;16(9):e1008148.
* [5] Reardon S. How the latest US travel ban could affect science. Nature News. 2017;550(7674):17.
* [6] Salomon D, Feldman MF. The future of conferences, today: Are virtual conferences a viable supplement to “live” conferences? EMBO reports. 2020;21(7):e50883.
* [7] Rich S, Diaconescu AO, Griffiths JD, Lankarany M. Ten simple rules for creating a brand-new virtual academic meeting (even amid a pandemic). PLOS Computational Biology. 2020;16(12):1–9. doi:10.1371/journal.pcbi.1008485.
* [8] Arnal A, Epifanio I, Gregori P, Martínez V. Ten Simple Rules for organizing a non-real-time web conference. PLOS Computational Biology. 2020;16(3):1–13. doi:10.1371/journal.pcbi.1007667.
* [9] Gichora NN, Fatumo SA, Ngara MV, Chelbat N, Ramdayal K, Opap KB, et al. Ten simple rules for organizing a virtual conference-anywhere. PLoS Computational Biology. 2010;6(2):e1000650.
* [10] Lortie CJ. Online conferences for better learning. Ecology and evolution. 2020;10(22):12442–12449.
* [11] Mabrouk PA. Survey study investigating the significance of conference participation to undergraduate research students. Journal of Chemical Education. 2009;86(11):1335.
* [12] Davis J, Alvarado C. Supporting undergraduates to make the most of conferences. ACM Inroads. 2017;8(3):32–35.
* [13] Wright HM, Tamer NB. Can sending first and second year computing students to technical conferences help retention? In: Proceedings of the 50th ACM Technical Symposium on Computer Science Education; 2019. p. 56–62.
* [14] Smith B. Mentoring at-risk students through the hidden curriculum of higher education. Lexington Books; 2013.
* [15] Martin JL. Ten simple rules to achieve conference speaker gender balance. PLoS Computational Biology. 2014;10(11):e1003903.
* [16] Le TT, Himmelstein DS, Anderson AAH, Gazzara MR, Greene CS. Analysis of ISCB honorees and keynotes reveals disparities. bioRxiv. 2020;.
* [17] Shishkova E, Kwiecien NW, Hebert AS, Westphall MS, Prenni JE, Coon JJ. Gender diversity in a STEM subfield – analysis of a large scientific society and its annual conferences. Journal of The American Society for Mass Spectrometry. 2017;28(12):2523–2531.
* [18] Tanner K, Allen D. Cultural competence in the college biology classroom. CBE – Life Sciences Education. 2007;6(4):251–258.
* [19] Association of American Colleges & Universities. Cultural Competency Resources: Teaching to Increase Diversity and Equity in STEM (TIDES); 2021. Available from: https://www.aacu.org/tides/cultural-competency.
* [20] Singleton KS, Tesfaye R, Dominguez EN, Dukes AJ. An open letter to past, current and future mentors of Black neuroscientists. Nature Reviews Neuroscience. 2020; p. 1–2.
* [21] Tulshyan R, Burey JA. Stop Telling Women They Have Imposter Syndrome. Harvard Business Review. 2021;.
* [22] Jackson L. The smiling philosopher: Emotional labor, gender, and harassment in conference spaces. Educational Philosophy and Theory. 2019;.
* [23] Favaro B, Oester S, Cigliano JA, Cornick LA, Hind EJ, Parsons E, et al. Your science conference should have a code of conduct. Frontiers in Marine Science. 2016;3:103.
|
# A Cable Knot and BPS-Series
John Chae Department of Physics and QMAP, UC Davis, 1 Shields Ave, Davis, CA,
95616, USA
<EMAIL_ADDRESS>
###### Abstract
A series invariant of a complement of a knot was introduced recently. The
invariant for several prime knots up to ten crossings have been explicitly
computed. We present the first example of a satellite knot, namely, a cable of
the figure eight knot, which has more than ten crossings. This cable knot
result provides nontrivial evidence for the conjectures for the series
invariant and demonstrates the robustness of integrality of the quantum
invariant under the cabling operation. Furthermore, we find interesting
effects of the cabling on the series invariant.
###### CONTENTS
1. 1 Introduction
2. 2 Background
1. 2.1 Satellites
2. 2.2 Quantum Torus and Recursion Ideal
3. 3 Knot Polynomials
1. 3.1 The Colored Jones Polynomial
2. 3.2 The Alexander Polynomial
4. 4 The Recursion Relation
5. 5 An Expansion of a Knot Complement
6. 6 Effects of the Cabling
7. A The Definitions of the Operators
## 1 Introduction
Inspired by the categorification of the Witten-Reshitikhin-Turaev invariant of
a closed oriented 3-manifold [37, 30, 29] in [15, 16], a two variable series
invariant $F_{K}(x,q)$ for a complement of a knot $M^{3}_{K}$ was introduced
in [13]. Although its rigorous definition is yet to be found, it possesses
various properties such as the Dehn surgery formula and the gluing formula.
This knot invariant $F_{K}$ takes the form111Implicitly, there is a choice of
group; originally, the group used is ${\rm SU}(2)$.
$\displaystyle F_{K}(x,q)=\frac{1}{2}\sum_{\begin{subarray}{c}m\geq 1\\\ m\
\text{odd}\end{subarray}}^{\infty}\big{(}x^{m/2}-x^{-m/2}\big{)}f_{m}(q)\in\frac{1}{2^{c}}q^{\Delta}\mathbb{Z}\big{[}x^{\pm
1/2}\big{]}\big{[}\big{[}q^{\pm 1}\big{]}\big{]},$ (1)
where $f_{m}(q)$ are Laurent series with integer coefficients222They can be
polynomials for monic Alexander polynomial of $K$ (See Section 3.2),
$c\in\mathbb{Z}_{+}$ and $\Delta\in\mathbb{Q}$. Moreover, $x$-variable is
associated to the relative ${\rm
Spin}^{c}\big{(}M^{3}_{K},T^{2}\big{)}$-structures, which is affinely
isomorphic to $H^{2}\big{(}M^{3}_{K},T^{2};\mathbb{Z}\big{)}\cong
H_{1}\big{(}M^{3}_{K};\mathbb{Z}\big{)}$; it has an infinite order, which is
reflected as a series in $F_{K}$. The rational constant $\Delta$ was
investigated in [14], which elucidated its intimate connection to the
d-invariant (or the correction term) in certain versions of the Heegaard Floer
homology ($HF^{\pm}$) for rational homology spheres. The physical
interpretation of the integer coefficients in $f_{m}(q)$ are number of BPS
states of 3d $\mathcal{N}=2$ supersymmetric quantum field theory on
$M^{3}_{K}$ together with boundary conditions on $\partial M^{3}_{K}$.
Furthermore, it was conjectured that $F_{K}$ also satisfies the
Melvin–Morton–Rozansky conjecture [24, 31, 32] (proven in [1]):
###### Conjecture 1.1 ([13, Conjecture 1.5]).
For a knot $K\subset$ $S^{3}$, the asymptotic expansion of the knot invariant
$F_{K}\big{(}x,q={\rm e}^{\hbar}\big{)}$ about $\hbar=0$ coincides with the
Melvin–Morton–Rozansky (MMR) expansion of the colored Jones polynomial in the
large color limit:
$\displaystyle\frac{F_{K}\big{(}x,q={\rm
e}^{\hbar}\big{)}}{x^{1/2}-x^{-1/2}}=\sum_{r=0}^{\infty}\frac{P_{r}(x)}{\Delta_{K}(x)^{2r+1}}\hbar^{r},$
(2)
where $x=e^{n\hbar}$ is fixed, n is the color of $K$,
$P_{r}(x)\in\mathbb{Q}\big{[}x^{\pm 1}\big{]}$, $P_{0}(x)=1$ and
$\Delta_{K}(x)$ is the (symmetrized) Alexander polynomial of $K$.
Additionally, motivated by the quantum volume conjecture/AJ-conjecture [7, 11]
(explained in Section 2.2), it was conjectured that $F_{K}$-series is
q-holonomic:
###### Conjecture 1.2 ([13, Conjecture 1.6]).
For any knot $K\subset$ $S^{3}$, the normalized series $f_{K}(x,q)$ satisfies
a linear recursion relation generated by the quantum A-polynomial of $K$
$\hat{A}_{K}(q,\hat{x},\hat{y})$:
$\displaystyle\hat{A}_{K}(q,\hat{x},\hat{y})f_{K}(x,q)=0,$ (3)
where $f_{K}:=F_{K}(x,q)/\big{(}x^{1/2}-x^{-1/2}\big{)}$.
The actions of $\hat{x}$ and $\hat{y}$ are
$\hat{x}f_{K}(x,q)=xf_{K}(x,q)\qquad\hat{y}f_{K}(x,q)=f_{K}(xq,q).$
For examples of the $F_{K}$-series, several prime knots up to ten crossings
have been analyzed [13, 27, 28]. They include the torus knots, the figure
eight knot in [13], and $\bm{5_{1}}$ in [28]. Positive braid knots
($\bm{10_{139}}$, $\bm{10_{152}}$), strongly quasipositive braids knots
($\text{m}(\bm{10_{145}}),\bm{10_{154}},\bm{10_{161}}$), double twist knots
(m($\bm{5_{2}}$), m($\bm{7_{3}}$), m($\bm{7_{4}}$)), and a few more prime
knots (m($\bm{7_{5}}$), m($\bm{8_{15}}$)) were examined in [27].
In this paper we verify the above conjectures by computing the $F_{K}$-series
for $(9,2)$-cabling of the figure eight knot and we compare our result to that
of the figure eight knot. Furthermore, we conjecture about the $F_{K}$-series
for a family of a cable knot of the figure eight.
The rest of the paper is organized as follows. In Section 2 we review the
satellite operation on a knot and the recursion ideal of the quantum torus. In
Section 3 we analyze knot polynomials of the cable knot of the figure eight.
In Section 4 we derive the recursion relation for the cable knot. Then we
deduce $\hbar$ expansion from the recursion in Section 5. Finally in Section 6
consequences of the cabling operation are discussed and we propose conjectures
about a family of a cable knot.
Acknowledgments. I would like to thank Sergei Gukov, Thang L$\hat{\text{e}}$
and Laura Starkston for helpful conversations. I am grateful to Ciprian
Manolescu for valuable suggestions on a draft of this paper. I am also
grateful to Colin Adams for valuable comments.
## 2 Background
### 2.1 Satellites
The satellite operation consists of a pattern knot P in the interior of the
solid torus $S^{1}\times D^{2}$, a companion knot $K^{\prime}$ in the $S^{3}$
and an canonical identification $h_{K^{\prime}}$
$h_{K^{\prime}}:S^{1}\times D^{2}\longrightarrow\nu(K^{\prime})\subset S^{3},$
(4)
where $\nu(K^{\prime})$ is the tubular neighborhood of $K^{\prime}$.
Figure 1: A pattern knot $P$ (left), companion $K^{\prime}$ (center) and
satellite knot $P(K^{\prime})$ (right).
A well-known example of satellite knots is a cable knot
$h_{K^{\prime}}(P)=C_{(r,s)}(K^{\prime})$ that is obtained by choosing P to be
the $(r,s)$-torus knot pushed into the interior of the $S^{1}\times D^{2}$.
This map $h_{K^{\prime}}$ has been investigated in [23, 25, 26].
### 2.2 Quantum Torus and Recursion Ideal
Let $\mathcal{T}$ be a quantum torus
$\mathcal{T}:=\mathbb{C}[t^{\pm 1}]\left\langle M^{\pm 1},L^{\pm
1}\right\rangle/(LM-t^{2}ML).$
The generators of the noncommutative ring $\mathcal{T}$ acts on a set of
discrete functions, which are colored Jones polynomials
$J_{K,n}\in\mathbb{Z}[t^{\pm 1}]$ in our context, as
$MJ_{K,n}=t^{2n}J_{K,n}\qquad LJ_{K,n}=J_{K,n+1}.$
The recursion(annihilator) ideal $\mathcal{A}_{K}$ of $J_{K,n}$ is the left
ideal $\mathcal{A}_{K}$ in $\mathcal{T}$ consisting of operators that
annihilates $J_{K,n}$:
$\mathcal{A}_{J_{K,n}}:=\left\\{\alpha_{K}\in\mathcal{T}\,|\,\alpha_{K}J_{K,n}=0\right\\}.$
As it turns out that $\mathcal{A}_{K}$ is not a principal ideal in general.
However, by adding inverse polynomials of t and M to $\mathcal{T}$ [7], we
obtain a principal ideal domain $\tilde{\mathcal{T}}$
$\tilde{\mathcal{T}}:=\left\\{\sum_{j\in\mathbb{Z}}a_{j}(M)L^{j}\Big{|}\,a_{j}(M)\in\mathbb{C}[t^{\pm
1}](M),\,a_{j}=\text{almost always}\quad 0\right\\}$
Using $\tilde{\mathcal{T}}$ we get a principal ideal
$\tilde{\mathcal{A}_{K}}:=\tilde{\mathcal{T}}\mathcal{A}_{K}$ generated by a
single polynomial $\hat{A}_{K}$
$\hat{A}_{K}(t,M,L)=\sum_{j=0}^{d}a_{j}(t,M)L^{j}.$
This $\hat{A}_{K}$ polynomial is a noncommutative deformation of a classical
A-polynomial of a knot [3] (see also [4]). Alternative approaches to obtain
$\hat{A}_{K}(t,M,L)$ are by quantizing the classical A-polynomial curve using
a twisted Alexander polynomial or applying the topological recursion [17]. A
conjecture called AJ conjecture/quantum volume conjecture was proposed in [7,
11] via different approaches:
###### Conjecture 2.1.
For any knot $K\subset$ $S^{3}$, $\hat{A}_{K}(t=-1,L,M)$ reduces to the
(classical) A-polynomial curve $A_{K}(L,M)$ up to a solely M-dependent overall
factor.
In other words, $J_{K,n}(t)$ satisfies a linear recursion relation generated
by $\hat{A}_{K}(t,M,L)$. This property of $J_{K,n}$ is often called
q-holonomic [9]. The conjecture was confirmed for a variety of knots [5, 7, 8,
10, 19, 22, 36, 34].
## 3 Knot Polynomials
In this section we will analyze the colored Jones polynomial and the Alexander
polynomial of a cable knot to show that the former satisfies the MMR expansion
and the latter is monic. Furthermore, the MMR expansion enables us to read off
the initial condition that is needed in Section 5.
### 3.1 The Colored Jones Polynomial
For (r,2)-cabling of the figure eight knot $\bm{4_{1}}$, we set $P=T(r,2)$ and
$K^{\prime}=\bm{4_{1}}$ in (4). The cabling formula for an unnormalized
$\mathfrak{sl}_{2}(\mathbb{C})$ colored Jones polynomial of a $(r,2)$-cabling
of a knot $K^{\prime}$ in $S^{3}$ is [35]
$\tilde{J}_{C_{(r,2)}(K^{\prime}),n}(q)=q^{-\frac{r}{2}\left(n^{2}-1\right)}\sum_{w=1}^{n}(-1)^{r(n-w)}q^{\frac{r}{2}w(w-1)}\tilde{J}_{K^{\prime},\,(2w-1)}(q),\quad\quad|r|>8\quad\text{and
odd}.$
Figure 2: (r,2)-cable of the figure eight knot.
Its application to $K=C_{(9,2)}(\bm{4_{1}})$333This cabling parameters
correspond to $\bm{9_{1}}$ for the pattern knot. We assume 0-framing for
$\bm{4_{1}}$., whose diagram has 25 crossings, is
$\displaystyle\tilde{J}_{K,n}(q)$
$\displaystyle=q^{-\frac{9}{2}\left(n^{2}-1\right)}\sum_{w=1}^{n}\left[(-1)^{(n-w)}q^{\frac{9}{2}w(w-1)}[2w-1]\sum_{r=0}^{2w-2}\prod_{k=1}^{r}\left(-q^{-k}-q^{k}+q^{1-2w}+q^{2w-1}\right)\right].$
Using the (0-framed) unknot $U$ value together with $q=t^{4}$
$J_{U,n}(t)=\frac{t^{2n}-t^{-2n}}{t^{2}-t^{-2}},$
the first few unknot normalized polynomials $J_{K,n}(q)$ are
$\displaystyle J_{K,1}(q)=1$ $\displaystyle
J_{K,2}(q)=q^{2}-q+\frac{1}{q^{4}}+\frac{1}{q^{6}}-\frac{1}{q^{7}}+\frac{1}{q^{8}}-\frac{1}{q^{9}}+\frac{1}{q^{12}}-\frac{1}{q^{13}}$
$\displaystyle J_{K,3}(q)$
$\displaystyle=q^{12}-q^{11}-q^{10}+q^{9}-q^{8}+q^{7}+q^{6}-q^{5}+q^{2}-1+\frac{1}{q^{8}}+\frac{1}{q^{11}}-\frac{1}{q^{13}}+\frac{1}{q^{14}}-\frac{1}{q^{16}}+\frac{1}{q^{17}}$
$\displaystyle-\frac{1}{q^{18}}-\frac{1}{q^{19}}+\frac{2}{q^{20}}-\frac{1}{q^{21}}+\frac{1}{q^{23}}-\frac{1}{q^{24}}+\frac{1}{q^{25}}+\frac{1}{q^{26}}-\frac{2}{q^{27}}-\frac{1}{q^{28}}+\frac{1}{q^{29}}-\frac{1}{q^{30}}+\frac{2}{q^{32}}-\frac{1}{q^{33}}$
$\displaystyle-\frac{1}{q^{34}}+\frac{1}{q^{35}}$
Their $\hbar$ series are
$\displaystyle J_{K,n}(e^{\hbar})$
$\displaystyle=1+\left(6-6n^{2}\right)\hbar^{2}+\left(-42+42n^{2}\right)\hbar^{3}+\left(\frac{801}{2}-462n^{2}+\frac{123}{2}n^{4}\right)\hbar^{4}$
(5)
$\displaystyle+\left(-\frac{8451}{2}+5173n^{2}-\frac{1895}{2}n^{4}\right)\hbar^{5}+\left(\frac{3111491}{60}-\frac{132779}{2}n^{2}+14986n^{4}-\frac{27281}{60}n^{6}\right)\hbar^{6}$
$\displaystyle+\left(-\frac{14631401}{20}+\frac{19399417}{20}n^{2}-\frac{3028829}{12}n^{4}+\frac{840097}{60}n^{6}\right)\hbar^{7}$
$\displaystyle+\left(\frac{39069313501}{3360}-\frac{950122877}{60}n^{2}+\frac{54585517}{12}n^{4}-\frac{1725671}{5}n^{6}+\frac{13273763}{3360}n^{8}\right)\hbar^{8}+\cdots$
We see that, at each $\hbar$ order, the degree of the polynomial in n is at
most the order of $\hbar$, which is an equivalent characterization of the MMR
expansion of the colored Jones polynomial of a knot. Secondly, as a
consequence of the cabling, odd powers of $\hbar$ appear in the expansion,
which are absent in the case of the figure eight knot [13]. Moreover, the
coefficient polynomials for the odd $\hbar$-powers have one lower degree
whereas the degree of the polynomials are the same for the even
$\hbar$-powers. Hence they are unaffected by the cabling operation.
### 3.2 The Alexander Polynomial
The cabling formula for the Alexander Polynomial of a knot $K$ is [18]
$\Delta_{C_{(p,q)}(K)}(t)=\Delta_{K}(t^{p})\Delta_{T_{(p,q)}}(t),\quad 2\leq
p<|q|\quad\text{gcd}(p,q)=1,$
where $\Delta(t)$ is the symmetrized Alexander polynomial and $T_{(p,q)}$ is
the (p,q) torus knot. Note that our convention for the parameters of the torus
knot are switched (i.e. $p\equiv 2,q\equiv r$). Applying the above formula to
$C_{(9,2)}(\bm{4_{1}})$, we get
$\displaystyle\Delta_{C_{(9,2)}{(\bm{4_{1}})}}(x)$
$\displaystyle=\Delta_{\bm{4_{1}}}(x^{2})\Delta_{T_{(2,9)}}(x)$
$\displaystyle=-x^{6}-\frac{1}{x^{6}}+x^{5}+\frac{1}{x^{5}}+2x^{4}+\frac{2}{x^{4}}-2x^{3}-\frac{2}{x^{3}}+x^{2}+\frac{1}{x^{2}}-x-\frac{1}{x}+1.$
From this Alexander polynomial its symmetric expansion about $x=0$ (in x) and
$x=\infty$ (in $1/x$) in the limit of $\hbar\rightarrow 0$ can be computed.
$\displaystyle\lim_{q\rightarrow 1}2F_{K}(x,q)$
$\displaystyle=2\,\text{s.e}\left(\frac{x^{1/2}-x^{-1/2}}{\Delta_{K}(x)}\right)$
$\displaystyle=x^{11/2}-\frac{1}{x^{11/2}}+2x^{15/2}-\frac{2}{x^{15/2}}+5x^{19/2}-\frac{5}{x^{19/2}}+13x^{23/2}-\frac{13}{x^{23/2}}$
$\displaystyle+34x^{27/2}-\frac{34}{x^{27/2}}-x^{29/2}+\frac{1}{x^{29/2}}+89x^{31/2}-\frac{89}{x^{31/2}}-2x^{33/2}+\frac{2}{x^{33/2}}$
$\displaystyle+233x^{35/2}-\frac{233}{x^{35/2}}-5x^{37/2}+\frac{5}{x^{37/2}}+610x^{39/2}-\frac{610}{x^{39/2}}+\cdots\in\mathbb{Z}\left[\left[x^{\pm
1/2}\right]\right]$ (6)
The coefficients in the expansions are integers and hence the Alexander
polynomial is monic, which is a necessary condition for $f_{m}(q)$’s in (1) to
be polynomials.
## 4 The Recursion Relation
The quantum (or noncommutative) A-polynomial of a class of cable knot
$C_{(r,2)}(\bm{4_{1}})$ in $S^{3}$ having minimal L-degree is given by [33]
$\hat{A}_{K}(t,M,L)=(L-1)B(t,M)^{-1}Q(t,M,L)\left(M^{r}L+t^{-2r}M^{-r}\right)\in\tilde{\mathcal{A}}_{K}$
(7)
where
$Q(t,M,L)=Q_{2}(t,M)L^{2}+Q_{1}(t,M)L+Q_{0}(t,M),\quad
B(t,M):=\sum_{j=0}^{2}c_{j}b(t,t^{2j+2}M^{2})$
$b(t,M)=\frac{M(1+t^{4}M^{2})(-1+t^{4}M^{4})(-t^{2}+t^{14}M^{4})}{t^{2}-t^{-2}}$
$c_{0}=\hat{P}_{0}(t,t^{4}M^{2})\hat{P}_{1}(t,t^{6}M^{2}),\quad
c_{1}=-\hat{P}_{1}(t,t^{2}M^{2})\hat{P}_{1}(t,t^{6}M^{2}),\quad
c_{2}=\hat{P}_{1}(t,t^{2}M^{2})\hat{P}_{2}(t,t^{4}M^{2}).$
The definitions of the operators $\hat{P}_{i}$ are written in Appendix A. For
$K=C_{(9,2)}(\bm{4_{1}})$, applying (7) to $f_{K}(x,q)$ together with
$x=q^{n}$ yields
$\alpha(x,q)F_{K}(x,q)+\beta(x,q)F_{K}(xq,q)+\gamma(x,q)F_{K}(xq^{2},q)+\delta(x,q)F_{K}(xq^{3},q)+F_{K}(xq^{4},q)=0,$
(8)
where $\alpha,\beta,\gamma,\delta$ functions and their $\hbar$ series are
documented in [2]. From (8) we find the recursion relation for $f_{m}$.
$\displaystyle f_{m+98}\,(q)$
$\displaystyle=\frac{-1}{q^{\frac{109+m}{2}}\left(1-q^{\frac{87+m}{2}}\right)}\Big{[}t_{2}\,f_{m+94}+t_{4}\,f_{m+90}+t_{6}\,f_{m+86}+t_{8}\,f_{m+82}+t_{9}\,f_{m+80}+t_{10}\,f_{m+78}$
(9)
$\displaystyle+t_{11}\,f_{m+76}+t_{12}\,f_{m+74}+t_{13}\,f_{m+72}+t_{14}\,f_{m+70}+t_{15}\,f_{m+68}+t_{16}\,f_{m+66}+t_{17}\,f_{m+64}$
$\displaystyle+t_{18}\,f_{m+62}+t_{19}\,f_{m+60}+t_{20}\,f_{m+58}+t_{21}\,f_{m+56}+t_{22}\,f_{m+54}+t_{23}\,f_{m+52}+t_{24}\,f_{m+50}$
$\displaystyle+t_{25}\,f_{m+48}+t_{26}\,f_{m+46}+t_{27}\,f_{m+44}+t_{28}\,f_{m+42}+t_{29}\,f_{m+40}+t_{30}\,f_{m+38}+t_{31}\,f_{m+36}$
$\displaystyle+t_{32}\,f_{m+34}+t_{33}\,f_{m+32}+t_{34}\,f_{m+30}+t_{35}\,f_{m+28}+t_{36}\,f_{m+26}+t_{37}\,f_{m+24}+t_{38}\,f_{m+22}$
$\displaystyle+t_{39}\,f_{m+20}+t_{40}\,f_{m+18}+t_{41}\,f_{m+16}+t_{43}\,f_{m+12}+t_{45}\,f_{m+8}+t_{47}\,f_{m+4}+t_{49}\,f_{m}\vphantom{1}\Big{]}\in\mathbb{Z}[q^{\pm
1}]$
where $t_{v}=t_{v}(q,q^{m})$’s are listed in [2]. Using this recursion and the
initial data documented in [2], $F_{K}(x,q)$ can be obtained to any desired
order in x.
## 5 An Expansion of a Knot Complement
We next compute a series expansion of the $F_{K}$ of complement of the cable
knot $K$. Specifically, a straightforward computation from (8) yields an
ordinary differential equation(ODE) for $P_{m}(x)$ at each $\hbar$ order.
Using the initial conditions for the ODEs obtained from (5)
$P_{1}(1)=0,\quad P_{2}(1)=6,\quad P_{3}(1)=-42,\quad
P_{4}(1)=\frac{801}{2},\quad P_{5}(1)=-\frac{8451}{2},\quad\cdots$
we find that
$\displaystyle P_{1}(x)$
$\displaystyle=\,5x^{12}+\frac{5}{x^{12}}-10x^{11}-\frac{10}{x^{11}}-13x^{10}-\frac{13}{x^{10}}+36x^{9}+\frac{36}{x^{9}}-10x^{8}-\frac{10}{x^{8}}-16x^{7}-\frac{16}{x^{7}}+15x^{6}$
$\displaystyle+\frac{15}{x^{6}}-14x^{5}-\frac{14}{x^{5}}+16x^{4}+\frac{16}{x^{4}}-18x^{3}-\frac{18}{x^{3}}+19x^{2}+\frac{19}{x^{2}}-20x-\frac{20}{x}+20$
$\displaystyle\newline P_{2}(x)$
$\displaystyle=\,\frac{25x^{24}}{2}+\frac{25}{2x^{24}}-50x^{23}-\frac{50}{x^{23}}-14x^{22}-\frac{14}{x^{22}}+306x^{21}+\frac{306}{x^{21}}-\frac{641x^{20}}{2}-\frac{641}{2x^{20}}-448x^{19}$
$\displaystyle-\frac{448}{x^{19}}+\frac{2011x^{18}}{2}+\frac{2011}{2x^{18}}-358x^{17}-\frac{358}{x^{17}}-522x^{16}-\frac{522}{x^{16}}+612x^{15}+\frac{612}{x^{15}}-\frac{589x^{14}}{2}$
$\displaystyle-\frac{589}{2x^{14}}+508x^{13}+\frac{508}{x^{13}}-\frac{3325x^{12}}{2}-\frac{3325}{2x^{12}}+1648x^{11}+\frac{1648}{x^{11}}+1538x^{10}+\frac{1538}{x^{10}}-3932x^{9}$
$\displaystyle-\frac{3932}{x^{9}}+1574x^{8}+\frac{1574}{x^{8}}+1670x^{7}+\frac{1670}{x^{7}}-1798x^{6}-\frac{1798}{x^{6}}+396x^{5}+\frac{396}{x^{5}}-\frac{1521x^{4}}{2}$
$\displaystyle-\frac{1521}{2x^{4}}+4082x^{3}+\frac{4082}{x^{3}}-\frac{6541x^{2}}{2}-\frac{6541}{2x^{2}}-8334x-\frac{8334}{x}+16831.$
The subsequent $P_{m}$’s are documented in [2]. Substituting them into (2)
results in
$\displaystyle 2F(x,e^{\hbar})$
$\displaystyle=\Big{(}x^{11/2}-\frac{1}{x^{11/2}}+2x^{15/2}-\frac{2}{x^{15/2}}+5x^{19/2}-\frac{5}{x^{19/2}}+13x^{23/2}-\frac{13}{x^{23/2}}+34x^{27/2}-\frac{34}{x^{27/2}}$
$\displaystyle-x^{29/2}+\frac{1}{x^{29/2}}+89x^{31/2}-\frac{89}{x^{31/2}}-2x^{33/2}+\frac{2}{x^{33/2}}+233x^{35/2}-\frac{233}{x^{35/2}}-5x^{37/2}$
$\displaystyle+\frac{5}{x^{37/2}}+\cdots\Big{)}$
$\displaystyle+\hbar\,\Big{(}5x^{11/2}-\frac{5}{x^{11/2}}+12x^{15/2}-\frac{12}{x^{15/2}}+35x^{19/2}-\frac{35}{x^{19/2}}+104x^{23/2}-\frac{104}{x^{23/2}}+306x^{27/2}$
$\displaystyle-\frac{306}{x^{27/2}}-15x^{29/2}+\frac{15}{x^{29/2}}+890x^{31/2}-\frac{890}{x^{31/2}}-36x^{33/2}+\frac{36}{x^{33/2}}+2563x^{35/2}-\frac{2563}{x^{35/2}}$
$\displaystyle-105x^{37/2}+\frac{105}{x^{37/2}}+\cdots\Big{)}$
$\displaystyle+\hbar^{2}\,\Big{(}\frac{25}{2}x^{11/2}-\frac{25}{2}\frac{1}{x^{11/2}}+36x^{15/2}-\frac{36}{x^{15/2}}+\frac{247}{2}x^{19/2}-\frac{247}{2}\frac{1}{x^{19/2}}+426x^{23/2}-\frac{426}{x^{23/2}}$
$\displaystyle+1441x^{27/2}-\frac{1441}{x^{27/2}}-\frac{225}{2}x^{29/2}+\frac{225}{2}\frac{1}{x^{29/2}}+4781x^{31/2}-\frac{4781}{x^{31/2}}-324x^{33/2}+\frac{324}{x^{33/2}}$
$\displaystyle+2563x^{35/2}-\frac{2563}{x^{35/2}}-\frac{2207}{2}x^{37/2}+\frac{2207}{2}\frac{1}{x^{37/2}}+\cdots\Big{)}.$
Comparing to the series of the figure eight knot [13], we notice that every
order of $\hbar$ appears in the above series whereas the series corresponding
to the figure eight knot consists of only even powers of $\hbar$ (i.e.
$P_{i}(x)=0$ for i odd). This difference is an effect of the torus knot whose
expansion involve all powers of $\hbar$ [13]. Furthermore, the x-terms begin
from $m=11$ instead of $m=1$ and there are gaps in their powers. Specifically,
$x^{\pm 13/2},x^{\pm 17/2},x^{\pm 21/2}$ and $x^{\pm 25/2}$ are absent. This
is a consequence of the structure of (6). A distinctive feature of the cable
knot is that from $x^{\pm 29/2}$ the coefficients are negative. Moreover, the
positive and the negative coefficients alternate from that x-power for all
$\hbar$ powers. These differences persist in the higher $\hbar$-orders, which
are recorded in [2]. We will see these differences in a manifest way in the
next section.
## 6 Effects of the Cabling
Since the initial data plays a core role in the recursion relation method, we
discuss their features for the cable knot and then propose conjectures about
them, which can be a useful guide for finding initial data for a family of the
cable knots.
In the initial data (see [2]) for the recursion relation (9), we notice
several differences from that of the figure eight knot [13]. Before discussing
them, let us begin with the properties of the $F_{K}$ that are preserved by
the cabling. The initial data consists of an odd number of terms and power of
q increases by one between every consecutive terms in a fixed $f_{m}$ for all
m’s, which are also true for $f_{99}$ and $f_{101}$. Additionally, the
reflection symmetry of coefficients is retained up to $f_{43}$ for positive
coefficients and up to $f_{61}$ for the negative ones but of course, those
$f_{m}$’s do not have the complete amphichiral structure. These invariant
properties are a remnant feature of the amphichiral property of the figure
eight knot.
A difference is that the nonzero initial data begins from $f_{11}$ and the
gaps between the powers of x is four up to $x^{27/2}$, which is in the
accordance with $f_{m}$’s. These features are direct consequences of the
symmetric expansion of the Alexander polynomial of the cable knot (6). In the
case of the figure eight its coefficient functions start from $f_{1}$ and
there are no such gaps. Another distinctive difference is that $f_{m}$’s
containing negative coefficients appear from $m=29$. Moreover, the positive
and negative coefficient $f_{m}$’s alternate from $f_{27}$ (i.e. positive
coefficients for $f_{27}$, $f_{31}$, $\ldots$ and negative coefficients for
$f_{29}$, $f_{33}$, $\ldots$). Furthermore, from $f_{47}$ the reflection
symmetry of the positive coefficients in the appropriate $f_{m}$’s is broken.
This phenomenon also occurs for the negative coefficient $f_{m}$’s from
$m=65$. Breaking of the symmetry is expected since the cable knot of the
figure eight is not amphichiral. The next difference is that the largest power
of q in the positive coefficient $f_{m}$’s for $m\geq 15$, the powers increase
by $2,2,3,3,4,4,\ldots,11,11$. For the negative coefficient case, the changes
are $4,4,5,5,6,6,\ldots,11,11$ from $m=33$. The smallest powers of $f_{m}$’s
having positive coefficients exhibit their changes as
$0,0,-1,-1,-2,-2,-3,-3,\ldots$ and for those with negative coefficients the
pattern is $2,2,1,1,0,0,-1,-1,-2,-2,\ldots$.
An universal feature of the negative coefficient $f_{m}$’s in the initial data
is that their coefficient modulo sign is determined by the positive
coefficient $f_{m}$’s. For example, the absolute value of the coefficients of
$f_{29}$ is same as that of $f_{11}$; $f_{33}$’s coefficients come from that
of $f_{15}$ up to sign and so forth. Hence coefficients of $f_{m}$ having
negative coefficients are determined by $f_{m-18}$. In fact, this peculiar
coefficient correlation also exists in the non-initial data $f_{101}$ whose
coefficients are correlated with that of $f_{83}$.
###### Conjecture 6.1.
For a class of a cable knot of the figure eight
$K_{r}=C_{(r,2)}(\bm{4_{1}})\subset S^{3}$, $r>8$ and odd having monic
Alexander polynomial, the coefficient functions
$\left\\{f_{m}(q)\in\mathbb{Z}[q^{\pm 1}]\right\\}$ in $F_{K_{r}}(x,q)$ can be
classified into two (disjoint) subsets: one of them consists of elements
having all positive coefficients $\left\\{f^{+}_{t}(q)\right\\}_{t\in I^{+}}$
and the other subset contains elements whose coefficients are all negative
$\left\\{f^{-}_{w}(q)\right\\}_{w\in I^{-}}$. Furthermore, for every element
in $\left\\{f^{-}_{w}(q)\right\\}$, its coefficients coincide with that of an
element in $\left\\{f^{+}_{t}(q)\right\\}$ up to sign.
###### Conjecture 6.2.
For the family of knots in Conjecture 6.1, every nonzero element in
$\left\\{f_{m}(q)\right\\}$ consists of an odd number of terms and power of q
increases(decreases) by one between every consecutive terms of the element.
Moreover, coefficients of an element
$f^{-}_{v}(q)\in\left\\{f^{-}_{w}(q)\right\\}$ agree with coefficients of
$f^{+}_{v-2r}(q)\in\left\\{f^{+}_{t}(q)\right\\}$ up to sign for all $v\in
I^{-}$.
## Appendix A The Definitions of the Operators
We list the definitions of the operators in the $\hat{A}$-polynomial (7) in
Section 4.
$\displaystyle Q_{2}(t,M)$
$\displaystyle=\,\hat{P}_{2}(t,t^{4}M^{2})\,\hat{P}_{1}(t,t^{2}M^{2})\,\hat{P}_{0}(t,t^{6}M^{2})$
$\displaystyle Q_{1}(t,M)$
$\displaystyle=\,\hat{P}_{0}(t,t^{4}M^{2})\,\hat{P}_{1}(t,t^{6}M^{2})\,\hat{P}_{2}(t,t^{2}M^{2})-\hat{P}_{1}(t,t^{6}M^{2})\,\hat{P}_{1}(t,t^{2}M^{2})\,\hat{P}_{1}(t,t^{4}M^{2})$
$\displaystyle+\hat{P}_{2}(t,t^{4}M^{2})\,\hat{P}_{1}(t,t^{2}M^{2})\,\hat{P}_{0}(t,t^{6}M^{2})$
$\displaystyle Q_{0}(t,M)$
$\displaystyle=\,\hat{P}_{0}(t,t^{4}M^{2})\,\hat{P}_{1}(t,t^{6}M^{2})\,\hat{P}_{0}(t,t^{2}M^{2}),$
$\displaystyle\hat{P}_{0}(t,M):$ $\displaystyle=\,t^{6}M^{4}(-1+t^{12}M^{4})$
$\displaystyle\hat{P}_{1}(t,M):$
$\displaystyle=\,-(-1+t^{4}M^{2})(1+t^{4}M^{2})\left(1-t^{4}M^{2}-t^{4}M^{4}-t^{12}M^{4}-t^{12}M^{4}-t^{12}M^{6}+t^{16}M^{8}\right)$
$\displaystyle\hat{P}_{2}(t,M):$ $\displaystyle=\,t^{10}M^{4}(-1+t^{4}M^{4}).$
## References
* [1] Bar-Natan D., Garoufalidis S., On the Melvin–Morton–Rozansky conjecture, Invent. Math. 125 (1996), 103–133.
* [2] Chae J., Ancillary files for “A Cable Knot and BPS Series”, arXiv:2101.11708.
* [3] Cooper D., Culler D M., Gillet H., Long D., Shalen P., Plane curves associated to character varieties of 3–manifolds, Invent. Math. 118 (1994), 47–84.
* [4] Cooper D., Long D., Remarks on the A-polynomial of a Knot, Journal of Knot Theory and Its Ramifications 05 (1996), 609–628.
* [5] Dimofte T., Gukov S., Lenells J., Zagier D., Exact results for perturbative Chern–Simons theory with complex gauge group, Communications in Number Theory and Physics 3 (2009). arXiv:0903.2472v1.
* [6] Ekholm T., Gruen A., Gukov S., Kucharski P., Park S., Sułkowski P., $\hat{Z}$ at large $N$: from curve counts to quantum modularity, arXiv:2005.13349.
* [7] Garoufalidis S., On the characteristic and deformation varieties of a knot, in Proceedings of the Casson Fest, Geom. Topol. Monogr., Vol. 7, Geom. Topol. Publ., Coventry, 2004, 291–309, arXiv:math.GT/0306230.
* [8] Garoufalidis S., Koutschan C., The Noncommutative A-Polynomial of $(-2,3,n)$ Pretzel Knots, Experimental Mathematics 21(3) (2012), 241–251, arXiv:1101.2844.
* [9] Garoufalidis S., Le T Q T., The colored Jones function is q–holonomic, Geometry and Topology 9 (2005), 1253–1293, arXiv:math/0309214.
* [10] Garoufalidis S., Sun X., The non-commutative A-polynomial of twist knots, Journal of Knot Theory and Its Ramifications 19 (2010), 1571–1595, arXiv:0802.4074.
* [11] Gukov S., Three-dimensional quantum gravity, Chern–Simons theory, and the A-polynomial, Comm. Math. Phys. 255 (2005), 577–627, arXiv:hep-th/0306165.
* [12] Gukov S., Hsin P.-S., Nakajima H., Park S., Pei D., Sopenko N., Rozansky–Witten geometry of Coulomb branches and logarithmic knot invariants, arXiv:2005.05347.
* [13] Gukov S., Manolescu C., A two-variable series for knot complements, arXiv:1904.06057.
* [14] Gukov S., Putrov P., Park S., Cobordism invariants from BPS q-series, arXiv:2009.11874
* [15] Gukov S., Pei D., Putrov P., Vafa C., BPS spectra and 3-manifold invariants, J. Knot Theory Ramifications 29 (2020), 2040003, 85 pages, arXiv:1701.06567.
* [16] Gukov S., Putrov P., Vafa C., Fivebranes and 3-manifold homology, J. High Energy Phys. 2017 (2017), no. 7, 071, 81 pages, arXiv:1602.05302
* [17] Gukov S., Sulkowski. P, A-polynomial, B-model, and quantization, J. High Energy Phys. 2012 (2012), no. 2, 070, arXiv:1108.0002
* [18] Hedden M., On knot Floer homology and cabling, Algebraic and Geometric Topology 5 (2005), 1197–1222, arXiv:math/0406402.
* [19] Hikami K., Difference equation of the colored Jones polynomial for torus knot, Journal of Knot Theory and Its Ramifications 15 (2004), 959–965, arXiv:math/0403224v1.
* [20] Khovanov M., A categorification of the Jones polynomial, Duke Math. J. 101 (2000), 359–426, arXiv:math.QA/9908171.
* [21] Khovanov M., A categorification of the colored Jones polynomial, J. Knot Theory Ramifications 14 (2005), 111–130, arXiv:math/0302060.
* [22] Le T Q T., Tran A., On the AJ conjecture for knots, arXiv:1111.5258.
* [23] Levine A., Non-surjective satellite operators and piecewise-linear concordance, Forum of Mathematics, Sigma 4e34 (2016), arXiv:1405.1125.
* [24] Melvin P.M., Morton H.R., The coloured Jones function, Comm. Math. Phys. 169 (1995), 501–520.
* [25] Miller A., Homomorphism obstructions for satellite maps, arXiv:1910.03461.
* [26] Miller A., Piccirillo L., Knot traces and concordance, Journal of Topology 11 (2018), 201–220. arXiv:1702.03974.
* [27] Park S., Large color $R$-matrix for knot complements and strange identities, arXiv:2004.02087.
* [28] Kucharski P., Quivers for 3-manifolds: the correspondence, BPS states, and 3d ${\mathcal{N}}=2$ theories, arXiv:2005.13394.
* [29] Reshetikhin N., Turaev V., Ribbon graphs and their invariants derived from quantum groups, Comm. Math. Phys. 127 (1990), 1–26.
* [30] Reshetikhin N., Turaev V., Invariants of $3$-manifolds via link polynomials and quantum groups, Invent. Math. 103 (1991), 547–597.
* [31] Rozansky L., Higher order terms in the Melvin–Morton expansion of the colored Jones polynomial, Comm. Math. Phys. 183 (1997), 291–306, arXiv:q-alg/9601009.
* [32] Rozansky L., The Universal R-Matrix, Burau Representaion and the Melvin-Morton Expansion of the Colored Jones Polynomial, Adv. Math 134 (1998), 1–31, arXiv:q-alg/9604005.
* [33] Ruppe D., The AJ-conjecture and cabled knots over the figure eight knot, Topology and its Applications 188 (2015), 27–50, arXiv:1405.3887.
* [34] Ruppe D., Zhang X., The AJ-conjecture and cabled knots over torus knots, Journal of Knot Theory and Its Ramifications 24 (2015), 27–50, arXiv:1403.1858.
* [35] Tran A., On the AJ conjecture for cables of the figure eight knot, New York Journal of Mathematics 20 (2014), 727–741, arXiv:1405.4055.
* [36] Tran A., On the AJ conjecture for cables of twist knots, arXiv:1409.6071.
* [37] Witten E., Quantum field theory and the Jones polynomial, Comm. Math. Phys. 121 (1989), 351–399.
|
# Tracking Short-Term Temporal Linguistic Dynamics to Characterize Candidate
Therapeutics for COVID-19 in the CORD-19 Corpus
James Powell<EMAIL_ADDRESS>0000-0002-3517-7485 Los Alamos National
LaboratoryP.O. Box 1663Los AlamosNew MexicoUSA87545 and Kari Sentz
<EMAIL_ADDRESS>Los Alamos National LaboratoryP.O. Box 1663Los AlamosNew
MexicoUSA87545
###### Abstract.
Scientific literature tends to grow as a function of funding and interest in a
given field. Mining such literature can reveal trends that may not be
immediately apparent. The CORD-19 corpus represents a growing corpus of
scientific literature associated with COVID-19. We examined the intersection
of a set of candidate therapeutics identified in a drug-repurposing study with
temporal instances of the CORD-19 corpus to determine if it was possible to
find and measure changes associated with them over time. We propose that the
techniques we used could form the basis of a tool to pre-screen new candidate
therapeutics early in the research process.
††conference: SenSys 2020: 18th ACM Conference on Embedded Networked Sensor
Systems; November 16-19, 2020; Yokohama, Japan
## 1\. Introduction
Diachronic word analysis is Natural Language Processing (NLP) technique for
characterizing the evolution of words over time. Often used for historical
linguistic studies, it can also be applied to scientific literature (Tshitoyan
et al., 2019) and can reveal early evidence of scientific discoveries before
they become widely known.
Drug-repurposing studies aim to identify existing drugs that might be useful
in treating other diseases. The availability of large amounts of data about
drugs and infectious agents such as viruses has enabled such studies to be
performed in-silico. In early 2020, a number of repurposing studies were
undertaken to identify potential treatments for COVID-19.
The CORD-19 corpus (Wang et al., 2020) was established in March 2020 as a
repository for research related to SARS-COV-2 and other coronaviruses. It
aggregates content from PubMed, bioRxiv, medRxiv, and other sources, and it is
updated with new publications on a regular basis. Figure 1 illustrates the
growth of CORD-19 through mid-2020.
Figure 1. Weekly growth of the CORD-19 corpus
Using CORD-19, we conducted a diachronic survey of candidate therapeutics
identified in one of the more exhaustive drug re-purposing studies conducted
to date for COVID-19, undertaken in February 2020 at Oak Ridge National
Laboratory. This study, detailed in (Smith and Smith, 2020), analyzed drugs in
the SWEETLEAD database for potential antiviral properties. The study produced
a dataset identifying over 9,000 existing approved drugs and supplements as
potential candidate therapeutics for COVID-19. Most importantly for our
purposes, this dataset included commonly used drug or supplement names for
each candidate.
Our survey considered the following questions:
* •
How many candidate therapeutics appear in CORD-19?
* •
Do references to the candidates change over time?
## 2\. Materials and Methods
Using temporal snapshots of CORD-19 spanning March 13 to June 30, we computed
frequency and semantic representations for each candidate therapeutic found in
the corpus. For each temporal instance, we computed TF/IDF (Term
Frequency/Inverse Document Frequency) score for each candidate, a common
metric to evaluate the relative importance of terms in a corpus.
Figure 2. TF/IDF and cosine embedding distances
To perform semantic analysis, we first computed diachronic word embeddings for
each temporal instance of the corpus. These embeddings were aligned with one
another to ensure that terms from each temporal instance were comparable. The
technique we used is based on TWEC [3]. It uses a negative sampling
optimization of softmax to maximize the probability that a set of words
surrounding word $w_{k}$ are representative of its context in time ($C^{t}$),
when multiplied by the mean of atemporal word embedding vectors from $u$ (the
compass) for the same set of context words around $w_{k}$.
$\mathnormal{\max_{\mathbf{C}^{t}}\log
P({w_{k}}|\gamma({w_{k}}))=\sigma(\vec{u}_{k}\cdot\vec{c}{\,{}^{t}_{\gamma(w_{k})}})}$
Since the TWEC embedding model did not account for phrases, we incoporated an
additional step to indentify them. Phrases (including drug names) were then
specially encoded to allow them to be treated like words. Figure 2 shows
TF/IDF verses the mean embedding distance to the compass for candidate
therapeutics.
Because diachronic embedding instances were aligned with one another, we were
able to isolate a given candidate and visualize its semantic trajectory over
time (Figure 3). As the trajectory is based on nearest neighbors at a given
time, subtle changes in semantic associations become apparent (Stewart et al.,
2017).
Figure 3. Semantic trajectory of ’acetazolamide’ (rank -2.4). Nodes along the
path represent the candidate embedding vector and its two closest terms at
time $t$
Semantic trajectory of ’acetazolamide’ (rank -2.4). Nodes along the path
represent the candidate embedding vector and its two closest terms at time $t$
## 3\. Results and Analysis
We detected 14% (1267) of the candidate therapeutics in CORD-19 at 3/13,
increasing to 26% (2361) by 6/30. For candidates detected in multiple adjacent
temporal instances of the corpus, we were able measure their changes over
time. We found that many candidates exhibited increases in frequency, and
stable or strengthening semantic associations. However, given the nature of
this corpus, we suspected some would exhibit other kinds of change over time.
We found that some candidate therapeutics exhibited different patterns of
semantic associations. Using heatmap visualizations as described in (Xu and
Crestani, 2017), we can illustrate two additional recurring patterns of
behavior. Some candidates exhibited weakening semantic associations over time
(Figure 4), while others exhibited an abrupt persistent shift to a different
pattern (Figure 5). Additionally, we found that these changes were not
strongly correlated with changes to a target’s frequency scores.
Figure 4. Example of weakening semantic associations for the candidate
therapeutic ivermectin
Figure 5. Example of disrupted semantic associations for the candidate
therapeutic acetazolamide
## 4\. Conclusion
Our diachronic survey of candidate therapeutics for COVID-19 in the CORD-19
corpus found that some exhibited weakening or abrupt changes of semantic
associations. We speculate that this could be related to the publication of
new research that positively or negatively affected consideration of a
candidate therapeutic as a treatment for COVID-19. Future work will
investigate how to detect and quantify these patterns, and to determine if
there are any correlations between a target’s rank and magnitude of change.
## References
* (1)
* Smith and Smith (2020) Micholas Smith and Jeremy C. Smith. 2020. Repurposing Therapeutics for Covid-19. https://doi.org/10.26434/chemrxiv.11871402.v3
* Stewart et al. (2017) Ian Stewart, Dustin Arendt, Eric Bell, and Svitlana Volkova. 2017. Measuring, predicting and visualizing short-term change in word representation and usage in vkontakte social network. _arXiv preprint arXiv:1703.07012_ (2017).
* Tshitoyan et al. (2019) Vahe Tshitoyan, John Dagdelen, Leigh Weston, Alexander Dunn, Ziqin Rong, Olga Kononova, Kristin A. Persson, Gerbrand Ceder, and Anubhav Jain. 2019. Unsupervised word embeddings capture latent knowledge from materials science literature. _Nature_ 571 (july 2019), 95–98. https://doi.org/10.1038/s41586-019-1335-8
* Wang et al. (2020) Lucy Lu Wang, Kyle Lo, Yoganand Chandrasekhar, Russell Reas, Jiangjiang Yang, Darrin Eide, Kathryn Funk, Rodney Kinney, Ziyang Liu, William Merrill, et al. 2020\. CORD-19: The Covid-19 Open Research Dataset. arXiv:2004.10706v2
* Xu and Crestani (2017) Zaikun Xu and Fabio Crestani. 2017. Temporal Semantic Analysis and Visualization of Words. In _Proceedings of the 8th Italian Information Retrieval Workshop, Lugano, Switzerland, June 05-07, 2017_ _(CEUR Workshop Proceedings)_ , Vol. 1911. CEUR-WS.org, 52–62. http://ceur-ws.org/Vol-1911/9.pdf
|
# Damage detection in operational wind turbine blades using a new approach
based on machine learning
Kartik Chandrasekhar<EMAIL_ADDRESS>Nevena Stevanovic add2
Elizabeth J. Cross add1 Nikolaos Dervilis add1 Keith Worden add1
Dynamics Research Group, Department of Mechanical Engineering, University of
Sheffield, Mappin Street, Sheffield, S1 3JD, UK. Siemens Gamesa Renewable
Energy A/S, Borupvej 16, 7330 Brande, Denmark
###### Abstract
The application of reliable structural health monitoring (SHM) technologies to
operational wind turbine blades is a challenging task, due to the uncertain
nature of the environments they operate in. In this paper, a novel SHM
methodology, which uses Gaussian Processes (GPs) is proposed. The methodology
takes advantage of the fact that the blades on a turbine are nominally
identical in structural properties and encounter the same environmental and
operational variables (EOVs). The properties of interest are the first
edgewise frequencies of the blades. The GPs are used to predict the edge
frequencies of one blade given that of another, after these relationships
between the pairs of blades have been learned when the blades are in a healthy
state. In using this approach, the proposed SHM methodology is able to
identify when the blades start behaving differently from one another over
time. To validate the concept, the proposed SHM system is applied to real
onshore wind turbine blade data, where some form of damage was known to have
taken place. X-bar control chart analysis of the residual errors between the
GP predictions and actual frequencies show that the system successfully
identified early onset of damage as early as six months before it was
identified and remedied.
###### keywords:
structural health monitoring , wind turbine blades , machine learning ,
Gaussian processes , SCADA
## 1 Introduction
Wind turbines rely on a number of structurally-critical components that may
directly affect their power generation capabilities. Of note are the tower,
the gearbox and generator (for variable-speed wind turbines), the magnetic
drive (for direct-drive wind turbines), various bearings (such as main
bearings), and perhaps most importantly, the blades [1]. Blades are among the
most expensive components - depending on their size; their manufacturing costs
range between 10% and 20% of total manufacturing costs [2]. In recent years,
blades have progressively become larger, to harvest more energy, and maximise
power production. At the same time, since larger blades are associated with
larger weight penalties, blade designs have continually been modified to keep
the penalties low. These compromises have led to more flexible blades, and
therefore, lower safety margins [3, 4].
Cracks and delaminations are reported to be the most common form of damage in
blades [3]; these typically occur close to the blade root since this is the
region where the blades experience the highest bending strains due to turbine
rotation; this is especially true when taking large blades into consideration.
Towards the end of the design lifetime of the blades, damage mechanisms become
accentuated during excessive levels of winds, lightning strikes, and ice
accumulation (which leads to rotor imbalance problems). It is extremely
important to avoid critical blade failures, because when damaged blades
liberate, they have the potential to damage not only the turbines they were
attached to, but also other turbines in their vicinity [5].
There have also been heavy investments in the offshore wind industry, whereby
wind turbines have increasingly been deployed on various seas and oceans,
where wind conditions are favourable for energy production [6]. However, this
freedom has come at a cost of ease of maintenance, since it is significantly
more expensive to send technicians to inspect offshore structures via
expensive transportation [7].
For all these reasons, Structural Health Monitoring (SHM) technologies have
become appealing to both wind turbine manufacturers and operators alike. The
use of SHM technologies has largely been driven by a number of factors:
1. 1.
minimisation of downtime,
2. 2.
maximisation of reliability,
3. 3.
minimisation of operation and maintenance costs, and
4. 4.
extension of design lifetimes of monitored structures.
Currently, most of the focus in SHM research on small-to-medium sized blades
has been in laboratory environments, where conditions are controlled.
Additionally, researchers have had the luxury of multiple sensors and data
acquisition systems with high sampling rates. Signals captured by these
sensors include strains [8, 9], velocities [10], accelerations [11], acoustic
emissions [8, 12, 13], guided waves [14, 15], thermographic images [16], and
digital image correlation (DIC) images [17], to name a few. This type of
monitoring has facilitated high-fidelity diagnostics (indication and location
of damage), as well as prognostics (remaining life prediction). There have
been numerous methods and procedures to aid in damage indication and
localisation; these include changes in stiffness [18, 19] and modal properties
[20], principal component analysis [9], image processing of thermographic [16]
and DIC images [17], etc. Machine learning algorithms, such as artificial
neural networks, have also been used previously [14].
The above-mentioned research subjects are extremely valuable, as they
illustrate various routes to tackle the ultimate goal of having a reliable SHM
system. However, operational wind turbines around the world face challenges
that cannot be replicated in laboratory environments. Due to this issue - at
least for the time being - certain methods can seem to be impractical. Some of
the challenges are listed below.
* 1.
Firstly, there are significant costs involved in applying numerous sensors
with high data-acquisition rates; these include the costs of the sensors
themselves, as well as costs associated with data storage and processing. One
must also consider the fact that there are hundreds of thousands of wind
turbines spread across the world.
* 2.
Furthermore, monitoring and diagnostic systems need to be designed to be
robust to the heavily nonstationary conditions that wind turbines face.
Nonstationarity primarily stems from constantly-changing wind and loading
conditions (gusts, turbulence, etc.), and taking these into account is
therefore a challenging task.
* 3.
Damage localisation is another challenge in operational wind turbines, since
this is a function of excitation frequency and number of sensors, as well as
data-acquisition rates and sensor placement. There are additional limitations
as well; for example, one cannot place piezoelectric sensors along the length
of the blades due to the risks associated with electromagnetic interference
and lightning strikes.
* 4.
Diagnostic methods also need to find a balance between true and false
positives, and this would involve a cost-benefit analysis. For example, in a
laboratory, a 3% false positive rate may be deemed acceptable, but in the real
world, if all turbines are stopped for 3% of the time, costs associated with
downtime will have to be borne by turbine operators. On the other hand, if it
does turn out that a damage indicator is correct, and the damage becomes
irreparable, costs associated with blade replacement can also be considerable.
This article proposes methodologies that address some of these challenges. As
mentioned above, when turbines are operating, the effects of environmental and
operational variables cause significant nonstationarity; this is especially
true of the frequency response of the blades. For example, high wind speeds
can cause faster rotation of the turbines, which lead to stiffer blades and
higher natural frequencies; higher temperatures lead to less stiff blades,
hence lowering the natural frequencies. The methodologies proposed in this
article take advantage of the fact that the EOVs act simultaneously on all
three blades. Therefore, when the monitored structural frequencies are viewed
relative to each other, the complex nonstationarities are transformed into a
simpler, and more explicable space. As a result of this transformation, it is
possible to learn relationships between these monitored variables (as well as
other EOVs). These relationships are learnt, in pairs, using probabilistic
regression in the form of Gaussian Processes (GPs). The GPs are then used for
predictive purposes, and the residuals between the actual signals and
predicted signals can be used as an informative damage indicator.
The SHM scheme proposed is summarised in the next section. This is followed by
a detailed review of the theory behind GPs. The article is complemented by
three case studies of onshore wind turbines, where the proposed methodologies
were applied, and would have successfully identified damage in advance of
critical failure. It should be noted that the versatility of this methodology
means that they can also be applied to offshore wind turbines.
## 2 Methodology and Theory
Figure 1 illustrates a summary of the SHM system that is presented in this
article.
Figure 1: Summary of proposed SHM system
As shown in Figure 1, the SHM system capability is divided into four
categories: data acquisition, pre-processing, machine learning, and
diagnostics and actions; these are discussed in detail next.
### 2.1 Data acquisition
Acceleration sensors, and sensors that measure the environmental and
operational variables (EOVs) measure the respective signals. The edge
frequencies of each blade (i.e. the frequencies of the first bending mode in
the blade edgewise direction) are estimated from the acceleration signals. In
this paper, the machine learning techniques are applied to Supervisory Control
and Data Acquisition (SCADA) data. For readers unfamiliar with the SCADA
system, it is a utility that acquires data from sensors mounted on a
structure, and sends them to a central computer for monitoring and control
purposes. The data statistics are logged at regular intervals, and these are
typically ten-minute averages [21], which implies that the SCADA data has a
low-pass filter applied to it. The variables used from the SCADA data include:
1. 1.
edge frequencies of each blade (in Hz),
2. 2.
ambient temperature (in °C),
3. 3.
power output (in W),
4. 4.
generator rotation speed (in RPM), and
5. 5.
pitch angle (in °).
Prior to data logging into the SCADA system, the individual signals are
usually sampled at a rate of 10 Hz (in some cases, higher sampling is also
possible). The aim of this work is to detect long term damage, which typically
occurs over the course of several weeks or months. Therefore, although the
data in the SCADA system have a low-pass filter applied to them, they are
still capable of identifying the long term degradation in the blade stiffness,
and hence the edge frequency. The benefit of using SCADA data over the higher-
resolution data is that it saves processing time quite considerably.
### 2.2 Pre-processing
In this part of the system, SCADA data from the respective channels were
downloaded from a number of wind turbines whose blades were known to have
incurred damage. These time-series were then filtered, and processed to remove
outliers within the data. Outliers are quite a common occurrence in SCADA
data, and typically occur in the data logging process. Furthermore, since the
time-series are logged in various dedicated channels, occasionally, it is
common for data to not be recorded at certain times in a given channel, whilst
data in other channels are being logged. It is possible to resample the data
such that all data channels have identical sampling rates. However, this
requires some form of interpolation, which could result in misleading
information. Therefore, it was decided that only data that had common
timestamps would used for the remainder of the processing that followed.
Note that all the data used for the purposes of this article have been
normalised prior to the application of the machine learning and signal
processing methods. This is performed according to,
$x_{norm}=\frac{x-\mu}{\sigma}$ (1)
where the $x$ are the data being normalised, $\mu$ is the mean, and $\sigma$
is the standard deviation of selected training data. The selection of training
data will be discussed shortly, but it is the same training data that will be
used in the training of the Gaussian Processes (GPs). There are two primary
reasons for normalising the data.
* 1.
To avoid numerical instabilities that may occur in various computations in GP
inference.
* 2.
To avoid disclosing confidential information regarding blade properties.
Next, the various features were analysed to assess whether they contributed to
the observed properties of the blade (i.e. the edge frequencies). Figure 2
shows a representative example of how the estimated edge frequencies of each
blade correlate to each other. Note that the three blades on the turbine have
been labelled A, B, and C.
(a)
(b)
(c)
Figure 2: Correlation between normalised edge frequencies for: (a) Blade A vs.
Blade B, (b) Blade B vs. Blade C, and (c) Blade C vs. Blade A.
Based on Figure 2, it is clear to see that these frequencies correlate well
with each other in a linear fashion. Over time, as the blades age, the
correlations generally worsen, and when damage occurs, the relationships
generally break down. Therefore, the primary motivation behind the SHM system
was to evaluate when the correlations worsened and/or broke down. It should be
noted here that for the examples of correlations shown in Figure 2, simple
methods such as Bayesian linear regression would be sufficient rather than the
more complicated GPs. However, as shall be seen later, the correlations
between the edge frequencies of each blade during normal condition are
typically not as linear and noise-free on other wind turbines (see Figure 12).
Since GPs are more robust to such scenarios (since the functional forms need
not be predetermined), they are the chosen method for this work.
Apart from the study of correlations between the respective edge frequencies,
correlations between other EOVs and the edge frequencies of each blade were
also studied. Figures 3 (a) to (d) illustrate exemplar relationships between
the EOVs (ambient temperature, power output, generator rotation speed, and
pitch angle, respectively) with respect to the estimated frequencies.
(a)
(b)
(c)
(d)
Figure 3: Correlation between normalised edge frequencies and normalised: (a)
ambient temperature, (b) power output, (c) generator rotation speed, and (d)
pitch angle.
As is indicated in Figure 3, it was found that, out of the EOVs considered,
only the ambient temperature had any systematic influence on the estimated
frequencies. When the estimated frequencies and temperatures are viewed over
time, it is evident that the cyclical nature of the estimated frequencies
occur due to the seasonal variations, as shown in Figure 4.
(a)
(b)
Figure 4: Time series of normalised: (a) frequency, and (b) ambient
temperature.
The use of these EOVs (other than ambient temperature) in the machine learning
tasks that follow were expected to lead to less accurate GP predictions since
they do not contribute to the observed behaviour of the edge frequencies. Due
to this fact, only the edge frequencies and ambient temperature features were
selected for further processing.
In the next section, the theory of Gaussian processes (GPs) is provided.
### 2.3 Gaussian Processes
Gaussian processes (GPs) implement a nonparametric probabilistic regression.
They are best understood in terms of a function space viewpoint [22]. Let
$\boldsymbol{X}$ be an $N\times d$ matrix of inputs
$\\{\boldsymbol{x}_{1},\boldsymbol{x}_{2},\cdots,\boldsymbol{x}_{N}\\}$, with
$N$ observations and $d$ variables (also referred to as dimensions),
$f(\boldsymbol{X})$ denote the underlying functions of the GP,
$\\{f(\boldsymbol{x}_{1}),f(\boldsymbol{x}_{2}),\cdots,f(\boldsymbol{x}_{N})\\}$,
and $\boldsymbol{y}$ be the output vector, which is actually observed, and can
be corrupted by noise,
$\boldsymbol{y}=f(\boldsymbol{X})+\boldsymbol{\varepsilon}$ (2)
where $\boldsymbol{\varepsilon}$ is a zero mean Gaussian noise process, with
variance ${\sigma_{N}}^{2}$. Note that for the sake of conciseness,
$f(\boldsymbol{X})$ will henceforth simply be denoted as $\boldsymbol{f}$.
In the Bayesian approach, the prior belief of the form of the GP,
$\boldsymbol{f}$ is utilised along with the likelihood function of the
observed data to calculate its posterior distribution. This can be expressed
as,
$\text{posterior}=\frac{\text{likelihood}\times\text{prior}}{\text{marginal
likelihood}}$ (3)
$p(\boldsymbol{f}|\boldsymbol{y},\boldsymbol{X})=\frac{p(\boldsymbol{y}|\boldsymbol{X},\boldsymbol{f})p(\boldsymbol{f}|\boldsymbol{X})}{p(\boldsymbol{y}|\boldsymbol{X})}=\frac{p(\boldsymbol{y}|\boldsymbol{X},\boldsymbol{f})p(\boldsymbol{f}|\boldsymbol{X})}{\int{p(\boldsymbol{y}|\boldsymbol{X},\boldsymbol{f})p(\boldsymbol{f}|\boldsymbol{X})d\boldsymbol{f}}}$
(4)
where $p(\boldsymbol{f}|\boldsymbol{y},\boldsymbol{X})$ is the posterior
distribution (probability of functions given the outputs and inputs,
$p(\boldsymbol{y}|\boldsymbol{X},\boldsymbol{f})$ is the likelihood function
(probability of outputs, given inputs and functions),
$p(\boldsymbol{f}|\boldsymbol{X})$ is the prior distribution (probability of
the functions, given inputs), and $p(\boldsymbol{y}|\boldsymbol{X})$ is the
marginal distribution (probability of outputs, given only the inputs).
$p(\boldsymbol{y}|\boldsymbol{X})$ is independent of the functions because it
is integrated over the entire functional space of $\boldsymbol{f}$.
In the framework of GPs, functions between inputs and outputs can be
completely defined by a mean function, $m(\boldsymbol{x})$, and a covariance
function, $k(\boldsymbol{x}_{p},\boldsymbol{x}_{q})$, and these are in turn
defined by hyperparameters [22]. Note that the subscripts are used to
distinguish one input vector from another. This definition is analogous to a
Gaussian distribution, which can be fully described using a mean and a
variance. GPs therefore provide a distribution of functions between inputs and
outputs. The mean and covariance functions are expressed as,
$m(\boldsymbol{x})=E[f(\boldsymbol{x})]$ (5)
$k(\boldsymbol{x}_{p},\boldsymbol{x}_{q})=E[(f(\boldsymbol{x}_{p})-m(\boldsymbol{x}_{p}))(f(\boldsymbol{x}_{q})-m(\boldsymbol{x}_{q}))]$
(6)
The covariance function is a vital ingredient of the GP, as it characterises
the general properties of the GP. Typically, when $\boldsymbol{x}_{p}$ and
$\boldsymbol{x}_{q}$ are almost identical,
$k(\boldsymbol{x}_{p},\boldsymbol{x}_{q})$ would yield high values (high
correlation), and as they grow in distance, the values reduce in magnitude
(low correlation). Due to this, when predictions are performed, new inputs
that geometrically lie close to previously observed inputs will have a greater
influence than those that lie far away.
#### 2.3.1 Covariance functions
The choice of covariance functions is one of fundamental importance. Depending
on the problem being solved, there are numerous covariance functions (also
referred to as kernels in the literature) available to use. Examples include
squared-exponential, Matérn, $\gamma$-exponential, rational quadratic, and the
Bayesian linear covariance functions [22]. These functions can be combined
using basic algebra (addition, multiplication, etc.) to form advanced
covariance functions. In the work presented in this article, only the squared-
exponential and Bayesian linear covariance functions are used, and hence the
discussion will be focussed on these.
The squared-exponential covariance function is defined as,
$k_{SE}(\boldsymbol{x}_{p},\boldsymbol{x}_{q})={\sigma_{f}}^{2}\exp\bigg{(}-\frac{(\boldsymbol{x}_{p}-\boldsymbol{x}_{q})^{T}(\boldsymbol{x}_{p}-\boldsymbol{x}_{q})}{2\lambda^{2}}\bigg{)}$
(7)
where ${\sigma_{f}}^{2}$ is a hyperparameter that represents the signal
variance, and $\lambda$ is a length-scale hyperparameter (which controls how
smooth the function appears). Note that
$(\boldsymbol{x}_{p}-\boldsymbol{x}_{q})^{T}(\boldsymbol{x}_{p}-\boldsymbol{x}_{q})$
is simply the squared Euclidean distance between input points.
The Bayesian linear covariance function is defined as,
$k_{BL}(\boldsymbol{x}_{p},\boldsymbol{x}_{q})={\sigma_{0}}^{2}\boldsymbol{x}_{p}\cdot\boldsymbol{x}_{q}$
(8)
where ${\sigma_{0}}^{2}$ is also a hyperparameter that represents the signal
variance. This covariance function is classed as a dot-product covariance
function, for obvious reasons.
Finally, input noise can also be taken into account, i.e. when observations
are noisy. This is defined via,
$k_{N}(\boldsymbol{x}_{p},\boldsymbol{x}_{q})={\sigma_{N}}^{2}\delta(\boldsymbol{x}_{p},\boldsymbol{x}_{q})={\sigma_{N}}^{2}\boldsymbol{I}$
(9)
where ${\sigma_{N}}^{2}$ is a hyperparameter that represents the variance of
the noise process and $\delta$ is the Kronecker-delta function, which can
simply be viewed as the identity matrix, $\boldsymbol{I}$.
For the GPs associated with this work, a sum of the above covariance functions
was considered to explain the observed data. There are three distinct
covariance matrices associated with the predictive GPs:
1. 1.
$\boldsymbol{K_{\theta}(X,X)}$: Covariance matrix between training data only.
For the sake of convenience, this will simply be denoted
$\boldsymbol{K_{\theta}}$.
$\boldsymbol{K_{\theta}}=k_{SE}(\boldsymbol{x}_{p},\boldsymbol{x}_{q})+k_{BL}(\boldsymbol{x}_{p},\boldsymbol{x}_{q})+k_{N}(\boldsymbol{x}_{p},\boldsymbol{x}_{q})$
(10)
2. 2.
$\boldsymbol{K_{*}(X,X^{*})}$: Cross covariance matrix between training and
test data. The asterisk superscript denotes test data. Note that
$\boldsymbol{K_{*}(X^{*},X)}$ can also be calculated separately, but it is
simply the transpose of $\boldsymbol{K_{*}(X,X^{*})}$. For conciseness,
$\boldsymbol{K_{*}(X,X^{*})}$ will be denoted $\boldsymbol{K_{*}}$.
$\boldsymbol{K_{*}}=k_{SE}(\boldsymbol{x}_{p},\boldsymbol{x}_{q}^{\boldsymbol{*}})+k_{BL}(\boldsymbol{x}_{p},\boldsymbol{x}_{q}^{\boldsymbol{*}})$
(11)
3. 3.
$\boldsymbol{K_{**}(X^{*},X^{*})}$: Covariance matrix between test data only.
This will simply be denoted $\boldsymbol{K_{**}}$.
$\boldsymbol{K_{**}}=k_{SE}(\boldsymbol{x}_{p}^{\boldsymbol{*}},\boldsymbol{x}_{q}^{\boldsymbol{*}})+k_{BL}(\boldsymbol{x}_{p}^{\boldsymbol{*}},\boldsymbol{x}_{q}^{\boldsymbol{*}})$
(12)
It should be noted that in the latter two covariance functions, the noise term
is not added for the prediction of the noise-free underlying function,
$\boldsymbol{f^{*}}$.
Now that the background of GPs has been introduced, it is necessary to discuss
how they can be used in regression problems. Matters will be elaborated using
the terms in the Bayes’ theorem: the prior, the likelihood, and the marginal
likelihood. The marginal likelihood is utilised in the training process, where
the hyperparameters that define the GP are learnt.
#### 2.3.2 The prior distribution
The prior to the functions, $\boldsymbol{f}$ has the form,
$\boldsymbol{f}\sim\mathcal{GP}(0,\boldsymbol{K})$ (13)
where the 0 indicates a zero-mean function. $\boldsymbol{K}$ is now the
covariance matrix, without the noise term. The benefit of incorporating a mean
function is that in the absence of training data in a given region of the
input space, the prediction of the GP will approach the mean trend that is
defined by the mean function. This is because it specifies a belief in the
relationship between inputs and outputs having a certain functional form. It
is noteworthy that it is not necessary to limit the form of the GP using a
mean. This is because, in practice, data is typically de-trended, and so, mean
trends are already removed [23]. In any case, the prior distribution gets
updated in each step of the training process, and so the mean function can be
set to zero. However, this does not imply that the output of the GP will have
zero mean, and is hence not a limiting issue. It has a distribution,
$p(\boldsymbol{f}|\boldsymbol{X})=\mathcal{N}(0,\boldsymbol{K})$ (14)
#### 2.3.3 The likelihood distribution
$\boldsymbol{f}$ is a realisation of the underlying GP with inputs,
$\boldsymbol{X}$. Because $\boldsymbol{y}$ is simply $\boldsymbol{f}$ with
additive Gaussian noise ($\mathcal{N}(0,{\sigma_{N}}^{2})$), the likelihood
distribution can be stated as,
$p(\boldsymbol{y}|\boldsymbol{X},\boldsymbol{f})=\mathcal{N}(\boldsymbol{f},{\sigma_{N}}^{2}\boldsymbol{I})$
(15)
#### 2.3.4 The marginal likelihood and hyperparameter learning
The denominator in equation (4) is important in estimating (or learning) the
hyperparameters. It is also normally distributed with the form,
$p(\boldsymbol{y}|\boldsymbol{X})=\mathcal{N}(0,\boldsymbol{K_{\theta}})$ (16)
where $\boldsymbol{K_{\theta}}$ is the covariance matrix with additive noise,
i.e. $\boldsymbol{K_{\theta}}=\boldsymbol{K}+{\sigma_{N}}^{2}\boldsymbol{I}$.
The marginal likelihood is usually expressed in terms of its logarithmic
transformation: the log marginal likelihood [22],
$\mathcal{L}=\log{}p(\boldsymbol{y}|\boldsymbol{X})=-\frac{1}{2}\boldsymbol{y}^{T}\boldsymbol{K_{\theta}}^{-1}\boldsymbol{y}-\frac{1}{2}\log{}|\boldsymbol{K_{\theta}}|-\frac{N}{2}\log{}(2\pi)$
(17)
where $N$ is the number of training points, and the other definitions are
consistent with the preceding discussion. When the hyperparameters of the
covariance functions are being learnt, the goal is to maximise the marginal
likelihood with respect to the hyperparameters. Alternatively, the negative
log marginal likelihood can also be minimised. If the hyperparameters of the
covariance functions can be combined in a tuple,
$\boldsymbol{\varphi}=\\{\sigma_{f},\lambda,\sigma_{0},\sigma_{N}\\}$, the
partial derivative of the negative log marginal likelihood can be expressed
as,
$\frac{\partial}{\partial\boldsymbol{\varphi}_{i}}(-\mathcal{L})=\frac{1}{2}tr\bigg{[}\boldsymbol{K_{\theta}}^{-1}\frac{\partial\boldsymbol{K_{\theta}}}{\partial\boldsymbol{\varphi}_{i}}\bigg{]}-\frac{1}{2}\boldsymbol{y}^{T}\boldsymbol{K_{\theta}}^{-1}\frac{\partial\boldsymbol{K_{\theta}}}{\partial\boldsymbol{\varphi}_{i}}\boldsymbol{K_{\theta}}^{-1}\boldsymbol{y}$
(18)
where $tr$ is the trace operator (i.e. sum of the diagonal elements of the
matrix). In practice, the negative log marginal likelihood of GPs may have
several local minima. Therefore, solving this problem becomes one of numerical
optimisation. There are several routes one can take, for example gradient
descent, differential evolution, Broyden-Fletcher-Goldfarb-Shanno algorithm,
etc. There are no right or wrong choices in the optimisation routines, but
each optimiser can have advantages and disadvantages associated with them. In
this work, the Nelder-Mead simplex algorithm was used as the optimiser [24].
#### 2.3.5 Posterior distribution
The posterior distribution can be obtained by combining equations (14), (15)
and (16), and using Bayes’ theorem - equation (3); this is given by,
$p(\boldsymbol{f}|\boldsymbol{X},\boldsymbol{y})=\mathcal{N}(\boldsymbol{K}^{T}\boldsymbol{K_{\theta}}^{-1}\boldsymbol{y},\boldsymbol{K}-\boldsymbol{K}^{T}\boldsymbol{K_{\theta}}^{-1}\boldsymbol{K})$
(19)
#### 2.3.6 Predictions using GPs
Thus far, the necessary ingredients of the GPs have been shown. The next task
is to use these for the main goal of the GPs: prediction. Let
$\boldsymbol{x^{*}}$ denote a test input vector, (i.e. a matrix which is now
going to be used to predict the outputs). The corresponding noise-free test
output is $\boldsymbol{f^{*}}$. Now, the observed (noisy) training outputs and
the noise-free test outputs are jointly Gaussian distributed according to the
prior distribution, and can be expressed as,
$\begin{bmatrix}\boldsymbol{y}\\\
\boldsymbol{f^{*}}\end{bmatrix}=\mathcal{N}\begin{pmatrix}0,\begin{bmatrix}\boldsymbol{K_{\theta}}&\boldsymbol{K_{*}}\\\
\boldsymbol{K_{*}}^{T}&\boldsymbol{K_{**}}\end{bmatrix}\end{pmatrix}$ (20)
where the respective covariance matrices were defined in equations (10), (11),
and (12). The corresponding conditional distribution can be expressed as,
$p(\boldsymbol{f^{*}}|\boldsymbol{X},\boldsymbol{y},\boldsymbol{x^{*}})=\mathcal{N}(\boldsymbol{\bar{f^{*}}},cov(\boldsymbol{f^{*}}))$
(21)
where the predictive mean, $\boldsymbol{\bar{f^{*}}}$ of the GP can be
calculated using,
$\boldsymbol{\bar{f^{*}}}=\boldsymbol{K_{*}}^{T}\boldsymbol{K_{\theta}}^{-1}\boldsymbol{y}$
(22)
and the predictive covariance, $cov(\boldsymbol{f^{*}})$ of the GP can be
calculated using,
$cov(\boldsymbol{f^{*}})=\boldsymbol{K_{**}}-\boldsymbol{K_{*}}^{T}\boldsymbol{K_{\theta}}^{-1}\boldsymbol{K_{*}}$
(23)
The diagonal of the matrix given by equation (23) is then used to give the
uncertainties (in terms of variance) in prediction. The noise variance can be
added to the predictive variance to take uncertainty related to noisy test
outputs into consideration, and hence instead of finding the predictive
distribution of $\boldsymbol{f^{*}}$, the predictive distribution of
$\boldsymbol{y^{*}}$ can be calculated. Note that this does not affect the
predictive mean, only the predictive covariance.
### 2.4 Application of machine learning
Figure 5 provides an overview of the flow of data in the machine learning
phase of the SHM scheme, which includes training, optimisation, and Gaussian
Process (GP) predictions.
Figure 5: Flow of data in the proposed methodology.
Figure 5 shows that there are a total of three GPs that perform the
predictions. However, before any predictions could be performed, the first
task was to learn the hyperparameters, $\varphi$ of the models that controlled
their predictive capabilities - and this entailed a training process. Here, a
subset of the selected feature data were selected for the training.
When choosing the training data, it was important to consider only data when
the blades were in a normal (healthy) condition. However, the conditions of
the blades were not known beforehand. It was therefore decided that training
data would only be selected from the first two years of the turbine operation,
assuming that the blades were in a normal condition throughout this period.
Within this period of two years, 2500 data points were randomly selected from
a uniform distribution. Choosing this two year time period for training
ensured that the selected data spanned the entire normal operating range of
the blades. The random selection ensured that there was no bias to any
specific operating condition in the training phase. It should be noted that
more training data would increase the computational costs of the GPs, since
computational costs are $\mathcal{O}(N^{3})$, where $N$ is the number of
training points. There are sparse methods available to reduce the
computational load associated with training, but these were not employed in
this work.
The estimated frequencies, $f$ of one blade (e.g. for blade A, $f_{A}$) were
combined with the ambient temperatures, $T$ to form a matrix of the training
inputs. Referring to Section 2.3, this is the matrix, $\boldsymbol{X}$. The
estimated frequencies of another blade (e.g. for blade B, $f_{B}$) formed the
training outputs. This is the vector, $\boldsymbol{y}$. in Section 2.3. These
training data were used to learn the hyperparameters ($\varphi$) for the
respective GPs via the Nelder-Mead method. During the learning phase, the
covariance matrices, equations (10 \- 12) were evaluated several times, which
is why it was stated that the computational costs of GPs increases
significantly with the number of training points. The hyperparameters selected
corresponded to those that yielded the largest maximum log marginal
likelihood, equation (17). The Nelder-Mead algorithm identifies this by
minimising the negative log marginal likelihood, i.e. the negative of equation
(17). These hyperparameters control the generalisation properties of the GPs,
i.e. how the GPs adapt to the nature of the problem being tested. The asterisk
symbols, ∗ shown in Figure 5 are used to denote test data, i.e. data not
included in the training set. The test data, $f^{*}$ and $T^{*}$ were combined
in a matrix (referred to as the matrix, $\boldsymbol{X^{*}}$ in Section 2.3),
and were then used in the various GPs (GPAB, GPBC, and GPCA), using the
respective hyperparameters ($\varphi_{AB}$, $\varphi_{BC}$, and
$\varphi_{CA}$), to evaluate the predicted frequencies, $\bar{f_{B}}$,
$\bar{f_{C}}$, and $\bar{f_{A}}$, respectively. The bar symbol ( $\bar{}$ )
indicates a prediction, rather than an actual estimate of the frequencies. The
predictions are referred to as $\boldsymbol{f^{*}}$ in Section 2.3, and are
calculated using equation (22). To elucidate,
* 1.
in GPAB, ${f_{A}}^{*}$ and $T^{*}$ were used to predict $\bar{f_{B}}$ using
hyperparameters $\varphi_{AB}$,
* 2.
in GPBC, ${f_{B}}^{*}$ and $T^{*}$ were used to predict $\bar{f_{C}}$ using
hyperparameters $\varphi_{BC}$, and
* 3.
in GPCA, ${f_{C}}^{*}$ and $T^{*}$ were used to predict $\bar{f_{A}}$ using
hyperparameters $\varphi_{CA}$.
### 2.5 Diagnostics and actions
Figure 6 shows the tasks involved in the diagnostics and actions section.
Figure 6: Diagnostics of GP residuals.
Following the frequency predictions for each blade via the respective GPs,
residual errors were calculated between the predicted frequencies and the
actual estimations (for example, the GP residual error for blade B, $re_{B}$
was defined as the difference between $\bar{f_{B}}$ and ${f_{B}}^{*}$).
The assessment of whether the blades were healthy or not was performed using
X-bar control charts. In the X-bar control charts, averages of the GP
residuals were taken at regular time intervals. For the purposes of this work,
averages were taken regularly over a 28 day interval. During a brief period
that followed the GP training phase detailed in the previous section,
3$\sigma$ thresholds were also calculated. These were evaluated over a period
of 6 months. The calculated averages of the X-bar control chart were then
compared with the 3$\sigma$ thresholds, and if the averages exceeded the
thresholds, it was concluded that the blades were no longer in a normal
condition.
During the development of this SHM system, maintenance actions were not taken.
In practice, once the GP residuals surpass the thresholds of the X-bar control
charts, warnings would be generated before alarms occur. In such a scenario,
maintenance technicians would be sent to inspect the blades, they would assess
the level of damage, and finally would decide whether to repair or replace the
turbines.
### 2.6 Summary
The following summarises the SHM system as detailed in the preceding sections.
1. 1.
Data acquisition.
2. 2.
Data cleansing (removal of outliers).
3. 3.
Feature selection.
4. 4.
GP training (identification of hyperparameters).
5. 5.
GP prediction.
6. 6.
Calculation of residual errors.
7. 7.
X-bar control chart analysis of GP residuals.
## 3 Case studies
Now that the diagnostic methodology and theory have been described in detail,
it is important to illustrate its performance. Prior to the results and
discussion of the case studies, the detailed steps of the employed machine
learning methodology are stated.
### 3.1 Detailed implementation
In this section, a detailed description is given/reiterated regarding the
implementation of the machine learning methods described thus far, referring
to the seven steps in Section 2.6. That is,
1. 1.
Data acquisition: In the following sections, three sets of data are
introduced. The data are either synthesised or acquired directly from
operational wind turbines. In the synthesised case, it is acknowledged that
the data does not reflect the true behaviour of blades. However, the method of
the data synthesis is described in detail to allow readers to reproduce them
and apply the machine learning methodology. In the two cases where the data is
acquired from operational wind turbines, a blade on each turbine shown was
known to be damaged, and as a result, it either needed to be repaired or
replaced. Damage can occur at various locations across the length of the
blade, and the types of damage in blades are predominantly cracks or
delaminations. In some cases, both occur at the same time. Due to
confidentiality issues, however, the specific damage types and locations are
not disclosed here.
2. 2.
Data cleansing (removal of outliers): In the case of the synthesised data,
outliers are not included in the data, hence this process is not conducted. In
the case of the real data, the standard deviations of the dataset are
calculated, and data points outside 3$\sigma$ of the data are iteratively
removed. This process is repeated until the difference in standard deviation
between subsequent iterations is less than 0.1. This value is empirically
identified to ensure that outliers due to blade damage are not removed. Note
that when an outlier is removed from one of the data channels, data from the
other channels with the same timestamp (which may not necessarily be outliers)
are also removed.
3. 3.
Feature selection: The main features of interest are the edge frequencies of
the blades. It is shown in Example 3 below that the inclusion of temperature
as a feature improves the performance of the GPs.
4. 4.
GP training (identification of hyperparameters): For each example shown, 2500
training points are selected from a period when the blades on the turbine are
considered ‘healthy’. In the case of data from the real turbines, this was
chosen to be the first two years of turbine operation. The training points,
indexed by the data point numbers, are sampled uniformly via that index.
Standard normalisation, equation (1) is applied to the entire dataset using
the mean and standard deviation values of the training data points. Next, the
training input and output edge frequency data are prepared into three pairs
(that is, $f_{A}$ and $f_{B}$, $f_{B}$ and $f_{C}$, and $f_{C}$ and $f_{A}$;
temperature, $T$ is also included for the training inputs). The
hyperparameters, $\phi_{AB}$, $\phi_{BC}$, and $\phi_{CA}$, are then
identified by minimising the negative log marginal likelihood, which is
equivalent to maximising log marginal likelihood, equation (17). The Nelder-
Mead algorithm is used to accomplish this.
5. 5.
GP prediction: The identified hyperparameters for each GP are used with the
rest of the edge frequency and temperature data (${f_{A}}^{*}$, ${f_{B}}^{*}$,
${f_{C}}^{*}$, and $T^{*}$) are used to evaluate the predicted edge
frequencies, $\bar{f_{A}}$, $\bar{f_{B}}$, and $\bar{f_{C}}$ using equation
(22). Note that for relatively large datasets such as those used in this work,
computer memory can easily be used up since the covariance matrices would be
too large. For this reason, the GP predictions are performed in smaller chunks
(1000 points at a time in this case).
6. 6.
Calculation of residual errors: The GP residuals are calculated by subtracting
the actual edge frequency data from the predicted edge frequencies. That is,
$\displaystyle re_{A}$ $\displaystyle=\bar{f_{A}}-{f_{A}}^{*}$ (24)
$\displaystyle re_{B}$ $\displaystyle=\bar{f_{B}}-{f_{B}}^{*}$ $\displaystyle
re_{C}$ $\displaystyle=\bar{f_{C}}-{f_{C}}^{*}$
7. 7.
X-bar control chart analysis of GP residuals: In this step, the X-bar control
charts of the residuals are calculated. This is performed by calculating
averages of the residual errors over periods spanning 28 days. As the GPs are
trained over the first two year period on the examples shown from real wind
turbines, it is important that the thresholds for the X-bar control chart
analyses are not evaluated over the same period. This is because, generally,
residual errors are low over training data since optimisation routines aim to
minimise these errors as much as possible when identifying the
hyperparameters. Thus, such practices can potentially lead to stringent
thresholds that may increase false alarm rates. For this reason, a six-month
period immediately after the initial two-year training period is chosen. As
the standard deviation of the errors during this period may be occasionally
large, the following technique is used to calculate robust thresholds. For
each of the residual error datasets over this six-month period, small subsets
are randomly selected several times (in this work, 20 times). For each subset
$i$, the mean $\mu_{i}$ and standard deviation $\sigma_{i}$ values are
calculated. Finally, averages of the mean and standard deviation data are
calculated, $\bar{\mu}$ and $\bar{\sigma}$, respectively. The thresholds,
$thr$ are then calculated using,
$thr=\bar{\mu}\pm 3\bar{\sigma}$ (25)
The residual errors for each blade are then compared to the upper and lower
thresholds for each blade. Once the residual errors exceed these thresholds,
the blade is assumed to be damaged.
### 3.2 Example 1: Synthesised data
In this example, a demonstration of the algorithm is described using some
synthesised data. It must be noted that due to confidentiality agreements in
effect, signals simulating the actual blade physics (and subsequent frequency
estimations of the blades) cannot be reported. Instead, the edge frequency
data of one blade is synthesised using a non-zero mean normal distribution.
The edge frequency data of the second blade is set to be equal to that of the
first blade with some additional noise. Finally, the third blade initially has
the same edge frequencies of the first blade, but ‘deteriorates’ over time by
means of a negative linear gradient. That is,
$f_{A}=\mathcal{N}(\mu_{A},\sigma^{2})$ (26)
$f_{B}=f_{A}+\mathcal{N}(0,\sigma^{2})$ (27)
$f_{C}=f_{A}+d_{C}+\mathcal{N}(0,\sigma^{2})$ (28)
where $d_{C}$ is a vector of zeros followed by a line with a negative
gradient. That is,
$d_{C}=\\{\boldsymbol{0},m\boldsymbol{x}+c\\}$ (29)
In this work, 360000 data points were generated. Only the first half of the
data points of Blade C are considered ‘healthy’. The length of the vector of
zeros in equation (29) is therefore 180000. The rest of the vector $d_{C}$,
which is indexed by the data point number, $\boldsymbol{x}$ (i.e. 180001 to
360000) has a gradient, $m$, of $-1\times 10^{-7}$ and a y-intercept value,
$c$ of 0.018. Whilst not realistic, $\mu_{A}$ is set to 10 and $\sigma^{2}$
has a value of $0.01$.
Figure 7 illustrates the synthesised edge frequencies of the three blades.
(a)
(b)
(c)
Figure 7: Synthesised edge frequency data for: (a) Blade A, (b) Blade B, and
(c) Blade C
Referring to Section 2.6, the above description explains the data acquisition
and feature selection processes. That is, the data acquired is As outliers do
not exist on the synthesised data,t covers the first three
The methodology, as described in Section 2, is applied. Figure 8 illustrates
the residual errors between the actual and predicted frequencies. Note that
due to data normalisation (which is required for numerical stability, as
discussed in Section 2), the residual errors are scaled. It is clear that the
‘damage’ is identified as the residual errors of blades A and C diverge on the
midway point. The residuals of blade B do not diverge since they do not rely
on data from the damaged blade C (actual or predicted). Note that in reality,
the dynamics of healthy blades may also be affected by the presence of a
damaged blade, and hence it may not always be possible to conclude which blade
contains damage.
(a)
(b)
(c)
Figure 8: Residual errors between synthesised and predicted data for: (a)
Blade A, (b) Blade B, and (c) Blade C
### 3.3 Example 2: Real turbine data
The correlation plots of this turbine were illustrated in Figures 2 and 3. In
this example, one of the blades on the turbine was found to have some form of
damage midway through Year 7 of turbine operation. Figure 9 illustrates the
comparisons between the actual frequency estimations and the GP predictions
for each blade. Note that on the following figures, the period between the
vertical red lines illustrate the training period for the GPs, the period
between the second vertical red line and the vertical green line illustrate
the period the $3\sigma$ thresholds were calculated, and the vertical black
lines indicate the date the damage was identified.
(a)
(b)
(c)
Figure 9: Comparison between actual estimated normalised edge frequencies and
GP predictions for: (a) Blade A, (b) Blade B, and (c) Blade C in Site A.
Figures 10 (a) - (c) illustrate the residual errors corresponding to the
results shown in Figures 9 (a) - (c), respectively.
(a)
(b)
(c)
Figure 10: Residual errors between actual estimated normalised edge
frequencies and GP predictions for: (a) Blade A, (b) Blade B, and (c) Blade C
in Site A.
From Figures 9 and 10, it is clear that the GP predictions become less
accurate as the date of damage approaches. As damage levels increase, and the
properties of the blades change, the inputs to the GPs become geometrically
further away than the inputs used during the training phase. Since machine
learning methods in general, and GPs in particular, should not be used for
extrapolation, the predictions become poorer, and hence, the residual errors
grow. On closer inspection of Figure 10, it can be seen that just prior to the
sharp increases in the GP residuals, there are small mean shifts that occur.
The corresponding X-bar control charts for Site A are shown in Figure 11. The
averages were completed once every four weeks.
(a)
(b)
(c)
Figure 11: X-bar control charts of GP residuals for: (a) Blade A, (b) Blade B,
and (c) Blade C in Site A.
The horizontal black lines indicate the calculated thresholds of the X-bar
control charts. On analysing these charts in Figure 11, it can be seen that
the GP residuals begin to indicate that the blade properties have changed
roughly 6 months before the blade was remedied/replaced.
It is noteworthy that following the identification of damage, and the
corresponding remedial action, there is a break down in the correlation
between the pairs of blades. This is because the mechanical properties of the
repaired/replaced blade are now different. Hence, whilst the residuals
approach the original thresholds, they do not settle within them. This fact
indicates that every time a remedial activity takes place, retraining is
essential.
### 3.4 Example 3: Real turbine data
The correlation plots for this turbine are shown in Figure 12. Note that the
temperature scales shown in these plots are normalised. In this example, all
three correlation plots are noisy, although there is the expected positive
correlation. The temperature effects are noticeable - at lower temperatures,
when the blades are stiffer, the edge frequencies are higher, whereas at
higher temperatures, the opposite is true.
(a)
(b)
(c)
Figure 12: Correlation between normalised edge frequencies for: (a) Blade A
vs. Blade B, (b) Blade B vs. Blade C, and (c) Blade C vs. Blade A in Site B.
Figure 12 illustrates how the GP predictions vary with the input edge
frequencies and temperature. It is interesting to note how the GPs have
identified a hidden structure within the noisy data - there are various bands
of predictions across the range of temperatures. These bands are both linear
(captured by the linear kernel) and nonlinear (captured by the squared-
exponential kernel).
(a)
(b)
(c)
Figure 13: Correlation between normalised edge frequencies for: (a) Blade A
vs. Blade B, (b) Blade B vs. Blade C, and (c) Blade C vs. Blade A in Site B.
Figures 14 and 15 compare the estimated and predicted edge frequencies. Once
again, the seasonal variations are captured well. However, the amplitudes of
the residual errors are high compared to those seen in Example 2, which is
expected given the noisy estimates and correlations. The residuals in Figure
15 (a) and (b) indicate that an event occurred that changed the structural
properties, and correlations between the blades around Year 4 as shown by the
change in structure of the residuals. As the damage event approaches, there is
a monotonic change in the residuals in Figure 15.
(a)
(b)
(c)
Figure 14: Comparison between actual estimated normalised edge frequencies,
and GP predictions for: (a) Blade A, (b) Blade B, and (c) Blade C in Site B.
(a)
(b)
(c)
Figure 15: Residual errors between actual estimated normalised edge
frequencies, and GP predictions for: (a) Blade A, (b) Blade B, and (c) Blade C
in Site B.
Figure 16 shows the results from this final case study. Once again, the
3$\sigma$ thresholds were exceeded in advance, this time only 3 months before
the remedial activities to fix the blade took place. Furthermore, the
amplitude of the threshold exceedance is not as pronounced as that seen in
Example 2. This is primarily attributed to the noisy features used.
(a)
(b)
(c)
Figure 16: X-bar control charts of GP residuals for: (a) Blade A, (b) Blade B,
and (c) Blade C in Site B.
## 4 Conclusions
In this article, a new diagnostic methodology for operational wind turbine
blades has been proposed. Gaussian Processes (GPs) are used to predict the
edge frequencies of one blade, given the edge frequencies of another blade and
the ambient temperature. This may be viewed as another way of applying the
Johansen procedure to cointegrate variables, as shown in [25]. The premise is
that as long as no damage has taken place on the blade, the relationship
between the edge frequencies of the respective blades should remain
consistent. However, when damage does take place, the predictive capabilities
of the GPs would be affected, and the predictions would not be accurate.
Hence, when damage takes place, the GP residuals (the difference between the
actual and predicted edge frequencies) would grow in amplitude, and would
hence act as informative damage indicators. X-bar control charts, along with
3$\sigma$ thresholds, were used to indicate whether the blades were healthy or
not.
A case study using synthesised data was used to illustrate the algorithm. Two
other case studies from real turbines, where one of the blades was reported to
sustain damage, were also shown to illustrate the diagnostic capabilities of
the methodology proposed in this article. The results showed the successful
implementation of this method whereby the respective damages were detected in
advance of critical failure.
The key outcomes of this work are:
* 1.
Although the individual edge frequency time series of the blades are
nonstationary, there exists a common and explicable trend between them, which
can be learned.
* 2.
Due to the noisy estimation of the edge frequencies, there may not be a one-
to-one mapping between the edge-frequency estimates of the pairs of blades.
However, as long as the nominal trend of this correlation is captured by the
GPs, the residual errors give sufficient information regarding the onset of
damage as long as these frequencies being monitored are sensitive to the
damage type and location.
* 3.
The addition of ambient temperature as a feature helps in the predictive
capabilities of the GPs, especially when the edge-frequency estimates are
noisy.
* 4.
The methodology of identifying correlations between two (or more) blades may
be applied to different types of signals (for example, strains and
displacements), although such signals may require additional pre-processing
and other complementary features.
## Acknowledgements
The authors would like to thank Siemens Gamesa Renewable Energy for funding
this research; without their input, this project would not have been possible.
The authors would also like to acknowledge the support of the following EPSRC
grants: EP/R004900/1, EP/S001565/1, and EP/R003645/1.
## References
* [1] F. P. G. Márquez, A. M. Tobias, J. M. P. Pérez, M. Papaelias, Condition monitoring of wind turbines: Techniques and methods, Renewable Energy 46 (2012) 169–178.
* [2] D. Li, S.-C. M. Ho, G. Song, L. Ren, H. Li, A review of damage detection methods for wind turbine blades, Smart Materials and Structures 24 (3) (2015) 033001.
* [3] C. C. Ciang, J.-R. Lee, H.-J. Bang, Structural health monitoring for a wind turbine system: a review of damage detection methods, Measurement Science and Technology 19 (12) (2008) 122001.
* [4] B. Yang, D. Sun, Testing, inspecting and monitoring technologies for wind turbine blades: A survey, Renewable and Sustainable Energy Reviews 22 (2013) 515–526.
* [5] R. Raišutis, E. Jasiūnienė, R. Šliteris, A. Vladišauskas, The review of non-destructive testing techniques suitable for inspection of the wind turbine blades, Ultragarsas 63 (2) (2008) 26–30.
* [6] J. Kaldellis, M. Kapsali, Shifting towards offshore wind energy - Recent activity and future development, Energy Policy 53 (2013) 136–148.
* [7] C. Röckmann, S. Lagerveld, J. Stavenuiter, Operation and maintenance costs of offshore wind farms and potential multi-use platforms in the dutch north sea, Aquaculture Perspective of Multi-Use Sites in the Open Ocean (2017) 97–113.
* [8] M. A. Rumsey, J. A. Paquette, Structural health monitoring of wind turbine blades, in: Smart Sensor Phenomena, Technology, Networks, and Systems 2008, Vol. 6933, International Society for Optics and Photonics, 2008\.
* [9] J. Sierra-Pérez, M. A. Torres-Arredondo, A. Güemes, Damage and nonlinearities detection in wind turbine blades based on strain field pattern recognition. FBGs, OBR and strain gauges comparison, Composite Structures 135 (2016) 156–166.
* [10] A. Ghoshal, M. J. Sundaresan, M. J. Schulz, P. F. Pai, Structural health monitoring techniques for wind turbine blades, Journal of Wind Engineering and Industrial Aerodynamics 85 (3) (2000) 309–324.
* [11] D. Adams, J. White, M. Rumsey, C. Farrar, Structural health monitoring of wind turbines: method and application to a HAWT, Wind Energy 14 (4) (2011) 603–623.
* [12] J.-R. Lee, H.-J. Shin, C. C. Chia, D. Dhital, D.-J. Yoon, Y.-H. Huh, Long distance laser ultrasonic propagation imaging system for damage visualization, Optics and Lasers in Engineering 49 (12) (2011) 1361–1371.
* [13] J. Tang, S. Soua, C. Mares, T.-H. Gan, An experimental study of acoustic emission methodology for in service condition monitoring of wind turbine blades, Renewable Energy 99 (2016) 170–179.
* [14] N. Dervilis, M. Choi, S. Taylor, R. Barthorpe, G. Park, C. Farrar, K. Worden, On damage diagnosis for a wind turbine blade using pattern recognition, Journal of Sound and Vibration 333 (6) (2014) 1833–1850.
* [15] S. G. Taylor, K. Farinholt, M. Choi, H. Jeong, J. Jang, G. Park, J.-R. Lee, M. D. Todd, Incipient crack detection in a composite wind turbine rotor blade, Journal of Intelligent Material Systems and Structures 25 (5) (2014) 613–620.
* [16] C. Galleguillos, A. Zorrilla, A. Jimenez, L. Diaz, A. Montiano, M. Barroso, A. Viguria, F. Lasagni, Thermographic non-destructive inspection of wind turbine blades using unmanned aerial systems, Plastics, Rubber and Composites 44 (3) (2015) 98–103.
* [17] C. Niezrecki, P. Avitabile, J. Chen, J. Sherwood, T. Lundstrom, B. LeBlanc, S. Hughes, M. Desmond, A. Beattie, M. Rumsey, et al., Inspection and monitoring of wind turbine blade-embedded wave defects during fatigue testing, Structural Health Monitoring 13 (6) (2014) 629–643.
* [18] M. Rumsey, J. Paquette, J. White, R. Werlink, A. Beattie, C. Pitchford, J. van Dam, Experimental results of structural health monitoring of wind turbine blades, in: 46th AIAA Aerospace Sciences Meeting and Exhibit, 2008.
* [19] J. Paquette, J. van Dam, S. Hughes, J. Johnson, Fatigue testing of 9 m carbon fiber wind turbine research blades, in: 46th AIAA Aerospace Sciences Meeting and Exhibit, 2008.
* [20] W. Yang, Z. Lang, W. Tian, Condition monitoring and damage location of wind turbine blades by frequency response transmissibility analysis, IEEE Transactions on Industrial Electronics 62 (10) (2015) 6558–6564.
* [21] K.-S. Wang, V. S. Sharma, Z.-Y. Zhang, SCADA data based condition monitoring of wind turbines, Advances in Manufacturing 2 (1) (2014) 61–69.
* [22] C. E. Rasmussen, C. K. I. Williams, Gaussian Processes in Machine Learning, The MIT Press, 2006.
* [23] K. P. Murphy, Machine Learning: A Probabilistic Perspective, MIT Press, 2012.
* [24] J. A. Nelder, R. Mead, A simplex method for function minimization, The computer journal 7 (4) (1965) 308–313.
* [25] E. J. Cross, K. Worden, Q. Chen, Cointegration: a novel approach for the removal of environmental trends in structural health monitoring data, in: Proceedings of the Royal Society of London A: Mathematical, Physical and Engineering Sciences, The Royal Society, 2011.
|
11institutetext: Ning Ge 22institutetext: School of Software, Beihang
University, Beijing, China
22email<EMAIL_ADDRESS>
_She was also with the Key Laboratory of Safety-Critical Software (Nanjing
University of Aeronautics and Astronautics), Ministry of Industry and
Information Technology._ 33institutetext: Guanghao Li 44institutetext: School
of Software, Beihang University, Beijing, China
44email<EMAIL_ADDRESS>55institutetext: Li Zhang 66institutetext:
School of Computer Science and Engineering, Beihang University, Beijing, China
66email<EMAIL_ADDRESS>
_Li Zhang is the corresponding author._ 77institutetext: Yi
Liu88institutetext: School of Computer Science and Engineering, Beihang
University, Beijing, China
88email<EMAIL_ADDRESS>
# Failure Prediction in Production Line Based on Federated Learning: An
Empirical Study
Ning Ge Guanghao Li Li Zhang Yi Liu
(Received: date / Accepted: date)
###### Abstract
Data protection across organizations is limiting the application of
centralized learning (CL) techniques. Federated learning (FL) enables multiple
participants to build a learning model without sharing data. Nevertheless,
there is very few research works on FL in intelligent manufacturing. This
paper presents the results of an empirical study on failure prediction in the
production line based on FL. This paper (1) designs Federated Support Vector
Machine (FedSVM) and Federated Random Forest (FedRF) algorithms for the
horizontal FL and vertical FL scenarios, respectively; (2) proposes an
experiment process for evaluating the effectiveness between the FL and CL
algorithms; (3) finds that the performance of FL and CL are not significantly
different on the global testing data, on the random partial testing data, and
on the estimated unknown Bosch data, respectively. The fact that the testing
data is heterogeneous enhances our findings. Our study reveals that FL can
replace CL for failure prediction.
###### Keywords:
Empirical study Federated learning Failure prediction Production line
Manufacturing Bosch dataset
## 1 Introduction
Artificial intelligence (AI) is one of the core techniques in the fourth
industrial revolution Zhong et al. (2017); Li et al. (2017). For instance, AI
techniques have been employed to improve the early detection or fault
prediction within production lines Tao et al. (2018); Kusiak (2017) in
intelligent manufacturing (IM). A prominent achievement is that in 2012, Intel
saved $3 million in manufacturing costs through the use of predictive
analytics to prioritize its silicon chip inspections Ronen and Burns . To this
end, the Bosch company has published a dataset of product quality prediction
on the data analysis competition platform
Kaggle111https://www.kaggle.com/c/bosch-production-line-performance. The
dataset reflects the relevant parameters and equipment operation of each
product during the production process, hoping to reduce defective products. By
using this dataset, several studies have proposed product quality prediction
methods based on centralized learning (CL) algorithms Carbery et al. (2019,
2018); Zhang et al. (2016); Khoza and Grobler (2019); Kotenko et al. (2019);
Mangal and Kumar (2016); Hebert (2016); Maurya (2016); Huang et al. (2019b);
Moldovan et al. (2019); Liu et al. (2020d). The results of these work revealed
two important facts. First, data preprocessing on the Bosch dataset has a
significant impact on the prediction result. Second, among the CL algorithms,
the Random Forest algorithm performed better than others when time-series
features are excluded Zhang et al. (2016); Khoza and Grobler (2019) and the
Long Short Term Memory (LSTM) network can improve the prediction result when
time-series features are included Huang et al. (2019b). CL methods usually
require that clients upload data to the server. The server trains an
integrated model based on the shared data. However, in real-world scenarios,
private data belongs to multiple independent organizations and cannot be
shared with others. Therefore, data sharing has arisen to be an obstacle to
the application of CL methods. As UC Berkeley pointed out in a report
published in 2017, learning from private data is one of the challenges
encountered by the current AI field Stoica et al. (2017).
In response to the above problem, Google proposed the concept of federated
learning (FL) in 2016 to decentralize the collaborative learning McMahan et
al. (2016). FL trains models across multiple decentralized devices holding
local private data samples, without exchanging or sharing these samples.
Clients then encrypt and transmit the trained model or parameters back to the
server to build an integrated model. On the one hand, FL addresses critical
issues such as data privacy and data security Yang et al. (2019a); Li et al.
(2020a). On the other hand, as FL trains model locally without removing the
private data features, the data samples can be more complete than CL. FL has
received widespread attention recently and has been applied in many fields.
For example, Google launched a project on the prediction of mobile keyboard
input in 2018 and obtained better prediction results using FL Hard et al.
(2018). Intel started to support the hardware architecture for
FL222https://www.intel.com/content/www/us/en/artificial-
intelligence/posts/federated-learning-for-medical-imaging.html. FL has also
been applied to various fields like medical systems Sheller et al. (2018b);
Brisimi et al. (2018); Boughorbel et al. (2019b); Huang et al. (2019a),
Internet finance Suzumura et al. (2019); Yang et al. (2019b), smart city
Samarakoon et al. (2019); Saputra et al. (2019); Liu et al. (2020b), edge
computing Bakopoulou et al. (2019), IOT Nguyen et al. (2019); Chen et al.
(2020), Cyber Physical Systems (CPS) Mowla et al. (2019); Aussel et al.
(2020), etc.
In the field of IM, the manufacturing process often crosses enterprises,
workshops, or even production lines. Due to data privacy and security, the
application of CL is becoming the bottleneck. The studies on the failure
prediction in the production line have been conducted under the assumption
that all data are shared. Under the premise of data protection between
production lines, these methods become unfeasible. Nevertheless, according to
our investigation Liu et al. (2020c), applying FL to the field of IM is still
in its infancy, and there are very few research works on it. Although FL has
achieved success in many fields, whether it can achieve similar performance
and replace CL in real-world manufacturing is still an open question.
Accordingly, this paper bridges this gap by reporting an empirical study of
defective product prediction in the production line by comparing FL with CL.
This study is the first attempt to design FL algorithms for failure prediction
in the production line.
In addressing our research goal, we construct a horizontal FL (HFL) scenario
and a vertical FL (VFL) scenario, respectively. In the former, we compared the
Federated Support Vector Machine (FedSVM) and the Support Vector Machine (SVM)
algorithms. In the latter, we compare the Federated Random Forest (FedRF) and
the Random Forest (RF) algorithms. Each group of contrast experiment focuses
on four research questions (RQs).
* $\circ$
_RQ1_ , we are interested in whether FL can replace CL on the whole Bosch
testing data. Hence, we ask whether the average performance of FedSVM (FedRF
resp.) is similar to that of SVM (RF resp.) on the whole Bosch testing data?
* $\circ$
_RQ2_ , we are interested in whether FL can replace CL on part of the given
Bosch testing data. Therefore, we ask whether the average performance of
FedSVM (FedRF resp.) is similar to that of SVM (RF resp.) on random partial
Bosch testing data?
* $\circ$
_RQ3_ , if FL is applied to predicting defective products on other unknown
Bosch data, whether FL can replace CL. Hence, we ask whether the average
performance of FedSVM (FedRF resp.) is similar to that of SVM (RF resp.) on
the estimated unknown Bosch testing data?
* $\circ$
_RQ4_ , we want to know whether the Bosch testing data is heterogeneous or
not. If the data is heterogeneous and the answers of RQ1-RQ3 are positive, it
means that the selected FL algorithms are robust. Therefore, we ask, is there
local heterogeneity within the given testing data?
Answers to these questions are helpful for manufacturing as they provide
insights into how to evaluate an FL algorithm can achieve similar effects as a
CL one in a real-world application, and then replace it. To answer these
questions, this paper makes methodological, substantive, and theoretical
contributions to the literature on failure prediction in the production line.
First, in terms of algorithm design, to conduct empirical research on the
failure prediction based on HFL and VFL, we first select SVM and RF from
previous works as CL baseline because both algorithms performed better than
others in the previous works. Then, we design FedSVM and FedRF algorithms for
our problem based on existing works Bakopoulou et al. (2019); Liu et al.
(2020a). In terms of novelty in the algorithm design, our work improved FedRF
by introducing optimal feature selection and pruning steps.
Second, we propose a set of experimental methodology to compare the average
performance of FL and CL in manufacturing from three aspects: on the whole
testing data, on the random partial testing data, and on the estimated unknown
testing data. To compare the performance of FL and CL on the estimated unknown
Bosch data, we follow the process of Measurement Systems Analysis (MSA) to fit
a Markov process model of prediction error $\mathtt{M_{F}}$ ($\mathtt{M_{C}}$
resp.) based on FL (CL resp.) and then compare the two fitting models. The
experimental methods also assess the heterogeneity of the testing data to
enhance the results of RQ1-RQ3 on whether FL can replace CL.
Last, our empirical research shows that our FL solution is not significantly
different from the CL solution for failure prediction, on the whole, on the
random partial, and on estimated unknown Bosch testing data. FL can replace CL
in the application of failure prediction within the manufacturing process.
In the remainder of this paper, we first introduce the background knowledge of
our empirical research and related work in Section 2, present the designed HFL
and VFL algorithms in Section 3, and outline the experimental methodology in
Section 4. Section 5 presents the results of our empirical study. Section 6
discusses the threat of validity. Section 7 discusses whether FL can replace
CL. Section 8 concludes this paper and looks forward to future work.
## 2 Background and Related Work
This section introduces previous works on Bosch product quality prediction
based on the CL algorithms and discusses the necessity of conducting empirical
research on the FL algorithms for this problem; and then, briefly explains the
principles of federated learning.
### 2.1 Failure Prediction in Bosch Production Line
Bosch is one of the worldwide leading manufacturing companies. It ensures high
quality of the production by monitoring its parts in the manufacturing
processes. Because Bosch records detailed data for each step on the assembly
lines, they can apply advanced techniques to improve the manufacturing
processes. To this end, Bosch has published a dataset on the Kaggle
competition platform to predict internal failures by thousands of measurements
and tests made for each component along the assembly line. Some studies have
analyzed the dataset and carried out approaches to predicting product quality
based on CL algorithms excluding time-series features Carbery et al. (2019,
2018); Zhang et al. (2016); Khoza and Grobler (2019); Kotenko et al. (2019);
Mangal and Kumar (2016); Hebert (2016); Maurya (2016) or including time-series
features Huang et al. (2019b); Moldovan et al. (2019); Liu et al. (2020d), as
summarized in Table 1. These works were conducted based on a set of learning
methods, including Logistic Regression (LR), Gradient Boosting Machine (GBM),
Random Forest (RF), Gradient Boosted Trees (GBT), Naive Bayes (NB), Bayesian
Network (BN), K-Nearest Neighbors (KNN), Support Vector Machines (SVM),
Multilayer Perceptron Classifier (MPC), Majority Voting (MV), Decision Tree
(DT), Statistical Process Control (SPC), etc.
Table 1: Research Works on Failure Prediction in Bosch Production Line Objectives | Ref | Learning Method | Contribution
---|---|---|---
Improving predictive model without time-series features | Carbery et al. (2018) | XGBoost, BN | BN model performs well in failure prediction.
Zhang et al. (2016) | RF, Gradient Boosting, LR, NB, DT | RF performs better on different clusters.
Khoza and Grobler (2019) | RF, SVM, NB, SPC | RF outperforms other models.
Kotenko et al. (2019) | SVM, KNN, Perceptron, LR, DT, MV | SVM and MV outperform other models.
Mangal and Kumar (2016) | LR, Extra Trees Classifier, RF, XGBoost | XGBoost and RF outperform other methods.
Hebert (2016) | LR, RF, XGBoost | RF and XGBoost can properly identify conditions leading to failure events.
Maurya (2016) | XGBoost | Optimize MCC by using GBM as a base classifier.
Improving predictive model with time-series features | Huang et al. (2019b) | Ontology-based LSTM neural network | Ontology-based LSTM neural network yields a better performance
Moldovan et al. (2019) | RF, GBT, NB, KNN, SVM, and MPC | The LSTM RNN model outperform others.
Liu et al. (2020d) | SP-LSTM models | The A-Bi-SP-LSTM model outperforms other models.
Carbery et al. Carbery et al. (2019) conducted a systematic feature analysis
on the Bosch dataset and used BN to predict product quality Carbery et al.
(2018). Zhang et al. Zhang et al. (2016) weakened data heterogeneity through
clustering, and then applied RF, Boosting, LR, NB, DT to each cluster. They
showed that RF performed better than other algorithms. Based on Zhang et al.
(2016), Khoza et al. Khoza and Grobler (2019) compared RF, NB, SVM, and SPC.
They also showed that RF performed better. Kotenko et al. Kotenko et al.
(2019) used SVM, KNN, LR, Perceptron, DT, and MV and showed that SVM achieved
relatively higher prediction accuracy. Mangal et al. Mangal and Kumar (2016)
conducted a visual analysis of the three types of data: categorical, numeric,
and time-series features on the Bosch dataset, and used LR, Extra Trees
Classifier, RF, XGBoost for quality prediction. Hebert Hebert (2016) found
that RF and XGBoost could properly identify conditions that lead to failure
events. Maurya Maurya (2016) optimized MCC measure by using GBM as a base
classifier.
The studies Moldovan et al. (2019); Huang et al. (2019b); Liu et al. (2020d)
have considered time-series features. Huang et al. Huang et al. (2019b)
constructed an LSTM network model based on the time-series features of the
data, which has great enlightening significance for other researchers.
Moldovan et al. Moldovan et al. (2019) found that the LSTM RNN model
outperformed other machine learning models. Liu et al. Liu et al. (2020d)
proposed an end-to-end unified quality prediction framework to capture
temporal interactions among the features of different processes. They showed
that the A-Bi-SP-LSTM model outperforms the existing data-driven methods.
Despite these studies on learning features for failure prediction, they have
been all conducted under the assumption that all data are shared. Under the
premise of data protection between production lines, these methods become
unpractical. An open question is whether there exist some FL algorithms that
can replace CL methods. Thus, our study is a first attempt to design FL
algorithms for failure prediction in the assembly line.
### 2.2 Overview of Federated Learning
According to the relationship between the datasets provided by the clients, we
usually classify FL into HFL, VFL, and federated transfer learning (TFL)
Kairouz et al. (2019). This article focuses on the HFL and VFL scenarios. In
HFL, data are partitioned by user IDs or device IDs. For example, in the
context of manufacturing, clients A and B are two independent factories having
the same production line structure and machine configuration. The data
features in each production line are roughly the same, while the product IDs
on their production lines do not overlap. VFL is applicable to the cases that
two datasets share the same sample ID space but differ in feature space. For
example, production lines A and B are independent, but one belongs to the
upstream, and the other belongs to the downstream of the entire production
line. Their product IDs are likely to be the same, which guarantees the
intersection of their product space. However, since both A and B record part
of manufacturing behavior, their feature spaces are very different. In this
paper, we design FedSVM as the HFL model, and design FedRF as the VFL model,
respectively, to compare with centralized SVM and RF models.
## 3 Federated SVM and Federated RF
This section introduces the design of the federation learning algorithms in
this work. Through the investigation in Sect. 2, we selected SVM and RF
models, which performed well on the Bosch dataset, as the baseline. In the HFL
scenario, we reuse the existing FedSVM algorithm to compare it with SVM. In
the VFL scenario, we improve the existing Federated Forest algorithm Liu et
al. (2020a) to compare with RF.
### 3.1 Federated SVM
As a supervised learning algorithm, SVM is suitable for classification and
regression analysis. The effectiveness of SVM mainly depends on how to select
1) the kernel, 2) the kernel’s parameters, and 3) the soft margin parameter.
In this work, we use FedSVM with a linear kernel. SVM specifies parameters and
intercepts, which can be directly weighted. In FL, the SVM models generated on
different clients can be integrated by averaging the parameters and intercepts
to meet the need of the server. The FedSVM algorithm is first proposed in
Bakopoulou et al. (2019) for mobile packet classification. We have adapted
their FedSVM to our failure prediction problem. Its training process is shown
in Fig. 1. The algorithm of the client and server is given by Algo. 1 and
Algo. 2, respectively. The main steps of FedSVM are explained hereafter.
Figure 1: The Training Process of FedSVM
* $\circ$
S1. Parameter initialization: The server ($S$) notifies each client ($C_{i}$)
to train models using the local data with random value of parameters and
intercepts ($w_{0}$).
* $\circ$
S2. Local training: $C_{i}$ trains the local SVM model independently.
* $\circ$
S3. Return local SVM parameters: $C_{i}$ returns locally trained parameters
and intercepts ($w^{i}_{t}$) in this round to $S$.
* $\circ$
S4. Compute global SVM parameters: $S$ averages the trained values in two
successive round $(\sum w^{i}_{t}+w_{t-1})/2$ as the value $w_{t}$ in the
global model $\mathtt{M_{G}}$.
* $\circ$
S5. Send global SVM parameter: $S$ sends the global value $w_{t}$ in the
$\mathtt{M_{G}}$ to $C_{i}$ to train the model of next round.
* $\circ$
S6. Repeat until convergence: Repeat steps S2 - S5 after a specified number of
iterations until convergence. The global model in the last round is sent to
all clients as the training result.
Algorithm 1 FedSVM - Client
1:Data set $D_{i}$ on client $i\ (i=1\ to\ k)$
2:Local SVM model on client $i$
3:Initialize $w_{0}^{i}$
4:$w_{1}^{i}\leftarrow$ SVM($w_{0}^{i}$, $D_{i}$) $\triangleright$ S2
5:Send $w_{1}^{i}$ to server $\triangleright$ S3
6:for $epoch\ t=2\ to\ N$ do $\triangleright$ S6
7: Receive global model $w_{t-1}$ $\triangleright$ S5
8: $w_{t}^{i}\leftarrow$ SVM($w_{t-1}$, $D_{i}$) $\triangleright$ S2
9: Send $w_{t}^{i}$ to server $\triangleright$ S3
10:end for
Algorithm 2 FedSVM - Server
1:Local SVM model
2:Global SVM model
3:Inform all clients start training $\triangleright$ S1
4:Receive $w_{1}^{i}$ from all clients
5:$w_{1}\leftarrow\sum_{i=1}^{k}\frac{w_{1}^{i}}{k}$ $\triangleright$ S4
6:Send $w_{1}$ to all clients $\triangleright$ S5
7:for $epoch\ t=2\ to\ N$ do $\triangleright$ S6
8: Receive $w_{t}^{i}$ from all clients
9: $w_{t}\leftarrow\frac{1}{2}(w_{t-1}+\sum_{i=1}^{k}\frac{w_{t}^{i}}{k})$
$\triangleright$ S4
10: Send $w_{t}$ to all clients $\triangleright$ S5
11:end for
### 3.2 Federated Random Forest
The Federated Forest algorithm was first proposed by Liu et al. in 2020 Liu et
al. (2020a). The random forest can be regarded as an integrated implementation
of the decision tree. The core steps of the decision tree (calculating the
purity of each feature, represented by Gini coefficients) only involve all
data in a single feature and the classification label. Each client in VFL can
do this on its local data. In our work, we have optimized the federated random
forest algorithm in Liu et al. (2020a) to get better accuracy and efficiency
by improving the steps of optimal feature selection and pruning. The improved
method is illustrated hereafter.
Optimal feature selection: The optimal feature is the one with the smallest
Gini coefficient according to the CART method in the decision tree. For
continuous variables, it is necessary to try to divide all possible values
when calculating the Gini coefficient. Here we reduce the calculation steps of
this function and improve the efficiency of the algorithm.
Pruning: There is no explicit pruning process in Liu et al. (2020a). Here we
add a pre-pruning function, which enhances robustness and improves the
efficiency of the overall algorithm. When the decision tree finds that the new
node will not improve the accuracy, it will change this node to a leaf node.
The training process of our improved FedRF algorithm is given in Fig. 2. The
algorithm of client and server is given by Algo. 3 and Algo. 4, respectively.
The main steps in its training process are explained hereafter.
Figure 2: The Training Process of FedRF
* $\circ$
S1. Select initial information: The server ($S$) randomly selects the feature
subset $\mathtt{F^{\prime}}$, the sample ID subset $\mathtt{D^{\prime}}$ and
test set $\mathtt{T}$, and notifies them to each client ($C_{i}$).
* $\circ$
S2. Local training and pruning: $C_{i}$ knows which features have been
selected, and does not know the number of features selected globally. $C_{i}$
calculates the Gini coefficient of each feature in $\mathtt{F^{\prime}}$.
Because the feature is only on the client, the client also knows the
classification labels of all samples, so the Gini coefficient calculated by
the client represents the global Gini coefficient. $C_{i}$ obtains the optimal
division feature and corresponding Gini coefficient, and retain the division
threshold of this feature. $C_{i}$ also checks whether the accuracy of the
decision tree is improved on this optimal feature partition, and if there is
no improvement, it judges that pre-pruning is needed.
* $\circ$
S3. Upload local Gini coefficient: $C_{i}$ uploads the Gini coefficient and
pre-pruning information to $S$.
* $\circ$
S4. Select global optimal features: $S$ selects the globally optimal feature.
* $\circ$
S5. Send optimal global features: If pruning is not required, $S$ notifies the
corresponding client to return the ID division result; otherwise, $S$ notifies
all clients to perform pruning processing to form a leaf node.
* $\circ$
S6. Send local decision: When pruning is not required, the notified client
returns the result of dividing the dataset with this feature (the ID set that
falls into the left and right subtrees).
* $\circ$
S7. Broadcast global decision: $S$ notifies other clients of the ID division
and forms the first node of the decision tree. $C_{i}$ knows the results of
each ID division, but only knows the existence of local features in the global
tree. $S$ knows the structure of the entire tree and the corresponding feature
names of each node. $S$ and $C_{i}$ recursively establish new nodes and their
left and right subtrees according to the current $\mathtt{F^{\prime}}$ and
$\mathtt{D^{\prime}}$ until they form leaf nodes.
* $\circ$
S8. Repeat until N tree built: Iteratively build N decision trees to form a
random forest.
Algorithm 3 FedRF - Client
1:Data set $D$, feature set $F_{i}$ on client $i\ (i=1\ to\ k)$
2:Local FedRF model on $Ci$
3:for $epoch\ t=1\ to\ N$ do
4: Receive $F_{i}^{{}^{\prime}}\subseteq F_{i}$, $D^{{}^{\prime}}\subseteq D$,
$T=D-D^{{}^{\prime}}$ from $S$
5: function TreeBuild($D^{{}^{\prime}},F_{i}^{{}^{\prime}},T$)
6: if all samples in $D^{{}^{\prime}}$ have the same label then
7: return $leaf\ node$
8: end if
9: $BestF_{i},GiniPara_{i},threshold_{i}\leftarrow$
Gini($D^{{}^{\prime}},F_{i}^{{}^{\prime}}$)
10: if $BestF_{i}$ cannot improve the accuracy on $T$ then
11: Send $needPruning_{i}$ to $S$ $\triangleright$ S2
12: end if
13: Send $BestF_{i},GiniPara_{i}$ to $S$ $\triangleright$ S3
14: if Receive ${}^{\prime}Prune^{\prime}$ message from $S$ then
15: return $leaf\ node$
16: else if Receive ${}^{\prime}Success^{\prime}$ message from $S$ then
17: Split $D^{{}^{\prime}}$ with $BestF_{i}$, get
$D_{l}^{{}^{\prime}},D_{r}^{{}^{\prime}},T_{l},T_{r}$
18: Send $D_{l}^{{}^{\prime}},D_{r}^{{}^{\prime}},T_{l},T_{r}$ to $S$
$\triangleright$ S6
19: leftTree $\leftarrow$
TreeBuild($D_{l}^{{}^{\prime}},F_{i}^{{}^{\prime}},T_{l}$)
20: rightTree $\leftarrow$
TreeBuild($D_{r}^{{}^{\prime}},F_{i}^{{}^{\prime}},T_{r}$) $\triangleright$ S7
21: else
22: Receive $D_{l}^{{}^{\prime}},D_{r}^{{}^{\prime}},T_{l},T_{r}$ from $S$
23: leftTree $\leftarrow$
TreeBuild($D_{l}^{{}^{\prime}},F_{i}^{{}^{\prime}},T_{l}$)
24: rightTree $\leftarrow$
TreeBuild($D_{r}^{{}^{\prime}},F_{i}^{{}^{\prime}},T_{r}$) $\triangleright$ S7
25: end if
26: return tree node
27: end function
28: Add this decision tree to forest
29:end for$\triangleright$ S8
Algorithm 4 FedRF - Server
1:Sample ID, feature names on all clients
2:Global FedForest model
3:for $epoch\ t=1\ to\ N$ do
4: for $i=1\ to\ k$ do
5: Send $F_{i}^{{}^{\prime}}\subseteq F_{i}$, $D^{{}^{\prime}}\subseteq D$,
$T=D-D^{{}^{\prime}}$ to $C_{i}$ $\triangleright$ S1
6: end for
7: function TreeBuild($D^{{}^{\prime}},T$)
8: if all samples in $D^{{}^{\prime}}$ have the same label then
9: return $leaf\ node$
10: end if
11: Receive $BestF_{i}^{k},GiniPara_{i}^{k},needPruning_{i}^{k}$
$\triangleright$ S3
12: $BestF_{s}=Max(BestF_{i=1}^{k},key=GiniPara_{i})$ $\triangleright$ S4
13: if $needPruning_{s}==True$ then
14: Send ${}^{\prime}Prune^{\prime}$ message to $C_{i}$
15: return $leaf\ node$
16: end if
17: Send ${}^{\prime}Success^{\prime}$ message to $C$ $\triangleright$ S5
18: Receive $D_{l}^{{}^{\prime}},D_{r}^{{}^{\prime}},T_{l},T_{r}$ from $C_{i}$
19: Send $D_{l}^{{}^{\prime}},D_{r}^{{}^{\prime}},T_{l},T_{r}$ to other
$C_{i}$
20: leftTree $\leftarrow$ TreeBuild($D_{l}^{{}^{\prime}},T_{l}$)
21: rightTree $\leftarrow$ TreeBuild($D_{r}^{{}^{\prime}},T_{r}$)
22: return tree node $\triangleright$ S7
23: end function
24: Add this decision tree to forest
25:end for$\triangleright$ S8
The prediction process is described as follows. $C_{i}$ analyzes the new data
ID according to the stored tree. For each sample, when an unknown node is
encountered, the sample enters the left and right subtrees. When the
encountered node is known (that means, the dividing feature of this node
belongs to this client’s dataset), the sample enters the corresponding subtree
according to the threshold of the feature. The final data that falls on the
leaf node _l_ of the global tree _t_ is the intersection of all client data
that falls on this leaf node.
We have experimented to evaluate the improvement of FedRF. In this experiment,
we constructed a VFL scenario with two clients sharing data features using the
Bosch dataset. There are 50 independent features in each client, which are
extracted after performing the principal component analysis (PCA) from
different production lines and the same 11154 samples. The experiment results
are shown in Table 2. It can be seen that we get better prediction results
than the federated random forest algorithm proposed in Liu et al. (2020a).
Table 2: Improvement of FedRF Algo. | ACC | PRE | F1 | MCC | AUC
---|---|---|---|---|---
FedForest Liu et al. (2020a) | 0.552 | 0.493 | 0.550 | 0.118 | 0.559
FedRF (Ours) | 0.825 | 0.802 | 0.883 | 0.593 | 0.866
## 4 Methodology
### 4.1 Experiment Overview and Research questions
This paper studies whether there is a big difference between the effectiveness
of FL and CL algorithms for failure prediction on the production line, and
then whether FL algorithms can replace CL algorithms on this problem. To this
end, we construct two production scenarios. One is HFL, where we compare
FedSVM and SVM. The other is VFL, where we compare the effectiveness of FedRF
and RF. We design four research questions (RQ), each of which concerns both
HFL and VFL. The logical association between the four RQs is depicted, as
shown in Fig. 3.
Figure 3: Experiment Overview and Research Questions
RQ1: On the whole Bosch testing dataset, is there a significant difference
between FedSVM and SVM? We ask the same question for FedRF and RF.
RQ1 is aimed at comparing the average performance of FL and CL algorithms on
the whole Bosch testing dataset. If the difference between the value of each
measurement is within the threshold $\delta$ ( $\delta=0.1$), it can be
considered that there is no significant difference between this pair of
algorithms on the Bosch testing dataset.
RQ2: On the random partial Bosch testing dataset, is there a significant
difference between FedSVM and SVM? We ask the same question for FedRF and RF.
RQ2 is aimed at comparing the average performance of the FL and CL algorithms
on the random partial Bosch testing dataset that contains consecutive time-
series samples. If the prediction results have no significant difference, it
means that the pair of algorithms perform similarly on the random partial
testing data.
RQ3: On the estimated unknown Bosch data, is there a significant difference
between FedSVM and SVM? We ask the same question for FedRF and RF.
RQ3 is aimed at comparing the average performance of FL and CL algorithms on
the estimated unknown Bosch data. To answer this RQ, we have designed a method
to compare the error distribution of FL and CL algorithm on the estimated
unknown data based on the given data, as shown in Fig. 4. This method consists
of four main steps (S1-S4), explained hereafter.
Figure 4: Comparing FL and CL algorithms on the Estimated Unknown Bosch Data
S1. Construct prediction table. We first construct a prediction table
containing the ground truth (GT) label, the prediction results (Pre) of CL and
FL algorithms of products ordered by timestamp (ts).
S2. Construct timed state sequence. The states represent the prediction error
compared to the GT labels: 1) _hit_ state: GT positive + Pre positive or GT
negative + Pre negative; 2) _miss_ state: GT positive + Pre negative; 3)
_mistake_ state: GT negative + Pre positive.
S3. Fit Markov models. We fit Markov model using the timed state sequence, and
obtain $\mathtt{M_{FL}}$ and $\mathtt{M_{CL}}$, respectively.
S4. Compare FL and CL Markov models. If the difference between the parameters
of $\mathtt{M_{FL}}$ and $\mathtt{M_{CL}}$ is within the threshold $\delta$
($\delta=0.1$), the difference between FL and CL predictive models will
maintain similar on estimated unknown Bosch data if the production line
structure, manufacturing process, and quality control methods are not changed.
RQ4: Is Bosch testing data heterogeneous? And what is the impact of data
heterogeneity on the results of RQ1-RQ3?
RQ4 analyzes the impact of data heterogeneity on the results of RQ1-RQ3. To
evaluate the heterogeneity of data, we first obtain N groups of randomly
selected continuous samples $\mathtt{D^{i}}$, and then uses the GT labels as
states (GT positive state and GT negative state) to construct the state
transition sequence $\mathtt{S^{i}_{GT}}$. Then we fit the Markov model
$\mathtt{M^{i}_{GT}}$ using $\mathtt{S^{i}_{GT}}$, and perform DBSCAN
clustering Ester et al. (1996) on the parameter matrix of
$\mathtt{M^{i}_{GT}}$. If the clustering result shows multiple different
clusters, it means that the data heterogeneity is strong. DBSCAN calculates
the distance between two parameter matrices by converting the two matrices
into a one-dimensional vector and calculating the angle between the two
vectors. If the included angle is less than the specified neighborhood
distance threshold, the two-parameter matrices can be regarded as on the same
cluster.
Based on the results of heterogeneity, the conclusions of RQ1-RQ3 (_Y_ means
there is no significant difference between FL and CL, _N_ means there is
significant difference between FL and CL) will be analyzed from the following
four aspects.
* $\circ$
If the data heterogeneity is strong, and the answers to RQ1-RQ3 are _Y_ , then
the conclusion will be further strengthened. In our manufacturing scenario,
the FL and CL algorithms can obtain similar prediction results and can replace
each other.
* $\circ$
If the data heterogeneity is strong and some answer in RQ1-RQ3 is _N_ , it
means that data heterogeneity may be one of the reasons for disturbing the
conclusions.
* $\circ$
If the data heterogeneity is weak and the answers to RQ1-RQ3 are _Y_ , it can
be concluded that the FL algorithm can replace the CL one under the premise of
data homogeneity.
* $\circ$
If the data heterogeneity is weak and some answer in RQ1-RQ3 is _N_ , it means
that the FL algorithm cannot replace CL even with homogeneous data.
### 4.2 Measurements
MSA is one of the commonly used analysis methods in manufacturing Montgomery
(2007). The learning-based failure prediction method can be seen as a product
quality measurement system. Taking the Bosch company’s product quality testing
results as a benchmark, we can analyze the FL and CL algorithms through the
MSA method. The evaluation measurements in MSA mainly include Accuracy (ACC),
Precision (PRE), and Stability. On this basis, F1, AUC, and Matthew’s
Correlation Coefficient (MCC) are also involved. The six measurements we use
in the empirical study are provided in Table 3, where P represents the number
of GT positive cases, N is the number of GT negative cases, TP (hit) is true
positive, TN (correct rejection) is true negative, FP (false alarm) is false
positive, and FN (miss) is false negative.
Table 3: Measurements Measurement | Formula
---|---
ACC | $\mathtt{\dfrac{TP+TN}{TP+FN+TN+FP}}$
PRE | $\mathtt{\dfrac{TP}{TP+FP}}$
F1 | $\mathtt{\dfrac{2TP}{2TP+FP+FN}}$
MCC | $\mathtt{\dfrac{TP\times TN-FP\times FN}{\sqrt{(TP+FP)(TP+FN)(TN+FP)(TN+FN)}}}$
AUC | The area enclosed by the ROC curve and the reference line
Stability | The tendency of accuracy for N groups of random selected continuous data group.
MCC is suitable for evaluating the prediction results of unbalanced data. AUC
is a commonly used two-category evaluation method, and its value is the area
enclosed by the ROC curve and the reference line. The x-axis of a ROC curve is
the false positive rate, and the y-axis of a ROC curve is the true positive
rate. It shows the relationship between clinical sensitivity and specificity
for every possible cut-off. It is a graph with: The x-axis showing
1-specificity ($\mathtt{=FP/(FP+TN)}$) The y-axis showing sensitivity
($\mathtt{=TP/(TP+FN)}$). The range of AUC value is between 0 and 1.
Stability characterizes the ability of the measurement system to maintain a
constant performance within a certain time range. We divide the test set into
N groups (N = 10) on average according to the time-series and calculate the
accuracy of these N groups to analyze the stability of the target algorithm.
### 4.3 Bosch Dataset
The Bosch dataset is one of the largest public manufacturing datasets (14.3Gb)
on Kaggle. The Bosch dataset consists of a training set (1184687 samples) and
a test set (1183748 samples). Each sample has three types of features:
categorical feature, numeric feature, and date feature, and a two-category
classification label indicating the faulty status of a product.
The date feature gives the time stamp of each site through which the product
passes. Based on the date feature, it is possible to analyze the timed
behavior. Existing studies have constructed LSTM neural network models based
on time-series information Huang et al. (2019b); Moldovan et al. (2019); Liu
et al. (2020d). Since we focus on empirical research on FL without time
information, the date features are excluded from the scope of our study.
Nevertheless, the time-series information can still provide time-continuity
data for empirical research on the estimated unknown data.
The data features are shown in Table 4. There are 1184687 products in the
dataset, including 1177808 positive samples (Pos) and 6879 negative ones
(Neg). In the feature space, the dataset has 968 features. These features are
collected from 51 workstations (S1-S51) on 4 production lines L0-L3 that
contain 168, 513, 42, and 245 features, respectively. L1_S24_F1695 indicates
that the feature No.1695 was observed at the No.24 workstation of the L1
production line. The label Response represents whether the final product is
qualified (label = 0) or not (label = 1). From the perspective of data items,
only a few items (0.58%) in the dataset result are qualified, and the data is
extremely unbalanced. The data is also extremely sparse, with only a few
workstations collecting data from most products Carbery et al. (2019).
Table 4: Bosch dataset Sample Num (1184687) | Features Num (968)
---|---
Positive | Negative | L0 | L1 | L2 | L3
1177808 | 6879 | 168 | 513 | 42 | 245
### 4.4 Experiment Scenarios of HFL and VFL
To compare FedSVM and SVM on the testing data, we form four workshops A, B, C,
and D that do not share manufacturing data but have the same structure and
configuration of the production line. Under the premise that the product
features are aligned, they contain different samples, and each sample has a
unique ID. Each workshop can be regarded as an independent client. Table 5
shows the data used in this experiment. The number of samples on the 4 clients
is basically the same. In the experiment, the data set is randomly distributed
to all clients on average, each of which possesses 3000 positive and 1500
negative samples. The four workshops have 713 common features of data.
Table 5: Data Description of HFL Scenario Sample Num. | Positive | Negative | Feature Num.
---|---|---|---
Total | 12000 | 6000 | 713 selected
Per client | 3000 | 1500
To compare FedRF and RF on the given testing data, it is assumed that the
first three production lines (L0, L1, L2) of the Bosch dataset belong to the
same organization O1, and the L3 production line belongs to another
organization O2. O1 and O2 are independent. Since the samples in O1 and O2
have the same ID and different characteristics, VFL is applied. The data set
used in the FedRF experiment is shown in Table 6. We extract the features
belonging to L0, L1, and L2 as a group, and the features belonging to L3 as
another group. Then we perform principal component analysis (PCA) to reduce
the feature dimension in O1/O2 from 723/245 to 22/22, respectively. According
to the overall analysis of PCA results, the first 22 dimensions can represent
more than 95% of the variance, so reducing to 22 dimensions, respectively,
which will not cause too much information loss Zhang et al. (2016). The
positive and negative sample numbers are 12000 and 6000, respectively.
Table 6: Data Description for VFL Scenario Sample Num | Positive | Negative | Feature Num
---|---|---|---
Total | 12000 | 6000 | 44 (F1 - F44)
Client O1 | 22 (F1 - F22)
Client O2 | 22 (F23 - F44)
## 5 Experiment Results
Our experiments investigate and answer four research questions.
### 5.1 RQ1: Comparison on Bosch Testing Data
#### 5.1.1 FedSVM vs. SVM
Table 7 shows the experimental results. To ensure that the data set and the
algorithm kernel are consistent, the baseline is the result of FedSVM when the
number of clients is set to 1, ie., FedSVM degenerates into SVM. Under this
premise, the results of FedSVM and SVM are compared. When analyzing the
stability, the experiment data is ordered by time-series and then divided into
10 groups on average. The measurements of each group are calculated. The
overall results of stability are shown in Fig. 5. The average value of
stability is above 0.7 for FedSVM and SVM. The variance of the stability is
given as Stability (Var) in Table 7.
Table 7: Experiment Results of RQ1 on FedSVM/SVM Algo. | ACC | PRE | F1 | MCC | AUC | Stab. (Var)
---|---|---|---|---|---|---
FedSVM | 0.825 | 0.859 | 0.869 | 0.607 | 0.807 | 0.005
SVM | 0.859 | 0.828 | 0.903 | 0.690 | 0.902 | 0.003
Diff | -0.034 | 0.031 | -0.046 | -0.083 | -0.095 | 0.002
Figure 5: FedSVM vs. SVM on Stability
The value difference of ACC, Precision, F1, MCC, AUC, and Stability (Var)
between FedSVM and SVM are all within a threshold value of 0.1, respectively.
Therefore, the conclusion of RQ1 on FedSVM and SVM is that _the prediction
results of FedSVM and SVM on the Bosch testing data are not significantly
different_.
In Figure 5, FedSVM performs slightly better than SVM on two test groups,
namely 1 and 7. Similar results are also shown for the Precision measure in
Table 7. This effect is only shown in the HFL scenario, where the clients
contribute local data features to the central server. If the amount of data on
each client is small, FL’s effect can be better than CL because FL expands the
number of IID data samples in the HFL scenario.
#### 5.1.2 FedRF vs. RF
The experimental results are shown in Table 8. FedRF degenerates into RF by
combining two clients so that the effect of FedRF and RF can be compared under
the premise of ensuring the consistency of the testing data and the algorithm
kernel. When analyzing the stability, the test set is divided into 10 groups
on average according to the time-series, and the measurements of each group
are calculated. The variance of the stability is given in Table 8, and the
overall results of stability are shown in Fig. 6.
Table 8: Experiment Results of RQ1 on FedRF/RF Algo. | ACC | PRE | F1 | MCC | AUC | Stab. (Var)
---|---|---|---|---|---|---
FedRF | 0.843 | 0.808 | 0.894 | 0.659 | 0.902 | 0.004
RF | 0.868 | 0.836 | 0.909 | 0.713 | 0.912 | 0.002
Diff | -0.025 | -0.028 | -0.015 | -0.054 | -0.010 | 0.002
Figure 6: FedRF vs. RF on Stability
The value difference of ACC, Precision, F1, MCC, AUC, and Stability (Var) on
each group are all within 0.1. Therefore, _the prediction results of FedRF and
RF on the Bosch testing data are not significantly different._
### 5.2 RQ2: Comparison on Random Partial Bosch Testing Data
In order to compare the average difference between the FL and CL algorithms in
randomly sampling partial data, we randomly select N groups (N=100) of testing
samples. Each group contains L time-continuous samples. The starting point S
of the group is randomly set. L is randomly generated from the interval [300,
1000].
#### 5.2.1 FedSVM vs. SVM
The line charts of the prediction results of FedSVM and SVM on 100 groups are
shown in Fig. 7. In all charts, the abscissa is the serial number of the 100
groups. The ordinate is the value of the measurement on each group. The solid
blue line and red dot-dash line represent the prediction result of FedSVM and
SVM, respectively. Through the charts, we can see the range and difference of
the FedSVM and SVM on the prediction results of each group.
(a) ACC Values of FedSVM and SVM
(b) PRE Values of FedSVM and SVM
(c) F1 Values of FedSVM and SVM
(d) MCC Values of FedSVM and SVM
(e) AUC Values of FedSVM and SVM
Figure 7: FedSVM vs. SVM on Random Partial Data
We draw a histogram to visualize the average difference between FedSVM and SVM
on each measurement for 100 groups of data, as shown in Fig. 8. Based on all
results, the difference values in ACC, PRE, and F1 are all within 0.1. The
difference values in MCC and AUC are mostly within 0.2. It is thus validated
that _the performance of FedSVM and SVM on random partial Bosch testing data
is not significantly different_.
(a) ACC Difference
(b) PRE Difference
(c) F1 Difference
(d) MCC Difference
(e) AUC Difference
Figure 8: Difference of FedSVM and SVM on Random Partial Testing Data
#### 5.2.2 FedRF vs. RF
The line charts of FedRF and RF on 100 groups of samples are shown in Fig. 9.
In all line charts, the abscissa is the number of groups. The ordinate is the
value of the measurements on each group. The solid blue line and red dot-dash
line represent the prediction result of the FedRF and RF, respectively.
(a) ACC Values of FedRF and RF
(b) PRE Values of FedRF and RF
(c) F1 Values of FedRF and RF
(d) MCC Values of FedRF and RF
(e) AUC Values of FedRF and RF
Figure 9: FedRF vs. RF on Random Partial Data
We draw a histogram of the differences between FedRF and RF on each
measurement for 100 groups of data, as shown in Fig. 10. Based on all results,
the difference values in ACC, PRE, F1, and AUC are all within 0.1. The
difference values in MCC are within 0.2, and the MCC values of over 80% groups
are within 0.1. It is thus validated that _the performance of FedRF and RF on
the random partial Bosch testing data is not significantly different._
(a) ACC Difference
(b) PRE Difference
(c) F1 Difference
(d) MCC Difference
(e) AUC Difference
Figure 10: Difference of FedRF and RF on Random Partial Testing Data
### 5.3 RQ3: Comparison on Estimated Unknown Bosch Data
This experiment is used to analyze whether the FL and CL algorithms can
maintain no significant difference on the estimated unknown Bosch data. This
experiment is conducted following the method presented in Sect. 4.
#### 5.3.1 FedSVM vs. SVM
We first constructed a timed state sequence indicating the prediction error
using FedSVM and SVM, respectively, based on which their Markov models are
fitted. The parameters of two Markov models are given in Table
LABEL:tab:RQ3-SVM_resa and LABEL:tab:RQ3-SVM_resa. The value difference
between each pair of parameters is calculated in Table LABEL:tab:RQ3-SVM_resc.
The average difference value between each pair of parameters is 0.054. The
maximum difference value is 0.096. It means that there exists no significant
difference between the two Markov models, and the prediction error between
FedSVM and SVM are almost equally distributed on the estimated unknown Bosch
data. Therefore, _the performance of FedSVM and SVM in the subsequent Bosch
production was not significantly different._
Table 9: Experiment Results of RQ3 on FedSVM/SVM
State | Hit | Miss | Mistake
---|---|---|---
Hit | 0.882 | 0.113 | 0.005
Miss | 0.715 | 0.280 | 0.005
Mistake | 0.813 | 0.062 | 0.125
(a) Markov Model Parameters of Prediction Error using FedSVM
| Hit | Miss | Mistake
---|---|---|---
Hit | 0.851 | 0.077 | 0.072
Miss | 0.690 | 0.225 | 0.085
Mistake | 0.717 | 0.125 | 0.158
(b) Markov Model Parameters of Prediction Error using SVM
| Hit | Miss | Mistake
---|---|---|---
Hit | 0.031 | 0.036 | 0.067
Miss | 0.025 | 0.055 | 0.080
Mistake | 0.096 | 0.063 | 0.033
(c) Comparison of Two Markov Model Parameters
#### 5.3.2 FedRF vs. RF
We first constructed a timed state sequence indicating the prediction error
using FedRF and RF, respectively, based on which their Markov models are
fitted. The parameters of two Markov models are given in Table
LABEL:tab:RQ3-RF_resa and LABEL:tab:RQ3-RF_resa. The value difference between
each pair of parameters is calculated in Table LABEL:tab:RQ3-RF_resc. The
average difference value between each pair of parameters is 0.035. The maximum
difference value is 0.100. It means that there exists no significant
difference between the two Markov models, and the prediction error between
FedRF and RF are equally distributed on the estimated unknown Bosch data.
Therefore, _the performance of FedRF and RF in the subsequent Bosch production
was not significantly different._
Table 10: Experiment Results of RQ3 on FedRF/RF
| Hit | Miss | Mistake
---|---|---|---
Hit | 0.869 | 0.130 | 0.001
Miss | 0.701 | 0.299 | 0.000
Mistake | 1.000 | 0.000 | 0.000
(a) Markov Model Parameters of Prediction Error using FedRF
| Hit | Miss | Mistake
---|---|---|---
Hit | 0.888 | 0.109 | 0.003
Miss | 0.735 | 0.263 | 0.002
Mistake | 0.900 | 0.100 | 0.000
(b) Markov Model Parameters of Prediction Error using RF
| Hit | Miss | Mistake
---|---|---|---
Hit | 0.019 | 0.021 | 0.002
Miss | 0.034 | 0.036 | 0.002
Mistake | 0.100 | 0.100 | 0.000
(c) Comparison of Two Markov Model Parameters
### 5.4 RQ4: Heterogeneity of Bosch Testing Data
This experiment is used to evaluate the heterogeneity of the testing data used
in the experiments of RQ1-RQ3. We follow the method proposed in Sect. 4 to
experiment. One hundred groups of time-ordered consecutive samples were
randomly selected. According to the GT label of the sample, the qualified
sample is set to state S0, and the unqualified sample is set to state S1,
forming a state sequence. For the state sequence of each group of samples, the
Markov model of k=1 step and k=2 steps are established, respectively. The
DBSCAN algorithm is used to cluster and analyze all the 1-step and 2-step
parameter matrices. If the clustering results show multiple different
clusters, it can explain that there is heterogeneity within the testing data.
#### 5.4.1 Heterogeneity of the Testing Data used in FedSVM
The experimental results are shown in Table 11. There are outliers in the test
results, so the sum of the samples in each cluster is less than 100. It can be
seen that the testing data is divided into 2 clusters. The difference between
clusters is no less than the distance threshold, which means that the data
used in the experiment of FedSVM is heterogeneous. It enforces the conclusion
that _FedSVM and SVM have no significant difference on heterogeneous
manufacturing data for the problem of failure prediction_.
Table 11: Experiment Results of RQ4 on FedSVM Step | Clusters | Distance threshold (∘) | Contour factor | Samples / cluster
---|---|---|---|---
k = 1 | 2 | 5 | 0.456 | 92 / 6
k = 2 | 2 | 8 | 0.421 | 89 / 6
According to the experiment results and the statement of RQ4 in Section 4.1,
the data heterogeneity in the VFL scenario is relatively strong and the answer
to RQ1-RQ3 are positive. This means in our VFL manufacturing scenarios, the FL
and CL algorithms can obtain similar prediction results and can thus replace
each other under the premise of data heterogeneity.
#### 5.4.2 Heterogeneity of the Testing Data used in FedRF
The experimental results are shown in Table 12. It can be seen that the
testing data is divided into 2 clusters. The difference between clusters is no
less than the distance threshold, which means that the data used in the
experiment of FedRF is heterogeneous. It enforces the conclusion that _FedRF
and RF have no significant difference on heterogeneous manufacturing data for
the problem of failure prediction_.
Table 12: Experiment Results of RQ4 on FedRF Step | Clusters | Distance threshold (∘) | Contour factor | Samples / cluster
---|---|---|---|---
k = 1 | 2 | 4 | 0.524 | 94 / 6
k = 2 | 2 | 8 | 0.473 | 78 / 22
According to the experiment results and the statement of RQ4 in Section 4.1,
the data heterogeneity in the HFL scenario is relatively strong and the answer
to RQ1-RQ3 are positive. This means in our HFL manufacturing scenarios, the FL
and CL algorithms can obtain similar prediction results and can thus replace
each other under the premise of data heterogeneity.
## 6 Threats to Validity
Following common guidelines for empirical studies Runeson and Höst (2009); Yin
(2017), we discuss threats to the validity of our study.
Threats to internal validity are mainly concerned with uncontrolled factors.
These factors may impact the results and reduce their creditability. In this
work, the main threat to internal validity is potential defects in the
implementation of our own algorithm and the reimplementation of other
algorithms (mainly the federated SVM and federated random forest algorithms).
To reduce this threat, we reimplement the baseline techniques by following
their papers. We carefully review all our code and experiment scripts by
several developers to ensure their correctness. However, there is always a
small possibility of defects, which introduces risk to the result’s
correctness.
Threats to external validity are mainly concerned with whether the performance
of our techniques can still hold in other experimental settings. In this work,
the main threat to external validity is the impact of data preprocessing
method. As we mentioned in Sect. 1, the existing works revealed that the data
preprocessing on the Bosch dataset has a significant impact on the prediction
result. In our work, we have applied PCA to find the principle features. For
example, in the VFL scenario, according to the PCA analysis, only 22 features
are kept in each client. But, these features represent more than 95% of the
variance of data. Another threat to external validity is the experimental
methodology we designed. In order to answer whether the FL algorithm can
replace the CL algorithm or not, we conduct experiment on the whole testing
data, on the partial random testing data, and on the estimated unknown data,
respectively. This method is aimed at simulating possible data set used in
manufacturing. Therefore, the answer tends to be complete.
The first threat to construct validity is the suitability of our evaluation
metrics. To reduce this risk, we follow the same measurements (ACC, precision,
F1, MCC, AUC, stability) as used in other works on failure prediction in
production line Carbery et al. (2019, 2018); Zhang et al. (2016); Khoza and
Grobler (2019); Kotenko et al. (2019); Mangal and Kumar (2016); Hebert (2016);
Maurya (2016); Huang et al. (2019b); Moldovan et al. (2019); Liu et al.
(2020d) to measure the quality of the learning model, which reduces the risk
of construction validity.
Another threat to construct validity is the threshold $\delta$ ( $\delta=0.1$
or $\delta=0.2$) used in the experiment. In the industry, the threshold value
is set according to the production needs. We have performed a hypothesis test
for the difference between the two approaches’ performances to explain that
the difference between FL and CL is within the threshold. The original
hypothesis H0 is difference $>\delta$, and the alternative hypothesis H1 is
difference $<\delta$, where the difference = CL – FL on the same random
partial testing data group. Suppose $\alpha$ = 0.05. The p-value of the
evaluation metrics for FedSVM vs. SVM and FedForest vs. RForest is shown in
Table 13 and Table 14, respectively. As $alpha>$ p-value for all metrics, the
alternative hypothesis H1 is accepted, which means that the difference between
CL and FL is less than the threshold $\delta$ ( $\delta=0.1$ or $\delta=0.2$).
Table 13: P-value of FedSVM vs. SVM ($\alpha$ = 0.05) Metrics | Threshold | Statistics | p-value
---|---|---|---
ACC | 0.1 | -22.48877 | 5.44616e-41
Precision | 0.1 | -83.57313 | 6.33511e-94
F1 | 0.1 | -17.48523 | 2.45813e-32
MCC | 0.2 | -20.44888 | 1.27870e-37
AUC | 0.2 | -26.63278 | 3.08126e-47
stability | 0.1 | -5.40692 | 0.00021
Table 14: P-value of FedForest vs. RForest ($\alpha$ = 0.05) Metrics | Threshold | Statistics | p-value
---|---|---|---
ACC | 0.1 | -34.44843 | 3.16706e-57
Precision | 0.1 | -48.40227 | 5.04976e-71
F1 | 0.1 | -75.16218 | 1.95987e-89
MCC | 0.2 | -12.90324 | 3.11933e-23
AUC | 0.2 | -16.77851 | 5.26331e-31
stability | 0.1 | -9.83780 | 2.04964e-06
## 7 Discussion
Due to data privacy and security, we are facing a bottleneck when applying CL
in the industry. In this work, we have adopted two FL algorithms to predict
failure in the production line. According to our investigation of existing
results of this problem, SVM and RF performed better on Bosch data among CL
algorithms that did not consider time series. Therefore, we choose Federated-
SVM and Federated-RF in the study. We expect that both will perform as well as
the CL algorithms. The results showed that our intuition is not wrong on the
given dataset.
An interesting question is whether FL can achieve the same or similar
prediction results as CL. If yes, what are the conditions (e.g., the the
dataset’s size)? Otherwise, what are the guidelines on when we can use FL in
place of CL? As our study is limited to the Bosch dataset from the
manufacturing industry, we have conducted a literature investigation to answer
this question. According to existing studies Kairouz et al. (2019); Duan et
al. (2019), FL’s effect is mainly related to the data distribution and data
amount. Two important facts are summarized as follows.
1. 1.
When the data is independent and identically distributed, the learning results
of FL and CL are similar. For instance, the work Lu et al. (2019); Chen et al.
(2020); Sozinov et al. (2018); Olivia et al. (2019) showed that when the
training data is independent and identically distributed (IID), the difference
between FL and CL is within 3%. If the amount of data on each client is small,
FL’s effect is better than CL because FL expands the number of IID data
samples Ickin et al. (2019); Bakopoulou et al. (2019); Guha et al. (2019);
Xinle et al. (2019); Shaoqi et al. (2020); Suzumura et al. (2019).
2. 2.
When the training data is unevenly distributed, FL may not achieve the same
effect as CL Sheller et al. (2018a); Hu et al. (2018); Feng et al. (2020);
Dianbo et al. (2018); Luca and M (2019); Chen et al. (2020); Boughorbel et al.
(2019a); Sozinov et al. (2018). The review article Kairouz et al. (2019)
claimed that if the amount of data on a client is small, the CL model may be
better than the FL model trained using data on multi-clients in the case of
uneven distribution of training data.
Besides, some works have shown that the effect of FL is also related to
encryption algorithms R et al. (2019); Tao et al. (2020); Feng et al. (2020);
Olivia et al. (2019); Lu et al. (2020); Li et al. (2019, 2020b). If the
encryption strength increases, the information loss on the data will be worse.
So that the effect of FL decreases. The trade-off between data privacy
protection and the accuracy of the trained model is still inevitable. It is
still considered an open question of how the effect of FL is related to other
factors.
## 8 Conclusion
There is increasing attention for federated learning in the manufacturing
industry, but few works have studied how federated learning methods perform in
practice. In this work, we have conducted an empirical study of comparing FL
and CL methods for problem of the failure prediction in the production line.
We constructed a horizontal and vertical federated learning scenario based on
the Bosch dataset. We implemented FedSVM to compare with SVM in the HFL
scenario and designed FedRF to compare it with RF in the VFL scenario. We also
designed an experiment process for evaluating the effectiveness of FL and CL
algorithms, which can be reused in future studies. Our results reveal that
FedSVM and FedRF can replace SVM and RF, respectively, for failure prediction
in the Bosch production line. Because the CL algorithm can be replaced by the
FL one (1) on the global testing dataset; (2) on the random partial testing
dataset; (3) on the estimated unknown Bosch data. The value of each evaluation
metric for FL and CL algorithms differ within 0.1. Moreover, the fact that the
testing data is heterogeneous enhances the above three conclusions.
The results of our empirical study reveal that FL can replace CL in some
applications and FL is a desirable way of protecting data in the manufacturing
industry. More FL techniques can also be investigated in other manufacturing
applications. The studies Moldovan et al. (2019); Huang et al. (2019b); Liu et
al. (2020d) have considered time-series features and showed that the Long
Short Term Memory (LSTM) network can improve the prediction result when time-
series features are included Huang et al. (2019b). A future work is to design
federated LSTM models to compare with centralized LSTM model, and study
whether FL can replace CL within this context.
###### Acknowledgements.
This work was supported by National Key Research and Development Program of
China Grant No. 2019YFB1703903, National Natural Science Foundation of China
Grant No. 61732019 and No. 61902011, and the Grant No. NJ2018014 of the Key
Laboratory of Safety-Critical Software (Nanjing University of Aeronautics and
Astronautics), Ministry of Industry and Information Technology.
## References
* Aussel et al. [2020] Nicolas Aussel, Sophie Chabridon, and Yohan Petetin. Combining federated and active learning for communication-efficient distributed failure prediction in aeronautics. _arXiv preprint arXiv:2001.07504_ , 2020.
* Bakopoulou et al. [2019] Evita Bakopoulou, Balint Tillman, and Athina Markopoulou. A federated learning approach for mobile packet classification. _arXiv preprint arXiv:1907.13113_ , 2019.
* Boughorbel et al. [2019a] Sabri Boughorbel, Fethi Jarray, Neethu Venugopal, Shabir Moosa, Haithum Elhadi, and Michel Makhlouf. Federated uncertainty-aware learning for distributed hospital EHR data. _arXiv preprint arXiv:1910.12191_ , 2019a.
* Boughorbel et al. [2019b] Sabri Boughorbel, Fethi Jarray, Neethu Venugopal, Shabir Moosa, Haithum Elhadi, and Michel Makhlouf. Federated uncertainty-aware learning for distributed hospital ehr data. _arXiv preprint arXiv:1910.12191_ , 2019b.
* Brisimi et al. [2018] Theodora S Brisimi, Ruidi Chen, Theofanie Mela, Alex Olshevsky, Ioannis Ch Paschalidis, and Wei Shi. Federated learning of predictive models from federated electronic health records. _International journal of medical informatics_ , 112:59–67, 2018.
* Carbery et al. [2018] Caoimhe M Carbery, Roger Woods, and Adele H Marshall. A bayesian network based learning system for modelling faults in large-scale manufacturing. In _2018 IEEE international conference on industrial technology (ICIT)_ , pages 1357–1362. IEEE, 2018.
* Carbery et al. [2019] Caoimhe M Carbery, Roger Woods, and Adele H Marshall. A new data analytics framework emphasising preprocessing of data to generate insights into complex manufacturing systems. _Proceedings of the Institution of Mechanical Engineers, Part C: Journal of Mechanical Engineering Science_ , 233(19-20):6713–6726, 2019.
* Chen et al. [2020] Dawei Chen, Linda Jiang Xie, BaekGyu Kim, Li Wang, Choong Seon Hong, Li-Chun Wang, and Zhu Han. Federated learning based mobile edge computing for augmented reality applications. In _2020 International Conference on Computing, Networking and Communications (ICNC)_ , pages 767–773. IEEE, 2020.
* Dianbo et al. [2018] Liu Dianbo, Miller Timothy, Sayeed Raheel, and Mandl Kenneth D. Fadl: Federated-autonomous deep learning for distributed electronic health record. _arXiv preprint arXiv:1811.11400_ , 2018.
* Duan et al. [2019] Moming Duan, Duo Liu, Xianzhang Chen, Yujuan Tan, Jinting Ren, Lei Qiao, and Liang Liang. Astraea: Self-balancing federated learning for improving classification accuracy of mobile deep learning applications. In _37th IEEE International Conference on Computer Design, ICCD 2019, Abu Dhabi, United Arab Emirates, November 17-20, 2019_ , pages 246–254. IEEE, 2019.
* Ester et al. [1996] Martin Ester, Hans-Peter Kriegel, Jörg Sander, Xiaowei Xu, et al. A density-based algorithm for discovering clusters in large spatial databases with noise. In _Kdd_ , volume 96, pages 226–231, 1996.
* Feng et al. [2020] Jie Feng, Can Rong, Funing Sun, Diansheng Guo, and Yong Li. PMF: A privacy-preserving human mobility prediction framework via federated learning. _IMWUT_ , 4(1):10:1–10:21, 2020.
* Guha et al. [2019] Roy Abhijit Guha, Siddiqui Shayan, Pölsterl Sebastian, Navab Nassir, and Wachinger Christian. Braintorrent: A peer-to-peer environment for decentralized federated learning. _arXiv preprint arXiv:1905.06731_ , 2019.
* Hard et al. [2018] Andrew Hard, Kanishka Rao, Rajiv Mathews, Swaroop Ramaswamy, Françoise Beaufays, Sean Augenstein, Hubert Eichner, Chloé Kiddon, and Daniel Ramage. Federated learning for mobile keyboard prediction. _arXiv preprint arXiv:1811.03604_ , 2018.
* Hebert [2016] Jeff Hebert. Predicting rare failure events using classification trees on large scale manufacturing data with complex interactions. In _2016 ieee international conference on big data (big data)_ , pages 2024–2028. IEEE, 2016.
* Hu et al. [2018] Binxuan Hu, Yujia Gao, Liang Liu, and Huadong Ma. Federated region-learning: An edge computing based framework for urban environment sensing. In _IEEE Global Communications Conference, GLOBECOM 2018, Abu Dhabi, United Arab Emirates, December 9-13, 2018_ , pages 1–7. IEEE, 2018\.
* Huang et al. [2019a] Li Huang, Andrew L Shea, Huining Qian, Aditya Masurkar, Hao Deng, and Dianbo Liu. Patient clustering improves efficiency of federated machine learning to predict mortality and hospital stay time using distributed electronic medical records. _Journal of Biomedical Informatics_ , 99:103291, 2019a.
* Huang et al. [2019b] Xin Huang, Cecilia Zanni-Merk, and Bruno Crémilleux. Enhancing deep learning with semantics: an application to manufacturing time series analysis. _Procedia Computer Science_ , 159:437–446, 2019b.
* Ickin et al. [2019] Selim Ickin, Konstantinos Vandikas, and Markus Fiedler. Privacy preserving qoe modeling using collaborative learning. In Pedro Casas, Florian Wamser, Fabian E. Bustamante, and David R. Choffnes, editors, _Proceedings of the 4th Internet-QoE Workshop on QoE-based Analysis and Management of Data Communication Networks, Internet-QoE@MobiCom 2019, Los Cabos, Mexico, October 21, 2019_ , pages 13–18. ACM, 2019.
* Kairouz et al. [2019] Peter Kairouz, H Brendan McMahan, Brendan Avent, Aurélien Bellet, Mehdi Bennis, Arjun Nitin Bhagoji, Keith Bonawitz, Zachary Charles, Graham Cormode, Rachel Cummings, et al. Advances and open problems in federated learning. _arXiv preprint arXiv:1912.04977_ , 2019.
* Khoza and Grobler [2019] Sibusiso C Khoza and Jacomine Grobler. Comparing machine learning and statistical process control for predicting manufacturing performance. In _EPIA Conference on Artificial Intelligence_ , pages 108–119. Springer, 2019.
* Kotenko et al. [2019] Igor Kotenko, Igor Saenko, and Alexander Branitskiy. Improving the performance of manufacturing technologies for advanced material processing using a big data and machine learning framework. _Materials Today: Proceedings_ , 11:380–385, 2019.
* Kusiak [2017] Andrew Kusiak. Smart manufacturing must embrace big data. _Nature_ , 544(7648):23–25, 2017.
* Li et al. [2017] Bo-hu Li, Bao-cun Hou, Wen-tao Yu, Xiao-bing Lu, and Chun-wei Yang. Applications of artificial intelligence in intelligent manufacturing: a review. _Frontiers of Information Technology & Electronic Engineering_, 18(1):86–96, 2017.
* Li et al. [2020a] Tian Li, Anit Kumar Sahu, Ameet Talwalkar, and Virginia Smith. Federated learning: Challenges, methods, and future directions. _IEEE Signal Processing Magazine_ , 37(3):50–60, 2020a.
* Li et al. [2019] Wenqi Li, Fausto Milletarì, Daguang Xu, Nicola Rieke, Jonny Hancox, Wentao Zhu, Maximilian Baust, Yan Cheng, Sébastien Ourselin, M. Jorge Cardoso, and Andrew Feng. Privacy-preserving federated brain tumour segmentation. In Heung-Il Suk, Mingxia Liu, Pingkun Yan, and Chunfeng Lian, editors, _Machine Learning in Medical Imaging - 10th International Workshop, MLMI 2019, Held in Conjunction with MICCAI 2019, Shenzhen, China, October 13, 2019, Proceedings_ , volume 11861 of _Lecture Notes in Computer Science_ , pages 133–141. Springer, 2019.
* Li et al. [2020b] Zengpeng Li, Vishal Sharma, and Saraju P. Mohanty. Preserving data privacy via federated learning: Challenges and solutions. _IEEE Consumer Electronics Magazine_ , 9(3):8–16, 2020b.
* Liu et al. [2020a] Yang Liu, Yingting Liu, Zhijie Liu, Yuxuan Liang, Chuishi Meng, Junbo Zhang, and Yu Zheng. Federated forest. _IEEE Transactions on Big Data_ , 2020a.
* Liu et al. [2020b] Yi Liu, JQ James, Jiawen Kang, Dusit Niyato, and Shuyu Zhang. Privacy-preserving traffic flow prediction: A federated learning approach. _IEEE Internet of Things Journal_ , 2020b.
* Liu et al. [2020c] Yi Liu, Li Zhang, Ning Ge, and Guanghao Li. A systematic literature review on federated learning: From a model quality perspective. _arXiv preprint arXiv:2012.01973_ , 2020c.
* Liu et al. [2020d] Zhenyu Liu, Donghao Zhang, Weiqiang Jia, Xianke Lin, and Hui Liu. An adversarial bidirectional serial–parallel lstm-based qtd framework for product quality prediction. _Journal of Intelligent Manufacturing_ , pages 1–19, 2020d.
* Lu et al. [2019] Sidi Lu, Yongtao Yao, and Weisong Shi. Collaborative learning on the edges: A case study on connected vehicles. In Irfan Ahmad and Swaminathan Sundararaman, editors, _2nd USENIX Workshop on Hot Topics in Edge Computing, HotEdge 2019, Renton, WA, USA, July 9, 2019_. USENIX Association, 2019.
* Lu et al. [2020] Yunlong Lu, Xiaohong Huang, Yueyue Dai, Sabita Maharjan, and Yan Zhang. Differentially private asynchronous federated learning for mobile edge computing in urban informatics. _IEEE Trans. Ind. Informatics_ , 16(3):2134–2143, 2020.
* Luca and M [2019] Corinzia Luca and Buhmann Joachim M. Variational federated multi-task learning. _arXiv preprint arXiv:1906.06268_ , 2019.
* Mangal and Kumar [2016] Ankita Mangal and Nishant Kumar. Using big data to enhance the bosch production line performance: A kaggle challenge. In _2016 IEEE International Conference on Big Data (Big Data)_ , pages 2029–2035. IEEE, 2016.
* Maurya [2016] Abhinav Maurya. Bayesian optimization for predicting rare internal failures in manufacturing processes. In _2016 IEEE International Conference on Big Data (Big Data)_ , pages 2036–2045. IEEE, 2016.
* McMahan et al. [2016] H Brendan McMahan, Eider Moore, Daniel Ramage, Seth Hampson, et al. Communication-efficient learning of deep networks from decentralized data. _arXiv preprint arXiv:1602.05629_ , 2016.
* Moldovan et al. [2019] Dorin Moldovan, Ionut Anghel, Tudor Cioara, and Ioan Salomie. Time series features extraction versus lstm for manufacturing processes performance prediction. In _2019 International Conference on Speech Technology and Human-Computer Dialogue (SpeD)_ , pages 1–10. IEEE, 2019.
* Montgomery [2007] Douglas C Montgomery. _Introduction to statistical quality control_. John Wiley & Sons, 2007.
* Mowla et al. [2019] Nishat I Mowla, Nguyen H Tran, Inshil Doh, and Kijoon Chae. Federated learning-based cognitive detection of jamming attack in flying ad-hoc network. _IEEE Access_ , 2019.
* Nguyen et al. [2019] Thien Duc Nguyen, Samuel Marchal, Markus Miettinen, Hossein Fereidooni, N Asokan, and Ahmad-Reza Sadeghi. Dïot: A federated self-learning anomaly detection system for iot. In _2019 IEEE 39th International Conference on Distributed Computing Systems (ICDCS)_ , pages 756–767. IEEE, 2019.
* Olivia et al. [2019] Choudhury Olivia, Gkoulalas-Divanis Aris, Salonidis Theodoros, Sylla Issa, Park Yoonyoung, Hsu Grace, and Das Amar. Differential privacy-enabled federated learning for sensitive health data. _arXiv preprint arXiv:1910.02578_ , 2019.
* R et al. [2019] Pfohl Stephen R, Dai Andrew M, and Heller Katherine. Federated and differentially private learning for electronic health records. _arXiv preprint arXiv:1911.05861_ , 2019.
* [44] AC Einat Ronen and Kenneth Burns. Utilizing predictive sales analytics with big data. Technical report, Intel, Tech. Rep., 2013.[Online]. Available: http://www. intel. com/content ….
* Runeson and Höst [2009] Per Runeson and Martin Höst. Guidelines for conducting and reporting case study research in software engineering. _Empirical software engineering_ , 14(2):131, 2009\.
* Samarakoon et al. [2019] Sumudu Samarakoon, Mehdi Bennis, Walid Saad, and Mérouane Debbah. Distributed federated learning for ultra-reliable low-latency vehicular communications. _IEEE Transactions on Communications_ , 2019.
* Saputra et al. [2019] Yuris Mulya Saputra, Dinh Thai Hoang, Diep N Nguyen, Eryk Dutkiewicz, Markus Dominik Mueck, and Srikathyayani Srikanteswara. Energy demand prediction with federated learning for electric vehicle networks. _arXiv preprint arXiv:1909.00907_ , 2019.
* Shaoqi et al. [2020] Chen Shaoqi, Xue Dongyu, Chuai Guohui, Yang Qiang, and Liu Qi. Fl-qsar: a federated learning based qsar prototype for collaborative drug discovery. _bioRxiv_ , 2020.
* Sheller et al. [2018a] Micah J. Sheller, G. Anthony Reina, Brandon Edwards, Jason Martin, and Spyridon Bakas. Multi-institutional deep learning modeling without sharing patient data: A feasibility study on brain tumor segmentation. In Alessandro Crimi, Spyridon Bakas, Hugo J. Kuijf, Farahani Keyvan, Mauricio Reyes, and Theo van Walsum, editors, _Brainlesion: Glioma, Multiple Sclerosis, Stroke and Traumatic Brain Injuries - 4th International Workshop, BrainLes 2018, Held in Conjunction with MICCAI 2018, Granada, Spain, September 16, 2018, Revised Selected Papers, Part I_ , volume 11383 of _Lecture Notes in Computer Science_ , pages 92–104. Springer, 2018a.
* Sheller et al. [2018b] Micah J Sheller, G Anthony Reina, Brandon Edwards, Jason Martin, and Spyridon Bakas. Multi-institutional deep learning modeling without sharing patient data: A feasibility study on brain tumor segmentation. In _International MICCAI Brainlesion Workshop_ , pages 92–104. Springer, 2018b.
* Sozinov et al. [2018] Konstantin Sozinov, Vladimir Vlassov, and Sarunas Girdzijauskas. Human activity recognition using federated learning. In Jinjun Chen and Laurence T. Yang, editors, _IEEE International Conference on Parallel & Distributed Processing with Applications, Ubiquitous Computing & Communications, Big Data & Cloud Computing, Social Computing & Networking, Sustainable Computing & Communications, ISPA/IUCC/BDCloud/SocialCom/SustainCom 2018, Melbourne, Australia, December 11-13, 2018_, pages 1103–1111. IEEE, 2018.
* Stoica et al. [2017] Ion Stoica, Dawn Song, Raluca Ada Popa, David Patterson, Michael W Mahoney, Randy Katz, Anthony D Joseph, Michael Jordan, Joseph M Hellerstein, Joseph E Gonzalez, et al. A berkeley view of systems challenges for ai. _arXiv preprint arXiv:1712.05855_ , 2017.
* Suzumura et al. [2019] Toyotaro Suzumura, Yi Zhou, Natahalie Barcardo, Guangnan Ye, Keith Houck, Ryo Kawahara, Ali Anwar, Lucia Larise Stavarache, Daniel Klyashtorny, Heiko Ludwig, et al. Towards federated graph learning for collaborative financial crimes detection. _arXiv preprint arXiv:1909.12946_ , 2019.
* Tao et al. [2018] Fei Tao, Qinglin Qi, Ang Liu, and Andrew Kusiak. Data-driven smart manufacturing. _Journal of Manufacturing Systems_ , 48:157–169, 2018.
* Tao et al. [2020] Qi Tao, Wu Fangzhao, Wu Chuhan, Huang Yongfeng, and Xie Xing. Fedrec: Privacy-preserving news recommendation with federated learning. _arXiv preprint arXiv:2003.09592_ , 2020.
* Xinle et al. [2019] Liang Xinle, Liu Yang, Chen Tianjian, Liu Ming, and Yang Qiang. Federated transfer reinforcement learning for autonomous driving. _arXiv preprint arXiv:1910.06001_ , 2019.
* Yang et al. [2019a] Qiang Yang, Yang Liu, Tianjian Chen, and Yongxin Tong. Federated machine learning: Concept and applications. _ACM Transactions on Intelligent Systems and Technology (TIST)_ , 10(2):1–19, 2019a.
* Yang et al. [2019b] Wensi Yang, Yuhang Zhang, Kejiang Ye, Li Li, and Cheng-Zhong Xu. Ffd: A federated learning based method for credit card fraud detection. In _International Conference on Big Data_ , pages 18–32. Springer, 2019b.
* Yin [2017] Robert K Yin. _Case study research and applications: Design and methods_. Sage publications, 2017.
* Zhang et al. [2016] Darui Zhang, Bin Xu, and Jasmine Wood. Predict failures in production lines: A two-stage approach with clustering and supervised learning. In _2016 IEEE international conference on big data (big data)_ , pages 2070–2074. IEEE, 2016.
* Zhong et al. [2017] Ray Y Zhong, Xun Xu, Eberhard Klotz, and Stephen T Newman. Intelligent manufacturing in the context of industry 4.0: a review. _Engineering_ , 3(5):616–630, 2017.
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.